Comment on Very large amounts of gaming gpus vs AI gpus

brucethemoose@lemmy.world ⁨3⁩ ⁨days⁩ ago

Be specific!

In other words, it depends, but the best option for a self hosted rig, OP, is probably:

Now run ik_llama.cpp, and you can server Deepseek 671B faster than you can read: github.com/ikawrakow/ik_llama.cpp

But there’s all sorts of niches. In a nutshell, you don’t think “How much do I need for AI?” But “What is my target use case and model?”

source
Sort:hotnewtop