It wouldn’t even matter. OP doesn’t understand how any of this works, and is instead just running rampant calling everything bullshit 😂
Comment on How to use GPUs over multiple computers for local AI?
Natanox@discuss.tchncs.de 1 week agoMay take a look at systems with the newer AMD SoC’s first. They utilize the systems’ RAM and come with a proper NPU, once ollama or mistral.rs are supporting those they might give you sufficient performance for your needs for way lower costs (incl. power consumption). Depending on how NPU support gets implemented it might even become possible to use NPU and GPU in tandem, that would probably enable pretty lowerful models to be run on consumer-grade hardware at reasonable speed.
just_another_person@lemmy.world 1 week ago
marauding_gibberish142@lemmy.dbzer0.com 1 week ago
I’d prefer that you reply with examples/an explanation of what I’m doing wrong instead of cursing
just_another_person@lemmy.world 1 week ago
Look at all your replies, chum. Everyone is already telling you.
marauding_gibberish142@lemmy.dbzer0.com 1 week ago
Thanks, but will NPUs integrated along with the CPU ever match the performance of a discrete GPU?
Natanox@discuss.tchncs.de 1 week ago
Depends on which GPU you compare it with, what model you use, what kind of RAM it has to work with, ecetera. NPU’s are purpose-built chips after all. Unfortunately the whole tech is still very young, so we’ll have to wait for stuff like ollama to introduce native support for an apples-to-apples comparison. The raw numbers to however do look promising.