Comment on How to use GPUs over multiple computers for local AI?

<- View Parent
Natanox@discuss.tchncs.de ⁨1⁩ ⁨week⁩ ago

May take a look at systems with the newer AMD SoC’s first. They utilize the systems’ RAM and come with a proper NPU, once ollama or mistral.rs are supporting those they might give you sufficient performance for your needs for way lower costs (incl. power consumption). Depending on how NPU support gets implemented it might even become possible to use NPU and GPU in tandem, that would probably enable pretty lowerful models to be run on consumer-grade hardware at reasonable speed.

source
Sort:hotnewtop