Comment on How to use GPUs over multiple computers for local AI?

MalReynolds@slrpnk.net ⁨2⁩ ⁨weeks⁩ ago

Sure, works fine for inference with tensor parallelism, USB4 / thunderbolt 4/5 is a better (40Gbit+ and already there) bet than ethernet (see distributed-llama). Trash for training / fine tuning, that needs higher inter GPU speed, or better a bigger GPU VRAM.

source
Sort:hotnewtop