Comment on How to use GPUs over multiple computers for local AI?
MalReynolds@slrpnk.net 2 weeks ago
Sure, works fine for inference with tensor parallelism, USB4 / thunderbolt 4/5 is a better (40Gbit+ and already there) bet than ethernet (see distributed-llama). Trash for training / fine tuning, that needs higher inter GPU speed, or better a bigger GPU VRAM.