Comment on Very large amounts of gaming gpus vs AI gpus

starshipwinepineapple@programming.dev ⁨1⁩ ⁨week⁩ ago

The table you’re referencing leaves out CUDA/ tensor cores (count+gen) which is a big part of the gpus, and also not factoring in type of memory. From the comments it looks like you want to use a large MoE model. You aren’t going to be able to just stack raw power and expect to be able to run this without major deterioration of performance if it runs at all.

Don’t forget your MoE model needs all-to-all communication for expert routing

source
Sort:hotnewtop