Comment on Very large amounts of gaming gpus vs AI gpus

<- View Parent
brucethemoose@lemmy.world ⁨3⁩ ⁨days⁩ ago

One last thing, I’ve heard mixed things about 235B, hence there might be a smaller, more optimal LLM for whatever you do, if it’s something targeted?

For instance, Kimi 72B is quite a good coding model: huggingface.co/moonshotai/Kimi-Dev-72B

It might fit in vllm (as an AWQ) with 2x 4090s. It and would easily fit as an exl3.

source
Sort:hotnewtop