Comment on The AI bubble is so big it's propping up the US economy (for now)

<- View Parent
brucethemoose@lemmy.world ⁨2⁩ ⁨days⁩ ago

I am referencing this: z.ai/blog/glm-4.5

The full GLM? Basically a 3090 or 4090 and a budget EPYC CPU. Or maybe 2 GPUs on a threadripper system.

GLM Air? Now this would work on a 16GB+ VRAM desktop, just slap in 96GB+ (maybe 64GB?) of fast RAM. Or the recent Framework desktop, or any mini PC/laptop with the 128GB Ryzen 395 config.

You’d download the weights, quantize yourself if needed, and run them in ik_llama.cpp (which should get support imminently).

github.com/ikawrakow/ik_llama.cpp/

source
Sort:hotnewtop