Comment on The AI bubble is so big it's propping up the US economy (for now)
brucethemoose@lemmy.world 2 days agoI am referencing this: z.ai/blog/glm-4.5
The full GLM? Basically a 3090 or 4090 and a budget EPYC CPU. Or maybe 2 GPUs on a threadripper system.
GLM Air? Now this would work on a 16GB+ VRAM desktop, just slap in 96GB+ (maybe 64GB?) of fast RAM. Or the recent Framework desktop, or any mini PC/laptop with the 128GB Ryzen 395 config.
You’d download the weights, quantize yourself if needed, and run them in ik_llama.cpp (which should get support imminently).