Comment on ChatGPT's new browser has potential, if you're willing to pay

<- View Parent
brucethemoose@lemmy.world ⁨1⁩ ⁨day⁩ ago

Not anymore. I can run GLM 4.6 on a Ryzen/single RTX 3090 at 7 tokens/s, and it runs rings around most API models. I can run 14-49Bs in more utilitarian cases that do just fine.

But again, it’s all ‘special interest tinkerer’ tier. You can’t do ollama run, you have to mess with exotic libraries and setups to squeeze out that kind of performance.

source
Sort:hotnewtop