Comment on I've just created c/Ollama!
EncryptKeeper@lemmy.world 1 day agoI’m going to go out in a limb and say they probably just want a comparable solution to Ollama.
Comment on I've just created c/Ollama!
EncryptKeeper@lemmy.world 1 day agoI’m going to go out in a limb and say they probably just want a comparable solution to Ollama.
brucethemoose@lemmy.world 1 day ago
OK.
Then LM Studio. With Qwen3 30B IQ4_XS, low temperature sampling, open web ui frontend if you wish.
That’s what I’m trying to say though, LLMs work a bajillion times better with just a little personal configuration. They are not “one click” magic boxes, they are specialized tools.
Random example: on a Mac? Grab an MLX distillation, it’ll be way faster and better.
Nvidia gaming PC? TabbyAPI with an exl3. Raspberry Pi? That’s important to know!
What do you ask it to do? Set timers? Look at pictures? Cooking recipes? Search the web? Do you need stuff fast or accurate?
This is one reason why ollama is so suboptimal, with the other being just bad defaults (Q4_0 quants, 2048 context, no imatrix or anything outside GGUF, bad sampling last I checked, chat template errors, bugs with certain models, I can go on…)