Comment on I've just created c/Ollama!

<- View Parent
brucethemoose@lemmy.world ⁨1⁩ ⁨day⁩ ago

8GB?

You might be able to run Qwen3 4B: huggingface.co/mlx-community/…/main

But honestly you don’t have enough RAM to spare, and even a small model might bog things down. I’d run Open Web UI or LM Studio with a free LLM API, like Gemini Flash, or pay a few bucks for something off openrouter. Or maybe Cerebras API.

source
Sort:hotnewtop