I’ve just re-discovered ollama and it’s come on a long way and has made a difficult task of locally hosting your own LLM to simply installing a deb! It also works for Windows and Mac, so can help everyone.
I’d like to see Lemmy become useful for specific technical sub branches instead of trying to find the best existing community, so I created c/ollama for everyone to discuss, ask questions, and help each other out with ollama!
So, please, join, subscribe and feel free to post, ask questions, post tips / projects, and help out where you can!
Thanks!
brucethemoose@lemmy.world 1 day ago
TBH you should fold this into localllama? Or open source AI?
I have very mixed (mostly bad) feelings on ollama. In a nutshell, they’re kinda Twitter attention grabbers that give zero credit/contribution to the underlying framework (llama.cpp). It’s also a highly suboptimal way for most people to run LLMs, especially if you’re willing to tweak.
They’re… slimey. I would always recommend Kobold.cpp, tabbyAPI, ik_llama.cpp, Aphrodite, any number of backends over them. Anything but ollama.
TheHobbyist@lemmy.zip 1 day ago
Indeed, Ollama is going a shady route. github.com/ggml-org/llama.cpp/pull/11016#issuecom…
I started playing with Ramalama (the name is a mouthful) and it works great. There is one or two more steps in the setup but I’ve achieved great performance and the project is making good use of standards (OCI, jinja, unmodified llama.cpp, from what I understand).
Go and check it out, they are compatible with models from HF and Ollama too.
github.com/containers/ramalama
southernbeaver@lemmy.world 1 day ago
What would you recommend to hook to my home assistant?
brucethemoose@lemmy.world 1 day ago
Totally depends on your hardware, and what you tend to ask it. What are you running?
TheHobbyist@lemmy.zip 1 day ago
Perhaps give Ramalama a try?
github.com/containers/ramalama
tal@lemmy.today 1 day ago
While I don’t think that llama.cpp is specifically a risk, I think that running generative AI software in a container is probably a good idea. It’s a rapidly-moving field with a lot of people contributing a lot of code that very quickly gets run on a lot of systems by a lot of people. There’s been malware that’s shown up in extensions for (for example) ComfyUI. And the software really doesn’t need to poke around at outside data.
Also, because the software has to touch the GPU, it needs a certain amount of outside access. Containerizing that takes some extra effort.
old.reddit.com/…/psa_please_secure_your_comfyui_i…
Ollama means sticking it in a Docker container, and that is, I think, a positive thing.
If there were a close analog, like some software package that could take a given LLM model and run in podman or Docker or something, I think that that’d be great. But I think that putting the software in a container is probably a good move relative to running it uncontainerized.
brucethemoose@lemmy.world 1 day ago
I don’t understand.
Ollama is not actually docker, right? It’s running the same llama.cpp engine, it’s just embedded inside the wrapper app, not containerized.
And basically every LLM project ships a docker container. I know for a fact llama.cpp, TabbyAPI, Aphrodite, vllm and sglang do.
You are 100% right about security though, in fact there’s a huge concern with compromised Python packages. This one almost got me: pytorch.org/blog/compromised-nightly-dependency/