Comment on What can I use for an offline, selfhosted LLM client, pref with images,charts, python code execution
TMP_NKcYUEoM7kXg4qYe@lemmy.world 2 weeks ago
You can tell Open Interpreter to run commands based on you human-language input. If you want local only LLM, you can pair it with Ollama. It works for “interactive” use where you’re asked for confirmation before a command is run.
I set this up in a VM because I wanted a full automatic coding “agent” which can run commands without my intervention and I did not want it to blow up main system. It did not really work though because as far as I know Open Interpreter does not have a way to “pipe” a command’s output back into the LLM so that it could create feedback with linters and stuff.
Another issue was that Starcoder2, which is the only LLM trained on permissive licensed code I could find, only has a 15B “human-like” model. The smaller models only speak code so I don’t know how that would work for agentic usage and the 15B is really slow running on DDR4 CPU. I think agents are cool though so I would like to try Aider which is a supposedly good open source agent and unlike Open Interpreter is not abandonware.
Thanks for coming to my blabering talk, hope this might be useful for someone.