Comment on Have you tried LocalGPT PrivateGPT or other similar alternatives to ChatGPT?
ReallyActuallyFrankenstein@lemmynsfw.com 10 months agoSecond, I found it very quickly that the amount of RAM you have is critical. My main server is a core i5 4th gen, and so I put AI software on another one of my servers which is a core i5 7th gen. You would think that the latter would work a lot better, but it had half the ram, and it basically wasn’t even able to get started.
Is there an amount of RAM that’s currently considered the bare minimum for CPU-only self-hosting?
exu@feditown.com 10 months ago
If you’re using llama.cpp, have a look at the GGUF models by TheBloke on huggingface. He puts approximate RAM required in the readme based on the quantisation level.
From personal experience I’d estimate 12G for 7B models based on how full RAM was with 16 gigs. For mixtral at least 32G.
ReallyActuallyFrankenstein@lemmynsfw.com 10 months ago
Thanks, appreciate it (I’m new to local text CPU models, I know it was a stupid question).