Comment on No, you don't need a 'very bespoke AOSP' to turn your phone into a Rabbit R1 — here's proof
GlassHalfHopeful@lemmy.ca 6 months agohuggingface! i found it once and never could remember again who hosted all those models. thank you!
i use a different device than homelab, but now I am curious what I may be able to achieve with my syno system. it’s hw weak, so probably not a lot. but i would like to give it a go. if it’s decent, i may consider another device for the purpose.
andrew0@lemmy.dbzer0.com 6 months ago
Good luck! You can try the huggingface-chat repo, or ollama with this web-ui. Both should be decent, as they have instructions to set up a docker container.
I believe the Llama 3 models are out there in a torrent somewhere, but I didn’t dig to find it. For the 70B model, you’ll probably need around 64GB of RAM available, but the 7B one should run fine with just 8GB. It will be somewhat slow though, compared to the ChatGPT experience. The self-attention mechanism can be parallelized, which is why you will see much better results on a GPU. According to some others that tested it, if you offload some stuff to RAM, you could see ~10-12 tokens per second on an RTX 3090 for certain 70B models. But more capable ones will be at less than 1 token per second, all depending on the context window you use.
If you don’t have a GPU available, just give the Phi-3 model a try :D If you quantize it to 4 bits, it can apparently get 12 tokens per second on an iPhone haha. It should play nice with pooling information from a search engine, or a vector database like milvus, qdrant or chroma.
GlassHalfHopeful@lemmy.ca 6 months ago
Thank you for all this! Much appreciated!