Comment on Llama 3.1 AI Models Have Officially Released

<- View Parent
admin@lemmy.my-box.dev ⁨2⁩ ⁨months⁩ ago

Oof - not on my 12gb 3060 it doesn’t :/ Even at 48k context and the Q4_K quantization, it’s ollama its doing a lot of offloading to the cpu. What kind of hardware are you running it on?

source
Sort:hotnewtop