As a general rule of thumb, you need about 1 GB per 1B parameters, so you’re looking at about 405 GB for the full size of the model.
Quantization can compress it down to 1/2 or 1/4 that, but “makes it stupider” as a result.
Comment on The first GPT-4-class AI model anyone can download has arrived: Llama 405B
chiisana@lemmy.chiisana.net 3 months agoWhat’s the resources requirements for the 405B model? I did some digging but couldn’t find any documentation during my cursory search.
As a general rule of thumb, you need about 1 GB per 1B parameters, so you’re looking at about 405 GB for the full size of the model.
Quantization can compress it down to 1/2 or 1/4 that, but “makes it stupider” as a result.
modeler@lemmy.world 3 months ago
Typically you need about 1GB graphics RAM for each billion parameters (i.e. one byte per parameter). This is a 408B parameter model. Ouch.
cheddar@programming.dev 3 months ago
Image
Deceptichum@quokk.au 3 months ago
Or you could run it via cpu and ram at a much slower rate.
errer@lemmy.world 3 months ago
Yeah uh let me just put in my 512GB ram stick…
Deceptichum@quokk.au 3 months ago
Samsung do make them.
Goodluck finding 512gb of VRAM.
bruhduh@lemmy.world 3 months ago
www.ebay.com/p/116332559 lga2011 motherboards quite cheap, insert 2 xeon 2696v4 44 threads each totalling at 88 threads and 8 ddr4 32gb sticks, it comes quite cheap actually, you can also install Nvidia p40 with 24gb each, you can max out this build for ai for under 2000$
chiisana@lemmy.chiisana.net 3 months ago
Finally! My dumb dumb 1TB ram server (4x E5-4640 + 32x32GB DDR3 ECC) can shine.
Siegfried@lemmy.world 3 months ago
At work we habe a small cluster totalling around 4TB of RAM.
TipRing@lemmy.world 3 months ago
When the 8 bit quants hit, you could probably lease a 128GB system on runpod.
1984@lemmy.today 3 months ago
Can you run this in a distributed manner, like with kubernetes and lots of smaller machines?
obbeel@lemmy.eco.br 3 months ago
According to huggingface, you can run a 34B model using 22.4GBs of RAM max. That’s a RTX 3090 Ti.
Longpork3@lemmy.nz 3 months ago
Hmm, I probably have that much distributed across my network… maybe I should look into some way of distributing it across multiple gpu
arefx@lemmy.ml 3 months ago
Ypu mean my 4090 isn’t good enough 🤣😂