Basicly I can run 9b models on my 16gb gpu mostly fine like getting responses of lets say 10 lines in a few seconds.
Bigger models if they don’t outright crash take for the same task then like 5x or 10x longer so long it isn’t even useful anymore
So very worse.