Comment on [deleted]

<- View Parent
hendrik@palaver.p3x.de ⁨5⁩ ⁨days⁩ ago

AI inference is memory-bound. So, memory bus width is the main bottleneck. I also do AI on an (old) CPU, but the CPU itself is mainly idle and waiting for the memory. I'd say it'll likely be very slow, like waiting 10 minutes for a longer answer. I believe all the AI people use Apple silicon because of the unified memory and it's bus width. Or some CPU with several memory lanes.

source
Sort:hotnewtop