Comment on Apple Develops Breakthrough Method for Running LLMs on iPhones
guitarsarereal@sh.itjust.works 10 months ago
Everyone likes to trash machine learning because the power requirements are high, but what they don’t realize is that we’re in the very first days of this technology. Every technology that got bundled together into your phone was equally as useless when it was first invented.
Literally once a week, I see some news story about AI researchers delivering an order of magnitude speedup in some aspect of AI inference. The technique described here apparently allows for a 20x speedup on GPU’s.
cybersandwich@lemmy.world 10 months ago
Whispercpp works off the ML cores on the m series chips. It’s faster than my 1080ti that I have in a server doing the same things–by orders of magnitude. And it sips power.
Purpose built chips can be super powerful for their specific purposes.