Comment on I'm looking for an article showing that LLMs don't know how they work internally
glizzyguzzler@lemmy.blahaj.zone 1 day agoI was channeling the Interstellar docking computer (“improper contact” in such a sassy voice) ;)
There is a distinction between data and an action you perform on data (matrix maths, codec algorithm, etc.). It’s literally completely different.
An audio codec (not a pipeline) is just actually doing math - just like the workings of an LLM. There’s plenty of work to be done after the audio codec decodes the m4a to get to tunes in your ears. Same for an LLM, sandwiching those matrix multiplications that make the magic happen are layers that crunch the prompts and assemble the tokens you see it spit out.
LLMs can’t think, that’s just the fact of how they work. The problem is that AI companies are happy to describe them in terms that make you think they can think to sell their product! I literally cannot be wrong that LLMs cannot think or reason, there’s no room for debate, it’s settled long ago. AI companies will string the LLMs together and let them chew for a while to try make themselves catch when they’re dropping bullshit. It’s still not thinking and reasoning though. They can be useful tools, but LLMs are just tools not sentient or verging on sentient
theunknownmuncher@lemmy.world 1 day ago
Incorrect. You might want to take an information theory class before speaking on subjects like this.
Correct. No one claimed they are “sentient” (you actually mean “sapient”, not “sentient”, but it’s fine as most people mix those up. And no, LLMs are not sapient either, and sapience has nothing to do with reasoning or logic, you’re just moving the goalpost)