Comment on I'm looking for an article showing that LLMs don't know how they work internally

<- View Parent
adespoton@lemmy.ca ⁨5⁩ ⁨days⁩ ago

No, they really don’t. It’s a large language model. Input cues instruct it as to which weighted path through the matrix to take. Those paths are complex enough that the human mind can’t hold all the branches and weights at the same time. But there’s no planning going on; the model can’t backtrack a few steps, consider different outcomes and run a meta analysis. Other reasoning models can do that, but not language models; language models are complex predictive translators.

source
Sort:hotnewtop