These language models don’t get the meaning of anything. They predict the next cluster of letters based on the clusters of letters that have come before. Sorry, but if it feels to you like they’re captured the meaning of something, you’re being bamboozled
brucethemoose@lemmy.world 1 week ago
It’s a metaphor.
They’re translating the input tokens to intent in the model’s middle layers, which is a bit more precise.