The problem is that itâs not actually counting anything. Itâs simply looking for some text somewhere in its database that relates to that word and the number of Râs in that word. Thereâs no mechanism within the LLM to actually count things. It is not designed with that function. This is not general AI, this is a Generative Adversarial Network thatâs using its vast vast store of text to put words together that sound like they answer the question that was asked.
Comment on AGI achieved đ¤
__dev@lemmy.world â¨3⊠â¨days⊠agoAnd yet they can seemingly spell and count (small numbers) just fine.
buddascrayon@lemmy.world â¨2⊠â¨days⊠ago
jsomae@lemmy.ml â¨3⊠â¨days⊠ago
what do you mean by spell fine? Theyâre just emitting the tokens for the words. Like, itâs not writing âstrawberry,â itâs writing tokens <302, 1618, 19772>, which correspond to st, raw, and berry respectively. If you ask it to put a space between each letter, that will disrupt the tokenization mechanism, and itâs going to be quite liable to making mistakes.
I donât think itâs really fair to say that the lookup 19772 -> berry counts as the LLM being able to spell, since the LLM isnât operating at that layer. It doesnât really emit letters directly. I would argue its inability to reliably spell words when you force it to go letter-by-letter or answer queries about how words are spelled is indicative of its poor ability to spell.
__dev@lemmy.world â¨2⊠â¨days⊠ago
I mean that when you ask them to spell a word they can list every character one at a time.
jsomae@lemmy.ml â¨2⊠â¨days⊠ago
Well thatâs a recent improvement. GPT3 was very bad at that, and GPT4 still makes mistakes.