Comment on The Irony of 'You Wouldn't Download a Car' Making a Comeback in AI Debates

<- View Parent
joshcodes@programming.dev ⁨3⁩ ⁨months⁩ ago

I think you’re anthropomorphising the tech tbh. It’s not a person or an animal, it’s a machine and cramming doesn’t work in the idea of neural networks. They’re a mathematical calculation over a vast multidimensional matrix, effectively solving a polynomial of an unimaginable order. So “cramming” as you put it doesn’t work because by definition an LLM cannot forget information because once it’s applied the calculations, it is in there forever. That information is supposed to be blended together. You have overfitting which would be inputting similar information and performing the similar calculations, and it would exhibit poor performance should it be asked do anything different to the training.

What Im arguing over here is language rather than a system so let’s do that and note the flaws. If we’re being intellectually honest we can agree that a flaw like that doesn’t represent true learning and shows a reliance on the training data, i.e. it cant learn unless it has seen similar data before and it cant find a way to learn better based on repeating unintelligible input.

In the example, it has statistically inferred that those are all the correct words to repeat in that order based on the prompt. This isn’t akin to anything human, people can’t repeat pages of text verbatim like this and no toddler can be tricked into repeating a random page from a random book as you say. The data is there, it’s encoded and referenced when the probability is high enough. As another commenter said, language itself is a powerful tool of rules and stipulations that provide guidelines for the machine, but it isn’t crafting its own sentences, it’s using everyone else’s.

Also, calling it “tricking the AI” isn’t really intellectually honest either, as in “it was tricked into exposing it still has the data encoded”. We can state it isn’t preferred or intended behaviour but the system still exhibits a reliance on the training data and the ability to replicate it almost exactly and therefore it is factually wrong to state that it doesn’t keep the training data in a usable format - which was my original point.

source
Sort:hotnewtop