Comment on AI’s Memorization Crisis | Large language models don’t “learn”—they copy. And that could change everything for the tech industry.

<- View Parent
TheBlackLounge@lemmy.zip ⁨2⁩ ⁨days⁩ ago

As you can learn from reading the article, they do also store the information itself.

They learn and store a compression algorithm that fits the data, then use it to store that data. The former part of this is not new, AI and compression theory go back decades. What’s new and surprising is that you can get the original work out of attention transformers. Even in traditional overfit models that isn’t a given. And attention transformers shine at generality, so it’s not evident that they should do this, but all models tested do it, so maybe it is even necessary?

Storing data isn’t a theoretical failure, some very useful AI algorithms do it by design. It’s a legal and ethical failure because openai etc have been claiming from the beginning that this isn’t happening, and it also provides proof of the pirated work it’s been trained on.

source
Sort:hotnewtop