Comment on AI models collapse when trained on recursively generated data
FaceDeer@fedia.io 3 months agoYou realize that those "billions of dollars" have actually resulted in a solution to this? "Model collapse" has been known about for a long time and further research figured out how to avoid it. Modern LLMs actually turn out better when they're trained on well-crafted and well-curated synthetic data.
Honestly, everyone seems to assume that machine learning researchers are simpletons who've never used a photocopier before.