Comment on Do you think Google execs keep a secret un-enshittified version of their search engine and LLM?

<- View Parent
partial_accumen@lemmy.world ⁨1⁩ ⁨day⁩ ago

Its also possible we’ve reached the limits of the training data.

This is my thinking too. I don’t know how to solve the problem either because datasets created after about 2022 likely are polluted with LLM results baked in. With even a 95% precision that means 5% hallucination baked into the dataset. I can’t imagine enough grounding is possible to mitigate that. As the years go forward the problem only gets worse because more LLM results will be fed back in as training data.

source
Sort:hotnewtop