Comment on Lies, Damned Lies, and LLMs: AI is a Con
auraithx@lemmy.dbzer0.com 1 week ago
Yes this was a specific problem with Gemini. They obviously tried to over correct for hallucinations and being too gullible, but it ended up making it certain of its hallucinations.
Hallucination rate for their latest model is 0.7%
github.com/vectara/hallucination-leaderboard
Should be <0.1% within a year
db0@lemmy.dbzer0.com 1 week ago
Hallucinations when summarizing are significantly lower than when generating code (since the original document would be in context)