Of course not, the whole point of disinformation is that it sounds correct, that’s AI’s bread and butter!
'Garbage in, garbage out': AI fails to debunk disinformation, study finds.
Submitted 4 weeks ago by Dot@feddit.org to technology@lemmy.world
Comments
Grandwolf319@sh.itjust.works 4 weeks ago
semperverus@lemmy.world 4 weeks ago
I dont think that the developers who came up with the processes for LLMs were really targeting that usecase. It just so happens that the limitations of LLMs also lend power to disinformatiom. LLMs don’t really know how to say “I don’t know”.
BertramDitore@lemm.ee 4 weeks ago
Think about it this way: remember those upside-down answer keys in the back of your grade school math textbook? Now imagine if those answer keys included just as many incorrect answers as correct ones. How would you know if you were right or wrong without asking your teacher? Until a LLM can guarantee a right answer, and back it up with real citations, it will continue to do more harm than good.
tehn00bi@lemmy.world 4 weeks ago
Certainly plausible. I’m sure they are trying to figure out how to get it to understand relationships between information now that it’s pretty good at statistical inference.
Zerlyna@lemmy.world 4 weeks ago
Supposedly ChatGPT had an update in September, but it doesn’t agree that Trump was found guilty in may 34 times. When I give it sources it says ok, but it doesn’t upload correct learned information.
Jrockwar@feddit.uk 4 weeks ago
That’s because it doesn’t learn, it’s a snapshot of its training data frozen in time.
I like Perplexity (a lot) because instead of using its data to answer your question, it uses your data to craft web searches, gather content, and summarise it into a response. It’s like a student that uses their knowledge to look for the answer in the books, instead of trying to answer from memory whether they know the answer or not.
It is not perfect, it does hallucinate from time to time, but it’s rare enough that I use it way more than regular web searches at this point. I can throw quite obscure questions at it and it will dig the answer for me.
As someone with ADHD with a somewhat compulsive need to understand random facts (e.g. “I need to know right now how the motor speed in a coffee grinder affects the taste of the coffee”) this is an absolute godsend.
I’m not affiliated or anything, and if anything better comes my way I’ll be happy to ditch it. But for now I really enjoy it.
Zerlyna@lemmy.world 4 weeks ago
Yes but I’m saying that snapshot in September is incorrect. Why is that. Is it rigged?
lemmeBe@sh.itjust.works 4 weeks ago
My experience is the same.
pennomi@lemmy.world 4 weeks ago
I think the next step in AI is learning how to control and direct the speech, rather than just make computers talk.
They are surprisingly good for being a mere statistical copycat of words on the internet. Whatever the second tier innovation is that jumps AI into true reasoning rather than pattern matching is going to be wild.
TommySoda@lemmy.world 4 weeks ago
THX1138@lemmy.ml 4 weeks ago
Shock pickachu
Hackworth@lemmy.world 4 weeks ago
In its latest audit of 10 leading chatbots, compiled in September, NewsGuard found that AI will repeat misinformation 18% of the time
MajorHavoc@programming.dev 4 weeks ago
I love that someone even bothered with a study.
jeansburger@lemmy.world 4 weeks ago
Confirmation of anecdotes or gut feelings is still science. At some point you need data rather than experience to help people and organizations change their perception (see: most big tech companies lighting billions of dollars on fire on generative AI).
GoodEye8@lemm.ee 4 weeks ago
Not to mention based on the numbers in the article I imagine the AI might actually do better than an average human would do. It wasn’t as much of a “duh” as I thought it would be.
MajorHavoc@programming.dev 4 weeks ago
Agreed!
I don’t mean sarcasticly, honestly. As you said, it’s still valuable science.
homesweethomeMrL@lemmy.world 4 weeks ago
That’s true. But still. Duh.
barsoap@lemm.ee 4 weeks ago
You also need that stuff to shut up pseudo-sceptics. Like, random example, posture having an influence on mood, there were actually psychologists denying that because either a) If there’s no study on some effect then it doesn’t exist, “literature realism” or b) some now-debunked theory of the past implied it, “incorrectness by association”. Just because you’re an atheist doesn’t mean that you should discount catholic opinions on beer brewing, they produce some good shit.
TheGrandNagus@lemmy.world 4 weeks ago
For many hundreds of years, blood-letting was an obvious thing to do. As was just giving people leeches for medical ailments. And ingesting mercury. We thought having sex with virgins would cure STDs. We thought doses of radiation was good for us. And tobacco. We thought it was obvious that the sun revolved around Earth.
It is enormously important to scientifically confirm things, even if they do seem obvious.
Flocklesscrow@lemm.ee 4 weeks ago
Uhh, we didn’t “think” a lot of those things- you’re describing marketing that some Company disseminated in order to shill their products. And many, many people paid the price in misery, or worse.