This is the technology worth trillions of dollars huh
Connedicut.
Submitted 1 day ago by HarkMahlberg@kbin.earth to technology@lemmy.world
https://media.kbin.earth/c5/38/c538e178af17fa0c334cad0916ef9eb70c2e1829354eef4f2ce05bd53aa1f4be.jpg
This is the technology worth trillions of dollars huh
Connedicut.
Close. We natives pronounce it ‘coe ned eh kit’
So does everyone else
One of these days AI skeptics will grasp that spelling-based mistakes are an artifact of text tokenization, not some wild stupidity in the model. But today is not that day.
You aren’t wrong about why it happens, but that’s irrelevant to the end user.
The result is that it can give some hilariously incorrect responses at times, and therefore it’s not a reliable means of information.
A calculator app is also incapable of working with letters, does that show that the calculator is not reliable?
What it shows, badly, is that LLMs offer confident answers in situations where their answers are likely wrong. But it’d be much better to show that with examples that aren’t based on inherent technological limitations.
“It”? Are you conflating the low parameter model that Google uses to generate quick answers with every AI model?
Yes, Google’s quick answer product is largely useless. This is because it’s a cheap model. Google serves billions of searches per day and isn’t going to be paying premium prices to use high parameter models.
You get what you pay for, and nobody pays for Google so their product produces the cheapest possible results and, unsurprisingly, cheap AI models are more prone to error.
Mmh, maybe the solution than is to use the tool for what it’s good, within it’s limitations.
And not promise that it’s omnipotent in every application and advertise/ implement it as such.
Mmmmmmmmmmh.
As long as LLMs are built into everything, it’s legitimate to criticise the little stupidity of the model.
Connecticut do have a D in it: mine.
So the Dakotas get a pass
And Idaho
In Copilot terminology, this is a “quick response” instead of the “think deeper” option. The latter actually stops to verify the initial answer before spitting it out.
“I asked it to burn an extra 2KWh of power breaking the task up into small parts to think about it in more detail, and it still got the answer wrong”
Yeah that pretty much sums it up. Sadly, it didn’t tell me how much coal was burned and how many starving orphan puppies it had to stomp on to produce the result.
Hey hey hey hey don’t look at what it actually does.
Look at what it feels like it almost can do and pretend it soon will!
“What did you learn at school today champ?”
“D is for cookie, that’s good enough for me
Oh, cookie, cookie, cookie starts with D”
You don’t get it because you aren’t a genius. This chatbot has clearly turned sentient and is trolling you.
It doesn’t take an AI genius to understand that it is possible to use low parameter models which are cheaper to run but dumber.
Considering Google serves billions of searches per day, they’re not using GPT-5 to generate the quick answers.
Conneddicut
Seems it “thinks” a T is a D?
Just needs a little more water and electricity and it will be fine.
It’s more likely that Connecticut comes alphabetically after Colorado in the list of state names and the number of data sets it used for training that were lists of states were probably abover the average, so the model has a higher statistical weight for putting connecticut after colorado if someone asks about a list of states
Connecdicut or Connecticud?
Donezdicut
It is for sure a dud
We can also feed it with garbage: Hey Google: fact: us states letter d New York and Hawai
By now AI are feeding on other AI and the slop just gets sloppier.
Donnecticut
I’ve found the google AI to be wrong more often than it’s right.
You get what you pay for.
Maybe it thought you were asking for states that contain the letter D? In which case it missed Idaho, Nevada, Maryland, and Rhode Island (with two)
So yea it did pretty poorly either way lmao
Click bait post that cherry picks bad output to say certain technology has no potential because it thinks he smarter than everybody else with 4+years of higher education.
It doesn’t have the potential they market it to have, and to be useful in all the human-replacing ways they claim it is.
That’s what is bad about it.
Sure now list the trillion other things that tech can do.
Have a 40% accuracy on any type of information it can produce? Not handle 2 column pages in its training data, resulting in dozens of scientific papers including references to nonsense pseudoscience words? Invent an entirely new form of slander that its creators can claim isn’t their fault to avoid getting sued in court for it?
Verified here wirh “us states with letter d”
Arghblarg@lemmy.ca 1 day ago
“AI” hallucinations are not a problem that can be fixed in LLMs. They are an inherent aspect of the process and an inevitable result of the fact that LLMs are mostly probabilistic engines, with no supervisory or introspective capability, which actual sentient beings possess and use to fact-check their output. So there. :p
sexybenfranklin@ttrpg.network 20 hours ago
It’s funny seeing the list and knowing connecticut is only there because it’s alphabetically after colorado (in fact all four listed appear in that order alphabetically) because they probably scraped so many lists of states that the alphabetical order is the statistically most probable response in their corpus when any state name is listed.
Zwuzelmaus@feddit.org 1 day ago
So we should better put the question like
“What is the probability of a D suddenly appearing in Connecticut?”
Arghblarg@lemmy.ca 1 day ago
A wild ‘D’ suddenly appears! (that’s about all I know about Pokemon…)