Let’s be honest though the majority of politicians are so terrible at their job, that this might actually be one of the rare occurrences where AI actually improves the work. But it is very susceptible to unknown influences.
‘We didn’t vote for ChatGPT’: Swedish Prime Minister under fire for using AI
Submitted 1 day ago by Davriellelouna@lemmy.world to technology@lemmy.world
Comments
Decq@lemmy.world 19 hours ago
breecher@sh.itjust.works 17 hours ago
Fuck no. Rather an incompetent politician than a hallucinating sycophant just telling you what you want to hear.
liuther9@feddit.nl 14 hours ago
Nah you are wrong and should use AI as a second opinion
Decq@lemmy.world 17 hours ago
I’m just making an objective observation. I don’t condone it. I rather we just have competent politicians. But it seems only people who can’t function elsewhere are drawn to the position…
squaresinger@lemmy.world 18 hours ago
That’s the big issue. If it was only about competence, I think throwing dice might yield better results than what many politicians are doing. But AI isn’t throwing dice but instead reproduces what the creators of the AI want to say.
interdimensionalmeme@lemmy.ml 18 hours ago
Creators of AI don’t quite have the technology to puppeteer their AI like this.
They can selects the input, they can bias the training, but if the model isn’t going to be lobotomized coming out
then they can’t really bend it toward any particular one opinionI’m sure in the future they’ll be able to adjust advertising manipulation in real time but not yet.
What is really sketchy is states and leaders relying on commercial models instead of public ones
I think states should train public models and release them for the public good
if only to undermine big tech bros and their nefarious influenceAnUnusualRelic@lemmy.world 17 hours ago
Depending on the AI, it will conclude that he ought to buy a new phone charger, deport all the foreigners, kill all the Jews or rewrite his legislation in Perl. It’s hard to say without more information.
caveman8000@lemmy.world 13 hours ago
Meanwhile the American president uses no intelligence at all. Artificial or otherwise
drmoose@lemmy.world 6 hours ago
Oh no man does research of course Americans are upset here lmao
HubertManne@piefed.social 1 day ago
I really don't get it. These things are brand new. How can anyone get so into these things so quickly. I don't take advice from people I barely know, much less ones that can be so easily and quickly reprogrammed.
kamenlady@lemmy.world 1 day ago
This is the unintentional uncanny valley for me in AI.
I ( was forced to ) use chatGTP for work. It can talk about everything and sounds very confident and seems reliable to always come up with something to help you solve your problems.
You talk with it about some niche content and suddenly have an ardent fan of said niche content responding. It surely knows every little bit of info of that niche and surprises you with funny, but apt quotes from your favorite show in the middle of conversations about something else.
This is just from a tiny bit of interaction, while at work.
I can imagine people completely overwhelmed, by having their thoughts confirmed and supported by something that seems so intelligent, responsive and remembers all your conversations. It literally remembers each word.
For many people it may be the first time in their life, that they experienced a positive response to their thoughts. Not only that, they also found someone eager to talk with you about it.
HubertManne@piefed.social 1 day ago
Everyones initial use of chatbots should be on the thing they are most knowledgable about so they can get an idea of how wrong it can be and how it can be useful but you have to treat it like some eager wet behind the ears intern just did for you.
greybeard@feddit.online 17 hours ago
One thing I struggle with AI is the answers it gives always seem plausable, but any time I quiz it on things I understand well, it seems to constantly get things slightly wrong. Which tells me it is getting everything slightly wrong, I just don't know enough to know it.
I see the same issue with TV. Anyone who works in a compicated field has felt the sting of watching a TV show fail to accurate represent it while most people watching just assume that's how your job works.
clif@lemmy.world 7 hours ago
Something I found today - ask it for the lyrics of your favorite song/artist. It will make something up based on the combination of the two and maybe a little of what it was trained on… Even really popular songs (I tried a niche one by Angelspit first then tried “Sweet Caroline” for more well known)
The new gpt-oss model replies with (paraphrased) “I can’t do that because it is copyrighted material” which I have a sneaking suspicion is intentional so there’s an excuse for not showing a very wrong answer to people who might start to doubt it’s ““intelligence””
… Like they give a flying fuck about copyright.
HubertManne@piefed.social 15 hours ago
This is where you have to check out the reference links it gives as if they were search results and the less you know the more you have to do it. I mean people have been webMDing for a long time. None of these things allow folks to stop critical thinking. If anything it requires it even more. This was actually one of my things with ai and work. The idea is for it to allow people with less knowledge to do things and to me its kinda the reverse.
FishFace@lemmy.world 14 hours ago
Because that’s what it is really trained for: to produce correct grammar and plausible sentences. It’s really an unbelievable leap from computer-generated text from preceding approaches where, in a matter of a few years, you went from little more than gibberish to stuff that’s so incredibly realistic that it can be mistaken for intelligent conversation, easily passing the Turing Test (I had to actually go to Wikipedia to check and, indeed, this was verified this year - note that this in particular is for recent models)
So you have something that is sufficiently realistic that it can appear to be a human conversation partner. Human beings aren’t (yet) well-equipped to deal with something which appears to be human but whose behaviour diverges from typical human behaviour so radically (most relevantly, it won’t readily admit to not knowing something).
HubertManne@piefed.social 14 hours ago
Its more than that. It takes the input and tries to interpret the bad grammar and sentences into search terms and finds link the correlate the highest to its interpretation and then gives back the response that summarizes the results with good grammar and plausible sentences. Again this is why I stress that you have to evaluate its response and sources. The sources are the real value in any query. Im actually not sure how much the chatbots give sources by default though as I know I have not gotten them and then asked for them and now I get them as a matter of course so im not sure if it learns that I want them or if they did a change to provide them when they had not before.
Perspectivist@feddit.uk 1 day ago
Anyone who has an immediate kneejerk reaction the moment someone mentions AI is no better than the people they’re criticizing. Horseshoe theory applies here too - the most vocal AI haters are just as out of touch as the people who treat everything an LLM says as gospel.
lime@feddit.nu 23 hours ago
here’s my kneejerk reaction: my prime minister is basing his decisions partly on the messages of an unknown foreign actor, and sending information about state internals to that unknown foreign actor.
whether it’s ai or not is a later issue.
Redex68@lemmy.world 19 hours ago
He explicitly states that no sensitive informarion gets used. If you believe that, then I have no issue with him additionally asking for a third opinion from an LLM.
audaxdreik@pawb.social 23 hours ago
Absolutely incorrect. Bullshit. And horseshoe theory itself is lately bullshit.
(Succinct response taken from Reddit post discussing the topic)
“Horseshoe Theory is slapping “theory” on a strawman to simplify WHY there’s crossover from two otherwise conflicting groups. It’s pseudo-intellectualizing it to make it seem smart.”
This ignores the many, many reasons we keep telling you why we find it dangerous, inaccurate, and distasteful. You don’t offer a counter argument in your response so I can only assume it’s along the lines of, “technology is inevitable, would you have said the same if the Internet?” Which is also a fallacious argument. But go ahead, give me something better if I assume wrong
I can easily see why people would be furious they’re elected leader is abdicating thought and responsibility to an often wrong, unaccountably biased chat bot.
Furthermore, your insistance continues to push an acceptance of AI on those who clearly don’t want it, contributing to the anger we feel at having it forced upon us
Perspectivist@feddit.uk 22 hours ago
You opened with a flat dismissal, followed by a quote from Reddit that didn’t explain why horseshoe theory is wrong - it just mocked it. That’s not an argument, that’s posturing.
From there, you shifted into responding to claims I never made. I didn’t argue that AI is flawless, inevitable, or beyond criticism. I pointed out that reflexive, emotional overreactions to AI are often as irrational as the blind techno-optimism they claim to oppose. That’s the context you ignored.
You then assumed what I must believe, invited yourself to argue against that imagined position, and finished with vague accusations about me “pushing acceptance” of something people “clearly don’t want.” None of that engages with what I actually said.
RememberTheApollo_@lemmy.world 23 hours ago
If someone says they got a second opinion from a physician known for being wrong half the time would you not wonder why they didn’t choose someone more reliable for something as their health? AI is notorious for providing incomplete, irrelevant, heavily slanted, or just plain wrong info. Why give it any level of trust to make national decisions? Might as well, I dunno…use a bible? Some would consider that trustworthy.
Perspectivist@feddit.uk 23 hours ago
I often ask ChatGPT for a second opinion, and the responses range from “not helpful” to “good point, I hadn’t thought of that.” It’s hit or miss. But just because half the time the suggestions aren’t helpful doesn’t mean it’s useless. It’s not doing the thinking for me - it’s giving me food for thought.
The problem isn’t taking into consideration what an LLM says - the problem is blindly taking it at its word.
yumyumsmuncher@feddit.uk 15 hours ago
Politicians and CEOs should be replaced with LLMs
Warl0k3@lemmy.world 14 hours ago
It can’t be any worse…
roofuskit@lemmy.world 1 day ago
It’s weird for a head of state to consult their mentally challenged imaginary friend?
Medic8teMe@lemmy.ca 1 day ago
William MacKenzie King, the longest serving Prime Minister in Canada used to commune with spirits via psychic mediums including those of his dead dogs.
I agree it’s weird.
MNByChoice@midwest.social 1 day ago
Didn’t Nancy Regan, wife of former USA President Ronald Regan, did this as well. (Ronald was apparently not mentally fit for the last few years as well.)
surewhynotlem@lemmy.world 1 day ago
Bad news friend. The number of atheist heads of state is surprisingly low.
roofuskit@lemmy.world 1 day ago
raspberriesareyummy@lemmy.world 6 hours ago
What a treasonist piece of shit.
tal@lemmy.today 1 day ago
“You have to be very careful,” Simone Fischer-Hübner, a computer science researcher at Karlstad University, told Aftonbladet, warning against using ChatGPT to work with sensitive information.
I mean, sending queries to a search engine or an LLM are about the same in terms of exposing one’s queries.
If the guy were complaining about information from an LLM not being cited or something, then I think I could see where he was coming from more.
j4yt33@feddit.org 23 hours ago
It’s a woman
alvyn@discuss.tchncs.de 22 hours ago
I’m not against the technology, I’m against people who runs it. I have problem with how they teach their LLMs on code, user data, music, books, webs all without author’s / user’s consent and worse even with authors / users explicit NO consent to scrape or to use it for learning. Another level is lack of security - ChatGPT chats available to everyone. Deep fakes everywhere, just see the latest Taylor Swift one. Sorry, but fuck you with all of this. There is lack of basic security, privacy and ignoring all of its danger. Only what that fucking AI firms want is easy, cheep and quick money. All that hype for nothing = means you cannot even rely on the output.
foenkyfjutschah@programming.dev 22 hours ago
yet you need these masses of input for the technology to exist. the business models that base on the technology aren’t sustainable even without payment of the input data.
Darkenfolk@sh.itjust.works 20 hours ago
Do we really need this technology to exist though? It’s unreliable and very niche as far as I have seen.
People say that it speeds up certain tasks, but it’s so unreliable that you need to error-check the whole thing afterwards.
Blackmist@feddit.uk 17 hours ago
Maybe it shouldn’t be a business model then.
alvyn@discuss.tchncs.de 22 hours ago
Of common, you justifying stealing by this bullshit?
UnderpantsWeevil@lemmy.world 1 day ago
It’s literally a cult.
UnfortunateShort@lemmy.world 1 day ago
It surely can’t hurt, if it’s to sanity check your highly paid advisors…
jonne@infosec.pub 1 day ago
Except those prompts are retained by OpenAI, and you don’t know who’s got access to that. They’ve had chats leak before.
Allemaniac@lemmy.world 22 hours ago
europe is fucking doomed
Sp00kyB00k@lemmy.world 19 hours ago
Because of this one incident. Good how you figured it out. So much smarter than the rest. … Get. out.
some_guy@lemmy.sdf.org 1 day ago
What a stupid fucking thing to admit. Let’s hope he’s out on his ass next election.
doctortofu@piefed.social 1 day ago
I would genuinely have more trust in a politician consulting I Ching than ChatGPT - at least the book of changes requires some thinking and interpretation, fucking hell...
Sigh, all the tech-bros that convinced the general public that AI are in any way intelligent or that they can think, reason or understand concepts like truth/lies or right/wrong should go on trial for crimes against humanity...
hisao@ani.social 1 day ago
Most people don’t care whether AI is intelligent, can it think or reason or understand concepts. What matters is that it can give helpful replies, and it does that a lot. In my experience maybe 1 reply out of 10 is seriously flawed, others are either mostly helpful or just tell me something I already knew until I reprompt for more, which, again, also works well most of the time (especially when you allow it to search for more information online). So if you wanted to say it’s dangerous in some ways, this is definitely not the proper way to say it, since neither it being dangerous nor it being right or wrong or helpful or useless has anything to do with intelligence, ability to think, reason, feel, comprehend or whatever.
kokesh@lemmy.world 1 day ago
You can see it on this government policies
Fleur_@aussie.zone 1 day ago
“That’s right voters I’m spineless and have no original ideas” -every politician
AbouBenAdhem@lemmy.world 1 day ago
The typical pattern for leaders is to get “second opinions” from advisors who end up telling them whatever they want to hear, so… maybe asking the equivalent of a magic 8 ball is a marginal improvement?
RobotZap10000@feddit.nl 1 day ago
I would rather have the politicians consult a plain old magic 8 ball than one controlled by Scam Altman.