“Malicious” keywords aren’t exclusively the problem, as the LLM cannot differentiate between “malicious” and “benign”. It’s been trivially easy to intentionally or accidentally hide misinformation in LLMs for a while now. Since they’re black boxes, it could be hard to identify. This is just a slightly more pointed example of data poisoning.
There is no threat to an LLM chatbot outputting text… unless that text is piped into something that can run commands. And who would be stupid enough to do that? Okay, besides vibe coders. And people dumb enough to use AI agents. And people rich enough to stupidly link those AI agents to their bank accounts.
Hond@piefed.social 1 day ago
First shame on OP for clickbaiting. Original title is just: Three clues that your LLM may be poisoned with a sleeper-agent back door
But:
WTF, why discredit your own article right at the beginning? Such a weird line.
TheBat@lemmy.world 1 day ago
That’s The Register for you. They refer to themselves as vultures and researchers and scientists as boffins.
alaphic@lemmy.world 1 day ago
Are you familiar with the term ‘tongue in cheek’? Or ‘hyperbole’? Cuz - I’m just sayin- I really doubt that even the yellow-est of rags would expect people to believe that we’re only a “bite my shiny metal ass” away from triggering a T2 style ‘Judgement Day’… I’d say it’s simply far more likely they were simply being facetious.
Now if it was NewsMax, on the other hand…
Hond@piefed.social 1 day ago
Yeah, i’m familiar with the concept of humor. No worries.
wuffah@lemmy.world 1 day ago
My personal theory is that it lends credibility to the idea that a “rogue AI” will destroy humanity instead of the billionaire broligarchs that wield it to control and surveil the masses.
RalfWausE@feddit.org 1 day ago
Its “The Register”.
CardboardVictim@piefed.social 1 day ago
Also there are three clues but it just explains the process a bit? Very strange article indeed.
hexagonwin@lemmy.sdf.org 1 day ago
kinda feels like they forgot to add ‘/s’