“Malicious” keywords aren’t exclusively the problem, as the LLM cannot differentiate between “malicious” and “benign”. It’s been trivially easy to intentionally or accidentally hide misinformation in LLMs for a while now. Since they’re black boxes, it could be hard to identify. This is just a slightly more pointed example of data poisoning.
There is no threat to an LLM chatbot outputting text… unless that text is piped into something that can run commands. And who would be stupid enough to do that? Okay, besides vibe coders. And people dumb enough to use AI agents. And people rich enough to stupidly link those AI agents to their bank accounts.
Hond@piefed.social 3 weeks ago
First shame on OP for clickbaiting. Original title is just: Three clues that your LLM may be poisoned with a sleeper-agent back door
But:
WTF, why discredit your own article right at the beginning? Such a weird line.
TheBat@lemmy.world 3 weeks ago
That’s The Register for you. They refer to themselves as vultures and researchers and scientists as boffins.
alaphic@lemmy.world 3 weeks ago
Are you familiar with the term ‘tongue in cheek’? Or ‘hyperbole’? Cuz - I’m just sayin- I really doubt that even the yellow-est of rags would expect people to believe that we’re only a “bite my shiny metal ass” away from triggering a T2 style ‘Judgement Day’… I’d say it’s simply far more likely they were simply being facetious.
Now if it was NewsMax, on the other hand…
Hond@piefed.social 3 weeks ago
Yeah, i’m familiar with the concept of humor. No worries.
wuffah@lemmy.world 3 weeks ago
My personal theory is that it lends credibility to the idea that a “rogue AI” will destroy humanity instead of the billionaire broligarchs that wield it to control and surveil the masses.
RalfWausE@feddit.org 3 weeks ago
Its “The Register”.
CardboardVictim@piefed.social 3 weeks ago
Also there are three clues but it just explains the process a bit? Very strange article indeed.
hexagonwin@lemmy.sdf.org 3 weeks ago
kinda feels like they forgot to add ‘/s’