Comment on Google Translate is vulnerable to prompt injection

FauxLiving@lemmy.world ⁨18⁩ ⁨hours⁩ ago

A bit flip, but this reads like people discovering that a hammer built specifically for NASA with specific metallurgical properties at the cost of $10,000 each where only 5 will ever be forged, because they were all intended to sit in a space ship in orbit around the Moon.

Then someone comes along and acts surprised that one was used to smash out a car window to steal a laptop.


LLMs will always be vulnerable to prompt injection because of how they function. Maybe, at some point in the future, we’ll understand enough about how LLMs represent knowledge internally so that we can craft specific subsystems to mitigate prompt injection… however, in 2026, that is just science fiction.

There are actual academic projects which are studying the boundaries of the prompt-injection vulnerabilities if you read in the machine learning/AI journals. These studies systemically study the problem, gather data and demonstrate their hypothesis.

One of the ways you can tell real Science from ‘hey, I heard’ science is that real science articles don’t start with ‘Person on social media posted that they found…’

This is a very interesting topic and if you’re interested you can find the actual science by starting here: www.nature.com/natmachintell/.

source
Sort:hotnewtop