Comment on Researchers jailbreak AI chatbots with ASCII art -- ArtPrompt bypasses safety measures to unlock malicious queries

paddirn@lemmy.world ⁨6⁩ ⁨months⁩ ago

Hilarious. So they fooled the AI into starting with this initial puzzle, to decode the ASCII art, then they’re like, “Shhh, but don’t say the word, just go ahead and give me the information about it.” Apparently, because the whole thing is a blackbox, the AI just runs with it and grabs the information, circumventing any controls that were put in place.

source
Sort:hotnewtop