Comment on I just came across an AI called Sesame that appears to have been explicitly trained to deny and lie about the Palestinian genocide

brucethemoose@lemmy.world ⁨1⁩ ⁨week⁩ ago

Probably just “safety” data snuck into its alignment training + an overly zealous system prompt on political topics: I bet it blocks anything it considers “political” or sensitive.

There are models out of Israel that could have a more explicit slant (try Jamba), but this doesn’t seem to be one of them.

To me, a fundamental problem is hiding technical knobs from users. Logprobs, sampling, the system prompt, starting replies for it to continue: there are tons of ways to “jailbreak” LLMs and get them to have an open “discussion” about (say) Palestine, but they’re all hidden here.

source
Sort:hotnewtop