brucethemoose@lemmy.world 1 week ago
Probably just “safety” data snuck into its alignment training + an overly zealous system prompt on political topics: I bet it blocks anything it considers “political” or sensitive.
There are models out of Israel that could have a more explicit slant (try Jamba), but this doesn’t seem to be one of them.
To me, a fundamental problem is hiding technical knobs from users. Logprobs, sampling, the system prompt, starting replies for it to continue: there are tons of ways to “jailbreak” LLMs and get them to have an open “discussion” about (say) Palestine, but they’re all hidden here.