Comment on AI Is Scheming, and Stopping It Won’t Be Easy, OpenAI Study Finds

<- View Parent
MentalEdge@sopuli.xyz ⁨3⁩ ⁨days⁩ ago

Seems like it’s a technical term, a bit like “hallucination”.

It refers to when an LLM will in some way try to deceive or manipulate the user interacting with it.

There’s hallucination, when a model “genuinely” claims something untrue is true.

This is about how a model might lie, even though the “chain of thought” shows it “knows” better.

source
Sort:hotnewtop