Comment on OpenAI’s latest model will block the ‘ignore all previous instructions’ loophole

<- View Parent
Hexarei@programming.dev ⁨3⁩ ⁨months⁩ ago

You can dislike the statement all you want, but they literally do not have a way to know things. They provide a convincing illusion of knowledge through statistical likelihood of the next token occurring, but they have no internal mechanism for looking up information.

They have no fact repositories to rely on.

They do not possess the ability to know what is and is not correct.

They cannot check documentation or verify that a function or library or API endpoint exists, even though they will confidently create calls to them.

They are statistical models, calculating how likely the next token is based on transformations in a many-dimensional space in which the relationships between existing tokens are treated as vectors in a process for determining the next token.

They have their uses, but relying on them for factual information (which includes knowledge of apis and libraries) is a bad idea. They are just as likely to provide realistic answers as they are to make up fake answers and present them as real.

They are good for inspiration or a jumping off point, but should always be fact checked and validated.

They’re fantastic at transforming data from one format to another, or extracting data from natural language written information. I’m even using one in a project to guess at filling in a form based on an incoming customer email.

source
Sort:hotnewtop