Comment on OpenAI confirms that AI writing detectors don’t work
cheesorist@lemmy.world 1 year ago
they never did, they never will.
Comment on OpenAI confirms that AI writing detectors don’t work
cheesorist@lemmy.world 1 year ago
they never did, they never will.
stevedidWHAT@lemmy.world 1 year ago
Why tho or are you trying to be vague on purpose
bioemerl@kbin.social 1 year ago
Because you're training a detector on something that is designed to emulate regular languages closest possible, and human speech has so much incredible variability that it's almost impossible to identify if someone or something has been written by an AI.
You can detect maybe your typical generic chat GPT type outputs, but you can characterize a conversation with chat GPT or any of the other much better local models (privacy and control are aspects which make them better) and after doing that you can get radically human seeming outputs that are totally different from anything chat GPT will output.
In short, given a static block of text it's going to be nearly impossible to detect if it's coming from an AI. It's just too difficult to problem, and if you're going to solve it it's going to be immediately obsolete the next time someone fine tunes their own model
stevedidWHAT@lemmy.world 1 year ago
Yeah this makes a lot of sense considering the vastness of language and it’s imperfections (English I’m mostly looking at you, ya inbred fuck)
Are there any other detection techniques that you know of? Wb forcing AI models to have a signature that is guaranteed to be indentifiable, permanent, and unique for each tuning produced? It’d have to be not directly noticeable but easy to calculate in order to prevent any “distractions” for the users.
Grimy@lemmy.world 1 year ago
The output is pure text so you would have to hide the signature in the response itself. On top of being useless since most users slightly modify the text after receiving it, it would probably have a negative effect on the quality. It’s also insanely complicated to train that kind of behavior into an llm.
bioemerl@kbin.social 1 year ago
Either AI remains entirely in the hands of fucks like open AI or this is impossible and easily removed. AI should be a free common use tool, not an extension of corporate control.
Eufalconimorph@discuss.tchncs.de 1 year ago
Because AIs are (partly) trained by making AI detectors. If an AI can be distinguished from a natural intelligence, it’s not good enough at emulating intelligence. If an AI detector can reliably distinguish AI from humans, the AI companies will use that detector to train their next AI.
stevedidWHAT@lemmy.world 1 year ago
I’m not sure I’m following your argument here - you keep switching between talking about AI and AI detectors. Each of the below are just numbered according to the order of your prior responses as sentences:
I’m having trouble finding your support for your claim
Theharpyeagle@lemmy.world 1 year ago
See Generative Adversarial Network (GAN). Basically, making new AI detectors will always be harder than beating current ones. AI detectors have to somehow find a new “tell”, the target AI need only train itself on the output of the detector to figure out how to trick it.
dack@lemmy.world 1 year ago
At a very high level, training is something like:
Step #2 is also exactly what an “AI detector” does. If someone is able to write code that reliably distinguishes between AI and human text, then AI developers would plug it in to that training step in order to improve their AI.
In other words, if some theoretical machine perfectly “knows” the difference between generated and human text, then the same machine can also be used to make text that is indistinguishable from human text.
sebi@lemmy.world 1 year ago
Because generative Neural Networks always have some random noise. Read more about it here
stevedidWHAT@lemmy.world 1 year ago
Isn’t that article about GANs?
Isn’t GPT not a GAN?
PetDinosaurs@lemmy.world 1 year ago
It almost certainly has some gan-like pieces.
Gans are part of the NN toolbox, like cnns and rnns and such.
Basically all commercial algorithms (not just nns, everything) are what I like to call “hybrid” methods, which means keep throwing different tools at it until things work well enough.
bioemerl@kbin.social 1 year ago
It's not even about diffusion models. Adversarial networks are basically obsolete