Am I missing something? The article seems to suggest it works via hidden text characters. Has OpenAI never heard of pasting text into a utf8 notepad before?
Am I the only one who rewrites most of ChatGPT’s output into my own words because it’s “voice” is garbage anyway? I ask it to write me a cover letter and that gives me a rough outline and some points to make, but I have to do massive editing to avoid redundancy, awkward phrasing, outright lies, etc.
I can’t imagine turning in raw ChatGPT output. I had one of my developers use AI to write code and submitted that shit raw and it was immediately obvious because some relatively simple code has really weird artifacts like overwriting a value that had no reason to even be touched.
CameronDev@programming.dev 3 months ago
The arstechnica article speculated it was more of a pattern of words thing.
I think it is lies, and doesn’t exist or work anywhere near as good as they claim. Or, its incredibly easy to bypass.
arstechnica.com/…/openai-has-the-tech-to-watermar…
deadcade@lemmy.deadca.de 3 months ago
Research on this topic exists, and it is possible to alter the output of an LLM in minor ways, that statistically “watermark” the results without drastically changing the quality of the output. OpenAI has probably implemented this into ChatGPT.
www.youtube.com/watch?v=2Kx9jbSMZqA
I think the tool exists, and is (at least close to) as good as they claim it is. They can’t release it, because once the public can tell with high accuracy whether ChatGPT wrote some text, another AI can be developed to circumvent detection from this method, making the tool useless.
CameronDev@programming.dev 3 months ago
That is a long video, is the paper published somewhere?
Im willing to accept that you can statistically “watermark” the text, but I’m not convinced that it would be tamper resistant, which is a large part of what makes a watermark useful. If it can’t survive an idiot with a thesaurus, its probably not gonna be terribly useful.
TheHobbyist@lemmy.zip 3 months ago
I think it exists and works but that its simply not in their best interest to have people use it and be found out that they used chatgpt, for OpenAI’s business/profit potential. I have nothing to back it up but have just lost all faith in OpenAI.
CameronDev@programming.dev 3 months ago
Im willing to believe it exists, but not that its any good. 99% is a crazy accuracy claim.
originalfrozenbanana@lemm.ee 3 months ago
They sell the cheating tool and the detection software
archomrade@midwest.social 3 months ago
Not to mention that it would be extremely difficult to implement an effective watermark on text below a certain size
There are hundreds of thousands of pixels in an image where you can hide a watermark, but in a text output of a paragraph or less there are only a couple hundred characters.
How precise is the watermark? Is it a specific sequence of characters? Is it a sequence of words? A number of characters in a row? Non-print characters?
How precise the watermark is will determine how easy it is to get around. I imagine some of the most important uses to detect would be twitter/social media influence bots where the output length is only 140 characters or less. I find it hard to imagine a watermark on output of that size being effective or reliable.