Comment on Asking ChatGPT to Repeat Words ‘Forever’ Is Now a Terms of Service Violation

<- View Parent
Jamie@jamie.moe ⁨11⁩ ⁨months⁩ ago

Speaking for LLMs, given that they operate on a next-token basis, there will be some statistical likelihood of spitting out original training data that can’t be avoided. The normal counter-argument being that in theory, the odds of a particular piece of training data coming back out intact for more than a handful of words should be extremely low.

Of course, in this case, Google’s researchers took advantage of the repeat discouragement mechanism to make that unlikelihood occur reliably.

source
Sort:hotnewtop