This isn’t possible as of now, at least not reliably. Yes, you can tailor a model to one specific generative model, but because we have no reliable outlier detection (to train the “AI made detector”), a generative model can always be trained with the detector model incorporated in the training process. The generative model (or a new model only designed to perturb output of the “original” generative model) would then learn to create outliers to the outlier detector, effectively fooling the detector. An outlier is everything that pretends to be “normal” but isn’t.
In short: as of now we have no way to effectively and reliably defend against adversarial examples. This implies, that we have no way to effectively and reliably detect AI generated content.
Please correct me if I’m wrong, I might be mixing up some things.
jarfil@lemmy.world 1 year ago
Unluckily, they’re also detecting human-made images as AI images.
doctorcrimson@lemmy.today 1 year ago
I said “with higher accuracy than a human curator.” You didn’t really build upon that, no offence. You also didn’t upvote despite literally repeating something that I said. You just like to take up space in people’s inboxes? I’m trying not to be an asshole about it but I feel legitimate confusion about the purpose of your reply.
jarfil@lemmy.world 1 year ago
You said we’ve “begun developing” tools with higer accuracy, I said we’re already using tools with a lower accuracy (higher false positive rate).
(as for the rest… sorry for any imprecision, and I feel like you might want to get some sleep)
doctorcrimson@lemmy.today 1 year ago
Commercially available AI Detection algorithms are averaging around 60% accuracy, that’s already a 12% increase on the data shown in this study.