Comment on Teen deepfake victim pushes for federal law targeting AI-generated explicit content
curiousaur@reddthat.com 9 months agoI don’t think you’re properly understanding the paradigm shift that’s coming with these models being open source and widely available while wearable AR smart glasses get better.
“You know Sharon is HR, look at this scandalous photo of her.”
“Uh, I’m seeing a live generated porno of everyone in this room right now, why would I care about that.”
aesthelete@lemmy.world 9 months ago
And I don’t think you’re fully understanding that the above is some type of fantasy you have, and will not actually be what the future is at all.
curiousaur@reddthat.com 9 months ago
It’s probably a bit of an exaggeration, but my point stands. It’s going to be so easy for anyone to see ai gen material of anyone else, no one is going to care anymore.
aesthelete@lemmy.world 9 months ago
I don’t even think that’s necessarily true. If you make it illegal and/or platforms ban it, you’re already taking a step toward making it more difficult to do.
I think throughout this thread you’re mistaking the technically possible for the probable or likely.
By making it illegal, you essentially eliminate the commercial incentive for making it easy. Every barrier to doing something makes it more unlikely that people will do it. I understand that there is an inherent motive for people to do it anyway, but, every hoop they have to jump through (e.g. setting up their “own, local AI”) reduces the likelihood of it happening.
People don’t even run their own email servers, music servers, video servers, etc. etc. etc…many don’t even store a local cache of regular porn…why the hell would most people bother themselves with setting up a local generative AI instance for this purpose?
Outlawing it and banning it from platforms makes it much more within the realm of the creepy basement weirdo rather than something that is as inevitably ubiquitous as you’re saying it will be.
curiousaur@reddthat.com 9 months ago
It’s not illegal to to work on, sell, or distribute the models. And making that illegal is what the first commenter said would be dangerous to do, since then regular people wouldn’t be able to compete with corporation’s abilities.
Once the models and portable hardware are good enough, and it’s just a matter of time, I think you’re underestimating how ubiquitous it will become.
Every teenage boy will have a pair of nudie glasses in the form of their smartphone running open source models, and you think they’re just going to not use them?