In the US companies are legally mandated to report specific things such as CSAM and other things if they come across it.
What the issue should be isn’t the fact that they are reporting it, the issue should be they have the capability to see it in the first place to be able to report it.
This isn’t me defending CSAM or anything like that but, in a decent storage system, google shouldn’t be able to even see what you have, let alone what the images actually are.
IAmNorRealTakeYourMeds@lemmy.world 3 weeks ago
severe mix feelings.
glad they caught him, but corporations casually snooping through your data and report whatever they want is definitely not an good thing
Dojan@pawb.social 3 weeks ago
Was a gay guy here in Sweden who got assaulted and kidnapped by masked police because some American company had found CSAM on his account while crawling through Yahoo email.
Only it wasn’t CSAM, the photos depicted the man’s 30 year old twinky boyfriend.
No restitution. No police were punished for assaulting a suspect proved innocent. The man and his boyfriend both were humiliated.
I’ve no mixed feelings about it. Spying through private data is entirely unforgivable. There are plenty of pesos out there who get caught and nothing happens anyway. They don’t need to violate innocent people’s privacy to do their job.
Like if the ends justify the means you can end all suffering in the world by just nuking everything. All problems solved.
HeyThisIsntTheYMCA@lemmy.world 3 weeks ago
Oh gods now you have me worried. 20 years ago I was a hundred pounds lighter and just a bag of skin holding a skeleton. There are some photos of me on my Google account that skinny. (also in your medical textbooks but anyways) and I also have photos of me now. We look like completely different people.
tidderuuf@lemmy.world 3 weeks ago
Microsoft has been doing this for years. It was with Onedrive at first but now that they’ve enabled “analytics” in every product that might connect to the internet they can have it all searched.
Supposedly it is first filtered by algorithms but that shit is still being uploaded somewhere other than your hard drive.
wizardbeard@lemmy.dbzer0.com 3 weeks ago
I believe it was in preview build versions of Win 7 or 10 where researchers found it was sending the generated thumbnails of images on your PC to Redmond (MS HQ). Can’t remember if they said it was for CSAM detection or just a debugging feature in the preview builds.
obvs@lemmy.world 3 weeks ago
Unfortunately, the negative effects from companies like Google turning in completely ethical people for doing things that should be completely legal and uncontroversial will do drastically more damage than the positive effects from said companies turning in the poorest of the pedophiles.
Zamboni_Driver@lemmy.ca 3 weeks ago
Example please
CompactFlax@discuss.tchncs.de 3 weeks ago
They’re suggesting it was automated hash based recognition.
I don’t have a problem with CSAM hash matching.
UnspecificGravity@piefed.social 3 weeks ago
Sure, until it starts flagging normal pictures with us janky AI and you get your door kicked in based on a warrant signed by Google.
IAmNorRealTakeYourMeds@lemmy.world 3 weeks ago
my issue is that we have a framework for corporations to scan all your data and inform the state. used to stop CSAM, but it’s a matter of state policy wether said structure will be used to fight discent.
org@lemmy.org 3 weeks ago
Doesn’t sound like hashes to me.
BlackLaZoR@lemmy.world 3 weeks ago
In EU it’s straight out forbidden
SaveTheTuaHawk@lemmy.ca 2 weeks ago
looking at account images or pedophilia?
SaveTheTuaHawk@lemmy.ca 2 weeks ago
Anyone with the public trust of dealing with patients needs to be scrutinized.
IAmNorRealTakeYourMeds@lemmy.world 2 weeks ago
I don’t disagree with that. but the feels like propaganda to destroy privacy and encourage a total surveillance state.