Comment on Telegram repeatedly refused to join child protection schemes
pressanykeynow@lemmy.world 2 months agoYou didn’t answer my questions.
What moderation do you want? And how would you prevent “moderation” from becoming censorship?
Aren’t there people whose job is to prevent crimes? Why some IT person who has no idea of crime need to do their job?
atrielienz@lemmy.world 2 months ago
Because your questions aren’t germane to the point I was making. In fact the first question “how would you prevent " moderation” from becoming censorship" is literally answered by my second comment. Facebook already does this with Facebook messenger. But even if they didn’t, Signal has functions to allow encryption.
So what you’re saying is, criminals who aren’t using encryption (on a platform where encryption features are readily available) don’t deserve to be moderated on a platform where their messages are using a company’s cloud bandwidth. Does the company not have rights? And if we agree that the company has rights then they also have to follow the law.
Yes there are people who’s jobs are to (not prevent because police and policing is reactionary not preventative) investigate, and try criminals in a court of law for crimes). This was a poor question to ask. You’re literally acting like we don’t employ thousands of people over various social media and messaging platforms to review and moderate things like CSAM.
The gist for me is criminals gonna do criminal things but at the end of the day these are our public spaces and just because I don’t want to be surveilled in public or live in a police state doesn’t mean that I want criminals not to be prosecuted for crimes they commit just because someone cares more about their bottom line than they do about moderation of a messaging platform they provide to the public.
We aren’t talking about end to end encrypted messages here. We’re talking about messages with no such encryption that can