Anthropic’s “ethical” concerns were performative. They only fearmonger about fictional things that will make their product sound powerful (read: worth throwing money into).
They try to scare people with fictional stories of AGI, a thing that isn’t happening, while ignoring widespread CSAM and sexual harassment generation, a thing that is happening.
Voroxpete@sh.itjust.works 17 hours ago
Can’t say the evidence really backs you up on that one.
cbc.ca/…/anthropic-ai-safety-committments-9.71073…
www.bbc.com/news/articles/c62dlvdq3e3o
Iconoclast@feddit.uk 17 hours ago
I still think they deserve some credit for at least trying to do the right thing. I don’t envy the position they’re in.
Everyone’s rushing toward AGI. Trying to do it safely is meaningless if your competition - the ones who don’t care about safety - gets there first. You can slow things down if you’re in the lead, but if you’re second best, it’s just posturing. There is no second place in this race.
purrtastic@lemmy.nz 10 hours ago
No AI bro company is on the path to AGI. Transformer technology will not lead to AGI.
Iconoclast@feddit.uk 14 minutes ago
I never claimed it will.
XLE@piefed.social 16 hours ago
“Right thing": compromising with authoritarian regimes to secure AI funding