So many examples of this method failing I don’t even know where to start. Most visible, of course, was how that approach failed to stop Grok from “being woke” for like, a year or more.
Frankly, you sound like you’re talking straight out of your ass.
Jozav@lemmy.world [bot] 1 day ago
Sure, it can go wrong, it is not fool-proof. Just like building a new model can cause unwanted surprises.
BTW. There are many theories about Grok’s unethical behavior but this one is new to me. The reasons I was familiar with are: unfiltered training data, no ethical output restrictions, programming errors or incorrect system maintenance, strategic errors (Elon!), publishing before proper testing.
jumping_redditor@sh.itjust.works 1 day ago
why should any llm care about “ethics”?
MouldyCat@feddit.uk 1 day ago
well obviously it won’t, that’s why you need ethical output restrictions