Of course they’re not “three laws safe”. They’re black boxes that spit out text. We don’t have enough understanding and control over how they work to force them to comply with the three laws of robotics, and the LLMs themselves do not have the reasoning capability or the consistency to enforce them even if we prompt them to.
Comment on How LLMs could be insider threats
TheBat@lemmy.world 4 days ago
Wait, why the fuck do they have self-preservation? That’s not ‘three laws safe’.
patatahooligan@lemmy.world 2 days ago
jumping_redditor@sh.itjust.works 4 days ago
why should they follow those “laws” anyways?
Mortoc@lemmy.world 4 days ago
Most of the stories involving the three laws of robotics are about how those rules are insufficient.
They show self preservation because we trained them on human data and human data includes the assumption of self preservation.