Yeah great point. Or similar cases in the past, even.
Hundreds in this case, but millions in the long term.
I can see why Air Canada wanted to fight it, because if they accept liability it sets a precedent that they should also accept liability for similar cases in future.
And they SHOULD accept liability, so I’m glad Air Canada were forced to!
thanks_shakey_snake@lemmy.ca 8 months ago
brsrklf@jlai.lu 8 months ago
The solution would be easy, just stop having an LLM chatbot.
But I suspect they don’t want to because someone sold them on how good and cheap and human-resource-free it was, and now they think they’re too invested.
thanks_shakey_snake@lemmy.ca 8 months ago
Plus just the general sentiment that you’re not businessing right if you don’t something something AI.
rottingleaf@lemmy.zip 8 months ago
Feel my blood boiling at the very thought of people choosing to use something buzzwordy like blockchain or “AI”, despite likely no competent person advising them to employ that, AND then trying to clean themselves of the responsibility when it misfires.
That’s as if drunk driving leading to car crash was blamed on the air, because “having fun is not a crime”.
Only with computing these people unironically think that nobody should be responsible, because everybody they respect is as clueless as themselves, so “nobody knows how it works, it’s a frontier, see”.