Isn’t this still subject to the same problem, where a system can lie about its inference chain by returning a plausible chain which wasn’t the actual chain used for the conclusion? (I’m thinking from the perspective of a consumer sending an API request, not the service provider directly accessing the model.)
Also:
Any time I see a highly technical post talking about AI and/or crypto, I imagine a skilled accountant living in the middle of mob territory. They may not be directly involved in any scams themselves, but they gotta know that their neighbors are crooked and a lot of their customers are gonna use their services in nefarious ways.
LainTrain@lemmy.dbzer0.com 1 year ago
I don’t understand what is exactly being verified there? Model integrity? Factors for “reasoning”?
AtHeartEngineer@lemmy.world 1 year ago
Integrity of the model, inputs, and outputs, but with the potential to hide either the inputs or the model and maintain verifiability.
LainTrain@lemmy.dbzer0.com 1 year ago
But what is meant by “integrity of the model, inputs and outputs”? I guess I don’t understand the attack vector, what’s the threat here? Someone messes with the model file or refines a model towards a specific malicious bias?