Certification of neural nets

Watermarks, cryptographic verification and other certificates of authenticity for our computation

November 26, 2024 — November 26, 2024

adversarial
approximation
Bayes
cryptography
game theory
Monte Carlo
neural nets
optimization
probabilistic algorithms
probability
security
statistics
Figure 1

Certifying NNs to be what they say they are. Various interesting challenges in this domain. I am not sure if this is well-specified category in itself. Possibly at some point I will separate the cryptographic verification from other certification ideas. Or maybe some other taxonomy? TBD

1 Ownership of models

Keyword: Proof-of-learning, …

(Garg et al. 2023; Goldwasser et al. 2022; Jia et al. 2021)

TBD

2 Proof of training

E.g. Abbaszadeh et al. (2024):

A zero-knowledge proof of training (zkPoT) enables a party to prove that they have correctly trained a committed model based on a committed dataset without revealing any additional information about the model or the dataset. An ideal zkPoT should offer provable security and privacy guarantees, succinct proof size and verifier runtime, and practical prover efficiency. In this work, we present , a zkPoT targeted for deep neural networks (DNNs) that achieves all these goals at once. Our construction enables a prover to iteratively train their model via (mini-batch) gradient descent, where the number of iterations need not be fixed in advance; at the end of each iteration, the prover generates a commitment to the trained model parameters attached with a succinct zkPoT, attesting to the correctness of the executed iterations. The proof size and verifier time are independent of the number of iterations.

3 Proof of robustness

Didn’t know this was a thing, but then I met Mahalakshmi Sabanayagam who graciously explained to me Gosch et al. (2024) and Sabanayagam et al. (2024).

4 References

Abbaszadeh, Pappas, Katz, et al. 2024. Zero-Knowledge Proofs of Training for Deep Neural Networks.”
Garg, Goel, Jha, et al. 2023. Experimenting with Zero-Knowledge Proofs of Training.” In Proceedings of the 2023 ACM SIGSAC Conference on Computer and Communications Security. CCS ’23.
Goldwasser, Kim, Vaikuntanathan, et al. 2022. Planting Undetectable Backdoors in Machine Learning Models.”
Gosch, Sabanayagam, Ghoshdastidar, et al. 2024. Provable Robustness of (Graph) Neural Networks Against Data Poisoning and Backdoor Attacks.”
Jia, Yaghini, Choquette-Choo, et al. 2021. Proof-of-Learning: Definitions and Practice.” In 2021 IEEE Symposium on Security and Privacy (SP).
Sabanayagam, Gosch, Günnemann, et al. 2024. “Exact Certification of (Graph) Neural Networks Against Label Poisoning.”