1. Adversarial vulnerability bounds for Gaussian process classification
- Author
-
Michael Thomas Smith, Kathrin Grosse, Michael Backes, and Mauricio A. Álvarez
- Subjects
FOS: Computer and information sciences ,Computer Science - Machine Learning ,Computer Science - Cryptography and Security ,ComputingMethodologies_PATTERNRECOGNITION ,Statistics - Machine Learning ,Artificial Intelligence ,Machine Learning (stat.ML) ,Cryptography and Security (cs.CR) ,Software ,Computer Science::Cryptography and Security ,Machine Learning (cs.LG) - Abstract
Machine learning (ML) classification is increasingly used in safety-critical systems. Protecting ML classifiers from adversarial examples is crucial. We propose that the main threat is that of an attacker perturbing a confidently classified input to produce a confident misclassification. To protect against this we devise an adversarial bound (AB) for a Gaussian process classifier, that holds for the entire input domain, bounding the potential for any future adversarial method to cause such misclassification. This is a formal guarantee of robustness, not just an empirically derived result. We investigate how to configure the classifier to maximise the bound, including the use of a sparse approximation, leading to the method producing a practical, useful and provably robust classifier, which we test using a variety of datasets., Comment: 10 pages + 2 pages references + 7 pages of supplementary. 12 figures. Submitted to AAAI
- Published
- 2022
- Full Text
- View/download PDF