
There is a more recent version of this eprint available. Click here to view it.
Smith, M.T., Grosse, K., Backes, M. et al. (1 more author) (Submitted: 2019) Adversarial vulnerability bounds for Gaussian process classification. arXiv. (Submitted)
Abstract
Machine learning (ML) classification is increasingly used in safety-critical systems. Protecting ML classifiers from adversarial examples is crucial. We propose that the main threat is that of an attacker perturbing a confidently classified input to produce a confident misclassification. To protect against this we devise an adversarial bound (AB) for a Gaussian process classifier, that holds for the entire input domain, bounding the potential for any future adversarial method to cause such misclassification. This is a formal guarantee of robustness, not just an empirically derived result. We investigate how to configure the classifier to maximise the bound, including the use of a sparse approximation, leading to the method producing a practical, useful and provably robust classifier, which we test using a variety of datasets.
Metadata
Item Type: | Article |
---|---|
Authors/Creators: |
|
Copyright, Publisher and Additional Information: | © 2019 The Author(s). For reuse permissions, please contact the Author(s). |
Dates: |
|
Institution: | The University of Sheffield |
Academic Units: | The University of Sheffield > Faculty of Engineering (Sheffield) > Department of Computer Science (Sheffield) |
Depositing User: | Symplectic Sheffield |
Date Deposited: | 17 Jan 2020 13:57 |
Last Modified: | 25 Nov 2022 17:21 |
Status: | Submitted |
Identification Number: | 10.48550/arXiv.1909.08864 |
Open Archives Initiative ID (OAI ID): | oai:eprints.whiterose.ac.uk:155247 |
Available Versions of this Item
- Adversarial vulnerability bounds for Gaussian process classification. (deposited 17 Jan 2020 13:57) [Currently Displayed]