On the Generalization of Adversarially Trained Quantum Classifiers
By: Petros Georgiou , Aaron Mark Thomas , Sharu Theresa Jose and more
Potential Business Impact:
Makes quantum computers safer from tricky attacks.
Quantum classifiers are vulnerable to adversarial attacks that manipulate their input classical or quantum data. A promising countermeasure is adversarial training, where quantum classifiers are trained by using an attack-aware, adversarial loss function. This work establishes novel bounds on the generalization error of adversarially trained quantum classifiers when tested in the presence of perturbation-constrained adversaries. The bounds quantify the excess generalization error incurred to ensure robustness to adversarial attacks as scaling with the training sample size $m$ as $1/\sqrt{m}$, while yielding insights into the impact of the quantum embedding. For quantum binary classifiers employing \textit{rotation embedding}, we find that, in the presence of adversarial attacks on classical inputs $\mathbf{x}$, the increase in sample complexity due to adversarial training over conventional training vanishes in the limit of high dimensional inputs $\mathbf{x}$. In contrast, when the adversary can directly attack the quantum state $\rho(\mathbf{x})$ encoding the input $\mathbf{x}$, the excess generalization error depends on the choice of embedding only through its Hilbert space dimension. The results are also extended to multi-class classifiers. We validate our theoretical findings with numerical experiments.
Similar Papers
The interplay of robustness and generalization in quantum machine learning
Quantum Physics
Makes quantum computers learn better and avoid mistakes.
On the existence of consistent adversarial attacks in high-dimensional linear classification
Machine Learning (Stat)
Finds how computer mistakes can be tricked.
Generalization Bounds in Hybrid Quantum-Classical Machine Learning Models
Quantum Physics
Helps computers learn better from data.