The interplay of robustness and generalization in quantum machine learning
By: Julian Berberich, Tobias Fellner, Christian Holm
Potential Business Impact:
Makes quantum computers learn better and avoid mistakes.
While adversarial robustness and generalization have individually received substantial attention in the recent literature on quantum machine learning, their interplay is much less explored. In this chapter, we address this interplay for variational quantum models, which were recently proposed as function approximators in supervised learning. We discuss recent results quantifying both robustness and generalization via Lipschitz bounds, which explicitly depend on model parameters. Thus, they give rise to a regularization-based training approach for robust and generalizable quantum models, highlighting the importance of trainable data encoding strategies. The practical implications of the theoretical results are demonstrated with an application to time series analysis.
Similar Papers
On the Generalization of Adversarially Trained Quantum Classifiers
Quantum Physics
Makes quantum computers safer from tricky attacks.
Adversarial Robustness in Distributed Quantum Machine Learning
Quantum Physics
Makes quantum computers safer from hackers.
Efficient and Accurate Estimation of Lipschitz Constants for Hybrid Quantum-Classical Decision Models
Quantum Physics
Makes smart computers learn and be fair.