Approximate equivariance via projection-based regularisation
By: Torben Berndt, Jan Stühmer
Potential Business Impact:
Makes AI learn faster and more accurately.
Equivariance is a powerful inductive bias in neural networks, improving generalisation and physical consistency. Recently, however, non-equivariant models have regained attention, due to their better runtime performance and imperfect symmetries that might arise in real-world applications. This has motivated the development of approximately equivariant models that strike a middle ground between respecting symmetries and fitting the data distribution. Existing approaches in this field usually apply sample-based regularisers which depend on data augmentation at training time, incurring a high sample complexity, in particular for continuous groups such as $SO(3)$. This work instead approaches approximate equivariance via a projection-based regulariser which leverages the orthogonal decomposition of linear layers into equivariant and non-equivariant components. In contrast to existing methods, this penalises non-equivariance at an operator level across the full group orbit, rather than point-wise. We present a mathematical framework for computing the non-equivariance penalty exactly and efficiently in both the spatial and spectral domain. In our experiments, our method consistently outperforms prior approximate equivariance approaches in both model performance and efficiency, achieving substantial runtime gains over sample-based regularisers.
Similar Papers
A Regularization-Guided Equivariant Approach for Image Restoration
CV and Pattern Recognition
Fixes blurry pictures by understanding shapes.
Equivariant Representation Learning for Symmetry-Aware Inference with Guarantees
Machine Learning (CS)
Teaches computers to learn from fewer examples.
Training Dynamics of Learning 3D-Rotational Equivariance
Machine Learning (CS)
Teaches computers to see 3D shapes perfectly.