On the Stability of Neural Networks in Deep Learning
By: Blaise Delattre
Potential Business Impact:
Makes computer brains more stable and trustworthy.
Deep learning has achieved remarkable success across a wide range of tasks, but its models often suffer from instability and vulnerability: small changes to the input may drastically affect predictions, while optimization can be hindered by sharp loss landscapes. This thesis addresses these issues through the unifying perspective of sensitivity analysis, which examines how neural networks respond to perturbations at both the input and parameter levels. We study Lipschitz networks as a principled way to constrain sensitivity to input perturbations, thereby improving generalization, adversarial robustness, and training stability. To complement this architectural approach, we introduce regularization techniques based on the curvature of the loss function, promoting smoother optimization landscapes and reducing sensitivity to parameter variations. Randomized smoothing is also explored as a probabilistic method for enhancing robustness at decision boundaries. By combining these perspectives, we develop a unified framework where Lipschitz continuity, randomized smoothing, and curvature regularization interact to address fundamental challenges in stability. The thesis contributes both theoretical analysis and practical methodologies, including efficient spectral norm computation, novel Lipschitz-constrained layers, and improved certification procedures.
Similar Papers
Stable neural networks and connections to continuous dynamical systems
Numerical Analysis
Makes computer brains harder to trick.
On the Stability of the Jacobian Matrix in Deep Neural Networks
Machine Learning (CS)
Makes smart computer programs learn better.
Understanding Learning Dynamics Through Structured Representations
Machine Learning (CS)
Makes AI learn faster and smarter with fewer mistakes.