Score: 0

Diagonal Linear Networks and the Lasso Regularization Path

Published: September 23, 2025 | arXiv ID: 2509.18766v1

By: Raphaël Berthier

Potential Business Impact:

Training computers learns like a math shortcut.

Business Areas:
Multi-level Marketing Sales and Marketing

Diagonal linear networks are neural networks with linear activation and diagonal weight matrices. Their theoretical interest is that their implicit regularization can be rigorously analyzed: from a small initialization, the training of diagonal linear networks converges to the linear predictor with minimal 1-norm among minimizers of the training loss. In this paper, we deepen this analysis showing that the full training trajectory of diagonal linear networks is closely related to the lasso regularization path. In this connection, the training time plays the role of an inverse regularization parameter. Both rigorous results and simulations are provided to illustrate this conclusion. Under a monotonicity assumption on the lasso regularization path, the connection is exact while in the general case, we show an approximate connection.

Page Count
29 pages

Category
Computer Science:
Machine Learning (CS)