Multi-step Predictive Coding Leads To Simplicity Bias
By: Aviv Ratzon, Omri Barak
Potential Business Impact:
Deep AI learns world's hidden rules.
Predictive coding is a framework for understanding the formation of low-dimensional internal representations mirroring the environment's latent structure. The conditions under which such representations emerge remain unclear. In this work, we investigate how the prediction horizon and network depth shape the solutions of predictive coding tasks. Using a minimal abstract setting inspired by prior work, we show empirically and theoretically that sufficiently deep networks trained with multi-step prediction horizons consistently recover the underlying latent structure, a phenomenon explained through the Ordinary Least Squares estimator structure and biases in learning dynamics. We then extend these insights to nonlinear networks and complex datasets, including piecewise linear functions, MNIST, multiple latent states and higher dimensional state geometries. Our results provide a principled understanding of when and why predictive coding induces structured representations, bridging the gap between empirical observations and theoretical foundations.
Similar Papers
Towards Scaling Deep Neural Networks with Predictive Coding: Theory and Practice
Machine Learning (CS)
Makes AI learn faster and use less energy.
Towards Scaling Deep Neural Networks with Predictive Coding: Theory and Practice
Machine Learning (CS)
Makes AI learn faster and use less power.
Introduction to Predictive Coding Networks for Machine Learning
Neural and Evolutionary Computing
Helps computers learn like brains to see pictures.