On the role of non-linear latent features in bipartite generative neural networks
By: Tony Bonnaire , Giovanni Catania , Aurélien Decelle and more
Potential Business Impact:
Improves computer memory recall by changing how it learns.
We investigate the phase diagram and memory retrieval capabilities of bipartite energy-based neural networks, namely Restricted Boltzmann Machines (RBMs), as a function of the prior distribution imposed on their hidden units - including binary, multi-state, and ReLU-like activations. Drawing connections to the Hopfield model and employing analytical tools from statistical physics of disordered systems, we explore how the architectural choices and activation functions shape the thermodynamic properties of these models. Our analysis reveals that standard RBMs with binary hidden nodes and extensive connectivity suffer from reduced critical capacity, limiting their effectiveness as associative memories. To address this, we examine several modifications, such as introducing local biases and adopting richer hidden unit priors. These adjustments restore ordered retrieval phases and markedly improve recall performance, even at finite temperatures. Our theoretical findings, supported by finite-size Monte Carlo simulations, highlight the importance of hidden unit design in enhancing the expressive power of RBMs.
Similar Papers
The unbearable lightness of Restricted Boltzmann Machines: Theoretical Insights and Biological Applications
Disordered Systems and Neural Networks
Helps computers learn from data better.
The Gaussian-Multinoulli Restricted Boltzmann Machine: A Potts Model Extension of the GRBM
Machine Learning (CS)
Teaches computers to remember and reason better.
Inferring Higher-Order Couplings with Neural Networks
Disordered Systems and Neural Networks
Finds hidden patterns in complex data.