PerNodeDrop: A Method Balancing Specialized Subnets and Regularization in Deep Neural Networks
By: Gelesh G Omathil, Sreeja CS
Potential Business Impact:
Makes computer learning more accurate on new information.
Deep neural networks possess strong representational capacity yet remain vulnerable to overfitting, primarily because neurons tend to co-adapt in ways that, while capturing complex and fine-grained feature interactions, also reinforce spurious and non-generalizable patterns that inflate training performance but reduce reliability on unseen data. Noise-based regularizers such as Dropout and DropConnect address this issue by injecting stochastic perturbations during training, but the noise they apply is typically uniform across a layer or across a batch of samples, which can suppress both harmful and beneficial co-adaptation. This work introduces PerNodeDrop, a lightweight stochastic regularization method. It applies per-sample, per-node perturbations to break the uniformity of the noise injected by existing techniques, thereby allowing each node to experience input-specific variability. Hence, PerNodeDrop preserves useful co-adaptation while applying regularization. This narrows the gap between training and validation performance and improves reliability on unseen data, as evident from the experiments. Although superficially similar to DropConnect, PerNodeDrop operates at the sample level. It drops weights at the sample level, not the batch level. An expected-loss analysis formalizes how its perturbations attenuate excessive co-adaptation while retaining predictive interactions. Empirical evaluations on vision, text, and audio benchmarks indicate improved generalization relative to the standard noise-based regularizer.
Similar Papers
A Combinatorial Theory of Dropout: Subnetworks, Graph Geometry, and Generalization
Machine Learning (CS)
Makes computer learning better by picking good parts.
Analytic theory of dropout regularization
Machine Learning (Stat)
Makes computer learning better by ignoring bad data.
Synaptic Pruning: A Biological Inspiration for Deep Learning Regularization
Machine Learning (CS)
Makes computer brains learn smarter and faster.