SmartMixed: A Two-Phase Training Strategy for Adaptive Activation Function Learning in Neural Networks
By: Amin Omidvar
Potential Business Impact:
Lets computer brains learn better ways to think.
The choice of activation function plays a critical role in neural networks, yet most architectures still rely on fixed, uniform activation functions across all neurons. We introduce SmartMixed, a two-phase training strategy that allows networks to learn optimal per-neuron activation functions while preserving computational efficiency at inference. In the first phase, neurons adaptively select from a pool of candidate activation functions (ReLU, Sigmoid, Tanh, Leaky ReLU, ELU, SELU) using a differentiable hard-mixture mechanism. In the second phase, each neuron's activation function is fixed according to the learned selection, resulting in a computationally efficient network that supports continued training with optimized vectorized operations. We evaluate SmartMixed on the MNIST dataset using feedforward neural networks of varying depths. The analysis shows that neurons in different layers exhibit distinct preferences for activation functions, providing insights into the functional diversity within neural architectures.
Similar Papers
Developing Training Procedures for Piecewise-linear Spline Activation Functions in Neural Networks
Machine Learning (CS)
Makes computer brains learn better and faster.
Make Haste Slowly: A Theory of Emergent Structured Mixed Selectivity in Feature Learning ReLU Networks
Machine Learning (CS)
Helps computers learn better by reusing parts.
Multi-Plasticity Synergy with Adaptive Mechanism Assignment for Training Spiking Neural Networks
Neural and Evolutionary Computing
Teaches computer brains to learn better, faster.