FlexAct: Why Learn when you can Pick?
By: Ramnath Kumar , Kyle Ritscher , Junmin Judy and more
Potential Business Impact:
Lets computers pick the best math for learning.
Learning activation functions has emerged as a promising direction in deep learning, allowing networks to adapt activation mechanisms to task-specific demands. In this work, we introduce a novel framework that employs the Gumbel-Softmax trick to enable discrete yet differentiable selection among a predefined set of activation functions during training. Our method dynamically learns the optimal activation function independently of the input, thereby enhancing both predictive accuracy and architectural flexibility. Experiments on synthetic datasets show that our model consistently selects the most suitable activation function, underscoring its effectiveness. These results connect theoretical advances with practical utility, paving the way for more adaptive and modular neural architectures in complex learning scenarios.
Similar Papers
SmartMixed: A Two-Phase Training Strategy for Adaptive Activation Function Learning in Neural Networks
Machine Learning (CS)
Lets computer brains learn better ways to think.
Task-Specific Activation Functions for Neuroevolution using Grammatical Evolution
Neural and Evolutionary Computing
Creates smarter computer brains that learn better.
DL101 Neural Network Outputs and Loss Functions
Machine Learning (CS)
Helps computers learn better by picking the right math.