Dynamic Learning Rate Scheduling based on Loss Changes Leads to Faster Convergence
By: Shreyas Subramanian, Bala Krishnamoorthy, Pranav Murthy
Despite significant advances in optimizers for training, most research works use common scheduler choices like Cosine or exponential decay. In this paper, we study \emph{GreedyLR}, a novel scheduler that adaptively adjusts the learning rate during training based on the current loss. To validate the effectiveness of our proposed scheduler, we conduct experiments on several NLP, CV, and LLM tasks with up to $7B$ parameters, including both fine-tuning and pre-training experiments. The results show that our approach outperforms several state-of-the-art schedulers in terms of accuracy, speed, and convergence. We also provide a theoretical analysis of the GreedyLR algorithm, including a proof of convergence and derivation of the optimal scaling factor $F$ that maximizes the convergence rate, along with experiments to show robustness of the algorithm to realistic noisy landscapes. Our scheduler is easy to implement, computationally efficient, and could be considered a good default scheduler for training.
Similar Papers
Cumulative Learning Rate Adaptation: Revisiting Path-Based Schedules for SGD and Adam
Machine Learning (CS)
Improves how computer learning programs learn faster.
Beyond Cosine Decay: On the effectiveness of Infinite Learning Rate Schedule for Continual Pre-training
Machine Learning (CS)
Teaches AI to learn new things without forgetting.
Accelerating SGDM via Learning Rate and Batch Size Schedules: A Lyapunov-Based Analysis
Machine Learning (CS)
Makes computer learning faster and more reliable.