Score: 2

Tune My Adam, Please!

Published: August 27, 2025 | arXiv ID: 2508.19733v2

By: Theodoros Athanasiadis , Steven Adriaensen , Samuel Müller and more

Potential Business Impact:

Makes computer learning faster and better.

Business Areas:
A/B Testing Data and Analytics

The Adam optimizer remains one of the most widely used optimizers in deep learning, and effectively tuning its hyperparameters is key to optimizing performance. However, tuning can be tedious and costly. Freeze-thaw Bayesian Optimization (BO) is a recent promising approach for low-budget hyperparameter tuning, but is limited by generic surrogates without prior knowledge of how hyperparameters affect learning. We propose Adam-PFN, a new surrogate model for Freeze-thaw BO of Adam's hyperparameters, pre-trained on learning curves from TaskSet, together with a new learning curve augmentation method, CDF-augment, which artificially increases the number of available training examples. Our approach improves both learning curve extrapolation and accelerates hyperparameter optimization on TaskSet evaluation tasks, with strong performance on out-of-distribution (OOD) tasks.

Repos / Data Links

Page Count
13 pages

Category
Computer Science:
Machine Learning (CS)