FoMEMO: Towards Foundation Models for Expensive Multi-objective Optimization
By: Yiming Yao , Fei Liu , Liang Zhao and more
Potential Business Impact:
Finds best solutions faster, even for new problems.
Expensive multi-objective optimization is a prevalent and crucial concern in many real-world scenarios, where sample-efficiency is vital due to the limited evaluations to recover the true Pareto front for decision making. Existing works either involve rebuilding Gaussian process surrogates from scratch for each objective in each new problem encountered, or rely on extensive past domain experiments for pre-training deep learning models, making them hard to generalize and impractical to cope with various emerging applications in the real world. To address this issue, we propose a new paradigm named FoMEMO (Foundation Models for Expensive Multi-objective Optimization), which enables the establishment of a foundation model conditioned on any domain trajectory and user preference, and facilitates fast in-context optimization based on the predicted preference-wise aggregation posteriors. Rather than accessing extensive domain experiments in the real world, we demonstrate that pre-training the foundation model with a diverse set of hundreds of millions of synthetic data can lead to superior adaptability to unknown problems, without necessitating any subsequent model training or updates in the optimization process. We evaluate our method across a variety of synthetic benchmarks and real-word applications, and demonstrate its superior generality and competitive performance compared to existing methods.
Similar Papers
Parametric Expensive Multi-Objective Optimization via Generative Solution Modeling
Machine Learning (CS)
Solves many hard problems faster, without re-testing.
Parametric Pareto Set Learning for Expensive Multi-Objective Optimization
Neural and Evolutionary Computing
Finds best answers for changing problems instantly.
MoFa: A Unified Performance Modeling Framework for LLM Pretraining
Distributed, Parallel, and Cluster Computing
Finds best way to train giant AI brains faster.