Preference-Driven Multi-Objective Combinatorial Optimization with Conditional Computation
By: Mingfeng Fan , Jianan Zhou , Yifeng Zhang and more
Potential Business Impact:
Helps computers solve hard problems better.
Recent deep reinforcement learning methods have achieved remarkable success in solving multi-objective combinatorial optimization problems (MOCOPs) by decomposing them into multiple subproblems, each associated with a specific weight vector. However, these methods typically treat all subproblems equally and solve them using a single model, hindering the effective exploration of the solution space and thus leading to suboptimal performance. To overcome the limitation, we propose POCCO, a novel plug-and-play framework that enables adaptive selection of model structures for subproblems, which are subsequently optimized based on preference signals rather than explicit reward values. Specifically, we design a conditional computation block that routes subproblems to specialized neural architectures. Moreover, we propose a preference-driven optimization algorithm that learns pairwise preferences between winning and losing solutions. We evaluate the efficacy and versatility of POCCO by applying it to two state-of-the-art neural methods for MOCOPs. Experimental results across four classic MOCOP benchmarks demonstrate its significant superiority and strong generalization.
Similar Papers
Multi-Objective Reward and Preference Optimization: Theory and Algorithms
Machine Learning (CS)
Teaches computers to make safe, smart choices.
BOPO: Neural Combinatorial Optimization via Best-anchored and Objective-guided Preference Optimization
Machine Learning (CS)
Solves hard puzzles much faster with smart computer learning.
UCPO: A Universal Constrained Combinatorial Optimization Method via Preference Optimization
Neural and Evolutionary Computing
Helps computers solve hard problems with fewer rules.