Towards Compositional Generalization of LLMs via Skill Taxonomy Guided Data Synthesis
By: Yifan Wei , Li Du , Xiaoyan Yu and more
Potential Business Impact:
Teaches computers to follow complex instructions better.
Large Language Models (LLMs) and agent-based systems often struggle with compositional generalization due to a data bottleneck in which complex skill combinations follow a long-tailed, power-law distribution, limiting both instruction-following performance and generalization in agent-centric tasks. To address this challenge, we propose STEPS, a Skill Taxonomy guided Entropy-based Post-training data Synthesis framework for generating compositionally challenging data. STEPS explicitly targets compositional generalization by uncovering latent relationships among skills and organizing them into an interpretable, hierarchical skill taxonomy using structural information theory. Building on this taxonomy, we formulate data synthesis as a constrained information maximization problem, selecting skill combinations that maximize marginal structural information within the hierarchy while preserving semantic coherence. Experiments on challenging instruction-following benchmarks show that STEPS outperforms existing data synthesis baselines, while also yielding improved compositional generalization in downstream agent-based evaluations.
Similar Papers
PolySkill: Learning Generalizable Skills Through Polymorphic Abstraction
Computation and Language
Agents learn useful skills that work everywhere.
Multimodal LLMs Do Not Compose Skills Optimally Across Modalities
Computation and Language
AI struggles to combine different skills to solve new problems.
How Does RL Post-training Induce Skill Composition? A Case Study on Countdown
Machine Learning (CS)
Teaches computers to build new skills from old ones.