Scalable Evaluation and Neural Models for Compositional Generalization
By: Giacomo Camposampiero , Pietro Barbiero , Michael Hersche and more
Potential Business Impact:
Teaches computers to understand new ideas.
Compositional generalization-a key open challenge in modern machine learning-requires models to predict unknown combinations of known concepts. However, assessing compositional generalization remains a fundamental challenge due to the lack of standardized evaluation protocols and the limitations of current benchmarks, which often favor efficiency over rigor. At the same time, general-purpose vision architectures lack the necessary inductive biases, and existing approaches to endow them compromise scalability. As a remedy, this paper introduces: 1) a rigorous evaluation framework that unifies and extends previous approaches while reducing computational requirements from combinatorial to constant; 2) an extensive and modern evaluation on the status of compositional generalization in supervised vision backbones, training more than 5000 models; 3) Attribute Invariant Networks, a class of models establishing a new Pareto frontier in compositional generalization, achieving a 23.43% accuracy improvement over baselines while reducing parameter overhead from 600% to 16% compared to fully disentangled counterparts. Our code is available at https://github.com/IBM/scalable-compositional-generalization.
Similar Papers
Scalable Evaluation and Neural Models for Compositional Generalization
Machine Learning (CS)
Teaches computers to understand new things from old.
Learning by Analogy: A Causal Framework for Composition Generalization
Machine Learning (CS)
Lets computers understand new ideas by breaking them down.
Composition-Incremental Learning for Compositional Generalization
CV and Pattern Recognition
Teaches computers to learn new things over time.