An Evaluation of Representation Learning Methods in Particle Physics Foundation Models
By: Michael Chen , Raghav Kansal , Abhijith Gandrakota and more
Potential Business Impact:
Teaches computers to understand tiny particles better.
We present a systematic evaluation of representation learning objectives for particle physics within a unified framework. Our study employs a shared transformer-based particle-cloud encoder with standardized preprocessing, matched sampling, and a consistent evaluation protocol on a jet classification dataset. We compare contrastive (supervised and self-supervised), masked particle modeling, and generative reconstruction objectives under a common training regimen. In addition, we introduce targeted supervised architectural modifications that achieve state-of-the-art performance on benchmark evaluations. This controlled comparison isolates the contributions of the learning objective, highlights their respective strengths and limitations, and provides reproducible baselines. We position this work as a reference point for the future development of foundation models in particle physics, enabling more transparent and robust progress across the community.
Similar Papers
Universally Converging Representations of Matter Across Scientific Foundation Models
Machine Learning (CS)
Models learn a common "language" for matter.
FM4NPP: A Scaling Foundation Model for Nuclear and Particle Physics
Machine Learning (CS)
Helps scientists find tiny particles faster.
Learning Reduced Representations for Quantum Classifiers
Quantum Physics
Helps quantum computers learn from huge amounts of data.