Learning to Validate Generative Models: a Goodness-of-Fit Approach
By: Pietro Cappelli , Gaia Grosso , Marco Letizia and more
Potential Business Impact:
Checks if AI models for science are trustworthy.
Generative models are increasingly central to scientific workflows, yet their systematic use and interpretation require a proper understanding of their limitations through rigorous validation. Classic approaches struggle with scalability, statistical power, or interpretability when applied to high-dimensional data, making it difficult to certify the reliability of these models in realistic, high-dimensional scientific settings. Here, we propose the use of the New Physics Learning Machine (NPLM), a learning based approach to goodness-of-fit testing inspired by the Neyman-Pearson construction, to test generative networks trained on high-dimensional scientific data. We demonstrate the performance of NPLM for validation in two benchmark cases: generative models trained on mixtures of Gaussian models with increasing dimensionality, and a public end-to-end generator for the Large Hadron Collider called FlashSim, trained on jet data, typical in the field of high-energy physics. We demonstrate that the NPLM can serve as a powerful validation method while also providing a means to diagnose sub-optimally modeled regions of the data.
Similar Papers
Comparing Generative Models with the New Physics Learning Machine
Machine Learning (Stat)
Checks if computer-made data looks real.
Neural Modular Physics for Elastic Simulation
Machine Learning (CS)
Makes computer simulations of stretchy things more real.
Limits To (Machine) Learning
Machine Learning (Stat)
Finds hidden money patterns machines miss.