Score: 1

Physically Interpretable Representation Learning with Gaussian Mixture Variational AutoEncoder (GM-VAE)

Published: November 26, 2025 | arXiv ID: 2511.21883v1

By: Tiffany Fan , Murray Cutforth , Marta D'Elia and more

BigTech Affiliations: Stanford University

Potential Business Impact:

Finds hidden patterns in messy science data.

Business Areas:
Image Recognition Data and Analytics, Software

Extracting compact, physically interpretable representations from high-dimensional scientific data is a persistent challenge due to the complex, nonlinear structures inherent in physical systems. We propose a Gaussian Mixture Variational Autoencoder (GM-VAE) framework designed to address this by integrating an Expectation-Maximization (EM)-inspired training scheme with a novel spectral interpretability metric. Unlike conventional VAEs that jointly optimize reconstruction and clustering (often leading to training instability), our method utilizes a block-coordinate descent strategy, alternating between expectation and maximization steps. This approach stabilizes training and naturally aligns latent clusters with distinct physical regimes. To objectively evaluate the learned representations, we introduce a quantitative metric based on graph-Laplacian smoothness, which measures the coherence of physical quantities across the latent manifold. We demonstrate the efficacy of this framework on datasets of increasing complexity: surface reaction ODEs, Navier-Stokes wake flows, and experimental laser-induced combustion Schlieren images. The results show that our GM-VAE yields smooth, physically consistent manifolds and accurate regime clustering, offering a robust data-driven tool for interpreting turbulent and reactive flow systems.

Country of Origin
🇺🇸 United States

Page Count
24 pages

Category
Computer Science:
Machine Learning (CS)