Towards Identifiability of Hierarchical Temporal Causal Representation Learning
By: Zijian Li , Minghao Fu , Junxian Huang and more
Potential Business Impact:
Lets computers understand complex patterns over time.
Modeling hierarchical latent dynamics behind time series data is critical for capturing temporal dependencies across multiple levels of abstraction in real-world tasks. However, existing temporal causal representation learning methods fail to capture such dynamics, as they fail to recover the joint distribution of hierarchical latent variables from \textit{single-timestep observed variables}. Interestingly, we find that the joint distribution of hierarchical latent variables can be uniquely determined using three conditionally independent observations. Building on this insight, we propose a Causally Hierarchical Latent Dynamic (CHiLD) identification framework. Our approach first employs temporal contextual observed variables to identify the joint distribution of multi-layer latent variables. Sequentially, we exploit the natural sparsity of the hierarchical structure among latent variables to identify latent variables within each layer. Guided by the theoretical results, we develop a time series generative model grounded in variational inference. This model incorporates a contextual encoder to reconstruct multi-layer latent variables and normalize flow-based hierarchical prior networks to impose the independent noise condition of hierarchical latent dynamics. Empirical evaluations on both synthetic and real-world datasets validate our theoretical claims and demonstrate the effectiveness of CHiLD in modeling hierarchical latent dynamics.
Similar Papers
Spatio-Temporal Hierarchical Causal Models
Machine Learning (Stat)
Finds hidden causes in changing data.
Spatio-Temporal Hierarchical Causal Models
Machine Learning (Stat)
Finds hidden causes in changing data.
Causal Discovery for Linear DAGs with Dependent Latent Variables via Higher-order Cumulants
Machine Learning (CS)
Finds hidden causes in data, even when they're tricky.