Causally-Aware Information Bottleneck for Domain Adaptation
By: Mohammad Ali Javidian
Potential Business Impact:
Fills in missing data in changing situations.
We tackle a common domain adaptation setting in causal systems. In this setting, the target variable is observed in the source domain but is entirely missing in the target domain. We aim to impute the target variable in the target domain from the remaining observed variables under various shifts. We frame this as learning a compact, mechanism-stable representation. This representation preserves information relevant for predicting the target while discarding spurious variation. For linear Gaussian causal models, we derive a closed-form Gaussian Information Bottleneck (GIB) solution. This solution reduces to a canonical correlation analysis (CCA)-style projection and offers Directed Acyclic Graph (DAG)-aware options when desired. For nonlinear or non-Gaussian data, we introduce a Variational Information Bottleneck (VIB) encoder-predictor. This approach scales to high dimensions and can be trained on source data and deployed zero-shot to the target domain. Across synthetic and real datasets, our approach consistently attains accurate imputations, supporting practical use in high-dimensional causal models and furnishing a unified, lightweight toolkit for causal domain adaptation.
Similar Papers
An Expectation-Maximization Algorithm for Domain Adaptation in Gaussian Causal Models
Machine Learning (CS)
Helps computers guess missing information accurately.
A Generalized Information Bottleneck Theory of Deep Learning
Machine Learning (CS)
Helps computers learn better by understanding feature connections.
Comprehensive Information Bottleneck for Unveiling Universal Attribution to Interpret Vision Transformers
CV and Pattern Recognition
Shows how computers make choices by looking everywhere.