Adversary-Free Counterfactual Prediction via Information-Regularized Representations
By: Shiqin Tang , Rong Feng , Shuxin Zhuang and more
Potential Business Impact:
Makes computer predictions fair even with biased data.
We study counterfactual prediction under assignment bias and propose a mathematically grounded, information-theoretic approach that removes treatment-covariate dependence without adversarial training. Starting from a bound that links the counterfactual-factual risk gap to mutual information, we learn a stochastic representation Z that is predictive of outcomes while minimizing I(Z; T). We derive a tractable variational objective that upper-bounds the information term and couples it with a supervised decoder, yielding a stable, provably motivated training criterion. The framework extends naturally to dynamic settings by applying the information penalty to sequential representations at each decision time. We evaluate the method on controlled numerical simulations and a real-world clinical dataset, comparing against recent state-of-the-art balancing, reweighting, and adversarial baselines. Across metrics of likelihood, counterfactual error, and policy evaluation, our approach performs favorably while avoiding the training instabilities and tuning burden of adversarial schemes.
Similar Papers
The Causal Round Trip: Generating Authentic Counterfactuals by Eliminating Information Loss
Machine Learning (CS)
Lets computers understand why things happen.
Counterfactual Reward Model Training for Bias Mitigation in Multimodal Reinforcement Learning
Machine Learning (CS)
Makes AI fairer by removing hidden biases.
Unifying Image Counterfactuals and Feature Attributions with Latent-Space Adversarial Attacks
Machine Learning (CS)
Shows why computers see what they see.