A Non-Bipartite Matching Framework for Difference-in-Differences with General Treatment Types
By: Siyu Heng, Yuan Huang, Hyunseung Kang
Potential Business Impact:
Finds true effects of changing things over time.
Difference-in-differences (DID) is one of the most widely used causal inference frameworks in observational studies. However, most existing DID methods are designed for binary treatments and cannot be readily applied to non-binary treatment settings. Although recent work has begun to extend DID to non-binary (e.g., continuous) treatments, these approaches typically require strong additional assumptions, including parametric outcome models or the presence of idealized comparison units with (nearly) static treatment levels over time (commonly called ``stayers'' or ``quasi-stayers''). In this technical note, we introduce a new non-bipartite matching framework for DID that naturally accommodates general treatment types (e.g., binary, ordinal, or continuous). Our framework makes three main contributions. First, we develop an optimal non-bipartite matching design for DID that jointly balances baseline covariates across comparable units (reducing bias) and maximizes contrasts in treatment trajectories over time (improving efficiency). Second, we establish a post-matching randomization condition, the design-based counterpart to the traditional parallel-trends assumption, which enables valid design-based inference. Third, we introduce the sample average DID ratio, a finite-population-valid and fully nonparametric causal estimand applicable to arbitrary treatment types. Our design-based approach that preserves the full treatment-dose information, avoids parametric assumptions, does not rely on the existence of stayers or quasi-stayers, and operates entirely within a finite-population framework, without appealing to hypothetical super-populations or outcome distributions.
Similar Papers
Estimating treatment effects with a unified semi-parametric difference-in-differences approach
Methodology
Finds true effects even with messy data.
Efficient Difference-in-Differences Estimation when Outcomes are Missing at Random
Methodology
Fixes studies when some information is missing.
Bias-Variance Tradeoff of Matching Prior to Difference-in-Differences When Parallel Trends is Violated
Methodology
Helps businesses make better decisions with data.