Testing Random Effects for Binomial Data
By: Lucas Kania, Larry Wasserman, Sivaraman Balakrishnan
Potential Business Impact:
Helps scientists combine study results more accurately.
In modern scientific research, small-scale studies with limited participants are increasingly common. However, interpreting individual outcomes can be challenging, making it standard practice to combine data across studies using random effects to draw broader scientific conclusions. In this work, we introduce an optimal methodology for assessing the goodness of fit between a given reference distribution and the distribution of random effects arising from binomial counts. Using the minimax framework, we characterize the smallest separation between the null and alternative hypotheses, called the critical separation, under the 1-Wasserstein distance that ensures the existence of a valid and powerful test. The optimal test combines a plug-in estimator of the Wasserstein distance with a debiased version of Pearson's chi-squared test. We focus on meta-analyses, where a key question is whether multiple studies agree on a treatment's effectiveness before pooling data. That is, researchers must determine whether treatment effects are homogeneous across studies. We begin by analyzing scenarios with a specified reference effect, such as testing whether all studies show the treatment is effective 80% of the time, and describe how the critical separation depends on the reference effect. We then extend the analysis to homogeneity testing without a reference effect and construct an optimal test by debiasing Cochran's chi-squared test. Finally, we illustrate how our proposed methodologies improve the construction of p-values and confidence intervals, with applications to assessing drug safety in the context of rare adverse outcomes and modeling political outcomes at the county level.
Similar Papers
A New Approach to the Nonparametric Behrens-Fisher Problem with Compatible Confidence Intervals
Methodology
Compares groups fairly, even with different data types.
Graph-theoretic Inference for Random Effects in High-dimensional Studies
Methodology
Finds hidden patterns in complex data.
Size-adaptive Hypothesis Testing for Fairness
Machine Learning (CS)
Checks if computer programs are unfair to groups.