Model-assisted analyses of cluster-randomized experiments

被引:28
作者
Su, Fangzhou [1 ]
Ding, Peng [1 ]
机构
[1] Univ Calif Berkeley, Dept Stat, Berkeley, CA 94720 USA
基金
美国国家科学基金会;
关键词
analysis of covariance; design-based inference; efficiency-robustness trade-off; group-randomized trial; potential outcomes; robust standard error; RECENT METHODOLOGICAL DEVELOPMENTS; REGRESSION ADJUSTMENTS; TRIALS; INFERENCE; NONCOMPLIANCE; ESTIMATOR; EDUCATION; DESIGN;
D O I
10.1111/rssb.12468
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
Cluster-randomized experiments are widely used due to their logistical convenience and policy relevance. To analyse them properly, we must address the fact that the treatment is assigned at the cluster level instead of the individual level. Standard analytic strategies are regressions based on individual data, cluster averages and cluster totals, which differ when the cluster sizes vary. These methods are often motivated by models with strong and unverifiable assumptions, and the choice among them can be subjective. Without any outcome modelling assumption, we evaluate these regression estimators and the associated robust standard errors from the design-based perspective where only the treatment assignment itself is random and controlled by the experimenter. We demonstrate that regression based on cluster averages targets a weighted average treatment effect, regression based on individual data is suboptimal in terms of efficiency and regression based on cluster totals is consistent and more efficient with a large number of clusters. We highlight the critical role of covariates in improving estimation efficiency and illustrate the efficiency gain via both simulation studies and data analysis. The asymptotic analysis also reveals the efficiency-robustness trade-off by comparing the properties of various estimators using data at different levels with and without covariate adjustment. Moreover, we show that the robust standard errors are convenient approximations to the true asymptotic standard errors under the design-based perspective. Our theory holds even when the outcome models are misspecified, so it is model-assisted rather than model-based. We also extend the theory to a wider class of weighted average treatment effects.
引用
收藏
页码:994 / 1015
页数:22
相关论文
共 51 条
[1]   When Should You Adjust Standard Errors for Clustering?* [J].
Abadie, Alberto ;
Athey, Susan ;
Imbens, Guido W. ;
Wooldridge, Jeffrey M. .
QUARTERLY JOURNAL OF ECONOMICS, 2022, 138 (01) :1-35
[2]  
Angrist JD, 2009, MOSTLY HARMLESS ECONOMETRICS: AN EMPIRICISTS COMPANION, P3
[3]  
Athey S., 2017, Handbook of economic field experiments, V1, P73
[4]   Clustering, Spatial Correlations, and Randomization Inference [J].
Barrios, Thomas ;
Diamond, Rebecca ;
Imbens, Guido W. ;
Kolesar, Michal .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2012, 107 (498) :578-591
[5]   Analyzing Two-Stage Experiments in the Presence of Interference [J].
Basse, Guillaume ;
Feller, Avi .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2018, 113 (521) :41-55
[6]   Lasso adjustments of treatment effect estimates in randomized experiments [J].
Bloniarz, Adam ;
Liu, Hanzhong ;
Zhang, Cun-Hui ;
Sekhon, Jasjeet S. ;
Yu, Bin .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2016, 113 (27) :7383-7390
[7]   A Practitioner's Guide to Cluster-Robust Inference [J].
Cameron, A. Colin ;
Miller, Douglas L. .
JOURNAL OF HUMAN RESOURCES, 2015, 50 (02) :317-372
[8]   RANDOMIZATION BY GROUP - FORMAL ANALYSIS [J].
CORNFIELD, J .
AMERICAN JOURNAL OF EPIDEMIOLOGY, 1978, 108 (02) :100-102
[9]   Causal inference from 2K factorial designs by using potential outcomes [J].
Dasgupta, Tirthankar ;
Pillai, Natesh S. ;
Rubin, Donald B. .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-STATISTICAL METHODOLOGY, 2015, 77 (04) :727-753
[10]   RANK TESTS IN UNMATCHED CLUSTERED RANDOMIZED TRIALS APPLIED TO A STUDY OF TEACHER TRAINING [J].
Ding, Peng ;
Keele, Luke .
ANNALS OF APPLIED STATISTICS, 2018, 12 (04) :2151-2174