Model selection and inference for estimation of causal parameters

被引:0
作者
Rothenhausler, Dominik [1 ]
机构
[1] Stanford Univ, Dept Stat, Stanford, CA 94305 USA
来源
ELECTRONIC JOURNAL OF STATISTICS | 2024年 / 18卷 / 02期
关键词
Causal inference; model selection; data fusion; efficiency; CROSS-VALIDATION;
D O I
10.1214/24-EJS2308
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
In causal inference there are often multiple reasonable estimators for a given target quantity. For example, one may reasonably use inverse probability weighting, an instrumental variables approach, or construct an estimate based on proxy outcomes if the actual outcome is difficult to measure. Ideally, the practitioner decides on an estimator before looking at the data. However, this might be challenging in practice since a priori it might not be clear to a practitioner how to choose the method. If the final model is chosen after peeking at the data, naive inferential procedures may fail. This raises the need for a model selection tool, with rigorous asymptotic guarantees. Since there is usually no loss function available in causal inference, standard model selection techniques do not apply. We propose a model selection procedure that estimates the squared pound 2- deviation of a finite-dimensional estimator from its target. The procedure relies on knowing an asymptotically unbiased (potentially highly variable) estimate of the parameter of interest. The resulting estimator is discontinuous and does not have a Gaussian limit distribution. Thus, standard asymptotic expansions do not apply. We derive asymptotically valid confidence intervals for low-dimensional settings that take into account the model selection step. The performance of the approach for estimation and inference for average treatment effects is evaluated on simulated data sets in low-dimensional settings, including experimental data, instrumental variables settings and observational data with selection on observables.
引用
收藏
页码:5449 / 5483
页数:35
相关论文
共 50 条
[21]   A Bayesian Model for Bivariate Causal Inference [J].
Kurthen, Maximilian ;
Ensslin, Torsten .
ENTROPY, 2020, 22 (01) :46
[22]   Covariate selection strategies for causal inference: Classification and comparison [J].
Witte, Janine ;
Didelez, Vanessa .
BIOMETRICAL JOURNAL, 2019, 61 (05) :1270-1289
[23]   An evolutionary algorithm for subset selection in causal inference models [J].
Cho, Wendy K. Tam .
JOURNAL OF THE OPERATIONAL RESEARCH SOCIETY, 2018, 69 (04) :630-644
[24]   Minimum Distance Estimation of Milky Way Model Parameters and Related Inference [J].
Banerjee, Sourabh ;
Basu, Ayanendranath ;
Bhattacharya, Sourabh ;
Bose, Smarajit ;
Chakrabarty, Dalia ;
Mukherjee, Soumendu Sundar .
SIAM-ASA JOURNAL ON UNCERTAINTY QUANTIFICATION, 2015, 3 (01) :91-115
[25]   Likelihood-based inference for bounds of causal parameters [J].
Lee, Woojoo ;
Sjolander, Arvid ;
Larsson, Anton ;
Pawitan, Yudi .
STATISTICS IN MEDICINE, 2018, 37 (30) :4695-4706
[26]   Complexity and Approximation Results for the Balance Optimization Subset Selection Model for Causal Inference in Observational Studies [J].
Sauppe, Jason J. ;
Jacobson, Sheldon H. ;
Sewell, Edward C. .
INFORMS JOURNAL ON COMPUTING, 2014, 26 (03) :547-566
[27]   A Causal Dirichlet Mixture Model for Causal Inference from Observational Data [J].
Lin, Adi ;
Lu, Jie ;
Xuan, Junyu ;
Zhu, Fujin ;
Zhang, Guangquan .
ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2020, 11 (03)
[28]   Applying the structural causal model framework for observational causal inference in ecology [J].
Arif, Suchinta ;
MacNeil, M. Aaron .
ECOLOGICAL MONOGRAPHS, 2023, 93 (01)
[29]   Outcome-adaptive lasso: Variable selection for causal inference [J].
Shortreed, Susan M. ;
Ertefaie, Ashkan .
BIOMETRICS, 2017, 73 (04) :1111-1122
[30]   Causal Inference under Interference and Model Uncertainty [J].
Zhang, Chi ;
Mohan, Karthika ;
Pearl, Judea .
CONFERENCE ON CAUSAL LEARNING AND REASONING, VOL 213, 2023, 213 :371-385