ADR: An Adversarial Approach to Learn Decomposed Representations for Causal Inference

被引:0
作者
Zheng, Xiangyu [1 ]
Tian, Guogang [1 ]
Wang, Sen [1 ]
Huang, Zhixiang [1 ]
机构
[1] JD Technol, Beijing, Peoples R China
来源
MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, PT II, ECML PKDD 2024 | 2024年 / 14942卷
关键词
Causal Inference; Observational Data; Representation Learning; Adversarial Learning;
D O I
10.1007/978-3-031-70344-7_16
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Estimating the individual treatment effect (ITE) from observational data is an important issue both theoretically and practically. While including all the pre-treatment covariates is the common practice for the inclusion of all possible confounders, it may aggravate the issue of data imbalance. In this paper, we theoretically show that including extra information would increase the variance lower bound. Based on the causal graph, we decompose the covariates into three components, namely instrumental variables (I), confounders (C), and adjustment variables (A). Both C and A should be included for the ITE estimation, while I should be avoided since it would aggravate the imbalance issue and contains no extra information for the ITE estimation. To facilitate the decomposed representation learning, we derive the probabilistic conditions for {I, C, A} from the graphical definitions, and theoretically show that such decomposition can be learned in an adversarial manner. Under the guidance of such theoretical justification, we propose the ADR algorithm, an adversarial learning approach to learn the decomposed representations and simultaneously estimate the treatment effect. The proposed algorithm can be applied to both categorical and numerical treatments and the effectiveness is assured by both theoretical analyses and empirical results. Experimental results on both synthetic and real data show that the ADR Algorithm is advantageous compared to the state-of-the-art methods. The theoretical analyses also provide a path to further explore the issue of decomposed representation learning for causal inference.
引用
收藏
页码:268 / 284
页数:17
相关论文
共 34 条
[1]  
Alaa Ahmed M, 2017, Advances in neural information processing systems, V30
[2]  
Chauhan VK, 2023, PR MACH LEARN RES, V206, P837
[3]  
Cramer H., 1999, Mathematical Methods of Statistics, V26
[4]  
Dorie Vincent, 2016, NPCI: Non-parametrics for Causal Inference
[5]  
Goodfellow IJ, 2014, ADV NEUR IN, V27, P2672
[6]   A decision support framework to implement optimal personalized marketing interventions [J].
Guelman, Leo ;
Guillen, Montserrat ;
Perez-Marin, Ana M. .
DECISION SUPPORT SYSTEMS, 2015, 72 :24-32
[7]   On the role of the propensity score in efficient semiparametric estimation of average treatment effects [J].
Hahn, JY .
ECONOMETRICA, 1998, 66 (02) :315-331
[8]  
Hassanpour N., 2020, INT C LEARNING REPRE
[9]  
Hassanpour N, 2019, PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P5880
[10]   Bayesian Nonparametric Modeling for Causal Inference [J].
Hill, Jennifer L. .
JOURNAL OF COMPUTATIONAL AND GRAPHICAL STATISTICS, 2011, 20 (01) :217-240