Significance Tests of Feature Relevance for a Black-Box Learner

被引:18
作者
Dai, Ben [1 ]
Shen, Xiaotong [2 ]
Pan, Wei [3 ]
机构
[1] Chinese Univ Hong Kong, Dept Stat, Hong Kong, Peoples R China
[2] Univ Minnesota, Sch Stat, Minneapolis, MN 55455 USA
[3] Univ Minnesota, Div Biostat, Minneapolis, MN 55455 USA
关键词
Adaptive splitting; black-box tests; combining; computational constraints; feature relevance; NEURAL-NETWORKS;
D O I
10.1109/TNNLS.2022.3185742
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
An exciting recent development is the uptake of deep neural networks in many scientific fields, where the main objective is outcome prediction with a black-box nature. Significance testing is promising to address the black-box issue and explore novel scientific insights and interpretations of the decision-making process based on a deep learning model. However, testing for a neural network poses a challenge because of its black-box nature and unknown limiting distributions of parameter estimates while existing methods require strong assumptions or excessive computation. In this article, we derive one-split and two-split tests relaxing the assumptions and computational complexity of existing black-box tests and extending to examine the significance of a collection of features of interest in a dataset of possibly a complex type, such as an image. The one-split test estimates and evaluates a black-box model based on estimation and inference subsets through sample splitting and data perturbation. The two-split test further splits the inference subset into two but requires no perturbation. Also, we develop their combined versions by aggregating the p-values based on repeated sample splitting. By deflating the bias-sd-ratio, we establish asymptotic null distributions of the test statistics and the consistency in terms of Type II error. Numerically, we demonstrate the utility of the proposed tests on seven simulated examples and six real datasets. Accompanying this article is our python library dnn-inference (https://dnn-inference.readthedocs.io/en/latest/) that implements the proposed tests.
引用
收藏
页码:1898 / 1911
页数:14
相关论文
共 36 条
[1]  
Arik SO, 2021, AAAI CONF ARTIF INTE, V35, P6679
[2]   The conditional permutation test for independence while controlling for confounders [J].
Berrett, Thomas B. ;
Wang, Yi ;
Barber, Rina Foygel ;
Samworth, Richard J. .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-STATISTICAL METHODOLOGY, 2020, 82 (01) :175-197
[3]   Random forests [J].
Breiman, L .
MACHINE LEARNING, 2001, 45 (01) :5-32
[5]   Panning for gold: "model-X' knockoffs for high dimensional controlled variable selection [J].
Candes, Emmanuel ;
Fan, Yingying ;
Janson, Lucas ;
Lv, Jinchi .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-STATISTICAL METHODOLOGY, 2018, 80 (03) :551-577
[6]  
Cappe O., 2006, Inference in Hidden Markov Models
[7]   Double/debiased machine learning for treatment and structural parameters [J].
Chernozhukov, Victor ;
Chetverikov, Denis ;
Demirer, Mert ;
Duflo, Esther ;
Hansen, Christian ;
Newey, Whitney ;
Robins, James .
ECONOMETRICS JOURNAL, 2018, 21 (01) :C1-C68
[8]   Discovering the anticancer potential of non-oncology drugs by systematic viability profiling [J].
Corsello, Steven M. ;
Nagari, Rohith T. ;
Spangler, Ryan D. ;
Rossen, Jordan ;
Kocak, Mustafa ;
Bryan, Jordan G. ;
Humeidi, Ranad ;
Peck, David ;
Wu, Xiaoyun ;
Tang, Andrew A. ;
Wang, Vickie M. ;
Bender, Samantha A. ;
Lemire, Evan ;
Narayan, Rajiv ;
Montgomery, Philip ;
Ben-David, Uri ;
Garvie, Colin W. ;
Chen, Yejia ;
Rees, Matthew G. ;
Lyons, Nicholas J. ;
McFarland, James M. ;
Wong, Bang T. ;
Wang, Li ;
Dumont, Nancy ;
O'Hearn, Patrick J. ;
Stefan, Eric ;
Doench, John G. ;
Harrington, Caitlin N. ;
Greulich, Heidi ;
Meyerson, Matthew ;
Vazquez, Francisca ;
Subramanian, Aravind ;
Roth, Jennifer A. ;
Bittker, Joshua A. ;
Boehm, Jesse S. ;
Mader, Christopher C. ;
Tsherniak, Aviad ;
Golub, Todd R. .
NATURE CANCER, 2020, 1 (02) :235-+
[9]  
Dodge Y., 2006, The Oxford dictionary of statistical terms
[10]  
Fahrmeir Ludwig, 2007, Regression