Conformal Prediction: A Gentle Introduction

被引:125
作者
Angelopoulos, Anastasios N. [1 ]
Bates, Stephen [1 ]
机构
[1] Univ Calif Berkeley, Berkeley, CA USA
来源
FOUNDATIONS AND TRENDS IN MACHINE LEARNING | 2023年 / 16卷 / 04期
基金
美国国家科学基金会;
关键词
CONFIDENCE MACHINES; INFORMATION-THEORY; REGRESSION; FOUNDATIONS; INFERENCE; QUANTILES; MODELS; BANDS;
D O I
10.1561/2200000101
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Black-box machine learning models are now routinely used in high-risk settings, like medical diagnostics, which demand uncertainty quantification to avoid consequential model failures. Conformal prediction (a.k.a. conformal inference) is a user-friendly paradigm for creating statistically rigorous uncertainty sets/intervals for the predictions of such models. Critically, the sets are valid in a distribution-free sense: they possess explicit, non-asymptotic guarantees even without distributional assumptions or model assumptions. One can use conformal prediction with any pre-trained model, such as a neural network, to produce sets that are guaranteed to contain the ground truth with a user-specified probability, such as 90%. It is easy-to-understand, easy-to-use, and general, applying naturally to problems arising in the fields of computer vision, natural language processing, deep reinforcement learning, and so on. This hands-on introduction is aimed to provide the reader a working understanding of conformal prediction and related distribution-free uncertainty quantification techniques with one self-contained document. We lead the reader through practical theory for and examples of conformal prediction and describe its extensions to complex machine learning tasks involving structured outputs, distribution shift, time-series, outliers, models that abstain, and more. Throughout, there are many explanatory illustrations, examples, and code samples in Python. With each code sample comes a Jupyter notebook implementing the method on a real-data example; the notebooks can be accessed and easily run by following the code footnotes.
引用
收藏
页码:494 / 591
页数:98
相关论文
共 126 条
[1]  
Aldous D., 1985, Ecole d'ete de probabilites de Saint-Flour, V8, P1
[2]  
Angelopoulos A.N., 2021, INT C LEARN REPR
[3]  
Angelopoulos A. N., 2021, ARXIV
[4]  
Angelopoulos A. N., 2022, arXiv
[5]  
Barber R. F., 2022, arXiv
[6]   PREDICTIVE INFERENCE WITH THE JACKKNIFE [J].
Barber, Rina Foygel ;
Candes, Emmanuel J. ;
Ramdas, Aaditya ;
Tibshirani, Ryan J. .
ANNALS OF STATISTICS, 2021, 49 (01) :486-507
[7]   The limits of distribution-free conditional predictive inference [J].
Barber, Rina Foygel ;
Candes, Emmanuel J. ;
Ramdas, Aaditya ;
Tibshirani, Ryan J. .
INFORMATION AND INFERENCE-A JOURNAL OF THE IMA, 2021, 10 (02) :455-482
[8]  
Bastani O., 2022, Advances in Neural Information Processing Systems
[9]  
Bates S., 2021, arXiv
[10]   Distribution-free, Risk-controlling Prediction Sets [J].
Bates, Stephen ;
Angelopoulos, Anastasios ;
Lei, Lihua ;
Malik, Jitendra ;
Jordan, Michael .
JOURNAL OF THE ACM, 2021, 68 (06)