Conformal Prediction: A Gentle Introduction

被引:81
作者
Angelopoulos, Anastasios N. [1 ]
Bates, Stephen [1 ]
机构
[1] Univ Calif Berkeley, Berkeley, CA USA
来源
FOUNDATIONS AND TRENDS IN MACHINE LEARNING | 2023年 / 16卷 / 04期
基金
美国国家科学基金会;
关键词
CONFIDENCE MACHINES; INFORMATION-THEORY; REGRESSION; FOUNDATIONS; INFERENCE; QUANTILES; MODELS; BANDS;
D O I
10.1561/2200000101
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Black-box machine learning models are now routinely used in high-risk settings, like medical diagnostics, which demand uncertainty quantification to avoid consequential model failures. Conformal prediction (a.k.a. conformal inference) is a user-friendly paradigm for creating statistically rigorous uncertainty sets/intervals for the predictions of such models. Critically, the sets are valid in a distribution-free sense: they possess explicit, non-asymptotic guarantees even without distributional assumptions or model assumptions. One can use conformal prediction with any pre-trained model, such as a neural network, to produce sets that are guaranteed to contain the ground truth with a user-specified probability, such as 90%. It is easy-to-understand, easy-to-use, and general, applying naturally to problems arising in the fields of computer vision, natural language processing, deep reinforcement learning, and so on. This hands-on introduction is aimed to provide the reader a working understanding of conformal prediction and related distribution-free uncertainty quantification techniques with one self-contained document. We lead the reader through practical theory for and examples of conformal prediction and describe its extensions to complex machine learning tasks involving structured outputs, distribution shift, time-series, outliers, models that abstain, and more. Throughout, there are many explanatory illustrations, examples, and code samples in Python. With each code sample comes a Jupyter notebook implementing the method on a real-data example; the notebooks can be accessed and easily run by following the code footnotes.
引用
收藏
页码:494 / 591
页数:98
相关论文
共 125 条
  • [1] Aldous D. J., 1985, Ecole d'Ete de Probabilites de Saint-Flour XIII-1983
  • [2] Angelopoulos A. N., 2021, INT C LEARN REPR
  • [3] Angelopoulos AN, 2021, ARXIV
  • [4] Angelopoulos Anastasios N., 2022, arXiv
  • [5] [Anonymous], 2003, P 20 INT C MACHINE L
  • [6] Barber R. F., 2022, arXiv
  • [7] PREDICTIVE INFERENCE WITH THE JACKKNIFE
    Barber, Rina Foygel
    Candes, Emmanuel J.
    Ramdas, Aaditya
    Tibshirani, Ryan J.
    [J]. ANNALS OF STATISTICS, 2021, 49 (01) : 486 - 507
  • [8] The limits of distribution-free conditional predictive inference
    Barber, Rina Foygel
    Candes, Emmanuel J.
    Ramdas, Aaditya
    Tibshirani, Ryan J.
    [J]. INFORMATION AND INFERENCE-A JOURNAL OF THE IMA, 2021, 10 (02) : 455 - 482
  • [9] Bastani Osbert, 2022, Advances in Neural Information Processing Systems
  • [10] Bates S., 2021, arXiv