Uncertainty Quantification for Text Classification

被引:2
作者
Zhang, Dell [1 ]
Sensoy, Murat [2 ]
Makrehchi, Masoud [1 ]
Taneva-Popova, Bilyana [3 ]
机构
[1] Thomson Reuters Labs, London, England
[2] Amazon Alexa AI, London, England
[3] Thomson Reuters Labs, Zug, Switzerland
来源
ADVANCES IN INFORMATION RETRIEVAL, ECIR 2023, PT III | 2023年 / 13982卷
关键词
Uncertainty quantification; Text classification;
D O I
10.1007/978-3-031-28241-6_38
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This half-day tutorial introduces modern techniques for practical uncertainty quantification specifically in the context of multiclass and multi-label text classification. First, we explain the usefulness of estimating aleatoric uncertainty and epistemic uncertainty for text classification models. Then, we describe several state-of-the-art approaches to uncertainty quantification and analyze their scalability to big text data: Virtual Ensemble in GBDT, Bayesian Deep Learning (including Deep Ensemble, Monte-Carlo Dropout, Bayes by Backprop, and their generalization Epistemic Neural Networks), as well as Evidential Deep Learning (including Prior Networks and Posterior Networks). Next, we discuss typical application scenarios of uncertainty quantification in text classification (including in-domain calibration, cross-domain robustness, and novel class detection). Finally, we list popular performance metrics for the evaluation of uncertainty quantification effectiveness in text classification. Practical hands-on examples/exercises are provided to the attendees for them to experiment with different uncertainty quantification methods on a few real-world text classification datasets such as CLINC150.
引用
收藏
页码:362 / 369
页数:8
相关论文
共 41 条
  • [1] A review of uncertainty quantification in deep learning: Techniques, applications and challenges
    Abdar, Moloud
    Pourpanah, Farhad
    Hussain, Sadiq
    Rezazadegan, Dana
    Liu, Li
    Ghavamzadeh, Mohammad
    Fieguth, Paul
    Cao, Xiaochun
    Khosravi, Abbas
    Acharya, U. Rajendra
    Makarenkov, Vladimir
    Nahavandi, Saeid
    [J]. INFORMATION FUSION, 2021, 76 : 243 - 297
  • [2] Arora G., 2020, P 1 WORKSHOP INSIGHT, P100, DOI 10.18653/v1/2020.insights-1.16
  • [3] Arora U, 2021, 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), P10687
  • [4] Blundell C, 2015, PR MACH LEARN RES, V37, P1613
  • [5] Cerutti F., 2022, IJCAI INT JOINT C AR, P5418, DOI 10.24963/ijcai.2022/760
  • [6] Handling epistemic and aleatory uncertainties in probabilistic circuits
    Cerutti, Federico
    Kaplan, Lance M.
    Kimmig, Angelika
    Sensoy, Murat
    [J]. MACHINE LEARNING, 2022, 111 (04) : 1259 - 1301
  • [7] Charpentier Bertrand, 2020, ADV NEUR IN, V33
  • [8] Foong AYK, 2019, Arxiv, DOI arXiv:1906.11537
  • [9] Gal Y, 2016, PR MACH LEARN RES, V48
  • [10] Gawlikowski J., 2021, ARXIV