A Multidisciplinary Survey and Framework for Design and Evaluation of Explainable AI Systems

被引:292
作者
Mohseni, Sina [1 ,3 ]
Zarei, Niloofar [1 ,3 ]
Ragan, Eric D. [2 ]
机构
[1] Texas A&M Univ, College Stn, TX 77843 USA
[2] Univ Florida, E301 CSE Bldg, Gainesville, FL 32611 USA
[3] B208 Langford Bldg,3137 TAMU, College Stn, TX 77840 USA
基金
美国国家科学基金会;
关键词
Explainable artificial intelligence (XAI); human-computer interaction (HCI); machine learning; explanation; transparency; VISUAL ANALYTICS; MENTAL MODELS; PART; EXPLANATION; TRUST; INTERPRETABILITY; ACCOUNTABILITY; VISUALIZATION; TRANSPARENCY; PREDICTION;
D O I
10.1145/3387166
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The need for interpretable and accountable intelligent systems grows along with the prevalence of artificial intelligence (AI) applications used in everyday life. Explainable AI (XAI) systems are intended to selfexplain the reasoning behind system decisions and predictions. Researchers from different disciplines work together to define, design, and evaluate explainable systems. However, scholars from different disciplines focus on different objectives and fairly independent topics of XAI research, which poses challenges for identifying appropriate design and evaluation methodology and consolidating knowledge across efforts. To this end, this article presents a survey and framework intended to share knowledge and experiences of XAI design and evaluation methods across multiple disciplines. Aiming to support diverse design goals and evaluation methods in XAI research, after a thorough review of XAI related papers in the fields of machine learning, visualization, and human-computer interaction, we present a categorization of XAI design goals and evaluation methods. Our categorization presents the mapping between design goals for different XAI user groups and their evaluation methods. From our findings, we develop a framework with step-by-step design guidelines paired with evaluation methods to close the iterative design and evaluation cycles in multidisciplinary XAI teams. Further, we provide summarized ready-to-use tables of evaluation methods and recommendations for different goals in XAI research.
引用
收藏
页数:45
相关论文
共 225 条
  • [1] Trends and Trajectories for Explainable, Accountable and Intelligible Systems: An HCI Research Agenda
    Abdul, Ashraf
    Vermeulen, Jo
    Wang, Danding
    Lim, Brian
    Kankanhalli, Mohan
    [J]. PROCEEDINGS OF THE 2018 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS (CHI 2018), 2018,
  • [2] Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
    Adadi, Amina
    Berrada, Mohammed
    [J]. IEEE ACCESS, 2018, 6 : 52138 - 52160
  • [3] Adebayo J, 2018, ADV NEUR IN, V31
  • [4] FairSight: Visual Analytics for Fairness in Decision Making
    Ahn, Yongsu
    Lin, Yu-Ru
    [J]. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2020, 26 (01) : 1086 - 1095
  • [5] Task-Driven Comparison of Topic Models
    Alexander, Eric
    Gleicher, Michael
    [J]. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2016, 22 (01) : 320 - 329
  • [6] Guidelines for Human-AI Interaction
    Amershi, Saleema
    Weld, Dan
    Vorvoreanu, Mihaela
    Fourney, Adam
    Nushi, Besmira
    Collisson, Penny
    Suh, Jina
    Iqbal, Shamsi
    Bennett, Paul N.
    Inkpen, Kori
    Teevan, Jaime
    Kikin-Gil, Ruth
    Horvitz, Eric
    [J]. CHI 2019: PROCEEDINGS OF THE 2019 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS, 2019,
  • [7] Power to the People: The Role of Humans in Interactive Machine Learning
    Amershi, Saleema
    Cakmak, Maya
    Knox, W. Bradley
    Kulesza, Todd
    [J]. AI MAGAZINE, 2014, 35 (04) : 105 - 120
  • [8] Amodei D., 2016, ARXIV160606565
  • [9] Seeing without knowing: Limitations of the transparency ideal and its application to algorithmic accountability
    Ananny, Mike
    Crawford, Kate
    [J]. NEW MEDIA & SOCIETY, 2018, 20 (03) : 973 - 989
  • [10] [Anonymous], 2017, 11 INT AAAI C WEB SO, DOI DOI 10.1609/ICWSM.V11I1.14898