Human attention guided explainable artificial intelligence for computer vision models

被引:2
|
作者
Liu, Guoyang [1 ,5 ]
Zhang, Jindi [3 ]
Chan, Antoni B. [4 ]
Hsiao, Janet H. [2 ,5 ]
机构
[1] Shandong Univ, Sch Integrated Circuits, Jinan, Peoples R China
[2] Hong Kong Univ Sci & Technol, Div Social Sci, Clearwater Bay, Hong Kong, Peoples R China
[3] Huawei Res, Hong Kong, Peoples R China
[4] City Univ Hong Kong, Dept Comp Sci, Kowloon Tong, Hong Kong, Peoples R China
[5] Univ Hong Kong, Dept Psychol, Pokfulam Rd, Hong Kong, Peoples R China
关键词
Object detection; XAI; Human attention; Deep learning; Saliency map; VISUAL-ATTENTION;
D O I
10.1016/j.neunet.2024.106392
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Explainable artificial intelligence (XAI) has been increasingly investigated to enhance the transparency of black -box artificial intelligence models, promoting better user understanding and trust. Developing an XAI that is faithful to models and plausible to users is both a necessity and a challenge. This work examines whether embedding human attention knowledge into saliency -based XAI methods for computer vision models could enhance their plausibility and faithfulness. Two novel XAI methods for object detection models, namely FullGrad-CAM and FullGrad-CAM++, were first developed to generate object -specific explanations by extending the current gradient -based XAI methods for image classification models. Using human attention as the objective plausibility measure, these methods achieve higher explanation plausibility. Interestingly, all current XAI methods when applied to object detection models generally produce saliency maps that are less faithful to the model than human attention maps from the same object detection task. Accordingly, human attention -guided XAI (HAG-XAI) was proposed to learn from human attention how to best combine explanatory information from the models to enhance explanation plausibility by using trainable activation functions and smoothing kernels to maximize the similarity between XAI saliency map and human attention map. The proposed XAI methods were evaluated on widely used BDD-100K, MS-COCO, and ImageNet datasets and compared with typical gradientbased and perturbation -based XAI methods. Results suggest that HAG-XAI enhanced explanation plausibility and user trust at the expense of faithfulness for image classification models, and it enhanced plausibility, faithfulness, and user trust simultaneously and outperformed existing state-of-the-art XAI methods for object detection models.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] Explainable Artificial Intelligence for Simulation Models
    Grigoryan, Gayane
    PROCEEDINGS OF THE 38TH ACM SIGSIM INTERNATIONAL CONFERENCE ON PRINCIPLES OF ADVANCED DISCRETE SIMULATION, ACM SIGSIM-PADS 2024, 2024, : 59 - 60
  • [2] Automated Human Vision Assessment using Computer Vision and Artificial Intelligence
    Van Eenwyk, Jonathan
    Agah, Arvin
    Cibis, Gerhard W.
    2008 IEEE INTERNATIONAL CONFERENCE ON SYSTEM OF SYSTEMS ENGINEERING (SOSE), 2008, : 317 - +
  • [3] A Systematic Review of Human-Computer Interaction and Explainable Artificial Intelligence in Healthcare With Artificial Intelligence Techniques
    Nazar, Mobeen
    Alam, Muhammad Mansoor
    Yafi, Eiad
    Su'ud, Mazliham Mohd
    IEEE ACCESS, 2021, 9 : 153316 - 153348
  • [4] Automated human vision assessment using computer vision and artificial intelligence
    Department of Electrical Engineering and Computer Science, University of Kansas, Lawrence, KS, United States
    不详
    IEEE Int. Conf. Syst. Syst. Eng., SoSE, 2008,
  • [5] Computer Vision With Explainable Artificial Intelligence for Visual Pollution Detection in the Kingdom of Saudi Arabia
    Al Mazroa, Alanoud
    Maray, Mohammed
    Alashjaee, Abdullah M.
    Alotaibi, Faiz Abdullah
    Alzahrani, Ahmad A.
    Alkharashi, Abdulwhab
    Alotaibi, Shoayee Dlaim
    Alnfiai, Mrim M.
    IEEE ACCESS, 2024, 12 : 193014 - 193027
  • [6] Artificial intelligence in laparoscopic cholecystectomy: does computer vision outperform human vision?
    Liu, Runwen
    An, Jingjing
    Wang, Ziyao
    Guan, Jingye
    Liu, Jie
    Jiang, Jingwen
    Chen, Zhimin
    Li, Hai
    Peng, Bing
    Wang, Xin
    ARTIFICIAL INTELLIGENCE SURGERY, 2022, 2 (02): : 80 - 92
  • [7] Locality Guided Neural Networks for Explainable Artificial Intelligence
    Tan, Randy
    Khan, Naimul
    Guan, Ling
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [8] Artificial intelligence, Computer Vision and Multimedia
    Zhang, Ning
    RISTI - Revista Iberica de Sistemas e Tecnologias de Informacao, 2016, 2016 (E8):
  • [9] Explainable Emotion Decoding for Human and Computer Vision
    Borriero, Alessio
    Milazzo, Martina
    Diano, Matteo
    Orsenigo, Davide
    Villa, Maria Chiara
    DiFazio, Chiara
    Tamietto, Marco
    Perotti, Alan
    EXPLAINABLE ARTIFICIAL INTELLIGENCE, PT II, XAI 2024, 2024, 2154 : 178 - 201
  • [10] HG-XAI: human-guided tool wear identification approach through augmentation of explainable artificial intelligence with machine vision
    Kumar, Aitha Sudheer
    Agarwal, Ankit
    Jansari, Vinita Gangaram
    Desai, K. A.
    Chattopadhyay, Chiranjoy
    Mears, Laine
    JOURNAL OF INTELLIGENT MANUFACTURING, 2024,