Knowledge Graphs for Empirical Concept Retrieval

被引:0
作者
Tetkova, Lenka [1 ]
Scheidt, Teresa Karen [1 ]
Fogh, Maria Mandrup [1 ]
Jorgensen, Ellen Marie Gaunby [1 ]
Nielsen, Finn Arup [1 ]
Hansen, Lars Kai [1 ]
机构
[1] Tech Univ Denmark, Sect Cognit Syst, DTU Compute, DK-2800 Lyngby, Denmark
来源
EXPLAINABLE ARTIFICIAL INTELLIGENCE, PT I, XAI 2024 | 2024年 / 2153卷
关键词
Concept-based Explainability; Knowledge Graphs; Concepts;
D O I
10.1007/978-3-031-63787-2_9
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Concept-based explainable AI is promising as a tool to improve the understanding of complex models at the premises of a given user, viz. as a tool for personalized explainability. An important class of concept-based explainability methods is constructed with empirically defined concepts, indirectly defined through a set of positive and negative examples, as in the TCAV approach. While it is appealing to the user to avoid formal definitions of concepts and their operationalization, it can be challenging to establish relevant concept datasets. Here, we address this challenge using general knowledge graphs (such as, e.g., Wikidata or WordNet) for comprehensive concept definition and present a workflow for user-driven data collection in both text and image domains. The concepts derived from knowledge graphs are defined interactively, providing an opportunity for personalization and ensuring that the concepts reflect the user's intentions. We test the retrieved concept datasets on two concept-based explainability methods, namely concept activation vectors (CAVs) and concept activation regions (CARs). We show that CAVs and CARs based on these empirical concept datasets provide robust and accurate explanations. Importantly, we also find good alignment between the models' representations of concepts and the structure of knowledge graphs, i.e., human representations. This supports our conclusion that knowledge graph-based concepts are relevant for XAI.
引用
收藏
页码:160 / 183
页数:24
相关论文
共 52 条
[1]  
Baevski A, 2022, PR MACH LEARN RES
[2]   Network Dissection: Quantifying Interpretability of Deep Visual Representations [J].
Bau, David ;
Zhou, Bolei ;
Khosla, Aditya ;
Oliva, Aude ;
Torralba, Antonio .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :3319-3327
[3]  
Brown TB, 2020, ADV NEUR IN, V33
[4]  
Christian B., 2021, The alignment problem: How can machines learn human values?
[5]  
Crabbe Jonathan, 2022, ADV NEURAL INFORM PR
[6]  
Demszky D, 2020, 58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), P4040
[7]  
Devlin J, 2019, Arxiv, DOI arXiv:1810.04805
[8]  
Dosovitskiy A, 2021, Arxiv, DOI arXiv:2010.11929
[9]  
Everingham M., The PASCAL Visual Object Classes Challenge 2007 (VOC2007) Results
[10]  
Fauconnier G., 1994, Mental Spaces: Aspects of Meaning Construction in Natural Language, DOI 10.1017/CBO9780511624582