Object Hallucination Detection in Large Vision Language Models via Evidential Conflict

被引:0
作者
Liu, Zhekun [1 ,2 ]
Huang, Tao [1 ,2 ]
Wang, Rui [1 ,2 ]
Jing, Liping [1 ,2 ]
机构
[1] Beijing Jiaotong Univ, Sch Comp Sci & Technol, Beijing, Peoples R China
[2] Beijing Key Lab Traff Data Anal & Min, Beijing, Peoples R China
来源
BELIEF FUNCTIONS: THEORY AND APPLICATIONS, BELIEF 2024 | 2024年 / 14909卷
关键词
LVLM; object hallucination; uncertainty quantification; Dempster-Shafer theory;
D O I
10.1007/978-3-031-67977-3_7
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Despite their remarkable ability to understand both textual and visual data, large vision-language models (LVLMs) still face issues with hallucination. This is particularly presented as the object hallucination, where the models inaccurately describe objects in the images. Current efforts mainly focus on detecting such erroneous behaviors through the semantic consistency of outputs via multiple inferences or by evaluating the entropy-based uncertainty of predictions. However, the former is resource-intensive, while the latter is often considered a less precise measure due to generally recognized overconfident predictions. To address the issue, we propose an object hallucination detection method based on evidential conflict. To be specific, we view the features in the last layer of the transformer decoder as evidence. Then, we combine the evidence based on Dempster's rule, following the approach presented in the work [6]. Hence, this enables us to detect hallucinations by evaluating the conflict among evidence. Preliminary experiments were conducted on a state-of-the-art LVLM, mPLUG-Owl2. Results show that our approach exhibits an enhancement over baseline methods, particularly in cases with highly uncertain inputs.
引用
收藏
页码:58 / 67
页数:10
相关论文
共 28 条
  • [1] Achiam OJ, 2023, Arxiv, DOI [arXiv:2303.08774, 10.48550/arXiv.2303.08774, DOI 10.48550/ARXIV.2303.08774]
  • [2] Anderljung M, 2023, Arxiv, DOI arXiv:2307.03718
  • [3] Brown TB, 2020, ADV NEUR IN, V33
  • [4] Dai Wenliang, 2024, Advances in Neural Information Processing Systems, V36
  • [5] Dempster AP, 2008, STUD FUZZ SOFT COMP, V219, P57
  • [6] Logistic regression, neural networks and Dempster-Shafer theory: A new perspective
    Denoeux, Thierry
    [J]. KNOWLEDGE-BASED SYSTEMS, 2019, 176 : 54 - 67
  • [7] Dosovitskiy A., 2021, P 9 INT C LEARN REPR
  • [8] Falke T, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P2214
  • [9] Gal Y, 2016, PR MACH LEARN RES, V48
  • [10] Guerreiro NM, 2023, 17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, P1059