Generalized attention-based deep multi-instance learning

被引:4
作者
Zhao, Lu [1 ]
Yuan, Liming [2 ]
Hao, Kun [1 ]
Wen, Xianbin [2 ]
机构
[1] Tianjin Chengjian Univ, Sch Comp & Informat Engn, 26 Jinjing Rd, Tianjin 300384, Peoples R China
[2] Tianjin Univ Technol, Sch Comp Sci & Engn, 391 Bin Shui Xi Dao Rd, Tianjin 300384, Peoples R China
基金
中国国家自然科学基金;
关键词
Multi-instance learning; Deep learning; Similarity-based loss; Attention-based pooling; Interpretability; NEURAL-NETWORKS;
D O I
10.1007/s00530-022-00992-w
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Attention-based deep multi-instance learning (MIL) is an effective and interpretable model. Its interpretability is attributed to the learnability of its inner attention-based MIL pooling. Its main problem is to learn a unique instance-level target concept for weighting instances. Another implicative issue is to assume that the bag and instance concepts are located in the same semantic space. In this paper, we relax these constraints as: (i) There exist multiple instance concepts; (ii) The bag and instance concepts live in different semantic spaces. Upon the two relaxed constraints, we propose a two-level attention-based MIL pooling that first learns several instance concepts in a low-level semantic space and subsequently captures the bag concept in a high-level semantic space. To effectively capture different types of instance concepts, we also present a new similarity-based loss. The experimental results show that our method achieves higher or very comparable performance with state-of-the-art methods on benchmark data sets and surpasses them in terms of performance and interpretability on a synthetic data set.
引用
收藏
页码:275 / 287
页数:13
相关论文
共 37 条