Mining multiple visual appearances of semantics for image annotation

被引:0
作者
Tan, Hung-Khoon [1 ]
Ngo, Chong-Wah [1 ]
机构
[1] City Univ Hong Kong, Dept Comp Sci, Kowloon, Hong Kong, Peoples R China
来源
ADVANCES IN MULTIMEDIA MODELING, PT 1 | 2007年 / 4351卷
关键词
image annotation; multiple-instance learning; Apriori;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper investigates the problem of learning the visual semantics of keyword categories for automatic image annotation. Supervised learning algorithms which learn only a single concept point of a category are limited in their effectiveness for image annotation. We propose to use data mining techniques to mine multiple concepts, where each concept may consist of one or more visual parts, to capture the diverse visual appearances of a single keyword category. For training, we use the Apriori principle to efficiently mine a set of frequent blobsets to capture the semantics of a rich and diverse visual category. Each concept is ranked based on a discriminative or diverse density measure. For testing, we propose a level-sensitive matching to rank words given an unannotated image. Our approach is effective, scales better during training and testing, and is efficient in terms of learning and annotation.
引用
收藏
页码:269 / 278
页数:10
相关论文
共 14 条
[1]  
[Anonymous], 2003, **NON-TRADITIONAL**
[2]  
Blei D., 2003, P 26 ANN INT ACM SIG, P127, DOI DOI 10.1145/860435.860460
[3]  
CARNEIRO G, 2005, CVPR, V2, P1463
[4]  
DUYGULU D, 2002, ECCV, P97
[5]  
Feng SL, 2004, PROC CVPR IEEE, P1002
[6]  
Ghoshal A., 2005, SIGIR 2005. Proceedings of the Twenty-Eighth Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, P544, DOI 10.1145/1076034.1076127
[7]  
JEON J, 2003, SIGIR, P119
[8]  
MARON O, 1998, ICML, P341
[9]  
MORI Y, 1999, IMAGE TO WORD TRANFO
[10]  
SHI R, 2006, CIVR, P102