A Multi-Modal Hashing Learning Framework for Automatic Image Annotation

被引:8
|
作者
Wang, Jiale [1 ]
Li, Guohui [1 ]
机构
[1] Huazhong Univ Sci & Technol, Sch Comp Sci & Technol, Wuhan 430074, Hubei, Peoples R China
关键词
Image annotation; Semi-Supervised Hashing; two-step method;
D O I
10.1109/DSC.2017.48
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Automatic Image Annotation (AIA) plays an important role in large-scaled intelligent image management and retrieval. Based on the correlation between image low-level features and high-level semantic concepts, images can be efficiently retrieved from large-scaled image dataset. Recently, many researchers leverage machine learning techniques to annotate images automatically. However, these methods still have many challenges regarding efficiency and scalability in the massive image dataset. Moreover, manually labeling massive images is a costly and time-consuming task, which is unacceptable in practical applications. Therefore, only a few labeled images can be obtained as samples in the training dataset. However, the tags associated with labeled and unlabeled images found on social network websites may be helpful for improving the performance of AIA. In this work, we propose a Multi-Modal Semantic Hash Learning framework named MMSHL for AIA. MMSHL seamlessly integrates multi-graph learning, multimodal correlation learning and latent semantic hashing learning into a joint optimization framework. Based on MMSHL, we annotate images using a two-step semi-supervised learning approach. Since our AIA method makes use of associated tags of images, good results can be achieved. Extensive experiments are performed based on two real-world datasets MIR Flickr and NUS-WIDE. Experimental results show that our framework can improve the performance of AIA effectively.
引用
收藏
页码:14 / 21
页数:8
相关论文
共 50 条
  • [1] Semantic relationships in multi-modal graphs for automatic image annotation
    Stathopoulos, Vassilios
    Urban, Jana
    Jose, Joemon
    ADVANCES IN INFORMATION RETRIEVAL, 2008, 4956 : 490 - 497
  • [2] A Framework for Enabling Unpaired Multi-Modal Learning for Deep Cross-Modal Hashing Retrieval
    Williams-Lekuona, Mikel
    Cosma, Georgina
    Phillips, Iain
    JOURNAL OF IMAGING, 2022, 8 (12)
  • [3] Multi-modal multi-concept-based deep neural network for automatic image annotation
    Xu, Haijiao
    Huang, Changqin
    Huang, Xiaodi
    Huang, Muxiong
    MULTIMEDIA TOOLS AND APPLICATIONS, 2019, 78 (21) : 30651 - 30675
  • [4] Multi-modal multi-concept-based deep neural network for automatic image annotation
    Haijiao Xu
    Changqin Huang
    Xiaodi Huang
    Muxiong Huang
    Multimedia Tools and Applications, 2019, 78 : 30651 - 30675
  • [5] SUPERVISED MULTI-MODAL TOPIC MODEL FOR IMAGE ANNOTATION
    Tran, Thu Hoai
    Choi, Seungjin
    2014 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2014,
  • [6] Multi-modal feature fusion for geographic image annotation
    Li, Ke
    Zou, Changqing
    Bu, Shuhui
    Liang, Yun
    Zhang, Jian
    Gong, Minglun
    PATTERN RECOGNITION, 2018, 73 : 1 - 14
  • [7] Sparse Multi-Modal Topical Coding for Image Annotation
    Song, Lingyun
    Luo, Minnan
    Liu, Jun
    Zhang, Lingling
    Qian, Buyue
    Li, Max Haifei
    Zheng, Qinghua
    NEUROCOMPUTING, 2016, 214 : 162 - 174
  • [8] Sparse Multi-Modal Hashing
    Wu, Fei
    Yu, Zhou
    Yang, Yi
    Tang, Siliang
    Zhang, Yin
    Zhuang, Yueting
    IEEE TRANSACTIONS ON MULTIMEDIA, 2014, 16 (02) : 427 - 439
  • [9] Multi-Modal Multi-Scale Deep Learning for Large-Scale Image Annotation
    Niu, Yulei
    Lu, Zhiwu
    Wen, Ji-Rong
    Xiang, Tao
    Chang, Shih-Fu
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2019, 28 (04) : 1720 - 1731
  • [10] Flexible Dual Multi-Modal Hashing for Incomplete Multi-Modal Retrieval
    Wei, Yuhong
    An, Junfeng
    INTERNATIONAL JOURNAL OF IMAGE AND GRAPHICS, 2024,