Deep Semantic Space with Intra-class Low-rank Constraint for Cross-modal Retrieval

被引:9
|
作者
Kang, Peipei [1 ]
Lin, Zehang [1 ]
Yang, Zhenguo [1 ,2 ]
Fang, Xiaozhao [3 ]
Li, Qing [4 ]
Liu, Wenyin [1 ]
机构
[1] Guangdong Univ Technol, Sch Comp Sci & Technol, Guangzhou, Guangdong, Peoples R China
[2] City Univ Hong Kong, Dept Comp Sci, Hong Kong, Peoples R China
[3] Guangdong Univ Technol, Dept Automat, Guangzhou, Guangdong, Peoples R China
[4] Hong Kong Polytech Univ, Dept Comp, Hong Kong, Peoples R China
来源
ICMR'19: PROCEEDINGS OF THE 2019 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL | 2019年
基金
中国国家自然科学基金;
关键词
cross-modal retrieval; deep neural networks; intra-class low-rank; semantic space;
D O I
10.1145/3323873.3325029
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
In this paper, a novel Deep Semantic Space learning model with Intra-class Low-rank constraint (DSSIL) is proposed for cross-modal retrieval, which is composed of two subnetworks for modality-specific representation learning, followed by projection layers for common space mapping. In particular, DSSIL takes into account semantic consistency to fuse the cross-modal data in a high-level common space, and constrains the common representation matrix within the same class to be low-rank, in order to induce the intra-class representations more relevant. More formally, two regularization terms are devised for the two aspects, which have been incorporated into the objective of DSSIL. To optimize the modality-specific subnetworks and the projection layers simultaneously by exploiting the gradient decent directly, we approximate the nonconvex low-rank constraint by minimizing a few smallest singular values of the intra-class matrix with theoretical analysis. Extensive experiments conducted on three public datasets demonstrate the competitive superiority of DSSIL for cross-modal retrieval compared with the state-of-the-art methods.
引用
收藏
页码:226 / 234
页数:9
相关论文
共 50 条
  • [21] Deep Multigraph Hierarchical Enhanced Semantic Representation for Cross-Modal Retrieval
    Zhu, Lei
    Zhang, Chengyuan
    Song, Jiayu
    Zhang, Shichao
    Tian, Chunwei
    Zhu, Xinghui
    IEEE MULTIMEDIA, 2022, 29 (03) : 17 - 26
  • [22] Label-Based Deep Semantic Hashing for Cross-Modal Retrieval
    Weng, Weiwei
    Wu, Jiagao
    Yang, Lu
    Liu, Linfeng
    Hu, Bin
    NEURAL INFORMATION PROCESSING (ICONIP 2019), PT III, 2019, 11955 : 24 - 36
  • [23] Deep Cross-modal Hashing Based on Intra-modal Similarity and Semantic Preservation
    Li T.
    Liu L.
    Data Analysis and Knowledge Discovery, 2023, 7 (05) : 105 - 115
  • [24] Deep Supervised Cross-modal Retrieval
    Zhen, Liangli
    Hu, Peng
    Wang, Xu
    Peng, Dezhong
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 10386 - 10395
  • [25] Semantic deep cross-modal hashing
    Lin, Qiubin
    Cao, Wenming
    He, Zhihai
    He, Zhiquan
    NEUROCOMPUTING, 2020, 396 (396) : 113 - 122
  • [26] Semantic consistency hashing for cross-modal retrieval
    Yao, Tao
    Kong, Xiangwei
    Fu, Haiyan
    Tian, Qi
    NEUROCOMPUTING, 2016, 193 : 250 - 259
  • [27] Analyzing semantic correlation for cross-modal retrieval
    Liang Xie
    Peng Pan
    Yansheng Lu
    Multimedia Systems, 2015, 21 : 525 - 539
  • [28] Analyzing semantic correlation for cross-modal retrieval
    Xie, Liang
    Pan, Peng
    Lu, Yansheng
    MULTIMEDIA SYSTEMS, 2015, 21 (06) : 525 - 539
  • [29] Specific class center guided deep hashing for cross-modal retrieval
    Shu, Zhenqiu
    Bai, Yibing
    Zhang, Donglin
    Yu, Jun
    Yu, Zhengtao
    Wu, Xiao-Jun
    INFORMATION SCIENCES, 2022, 609 : 304 - 318
  • [30] CROSS-MODAL LEARNING TO RANK WITH ADAPTIVE LISTWISE CONSTRAINT
    Qu, Guangzhuo
    Xiao, Jing
    Zhu, Jia
    Cao, Yang
    Huang, Changqin
    2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2018, : 1658 - 1662