Online Fast Adaptive Low-Rank Similarity Learning for Cross-Modal Retrieval

被引:16
|
作者
Wu, Yiling [1 ]
Wang, Shuhui [1 ]
Huang, Qingming [1 ,2 ]
机构
[1] Chinese Acad Sci, Key Lab Intelligent Informat Proc, Inst Comp Technol, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Sch Comp Sci & Technol, Beijing 101408, Peoples R China
基金
中国国家自然科学基金;
关键词
Semantics; Correlation; Training; Data models; Visualization; Adaptation models; Fasteners; Cross-modality learning; similarity function learning; online learning; low-rank matrix; IMAGES;
D O I
10.1109/TMM.2019.2942494
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The semantic similarity among cross-modal data objects, e.g., similarities between images and texts, are recognized as the bottleneck of cross-modal retrieval. However, existing batch-style correlation learning methods suffer from prohibitive time complexity and extra memory consumption in handling large-scale high dimensional cross-modal data. In this paper, we propose a Cross-Modal Online Low-Rank Similarity function learning (CMOLRS) method, which learns a low-rank bilinear similarity measurement for cross-modal retrieval. We model the cross-modal relations by relative similarities on the training data triplets and formulate the relative relations as convex hinge loss. By adapting the margin in hinge loss with pair-wise distances in feature space and label space, CMOLRS effectively captures the multi-level semantic correlation and adapts to the content divergence among cross-modal data. Imposed with a low-rank constraint, the similarity function is trained by online learning in the manifold of low-rank matrices. The low-rank constraint not only endows the model learning process with faster speed and better scalability, but also improves the model generality. We further propose fast-CMOLRS combining multiple triplets for each query instead of standard process using single triplet at each model update step, which further reduces the times of gradient updates and retractions. Extensive experiments are conducted on four public datasets, and comparisons with state-of-the-art methods show the effectiveness and efficiency of our approach.
引用
收藏
页码:1310 / 1322
页数:13
相关论文
共 50 条
  • [1] ONLINE LOW-RANK SIMILARITY FUNCTION LEARNING WITH ADAPTIVE RELATIVE MARGIN FOR CROSS-MODAL RETRIEVAL
    Wu, Yiling
    Wang, Shuhui
    Zhang, Weigang
    Huang, Qingming
    2017 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2017, : 823 - 828
  • [2] Reconstruction regularized low-rank subspace learning for cross-modal retrieval
    Wu, Jianlong
    Xie, Xingxu
    Nie, Liqiang
    Lin, Zhouchen
    Zha, Hongbin
    PATTERN RECOGNITION, 2021, 113
  • [3] Online Asymmetric Similarity Learning for Cross-Modal Retrieval
    Wu, Yiling
    Wang, Shuhui
    Huang, Qingming
    30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 3984 - 3993
  • [4] HCMSL: Hybrid Cross-modal Similarity Learning for Cross-modal Retrieval
    Zhang, Chengyuan
    Song, Jiayu
    Zhu, Xiaofeng
    Zhu, Lei
    Zhang, Shichao
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2021, 17 (01)
  • [5] Deep Hashing Similarity Learning for Cross-Modal Retrieval
    Ma, Ying
    Wang, Meng
    Lu, Guangyun
    Sun, Yajun
    IEEE ACCESS, 2024, 12 : 8609 - 8618
  • [6] Deep Semantic Space with Intra-class Low-rank Constraint for Cross-modal Retrieval
    Kang, Peipei
    Lin, Zehang
    Yang, Zhenguo
    Fang, Xiaozhao
    Li, Qing
    Liu, Wenyin
    ICMR'19: PROCEEDINGS OF THE 2019 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, 2019, : 226 - 234
  • [7] CROSS-MODAL LEARNING TO RANK WITH ADAPTIVE LISTWISE CONSTRAINT
    Qu, Guangzhuo
    Xiao, Jing
    Zhu, Jia
    Cao, Yang
    Huang, Changqin
    2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2018, : 1658 - 1662
  • [8] Hashing for Cross-Modal Similarity Retrieval
    Liu, Yao
    Yuan, Yanhong
    Huang, Qiaoli
    Huang, Zhixing
    2015 11TH INTERNATIONAL CONFERENCE ON SEMANTICS, KNOWLEDGE AND GRIDS (SKG), 2015, : 1 - 8
  • [9] Adaptive Adversarial Learning based cross-modal retrieval
    Li, Zhuoyi
    Lu, Huibin
    Fu, Hao
    Wang, Zhongrui
    Gu, Guanghun
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 123
  • [10] DRSL: Deep Relational Similarity Learning for Cross-modal Retrieval
    Wang, Xu
    Hu, Peng
    Zhen, Liangli
    Peng, Dezhong
    INFORMATION SCIENCES, 2021, 546 : 298 - 311