CROSS-MODAL LEARNING TO RANK WITH ADAPTIVE LISTWISE CONSTRAINT

被引:0
作者
Qu, Guangzhuo [1 ]
Xiao, Jing [1 ]
Zhu, Jia [1 ]
Cao, Yang [1 ]
Huang, Changqin [2 ]
机构
[1] South China Normal Univ, Sch Comp Sci, Guangzhou, Guangdong, Peoples R China
[2] South China Normal Univ, Sch Informat Technol Educ, Guangzhou, Guangdong, Peoples R China
来源
2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP) | 2018年
基金
中国国家自然科学基金;
关键词
Cross-modal retrieval; common space; adaptive listwise theory; cross-modal learning to rank;
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Multi-modal data lies on heterogeneous feature spaces, which brings a significant challenge to cross-modal retrieval. Some works have been proposed to cope with this problem by learning a common subspace. However, previous methods often learn the common subspace by enhancing the relation between embedded features and relevant class labels but ignore the relation between embedded features and irrelevant class labels. Additionally, most methods assume that irrelevant samples are of equal importance. Considering this, we propose to train an optimal common embedding space via cross-modal learning to rank with adaptive listwise constraint (CMAL(2)R) based on two-branch neural networks. The listwise loss function in CMAL(2)R adaptively assigns larger margins to harder irrelevant samples, strengthening the relation between embedded features and irrelevant class labels. Experiments on Wikipedia and Pascal datasets demonstrate the effectiveness for bi-directional image-text retrieval.
引用
收藏
页码:1658 / 1662
页数:5
相关论文
共 50 条
  • [31] Deep adversarial metric learning for cross-modal retrieval
    Xing Xu
    Li He
    Huimin Lu
    Lianli Gao
    Yanli Ji
    World Wide Web, 2019, 22 : 657 - 672
  • [32] Deep Hashing Similarity Learning for Cross-Modal Retrieval
    Ma, Ying
    Wang, Meng
    Lu, Guangyun
    Sun, Yajun
    IEEE ACCESS, 2024, 12 : 8609 - 8618
  • [33] Scalable Deep Multimodal Learning for Cross-Modal Retrieval
    Hu, Peng
    Zhen, Liangli
    Peng, Dezhong
    Liu, Pei
    PROCEEDINGS OF THE 42ND INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '19), 2019, : 635 - 644
  • [34] Dual Subspaces with Adversarial Learning for Cross-Modal Retrieval
    Xia, Yaxian
    Wang, Wenmin
    Han, Liang
    ADVANCES IN MULTIMEDIA INFORMATION PROCESSING, PT I, 2018, 11164 : 654 - 663
  • [35] Semantic supervised learning based Cross-Modal Retrieval
    Li, Zhuoyi
    Fu, Hao
    Gu, Guanghua
    PROCEEDINGS OF THE ACM TURING AWARD CELEBRATION CONFERENCE-CHINA 2024, ACM-TURC 2024, 2024, : 207 - 209
  • [36] Discriminative semantic transitive consistency for cross-modal learning
    Parida, Kranti Kumar
    Sharma, Gaurav
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2022, 219
  • [37] LEARNING A CROSS-MODAL HASHING NETWORK FOR MULTIMEDIA SEARCH
    Liong, Venice Erin
    Lu, Jiwen
    Tan, Yap-Peng
    2017 24TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2017, : 3700 - 3704
  • [38] Graph Embedding Learning for Cross-Modal Information Retrieval
    Zhang, Youcai
    Gu, Xiaodong
    NEURAL INFORMATION PROCESSING (ICONIP 2017), PT III, 2017, 10636 : 594 - 601
  • [39] Adaptive Asymmetric Supervised Cross-Modal Hashing with consensus matrix
    Li, Yinan
    Long, Jun
    Huang, Youyuan
    Yang, Zhan
    INFORMATION PROCESSING & MANAGEMENT, 2025, 62 (03)
  • [40] Adaptive Marginalized Semantic Hashing for Unpaired Cross-Modal Retrieval
    Luo, Kaiyi
    Zhang, Chao
    Li, Huaxiong
    Jia, Xiuyi
    Chen, Chunlin
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 9082 - 9095