Multi-Task Consistency-Preserving Adversarial Hashing for Cross-Modal Retrieval

被引:159
作者
Xie, De [1 ]
Deng, Cheng [1 ]
Li, Chao [1 ]
Liu, Xianglong [2 ]
Tao, Dacheng [3 ,4 ]
机构
[1] Xidian Univ, Sch Elect Engn, Xian 710071, Peoples R China
[2] Beihang Univ, State Key Lab Software Dev Environm, Beijing 100191, Peoples R China
[3] Univ Sydney, UBTECH Sydney Artificial Intelligence Ctr, Fac Engn, Darlington, NSW 2008, Australia
[4] Univ Sydney, Sch Comp Sci, Fac Engn, Darlington, NSW 2008, Australia
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Cross-modal retrieval; hashing; consistency-preserving; adversarial; multi-task; NETWORK;
D O I
10.1109/TIP.2020.2963957
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Owing to the advantages of low storage cost and high query efficiency, cross-modal hashing has received increasing attention recently. As failing to bridge the inherent modality gap between modalities, most existing cross-modal hashing methods have limited capability to explore the semantic consistency information between different modality data, leading to unsatisfactory search performance. To address this problem, we propose a novel deep hashing method named Multi-Task Consistency-Preserving Adversarial Hashing (CPAH) to fully explore the semantic consistency and correlation between different modalities for efficient cross-modal retrieval. First, we design a consistency refined module (CR) to divide the representations of different modality into two irrelevant parts, i.e., modality-common and modality-private representations. Then, a multi-task adversarial learning module (MA) is presented, which can make the modality-common representation of different modalities close to each other on feature distribution and semantic consistency. Finally, the compact and powerful hash codes can be generated from modality-common representation. Comprehensive evaluations conducted on three representative cross-modal benchmark datasets illustrate our method is superior to the state-of-the-art cross-modal hashing methods.
引用
收藏
页码:3626 / 3637
页数:12
相关论文
共 48 条
[1]  
[Anonymous], IEEE T CYBERN
[2]  
[Anonymous], representation learning with deep convolutional generative
[3]  
[Anonymous], 2016, ADV NEUR IN
[4]  
[Anonymous], P CVPR
[5]  
[Anonymous], ADV NEURAL INFORM PR
[6]  
[Anonymous], 2015, IEEE T IMAGE PROCESS, DOI DOI 10.1109/TIP.2015.2467315
[7]  
[Anonymous], 2014, Advances in Neural Information Processing Systems
[8]   Representation Learning: A Review and New Perspectives [J].
Bengio, Yoshua ;
Courville, Aaron ;
Vincent, Pascal .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2013, 35 (08) :1798-1828
[9]  
Bronstein MM, 2010, PROC CVPR IEEE, P3594, DOI 10.1109/CVPR.2010.5539928
[10]   HashGAN: Deep Learning to Hash with Pair Conditional Wasserstein GAN [J].
Cao, Yue ;
Liu, Bin ;
Long, Mingsheng ;
Wang, Jianmin .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :1287-1296