Unequal adaptive visual recognition by learning from multi-modal data

被引:11
作者
Cai, Ziyun [1 ]
Zhang, Tengfei [1 ]
Jing, Xiao-Yuan [2 ,3 ,4 ]
Shao, Ling [5 ]
机构
[1] Nanjing Univ Posts & Telecommun, Coll Automat, Nanjing 210023, Peoples R China
[2] Wuhan Univ, Sch Comp, Wuhan 430072, Peoples R China
[3] Nanjing Univ, State Key Lab Novel Software Technol, Nanjing, Jiangsu, Peoples R China
[4] Guangdong Univ Petrochem Technol, Sch Comp, Maoming, Peoples R China
[5] Incept Inst Artificial Intelligence, Abu Dhabi, U Arab Emirates
基金
美国国家科学基金会;
关键词
RGB-D data; Domain adaptation; Visual categorization; Unequal categories;
D O I
10.1016/j.ins.2022.03.076
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Conventional domain adaptation tries to leverage knowledge obtained from the single source domain to recognize the data in the target domain, where only one modality exists in the source domain. This neglects the scenario that source domain can be acquired from multi-modal data, such as RGB data and depth data. In addition, conventional domain adaptation approaches generally assume source and target domains have the identical number of categories, which is quite restrict for real-world applications. In practice, the number of categories in the target domain is often less than that in the source domain. In this work, we focus on a more practical and challenging task that recognizes RGB data by learning from RGB-D data under an unequal label scenario, which suffers from three challenges: i) the addition of depth information, ii) the domain mismatch problem and iii) the negative transfer caused by unequal label numbers. Our main contribution is a novel method, referred to as unequal Distribution Visual-Depth Adaption (uDVDA), which takes advantage of depth data and handles domain mismatch problem under label inequality, simultaneously. Experiments show that uDVDA outperforms state-of-the-art models on different datasets, especially under unequal label scenario.(c) 2022 Elsevier Inc. All rights reserved.
引用
收藏
页码:1 / 21
页数:21
相关论文
共 41 条
[1]  
Bo LF, 2011, IEEE INT C INT ROBOT, P821, DOI 10.1109/IROS.2011.6048717
[2]   Open Set Domain Adaptation for Image and Action Recognition [J].
Busto, Pau Panareda ;
Iqbal, Ahsan ;
Gall, Juergen .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2020, 42 (02) :413-429
[3]   Adaptive RGB Image Recognition by Visual-Depth Embedding [J].
Cai, Ziyun ;
Long, Yang ;
Shao, Ling .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2018, 27 (05) :2471-2483
[4]  
Cai Ziyun, 2020, IEEE T CYBERN
[5]   Learning to Transfer Examples for Partial Domain Adaptation [J].
Cao, Zhangjie ;
You, Kaichao ;
Long, Mingsheng ;
Wang, Jianmin ;
Yang, Qiang .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :2980-2989
[6]  
Chen Minming, 2009, UILUENG092215
[7]  
Chen Z., 2020, IEEE C COMP VIS PATT, P12706
[8]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[9]  
Donahue J, 2014, PR MACH LEARN RES, V32
[10]  
Farquhar J.D. R., 2005, NIPS