Vector-Decomposed Disentanglement for Domain-Invariant Object Detection

被引:77
作者
Wu, Aming [1 ,2 ]
Liu, Rui [1 ,2 ]
Han, Yahong [1 ,2 ,3 ]
Zhu, Linchao [4 ]
Yang, Yi [4 ]
机构
[1] Tianjin Univ, Coll Intelligence & Comp, Tianjin, Peoples R China
[2] Tianjin Univ, Tianjin Key Lab Machine Learning, Tianjin, Peoples R China
[3] Peng Cheng Lab, Shenzhen, Peoples R China
[4] Univ Technol Sydney, AAII, ReLER Lab, Sydney, NSW, Australia
来源
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021) | 2021年
关键词
D O I
10.1109/ICCV48922.2021.00921
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
To improve the generalization of detectors, for domain adaptive object detection (DAOD), recent advances mainly explore aligning feature-level distributions between the source and single-target domain, which may neglect the impact of domain-specific information existing in the aligned features. Towards DAOD, it is important to extract domain-invariant object representations. To this end, in this paper, we try to disentangle domain-invariant representations from domain-specific representations. And we propose a novel disentangled method based on vector decomposition. Firstly, an extractor is devised to separate domain-invariant representations from the input, which are used for extracting object proposals. Secondly, domain-specific representations are introduced as the differences between the input and domain-invariant representations. Through the difference operation, the gap between the domain-specific and domain-invariant representations is enlarged, which promotes domain-invariant representations to contain more domain-irrelevant information. In the experiment, we separately evaluate our method on the single- and compound-target case. For the single-target case, experimental results of four domain-shift scenes show our method obtains a significant performance gain over baseline methods. Moreover, for the compound-target case (i.e., the target is a compound of two different domains without domain labels), our method outperforms baseline methods by around 4%, which demonstrates the effectiveness of our method.
引用
收藏
页码:9322 / 9331
页数:10
相关论文
共 46 条
[11]  
Ganin Y., 2014, ARXIV14097495
[12]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[13]  
He Zhenwei, 2020, EUR C COMP VIS
[14]  
He Zhenwei, 2019, ARXIV190710343
[15]  
Higgins I, 2018, Towards a definition of disentangled representations
[16]   Cross-Domain Weakly-Supervised Object Detection through Progressive Domain Adaptation [J].
Inoue, Naoto ;
Furuta, Ryosuke ;
Yamasaki, Toshihiko ;
Aizawa, Kiyoharu .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :5001-5009
[17]   Self-Training and Adversarial Background Regularization for Unsupervised Domain Adaptive One-Stage Object Detection [J].
Kim, Seunghyeon ;
Choi, Jaehoon ;
Kim, Taekyung ;
Kim, Changick .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :6091-6100
[18]   Diversify and Match: A Domain Adaptive Representation Learning Paradigm for Object Detection [J].
Kim, Taekyung ;
Jeong, Minki ;
Kim, Seunghyeon ;
Choi, Seokeon ;
Kim, Changick .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :12448-12457
[19]   Diverse Image-to-Image Translation via Disentangled Representations [J].
Lee, Hsin-Ying ;
Tseng, Hung-Yu ;
Huang, Jia-Bin ;
Singh, Maneesh ;
Yang, Ming-Hsuan .
COMPUTER VISION - ECCV 2018, PT I, 2018, 11205 :36-52
[20]   Joint Adversarial Domain Adaptation [J].
Li, Shuang ;
Liu, Chi Harold ;
Xie, Binhui ;
Su, Limin ;
Ding, Zhengming ;
Huang, Gao .
PROCEEDINGS OF THE 27TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA (MM'19), 2019, :729-737