Unsupervised Cross-domain Object Detection via Multiple Domain Randomization

被引:1
作者
Luo, Fang [1 ]
Liu, Jie [1 ]
Ho, George To Sum [2 ]
Yan, Kun [1 ]
机构
[1] Wuhan Univ Technol, Coll Comp Sci & Artificial Intelligence, Wuhan, Peoples R China
[2] Hang Seng Univ Hong Kong, Coll Dept Supply Chain & Informat Management, Hong Kong, Peoples R China
来源
PROCEEDINGS OF THE 2024 27 TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, CSCWD 2024 | 2024年
关键词
object detection; domain shift; unsupervised domain adaptation; image translation; domain randomization;
D O I
10.1109/CSCWD61410.2024.10580650
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Cross-domain detection refers to the challenge of detecting objects or patterns belonging to different domains or contexts. The cross-domain detection problem arises when the training data and test data do not subject to independent and identical distribution, leading to a significant decrease in the performance of existing object detection methods. In order to address the aforementioned cross-domain detection problem, this paper proposes an unsupervised cross-domain object detection method based on multi-domain randomization. Firstly, the method utilizes Cycle-GAN to generate multiple randomized domains, enabling the comprehensive learning of the target domain overall's feature distribution. Then, a domain randomization parameter callback module is devised to retain the key detection information of the object, thereby improving the model's stability. Additionally, to alleviate the problem of domain bias and inconsistency between data and labels, a source domain consistency loss is incorporated to enhance the convergence speed of the model and amplify the semantic information embedded within the features. The experimental results on multiple cross-domain datasets show that the proposed method outperforms existing unsupervised cross-domain object detection algorithms in terms of cross-domain detection performance.
引用
收藏
页码:845 / 851
页数:7
相关论文
共 19 条
[11]   Unsupervised Domain Adaptation of Object Detectors: A Survey [J].
Oza, Poojan ;
Sindagi, Vishwanath A. ;
Vibashan, V. S. ;
Patel, Vishal M. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (06) :4018-4040
[12]   Contrastive Learning for Unpaired Image-to-Image Translation [J].
Park, Taesung ;
Efros, Alexei A. ;
Zhang, Richard ;
Zhu, Jun-Yan .
COMPUTER VISION - ECCV 2020, PT IX, 2020, 12354 :319-345
[13]   Diffusion Autoencoders: Toward a Meaningful and Decodable Representation [J].
Preechakul, Konpat ;
Chatthee, Nattanat ;
Wizadwongsa, Suttisak ;
Suwajanakorn, Supasorn .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, :10609-10619
[14]   Strong-Weak Distribution Alignment for Adaptive Object Detection [J].
Saito, Kuniaki ;
Ushiku, Yoshitaka ;
Harada, Tatsuya ;
Saenko, Kate .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :6949-6958
[15]   Semantic Foggy Scene Understanding with Synthetic Data [J].
Sakaridis, Christos ;
Dai, Dengxin ;
Van Gool, Luc .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2018, 126 (09) :973-992
[16]   SeaShips: A Large-Scale Precisely Annotated Dataset for Ship Detection [J].
Shao, Zhenfeng ;
Wu, Wenjing ;
Wang, Zhongyuan ;
Du, Wan ;
Li, Chengyuan .
IEEE TRANSACTIONS ON MULTIMEDIA, 2018, 20 (10) :2593-2604
[17]   Cross-domain Object Detection through Coarse-to-Fine Feature Adaptation [J].
Zheng, Yangtao ;
Huang, Di ;
Liu, Songtao ;
Wang, Yunhong .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :13763-13772
[18]   Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks [J].
Zhu, Jun-Yan ;
Park, Taesung ;
Isola, Phillip ;
Efros, Alexei A. .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :2242-2251
[19]   Adapting Object Detectors via Selective Cross-Domain Alignment [J].
Zhu, Xinge ;
Pang, Jiangmiao ;
Yang, Ceyuan ;
Shi, Jianping ;
Lin, Dahua .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :687-696