Domain-aware triplet loss in domain generalization

被引:3
作者
Guo, Kaiyu [1 ]
Lovell, Brian C. [1 ]
机构
[1] Univ Queensland, Brisbane, Qld 4072, Australia
关键词
Domain generalization; Contrastive learning; Domain dispersion;
D O I
10.1016/j.cviu.2024.103979
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Despite the considerable advances in deep learning for object recognition, there are still several factors that hinder the performance of deep learning models. One of these factors is domain shift, which occurs due to variations in the distribution of the testing and training data. This paper addresses the issue of compact feature clustering in domain generalization, with the aim of optimizing the embedding space from multi -domain data. Specifically, we propose a domain -aware triplet loss for domain generalization, which not only facilitates clustering of similar semantic features but also disperses features that arise from the domain. Unlike previous methods that focus on aligning distributions, our algorithm disperses domain information in the embedding space. Our approach is based on the assumption that embedding features can be clustered based on domain information, which is supported mathematically and empirically in this paper. Furthermore, in our investigation of feature clustering in domain generalization, we observe that the factors that influence the convergence of metric learning loss in domain generalization are more significant than the pre -defined domains. To address this issue, we utilize two methods to normalize the embedding space and reduce the internal covariate shift of the embedding features. Our ablation study illustrates the effectiveness of our algorithm. Additionally, our experiments on benchmark datasets, including PACS, VLCS, and Office -Home, demonstrate that our method outperforms related approaches that focus on domain discrepancy. Notably, our results on RegnetY-16GF are substantially better than state-of-the-art methods on the benchmark datasets. Our code is available at https://github.com/workerbcd/DCT.
引用
收藏
页数:8
相关论文
共 53 条
[1]  
Albuquerque, 2019, ARXIV
[2]  
[Anonymous], 2006, P IEEE COMP SOC C CO, P1735
[3]  
Arjovsky M, 2020, Arxiv, DOI [arXiv:1907.02893, 10.48550/arXiv.1907.02893]
[4]  
Blanchard G., 2011, ADV NEURAL INFORM PR, V24
[5]  
Cha J, 2021, ADV NEUR IN
[6]  
Cha J, 2022, LECT NOTES COMPUT SC, V13683, P440
[7]   A Style and Semantic Memory Mechanism for Domain Generalization [J].
Chen, Yang ;
Wang, Yu ;
Pan, Yingwei ;
Yao, Ting ;
Tian, Xinmei ;
Mei, Tao .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :9144-9153
[8]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[9]  
Dosovitskiy Alexey, 2021, ICLR
[10]   Unbiased Metric Learning: On the Utilization of Multiple Datasets and Web Images for Softening Bias [J].
Fang, Chen ;
Xu, Ye ;
Rockmore, Daniel N. .
2013 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2013, :1657-1664