Exploring Explicitly Disentangled Features for Domain Generalization

被引:11
|
作者
Li, Jingwei [1 ,2 ]
Li, Yuan [1 ,2 ]
Wang, Huanjie [1 ,2 ]
Liu, Chengbao [1 ]
Tan, Jie [1 ,2 ]
机构
[1] Chinese Acad Sci, Inst Automat, Beijing 100080, Peoples R China
[2] Univ Chinese Acad Sci, Sch Artificial Intelligence, Beijing 101408, Peoples R China
基金
中国国家自然科学基金;
关键词
Domain generalization; feature disentanglement; Fourier transform; data augmentation;
D O I
10.1109/TCSVT.2023.3269534
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Domain generalization (DG) is a challenging task that aims to train a robust model with only labeled source data and can generalize well on unseen target data. The domain gap between the source and target data may degrade the performance. A plethora of methods resort to obtaining domain-invariant features to overcome the difficulties. However, these methods require sophisticated network designs or training strategies, causing inefficiency and complexity. In this paper, we first analyze and reclassify the features into two categories, i.e., implicitly disentangled ones and explicitly disentangled counterparts. Since we aim to design a generic algorithm for DG to alleviate the problems mentioned above, we focus more on the explicitly disentangled features due to their simplicity and interpretability. We find out that the shape features of images are simple and elegant choices based on our analysis. We extract the shape features from two aspects. In the aspect of networks, we propose Multi-Scale Amplitude Mixing (MSAM) to strengthen shape features at different layers of the network by Fourier transform. In the aspect of inputs, we propose a new data augmentation method called Random Shape Warping (RSW) to facilitate the model to concentrate more on the global structures of the objects. RSW randomly distorts the local parts of the images and keeps the global structures unchanged, which can further improve the robustness of the model. Our methods are simple yet efficient and can be conveniently used as plug-and-play modules. They can outperform state-of-the-art (SOTA) methods without bells and whistles.
引用
收藏
页码:6360 / 6373
页数:14
相关论文
共 50 条
  • [21] Comparing Handcrafted Features and Deep Neural Representations for Domain Generalization in Human Activity Recognition
    Bento, Nuno
    Rebelo, Joana
    Barandas, Marilia
    Carreiro, Andre, V
    Campagner, Andrea
    Cabitza, Federico
    Gamboa, Hugo
    SENSORS, 2022, 22 (19)
  • [22] Disentangled Representation Learning with Causality for Unsupervised Domain Adaptation
    Wang, Shanshan
    Chen, Yiyang
    He, Zhenwei
    Yang, Xun
    Wang, Mengzhu
    You, Quanzeng
    Zhang, Xingyi
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 2918 - 2926
  • [23] Research on the improvement of domain generalization by the fusion of invariant features and sharpness-aware minimization
    Yang, Yixuan
    Dong, Mingrong
    Zeng, Kai
    Shen, Tao
    JOURNAL OF SUPERCOMPUTING, 2025, 81 (01)
  • [24] Arbitrary Font Generation by Encoder Learning of Disentangled Features
    Lee, Jeong-Sik
    Baek, Rock-Hyun
    Choi, Hyun-Chul
    SENSORS, 2022, 22 (06)
  • [25] Inter-Class and Inter-Domain Semantic Augmentation for Domain Generalization
    Wang, Mengzhu
    Liu, Yuehua
    Yuan, Jianlong
    Wang, Shanshan
    Wang, Zhibin
    Wang, Wei
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 1338 - 1347
  • [26] Domain Generalization via Frequency-domain-based Feature Disentanglement and Interaction
    Wang, Jingye
    Du, Ruoyi
    Chang, Dongliang
    Liang, Kongming
    Ma, Zhanyu
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 4821 - 4829
  • [27] Style Augmentation and Domain-Aware Parametric Contrastive Learning for Domain Generalization
    Li, Mingkang
    Zhang, Jiali
    Zhang, Wen
    Gong, Lu
    Zhang, Zili
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT IV, KSEM 2023, 2023, 14120 : 211 - 224
  • [28] Domain Generalization via Adversarially Learned Novel Domains
    Zhe, Yu
    Fukuchi, Kazuto
    Akimoto, Youhei
    Sakuma, Jun
    IEEE ACCESS, 2022, 10 : 101855 - 101868
  • [29] Feature-Based Style Randomization for Domain Generalization
    Wang, Yue
    Qi, Lei
    Shi, Yinghuan
    Gao, Yang
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (08) : 5495 - 5509
  • [30] NormAUG: Normalization-Guided Augmentation for Domain Generalization
    Qi, Lei
    Yang, Hongpeng
    Shi, Yinghuan
    Geng, Xin
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 1419 - 1431