Rethinking Domain Generalization: Discriminability and Generalizability

被引:2
作者
Long, Shaocong [1 ]
Zhou, Qianyu [1 ]
Ying, Chenhao [1 ,2 ]
Ma, Lizhuang [1 ]
Luo, Yuan [1 ,2 ]
机构
[1] Shanghai Jiao Tong Univ, Dept Comp Sci & Engn, Shanghai 200240, Peoples R China
[2] Shanghai Jiao Tong Univ, Blockchain Adv Res Ctr, Wuxi 214101, Jiangsu, Peoples R China
关键词
Domain generalization; representation learning; discriminability; generalizability; transfer learning;
D O I
10.1109/TCSVT.2024.3422887
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Domain generalization (DG) endeavours to develop robust models that possess strong generalizability while preserving excellent discriminability. Nonetheless, pivotal DG techniques tend to improve the feature generalizability by learning domain-invariant representations, inadvertently overlooking the feature discriminability. On the one hand, the simultaneous attainment of generalizability and discriminability of features presents a complex challenge, often entailing inherent contradictions. This challenge becomes particularly pronounced when domain-invariant features manifest reduced discriminability owing to the inclusion of unstable factors, i.e., spurious correlations. On the other hand, prevailing domain-invariant methods can be categorized as category-level alignment, susceptible to discarding indispensable features possessing substantial generalizability and narrowing intra-class variations. To surmount these obstacles, we rethink DG from a new perspective that concurrently imbues features with formidable discriminability and robust generalizability, and present a novel framework, namely, Discriminative Microscopic Distribution Alignment (DMDA). DMDA incorporates two core components: Selective Channel Pruning (SCP) and Micro-level Distribution Alignment (MDA). Concretely, SCP attempts to curtail redundancy within neural networks, prioritizing stable attributes conducive to accurate classification. This approach alleviates the adverse effect of spurious domain-invariance and amplifies the feature discriminability. Besides, MDA accentuates micro-level alignment within each class, going beyond mere category-level alignment. This strategy accommodates sufficient generalizable features and facilitates within-class variations. Extensive experiments on four benchmark datasets corroborate that DMDA achieves comparable results to state-of-the-art methods in DG, underscoring the efficacy of our method. The source code will be available at https://github.com/longshaocong/DMDA.
引用
收藏
页码:11783 / 11797
页数:15
相关论文
共 93 条
  • [21] Hendrycks D., Dietterich T., Benchmarking neural network robustness to common corruptions and perturbations, Proc. Int. Conf. Learn. Represent., (2018)
  • [22] Taori R., Dave A., Shankar V., Measuring robustness to natural distribution shifts in image classification, Proc. Adv. Neural Inf. Process. Syst. (NeurIPS), pp. 18583-18599, (2020)
  • [23] Beery S., Van Horn G., Perona P., Recognition in terra incognita, Proc. Eur. Conf. Comput. Vis. (ECCV), pp. 456-473, (2018)
  • [24] Scholkopf B., Et al., Towards causal representation learning, (2021)
  • [25] Gu Q., Et al., PIT: Position-invariant transform for cross-FoV domain adaptation, Proc. IEEE/CVF Int. Conf. Comput. Vis. (ICCV), pp. 8761-8770, (2021)
  • [26] Fu H., Gong M., Wang C., Batmanghelich K., Zhang K., Tao D., Geometry-consistent generative adversarial networks for onesided unsupervised domain mapping, Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 2427-2436, (2019)
  • [27] Pearl J., A probabilistic calculus of actions, (2013)
  • [28] Wei Y., Yang L., Han Y., Hu Q., Multi-source collaborative contrastive learning for decentralized domain adaptation, IEEE Trans. Circuits Syst. Video Technol., 33, 5, pp. 2202-2216, (2023)
  • [29] Meng R., Et al., Attention diversification for domain generalization, Proc. Eur. Conf. Comput. Vis. (ECCV), pp. 322-340, (2022)
  • [30] Ganin Y., Et al., Domain-adversarial training of neural networks, J. Mach. Learn. Res., 17, 1, pp. 1-35, (2016)