Constrained Maximum Cross-Domain Likelihood for Domain Generalization

被引:6
作者
Lin, Jianxin [1 ,2 ]
Tang, Yongqiang [1 ,2 ]
Wang, Junping [1 ,2 ]
Zhang, Wensheng [1 ,2 ]
机构
[1] Chinese Acad Sci, Inst Automat, State Key Lab Multimodal Artificial Intelligence, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Sch Artificial Intelligence, Beijing 100049, Peoples R China
基金
中国国家自然科学基金;
关键词
Distribution shift; domain adaptation; domain generalization; domain-invariant representation; joint distribution alignment;
D O I
10.1109/TNNLS.2023.3292242
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As a recent noticeable topic, domain generalization aims to learn a generalizable model on multiple source domains, which is expected to perform well on unseen test domains. Great efforts have been made to learn domain-invariant features by aligning distributions across domains. However, existing works are often designed based on some relaxed conditions which are generally hard to satisfy and fail to realize the desired joint distribution alignment. In this article, we propose a novel domain generalization method, which originates from an intuitive idea that a domain-invariant classifier can be learned by minimizing the Kullback-Leibler (KL)-divergence between posterior distributions from different domains. To enhance the generalizability of the learned classifier, we formalize the optimization objective as an expectation computed on the ground-truth marginal distribution. Nevertheless, it also presents two obvious deficiencies, one of which is the side-effect of entropy increase in KL-divergence and the other is the unavailability of ground-truth marginal distributions. For the former, we introduce a term named maximum in-domain likelihood to maintain the discrimination of the learned domain-invariant representation space. For the latter, we approximate the ground-truth marginal distribution with source domains under a reasonable convex hull assumption. Finally, a constrained maximum cross-domain likelihood (CMCL) optimization problem is deduced, by solving which the joint distributions are naturally aligned. An alternating optimization strategy is carefully designed to approximately solve this optimization problem. Extensive experiments on four standard benchmark datasets, i.e., Digits-DG, PACS, Office-Home, and miniDomainNet, highlight the superior performance of our method.
引用
收藏
页码:2013 / 2027
页数:15
相关论文
共 86 条
[1]  
Ahuja K, 2021, ADV NEUR IN, V34
[2]  
Albuquerque I, 2021, Arxiv, DOI arXiv:1911.00804
[3]  
Arjovsky M, 2020, Arxiv, DOI [arXiv:1907.02893, 10.48550/arXiv.1907.02893]
[4]   NAS-OoD: Neural Architecture Search for Out-of-Distribution Generalization [J].
Bai, Haoyue ;
Zhou, Fengwei ;
Hong, Lanqing ;
Ye, Nanyang ;
Chan, S. -H. Gary ;
Li, Zhenguo .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :8300-8309
[5]  
Balaji Y, 2018, ADV NEUR IN, V31
[6]  
Blanchard G, 2021, J MACH LEARN RES, V22
[7]   Domain Generalization by Solving Jigsaw Puzzles [J].
Carlucci, Fabio M. ;
D'Innocente, Antonio ;
Bucci, Silvia ;
Caputo, Barbara ;
Tommasi, Tatiana .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :2224-2233
[8]   Learning to Balance Specificity and Invariance for In and Out of Domain Generalization [J].
Chattopadhyay, Prithvijit ;
Balaji, Yogesh ;
Hoffman, Judy .
COMPUTER VISION - ECCV 2020, PT IX, 2020, 12354 :301-318
[9]  
Dou Q, 2019, ADV NEUR IN, V32
[10]  
Du Y., 2021, P INT C LEARN REPR, P1