Partially-Labeled Domain Generalization via Multi-Dimensional Domain Adaptation

被引:0
作者
Ye, Feiyang [1 ,3 ]
Bao, Jianghan [2 ]
Zhang, Yu [1 ,4 ]
机构
[1] Southern Univ Sci & Technol, Dept Comp Sci & Engn, Shenzhen, Peoples R China
[2] Southeast Univ, Sch Informat Sci & Engn, Nanjing, Peoples R China
[3] Univ Technol Sydney, Australian Artificial Intelligence Inst, Sydney, NSW, Australia
[4] Peng Cheng Lab, Shenzhen, Peoples R China
来源
2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN | 2023年
关键词
domain generalization; domain adaptation; knowledge distillation;
D O I
10.1109/IJCNN54540.2023.10191532
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Domain generalization deals with a challenging setting where several labeled source domains are given, and the goal is to train machine learning models that can generalize to an unseen test domain. However, in practice, labeled samples are often difficult and expensive to obtain. Thus the source domains would not always be labeled. When only some source domains are labeled and others are unlabeled, we formally introduce this domain generalization problem as Partially-Labeled Domain Generalization (PLDG). In this paper, we study the most challenging setting in PLDG problems, where only one source domain is labeled and a few unlabeled source domains are available. To enable generalization, we assume that all source domains follow certain domain index information that can reflect their domain relationships. With this domain index information, we propose a Multi-Dimensional Domain Adaptation (MDDA) method to address this PLDG problem. Specifically, the MDDA method first trains multiple domain adaptation models to adapt from the labeled source domain to all the unlabeled source domains via adversarial learning. Then those domain adaptation models and the source-only model trained on the labeled source domain only are distilled into the target model used for the unseen target domain. Theoretically, we provide a generalization bound of the MDDA method. The experiments on four real-world datasets demonstrate the effectiveness of the proposed MDDA method.
引用
收藏
页数:8
相关论文
共 45 条
[1]  
Albuquerque I., 2019, arXiv
[2]  
[Anonymous], PROC CVPR IEEE
[3]  
Balaji Y, 2018, ADV NEUR IN, V31
[4]   Patient Subtyping via Time-Aware LSTM Networks [J].
Baytas, Inci M. ;
Xiao, Cao ;
Zhang, Xi ;
Wang, Fei ;
Jain, Anil K. ;
Zhou, Jiayu .
KDD'17: PROCEEDINGS OF THE 23RD ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2017, :65-74
[5]   A theory of learning from different domains [J].
Ben-David, Shai ;
Blitzer, John ;
Crammer, Koby ;
Kulesza, Alex ;
Pereira, Fernando ;
Vaughan, Jennifer Wortman .
MACHINE LEARNING, 2010, 79 (1-2) :151-175
[6]  
Ben -David Shai, 2006, Advances in Neural Information Processing Systems, V19, DOI DOI 10.7551/MITPRESS/7503.003.0022
[7]  
Bertin-Mahieux T., 2011, P 12 INT C MUSIC INF
[8]  
Bucila C., 2006, SIGKDD, P535, DOI [DOI 10.1145/1150402.1150464, 10.1145/1150402.1150464]
[9]  
David S. B., 2010, P AISTATS, P129
[10]  
Dean J., 2015, NIPS DEEP LEARNING R