Source Free Graph Unsupervised Domain Adaptation

被引:3
作者
Mao, Haitao [1 ,2 ]
Dut, Lun [2 ]
Zheng, Yujia [3 ]
Fu, Qiang [2 ]
Li, Zelin [2 ]
Chen, Xu [2 ]
Han, Shi [2 ]
Zhang, Dongmei [2 ]
机构
[1] Michigan State Univ, E Lansing, MI 48824 USA
[2] Microsoft, Beijing, Peoples R China
[3] Carnegie Mellon Univ, Pittsburgh, PA 15213 USA
来源
PROCEEDINGS OF THE 17TH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING, WSDM 2024 | 2024年
关键词
graph representation learning; unsupervised domain adaptation; transfer learning;
D O I
10.1145/3616855.3635802
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Graph Neural Networks (GNNs) have achieved great success on a variety of tasks with graph-structural data, among which node classification is an essential one. Unsupervised Graph Domain Adaptation (UGDA) shows its practical value of reducing the labeling cost for node classification. It leverages knowledge from a labeled graph (i.e., source domain) to tackle the same task on another unlabeled graph (i.e., target domain). Most existing UGDA methods heavily rely on the labeled graph in the source domain. They utilize labels from the source domain as the supervision signal and are jointly trained on both the source graph and the target graph. However, in some real-world scenarios, the source graph is inaccessible because of privacy issues. Therefore, we propose a novel scenario named Source Free Unsupervised Graph Domain Adaptation (SFUGDA). In this scenario, the only information we can leverage from the source domain is the well-trained source model, without any exposure to the source graph and its labels. As a result, existing UGDA methods are not feasible anymore. To address the non-trivial adaptation challenges in this practical scenario, we propose a model-agnostic algorithm called SOGA for domain adaptation to fully exploit the discriminative ability of the source model while preserving the consistency of structural proximity on the target graph. We prove the effectiveness of the proposed algorithm both theoretically and empirically. The experimental results on four cross-domain tasks show consistent improvements in the Macro-F1 score and Macro-AUC.
引用
收藏
页码:520 / 528
页数:9
相关论文
共 40 条
[1]   Integrating structured biological data by Kernel Maximum Mean Discrepancy [J].
Borgwardt, Karsten M. ;
Gretton, Arthur ;
Rasch, Malte J. ;
Kriegel, Hans-Peter ;
Schoelkopf, Bernhard ;
Smola, Alex J. .
BIOINFORMATICS, 2006, 22 (14) :E49-E57
[2]  
Dai WY, 2007, KDD-2007 PROCEEDINGS OF THE THIRTEENTH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, P210
[3]  
Defferrard M, 2016, ADV NEUR IN, V29
[4]   Learning Structural Node Embeddings via Diffusion Wavelets [J].
Donnat, Claire ;
Zitnik, Marinka ;
Hallac, David ;
Leskovec, Jure .
KDD'18: PROCEEDINGS OF THE 24TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2018, :1320-1329
[5]  
Feng J., 2020, PMLR, P6028
[6]  
Ganin Y, 2015, PR MACH LEARN RES, V37, P1180
[7]   node2vec: Scalable Feature Learning for Networks [J].
Grover, Aditya ;
Leskovec, Jure .
KDD'16: PROCEEDINGS OF THE 22ND ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2016, :855-864
[8]  
Hamilton WL, 2017, ADV NEUR IN, V30
[9]  
Jiang Jing., 2007, P 45 ANN M ASS COMPU, P264
[10]   Active Domain Transfer on Network Embedding [J].
Jin, Lichen ;
Zhang, Yizhou ;
Song, Guojie ;
Jin, Yilun .
WEB CONFERENCE 2020: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW 2020), 2020, :2683-2689