Hierarchical Graph Contrastive Learning via Debiasing Noise Samples with Adaptive Repelling Ratio

被引:2
作者
Liu, Peishuo [1 ]
Zhou, Cangqi [1 ]
Zhang, Jing [2 ]
Li, Qianmu [1 ]
Hu, Dianming [3 ]
机构
[1] Nanjing Univ Sci & Technol, Nanjing, Peoples R China
[2] Southeast Univ, Nanjing, Peoples R China
[3] SenseDeal Intelligent Technol Co Ltd, Beijing, Peoples R China
来源
23RD IEEE INTERNATIONAL CONFERENCE ON DATA MINING, ICDM 2023 | 2023年
基金
中国国家自然科学基金;
关键词
Unsupervised Graph Learning; Graph Contrastive Learning; Graph Neural Network;
D O I
10.1109/ICDM58522.2023.00051
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In recent years, the field of unsupervised graph representation learning has witnessed the emergence of graph contrastive learning (GCL) as a highly successful approach. GCL excels in learning graph representations by effectively bringing positive sample pairs into proximity while simultaneously pushing negative sample pairs apart in the representation space, without any manual labels. Graph-structured data to be learned inherently exhibits a critical hierarchical structure, which is crucial for organizing and managing graphs. Leveraging this attribute enhances the accuracy of graph representation outcomes. However, current GCL methods tend to overlook the hierarchical structure, which can result in sampling bias during node selection. Nodes of the same semantics can potentially he sampled as negative pairs. To overcome these limitations, we present a novel framework, Hierarchical Graph Contrastive Learning via Debiasing Noise Samples with Adaptive Repelling Ratio (HGClear). Our framework enables the simultaneous learning of node representations and the graph hierarchy in an end -to-end manner. During the process of method design, we discovered that the accuracy of node category prediction significantly affects representation results. To remove the bias caused by noise samples, we introduced a module to handle boundary nodes (i.e., noise samples) that are vulnerable to mislabeling. Specifically, we introduce a hierarchy detection module that captures both coarse-grained views and category attributes of nodes. Leveraging these results, we can identify boundary nodes and establish varying repelling ratios based on category labels, replacing the conventional temperature coefficient in the contrastive loss. Simultaneously incorporating an intra-view node contrast module not only eliminates the bias resulting from noise samples but also enhances the uniqueness of node representations. Numerous experiments on node classification datasets show that HGClear produces encouraging results and outperforms some state-of-the-art methods.
引用
收藏
页码:418 / 427
页数:10
相关论文
共 40 条
[1]  
Bhagat S, 2011, Arxiv, DOI arXiv:1101.3291
[2]  
Bianchi FM, 2020, PR MACH LEARN RES, V119
[3]  
Cho HYS, 2021, Arxiv, DOI arXiv:2105.08793
[4]  
Gilmer Justin., 2017, INT C MACHINE LEARNI
[5]   node2vec: Scalable Feature Learning for Networks [J].
Grover, Aditya ;
Leskovec, Jure .
KDD'16: PROCEEDINGS OF THE 22ND ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2016, :855-864
[6]   DISCERN: diversity-based selection of centroids for k-estimation and rapid non-stochastic clustering [J].
Hassani, Ali ;
Iranmanesh, Amir ;
Eftekhari, Mahdi ;
Salemi, Abbas .
INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2021, 12 (03) :635-649
[7]   X-GOAL: Multiplex Heterogeneous Graph Prototypical Contrastive Learning [J].
Jing, Baoyu ;
Feng, Shengyu ;
Xiang, Yuejia ;
Chen, Xi ;
Chen, Yu ;
Tong, Hanghang .
PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, :894-904
[8]  
Khosla P, 2020, ADV NEUR IN, V33
[9]  
Kingsbury D, 2015, P1, DOI [10.48550/arXiv.1312.6114, DOI 10.48550/ARXIV.1412.6980, 10.48550/arXiv.1412.6980, DOI 10.1021/bk-2015-1214.ch001]
[10]   Graph Communal Contrastive Learning [J].
Li, Bolian ;
Jing, Baoyu ;
Tong, Hanghang .
PROCEEDINGS OF THE ACM WEB CONFERENCE 2022 (WWW'22), 2022, :1203-1213