End-to-end variational graph clustering with local structural preservation

被引:5
作者
Guo, Lin [1 ]
Dai, Qun [1 ]
机构
[1] Nanjing Univ Aeronaut & Astronaut, Coll Comp Sci & Technol, Nanjing 211106, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Graph convolutional neural network; Variational graph embedding; Graph clustering; Variational graph auto-encoder;
D O I
10.1007/s00521-021-06639-7
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Graph clustering, a basic problem in machine learning and artificial intelligence, facilitates a variety of real-world applications. How to perform a task of graph clustering, with a relatively high-quality optimization decision and an effective yet efficient way to use graph information, to obtain a more excellent assignment for discrete points is not an ordinary challenge that troubles scholars. Often, many preeminent works on graph clustering neglect an essential element that the defined clustering loss may destroy the feature space. This is also a vital factor that leads to unrepresentative nonsense features that generate poor partitioning decisions. Here, we propose an end-to-end variational graph clustering (EVGC) algorithm focusing on preserving the original information of the graph. Specifically, the KL loss with an auxiliary distribution serves as a specific guide to manipulate the embedding space, and consequently disperse data points. A graph auto-encoder plays a propulsive role in maximumly retaining the local structure of the generative distribution of the graph. And each node is represented as a Gaussian distribution in dealing with separating the true embedding position and uncertainty from the graph. Experimental results reveal the importance of preserving local structure, and our EVGC system outperforms state-of-the-art approaches.
引用
收藏
页码:3767 / 3782
页数:16
相关论文
共 34 条
[11]   Finding community structure in networks using the eigenvectors of matrices [J].
Newman, M. E. J. .
PHYSICAL REVIEW E, 2006, 74 (03)
[12]  
Nigam K., 2000, Proceedings of the Ninth International Conference on Information and Knowledge Management. CIKM 2000, P86, DOI 10.1145/354756.354805
[13]  
Pan SR, 2018, PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P2609
[14]   DeepWalk: Online Learning of Social Representations [J].
Perozzi, Bryan ;
Al-Rfou, Rami ;
Skiena, Steven .
PROCEEDINGS OF THE 20TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING (KDD'14), 2014, :701-710
[15]  
Rezende DJ, 2014, PR MACH LEARN RES, V32, P1278
[16]  
Salha G., 2020, ARXIV PREPRINT ARXIV
[17]   Performance and Convergence Analysis of Modified C-Means Using Jeffreys-Divergence for Clustering [J].
Seal, Ayan ;
Karlekar, Aditya ;
Krejcar, Ondrej ;
Herrera-Viedma, Enrique .
INTERNATIONAL JOURNAL OF INTERACTIVE MULTIMEDIA AND ARTIFICIAL INTELLIGENCE, 2021, 7 (02) :141-149
[18]   Optimized gene selection and classification of cancer from microarray gene expression data using deep learning [J].
Shah, Shamveel Hussain ;
Iqbal, Muhammad Javed ;
Ahmad, Iftikhar ;
Khan, Suleman ;
Rodrigues, Joel J. P. C. .
NEURAL COMPUTING & APPLICATIONS, 2020,
[19]   An Enhanced Spectral Clustering Algorithm with S-Distance [J].
Sharma, Krishna Kumar ;
Seal, Ayan ;
Herrera-Viedma, Enrique ;
Krejcar, Ondrej .
SYMMETRY-BASEL, 2021, 13 (04)
[20]   Spectral embedded generalized mean based κ-nearest neighbors clustering with S-distance [J].
Sharma, Krishna Kumar ;
Seal, Ayan .
EXPERT SYSTEMS WITH APPLICATIONS, 2021, 169