Self-supervised graph convolutional clustering by preserving latent distribution

被引:18
作者
Kou, Shiwen [1 ]
Xia, Wei [1 ]
Zhang, Xiangdong [1 ]
Gao, Quanxue [1 ]
Gao, Xinbo [2 ,3 ]
机构
[1] Xidian Univ, State Key Lab Integrated Serv Networks, Xian 710071, Shaanxi, Peoples R China
[2] Xidian Univ, Sch Elect Engn, Xian 710071, Shaanxi, Peoples R China
[3] Chongqing Univ Posts & Telecommun, Chongqing Key Lab Image Cognit, Chongqing 400065, Peoples R China
基金
中国国家自然科学基金;
关键词
Graph convolutional network; Node clustering; Self-supervised learning; Distribution consistency; NEURAL-NETWORK;
D O I
10.1016/j.neucom.2021.01.082
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Node clustering has attracted widespread concern due to the fact that graph structure helps to seek the geometric structure of data. Existing graph convolutional networks (GCNs) based node clustering methods usually contain two steps. First, learning latent nodes representations. Second, inferring the unknown labels. Despite the promising preliminary results, it cannot take full advantage of the information embedded in current pseudo clustering labels, resulting in suboptimal performance. In fact, it is the truth we do not know the true label of the node in clustering, but the pseudo clustering label of the node generated by each iteration is known. Under this condition, although the pseudo labels of some nodes are incorrect, the partial correct label information is useful and significant. Moreover, GCNs based node clustering aims to learn satisfactory low dimensional node representations which are usually used for clustering. However, existing GCNs based clustering methods fail to take into account distribution consistency between the raw data space and the latent space of nodes representations, resulting in insufficient representations. To address the above problems, we propose a novel GCN-based node clustering method. By introducing a self-supervision module, it can employ the good property of pseudo clustering labels to self-supervise the learning of node representations. Moreover, a latent distribution preserving term, which is measured by KL divergence, is employed to help the latent representations of the same sample to further have a consistent distribution in dimension space as well as in the original dimension space. We formulate the above two concerns into a unified optimization framework. Experimental results on several public datasets indicate our method is effective compared with the state-of-the-art node clustering methods. (c) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页码:218 / 226
页数:9
相关论文
共 41 条
[1]  
Belkin M, 2002, ADV NEUR IN, V14, P585
[2]  
Bruna J., ARXIV PREPRINT ARXIV
[3]  
Cao SS, 2016, AAAI CONF ARTIF INTE, P1145
[4]  
Davulcu H., ARXIV PREPRINT ARXIV
[5]  
Defferrard M, 2016, ADV NEUR IN, V29
[6]  
Gao QX, 2020, AAAI CONF ARTIF INTE, V34, P3930
[7]   Enhanced Tensor RPCA and its Application [J].
Gao, Quanxue ;
Zhang, Pu ;
Xia, Wei ;
Xie, Deyan ;
Gao, Xinbo ;
Tao, Dacheng .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2021, 43 (06) :2133-2140
[8]   Automatic Chemical Design Using a Data-Driven Continuous Representation of Molecules [J].
Gomez-Bombarelli, Rafael ;
Wei, Jennifer N. ;
Duvenaud, David ;
Hernandez-Lobato, Jose Miguel ;
Sanchez-Lengeling, Benjamin ;
Sheberla, Dennis ;
Aguilera-Iparraguirre, Jorge ;
Hirzel, Timothy D. ;
Adams, Ryan P. ;
Aspuru-Guzik, Alan .
ACS CENTRAL SCIENCE, 2018, 4 (02) :268-276
[9]  
Gori M, 2005, IEEE IJCNN, P729
[10]  
He X., 2020, PROC 28 ACM INT C MU, P3541