Long Short-Term Graph Memory Against Class-imbalanced Over-smoothing

被引:1
作者
Yang, Liang [1 ]
Wang, Jiayi [1 ]
Zhang, Tingting [2 ]
He, Dongxiao [3 ]
Wang, Chuan [4 ]
Guo, Yuanfang [5 ]
Cao, Xiaochun [6 ]
Niu, Bingxin [1 ]
Wang, Zhen [7 ]
机构
[1] Hebei Univ Technol, Sch Artificial Intelligence, Tianjin, Peoples R China
[2] Army Engn Univ, Coll Command & Control Engn, Nanjing, Peoples R China
[3] Tianjin Univ, Coll Intelligence & Comp, Tianjin, Peoples R China
[4] Chinese Acad Sci, Inst Informat Engn, State Key Lab Informat Secur, Beijing, Peoples R China
[5] Beihang Univ, Sch Comp Sci & Engn, Beijing, Peoples R China
[6] Sun Yat Sen Univ, Sch Cyber Sci & Technol, Shenzhen Campus, Shenzhen, Peoples R China
[7] Northwestern Polytech Univ, Opt & Elect iOPEN, Sch Cybersecur, Xian, Peoples R China
来源
PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023 | 2023年
基金
中国国家自然科学基金;
关键词
Graph Neural Networks; Long Short-Term Memory Networks; Deep Models;
D O I
10.1145/3581783.3612566
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Most Graph Neural Networks (GNNs) follow the message-passing scheme. Residual connection is an effective strategy to tackle GNNs' over-smoothing issue and performance reduction issue on non-homophilic networks. Unfortunately, the coarse-grained residual connection still suffers from class-imbalanced over-smoothing issue, due to the fixed and linear combination of topology and attribute in node representation learning. To make the combination flexible to capture complicated relationship, this paper reveals that the residual connection needs to be node-dependent, layer-dependent, and related to both topology and attribute. To alleviate the difficulty in specifying complicated relationship, this paper presents a novel perspective on GNNs, i.e., the representations of one node in different layers can be seen as a sequence of states. From this perspective, existing residual connections are not flexible enough for sequence modeling. Therefore, a novel node-dependent residual connection, i.e., Long Short-Term Graph Memory Network (LSTGM) is proposed to employ Long Short-Term Memory (LSTM), to model the sequence of node representation. To make the graph topology fully employed, LSTGM innovatively enhances the updated memory and three gates with graph topology. A speedup version is also proposed for effective training. Experimental evaluations on real-world datasets demonstrate their effectiveness in preventing over-smoothing issue and handling networks with heterophily.
引用
收藏
页码:2955 / 2963
页数:9
相关论文
共 43 条
  • [1] Balcilar Muhammet, 2021, ICLR
  • [2] Bo DY, 2021, AAAI CONF ARTIF INTE, V35, P3950
  • [3] Chen Ming, 2020, P MACHINE LEARNING R, V119
  • [4] TSSRGCN: Temporal Spectral Spatial Retrieval Graph Convolutional Network for Traffic Flow Forecasting
    Chen, Xu
    Zhang, Yuanxing
    Du, Lun
    Fang, Zheng
    Ren, Yi
    Bian, Kaigui
    Xie, Kunqing
    [J]. 20TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2020), 2020, : 954 - 959
  • [5] Chien E., 2021, INT C LEARNING REPRE
  • [6] Gilmer Justin, 2017, P MACHINE LEARNING R, V70
  • [7] Graves A, 2012, STUD COMPUT INTELL, V385, P1, DOI [10.1162/neco.1997.9.1.1, 10.1007/978-3-642-24797-2]
  • [8] Guo SN, 2019, AAAI CONF ARTIF INTE, P922
  • [9] Hamilton WL, 2017, ADV NEUR IN, V30
  • [10] Motor Skills Learning for CNT Pick-up of Micro-Nano Robotic Manipulator in SEM
    Han, Chirui
    Zhang, Lue
    Yang, Zhan
    [J]. 2021 IEEE 21ST INTERNATIONAL CONFERENCE ON NANOTECHNOLOGY (IEEE NANO 2021), 2021, : 217 - 220