A two-stage co-adversarial perturbation to mitigate out-of-distribution generalization of large-scale graph

被引:0
作者
Wang, Yili [1 ]
Xue, Haotian [1 ]
Wang, Xin [1 ]
机构
[1] Jilin Univ, Sch Artificial Intelligence, Changchun 130012, Peoples R China
基金
中国国家自然科学基金;
关键词
Graph neural network; Adversarial training; Graph out-of-distribution; NETWORK;
D O I
10.1016/j.eswa.2024.124472
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In the realm of graph out-of-distribution (OOD), despite recent strides in advancing graph neural networks (GNNs) for the modeling of graph data, training GNNs on large-scale datasets presents a formidable hurdle due to the pervasive challenge of overfitting. To address these issues, researchers have explored adversarial training, a technique that enriches training data with worst-case adversarial examples. However, while prior work on adversarial training primarily focuses on safeguarding GNNs against malicious attacks, its potential to enhance the OOD generalization abilities of GNNs in the context of graph analytics remains less explored. In our research, we delve into the inner workings of GNNs by examining the landscapes of weight and feature losses, which respectively illustrate how the loss function changes concerning model weights and node features. Our investigation reveals a noteworthy phenomenon: GNNs are inclined to become trapped in sharp local minima within these loss landscapes, resulting in suboptimal OOD generalization performance. To address this challenge, we introduce the concept of co-adversarial perturbation optimization, which considers both model weights and node features, and we design an alternating adversarial perturbation algorithm for graph out-of-distribution generalization. This algorithm operates iteratively, smoothing the weight and feature loss landscapes alternately. Moreover, our training process unfolds in two distinct stages. The first stage centers on standard cross-entropy minimization, ensuring rapid convergence of GNN models. In the second stage, we employ our alternating adversarial training strategy to prevent the models from becoming ensnared in locally sharp minima. Our extensive experiments provide compelling evidence that our CAP approach can generally enhance the OOD generalization performance of GNNs across a diverse range of large-scale graphs.
引用
收藏
页数:11
相关论文
共 56 条
  • [1] Bevilacqua Beatrice, 2021, PMLR, P837
  • [2] Bruna Joan, 2014, P ICLR
  • [3] Chen TL, 2022, Arxiv, DOI arXiv:2108.10521
  • [4] Chen TL, 2021, Arxiv, DOI arXiv:2103.12171
  • [5] Chen X., 2021, arXiv
  • [6] Chen ZY, 2024, AAAI CONF ARTIF INTE, P8320
  • [7] Cluster-GCN: An Efficient Algorithm for Training Deep and Large Graph Convolutional Networks
    Chiang, Wei-Lin
    Liu, Xuanqing
    Si, Si
    Li, Yang
    Bengio, Samy
    Hsieh, Cho-Jui
    [J]. KDD'19: PROCEEDINGS OF THE 25TH ACM SIGKDD INTERNATIONAL CONFERENCCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2019, : 257 - 266
  • [8] Dai HJ, 2018, PR MACH LEARN RES, V80
  • [9] Adversarial training regularization for negative sampling based network embedding
    Dai, Quanyu
    Shen, Xiao
    Zheng, Zimu
    Zhang, Liang
    Li, Qiang
    Wang, Dan
    [J]. INFORMATION SCIENCES, 2021, 579 : 199 - 217
  • [10] Deng ZJ, 2019, Arxiv, DOI arXiv:1902.09192