Efficient graph neural architecture search using Monte Carlo Tree search and prediction network

被引:5
作者
Deng, TianJin [1 ]
Wu, Jia [1 ,2 ]
机构
[1] Univ Elect Sci & Technol China, Sch Informat & Software Engn, Chengdu, Peoples R China
[2] 4,Sect 2,North Jianshe Rd, Chengdu 610054, Peoples R China
基金
中国国家自然科学基金;
关键词
Graph neural architecture search; Graph Neural Networks; Monte Carlo Tree search; Prediction network; Reinforcement learning; GO;
D O I
10.1016/j.eswa.2022.118916
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Graph Neural Networks (GNNs) have emerged recently as a powerful way of dealing with non-Euclidean data on graphs, such as social networks and citation networks. Despite their success, obtaining optimal graph neural networks requires immense manual work and domain knowledge. Inspired by the strong searching capability of neural architecture search in CNN, a few attempts automatically search optimal GNNs that rival the best human-invented architectures. However, existing Graph Neural Architecture Search (GNAS) approaches face two challenges: (1) Sampling GNNs across the entire search space results in low search efficiency, particularly in large search spaces. (2) It is pretty costly to evaluate GNNs by training architectures from scratch. To overcome these challenges, this paper proposes an Efficient Graph Neural Architecture Search (EGNAS) method based on Monte Carlo Tree Search (MCTS) and a prediction network. Specifically, EGNAS first uses MCTS to recursively partition the entire search space into good or bad search regions. Then, the reinforcement learning-based search strategy (also called the agent) is applied to sample GNNs in those good search regions, which prevents overly exploring complex architectures and bad-performance regions, thus improving sampling efficiency. To reduce the evaluation cost, we use a prediction network to estimate the performance of GNNs. We alternately use ground-truth accuracy (by training GNNs from scratch) and prediction accuracy (by the prediction network) to update the search strategy to avoid inaccuracies caused by long-term use of the prediction network. Furthermore, to improve the training efficiency and stability, the agent is trained by a variant of Proximal Policy Optimization. Experiments show that EGNAS can search for better GNNs in the promising search region in a shorter search time, with an accuracy of 83.5%, 73.3%, 79.6%, and 94.5% on Cora, Citeseer, Pubmed, and Photo datasets, respectively In particular, compared to the most popular GNAS algorithm, our EGNAS-NP without using the prediction network achieves an accuracy of 83.6% on Cora, 73.5% on Citeseer, 79.9% on Pubmed, and 94.6% on Photo, with a relative improvement of 0.6%, 0.2%, 0.7%, and 0.6%.
引用
收藏
页数:15
相关论文
共 71 条
[1]   Computing Graph Neural Networks: A Survey from Algorithms to Accelerators [J].
Abadal, Sergi ;
Jain, Akshay ;
Guirado, Robert ;
Lopez-Alonso, Jorge ;
Alarcon, Eduard .
ACM COMPUTING SURVEYS, 2022, 54 (09)
[2]  
Abu-El-Haifa S, 2019, PR MACH LEARN RES, V97
[3]   Finite-time analysis of the multiarmed bandit problem [J].
Auer, P ;
Cesa-Bianchi, N ;
Fischer, P .
MACHINE LEARNING, 2002, 47 (2-3) :235-256
[4]  
Baker B., 2016, Designing Neural Network Architectures using Reinforcement Learning
[5]  
Bergstra J., 2011, Advances in Neural Information Processing Systems, V24, P2546
[6]   Graph Neural Networks With Convolutional ARMA Filters [J].
Bianchi, Filippo Maria ;
Grattarola, Daniele ;
Livi, Lorenzo ;
Alippi, Cesare .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (07) :3496-3507
[7]  
Brock A., 2018, INT C LEARNING REPRE
[8]  
Bruna Joan., 2014, 2 INT C LEARN REPR I
[9]  
Cai Han, 2019, INT C LEARNING REPRE
[10]   Rethinking Graph Neural Architecture Search from Message-passing [J].
Cai, Shaofei ;
Li, Liang ;
Deng, Jincan ;
Zhang, Beichen ;
Zha, Zheng-Jun ;
Su, Li ;
Huang, Qingming .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :6653-6662