Scalable Spatiotemporal Graph Neural Networks

被引:0
作者
Cini, Andrea [1 ]
Marisca, Ivan [1 ]
Bianchi, Filippo Maria [2 ,3 ]
Alippi, Cesare [1 ,4 ]
机构
[1] Univ Svizzera Italiana, Swiss AI Lab IDSIA, Lugano, Switzerland
[2] UiT Arctic Univ Norway, Tromso, Norway
[3] NORCE Norwegian Res Ctr, Bergen, Norway
[4] Politecn Milan, Milan, Italy
来源
THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 6 | 2023年
基金
芬兰科学院; 瑞士国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neural forecasting of spatiotemporal time series drives both research and industrial innovation in several relevant application domains. Graph neural networks (GNNs) are often the core component of the forecasting architecture. However, in most spatiotemporal GNNs, the computational complexity scales up to a quadratic factor with the length of the sequence times the number of links in the graph, hence hindering the application of these models to large graphs and long temporal sequences. While methods to improve scalability have been proposed in the context of static graphs, few research efforts have been devoted to the spatiotemporal case. To fill this gap, we propose a scalable architecture that exploits an efficient encoding of both temporal and spatial dynamics. In particular, we use a randomized recurrent neural network to embed the history of the input time series into high-dimensional state representations encompassing multi-scale temporal dynamics. Such representations are then propagated along the spatial dimension using different powers of the graph adjacency matrix to generate node embeddings characterized by a rich pool of spatiotemporal features. The resulting node embeddings can be efficiently pre-computed in an unsupervised manner, before being fed to a feed-forward decoder that learns to map the multi-scale spatiotemporal representations to predictions. The training procedure can then be parallelized node-wise by sampling the node embeddings without breaking any dependency, thus enabling scalability to large networks. Empirical results on relevant datasets show that our approach achieves results competitive with the state of the art, while dramatically reducing the computational burden.
引用
收藏
页码:7218 / 7226
页数:9
相关论文
共 33 条
[1]   A gentle introduction to deep learning for graphs [J].
Bacciu, Davide ;
Errica, Federico ;
Micheli, Alessio ;
Podda, Marco .
NEURAL NETWORKS, 2020, 129 :203-221
[2]   Reservoir Computing Approaches for Representation and Classification of Multivariate Time Series [J].
Bianchi, Filippo Maria ;
Scardapane, Simone ;
Lokse, Sigurd ;
Jenssen, Robert .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (05) :2169-2179
[3]   Cluster-GCN: An Efficient Algorithm for Training Deep and Large Graph Convolutional Networks [J].
Chiang, Wei-Lin ;
Liu, Xuanqing ;
Si, Si ;
Li, Yang ;
Bengio, Samy ;
Hsieh, Cho-Jui .
KDD'19: PROCEEDINGS OF THE 25TH ACM SIGKDD INTERNATIONAL CONFERENCCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2019, :257-266
[4]  
Cini A, 2022, INT C LEARN REPR
[5]  
Commission for Energy Regulation, 2016, CER Smart Me-tering Project-Electricity Customer Behaviour Trial, 2009-2010 dataset
[6]  
Frasca F., 2020, ICML 2020 WORKSH GRA
[7]   Design of deep echo state networks [J].
Gallicchio, Claudio ;
Micheli, Alessio ;
Pedrelli, Luca .
NEURAL NETWORKS, 2018, 108 :33-47
[8]   Deep reservoir computing: A critical experimental analysis [J].
Gallicchio, Claudio ;
Micheli, Alessio ;
Pedrelli, Luca .
NEUROCOMPUTING, 2017, 268 :87-99
[9]   Spatio-Temporal Multi-graph Networks for Demand Forecasting in Online Marketplaces [J].
Gandhi, Ankit ;
Aakanksha ;
Kaveri, Sivaramakrishnan ;
Chaoji, Vineet .
MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2021: APPLIED DATA SCIENCE TRACK, PT IV, 2021, 12978 :187-203
[10]  
Gao Jianfei, 2022, PR MACH LEARN RES