OnsitNet: A memory-capable online time series forecasting model incorporating a self-attention mechanism

被引:1
作者
Liu, Hui [1 ,2 ,3 ]
Wang, Zhengkai [1 ]
Dong, Xiyao [1 ]
Du, Junzhao [1 ,2 ,3 ]
机构
[1] Xidian Univ, Sch Comp Sci & Technol, Xian 710126, Peoples R China
[2] Minist Educ, Engn Res Ctr Blockchain Technol Applicat & Evaluat, Xian 710126, Peoples R China
[3] Key Lab Smart Human Comp Interact & Wearable Techn, Xian 710126, Peoples R China
基金
中国国家自然科学基金;
关键词
Time series forecasting; Online learning; Offline learning; Self-attention mechanism; ITransformer;
D O I
10.1016/j.eswa.2024.125231
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Traditional time series (TS) forecasting models are based on fixed, static datasets and lack scalability when faced with the continuous influx of data in real-world scenarios. Real-time online learning of data streams is crucial for improving forecasting efficiency. However, few studies focus on online TS forecasting, and existing approaches have several limitations. Most current online TS forecasting models merely train on data streams and are ineffective in handling concept drift scenarios. Furthermore, they often fail to adequately consider dependencies between variables and do not leverage the robust modeling capabilities of offline models. Therefore, we propose an innovative online learning method called OnsitNet. It consists of multiple learning modules that progressively expand the receptive field of convolutional kernels within the learning modules using an exponentially growing dilation factor, aiding in the capture of multi-scale data features. Within the learning modules, we propose an online learning strategy focusing on memorizing concept drift scenarios, with a fast learner, memorizer, and Pearson trigger. The Pearson trigger activates dynamic interaction between the fast learner and memorizer by detecting new data patterns, facilitating online rapid learning of data streams. To capture the dependencies between variables, we propose a new model, SITransformer, which is a streamlined version of the offline model ITransformer. Unlike the traditional Transformer, it reverses the roles of the feed-forward network and the attention mechanism. This inverted architecture is more effective at learning the correlations between variables. Experimental results on five real-world datasets show OnsitNet achieves lower online prediction errors, enabling timely and effective forecasting of future trends in TS data.
引用
收藏
页数:13
相关论文
共 50 条
  • [41] Self-Attention Causal Dilated Convolutional Neural Network for Multivariate Time Series Classification and Its Application
    Yang, Wenbiao
    Xia, Kewen
    Wang, Zhaocheng
    Fan, Shurui
    Li, Ling
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 122
  • [42] ResDeepSurv: A Survival Model for Deep Neural Networks Based on Residual Blocks and Self-attention Mechanism
    Wang, Yuchen
    Kong, Xianchun
    Bi, Xiao
    Cui, Lizhen
    Yu, Hong
    Wu, Hao
    INTERDISCIPLINARY SCIENCES-COMPUTATIONAL LIFE SCIENCES, 2024, 16 (02) : 405 - 417
  • [43] Time Series Forecasting and Classification Models Based on Recurrent with Attention Mechanism and Generative Adversarial Networks
    Zhou, Kun
    Wang, Wenyong
    Hu, Teng
    Deng, Kai
    SENSORS, 2020, 20 (24) : 1 - 20
  • [44] Improving Ship Fuel Consumption and Carbon Intensity Prediction Accuracy Based on a Long Short-Term Memory Model with Self-Attention Mechanism
    Wang, Zhihuan
    Lu, Tianye
    Han, Yi
    Zhang, Chunchang
    Zeng, Xiangming
    Li, Wei
    APPLIED SCIENCES-BASEL, 2024, 14 (18):
  • [45] Code Defect Detection Model with Multi-layer Bi-directional Long Short Term Memory based on Self-Attention Mechanism
    Hou, Cong
    Sun, Yue
    Li, Lin
    Chen, Wei
    Xu, Xiaotian
    PROCEEDINGS OF 2023 7TH INTERNATIONAL CONFERENCE ON ELECTRONIC INFORMATION TECHNOLOGY AND COMPUTER ENGINEERING, EITCE 2023, 2023, : 1656 - 1660
  • [46] Effective Construction of a Reflection Angle Prediction Model for Reflectarrays Using the Hadamard Product Self-Attention Mechanism
    Lin, Yi-Le
    Chen, You-Cheng
    Liu, Alan
    2022 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2022, : 1298 - 1303
  • [47] A new hybrid model for multi-step WTI futures price forecasting based on self-attention mechanism and spatial-temporal graph neural network
    Zhao, Geya
    Xue, Minggao
    Cheng, Li
    RESOURCES POLICY, 2023, 85
  • [48] MPM: Multi Patterns Memory Model for Short-Term Time Series Forecasting
    Wang, Dezheng
    Liu, Rongjie
    Chen, Congyan
    Li, Shihua
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2025, 37 (01) : 438 - 448
  • [49] A Model Selection Approach for Time Series Forecasting: Incorporating Google Trends Data in Australian Macro Indicators
    Karim, Ali Abdul
    Pardede, Eric
    Mann, Scott
    ENTROPY, 2023, 25 (08)
  • [50] Candidate point selection using a self-attention mechanism for generating a smooth volatility surface under the SABR model
    Kim, Hyeonuk
    Park, Kyunghyun
    Jeon, Junkee
    Song, Changhoon
    Bae, Jungwoo
    Kim, Yongsik
    Kang, Myungjoo
    EXPERT SYSTEMS WITH APPLICATIONS, 2021, 173