Improving Accuracy and Efficiency in Time Series Forecasting with an Optimized Transformer Model

被引:0
作者
Chen, Junhong [1 ]
Dai, Hong [1 ]
Wang, Shuang [1 ]
Liu, Chengrui [1 ]
机构
[1] Univ Sci & Technol Liaoning, Sch Comp Sci & Software Engn, Anshan, Liaoning, Peoples R China
关键词
Time series forecasting; Deep learning; Transformer-based models; Self-Attention; NEURAL-NETWORK; PREDICTION;
D O I
暂无
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
Time series forecasting (TSF) is a prevalent research task in various fields such as medicine, transportation, environment, network detection, finance, and others. The TSF task aims to identify underlying patterns in data and make relatively accurate estimates of future data based on known values. In recent years, deep learning models have gained popularity for TSF tasks due to their capability to capture internal information effectively. However, traditional deep-learning models encounter difficulties when parallelizing data calculations, leading to error accumulation and reduced forecasting accuracy. Additionally, when dealing with excessively long input data, traditional deep learning models may experience performance degradation despite providing sufficient information and making it arduous to predict future data. Transformer-based models, with Self-Attention as the core, have shown the ability to facilitate global information interaction and enhance prediction efficiency. Nonetheless, they may encounter problems with significant and redundant parameters, causing unnecessary time overhead. To overcome these challenges, we propose a novel model called VarSeg-Trans, which incorporates three key optimizations: the cut-up mechanism, the variables-isolating mechanism, and an improved attention calculation method to enhance the transformer model's performance. Specifically, the cut-up mechanism enables the model to process longer input sequences, the variables-isolating mechanism mitigates overfitting, and the improved attention method leverages sequence information more effectively. Compared to other baseline TSF models and previous Transformer-based models, VarSeg-Trans has achieved an average reduction of 9% in MSE and MAE, along with a 3% increase in the coefficient of determination R2. This trend is substantiated by consistent results across multiple experimental trials.
引用
收藏
页码:1 / 11
页数:11
相关论文
共 35 条
[1]   A Review on Deep Learning with Focus on Deep Recurrent Neural Network for Electricity Forecasting in Residential Building [J].
Abdulrahman, Mustapha Lawal ;
Ibrahim, Kabiru Musa ;
Gital, Abdusalam Yau ;
Zambuk, Fatima Umar ;
Ja'afaru, Badamasi ;
Yakubu, Zahraddeen Ismail ;
Ibrahim, Abubakar .
10TH INTERNATIONAL YOUNG SCIENTISTS CONFERENCE IN COMPUTATIONAL SCIENCE (YSC2021), 2021, 193 :141-154
[2]  
Bai L, 2020, ADV NEUR IN, V33
[3]   Short-term water quality variable prediction using a hybrid CNN-LSTM deep learning model [J].
Barzegar, Rahim ;
Aalami, Mohammad Taghi ;
Adamowski, Jan .
STOCHASTIC ENVIRONMENTAL RESEARCH AND RISK ASSESSMENT, 2020, 34 (02) :415-433
[4]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[5]   UTNet: A Hybrid Transformer Architecture for Medical Image Segmentation [J].
Gao, Yunhe ;
Zhou, Mu ;
Metaxas, Dimitris N. .
MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2021, PT III, 2021, 12903 :61-71
[6]   Pre-trained models: Past, present and future [J].
Han, Xu ;
Zhang, Zhengyan ;
Ding, Ning ;
Gu, Yuxian ;
Liu, Xiao ;
Huo, Yuqi ;
Qiu, Jiezhong ;
Yao, Yuan ;
Zhang, Ao ;
Zhang, Liang ;
Han, Wentao ;
Huang, Minlie ;
Jin, Qin ;
Lan, Yanyan ;
Liu, Yang ;
Liu, Zhiyuan ;
Lu, Zhiwu ;
Qiu, Xipeng ;
Song, Ruihua ;
Tang, Jie ;
Wen, Ji-Rong ;
Yuan, Jinhui ;
Zhao, Wayne Xin ;
Zhu, Jun .
AI OPEN, 2021, 2 :225-250
[7]  
Ioffe S, 2015, PR MACH LEARN RES, V37, P448
[8]   ARIMA Modeling With Intervention to Forecast and Analyze Chinese Stock Prices [J].
Jarrett, Jeffrey E. ;
Kyper, Eric .
INTERNATIONAL JOURNAL OF ENGINEERING BUSINESS MANAGEMENT, 2011, 3 (03) :53-58
[9]  
Jiang N, 2021, ENG LET, V29, P765
[10]   Short-Term Residential Load Forecasting Based on LSTM Recurrent Neural Network [J].
Kong, Weicong ;
Dong, Zhao Yang ;
Jia, Youwei ;
Hill, David J. ;
Xu, Yan ;
Zhang, Yuan .
IEEE TRANSACTIONS ON SMART GRID, 2019, 10 (01) :841-851