TCAMS-Trans: Efficient temporal-channel attention multi-scale transformer for net load forecasting

被引:5
作者
Zhang, Qingyong [1 ]
Zhou, Shiyang [1 ]
Xu, Bingrong [1 ]
Li, Xinran [1 ]
机构
[1] Wuhan Univ Technol, Sch Automat, Wuhan 430070, Peoples R China
基金
中国国家自然科学基金;
关键词
Net load forecasting; Temporal-channel attention block; Transformer; Deep learning; Renewable energy sources;
D O I
10.1016/j.compeleceng.2024.109415
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Accurate net load forecasting contributes to increasing the integration of renewable energy sources and reducing the operating cost of the power grid. In recent years, deep learning models have been applied in net load forecasting. However, there is still a large scope for them to improve the extraction of complex multi-scale temporal features in the net load. In addition, most researches do not establish explicit mechanisms to consider the impact of relevant factors on net load forecasting, such as renewable energy outputs and load demand. Therefore, in this paper, we propose a multivariate Transformer-based model, named TCAMS-Trans, to achieve direct net load forecasting. First, TCAMS-Trans exploits multi-scale patch-based variable-independent embedding to model potential multi-scale patterns in the load data. Subsequently, TCAMS-Trans designs a new hierarchical encoder-decoder pair configured with temporal-channel attention block (TCA block) to further extract multi-scale features and capture the connections between the net load and the relevant factors. The temporal attention in TCA block is used to mine complex temporal dependency along the time dimension, and the channel attention is used to capture the inter-variable dependency along the variable dimension. Experiments conducted on three publicly available datasets show that TCAMS-Trans has higher prediction accuracy than the current state-of-the-art models at different prediction lengths. The model proposed in this paper provides a new powerful solution for net load forecasting.
引用
收藏
页数:16
相关论文
共 41 条