A Transformer-based Framework for Multivariate Time Series Representation Learning

被引:607
作者
Zerveas, George [1 ]
Jayaraman, Srideepika [2 ]
Patel, Dhaval [2 ]
Bhamidipaty, Anuradha [2 ]
Eickhoff, Carsten [1 ]
机构
[1] Brown Univ, Providence, RI 02912 USA
[2] IBM Res, Yorktown Hts, NY USA
来源
KDD '21: PROCEEDINGS OF THE 27TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING | 2021年
关键词
transformer; deep learning; multivariate time series; unsupervised learning; self-supervised learning; framework; regression; classification; imputation;
D O I
10.1145/3447548.3467401
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present a novel framework for multivariate time series representation learning based on the transformer encoder architecture. The framework includes an unsupervised pre-training scheme, which can offer substantial performance benefits over fully supervised learning on downstream tasks, both with but even without lever-aging additional unlabeled data, i.e., by reusing the existing data samples. Evaluating our framework on several public multivariate time series datasets from various domains and with diverse characteristics, we demonstrate that it performs significantly better than the best currently available methods for regression and classification, even for datasets which consist of only a few hundred training samples. Given the pronounced interest in unsupervised learning for nearly all domains in the sciences and in industry, these findings represent an important landmark, presenting the first unsupervised method shown to push the limits of state-of-the-art performance for multivariate time series regression and classification.
引用
收藏
页码:2114 / 2124
页数:11
相关论文
共 34 条
[1]  
[Anonymous], 2019, IEEE IJCNN, DOI DOI 10.1007/S10618-020-00710-Y
[2]  
Bagnall A., 2018, UEA MULTIVARIATE TIM
[3]   The great time series classification bake off: a review and experimental evaluation of recent algorithmic advances [J].
Bagnall, Anthony ;
Lines, Jason ;
Bostrom, Aaron ;
Large, James ;
Keogh, Eamonn .
DATA MINING AND KNOWLEDGE DISCOVERY, 2017, 31 (03) :606-660
[4]  
Beltagy I., 2020, Longformer: The Long-Document Transformer, V2004, P05150, DOI DOI 10.48550/ARXIV.2004.05150
[5]  
Bergmeir C., 2020, ARXIV200612672
[6]   Learning representations of multivariate time series with missing data [J].
Bianchi, Filippo Maria ;
Livi, Lorenzo ;
Mikalsen, Karl Oyvind ;
Kampffmeyer, Michael ;
Jenssen, Robert .
PATTERN RECOGNITION, 2019, 96
[7]  
Brown TB, 2020, ADV NEUR IN, V33
[8]  
Dai ZH, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P2978
[9]  
De Brouwer E, 2019, ADV NEUR IN, V32
[10]   ROCKET: exceptionally fast and accurate time series classification using random convolutional kernels [J].
Dempster, Angus ;
Petitjean, Francois ;
Webb, Geoffrey, I .
DATA MINING AND KNOWLEDGE DISCOVERY, 2020, 34 (05) :1454-1495