Time Series Classification With Large Language Models via Linguistic Scaffolding

被引:0
|
作者
Jang, Hyeongwon [1 ,2 ]
Yong Yang, June [2 ]
Hwang, Jaeryong [3 ]
Yang, Eunho [2 ,4 ]
机构
[1] Seoul Natl Univ, Dept Math Sci, Seoul 08826, South Korea
[2] Korea Adv Inst Sci & Technol, Kim Jaechul Grad Sch Artificial Intelligence, Daejeon ] 34141, South Korea
[3] Republ Korea Naval Acad, Dept Cyber Sci, Changwon Si 51704, South Korea
[4] AITRICS, Seoul 06221, South Korea
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Time series analysis; Transformers; Linguistics; Large language models; Semantics; Cognition; Predictive models; Vectors; Sensors; Encoding; Artificial neural networks; Classification algorithms; deep learning; irregular time series classification; pre-trained language models; time series classification;
D O I
10.1109/ACCESS.2024.3475471
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Time series classification requires specialized models that can effectively capture temporal structures. Consequently, Large Language Models (LLMs) have emerged as promising candidates due to their proficiency in sequence modeling and semantic reasoning. However, converting time series data into text results in sequences that exceed the maximum token limit, necessitating truncation or the removal of word embeddings for fixed-length time series embeddings. This restriction not only sacrifices semantic reasoning capabilities accessed through natural language but also limits the ability to handle temporal irregularities. To overcome these challenges, we propose the Language-Scaffolded Time Series Transformer (LSTST), which combines linguistic components and time series embeddings to effectively harness LLMs while overcoming dimensional constraints. Our Language Scaffold reformulates time series classification as a contextual question-answering task, with time series embeddings as context, facilitating the LLM to utilize its inherent semantic knowledge. Moreover, the preserved linguistic structure allows a dynamic number of input context embeddings with real-time positional encoding, handling length restrictions and irregularity in the temporal dimension. Through experiments, we show that LSTST achieves state-of-the-art performance on regular time series classification and also handles irregular time series without any model modifications.
引用
收藏
页码:170387 / 170398
页数:12
相关论文
共 50 条
  • [11] Algebra Error Classification with Large Language Models
    McNichols, Hunter
    Zhang, Mengxue
    Lan, Andrew
    ARTIFICIAL INTELLIGENCE IN EDUCATION, AIED 2023, 2023, 13916 : 365 - 376
  • [12] Finding Foundation Models for Time Series Classification with a PreText Task
    Ismail-Fawaz, Ali
    Devanne, Maxime
    Berretti, Stefano
    Weber, Jonathan
    Forestier, Germain
    TRENDS AND APPLICATIONS IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2024 WORKSHOPS, RAFDA AND IWTA, 2024, 14658 : 123 - 135
  • [13] CoLLM: Integrating Collaborative Embeddings Into Large Language Models for Recommendation
    Zhang, Yang
    Feng, Fuli
    Zhang, Jizhi
    Bao, Keqin
    Wang, Qifan
    He, Xiangnan
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2025, 37 (05) : 2329 - 2340
  • [14] Deep Learning Models for Time Series Forecasting: A Review
    Li, Wenxiang
    Law, K. L. Eddie
    IEEE ACCESS, 2024, 12 : 92306 - 92327
  • [15] A Review of Deep Learning Models for Time Series Prediction
    Han, Zhongyang
    Zhao, Jun
    Leung, Henry
    Ma, King Fai
    Wang, Wei
    IEEE SENSORS JOURNAL, 2021, 21 (06) : 7833 - 7848
  • [16] Self-Bidirectional Decoupled Distillation for Time Series Classification
    Xiao Z.
    Xing H.
    Qu R.
    Li H.
    Feng L.
    Zhao B.
    Yang J.
    IEEE Transactions on Artificial Intelligence, 2024, 5 (08): : 1 - 11
  • [17] Heterogeneous Feature Based Time Series Classification With Attention Mechanism
    Zhang, Hanbo
    Wang, Peng
    Liang, Shen
    Zhou, Tongming
    Wang, Wei
    IEEE ACCESS, 2023, 11 : 19373 - 19384
  • [18] iTimes: Investigating Semisupervised Time Series Classification via Irregular Time Sampling
    Liu, Xuxin
    Zhang, Fengbin
    Liu, Han
    Fan, Haoyi
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2023, 19 (05) : 6930 - 6938
  • [19] Large Language Models are Not Models of Natural Language: They are Corpus Models
    Veres, Csaba
    IEEE ACCESS, 2022, 10 : 61970 - 61979
  • [20] LLMental: Classification of Mental Disorders with Large Language Models
    Nowacki, Arkadiusz
    Sitek, Wojciech
    Rybinski, Henryk
    FOUNDATIONS OF INTELLIGENT SYSTEMS, ISMIS 2024, 2024, 14670 : 35 - 44