Time Series Classification With Large Language Models via Linguistic Scaffolding

被引:0
|
作者
Jang, Hyeongwon [1 ,2 ]
Yong Yang, June [2 ]
Hwang, Jaeryong [3 ]
Yang, Eunho [2 ,4 ]
机构
[1] Seoul Natl Univ, Dept Math Sci, Seoul 08826, South Korea
[2] Korea Adv Inst Sci & Technol, Kim Jaechul Grad Sch Artificial Intelligence, Daejeon ] 34141, South Korea
[3] Republ Korea Naval Acad, Dept Cyber Sci, Changwon Si 51704, South Korea
[4] AITRICS, Seoul 06221, South Korea
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Time series analysis; Transformers; Linguistics; Large language models; Semantics; Cognition; Predictive models; Vectors; Sensors; Encoding; Artificial neural networks; Classification algorithms; deep learning; irregular time series classification; pre-trained language models; time series classification;
D O I
10.1109/ACCESS.2024.3475471
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Time series classification requires specialized models that can effectively capture temporal structures. Consequently, Large Language Models (LLMs) have emerged as promising candidates due to their proficiency in sequence modeling and semantic reasoning. However, converting time series data into text results in sequences that exceed the maximum token limit, necessitating truncation or the removal of word embeddings for fixed-length time series embeddings. This restriction not only sacrifices semantic reasoning capabilities accessed through natural language but also limits the ability to handle temporal irregularities. To overcome these challenges, we propose the Language-Scaffolded Time Series Transformer (LSTST), which combines linguistic components and time series embeddings to effectively harness LLMs while overcoming dimensional constraints. Our Language Scaffold reformulates time series classification as a contextual question-answering task, with time series embeddings as context, facilitating the LLM to utilize its inherent semantic knowledge. Moreover, the preserved linguistic structure allows a dynamic number of input context embeddings with real-time positional encoding, handling length restrictions and irregularity in the temporal dimension. Through experiments, we show that LSTST achieves state-of-the-art performance on regular time series classification and also handles irregular time series without any model modifications.
引用
收藏
页码:170387 / 170398
页数:12
相关论文
共 50 条
  • [1] Time-Series Large Language Models: A Systematic Review of State-of-the-Art
    Abdullahi, Shamsu
    Danyaro, Kamaluddeen Usman
    Zakari, Abubakar
    Aziz, Izzatdin Abdul
    Zawawi, Noor Amila Wan Abdullah
    Adamu, Shamsuddeen
    IEEE ACCESS, 2025, 13 : 30235 - 30261
  • [2] Large Language Models on Graphs: A Comprehensive Survey
    Jin, Bowen
    Liu, Gang
    Han, Chi
    Jiang, Meng
    Ji, Heng
    Han, Jiawei
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (12) : 8622 - 8642
  • [3] Large language models and linguistic intentionality
    Grindrod, Jumbly
    SYNTHESE, 2024, 204 (02)
  • [4] Time Series Classification with Linguistic Summaries
    Kaczmarek, Katarzyna
    Hryniewicz, Olgierd
    PROCEEDINGS OF THE 2015 CONFERENCE OF THE INTERNATIONAL FUZZY SYSTEMS ASSOCIATION AND THE EUROPEAN SOCIETY FOR FUZZY LOGIC AND TECHNOLOGY, 2015, 89 : 471 - 477
  • [5] Can Large Language Models be Anomaly Detectors for Time Series?
    Alnegheimish, Sarah
    Nguyen, Linh
    Berti-Equille, Laure
    Veeramachaneni, Kalyan
    2024 IEEE 11TH INTERNATIONAL CONFERENCE ON DATA SCIENCE AND ADVANCED ANALYTICS, DSAA 2024, 2024, : 218 - 227
  • [6] Large margin mixture of AR models for time series classification
    Kini, B. Venkataramana
    Sekhar, C. Chandra
    APPLIED SOFT COMPUTING, 2013, 13 (01) : 361 - 371
  • [7] Linguistic Intelligence in Large Language Models for Telecommunications
    Ahmed, Tasnim
    Piovesan, Nicola
    De Domenico, Antonio
    Choudhury, Salimur
    2024 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS WORKSHOPS, ICC WORKSHOPS 2024, 2024, : 1237 - 1243
  • [8] Domain Adaptation of Time Series Classification
    Wang, Xinli
    Sun, Rencheng
    IEEE ACCESS, 2024, 12 : 38456 - 38472
  • [9] Synthetic Time Series Generation for Decision Intelligence Using Large Language Models
    Grigoras, Alexandru
    Leon, Florin
    MATHEMATICS, 2024, 12 (16)
  • [10] LegalReasoner: A Multi-Stage Framework for Legal Judgment Prediction via Large Language Models and Knowledge Integration
    Wang, Xuran
    Zhang, Xinguang
    Hoo, Vanessa
    Shao, Zhouhang
    Zhang, Xuguang
    IEEE ACCESS, 2024, 12 : 166843 - 166854