Time-Series Large Language Models: A Systematic Review of State-of-the-Art

被引:0
作者
Abdullahi, Shamsu [1 ,2 ]
Danyaro, Kamaluddeen Usman [1 ]
Zakari, Abubakar [1 ,3 ]
Aziz, Izzatdin Abdul [1 ]
Zawawi, Noor Amila Wan Abdullah [4 ]
Adamu, Shamsuddeen [1 ]
机构
[1] Univ Teknol PETRONAS, Dept Comp & Informat Sci, Seri Iskandar 32610, Perak, Malaysia
[2] Hassan Usman Katsina Polytech, Dept Comp Sci, Katsina 820102, Katsina State, Nigeria
[3] Aliko Dangote Univ Sci & Technol, Dept Comp Sci, Kano 713101, Nigeria
[4] Univ Teknol PETRONAS, Civil & Environm Engn Dept, Seri Iskandar 32610, Perak, Malaysia
来源
IEEE ACCESS | 2025年 / 13卷
关键词
Time series analysis; Measurement; Tokenization; Systematic literature review; Systematics; Databases; Transformers; Search problems; Anomaly detection; Time-series; large language models; forecasting; tokenization; time-series LLMs; PREDICTION;
D O I
10.1109/ACCESS.2025.3535782
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Large Language Models (LLMs) have transformed Natural Language Processing (NLP) and Software Engineering by fostering innovation, streamlining processes, and enabling data-driven decision-making. Recently, the adoption of LLMs in time-series analysis has catalyzed the emergence of time-series LLMs, a rapidly evolving research area. Existing reviews provide foundational insights into time-series LLMs but lack a comprehensive examination of recent advancements and do not adequately address critical challenges in this domain. This Systematic Literature Review (SLR) bridges these gaps by analysing state-of-the-art contributions in time-series LLMs, focusing on architectural innovations, tokenisation strategies, tasks, datasets, evaluation metrics, and unresolved challenges. Using a rigorous methodology based on PRISMA guidelines, over 700 studies from 2020 to 2024 were reviewed, with 59 relevant studies selected from journals, conferences, and workshops. Key findings reveal advancements in architectures and novel tokenization strategies tailored for temporal data. Forecasting dominates the identified tasks with 79.66% of the selected studies, while classification and anomaly detection remain underexplored. Furthermore, the analysis reveals a strong reliance on datasets from the energy and transportation domains, highlighting the need for more diverse datasets. Despite these advancements, significant challenges persist, including tokenization inefficiencies, prediction hallucinations, and difficulties in modelling long-term dependencies. These issues hinder the robustness, scalability, and adaptability of time-series LLMs across diverse applications. To address these challenges, this SLR outlines a research roadmap emphasizing the improvement of tokenization methods, the development of mechanisms for capturing long-term dependencies, the mitigation of hallucination effects, and the design of scalable, interpretable models for diverse time-series tasks.
引用
收藏
页码:30235 / 30261
页数:27
相关论文
共 50 条
  • [11] A systematic review on the state-of-the-art strategies for protein representation
    Yue, Zi-Xuan
    Yan, Tian-Ci
    Xu, Hong-Quan
    Liu, Yu-Hong
    Hong, Yan-Feng
    Chen, Gong-Xing
    Xie, Tian
    Tao, Lin
    COMPUTERS IN BIOLOGY AND MEDICINE, 2023, 152
  • [12] A Survey on Time-Series Pre-Trained Models
    Ma, Qianli
    Liu, Zhen
    Zheng, Zhenjing
    Huang, Ziyang
    Zhu, Siying
    Yu, Zhongzhong
    Kwok, James T.
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (12) : 7536 - 7555
  • [13] Comparing classic time series models and state-of-the-art time series neural networks for forecasting as-fired liquor properties
    Ng, Jerry
    Lawryshyn, Yuri
    Demartini, Nikolai
    NORDIC PULP & PAPER RESEARCH JOURNAL, 2025, 40 (01) : 33 - 45
  • [14] Large language models for code completion: A systematic literature review
    Husein, Rasha Ahmad
    Aburajouh, Hala
    Catal, Cagatay
    COMPUTER STANDARDS & INTERFACES, 2025, 92
  • [15] Financial inclusion and fintech: a state-of-the-art systematic literature review
    Ha, Dao
    Le, Phuong
    Nguyen, Duc Khuong
    FINANCIAL INNOVATION, 2025, 11 (01)
  • [17] Transformers and large language models in healthcare: A review
    Nerella, Subhash
    Bandyopadhyay, Sabyasachi
    Zhang, Jiaqing
    Contreras, Miguel
    Siegel, Scott
    Bumin, Aysegul
    Silva, Brandon
    Sena, Jessica
    Shickel, Benjamin
    Bihorac, Azra
    Khezeli, Kia
    Rashidi, Parisa
    ARTIFICIAL INTELLIGENCE IN MEDICINE, 2024, 154
  • [18] Benchmarking State-of-the-Art Large Language Models for Migraine Patient Education: Performance Comparison of Responses to Common Queries
    Li, Linger
    Li, Pengfei
    Wang, Kun
    Zhang, Liang
    Ji, Hongwei
    Zhao, Hongqin
    JOURNAL OF MEDICAL INTERNET RESEARCH, 2024, 26
  • [19] LLaMA 3 vs. State-of-the-Art Large Language Models: Performance in Detecting Nuanced Fake News
    Repede, Stefan Emil
    Brad, Remus
    COMPUTERS, 2024, 13 (11)
  • [20] State of the Art of Machine Learning Models in Energy Systems, a Systematic Review
    Mosavi, Amir
    Salimi, Mohsen
    Ardabili, Sina Faizollahzadeh
    Rabczuk, Timon
    Shamshirband, Shahaboddin
    Varkonyi-Koczy, Annamaria R.
    ENERGIES, 2019, 12 (07)