Processing Natural Language on Embedded Devices: How Well Do Transformer Models Perform?

被引:0
|
作者
Sarkar, Souvika [1 ]
Babar, Mohammad Fakhruddin [2 ]
Hassan, Md Mahadi [1 ]
Hasan, Monowar [2 ]
Santu, Shubhra Kanti Karmaker [1 ]
机构
[1] Auburn Univ, Auburn, AL 36849 USA
[2] Washington State Univ, Pullman, WA USA
来源
PROCEEDINGS OF THE 15TH ACM/SPEC INTERNATIONAL CONFERENCE ON PERFORMANCE ENGINEERING, ICPE 2024 | 2024年
基金
美国国家科学基金会;
关键词
Transformers; Embedded Systems; NLP; Language Models;
D O I
10.1145/3629526.3645054
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Transformer-based language models such as BERT and its variants are primarily developed with compute-heavy servers in mind. Despite the great performance of BERT models across various NLP tasks, their large size and numerous parameters pose substantial obstacles to offline computation on embedded systems. Lighter replacements of such language models (e.g., DistilBERT and TinyBERT) often sacrifice accuracy, particularly for complex NLP tasks. Until now, it is still unclear (a) whether the state-of-the-art language models, viz., BERT and its variants are deployable on embedded systems with a limited processor, memory, and battery power and (b) if they do, what are the "right" set of configurations and parameters to choose for a given NLP task. This paper presents a performance study of transformer language models under different hardware configurations and accuracy requirements and derives empirical observations about these resource/accuracy trade-offs. In particular, we study how the most commonly used BERT-based language models (viz., BERT, RoBERTa, DistilBERT, and TinyBERT) perform on embedded systems. We tested them on four off-the-shelf embedded platforms (Raspberry Pi, Jetson, UP2, and UDOO) with 2 GB and 4 GB memory (i.e., a total of eight hardware configurations) and four datasets (i.e., HuRIC, GoEmotion, CoNLL, WNUT17) running various NLP tasks. Our study finds that executing complex NLP tasks (such as "sentiment" classification) on embedded systems is feasible even without any GPUs (e.g., Raspberry Pi with 2 GB of RAM). Our findings can help designers understand the deployability and performance of transformer language models, especially those based on BERT architectures.
引用
收藏
页码:211 / 222
页数:12
相关论文
共 32 条
  • [21] Knowledge-Based Verification of Concatenative Programming Patterns Inspired by Natural Language for Resource-Constrained Embedded Devices
    Gaglio, Salvatore
    Lo Re, Giuseppe
    Martorella, Gloria
    Peri, Daniele
    SENSORS, 2021, 21 (01) : 1 - 23
  • [22] From explainable to interpretable deep learning for natural language processing in healthcare: How far from reality?
    Huang, Guangming
    Li, Yingya
    Jameel, Shoaib
    Long, Yunfei
    Papanastasiou, Giorgos
    COMPUTATIONAL AND STRUCTURAL BIOTECHNOLOGY JOURNAL, 2024, 24 : 362 - 373
  • [23] Qualitative and quantitative research in the humanities and social sciences: how natural language processing (NLP) can help
    Franzosi R.
    Dong W.
    Dong Y.
    Quality & Quantity, 2022, 56 (4) : 2751 - 2781
  • [24] Optimizing Resource Allocation in Cloud for Large-Scale Deep Learning Models in Natural Language Processing
    Dhopavkar, Gauri
    Welekar, Rashmi R.
    Ingole, Piyush K.
    Vaidya, Chandu
    Wankhade, Shalini Vaibhav
    Vasgi, Bharati P.
    JOURNAL OF ELECTRICAL SYSTEMS, 2023, 19 (03) : 62 - 77
  • [25] A Natural Language Processing (NLP) Framework for Embedded Systems to Automatically Extract Verification Aspects from Textual Design Requirements
    Anwar, Muhammad Waseem
    Ahsan, Imran
    Azam, Farooque
    Butt, Wasi Haider
    Rashid, Muhammad
    PROCEEDINGS OF 2020 12TH INTERNATIONAL CONFERENCE ON COMPUTER AND AUTOMATION ENGINEERING (ICCAE 2020), 2020, : 7 - 12
  • [26] Exploring the applications of natural language processing and language models for production, planning, and control activities of SMEs in industry 4.0: a systematic literature review
    Mathieu, Bourdin
    Anas, Neumann
    Thomas, Paviot
    Robert, Pellerin
    Samir, Lamouri
    JOURNAL OF INTELLIGENT MANUFACTURING, 2024,
  • [27] Enhancing ASD detection accuracy: a combined approach of machine learning and deep learning models with natural language processing
    Rubio-Martin, Sergio
    Garcia-Ordas, Maria Teresa
    Bayon-Gutierrez, Martin
    Prieto-Fernandez, Natalia
    Benitez-Andrades, Jose Alberto
    HEALTH INFORMATION SCIENCE AND SYSTEMS, 2024, 12 (01)
  • [28] An Empirical Evaluation of Prompting Strategies for Large Language Models in Zero-Shot Clinical Natural Language Processing: Algorithm Development and Validation Study
    Sivarajkumar, Sonish
    Kelley, Mark
    Samolyk-Mazzanti, Alyssa
    Visweswaran, Shyam
    Wang, Yanshan
    JMIR MEDICAL INFORMATICS, 2024, 12
  • [29] Variability in Large Language Models' Responses to Medical Licensing and Certification Examinations. Comment on "How Does ChatGPT Perform on the United States Medical Licensing Examination? The Implications of Large Language Models for Medical Education and Knowledge Assessment"
    Epstein, Richard H.
    Dexter, Franklin
    JMIR MEDICAL EDUCATION, 2023, 9
  • [30] From rule-based models to deep learning transformers architectures for natural language processing and sign language translation systems: survey, taxonomy and performance evaluation
    Shahin, Nada
    Ismail, Leila
    ARTIFICIAL INTELLIGENCE REVIEW, 2024, 57 (10)