Dependability and Protection of Transformer Models Against Soft Errors on Text Embeddings

被引:1
|
作者
Gao, Zhen [1 ]
Liu, Shuang [1 ]
Reviriego, Pedro [2 ]
Liu, Shanshan [3 ]
Lombardi, Fabrizio [4 ]
机构
[1] Tianjin Univ, Sch Elect & Informat Engn, Tianjin 300072, Peoples R China
[2] Univ Politecn Madrid, ETSI Telecomunicac, Madrid 28040, Spain
[3] Univ Elect Sci & Technol China, Sch Informat & Commun Engn, Chengdu 611731, Sichuan, Peoples R China
[4] Northeastern Univ, Dept Elect & Comp Engn, Boston, MA 02215 USA
关键词
Transformers; Vectors; Encoding; Bidirectional control; Bit error rate; Protection; Natural language processing; Text summarization; Semantics; Predictive models; embedding parameters; soft errors; dependability; BERT; T5; CLIP; CODES;
D O I
10.1109/TDMR.2024.3478753
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Transformers have achieved remarkable success in diverse fields such as Natural Language Processing (NLP) and computer vision (CV). For pre-trained Transformer models involving text processing, embedding representations are important parameters, incurring a large volume of memory. Soft errors on embedding vectors can lead to incorrect inputs to Transformers, and if not corrected in time, accumulated errors may produce undesirable outcomes. This paper considers the dependability of text related Transformer models to accumulated errors on embedding parameters and takes three typical models in different applications as case studies: BERT based sentence emotion classification, T5 based text summarization, and CLIP based image classification. We first evaluate the dependability of the three models by injecting bit errors on embedding parameters; only errors on a few critical bits affect model performance. Based on this finding, we first propose an efficient selective protection for embedding parameters with small values, and then through scaling, we extend the scheme for models with large embedding parameters. Extensive simulation results show that the proposed protection scheme can effectively remove the impact of soft errors on task performance. In particular, the complexity overhead of the proposed scheme is negligible, and the additional memory overhead as encountered in the SEC scheme is avoided.
引用
收藏
页码:54 / 65
页数:12
相关论文
共 50 条
  • [21] HARDENING RAMS AGAINST SOFT ERRORS
    BRODSKY, M
    ELECTRONICS, 1980, 53 (10): : 117 - 122
  • [22] System-Level Hardware-Based Protection of Memories against Soft-Errors
    Gherman, Valentin
    Evain, Samuel
    Cartron, Mickael
    Seymour, Nathaniel
    Bonhomme, Yannick
    DATE: 2009 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION, VOLS 1-3, 2009, : 1222 - 1225
  • [23] Using Word Embeddings with Linear Models for Short Text Classification
    Krzywicki, Alfred
    Heap, Bradford
    Bain, Michael
    Wobcke, Wayne
    Schmeidl, Susanne
    AI 2018: ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, 11320 : 819 - 827
  • [24] Protection against soft errors in the space environment: A finite impulse response (FIR) filter case study
    Maestro, J. A.
    Reviriego, P.
    Reyes, R.
    Ruano, O.
    INTEGRATION-THE VLSI JOURNAL, 2009, 42 (02) : 128 - 136
  • [25] FPGA-Based Reliable Fault Secure Design for Protection against Single and Multiple Soft Errors
    Shaker, Manar N.
    Hussien, Ahmed
    Alkady, Gehad I.
    Amer, Hassanein H.
    Adly, Ihab
    ELECTRONICS, 2020, 9 (12) : 1 - 13
  • [26] HARDENING RAMS AGAINST SOFT ERRORS.
    Brodsky, Mark
    Electronics, 1980, 53 (10): : 117 - 122
  • [27] Power-Efficient Protection from Soft Errors
    Lee, Kyoungwoo
    Shrivastava, Aviral
    Jeyapaul, Reiley
    2013 26TH INTERNATIONAL CONFERENCE ON VLSI DESIGN AND 2013 12TH INTERNATIONAL CONFERENCE ON EMBEDDED SYSTEMS (VLSID), 2013, : LXXI - LXXII
  • [28] PROTECTION AGAINST EXTERNAL ERRORS IN A DEDICATED SYSTEM
    BELLON, C
    SAUCIER, G
    IEEE TRANSACTIONS ON COMPUTERS, 1982, 31 (04) : 311 - 317
  • [29] Text Representation Models based on the Spatial Distributional Properties of Word Embeddings
    Unnam, Narendra Babu
    Reddy, P. Krishna
    Pandey, Amit
    Manwani, Naresh
    PROCEEDINGS OF 7TH JOINT INTERNATIONAL CONFERENCE ON DATA SCIENCE AND MANAGEMENT OF DATA, CODS-COMAD 2024, 2024, : 603 - 604
  • [30] Assessing the Effectiveness of Multilingual Transformer-based Text Embeddings for Named Entity Recognition in Portuguese
    de Lima Santos, Diego Bernardes
    de Carvalho Dutra, Frederico Giffoni
    Parreiras, Fernando Silva
    Brandao, Wladmir Cardoso
    PROCEEDINGS OF THE 23RD INTERNATIONAL CONFERENCE ON ENTERPRISE INFORMATION SYSTEMS (ICEIS 2021), VOL 1, 2021, : 473 - 483