ERA-LSTM: An Efficient ReRAM-Based Architecture for Long Short-Term Memory

被引:29
|
作者
Han, Jianhui [1 ,2 ]
Liu, He [2 ,3 ]
Wang, Mingyu [1 ,2 ]
Li, Zhaolin [2 ,3 ]
Zhang, Youhui [2 ,3 ]
机构
[1] Tsinghua Univ, Inst Microelect, Beijing 100084, Peoples R China
[2] Beijing Natl Res Ctr Informat Sci & Technol, Beijing 100084, Peoples R China
[3] Tsinghua Univ, Dept Comp Sci & Technol, Beijing 100084, Peoples R China
基金
中国博士后科学基金;
关键词
Long short-term memory (LSTM); resistive random-access memory (ReRAM); processing in memory (PIM); approximate computing; accelerator; NETWORKS; DESIGN; PULSE;
D O I
10.1109/TPDS.2019.2962806
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Processing-in-memory (PIM) architecture based on resistive random access memory (ReRAM) crossbars is a promising solution to the memory bottleneck that long short-term memory (LSTM) faces. Based on the dataflow analysis of the LSTM computing paradigm, this article proposes to adopt the ReRAM-based analog approximate computing to conduct the LSTM-specific element-wise computation. Combined with the dot-product computation implemented with ReRAM crossbars, a new LSTM processing tile is designed to significantly reduce the demand for analog-to-digital converters (ADCs), which is the major part of power consumption of existing designs. Next, we elaborate on a mapping scheme to efficiently deploy large-scale LSTM onto multiple processing tiles. Finally, an architecture enhancement is proposed to support crossbar-friendly LSTM pruning to further improve efficiency. This overall design, named ERA-LSTM, is presented. Our evaluation shows that it can outperform two state-of-the-art FPGA-based LSTM accelerators by 103.6 and 35.9 times, respectively; compared with a state-of-the-art ReRAM-based LSTM accelerator with digital element-wise computation, it is 6.1 times more efficient. Moreover, our experiments demonstrate that the impact of hardware constraints and approximation errors on the inference accuracy can be effectively reduced by the proposed fine-tuning scheme and by optimizing the design of the approximator.
引用
收藏
页码:1328 / 1342
页数:15
相关论文
共 50 条
  • [1] E-LSTM: An Efficient Hardware Architecture for Long Short-Term Memory
    Wang, Meiqi
    Wang, Zhisheng
    Lu, Jinming
    Lin, Jun
    Wang, Zhongfeng
    IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS, 2019, 9 (02) : 280 - 291
  • [2] A Reversible-Logic based Architecture for Long Short-Term Memory (LSTM) Network
    Khalil, Kasem
    Dey, Bappaditya
    Kumar, Ashok
    Bayoumi, Magdy
    2021 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2021,
  • [3] A ReRAM-Based Processing-In-Memory Architecture for Hyperdimensional Computing
    Liu, Cong
    Wu, Kaibo
    Liu, Haikun
    Jin, Hai
    Liao, Xiaofei
    Duan, Zhuohui
    Xu, Jiahong
    Li, Huize
    Zhang, Yu
    Yang, Jing
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2025, 44 (02) : 512 - 524
  • [4] ReRAM-based Processing-in-Memory Architecture for Blockchain Platforms
    Wang, Fang
    Shen, Zhaoyan
    Han, Lei
    Shao, Zili
    24TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC 2019), 2019, : 615 - 620
  • [5] Long short-term memory (LSTM)-based news classification model
    Liu, Chen
    PLOS ONE, 2024, 19 (05):
  • [6] Lane Position Detection Based on Long Short-Term Memory (LSTM)
    Yang, Wei
    Zhang, Xiang
    Lei, Qian
    Shen, Dengye
    Xiao, Ping
    Huang, Yu
    SENSORS, 2020, 20 (11)
  • [7] Efficient Neural Architecture Search for Long Short-Term Memory Networks
    Abed, Hamdi
    Gyires-Toth, Balint
    2021 IEEE 19TH WORLD SYMPOSIUM ON APPLIED MACHINE INTELLIGENCE AND INFORMATICS (SAMI 2021), 2021, : 287 - 292
  • [8] RETRANSFORMER: ReRAM-based Processing-in-Memory Architecture for Transformer Acceleration
    Yang, Xiaoxuan
    Yan, Bonan
    Li, Hai
    Chen, Yiran
    2020 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED-DESIGN (ICCAD), 2020,
  • [9] A Novel ReRAM-based Processing-in-Memory Architecture for Graph Computing
    Han, Lei
    Shen, Zhaoyan
    Shao, Zili
    Huang, H. Howie
    Li, Tao
    2017 IEEE 6TH NON-VOLATILE MEMORY SYSTEMS AND APPLICATIONS SYMPOSIUM (NVMSA 2017), 2017,
  • [10] Arabic Language Opinion Mining Based on Long Short-Term Memory (LSTM)
    Setyanto, Arief
    Laksito, Arif
    Alarfaj, Fawaz
    Alreshoodi, Mohammed
    Kusrini
    Oyong, Irwan
    Hayaty, Mardhiya
    Alomair, Abdullah
    Almusallam, Naif
    Kurniasari, Lilis
    APPLIED SCIENCES-BASEL, 2022, 12 (09):