Architecture of neural processing unit for deep neural networks

被引:15
|
作者
Lee, Kyuho J. [1 ]
机构
[1] Ulsan Natl Inst Sci & Technol, Artificial Intelligence Grad Sch, Sch Elect & Comp Engn, Ulsan, South Korea
来源
HARDWARE ACCELERATOR SYSTEMS FOR ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING | 2021年 / 122卷
基金
新加坡国家研究基金会;
关键词
ACCELERATOR;
D O I
10.1016/bs.adcom.2020.11.001
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep Neural Networks (DNNs) have become a promising solution to inject AI in our daily lives from self-driving cars, smartphones, games, drones, etc. In most cases, DNNs were accelerated by server equipped with numerous computing engines, e.g., GPU, but recent technology advance requires energy-efficient acceleration of DNNs as the modern applications moved down to mobile computing nodes. Therefore, Neural Processing Unit (NPU) architectures dedicated to energy-efficient DNN acceleration became essential. Despite the fact that training phase of DNN requires precise number representations, many researchers proved that utilizing smaller bit-precision is enough for inference with low-power consumption. This led hardware architects to investigate energy-efficient NPU architectures with diverse HW-SW co-optimization schemes for inference. This chapter provides a review of several design examples of latest NPU architecture for DNN, mainly about inference engines. It also provides a discussion on the new architectural researches of neuromorphic computers and processing-in-memory architecture, and provides perspectives on the future research directions.
引用
收藏
页码:217 / 245
页数:29
相关论文
共 50 条
  • [1] Hardware Efficient Convolution Processing Unit for Deep Neural Networks
    Hazarika, Anakhi
    Poddar, Soumyajit
    Rahaman, Hafizur
    2019 2ND INTERNATIONAL SYMPOSIUM ON DEVICES, CIRCUITS AND SYSTEMS (ISDCS 2019), 2019,
  • [2] A Heterogeneous Architecture for the Vision Processing Unit with a Hybrid Deep Neural Network Accelerator
    Liu, Peng
    Yang, Zikai
    Kang, Lin
    Wang, Jian
    MICROMACHINES, 2022, 13 (02)
  • [3] Architecture Disentanglement for Deep Neural Networks
    Hu, Jie
    Cao, Liujuan
    Tong, Tong
    Ye, Qixiang
    Zhang, Shengchuan
    Li, Ke
    Huang, Feiyue
    Shao, Ling
    Ji, Rongrong
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 652 - 661
  • [4] An architecture of fuzzy neural networks for linguistic processing
    Bortolan, G
    FUZZY SETS AND SYSTEMS, 1998, 100 (1-3) : 197 - 215
  • [6] The Impact of Architecture on the Deep Neural Networks Training
    Rozycki, Pawel
    Kolbusz, Janusz
    Malinowski, Aleksander
    Wilamowski, Bogdan
    2019 12TH INTERNATIONAL CONFERENCE ON HUMAN SYSTEM INTERACTION (HSI), 2019, : 41 - 46
  • [7] Extending Neural Processing Unit and Compiler for Advanced Binarized Neural Networks
    Song, Minjoon
    Asim, Faaiz
    Lee, Jongeun
    29TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE, ASP-DAC 2024, 2024, : 115 - 120
  • [8] Hardware Architecture Exploration for Deep Neural Networks
    Zheng, Wenqi
    Zhao, Yangyi
    Chen, Yunfan
    Park, Jinhong
    Shin, Hyunchul
    ARABIAN JOURNAL FOR SCIENCE AND ENGINEERING, 2021, 46 (10) : 9703 - 9712
  • [9] LiteLSTM Architecture for Deep Recurrent Neural Networks
    Elsayed, Nelly
    ElSayed, Zag
    Maida, Anthony S.
    2022 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS 22), 2022, : 1304 - 1308
  • [10] An Architecture to Accelerate Convolution in Deep Neural Networks
    Ardakani, Arash
    Condo, Carlo
    Ahmadi, Mehdi
    Gross, Warren J.
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2018, 65 (04) : 1349 - 1362