Architecture of neural processing unit for deep neural networks

被引:15
|
作者
Lee, Kyuho J. [1 ]
机构
[1] Ulsan Natl Inst Sci & Technol, Artificial Intelligence Grad Sch, Sch Elect & Comp Engn, Ulsan, South Korea
来源
HARDWARE ACCELERATOR SYSTEMS FOR ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING | 2021年 / 122卷
基金
新加坡国家研究基金会;
关键词
ACCELERATOR;
D O I
10.1016/bs.adcom.2020.11.001
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep Neural Networks (DNNs) have become a promising solution to inject AI in our daily lives from self-driving cars, smartphones, games, drones, etc. In most cases, DNNs were accelerated by server equipped with numerous computing engines, e.g., GPU, but recent technology advance requires energy-efficient acceleration of DNNs as the modern applications moved down to mobile computing nodes. Therefore, Neural Processing Unit (NPU) architectures dedicated to energy-efficient DNN acceleration became essential. Despite the fact that training phase of DNN requires precise number representations, many researchers proved that utilizing smaller bit-precision is enough for inference with low-power consumption. This led hardware architects to investigate energy-efficient NPU architectures with diverse HW-SW co-optimization schemes for inference. This chapter provides a review of several design examples of latest NPU architecture for DNN, mainly about inference engines. It also provides a discussion on the new architectural researches of neuromorphic computers and processing-in-memory architecture, and provides perspectives on the future research directions.
引用
收藏
页码:217 / 245
页数:29
相关论文
共 50 条
  • [31] Joint device architecture algorithm codesign of the photonic neural processing unit
    Pei, Li
    Xi, Zeya
    Bai, Bing
    Wang, Jianshuai
    Zheng, Jingjing
    Li, Jing
    Ning, Tigang
    ADVANCED PHOTONICS NEXUS, 2023, 2 (03):
  • [32] On the Importance of Network Architecture in Training Very Deep Neural Networks
    Chi, Zhizhen
    Li, Hongyang
    Wang, Jingjing
    Lu, Huchuan
    2016 IEEE INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING, COMMUNICATIONS AND COMPUTING (ICSPCC), 2016,
  • [33] Automatic Generation of Dynamic Inference Architecture for Deep Neural Networks
    Zhao, Shize
    He, Liulu
    Xie, Xiaoru
    Lin, Jun
    Wang, Zhongfeng
    2021 IEEE WORKSHOP ON SIGNAL PROCESSING SYSTEMS (SIPS 2021), 2021, : 117 - 122
  • [34] Architecture-Preserving Provable Repair of Deep Neural Networks
    Tao, Zhe
    Nawas, Stephanie
    Mitchell, Jacqueline
    Thakur, Aditya V.
    PROCEEDINGS OF THE ACM ON PROGRAMMING LANGUAGES-PACMPL, 2023, 7 (PLDI):
  • [35] Developing a Hybrid Network Architecture for Deep Convolutional Neural Networks
    Sayan, H. Huseyin
    Tekgozoglu, O. Faruk
    Sonmez, Yusuf
    Turan, Bilal
    ARTIFICIAL INTELLIGENCE AND APPLIED MATHEMATICS IN ENGINEERING PROBLEMS, 2020, 43 : 750 - 757
  • [36] DeepRecon: Dynamically Reconfigurable Architecture for Accelerating Deep Neural Networks
    Rzayev, Tayyar
    Moradi, Saber
    Albonesi, David H.
    Manohar, Rajit
    2017 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2017, : 116 - 124
  • [37] A novel softplus linear unit for deep convolutional neural networks
    Zhao, Huizhen
    Liu, Fuxian
    Li, Longyue
    Luo, Chang
    APPLIED INTELLIGENCE, 2018, 48 (07) : 1707 - 1720
  • [38] Parametric Exponential Linear Unit for Deep Convolutional Neural Networks
    Trottier, Ludovic
    Giguere, Philippe
    Chaib-draa, Brahim
    2017 16TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA), 2017, : 207 - 214
  • [39] Neural Architecture Search for Spiking Neural Networks
    Kim, Youngeun
    Li, Yuhang
    Park, Hyoungseob
    Venkatesha, Yeshwanth
    Panda, Priyadarshini
    COMPUTER VISION, ECCV 2022, PT XXIV, 2022, 13684 : 36 - 56
  • [40] Deep neural networks predicting oil movement in a development unit
    Temirchev, P.
    Simonov, M.
    Kostoev, R.
    Burnaev, E.
    Oseledets, I
    Akhmetov, A.
    Margarit, A.
    Sitnikov, A.
    Koroteev, D.
    JOURNAL OF PETROLEUM SCIENCE AND ENGINEERING, 2020, 184