Energy-Efficient DNN Training Processorson Micro-AI Systems

被引:5
作者
Han, Donghyeon [1 ]
Kang, Sanghoon [1 ]
Kim, Sangyeob [1 ]
Lee, Juhyoung [1 ]
Yoo, Hoi-Jun [1 ]
机构
[1] Korea Adv Inst Sci & Technol, Sch Elect Engn, Daejeon 34141, South Korea
来源
IEEE OPEN JOURNAL OF THE SOLID-STATE CIRCUITS SOCIETY | 2022年 / 2卷
关键词
Training; Program processors; Artificial intelligence; System-on-chip; Solid state circuits; Performance evaluation; Energy efficiency; Backpropagation (BP); backward unlocking (BU); bit-precision optimization; deep neural network (DNN) training; reading transposed weight; sparsity exploitation; FACE RECOGNITION PROCESSOR; LEARNING PROCESSOR; LOW-POWER; ACCELERATOR; TUTORIAL;
D O I
10.1109/OJSSCS.2022.3219034
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Many edge/mobile devices are now able to utilize deep neural networks (DNNs) thanks to the development of mobile DNN accelerators. Mobile DNN accelerators overcame the problems of limited computing resources and battery capacity by realizing energy-efficient inference. However, its passive behavior makes it difficult for DNN to provide active customization for individual users or its service environment. The importance of on-chip training is rising more and more to provide active interaction between DNN processors and ever-changing surroundings or conditions. Despite its advantages, the DNN training has more constraints than the inference such that it was considered impractical to be realized on mobile/edge devices. Recently, there are many trials to realize mobile DNN training, and a number of prior works will be summarized. First, it arranges the new challenges of the DNN accelerator induced by training functionality and discusses new hardware features related to the challenges. Second, it explains algorithm-hardware co-optimization methods and explains why it becomes mainstream in mobile DNN training research. Third, it compares the main differences between the conventional inference accelerators and recent training processors. Finally, the conclusion is made by proposing the future directions of the DNN training processor in micro-AI systems.
引用
收藏
页码:259 / 275
页数:17
相关论文
共 50 条
[41]   HunterPlus: AI based energy-efficient task scheduling for cloud-fog computing environments [J].
Iftikhar, Sundas ;
Ahmad, Mirza Mohammad Mufleh ;
Tuli, Shreshth ;
Chowdhury, Deepraj ;
Xu, Minxian ;
Gill, Sukhpal Singh ;
Uhlig, Steve .
INTERNET OF THINGS, 2023, 21
[42]   Energy-Efficient Resource Allocation in Generative AI-Aided Secure Semantic Mobile Networks [J].
Zheng, Jie ;
Du, Baoxia ;
Du, Hongyang ;
Kang, Jiawen ;
Niyato, Dusit ;
Zhang, Haijun .
IEEE TRANSACTIONS ON MOBILE COMPUTING, 2024, 23 (12) :11422-11435
[43]   A Scoping Review of Energy-Efficient Driving Behaviors and Applied State-of-the-Art AI Methods [J].
Ma, Zhipeng ;
Jorgensen, Bo Norregaard ;
Ma, Zheng .
ENERGIES, 2024, 17 (02)
[44]   Energy-efficient computing-in-memory architecture for AI processor: device, circuit, architecture perspective [J].
Chang, Liang ;
Li, Chenglong ;
Zhang, Zhaomin ;
Xiao, Jianbiao ;
Liu, Qingsong ;
Zhu, Zhen ;
Li, Weihang ;
Zhu, Zixuan ;
Yang, Siqi ;
Zhou, Jun .
SCIENCE CHINA-INFORMATION SCIENCES, 2021, 64 (06)
[45]   Workflow Provenance in the Computing Continuum for Responsible, Trustworthy, and Energy-Efficient AI [J].
Souza, Renan ;
Caino-Lores, Silvina ;
Coletti, Mark ;
Skluzacek, Tyler J. ;
Costan, Alexandru ;
Suter, Frederic ;
Mattoso, Marta ;
da Silva, Rafael Ferreira .
2024 IEEE 20TH INTERNATIONAL CONFERENCE ON E-SCIENCE, E-SCIENCE 2024, 2024,
[46]   Memory-Based Computing for Energy-Efficient AI: Grand Challenges [J].
Karimzadeh, Foroozan ;
Imani, Mohsen ;
Asgari, Bahar ;
Cao, Ningyuan ;
Lin, Yingyan ;
Fang, Yan .
2023 IFIP/IEEE 31ST INTERNATIONAL CONFERENCE ON VERY LARGE SCALE INTEGRATION, VLSI-SOC, 2023, :110-117
[47]   Background Noise Adaptive Energy-Efficient Keywords Recognition Processor With Reusable DNN and Reconfigurable Architecture [J].
He, Guoqiang ;
Ding, Xiaoling ;
Zhou, Minghao ;
Liu, Bo ;
Li, Li .
IEEE ACCESS, 2022, 10 :17819-17827
[48]   PL-NPU: An Energy-Efficient Edge-Device DNN Training Processor With Posit-Based Logarithm-Domain Computing [J].
Wang, Yang ;
Deng, Dazheng ;
Liu, Leibo ;
Wei, Shaojun ;
Yin, Shouyi .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2022, 69 (10) :4042-4055
[49]   InterGrad: Energy-Efficient Training of Convolutional Neural Networks via Interleaved Gradient Scheduling [J].
Unnikrishnan, Nanda K. ;
Parhi, Keshab K. .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2023, 70 (05) :1949-1962
[50]   Survey on Optimization Models for Energy-Efficient Computing Systems [J].
Jozefowska, Joanna ;
Nowak, Mariusz ;
Rozycki, Rafal ;
Waligora, Grzegorz .
ENERGIES, 2022, 15 (22)