Neural Architecture Search and Hardware Accelerator Co-Search: A Survey

被引:21
作者
Sekanina, Lukas [1 ]
机构
[1] Brno Univ Technol, Fac Informat Technol, Brno 61266, Czech Republic
关键词
Neurons; Computer architecture; Hardware acceleration; Convolutional neural networks; Training; Optimization; Benchmark testing; Automated design; classification; co-design; deep neural network; hardware accelerator; neural architecture search; optimization; GENETIC ALGORITHM; NETWORKS; ENERGY; OPTIMIZATION; EVOLUTION; DESIGN; MODEL; NAS;
D O I
10.1109/ACCESS.2021.3126685
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep neural networks (DNN) are now dominating in the most challenging applications of machine learning. As DNNs can have complex architectures with millions of trainable parameters (the so-called weights), their design and training are difficult even for highly qualified experts. In order to reduce human effort, neural architecture search (NAS) methods have been developed to automate the entire design process. The NAS methods typically combine searching in the space of candidate architectures and optimizing (learning) the weights using a gradient method. In this paper, we survey the key elements of NAS methods that - to various extents - consider hardware implementation of the resulting DNNs. We classified these methods into three major classes: single-objective NAS (no hardware is considered), hardware-aware NAS (DNN is optimized for a particular hardware platform), and NAS with hardware co-optimization (hardware is directly co-optimized with DNN as a part of NAS). Compared to previous surveys, we emphasize the multi-objective design approach that must be adopted in NAS and focus on co-design algorithms developed for concurrent optimization of DNN architectures and hardware platforms. As most research in this area deals with NAS for image classification using convolutional neural networks, we follow this trajectory in our paper. After reading the paper, the reader should understand why and how NAS and hardware co-optimization are currently used to build cutting-edge implementations of DNNs.
引用
收藏
页码:151337 / 151362
页数:26
相关论文
共 177 条
[1]  
Abadi M, 2016, PROCEEDINGS OF OSDI'16: 12TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, P265
[2]   Best of Both Worlds: AutoML Codesign of a CNN and its Hardware Accelerator [J].
Abdelfattah, Mohamed S. ;
Dudziak, Lukasz ;
Chau, Thomas ;
Lee, Royson ;
Kim, Hyeji ;
Lane, Nicholas D. .
PROCEEDINGS OF THE 2020 57TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2020,
[3]   APNAS: Accuracy-and-Performance-Aware Neural Architecture Search for Neural Hardware Accelerators [J].
Achararit, Paniti ;
Hanif, Muhammad Abdullah ;
Putra, Rachmad Vidya Wicaksana ;
Shafique, Muhammad ;
Hara-Azumi, Yuko .
IEEE ACCESS, 2020, 8 :165319-165334
[4]   True North: Design and Tool Flow of a 65 mW 1 Million Neuron Programmable Neurosynaptic Chip [J].
Akopyan, Filipp ;
Sawada, Jun ;
Cassidy, Andrew ;
Alvarez-Icaza, Rodrigo ;
Arthur, John ;
Merolla, Paul ;
Imam, Nabil ;
Nakamura, Yutaka ;
Datta, Pallab ;
Nam, Gi-Joon ;
Taba, Brian ;
Beakes, Michael ;
Brezzo, Bernard ;
Kuang, Jente B. ;
Manohar, Rajit ;
Risk, William P. ;
Jackson, Bryan ;
Modha, Dharmendra S. .
IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2015, 34 (10) :1537-1557
[5]  
[Anonymous], 2021, INTEL OPTIMIZED MATH
[6]  
[Anonymous], 2019, PR MACH LEARN RES
[7]  
[Anonymous], 2016, INT C LEARN REPR ICL
[8]  
[Anonymous], 2021, NVIDIA deep learning accelerator (NVDLA)
[9]  
[Anonymous], 2021, MAESTRO OPEN SOURCE
[10]  
[Anonymous], 2021, INTEL MOVIDIUS VISIO