共 1 条
Rapid Design Space Exploration of Near-Optimal Memory-Reduced DCNN Architecture using Multiple Model Compression Techniques
被引:1
|作者:
Byun, Younghoon
[1
]
Lee, Youngjoo
[1
]
机构:
[1] POSTECH, Dept Elect Engn, Pohang, South Korea
来源:
2021 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS)
|
2021年
关键词:
neural networks;
pruning;
channel scaling;
energy optimization;
design space exploration;
D O I:
10.1109/ISCAS51556.2021.9401489
中图分类号:
TM [电工技术];
TN [电子技术、通信技术];
学科分类号:
0808 ;
0809 ;
摘要:
In spite of the attractive accuracy, it is hard to use a deep convolutional neural network (DCNN) directly at the resource-limited devices due to the energy-consuming memory overheads, and thus the aggressive compression schemes are essentially utilized in practice to reduce the DCNN model size. As the recent methods have been individually developed, however, it is inevitable to exhaustively find the optimal combination of different approaches, requiring an enormous amount of search time. Given the complex baseline network, in this work, we introduce a rapid and systematic way to find the near-optimal memory-reduced DCNN option using multiple compression schemes together. We first precisely observe the accuracy-size trade-off of each method and make a novel interpolating scheme to speculate the accuracy of an arbitrary combination. We then present an iterative search algorithm to minimize the number of network evaluations for finding the memory-efficient DCNN structure satisfying the required accuracy. Experimental results reveal that our framework provides a similar compression level to the naive full-search strategy with three popular optimization methods while saving the search time by 7.35 times.
引用
收藏
页数:5
相关论文