ATA: Attentional Non-Linear Activation Function Approximation for VLSI-Based Neural Networks

被引:5
作者
Wei, Linyu [1 ]
Cai, Jueping [1 ]
Wang, Wuzhuang [1 ]
机构
[1] Xidian Univ, State Key Lab Wide Bandgap Semicond Technol Disci, Xian 710071, Peoples R China
关键词
Hardware; Fitting; Table lookup; Approximation methods; Sensitivity; Feature extraction; Function approximation; Neural networks; activation function; attention mechanism; hardware implementation; SIGMOID FUNCTION; IMPLEMENTATION;
D O I
10.1109/LSP.2021.3067188
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In this letter, we present an attentional non-linear activation function approximation method called ATA for VLSI-based neural networks. Unlike other approximation methods that pursue the low hardware resources with a high recognition accuracy loss, the ATA utilizes the pixel attention to focus on the important features to keep the recognition accuracy and reduce resource cost. Specifically, attention applied in the activation function is realized by the approximated activation functions with different fitting errors for VLSI-based neural networks. The important features are highlighted by the piecewise linear function and improved look-up table with low fitting error, while the trivial features are ignored with the large fitting error. Experimental results demonstrate that the ATA outperforms other state-of-the-art approximation methods in recognition accuracy, power and area.
引用
收藏
页码:793 / 797
页数:5
相关论文
共 22 条
[1]   Low-error digital hardware implementation of artificial neuron activation functions and their derivative [J].
Armato, A. ;
Fanucci, L. ;
Scilingo, E. P. ;
De Rossi, D. .
MICROPROCESSORS AND MICROSYSTEMS, 2011, 35 (06) :557-567
[2]   Large-Margin Classification in Infinite Neural Networks [J].
Cho, Youngmin ;
Saul, Lawrence K. .
NEURAL COMPUTATION, 2010, 22 (10) :2678-2697
[3]  
Clevert D.-A., 2016, 4 INT C LEARN REPR I
[4]   Controlled accuracy approximation of sigmoid function for efficient FPGA-based implementation of artificial neurons [J].
del Campo, I. ;
Finker, R. ;
Echanobe, J. ;
Basterretxea, K. .
ELECTRONICS LETTERS, 2013, 49 (25) :1598-1600
[5]   Dual Attention Network for Scene Segmentation [J].
Fu, Jun ;
Liu, Jing ;
Tian, Haijie ;
Li, Yong ;
Bao, Yongjun ;
Fang, Zhiwei ;
Lu, Hanqing .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :3141-3149
[6]  
Im D., 2020, IEEE T CIRCUITS SY 1, V67, P20
[7]   Deep neural networks with a set of node-wise varying activation functions [J].
Jang, Jinhyeok ;
Cho, Hyunjoong ;
Kim, Jaehong ;
Lee, Jaeyeon ;
Yang, Seungjoon .
NEURAL NETWORKS, 2020, 126 :118-131
[8]  
Kim Y, 2014, IEEE ASIAN SOLID STA, P213, DOI 10.1109/ASSCC.2014.7008898
[9]   xUnit: Learning a Spatial Activation Function for Efficient Image Restoration [J].
Kligvasser, Idan ;
Shaham, Tamar Rott ;
Michaeli, Tomer .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :2433-2442
[10]   High Speed VLSI Implementation of the Hyperbolic Tangent Sigmoid Function [J].
Leboeuf, Karl ;
Namin, Ashkan Hosseinzadeh ;
Muscedere, Roberto ;
Wu, Huapeng ;
Ahmadi, Majid .
THIRD 2008 INTERNATIONAL CONFERENCE ON CONVERGENCE AND HYBRID INFORMATION TECHNOLOGY, VOL 1, PROCEEDINGS, 2008, :1070-1073