Area and power efficient pipelined hybrid merged adders for customized deep learning framework for FPGA implementation

被引:9
作者
Kowsalya, T. [1 ]
机构
[1] Muthayammal Engn Coll, Dept Elect & Commun Engn, Namakkal, Tamil Nadu, India
关键词
Deep learning framework; Phmac; Gpu; Neural networks; Optimization;
D O I
10.1016/j.micpro.2019.102906
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
With the rapid growth of deep learning and neural network algorithms, various fields such as communication, Industrial automation, computer vision system and medical applications have seen the drastic improvements in recent years. However, deep learning and neural network models are increasing day by day, while model parameters are used for representing the models. Although the existing models use efficient GPU for accommodating these models, their implementation in the dedicated embedded devices needs more optimization which remains a real challenge for researchers. Thus paper, carries an investigation of deep learning frameworks, more particularly as review of adders implemented in the deep learning framework. A new pipelined hybrid merged adders (PHMAC) optimized for FPGA architecture which has more efficient in terms of area and power is presented. The proposed adders represent the integration of the principle of carry select and carry look ahead principle of adders in which LUT is re-used for the different inputs which consume less power and provide effective area utilization. The proposed adders were investigated on different FPGA architectures in which the power and area were analyzed. Comparison of the proposed adders with the other adders such as carry select adders (CSA), carry look ahead adder (CLA), Carry skip adders and Koggle Stone adders has been made and results have proved to be highly vital into a 50% reduction in the area, power and 45% when compared with above mentioned traditional adders. (C) 2019 Elsevier B.V. All rights reserved.
引用
收藏
页数:9
相关论文
共 12 条
[1]  
[Anonymous], 2015, P 2015 ACM SIGDA INT
[2]  
[Anonymous], [No title captured]
[3]  
[Anonymous], [No title captured]
[4]  
[Anonymous], [No title captured]
[5]  
[Anonymous], [No title captured]
[6]   Fast inference of deep neural networks in FPGAs for particle physics [J].
Duarte, J. ;
Han, S. ;
Harris, P. ;
Jindariani, S. ;
Kreinar, E. ;
Kreis, B. ;
Ngadiuba, J. ;
Pierini, M. ;
Rivera, R. ;
Tran, N. ;
Wu, Z. .
JOURNAL OF INSTRUMENTATION, 2018, 13
[7]   On Hard Adders and Carry Chains in FPGAs [J].
Luu, Jason ;
McCullough, Conor ;
Wang, Sen ;
Huda, Safeen ;
Yan, Bo ;
Chiasson, Charles ;
Kent, Kenneth B. ;
Anderson, Jason ;
Rose, Jonathan ;
Betz, Vaughn .
2014 IEEE 22ND ANNUAL INTERNATIONAL SYMPOSIUM ON FIELD-PROGRAMMABLE CUSTOM COMPUTING MACHINES (FCCM 2014), 2014, :52-59
[8]   High-Efficiency Convolutional Ternary Neural Networks with Custom Adder Trees and Weight Compression [J].
Prost-Boucle, Adrien ;
Bourge, Alban ;
Petrot, Frederic .
ACM TRANSACTIONS ON RECONFIGURABLE TECHNOLOGY AND SYSTEMS, 2018, 11 (03)
[9]  
Rahman A, 2017, DES AUT TEST EUROPE, P1147, DOI 10.23919/DATE.2017.7927162
[10]   Toolflows for Mapping Convolutional Neural Networks on FPGAs: A Survey and Future Directions [J].
Venieris, Stylianos I. ;
Kouris, Alexandros ;
Bouganis, Christos-Savvas .
ACM COMPUTING SURVEYS, 2018, 51 (03)