NICE: Noise Injection and Clamping Estimation for Neural Network Quantization

被引:5
|
作者
Baskin, Chaim [1 ]
Zheltonozhkii, Evgenii [1 ]
Rozen, Tal [2 ]
Liss, Natan [2 ]
Chai, Yoav [3 ]
Schwartz, Eli [3 ]
Giryes, Raja [3 ]
Bronstein, Alexander M. [1 ]
Mendelson, Avi [1 ]
机构
[1] Technion, Dept Comp Sci, IL-3200003 Haifa, Israel
[2] Technion, Dept Elect Engn, IL-3200003 Haifa, Israel
[3] Tel Aviv Univ, Sch Elect Engn, IL-6997801 Tel Aviv, Israel
关键词
neural networks; low power; quantization; CNN architecture;
D O I
10.3390/math9172144
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Convolutional Neural Networks (CNNs) are very popular in many fields including computer vision, speech recognition, natural language processing, etc. Though deep learning leads to groundbreaking performance in those domains, the networks used are very computationally demanding and are far from being able to perform in real-time applications even on a GPU, which is not power efficient and therefore does not suit low power systems such as mobile devices. To overcome this challenge, some solutions have been proposed for quantizing the weights and activations of these networks, which accelerate the runtime significantly. Yet, this acceleration comes at the cost of a larger error unless spatial adjustments are carried out. The method proposed in this work trains quantized neural networks by noise injection and a learned clamping, which improve accuracy. This leads to state-of-the-art results on various regression and classification tasks, e.g., ImageNet classification with architectures such as ResNet-18/34/50 with as low as 3 bit weights and activations. We implement the proposed solution on an FPGA to demonstrate its applicability for low-power real-time applications. The quantization code will become publicly available upon acceptance.
引用
收藏
页数:12
相关论文
共 50 条
  • [31] Initialization Seeds Facilitating Neural Network Quantization
    Fei, Wen
    Ding, Li
    Pan, Yu
    Dai, Wenrui
    Li, Chenglin
    Zou, Junni
    Xiong, Hongkai
    2024 DATA COMPRESSION CONFERENCE, DCC, 2024, : 554 - 554
  • [32] Image Noise Level Estimation for Rice Noise Based On Extended ELM Neural Network Training Algorithm
    Yang, Xiaohui
    Xu, Kaiwei
    Xu, Shaoping
    Liu, Peter Xiaoping
    IEEE ACCESS, 2019, 7 : 1943 - 1951
  • [33] Noise Level Estimation Algorithm Using Convolutional Neural Network-Based Noise Separation Model
    Xu S.
    Liu T.
    Li C.
    Tang Y.
    Hu L.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2019, 56 (05): : 1060 - 1070
  • [34] Neural network quantization in federated learning at the edge
    Tonellotto, Nicola
    Gotta, Alberto
    Nardini, Franco Maria
    Gadler, Daniele
    Silvestri, Fabrizio
    INFORMATION SCIENCES, 2021, 575 : 417 - 436
  • [35] Convolutional Neural Network Accelerator with Vector Quantization
    Lee, Heng
    Wu, Yi-Heng
    Lin, Yu-Sheng
    Chien, Shao-Yi
    2019 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2019,
  • [36] A novel fault diagnosis method based on convolutional neural network with adaptive noise injection
    Xiao, Lei
    Wang, Jun
    Liu, Ximing
    Sun, Huanan
    Zhao, Hailong
    MEASUREMENT SCIENCE AND TECHNOLOGY, 2025, 36 (03)
  • [37] An adaptive vector quantization based on neural network
    Qiu, BS
    Qi, JQ
    An, P
    Zhang, DC
    ICSP '96 - 1996 3RD INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING, PROCEEDINGS, VOLS I AND II, 1996, : 1413 - 1416
  • [38] Advances in the Neural Network Quantization: A Comprehensive Review
    Wei, Lu
    Ma, Zhong
    Yang, Chaojie
    Yao, Qin
    APPLIED SCIENCES-BASEL, 2024, 14 (17):
  • [39] ACCOUNTING FOR QUANTIZATION NOISE IN ONLINE CORRELATION NOISE ESTIMATION FOR DISTRIBUTED VIDEO CODING
    Slowack, Jurgen
    Mys, Stefaan
    Skorupa, Jozef
    Lambert, Peter
    Van de Walle, Rik
    Grecos, Christos
    PCS: 2009 PICTURE CODING SYMPOSIUM, 2009, : 37 - +
  • [40] QUANTIZATION AND CIRCUIT NOISE EVALUATION FOR DIGITAL TELEPHONY NETWORK
    KITAWAKI, N
    ITOH, K
    KAKEHI, K
    REVIEW OF THE ELECTRICAL COMMUNICATIONS LABORATORIES, 1983, 31 (05): : 596 - 602