Latent Weight Quantization for Integerized Training of Deep Neural Networks

被引:0
|
作者
Fei, Wen [1 ]
Dai, Wenrui [2 ]
Zhang, Liang [3 ]
Zhang, Luoming [4 ]
Li, Chenglin [1 ]
Zou, Junni [2 ]
Xiong, Hongkai [1 ]
机构
[1] Shanghai Jiao Tong Univ, Dept Elect Engn, Shanghai 200240, Peoples R China
[2] Shanghai Jiao Tong Univ, Dept Comp Sci & Engn, Shanghai 200240, Peoples R China
[3] Donghua Univ, Sch Comp Sci & Technol, Shanghai 201620, Peoples R China
[4] Zhejiang Univ, Key Lab Biomed Engn, Minist Educ, Hangzhou 310027, Peoples R China
基金
中国国家自然科学基金;
关键词
Quantization (signal); Training; Perturbation methods; Memory management; Hardware; Trajectory; Random access memory; Graphics processing units; Computational modeling; Noise; Integerized training; deep neural network quantization; latent weight; dual quantizer; large language models;
D O I
10.1109/TPAMI.2025.3527498
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Existing methods for integerized training speed up deep learning by using low-bitwidth integerized weights, activations, gradients, and optimizer buffers. However, they overlook the issue of full-precision latent weights, which consume excessive memory to accumulate gradient-based updates for optimizing the integerized weights. In this paper, we propose the first latent weight quantization schema for general integerized training, which minimizes quantization perturbation to training process via residual quantization with optimized dual quantizer. We leverage residual quantization to eliminate the correlation between latent weight and integerized weight for suppressing quantization noise. We further propose dual quantizer with optimal nonuniform codebook to avoid frozen weight and ensure statistically unbiased training trajectory as full-precision latent weight. The codebook is optimized to minimize the disturbance on weight update under importance guidance and achieved with a three-segment polyline approximation for hardware-friendly implementation. Extensive experiments show that the proposed schema allows integerized training with lowest 4-bit latent weight for various architectures including ResNets, MobileNetV2, and Transformers, and yields negligible performance loss in image classification and text generation. Furthermore, we successfully fine-tune Large Language Models with up to 13 billion parameters on one single GPU using the proposed schema.
引用
收藏
页码:2816 / 2832
页数:17
相关论文
共 50 条
  • [31] The Hardware Impact of Quantization and Pruning for Weights in Spiking Neural Networks
    Schaefer, Clemens J. S.
    Taheri, Pooria
    Horeni, Mark
    Joshi, Siddharth
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2023, 70 (05) : 1789 - 1793
  • [32] Defenses Against Byzantine Attacks in Distributed Deep Neural Networks
    Xia, Qi
    Tao, Zeyi
    Li, Qun
    IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2021, 8 (03): : 2025 - 2035
  • [33] Evaluations on Deep Neural Networks Training Using Posit Number System
    Lu, Jinming
    Fang, Chao
    Xu, Mingyang
    Lin, Jun
    Wang, Zhongfeng
    IEEE TRANSACTIONS ON COMPUTERS, 2021, 70 (02) : 174 - 187
  • [34] A Gradient-Guided Evolutionary Approach to Training Deep Neural Networks
    Yang, Shangshang
    Tian, Ye
    He, Cheng
    Zhang, Xingyi
    Tan, Kay Chen
    Jin, Yaochu
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (09) : 4861 - 4875
  • [35] An Efficient Learning Algorithm for Direct Training Deep Spiking Neural Networks
    Zhu, Xiaolei
    Zhao, Baixin
    Ma, De
    Tang, Huajin
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2022, 14 (03) : 847 - 856
  • [36] General Bitwidth Assignment for Efficient Deep Convolutional Neural Network Quantization
    Fei, Wen
    Dai, Wenrui
    Li, Chenglin
    Zou, Junni
    Xiong, Hongkai
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (10) : 5253 - 5267
  • [37] Traffic Network Flow Prediction Using Parallel Training for Deep Convolutional Neural Networks on Spark Cloud
    Zhang, Yongnan
    Zhou, Yonghua
    Lu, Huapu
    Fujita, Hamido
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2020, 16 (12) : 7369 - 7380
  • [38] Multi-Valued Quantization Neural Networks toward Hardware Implementation
    Aratani, Yoshiya
    Jye, Yeoh Yoeng
    Suzuki, Akihiro
    Shuto, Daisuke
    Morie, Takashi
    Tamukoh, Hakaru
    ICAROB 2017: PROCEEDINGS OF THE 2017 INTERNATIONAL CONFERENCE ON ARTIFICIAL LIFE AND ROBOTICS, 2017, : P132 - P135
  • [39] Dynamic Memory Management for GPU-based training of Deep Neural Networks
    Shriram, S. B.
    Garg, Anshuj
    Kulkarni, Purushottam
    2019 IEEE 33RD INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM (IPDPS 2019), 2019, : 200 - 209
  • [40] A General Multiple Data Augmentation Based Framework for Training Deep Neural Networks
    Hu, Binyan
    Sun, Yu
    Qin, A. K.
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,