aHCQ: Adaptive Hierarchical Clustering Based Quantization Framework for Deep Neural Networks

被引:0
|
作者
Hu, Jiaxin [1 ]
Rao, Weixiong [1 ]
Zhao, Qinpei [1 ]
机构
[1] Tongji Univ, Sch Software Engn, Shanghai, Peoples R China
来源
ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2021, PT II | 2021年 / 12713卷
基金
中国国家自然科学基金; 上海市自然科学基金;
关键词
Deep neural network; Hierarchical clustering; Network quantization; Compression rate;
D O I
10.1007/978-3-030-75765-6_17
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
For deep neural networks (DNNs), a high model accuracy is usually the main focus. However, millions of model parameters commonly lead to high space overheads, especially parameter redundancy. By maintaining network weights with less bit-widths, network quantization has been used to compress DNNs for lower space costs. However, existing quantization methods cannot well optimally balance the model size and the accuracy, thus they suffer from the accuracy loss more or less. Besides, though few of existing quantization techniques can adaptively determine layers quantization bit-widths, they either give little consideration on the relations of different DNN layers, or are designed for special hardware environment that are not universal in broad computer fields. To overcome these issues, we propose an adaptive Hierarchical Clustering based Quantization (aHCQ) framework. The aHCQ can find a largely compressed model from the quantization of each layer and take only little loss on the model accuracy. It is shown from the experiments that the aHCQ can achieve 11.4x and 8.2x model compression rates with only around 0.5% drop of the model accuracy.
引用
收藏
页码:207 / 218
页数:12
相关论文
共 50 条
  • [21] SHIP TRAJECTORY CLUSTERING METHOD BASED ON DEEP NEURAL NETWORKS
    Cup, Ying
    Xiong, Lian
    Liao, Hongzhou
    Dai, Xiang
    Gao, Xiang
    Chen, Huaixin
    UNIVERSITY POLITEHNICA OF BUCHAREST SCIENTIFIC BULLETIN SERIES C-ELECTRICAL ENGINEERING AND COMPUTER SCIENCE, 2022, 84 (03): : 71 - 84
  • [22] SHIP TRAJECTORY CLUSTERING METHOD BASED ON DEEP NEURAL NETWORKS
    Cui, Ying
    Xiong, Lian
    Liao, Hongzhou
    Dai, Xiang
    Gao, Xiang
    Chen, Huaixin
    UPB Scientific Bulletin, Series C: Electrical Engineering and Computer Science, 2022, 84 (03): : 71 - 84
  • [23] Hierarchical nucleation in deep neural networks
    Doimo, Diego
    Glielmo, Aldo
    Ansuini, Alessio
    Laio, Alessandro
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [24] A cellular neural network for clustering-based adaptive quantization in subband video compression
    Chen, CW
    Chen, LL
    Luo, JB
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 1996, 6 (06) : 688 - 692
  • [25] Cellular neural network for clustering-based adaptive quantization in subband video compression
    Univ of Rochester, Rochester, United States
    IEEE Trans Circuits Syst Video Technol, 6 (688-692):
  • [26] Image Compression Based on Hierarchical Clustering Vector Quantization
    Wang, Shi
    Ye, Long
    Zhong, Wei
    Zhang, Qin
    MULTIMEDIA AND SIGNAL PROCESSING, 2012, 346 : 120 - 128
  • [27] Online Adaptive Hierarchical Clustering in a Decision Tree Framework
    Basak, Jayanta
    JOURNAL OF PATTERN RECOGNITION RESEARCH, 2011, 6 (02): : 201 - 229
  • [28] ADAPTIVE LAYERWISE QUANTIZATION FOR DEEP NEURAL NETWORK COMPRESSION
    Zhu, Xiaotian
    Zhou, Wengang
    Li, Houqiang
    2018 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2018,
  • [29] Neuroevolution with box mutation: An adaptive and modular framework for evolving deep neural networks
    Santos, Frederico J. J. B.
    Goncalves, Ivo
    Castelli, Mauro
    APPLIED SOFT COMPUTING, 2023, 147
  • [30] Skipout: An Adaptive Layer-Level Regularization Framework for Deep Neural Networks
    Moayed, Hojjat
    Mansoori, Eghbal G.
    IEEE ACCESS, 2022, 10 : 62391 - 62401