Compressing Deep Neural Networks for Recognizing Places

被引:0
|
作者
Saha, Soham [1 ]
Varma, Girish [1 ]
Jawahar, C. V. [1 ]
机构
[1] Int Inst Informat Technol, KCIS, CVIT, Hyderabad, India
来源
PROCEEDINGS 2017 4TH IAPR ASIAN CONFERENCE ON PATTERN RECOGNITION (ACPR) | 2017年
关键词
Visual Place Recognition; Model Compression; Image Retrieval;
D O I
10.1109/ACPR.2017.154
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Visual place recognition on low memory devices such as mobile phones and robotics systems is a challenging problem. The state of the art models for this task uses deep learning architectures having close to 100 million parameters which takes over 400MB of memory. This makes these models infeasible to be deployed in low memory devices and gives rise to the need of compressing them. Hence we study the effectiveness of model compression techniques like trained quantization and pruning for reducing the number of parameters on one of the best performing image retrieval models called NetVLAD. We show that a compressed network can be created by starting with a model pre-trained for the task of visual place recognition and then fine-tuning it via trained pruning and quantization. The compressed model is able to produce the same mAP as the original uncompressed network. We achieve almost 50% parameter pruning with no loss in mAP and 70% pruning with close to 2% mAP reduction, while also performing 8-bit quantization. Furthermore, together with 5-bit quantization, we perform about 50% parameter reduction by pruning and get only about 3% reduction in mAP. The resulting compressed networks have sizes of around 30MB and 65MB which makes them easily usable in memory constrained devices.
引用
收藏
页码:352 / 357
页数:6
相关论文
共 50 条
  • [1] COMPRESSING DEEP NEURAL NETWORKS FOR EFFICIENT SPEECH ENHANCEMENT
    Tan, Ke
    Wang, DeLiang
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 8358 - 8362
  • [2] COMPRESSING DEEP NEURAL NETWORKS FOR EFFICIENT VISUAL INFERENCE
    Ge, Shiming
    Luo, Zhao
    Zhao, Shengwei
    Jin, Xin
    Zhang, Xiao-Yu
    2017 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2017, : 667 - 672
  • [3] Accelerating and Compressing Deep Neural Networks for Massive MIMO CSI Feedback
    Erak, Omar
    Abou-Zeid, Hatem
    ICC 2023-IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS, 2023, : 1029 - 1035
  • [4] Improving shallow neural network by compressing deep neural network
    Carvalho, Marcus
    Pratama, Mahardhika
    2018 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI), 2018, : 1382 - 1387
  • [5] Compressing neural networks via formal methods
    Ressi, Dalila
    Romanello, Riccardo
    Rossi, Sabina
    Piazza, Carla
    NEURAL NETWORKS, 2024, 178
  • [6] Compressing Convolutional Neural Networks in the Frequency Domain
    Chen, Wenlin
    Wilson, James
    Tyree, Stephen
    Weinberger, Kilian Q.
    Chen, Yixin
    KDD'16: PROCEEDINGS OF THE 22ND ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2016, : 1475 - 1484
  • [7] Compressing neural networks with two-layer decoupling
    De Jonghe, Joppe
    Usevich, Konstantin
    Dreesen, Philippe
    Ishteva, Mariya
    2023 IEEE 9TH INTERNATIONAL WORKSHOP ON COMPUTATIONAL ADVANCES IN MULTI-SENSOR ADAPTIVE PROCESSING, CAMSAP, 2023, : 226 - 230
  • [8] Compressing Neural Networks With Inter Prediction and Linear Transformation
    Lee, Kang-Ho
    Bae, Sung-Ho
    IEEE ACCESS, 2021, 9 : 69601 - 69608
  • [9] A Generalist Reinforcement Learning Agent for Compressing Convolutional Neural Networks
    Gonzalez-Sahagun, Gabriel
    Conant-Pablos, Santiago Enrique
    Ortiz-Bayliss, Jose Carlos
    Cruz-Duarte, Jorge M.
    IEEE ACCESS, 2024, 12 : 51100 - 51114
  • [10] Compressing Deep Neural Network: A Black-Box System Identification Approach
    Sahu, Ishan
    Pal, Arpan
    Ukil, Arijit
    Majumdar, Angshul
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,