Compressing Deep Neural Networks for Recognizing Places

被引:0
|
作者
Saha, Soham [1 ]
Varma, Girish [1 ]
Jawahar, C. V. [1 ]
机构
[1] Int Inst Informat Technol, KCIS, CVIT, Hyderabad, India
来源
PROCEEDINGS 2017 4TH IAPR ASIAN CONFERENCE ON PATTERN RECOGNITION (ACPR) | 2017年
关键词
Visual Place Recognition; Model Compression; Image Retrieval;
D O I
10.1109/ACPR.2017.154
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Visual place recognition on low memory devices such as mobile phones and robotics systems is a challenging problem. The state of the art models for this task uses deep learning architectures having close to 100 million parameters which takes over 400MB of memory. This makes these models infeasible to be deployed in low memory devices and gives rise to the need of compressing them. Hence we study the effectiveness of model compression techniques like trained quantization and pruning for reducing the number of parameters on one of the best performing image retrieval models called NetVLAD. We show that a compressed network can be created by starting with a model pre-trained for the task of visual place recognition and then fine-tuning it via trained pruning and quantization. The compressed model is able to produce the same mAP as the original uncompressed network. We achieve almost 50% parameter pruning with no loss in mAP and 70% pruning with close to 2% mAP reduction, while also performing 8-bit quantization. Furthermore, together with 5-bit quantization, we perform about 50% parameter reduction by pruning and get only about 3% reduction in mAP. The resulting compressed networks have sizes of around 30MB and 65MB which makes them easily usable in memory constrained devices.
引用
收藏
页码:352 / 357
页数:6
相关论文
共 50 条
  • [31] Fast and Robust Compression of Deep Convolutional Neural Networks
    Wen, Jia
    Yang, Liu
    Shen, Chenyang
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2020, PT II, 2020, 12397 : 52 - 63
  • [32] Compressing Deep Image Super-resolution Models
    Jiang, Yuxuan
    Nawala, Jakub
    Zhang, Fan
    Bull, David
    2024 PICTURE CODING SYMPOSIUM, PCS 2024, 2024,
  • [33] Compacting Deep Neural Networks for Internet of Things: Methods and Applications
    Zhang, Ke
    Ying, Hanbo
    Dai, Hong-Ning
    Li, Lin
    Peng, Yuanyuan
    Guo, Keyi
    Yu, Hongfang
    IEEE INTERNET OF THINGS JOURNAL, 2021, 8 (15): : 11935 - 11959
  • [34] Filter Pruning via Feature Discrimination in Deep Neural Networks
    He, Zhiqiang
    Qian, Yaguan
    Wang, Yuqi
    Wang, Bin
    Guan, Xiaohui
    Gu, Zhaoquan
    Ling, Xiang
    Zeng, Shaoning
    Wang, Haijiang
    Zhou, Wujie
    COMPUTER VISION, ECCV 2022, PT XXI, 2022, 13681 : 245 - 261
  • [35] Training Integer-Only Deep Recurrent Neural Networks
    Nia V.P.
    Sari E.
    Courville V.
    Asgharian M.
    SN Computer Science, 4 (5)
  • [36] Vector Quantization of Deep Convolutional Neural Networks With Learned Codebook
    Yang, Siyuan
    Mao, Yongyi
    2022 17TH CANADIAN WORKSHOP ON INFORMATION THEORY (CWIT), 2022, : 39 - 44
  • [37] Soft Taylor Pruning for Accelerating Deep Convolutional Neural Networks
    Rong, Jintao
    Yu, Xiyi
    Zhang, Mingyang
    Ou, Linlin
    IECON 2020: THE 46TH ANNUAL CONFERENCE OF THE IEEE INDUSTRIAL ELECTRONICS SOCIETY, 2020, : 5343 - 5349
  • [38] Compressing Deep Models using Multi Tensor Train Decomposition
    Yang, Xin
    Sun, Weize
    Huang, Lei
    Chen, Shaowu
    ICCAIS 2019: THE 8TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND INFORMATION SCIENCES, 2019,
  • [39] Training deep neural networks for wireless sensor networks using loosely and weakly labeled images
    Zhou, Qianwei
    Chen, Yuhang
    Li, Baoqing
    Li, Xiaoxin
    Zhou, Chen
    Huang, Jingchang
    Hu, Haigen
    NEUROCOMPUTING, 2021, 427 : 64 - 73
  • [40] An Improved Hashing Method for Image Retrieval Based on Deep Neural Networks
    Chen, Qiu
    Wang, Weidong
    Lee, Feifei
    2018 INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING AND MACHINE LEARNING (SPML 2018), 2018, : 78 - 83