Compressing Deep Neural Networks for Recognizing Places

被引:0
|
作者
Saha, Soham [1 ]
Varma, Girish [1 ]
Jawahar, C. V. [1 ]
机构
[1] Int Inst Informat Technol, KCIS, CVIT, Hyderabad, India
来源
PROCEEDINGS 2017 4TH IAPR ASIAN CONFERENCE ON PATTERN RECOGNITION (ACPR) | 2017年
关键词
Visual Place Recognition; Model Compression; Image Retrieval;
D O I
10.1109/ACPR.2017.154
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Visual place recognition on low memory devices such as mobile phones and robotics systems is a challenging problem. The state of the art models for this task uses deep learning architectures having close to 100 million parameters which takes over 400MB of memory. This makes these models infeasible to be deployed in low memory devices and gives rise to the need of compressing them. Hence we study the effectiveness of model compression techniques like trained quantization and pruning for reducing the number of parameters on one of the best performing image retrieval models called NetVLAD. We show that a compressed network can be created by starting with a model pre-trained for the task of visual place recognition and then fine-tuning it via trained pruning and quantization. The compressed model is able to produce the same mAP as the original uncompressed network. We achieve almost 50% parameter pruning with no loss in mAP and 70% pruning with close to 2% mAP reduction, while also performing 8-bit quantization. Furthermore, together with 5-bit quantization, we perform about 50% parameter reduction by pruning and get only about 3% reduction in mAP. The resulting compressed networks have sizes of around 30MB and 65MB which makes them easily usable in memory constrained devices.
引用
收藏
页码:352 / 357
页数:6
相关论文
共 50 条
  • [41] SFP: Similarity-based filter pruning for deep neural networks
    Li, Guoqing
    Li, Rengang
    Li, Tuo
    Shen, Chaoyao
    Zou, Xiaofeng
    Wang, Jiuyang
    Wang, Changhong
    Li, Nanjun
    INFORMATION SCIENCES, 2025, 689
  • [42] Compression of Deep Neural Networks by combining pruning and low rank decomposition
    Goyal, Saurabh
    Choudhury, Anamitra Roy
    Sharma, Vivek
    2019 IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM WORKSHOPS (IPDPSW), 2019, : 952 - 958
  • [43] Studying the plasticity in deep convolutional neural networks using random pruning
    Mittal, Deepak
    Bhardwaj, Shweta
    Khapra, Mitesh M.
    Ravindran, Balaraman
    MACHINE VISION AND APPLICATIONS, 2019, 30 (02) : 203 - 216
  • [44] Acceleration of Deep Convolutional Neural Networks Using Adaptive Filter Pruning
    Singh, Pravendra
    Verma, Vinay Kumar
    Rai, Piyush
    Namboodiri, Vinay P.
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2020, 14 (04) : 838 - 847
  • [45] Deep compression of convolutional neural networks with low-rank approximation
    Astrid, Marcella
    Lee, Seung-Ik
    ETRI JOURNAL, 2018, 40 (04) : 421 - 434
  • [46] A Comprehensive Survey on Model Quantization for Deep Neural Networks in Image Classification
    Rokh, Babak
    Azarpeyvand, Ali
    Khanteymoori, Alireza
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2023, 14 (06)
  • [47] Studying the plasticity in deep convolutional neural networks using random pruning
    Deepak Mittal
    Shweta Bhardwaj
    Mitesh M. Khapra
    Balaraman Ravindran
    Machine Vision and Applications, 2019, 30 : 203 - 216
  • [48] Compressing recurrent neural network models through principal component analysis
    Qi, Haobo
    Cao, Jingxuan
    Chen, Shichong
    Zhou, Jing
    STATISTICS AND ITS INTERFACE, 2023, 16 (03) : 397 - 407
  • [49] Transforming Large-Size to Lightweight Deep Neural Networks for IoT Applications
    Mishra, Rahul
    Gupta, Hari
    ACM COMPUTING SURVEYS, 2023, 55 (11)
  • [50] Low-Complexity Deep Neural Networks for Image Object Classification and Detection
    Hsiao, Shen-Fu
    Zhan, Jing-Fu
    Lin, Chih-Chien
    2019 IEEE ASIA PACIFIC CONFERENCE ON CIRCUITS AND SYSTEMS (APCCAS 2019), 2019, : 313 - 316