An Efficient Approach to Escalate the Speed of Training Convolution Neural Networks

被引:1
作者
Pabitha, P. [1 ]
Jayasimhan, Anusha [1 ]
机构
[1] Anna Univ, Madras Inst Technol Campus, Dept Comp Technol, Chennai 600044, India
关键词
CNN; deep learning; image classification; model compression;
D O I
10.23919/JCC.fa.2022-0639.202402
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
Deep neural networks excel at image identification and computer vision applications such as visual product search, facial recognition, medical image analysis, object detection, semantic segmentation, instance segmentation, and many others. In image and video recognition applications, convolutional neural networks (CNNs) are widely employed. These networks provide better performance but at a higher cost of computation. With the advent of big data, the growing scale of datasets has made processing and model training a time-consuming operation, resulting in longer training times. Moreover, these large scale datasets contain redundant data points that have minimum impact on the final outcome of the model. To address these issues, an accelerated CNN system is proposed for speeding up training by eliminating the noncritical data points during training alongwith a model compression method. Furthermore, the identification of the critical input data is performed by aggregating the data points at two levels of granularity which are used for evaluating the impact on the model output. Extensive experiments are conducted using the proposed method on CIFAR-10 dataset on ResNet models giving a 40% reduction in number of FLOPs with a degradation of just 0.11% accuracy.
引用
收藏
页码:258 / 269
页数:12
相关论文
共 29 条
  • [1] [Anonymous], 2013, The national security agency: Missions, authorities, oversight and partnerships
  • [2] [Anonymous], 2007, IDC IVIEW IDC ANAL F
  • [3] [Anonymous], 2009, P 26 ANN INT C MACHI, DOI [DOI 10.1145/1553374.1553380, 10.1145/1553374.155338]
  • [4] Birodkar V, 2019, Arxiv, DOI arXiv:1901.11409
  • [5] de Jorge P., 2021, INT C LEARNING REPRE
  • [6] More is Less: A More Complicated Network with Less Inference Complexity
    Dong, Xuanyi
    Huang, Junshi
    Yang, Yi
    Yan, Shuicheng
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 1895 - 1903
  • [7] Fan Y., 2017, P 5 INT C LEARNING R
  • [8] Howard AG, 2017, Arxiv, DOI arXiv:1704.04861
  • [9] Attention Based Pruning for Shift Networks
    Hacene, Ghouthi Boukli
    Lassance, Carlos
    Gripon, Vincent
    Courbariaux, Matthieu
    Bengio, Yoshua
    [J]. 2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 4054 - 4061
  • [10] Accelerating Deep Learning Systems via Critical Set Identification and Model Compression
    Han, Rui
    Liu, Chi Harold
    Li, Shilin
    Wen, Shilin
    Liu, Xue
    [J]. IEEE TRANSACTIONS ON COMPUTERS, 2020, 69 (07) : 1059 - 1070