A Lightweight UAV Object Detection Algorithm Based on Iterative Sparse Training

被引:0
|
作者
Hou X. [1 ]
Qu G. [2 ]
Wei D. [2 ]
Zhang J. [3 ]
机构
[1] Institute of Computing Technology, Chinese Academy of Sciences, Beijing
[2] Chinese Aeronautical Radio Electronics Research Institute, Shanghai
[3] School of Information and Communication Engineering, Beijing University of Posts and Telecommunications, Beijing
来源
Jisuanji Yanjiu yu Fazhan/Computer Research and Development | 2022年 / 59卷 / 04期
基金
中国国家自然科学基金;
关键词
Data enhancement; Iterative sparse training; Low precision loss; Model compression; YOLOv3;
D O I
10.7544/issn1000-1239.20200986
中图分类号
学科分类号
摘要
With the maturity of UAV (unmanned aerial vehicle) technology, vehicles equipped with cameras are widely used in various fields, such as security and surveillance, aerial photography and infrastructure inspection. It is important to automatically and efficiently analyze and understand the visual data collected from vehicles. The object detection algorithm based on deep convolutional neural network has made amazing achievements in many practical applications, but it is often accompanied by great resource consumption and memory occupation. Thus, it is challenging to run deep convolutional neural networks directly on embedded devices with limited computing power carried by vehicles, which leads to high latency. In order to meet these challenges, a novel pruning algorithm based on iterative sparse training is proposed to improve the computational effectiveness of the classic object detection network YOLOv3 (you only look once). At the same time, different data enhancement methods and related optimization means are combined to ensure that the precision error of the detector before and after compression is within an acceptable range. Experimental results indicate that the pruning scheme based on iterative sparse training proposed in this paper achieves a considerable compression rate of YOLOv3 within slightly decline in precision. The original YOLOv3 model contains 61.57 MB weights and requires 139.77GFLOPS(floating-point operations). With 98.72% weights and 90.03% FLOPS reduced, our model still maintains a decent accuracy, with only 2.0% mAP(mean average precision) loss, which provides support for real-time application of UAV object detection. © 2022, Science Press. All right reserved.
引用
收藏
页码:882 / 893
页数:11
相关论文
共 48 条
  • [1] Bhaskaranand M, Gibson J D., Low-complexity video encoding for UAV reconnaissance and surveillance, Proc of IEEE 2011-MILCOM, pp. 1633-1638, (2011)
  • [2] Madawalagama S, Munasinghe N, Dampegama S, Et al., Low cost aerial mapping with consumer grade drones, Coordinates, 8, 4, pp. 13-18, (2016)
  • [3] Sa I, Hrabar S, Corke P., Outdoor flight testing of a pole inspection UAV incorporating high-speed vision, Proc of Field and Service Robotics, pp. 107-121, (2015)
  • [4] Tian Zhi, Shen Chunhua, Chen Hao, Et al., Fcos: Fully convolutional one-stage object detection, Proc of the IEEE Int Conf on Computer Vision, pp. 9627-9636, (2019)
  • [5] Shen Zhiqiang, Liu Zhuang, Li Jianguo, Et al., Dsod: Learning deeply supervised object detectors from scratch, Proc of the IEEE Int Conf on Computer Vision, pp. 1919-1927, (2017)
  • [6] He Kaiming, Gkioxari G, Dollar P, Et al., Mask R-CNN, Proc of the IEEE Int Conf on Computer Vision, pp. 2961-2969, (2017)
  • [7] Tan Mingxing, Pang Ruoming, Le Q V., Efficientdet: Scalable and efficient object detection, Proc of the IEEE Conf on Computer Vision and Pattern Recognition, pp. 10781-10790, (2020)
  • [8] Zhou Xingyi, Wang Dequan, Krahenbuhl P., Objects as points, (2019)
  • [9] Law H, Deng Jia, CornerNet: Detecting objects as paired keypoints, Proc of the European Conf on Computer Vision (ECCV), pp. 734-750, (2018)
  • [10] Duan Kaiwen, Bai Song, Xie Lingxi, Et al., CenterNet: Object detection with keypoint triplets, (2019)