RCVNet: A bird damage identification network for power towers based on fusion of RF images and visual images

被引:7
作者
Gao, Wei [1 ,2 ]
Wu, Yangming [1 ]
Hong, Cui [1 ]
Wai, Rong-Jong [3 ]
Fan, Cheng-Tao [1 ]
机构
[1] Fuzhou Univ, Coll Elect Engn & Automat, Fuzhou 350108, Fujian, Peoples R China
[2] Fuzhou Univ, Dept Elect Engn, Zhicheng Coll, Fuzhou 350002, Fujian, Peoples R China
[3] Natl Taiwan Univ Sci & Technol, Dept Elect & Comp Engn, Taipei City 10607, Taiwan
关键词
Target recognition; Visual image; RF image; Sensor fusion; Deep convolutional neural network; OBSTACLE DETECTION; RADAR; DATASET; CAMERA;
D O I
10.1016/j.aei.2023.102104
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The technology for identifying birds around power towers using cameras alone is still susceptible to environmental interference. This paper proposes a new bird damage recognition network, RCVNet, which addresses this issue by fusing radio-frequency (RF) images and visual images. The network employs a feature layer fusion approach that accurately identifies bird damages in the monitoring area. Initially, RCVNet takes a group of RF and visual images as input. Then, through a series of convolutional neural networks (CNNs), birds are identified and located. To overcome challenges in recognizing small targets, several improved modules such as crosssupervised fusion network (CSF-net), posture deformable convolution (PDF), small-target attention fusion mechanism (SAFM), and Tiny-YOLOHead are introduced throughout RCVNet, improving surface information utilization and small feature retention rates. Finally, a bird damage discrimination strategy is developed based on the recognition outcomes of birds. As there is currently no public dataset available for RCVNet training, a new bird dataset called CRB2022, which includes RF and visual images, was gathered. Through large-scale experiments utilizing these methods, RCVNet effectively identifies birds, achieving a mean average precision of 79.34% and a mean average recall of 83.29%. Additionally, the discrimination rate of the utilized strategy can reach up to 98%.
引用
收藏
页数:15
相关论文
共 49 条
[1]   Insect detection from imagery using YOLOv3-based adaptive feature fusion convolution network [J].
Amrani, Abderraouf ;
Sohel, Ferdous ;
Diepeveen, Dean ;
Murray, David ;
Jones, Michael G. K. .
CROP & PASTURE SCIENCE, 2023, 74 (06) :615-627
[2]  
Barnes D, 2020, IEEE INT CONF ROBOT, P6433, DOI [10.1109/ICRA40945.2020.9196884, 10.1109/icra40945.2020.9196884]
[3]   Deep learning-based method for vision-guided robotic grasping of unknown objects [J].
Bergamini, Luca ;
Sposato, Mario ;
Pellicciari, Marcello ;
Peruzzini, Margherita ;
Calderara, Simone ;
Schmidt, Juliana .
ADVANCED ENGINEERING INFORMATICS, 2020, 44
[4]   nuScenes: A multimodal dataset for autonomous driving [J].
Caesar, Holger ;
Bankiti, Varun ;
Lang, Alex H. ;
Vora, Sourabh ;
Liong, Venice Erin ;
Xu, Qiang ;
Krishnan, Anush ;
Pan, Yu ;
Baldan, Giancarlo ;
Beijbom, Oscar .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :11618-11628
[5]   Object detection using depth completion and camera-LiDAR fusion for autonomous driving [J].
Carranza-Garcia, Manuel ;
Javier Galan-Sales, F. ;
Maria Luna-Romera, Jose ;
Riquelme, Jose C. .
INTEGRATED COMPUTER-AIDED ENGINEERING, 2022, 29 (03) :241-258
[6]   Spatial Attention Fusion for Obstacle Detection Using MmWave Radar and Vision Sensor [J].
Chang, Shuo ;
Zhang, Yifan ;
Zhang, Fan ;
Zhao, Xiaotong ;
Huang, Sai ;
Feng, Zhiyong ;
Wei, Zhiqing .
SENSORS, 2020, 20 (04)
[7]   Deformable Convolutional Networks [J].
Dai, Jifeng ;
Qi, Haozhi ;
Xiong, Yuwen ;
Li, Yi ;
Zhang, Guodong ;
Hu, Han ;
Wei, Yichen .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :764-773
[8]  
Ge Z, 2021, Arxiv, DOI [arXiv:2107.08430, DOI 10.48550/ARXIV.2107.08430]
[9]   Vision meets robotics: The KITTI dataset [J].
Geiger, A. ;
Lenz, P. ;
Stiller, C. ;
Urtasun, R. .
INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2013, 32 (11) :1231-1237
[10]  
He KM, 2017, IEEE I CONF COMP VIS, P2980, DOI [10.1109/TPAMI.2018.2844175, 10.1109/ICCV.2017.322]