UAV target following in complex occluded environments with adaptive multi-modal fusion

被引:4
作者
Xu, Lele [1 ,2 ]
Wang, Teng [1 ]
Cai, Wenzhe [1 ]
Sun, Changyin [1 ,2 ]
机构
[1] Southeast Univ, Dept Automat, Nanjing 210018, Peoples R China
[2] Peng Cheng Lab, Shenzhen 518055, Peoples R China
基金
中国博士后科学基金; 中国国家自然科学基金;
关键词
UAV target following; Multi-modal fusion; Deep reinforcement learning; Complex occluded environments; TRACKING; NAVIGATION;
D O I
10.1007/s10489-022-04317-2
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Nowadays, deep reinforcement learning (DRL) has made remarkable achievements in unmanned aerial vehicle (UAV) target following. However, current DRL-based solutions only exploit appearance features to recognize and follow the target, and thus suffer from loss of the target in complex occluded environments. To that end, a novel target following solution based on adaptive appearance-motion feature fusion is proposed. First, we follow prior works to exploit convolutional neural network to extract appearance features of the target from the observation image captured from a down-looking camera. Meanwhile, we innovatively leverage action sequences of UAV to explicitly encode the motion features of the target. An attention module is subsequently introduced to adaptively select relevant useful features which serve as environment states and are fed into the decision module to produce the motion action of UAV. The whole network is trained using Deep Q-Network to learn the motion policy from observation in an end-to-end manner. We perform simulation experiments on Virtual Robot Experimentation Platform. Extensive experimental results demonstrate that: (1) Our proposed method achieves higher tracking accuracy and longer tracking time in various environments compared to state-of-the-art approaches; (2) The learned DRL policy could be well generalized to unseen environments.
引用
收藏
页码:16998 / 17014
页数:17
相关论文
共 43 条
[1]   Autonomous Navigation in Complex Environments with Deep Multimodal Fusion Network [J].
Anh Nguyen ;
Ngoc Nguyen ;
Kim Tran ;
Tjiputra, Erman ;
Tran, Quang D. .
2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, :5824-5830
[2]  
Ates U., 2020, 2020 INNOVATIONS INT, P1
[3]  
Bhagat S, 2020, INT CONF UNMAN AIRCR, P694, DOI [10.1109/icuas48674.2020.9213856, 10.1109/ICUAS48674.2020.9213856]
[4]  
Chen J, 2016, 2016 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS 2016), P446, DOI 10.1109/IROS.2016.7759092
[5]   The effects of challenge and threat states on coping flexibility: evidence from framing and exemplar priming [J].
Chen, Luxi ;
Qu, Li .
ANXIETY STRESS AND COPING, 2023, 36 (02) :163-183
[6]   UAV-UGV Multi-robot System for Warehouse Inventory: Scheduling Issues [J].
Guinand, Frederic ;
Guerin, Francois ;
Petitprez, Etienne .
COMPUTATIONAL COLLECTIVE INTELLIGENCE (ICCCI 2021), 2021, 12876 :241-254
[7]   Autonomous navigation and sensorless obstacle avoidance for UGV with environment information from UAV [J].
Harikumar, K. ;
Bera, Titas ;
Bardhan, Rajarshi ;
Sundaram, Suresh .
2018 SECOND IEEE INTERNATIONAL CONFERENCE ON ROBOTIC COMPUTING (IRC), 2018, :266-269
[8]   High-Speed Tracking with Kernelized Correlation Filters [J].
Henriques, Joao F. ;
Caseiro, Rui ;
Martins, Pedro ;
Batista, Jorge .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2015, 37 (03) :583-596
[9]   Deep Reinforcement Learning for UAV Navigation Through Massive MIMO Technique [J].
Huang, Hongji ;
Yang, Yuchun ;
Wang, Hong ;
Ding, Zhiguo ;
Sari, Hikmet ;
Adachi, Fumiyuki .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2020, 69 (01) :1117-1121
[10]   Learning Aberrance Repressed Correlation Filters for Real-Time UAV Tracking [J].
Huang, Ziyuan ;
Fu, Changhong ;
Li, Yiming ;
Lin, Fuling ;
Lu, Peng .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :2891-2900