Deep learning for unambiguous pose estimation of a non-cooperative fixed-wing UAV

被引:0
作者
Herrera, Leonardo [1 ]
Kim, Jae Jun [1 ]
Agrawal, Brij N. [1 ]
机构
[1] Naval Postgrad Sch, Dept Mech & Aerosp Engn, 1 Univ Circle, Monterey, CA 93943 USA
关键词
Deep learning; Pose estimation; UAV; Radar cues; Pose ambiguity; alpha beta filter;
D O I
10.1007/s00138-024-01630-3
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Image-based unmanned aerial vehicle (UAV) pose estimation is carried out by deep learning (DL) to automate the UAV aim-point selection for a laser beam control research testbed (LBCRT). DL models are proposed to estimate the UAV pose from available UAV target images and serve as a new sensor within the testbed. Such models are designed to address the problem of pose ambiguity, which arises from different 3D UAV poses looking similar in a 2D image plane. The models are trained under three datasets: synthetic data, synthetic data combined with real-world data, and real-world data. Afterward, the models are compared based on their ability to infer a real-world UAV trajectory in order to analyze the impact of using synthetic data during training. Quantitative results show that the proposed DL models solve the pose ambiguity problem when trained with appropriate data. In average, the models trained on real-world data had the lowest mean angle error, followed by those trained on combined data, and those trained solely on synthetic data. This shows that synthetic data should be carefully selected to bridge the gap when real-world data is unavailable or scarce. Customized real-world and synthetic data were created for the research. Real-world data are short wave infra-red (SWIR) images of a 3D printed UAV model with varying poses and associated pose labels. In addition, synthetic data are UAV images with varying poses and associated labels created by simulation. For this research, 77,077 real-world and 100,000 synthetic data were created.
引用
收藏
页数:11
相关论文
共 23 条
[1]  
Brändle M, 2014, IEEE ASME INT C ADV, P676, DOI 10.1109/AIM.2014.6878157
[2]   EPro-PnP: Generalized End-to-End Probabilistic Perspective-n-Points for Monocular Object Pose Estimation [J].
Chen, Hansheng ;
Wang, Pichao ;
Wang, Fan ;
Tian, Wei ;
Xiong, Lu ;
Li, Hao .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, :2771-2780
[3]  
Farrell J.A., 2008, AIDED NAVIGATION GPS
[4]   The Aircraft Pose Estimation Based on a Convolutional Neural Network [J].
Fu, Daoyong ;
Li, Wei ;
Han, Songchen ;
Zhang, Xinyan ;
Zhan, Zhaohuan ;
Yang, Menglong .
MATHEMATICAL PROBLEMS IN ENGINEERING, 2019, 2019
[5]   Deep Learning-Based Pose Estimation of Apples for Inspection in Logistic Centers Using Single-Perspective Imaging [J].
Giefer, Lino Antoni ;
Castellanos, Juan Daniel Arango ;
Babr, Mohammad Mohammadzadeh ;
Freitag, Michael .
PROCESSES, 2019, 7 (07)
[6]  
Grewal M. S., 2007, GLOBAL POSITIONING S
[7]  
He Kaiming., P IEEE INT C COMPUTE, P2961
[8]   Deep Learning Based Object Attitude Estimation for a Laser Beam Control Research Testbed [J].
Herrera, Leonardo ;
Jun, Kim Jae ;
Baker, Jeffrey ;
Agrawal, Brij N. .
APPLIED ARTIFICIAL INTELLIGENCE, 2023, 37 (01)
[9]   Deep Keypoint-Based Camera Pose Estimation with Geometric Constraints [J].
Jau, You-Yi ;
Zffu, Rui ;
Su, Hao ;
Chandraker, Manmohan .
2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, :4950-4957
[10]   PoseNet: A Convolutional Network for Real-Time 6-DOF Camera Relocalization [J].
Kendall, Alex ;
Grimes, Matthew ;
Cipolla, Roberto .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :2938-2946