Fast and High-Quality Monocular Depth Estimation with Optical Flow for Autonomous Drones

被引:2
作者
Shimada, Tomoyasu [1 ]
Nishikawa, Hiroki [2 ]
Kong, Xiangbo [1 ]
Tomiyama, Hiroyuki [1 ]
机构
[1] Ritsumeikan Univ, Grad Sch Sci & Engn, Kusatsu 5258577, Japan
[2] Osaka Univ, Grad Sch Informat Sci & Technol, Suita 5650871, Japan
基金
日本学术振兴会;
关键词
depth estimation; optical flow attention; perceptual discriminator;
D O I
10.3390/drones7020134
中图分类号
TP7 [遥感技术];
学科分类号
081102 ; 0816 ; 081602 ; 083002 ; 1404 ;
摘要
Recent years, autonomous drones have attracted attention in many fields due to their convenience. Autonomous drones require precise depth information so as to avoid collision to fly fast and both of RGB image and LiDAR point cloud are often employed in applications based on Convolutional Neural Networks (CNNs) to estimate the distance to obstacles. Such applications are implemented onboard embedded systems. In order to precisely estimate the depth, such CNN models are in general so complex to extract many features that the computational complexity increases, requiring long inference time. In order to solve the issue, we employ optical flow to aid in-depth estimation. In addition, we propose a new attention structure that makes maximum use of optical flow without complicating the network. Furthermore, we achieve improved performance without modifying the depth estimator by adding a perceptual discriminator in training. The proposed model is evaluated through accuracy, error, and inference time on the KITTI dataset. In the experiments, we have demonstrated the proposed method achieves better performance by up to 34% accuracy, 55% error reduction and 66% faster inference time on Jetson nano compared to previous methods. The proposed method is also evaluated through a collision avoidance in simulated drone flight and achieves the lowest collision rate of all estimation methods. These experimental results show the potential of proposed method to be used in real-world autonomous drone flight applications.
引用
收藏
页数:17
相关论文
共 47 条
[1]   Face Depth Estimation With Conditional Generative Adversarial Networks [J].
Arslan, Abdullah Taha ;
Seke, Erol .
IEEE ACCESS, 2019, 7 :23222-23231
[2]  
Baby Alba Terese, 2020, 2020 Advanced Computing and Communication Technologies for High Performance Applications (ACCTHPA), P125, DOI 10.1109/ACCTHPA49271.2020.9213233
[3]   AdaBins: Depth Estimation Using Adaptive Bins [J].
Bhat, Shariq Farooq ;
Alhashim, Ibraheem ;
Wonka, Peter .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :4008-4017
[4]  
Eigen D, 2014, ADV NEUR IN, V27
[5]   Predicting Depth, Surface Normals and Semantic Labels with a Common Multi-Scale Convolutional Architecture [J].
Eigen, David ;
Fergus, Rob .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :2650-2658
[6]   Two-frame motion estimation based on polynomial expansion [J].
Farnebäck, G .
IMAGE ANALYSIS, PROCEEDINGS, 2003, 2749 :363-370
[7]  
Figetakis E., 2021, P IEEE INT C COMM WO, P1
[8]   Survey on UAV Cellular Communications: Practical Aspects, Standardization Advancements, Regulation, and Security Challenges [J].
Fotouhi, Azade ;
Qiang, Haoran ;
Ding, Ming ;
Hassan, Mahbub ;
Giordano, Lorenzo Galati ;
Garcia-Rodriguez, Adrian ;
Yuan, Jinhong .
IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2019, 21 (04) :3417-3442
[9]   A Review on IoT Deep Learning UAV Systems for Autonomous Obstacle Detection and Collision Avoidance [J].
Fraga-Lamas, Paula ;
Ramos, Lucia ;
Mondejar-Guerra, Victor ;
Fernandez-Carames, Tiago M. .
REMOTE SENSING, 2019, 11 (18)
[10]  
Fuhrman T, 2019, IEEE INT C INT ROBOT, P6052, DOI [10.1109/IROS40897.2019.8967587, 10.1109/iros40897.2019.8967587]