Deep Learning Approaches Assessment for Underwater Scene Understanding and Egomotion Estimation

被引:1
作者
Teixeira, Bernardo [1 ,3 ]
Silva, Hugo [1 ]
Matos, Anibal [1 ,3 ]
Silva, Eduardo [1 ,2 ]
机构
[1] Porto Polytech Inst, INESC IBC Inst Syst & Comp Engn Technol & Sci, Porto, Portugal
[2] Porto Polytech Inst, ISEP Sch Engn, Porto, Portugal
[3] Univ Porto, FEUP Fac Engn, Porto, Portugal
来源
OCEANS 2019 MTS/IEEE SEATTLE | 2019年
关键词
Artificial intelligence; Computer vision; Deep learning; Visual Odometry; Robot navigation;
D O I
10.23919/oceans40490.2019.8962872
中图分类号
U6 [水路运输]; P75 [海洋工程];
学科分类号
0814 ; 081505 ; 0824 ; 082401 ;
摘要
This paper address the use of deep learning approaches for visual based navigation in confined underwater environments. State-of-the-art algorithms have shown the tremendous potential deep learning architectures can have for visual navigation implementations, though they are still mostly outperformed by classical feature-based techniques. In this work, we apply current state-of-the-art deep learning methods for visual-based robot navigation to the more challenging underwater environment, providing both an underwater visual dataset acquired in real operational mission scenarios and an assessment of state-of-the-art algorithms on the underwater context. We extend current work by proposing a novel pose optimization architecture for the purpose of correcting visual odometry estimate drift using a Visual-Inertial fusion network, consisted of a neural network architecture anchored on an Inertial supervision learning scheme. Our Visual-Inertial Fusion Network was shown to improve results an average of 50% for trajectory estimates, also producing more visually consistent trajectory estimates for both our underwater application scenarios.
引用
收藏
页数:9
相关论文
共 50 条
[41]   Deep learning for visual understanding: A review [J].
Guo, Yanming ;
Liu, Yu ;
Oerlemans, Ard ;
Lao, Songyang ;
Wu, Song ;
Lew, Michael S. .
NEUROCOMPUTING, 2016, 187 :27-48
[42]   Advancing Grapevine Variety Identification: A Systematic Review of Deep Learning and Machine Learning Approaches [J].
Carneiro, Gabriel A. ;
Cunha, Antonio ;
Aubry, Thierry J. ;
Sousa, Joaquim .
AGRIENGINEERING, 2024, 6 (04) :4851-4888
[43]   Classifying Motorcycle Rider Helmet on a Low Light Video Scene using Deep Learning [J].
Tomas, John Paul Q. ;
Doma, Bonifacio T. .
INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2024, 15 (02) :31-38
[44]   Comparing Human Pose Estimation through deep learning approaches: An overview [J].
Dibenedetto, Gaetano ;
Sotiropoulos, Stefanos ;
Polignano, Marco ;
Cavallo, Giuseppe ;
Lops, Pasquale .
COMPUTER VISION AND IMAGE UNDERSTANDING, 2025, 252
[45]   Deep Learning Based Channel Covariance Matrix Estimation With User Location and Scene Images [J].
Xu, Weihua ;
Gao, Feifei ;
Zhang, Jianhua ;
Tao, Xiaoming ;
Alkhateeb, Ahmed .
IEEE TRANSACTIONS ON COMMUNICATIONS, 2021, 69 (12) :8145-8158
[46]   A Study of Deep Learning Approaches and Loss Functions for Abundance Fractions Estimation [J].
Lodhi, Vaibhav ;
Biswas, Arindam ;
Chakravarty, Debashish ;
Mitra, Pabitra .
2021 11TH WORKSHOP ON HYPERSPECTRAL IMAGING AND SIGNAL PROCESSING: EVOLUTION IN REMOTE SENSING (WHISPERS), 2021,
[47]   Advancing Underwater Vision: A Survey of Deep Learning Models for Underwater Object Recognition and Tracking [J].
Elmezain, Mahmoud ;
Saad Saoud, Lyes ;
Sultan, Atif ;
Heshmat, Mohamed ;
Seneviratne, Lakmal ;
Hussain, Irfan .
IEEE ACCESS, 2025, 13 :17830-17867
[48]   Deep Learning Approaches for Understanding Adverse Drug Reaction: Short Literature Review [J].
Zyani, Chaimaa ;
Nfaoui, El Habib .
DIGITAL TECHNOLOGIES AND APPLICATIONS, ICDTA 2024, VOL 4, 2024, 1101 :536-544
[49]   Video scene analysis: an overview and challenges on deep learning algorithms [J].
Qaisar Abbas ;
Mostafa E. A. Ibrahim ;
M. Arfan Jaffar .
Multimedia Tools and Applications, 2018, 77 :20415-20453
[50]   Deep Learning for Scene Recognition from Visual Data: A Survey [J].
Matei, Alina ;
Glavan, Andreea ;
Talavera, Estefania .
HYBRID ARTIFICIAL INTELLIGENT SYSTEMS, HAIS 2020, 2020, 12344 :763-773