Geometric-driven structure recovery from a single omnidirectional image based on planar depth map learning

被引:2
作者
Meng, Ming [1 ]
Xiao, Likai [2 ]
Zhou, Zhong [2 ,3 ]
机构
[1] Commun Univ China, Sch Data Sci & Media Intelligence, Beijing 100024, Peoples R China
[2] Beihang Univ, Sch Comp Sci & Engn, State Key Lab Virtual Real Technol & Syst, Beijing 100191, Peoples R China
[3] Zhongguancun Lab, Beijing 102629, Peoples R China
关键词
Structure recovery; Omnidirectional image; Planar depth map learning; Distortion-aware learning; CONTEXT; MODEL;
D O I
10.1007/s00521-023-09025-7
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Scene structure recovery is a crucial process for assisting scene reconstruction and understanding by extracting vital scene structure information and has been widely used in smart city, VR/AR and intelligent robot navigation. Omnidirectional image with a 180 degrees or 360 degrees field of view (FoV) provides greater visual information, making them a significant research topic in computer vision and computational photography. However, indoor omnidirectional scene structure recovery faces challenges like severe occlusion of critical local regions caused by cluttered objects and large nonlinear distortion. To address these limitations, we propose a geometric-driven indoor structure recovery method based on planar depth map learning, aiming to mitigate the interference caused by occlusions in critical local regions. Our approach involves designing an OmniPDMNet, a planar depth map learning network for omnidirectional image, which uses upsampling and a feature-based objective loss function to accurately estimate high-precision planar depth map. Furthermore, we leverage prior knowledge from the omnidirectional depth map and introduce it into the structure recovery network (OmniSRNet) to extract global structural features and enhance the overall quality of structure recovery. We also introduce a distortion-aware module for feature extraction from omnidirectional image, allowing adaptability to omnidirectional geometric distortion and enhancing the performance of both OmniPDMNet and OmniSRNet. Finally, we conduct extensive experiments on omnidirectional dataset focusing on planar depth and structure recovery demonstrate that our proposed method achieves state-of-the-art performance.
引用
收藏
页码:24407 / 24433
页数:27
相关论文
共 67 条
[1]   An AI-Based System Offering Automatic DR-Enhanced AR for Indoor Scenes [J].
Albanis, Georgios ;
Gkitsas, Vasileios ;
Zioulis, Nikolaos ;
Onsori-Wechtitsch, Stefanie ;
Whitehand, Richard ;
Strom, Per ;
Zarpalas, Dimitrios .
ADVANCED INTELLIGENT VIRTUAL REALITY TECHNOLOGIES, AIVR 2022, 2023, 330 :187-199
[2]  
Armeni I., 2017, ARXIV170201105
[3]   Smarter real estate marketing using virtual reality to influence potential homebuyers' emotions and purchase intention [J].
Azmi, Athira ;
Ibrahim, Rahinah ;
Ghafar, Maszura Abdul ;
Rashidi, Ali .
SMART AND SUSTAINABLE BUILT ENVIRONMENT, 2022, 11 (04) :870-890
[4]  
Buschel W., 2021, P 2021 CHI C HUM FAC, DOI [10.1145/3411764.3445651, DOI 10.1145/3411764.3445651]
[5]   Distortion-Aware Monocular Depth Estimation for Omnidirectional Images [J].
Chen, Hong-Xiang ;
Li, Kunhong ;
Fu, Zhiheng ;
Li, Mengyi ;
Chen, Zonghao ;
Guo, Yulan .
IEEE SIGNAL PROCESSING LETTERS, 2021, 28 (28) :334-338
[6]  
Cheng XJ, 2020, IEEE INT CONF ROBOT, P589, DOI [10.1109/ICRA40945.2020.9197123, 10.1109/icra40945.2020.9197123]
[7]  
Chhikara Prateek, 2023, CODS-COMAD '23: Proceedings of the 6th Joint International Conference on Data Science & Management of Data (10th ACM IKDD CODS and 28th COMAD), P223, DOI 10.1145/3570991.3571060
[8]  
Coughlan James, 2000, Advances in Neural Information Processing Systems, V13
[9]   Fusing surveillance videos and three-dimensional scene: A mixed reality system [J].
Cui, Xiaoliang ;
Khan, Dawar ;
He, Zhenbang ;
Cheng, Zhanglin .
COMPUTER ANIMATION AND VIRTUAL WORLDS, 2023, 34 (01)
[10]   Visual Computing in 360°: Foundations, Challenges, and Applications [J].
da Silveira, Thiago L. T. ;
Jung, Claudio R. .
2022 35TH SIBGRAPI CONFERENCE ON GRAPHICS, PATTERNS AND IMAGES (SIBGRAPI 2022), 2022, :302-307