Light field reconstruction using hierarchical features fusion

被引:9
作者
Hu, Zexi [1 ]
Chung, Yuk Ying [1 ]
Ouyang, Wanli [2 ]
Chen, Xiaoming [3 ]
Chen, Zhibo [4 ]
机构
[1] Univ Sydney, Sch Comp Sci, Sydney, NSW 2006, Australia
[2] Univ Sydney, Sch Elect & Informat Engn, Sydney, NSW 2006, Australia
[3] Univ Sci & Technol China, Inst Adv Technol, Hefei 230026, Peoples R China
[4] Univ Sci & Technol China, CAS Key Lab Technol Geospatial Informat Proc & Ap, Hefei 230027, Peoples R China
关键词
Light field; Deep learning; Neural network; Image processing; DEPTH; NETWORK;
D O I
10.1016/j.eswa.2020.113394
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Light field imagery has attracted increasing attention for its capacity of simultaneously capturing intensity values of light rays from multiple directions. Such imagery technique has become widely accessible with the emergence of consumer-grade devices, e.g. Lytro, and the Virtual Reality (VR) / Augmented Reality (AR) areas. Light field reconstruction is a critical topic to mitigate the trade-off problem between the spatial and angular resolutions. Learning-based methods have attained outstanding performance among the recently proposed methods, however, the state-of-the-art methods still suffer from heavy artifacts in the case of occlusion. This is likely to be a consequence of failure in capturing the semantic information from the limited spatial receptive field during training. It is crucial for light field reconstruction to learn semantic features and understand a wider context in both the angular and spatial dimensions. To address this issue, we introduce a novel end-to-end U-Net with SAS network (U-SAS-Net) to extract and fuse hierarchical features, both local and semantic, from a relatively large receptive field while establishing the relation of the correlated sub-aperture images. Experimental results on extensive light field datasets demonstrate that our method produces a state-of-the-art performance that exceeds the previous works by more than 0.6 dB PSNR with the fused hierarchical features, especially the semantic features for handling scenes with occlusion and the local features for recovering the rich details. Meanwhile, our method is at a substantially lower cost which takes 48% parameters and less than 10% computation of the previous state-of-the-art method. (C) 2020 Elsevier Ltd. All rights reserved.
引用
收藏
页数:11
相关论文
共 39 条
[1]  
Abadi Martin, 2016, Proceedings of OSDI '16: 12th USENIX Symposium on Operating Systems Design and Implementation. OSDI '16, P265
[2]   Hybrid light field imaging for improved spatial resolution and depth range [J].
Alam, M. Zeshan ;
Gunturk, Bahadir K. .
MACHINE VISION AND APPLICATIONS, 2018, 29 (01) :11-22
[3]  
Ali M, 2014, 2014 IEEE INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING, COMMUNICATIONS AND COMPUTING (ICSPCC), P184, DOI 10.1109/ICSPCC.2014.6986179
[4]   Accurate Light Field Depth Estimation With Superpixel Regularization Over Partially Occluded Regions [J].
Chen, Jie ;
Hou, Junhui ;
Ni, Yun ;
Chau, Lap-Pui .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2018, 27 (10) :4889-4900
[5]  
Chollet F., 2015, KERAS
[6]  
Fiss J., 2014, Computational Photography (ICCP), 2014 IEEE International Conference on, P1
[7]  
Georgiev T. G., 2006, Rendering Techn., V2006
[8]  
HE KM, 2016, PROC CVPR IEEE, P770, DOI DOI 10.1109/CVPR.2016.90
[9]  
Heber S., 2016, P BRIT MACH VIS C BM, V1, DOI DOI 10.5244/C.30.37
[10]   Neural EPI-volume Networks for Shape from Light Field [J].
Heber, Stefan ;
Yu, Wei ;
Pock, Thomas .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :2271-2279