Light field reconstruction using hierarchical features fusion

被引:9
作者
Hu, Zexi [1 ]
Chung, Yuk Ying [1 ]
Ouyang, Wanli [2 ]
Chen, Xiaoming [3 ]
Chen, Zhibo [4 ]
机构
[1] Univ Sydney, Sch Comp Sci, Sydney, NSW 2006, Australia
[2] Univ Sydney, Sch Elect & Informat Engn, Sydney, NSW 2006, Australia
[3] Univ Sci & Technol China, Inst Adv Technol, Hefei 230026, Peoples R China
[4] Univ Sci & Technol China, CAS Key Lab Technol Geospatial Informat Proc & Ap, Hefei 230027, Peoples R China
关键词
Light field; Deep learning; Neural network; Image processing; DEPTH; NETWORK;
D O I
10.1016/j.eswa.2020.113394
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Light field imagery has attracted increasing attention for its capacity of simultaneously capturing intensity values of light rays from multiple directions. Such imagery technique has become widely accessible with the emergence of consumer-grade devices, e.g. Lytro, and the Virtual Reality (VR) / Augmented Reality (AR) areas. Light field reconstruction is a critical topic to mitigate the trade-off problem between the spatial and angular resolutions. Learning-based methods have attained outstanding performance among the recently proposed methods, however, the state-of-the-art methods still suffer from heavy artifacts in the case of occlusion. This is likely to be a consequence of failure in capturing the semantic information from the limited spatial receptive field during training. It is crucial for light field reconstruction to learn semantic features and understand a wider context in both the angular and spatial dimensions. To address this issue, we introduce a novel end-to-end U-Net with SAS network (U-SAS-Net) to extract and fuse hierarchical features, both local and semantic, from a relatively large receptive field while establishing the relation of the correlated sub-aperture images. Experimental results on extensive light field datasets demonstrate that our method produces a state-of-the-art performance that exceeds the previous works by more than 0.6 dB PSNR with the fused hierarchical features, especially the semantic features for handling scenes with occlusion and the local features for recovering the rich details. Meanwhile, our method is at a substantially lower cost which takes 48% parameters and less than 10% computation of the previous state-of-the-art method. (C) 2020 Elsevier Ltd. All rights reserved.
引用
收藏
页数:11
相关论文
共 39 条
[11]   Convolutional Networks for Shape from Light Field [J].
Heber, Stefan ;
Pock, Thomas .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :3746-3754
[12]  
Huang F.-C., 2015, P C SIGGRAPH EM TECH
[13]   Depth from a Light Field Image with Learning-Based Matching Costs [J].
Jeon, Hae-Gon ;
Park, Jaesik ;
Choe, Gyeongmin ;
Park, Jinsun ;
Bok, Yunsu ;
Tai, Yu-Wing ;
Kweon, In So .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2019, 41 (02) :297-310
[14]  
Jeon HG, 2015, PROC CVPR IEEE, P1547, DOI 10.1109/CVPR.2015.7298762
[15]   Learning-Based View Synthesis for Light Field Cameras [J].
Kalantari, Nima Khademi ;
Wang, Ting-Chun ;
Ramamoorthi, Ravi .
ACM TRANSACTIONS ON GRAPHICS, 2016, 35 (06)
[16]  
Kim J, 2016, PROC CVPR IEEE, P1637, DOI [10.1109/CVPR.2016.181, 10.1109/CVPR.2016.182]
[17]  
LAI WS, 2017, PROC CVPR IEEE, P5835, DOI DOI 10.1109/CVPR.2017.618
[18]   Improved image classification with 4D light-field and interleaved convolutional neural network [J].
Lu, Zhicheng ;
Yeung, Henry W. F. ;
Qu, Qiang ;
Chung, Yuk Ying ;
Chen, Xiaoming ;
Chen, Zhibo .
MULTIMEDIA TOOLS AND APPLICATIONS, 2019, 78 (20) :29211-29227
[19]  
Ng R., 2005, LIGHT FIELD PHOTOGRA
[20]   Plenoptic Layer-Based Modeling for Image Based Rendering [J].
Pearson, James ;
Brookes, Mike ;
Dragotti, Pier Luigi .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2013, 22 (09) :3405-3419