A Unified Learning-Based Framework for Light Field Reconstruction From Coded Projections

被引:17
作者
Vadathya, Anil Kumar [1 ]
Girish, Sharath [2 ]
Mitra, Kaushik [3 ]
机构
[1] Rice Univ, Dept Elect & Comp Engn, Houston, TX 77005 USA
[2] Univ Maryland, Dept Comp Sci, College Pk, MD 20740 USA
[3] Indian Inst Technol Madras, Dept Elect Engn, Chennai 600036, Tamil Nadu, India
关键词
Light field resolution trade-off; compressive light field imaging; coded aperture photography; disparity based view synthesis; DEPTH; NETWORK;
D O I
10.1109/TCI.2019.2948780
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Light fields present a rich way to represent the 3D world by capturing the spatio-angular dimensions of the visual signal. However, the popular way of capturing light fields (LF) via a plenoptic camera presents a spatio-angular resolution trade-off. To address this issue, computational imaging techniques such as compressive light field and programmable coded aperture have been proposed, which reconstruct full sensor resolution LF from coded projections of the LF. Here, we present a unified learning framework that can reconstruct LF from a variety of multiplexing schemes with minimal number of coded images as input. We consider three light field capture schemes: heterodyne capture scheme with code placed near the sensor, coded aperture scheme with code at the camera aperture and finally the dual exposure scheme of capturing a focus-defocus pair where there is no explicit coding. Our algorithm consists of three stages: Firstly, we recover the all-in-focus image from the coded image. Secondly, we estimate the disparity maps for all the LF views from the coded image and the all-in-focus image. And finally, we render the LF by warping the all-in-focus image using the estimated disparity maps. We show that our proposed learning algorithm performs either on par with or better than the state-of-the-art methods for all the three multiplexing schemes. LF from focus-defocus pair is especially attractive as it requires no hardware modification and produces LF reconstructions that are comparable to the current state of the art learning-based view synthesis approaches from multiple images. Thus, our work paves the way for capturing full-resolution LF using conventional cameras such as DSLRs and smartphones.
引用
收藏
页码:304 / 316
页数:13
相关论文
共 40 条
[1]  
[Anonymous], 2016, ICLR
[2]  
[Anonymous], 2015, ACS SYM SER
[3]  
[Anonymous], 2015, ARXIV150301903
[4]  
[Anonymous], 2009, IEEE T PATTERN ANAL
[5]   Compressive Light Field Imaging [J].
Ashok, Amit ;
Neifeld, Mark A. .
THREE-DIMENSIONAL IMAGING, VISUALIZATION, AND DISPLAY 2010 AND DISPLAY TECHNOLOGIES AND APPLICATIONS FOR DEFENSE, SECURITY, AND AVIONICS IV, 2010, 7690
[6]   Compressive Light Field Sensing [J].
Babacan, S. Derin ;
Ansorge, Reto ;
Luessi, Martin ;
Ruiz Mataran, Pablo ;
Molina, Rafael ;
Katsaggelos, Aggelos K. .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2012, 21 (12) :4746-4757
[7]   Joint View Expansion and Filtering for Automultiscopic 3D Displays [J].
Didyk, Piotr ;
Sitthi-Amorn, Pitchaya ;
Freeman, William ;
Durand, Fredo ;
Matusik, Wojciech .
ACM TRANSACTIONS ON GRAPHICS, 2013, 32 (06)
[8]   DeepStereo: Learning to Predict New Views from the World's Imagery [J].
Flynn, John ;
Neulander, Ivan ;
Philbin, James ;
Snavely, Noah .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :5515-5524
[9]   Unsupervised CNN for Single View Depth Estimation: Geometry to the Rescue [J].
Garg, Ravi ;
VijayKumar, B. G. ;
Carneiro, Gustavo ;
Reid, Ian .
COMPUTER VISION - ECCV 2016, PT VIII, 2016, 9912 :740-756
[10]  
Georgiev T. G., 2006, Rendering Techn., V2006