Learning an epipolar shift compensation for light field image super-resolution

被引:19
作者
Wang, Xinya [1 ]
Ma, Jiayi [1 ]
Yi, Peng [2 ]
Tian, Xin [1 ]
Jiang, Junjun [3 ]
Zhang, Xiao-Ping [4 ]
机构
[1] Wuhan Univ, Elect Informat Sch, Wuhan 430072, Peoples R China
[2] Wuhan Univ, Sch Comp, Wuhan 430072, Peoples R China
[3] Harbin Inst Technol, Sch Comp Sci & Technol, Harbin 150001, Peoples R China
[4] Ryerson Univ, Dept Elect Comp & Biomed Engn, Toronto, ON M5B 2K3, Canada
基金
加拿大自然科学与工程研究理事会; 中国国家自然科学基金;
关键词
Light field; Super-resolution; Multi-view fusion; Dynamic deformable convolution; FEATURE FUSION; RESOLUTION; NETWORK;
D O I
10.1016/j.inffus.2021.10.005
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Light field imaging has drawn broad attention since the advent of practical light field capturing systems that facilitate a wide range of applications in computer vision. However, existing learning-based methods for improving the spatial resolution of light field images neglect the shifts in the sub-pixel domain that are widely used by super-resolution techniques, thus, fail in recovering rich high-frequency information. To fully exploit the shift information, our method attempts to learn an epipolar shift compensation for light field image super-resolution that allows the restored light field image to be angular coherent with the enhancement of spatial resolution. The proposed method first utilizes the rich surrounding views along some typical epipolar directions to explore the inter-view correlations. We then implement feature-level registration to capture accurate sub-pixel shifts of central view, which is constructed by the compensation module equipped with dynamic deformable convolution. Finally, the complementary information from different spatial directions is fused to provide high-frequency details for the target view. By taking each sub-aperture image as a central view, our method could be applied for light field images with any angular resolution. Extensive experiments on both synthetic and real scene datasets demonstrate the superiority of our method over the state-of-the-art qualitatively and quantitatively. Moreover, the proposed method shows good performance in preserving the inherent epipolar structures in light field images. Specifically, our LFESCN method outperforms the state-of-the-art method with about 0.7 dB (PSNR) on average.
引用
收藏
页码:188 / 199
页数:12
相关论文
共 64 条
[1]  
Alain M, 2018, IEEE IMAGE PROC, P2501, DOI 10.1109/ICIP.2018.8451162
[2]  
[Anonymous], 2016, Advances in Neural Information Processing Systems
[3]  
[Anonymous], 2013, P VMV
[4]   The Light Field Camera: Extended Depth of Field, Aliasing, and Superresolution [J].
Bishop, Tom E. ;
Favaro, Paolo .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2012, 34 (05) :972-986
[5]  
Bishop Tom E., 2009, P IEEE INT C COMPUTA, P1
[6]   Soft-NMS - Improving Object Detection With One Line of Code [J].
Bodla, Navaneeth ;
Singh, Bharat ;
Chellappa, Rama ;
Davis, Larry S. .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :5562-5570
[7]  
Capel D, 2001, PROC CVPR IEEE, P627
[8]   Deep Cross-Modal Audio-Visual Generation [J].
Chen, Lele ;
Srivastava, Sudhanshu ;
Duan, Zhiyao ;
Xu, Chenliang .
PROCEEDINGS OF THE THEMATIC WORKSHOPS OF ACM MULTIMEDIA 2017 (THEMATIC WORKSHOPS'17), 2017, :349-357
[9]   Light Field Super-Resolution By Jointly Exploiting Internal and External Similarities [J].
Cheng, Zhen ;
Xiong, Zhiwei ;
Liu, Dong .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2020, 30 (08) :2604-2616
[10]  
Deudon Michel, 2020, HIGHRES NET RECURSIV