FewarNet: An Efficient Few-Shot View Synthesis Network Based on Trend Regularization

被引:0
作者
Song, Chenxi [1 ]
Wang, Shigang [1 ]
Wei, Jian [1 ]
Zhao, Yan [1 ]
机构
[1] Jilin Univ, Coll Commun Engn, Changchun 130012, Peoples R China
基金
中国国家自然科学基金;
关键词
Three-dimensional displays; Market research; Cameras; Costs; Geometry; Training; Estimation; Depth estimation; few-shot view synthesis; regularization constraint; prior depth; VIDEO;
D O I
10.1109/TCSVT.2024.3395447
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Novel view synthesis from existing inputs remains a research focus in computer vision. Predicting views becomes more challenging when only a limited number of views are available. This challenge is commonly referred to as the few-shot view synthesis problem. Recently, various strategies have emerged for few-shot view synthesis, such as transfer learning, depth supervision, and regularization constraints. However, transfer learning relies on massive scene data, depth supervision is affected by input depth quality, and regularization causes increased computational costs or impaired generalization. To address these issues, we propose a new few-shot view synthesis framework called FewarNet that introduces trend regularization to leverage depth structural features and a warping loss to supervise depth estimation, possessing the advantages of existing few-shot strategies, enabling high-quality novel view prediction with generalization and efficiency. Specifically, FewarNet consists of three stages: fusion, warping, and rectification. In the fusion stage, a fusion network is introduced to estimate depths using scene priors from coarse depths. In the warping stage, the predicted depths are used to guide the warping of the input views, and a distance-weighted warping loss is proposed to correctly guide depth estimation. To further improve prediction accuracy, we propose trend regularization which imposes penalties on depth variation trends to provide depth structural constraints. In the rectification stage, a rectification network is introduced to refine occluded regions in each warped view to generate novel views. Additionally, a rapid view synthesis strategy that leverages depth interpolation is designed to improve efficiency. We validate the method's effectiveness and generalization on various datasets. Given the same sparse inputs, our method demonstrates superior performance in quality and efficiency over state-of-the-art few-shot view synthesis methods.
引用
收藏
页码:9264 / 9280
页数:17
相关论文
共 65 条
  • [1] Mip-NeRF: A Multiscale Representation for Anti-Aliasing Neural Radiance Fields
    Barron, Jonathan T.
    Mildenhall, Ben
    Tancik, Matthew
    Hedman, Peter
    Martin-Brualla, Ricardo
    Srinivasan, Pratul P.
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 5835 - 5844
  • [2] Birkl R, 2023, Arxiv, DOI arXiv:2307.14460
  • [3] FWD: Real-time Novel View Synthesis with Forward Warping and Depth
    Cao, Ang
    Rockwell, Chris
    Johnson, Justin
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 15692 - 15703
  • [4] Image-based rendering and synthesis - Technological advances and challenges
    Chan, S. C.
    Shum, Heung-Yeung
    Ng, King-To
    [J]. IEEE SIGNAL PROCESSING MAGAZINE, 2007, 24 (06) : 22 - 33
  • [5] MVSNeRF: Fast Generalizable Radiance Field Reconstruction from Multi-View Stereo
    Chen, Anpei
    Xu, Zexiang
    Zhao, Fuqiang
    Zhang, Xiaoshuai
    Xiang, Fanbo
    Yu, Jingyi
    Su, Hao
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 14104 - 14113
  • [6] Stereo Radiance Fields (SRF): Learning View Synthesis for Sparse Views of Novel Scenes
    Chibane, Julian
    Bansal, Aayush
    Lazova, Verica
    Pons-Moll, Gerard
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 7907 - 7916
  • [7] Depth-supervised NeRF: Fewer Views and Faster Training for Free
    Deng, Kangle
    Liu, Andrew
    Zhu, Jun-Yan
    Ramanan, Deva
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 12872 - 12881
  • [8] Dosovitskiy A., 2021, arXiv
  • [9] Neural Radiance Flow for 4D View Synthesis and Video Processing
    Du, Yilun
    Zhang, Yinan
    Yu, Hong-Xing
    Tenenbaum, Joshua B.
    Wu, Jiajun
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 14304 - 14314
  • [10] DeepView: View synthesis with learned gradient descent
    Flynn, John
    Broxton, Michael
    Debevec, Paul
    DuVall, Matthew
    Fyffe, Graham
    Overbeck, Ryan
    Snavely, Noah
    Tucker, Richard
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 2362 - 2371