Gen-NeRF: Efficient and Generalizable Neural Radiance Fields via Algorithm-Hardware Co-Design

被引:4
作者
Fu, Yonggan [1 ]
Ye, Zhifan [1 ]
Yuan, Jiayi [2 ]
Zhang, Shunyao [2 ]
Li, Sixu [1 ]
You, Haoran [1 ]
Lin, Yingyan [1 ]
机构
[1] Georgia Inst Technol, Atlanta, GA 30332 USA
[2] Rice Univ, Houston, TX USA
来源
PROCEEDINGS OF THE 2023 THE 50TH ANNUAL INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE, ISCA 2023 | 2023年
基金
美国国家科学基金会;
关键词
Neural Radiance Field; Hardware Accelerator; VR;
D O I
10.1145/3579371.3589109
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Novel view synthesis is an essential functionality for enabling immersive experiences in various Augmented- and Virtual-Reality (AR/VR) applications, for which Neural Radiance Field (NeRF) has emerged as the state-of-the-art (SOTA) technique. In particular, generalizable NeRFs have gained increasing popularity thanks to their cross-scene generalization capability, which enables NeRFs to be instantly serviceable for new scenes without per-scene training. Despite their promise, generalizable NeRFs aggravate the prohibitive complexity of NeRFs due to their required extra memory accesses needed to acquire scene features, causing NeRFs' ray marching process to be memory-bounded. To tackle this dilemma, existing sparsity-exploitation techniques for NeRFs fall short, because they require knowledge of the sparsity distribution of the target 3D scene which is unknown when generalizing NeRFs to a new scene. To this end, we propose Gen-NeRF, an algorithm-hardware co-design framework dedicated to generalizable NeRF acceleration, which aims to win both rendering efficiency and generalization capability in NeRFs. To the best of our knowledge, Gen-NeRF is the first to enable real-time generalizable NeRFs, demonstrating a promising NeRF solution for next-generation AR/VR devices. On the algorithm side, Gen-NeRF integrates a coarse-then-focus sampling strategy, leveraging the fact that different regions of a 3D scene contribute differently to the rendered pixels depending on where the objects are located in the scene, to enable sparse yet effective s ampling. In addition, Gen-NeRF replaces the ray transformer, which is generally included in SOTA generalizable NeRFs to enhance density estimation, with a novel Ray-Mixer module to reduce workload heterogeneity. On the hardware side, Gen-NeRF highlights an accelerator micro-architecture dedicated to accelerating the resulting model workloads from our Gen-NeRF algorithm to maximize the data reuse opportunities among different rays by making use of their epipolar geometric relationship. Furthermore, our Gen-NeRF accelerator features a customized dataflow to enhance data locality during point-to-hardware mapping and an optimized scene feature storage strategy to minimize memory bank conflicts across camera rays of NeRFs. Extensive experiments validate the effectiveness of our proposed Gen-NeRF framework in enabling real-time and generalizable novel view synthesis.
引用
收藏
页码:1038 / 1049
页数:12
相关论文
共 50 条
  • [11] Hartley R., 2003, MULTIPLE VIEW GEOMET, DOI 10.1016/S0143-8166(01)00145-2
  • [12] Baking Neural Radiance Fields for Real-Time View Synthesis
    Hedman, Peter
    Srinivasan, Pratul P.
    Mildenhall, Ben
    Barron, Jonathan T.
    Debevec, Paul
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 5855 - 5864
  • [13] Local Implicit Grid Representations for 3D Scenes
    Jiang, Chiyu ''Max''
    Sud, Avneesh
    Makadia, Ameesh
    Huang, Jingwei
    Niessner, Matthias
    Funkhouser, Thomas
    [J]. 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 6000 - 6009
  • [14] GeoNeRF: Generalizing NeRF with Geometry Priors
    Johari, Mohammad Mahdi
    Lepoittevin, Yann
    Fleuret, Francois
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 18344 - 18354
  • [15] Ramulator: A Fast and Extensible DRAM Simulator
    Kim, Yoongu
    Yang, Weikun
    Mutlu, Onur
    [J]. IEEE COMPUTER ARCHITECTURE LETTERS, 2016, 15 (01) : 45 - 49
  • [16] RT-NeRF: Real-Time On-Device Neural Radiance Fields Towards Immersive AR/VR Rendering
    Li, Chaojian
    Li, Sixu
    Zhao, Yang
    Zhu, Wenbo
    Lin, Yingyan
    [J]. 2022 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED DESIGN, ICCAD, 2022,
  • [17] Neural Scene Flow Fields for Space-Time View Synthesis of Dynamic Scenes
    Li, Zhengqi
    Niklaus, Simon
    Snavely, Noah
    Wang, Oliver
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 6494 - 6504
  • [18] Neural Rays for Occlusion-aware Image-based Rendering
    Liu, Yuan
    Peng, Sida
    Liu, Lingjie
    Wang, Qianqian
    Wang, Peng
    Theobalt, Christian
    Zhou, Xiaowei
    Wang, Wenping
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 7814 - 7823
  • [19] Meta, 2022, Meta Quest Pro
  • [20] Mildenhall Ben, 2020, Computer Vision - ECCV 2020. 16th European Conference. Proceedings. Lecture Notes in Computer Science (LNCS 12346), P405, DOI 10.1007/978-3-030-58452-8_24