ControlNeRF: Text-Driven 3D Scene Stylization via Diffusion Model

被引:0
|
作者
Chen, Jiahui [1 ]
Yang, Chuanfeng [1 ]
Li, Kaiheng [1 ]
Wu, Qingqiang [1 ]
Hong, Qingqi [1 ]
机构
[1] Xiamen Univ, Dept Digital Media Technol, Xiamen, Peoples R China
来源
ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING-ICANN 2024, PT II | 2024年 / 15017卷
关键词
Stylization; Neural Radiance Fields; Diffusion Model; View Synthesis;
D O I
10.1007/978-3-031-72335-3_27
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
3D scene stylization aims to generate artistically rendered images from various viewpoints within a 3D space while ensuring style consistency regardless of the viewing angle. Traditional 2D methods usually used in this field struggle with maintaining this consistency when applied to 3D environments. To address this issue, we propose a novel approach named ControlNeRF, which employs a customized conditional diffusion model, ControlNet, and introduces latent variables, obtaining a stylized appearance throughout the scene solely driven by text. Specifically, this text-driven approach effectively overcomes the inconveniences associated with using images as style cues, and it not only achieves a high degree of stylistic consistency across various viewpoints but also produces high-quality images. We have conducted rigorous testing on ControlNeRF with diverse styles, which has confirmed these outcomes. Our approach not only advances the field of 3D scene stylization but also opens new possibilities for artistic expression and digital imaging.
引用
收藏
页码:395 / 406
页数:12
相关论文
共 50 条
  • [31] Advances in text-guided 3D editing: a survey
    Lu, Lihua
    Li, Ruyang
    Zhang, Xiaohui
    Wei, Hui
    Du, Guoguang
    Wang, Binqiang
    ARTIFICIAL INTELLIGENCE REVIEW, 2024, 57 (12)
  • [32] DTF-diffusion: A 3D equivariant diffusion generation model based on ligand-target information fusion
    Wang, Jianxin
    Zhu, Yongxin
    Liu, Yushuang
    Yu, Bin
    COMPUTATIONAL BIOLOGY AND CHEMISTRY, 2025, 117
  • [33] DiffSurf: A Transformer-Based Diffusion Model for Generating and Reconstructing 3D Surfaces in Pose
    Yoshiyasu, Yusuke
    Sun, Leyuan
    COMPUTER VISION-ECCV 2024, PT LXXXII, 2025, 15140 : 246 - 264
  • [34] Anything to Glyph: Artistic Font Synthesis via Text-to-Image Diffusion Model
    Wang, ChangShuo
    Wu, Lei
    Liu, XiaoLe
    Li, Xiang
    Meng, Lei
    Meng, Xiangxu
    PROCEEDINGS OF THE SIGGRAPH ASIA 2023 CONFERENCE PAPERS, 2023,
  • [35] Monocular thermal SLAM with neural radiance fields for 3D scene reconstruction
    Wu, Yuzhen
    Wang, Lingxue
    Zhang, Lian
    Chen, Mingkun
    Zhao, Wenqu
    Zheng, Dezhi
    Cai, Yi
    NEUROCOMPUTING, 2025, 617
  • [36] A multi-condition denoising diffusion probabilistic model controls the reconstruction of 3D digital rocks
    Luo, Xin
    Sun, Jianmeng
    Zhang, Ran
    Chi, Peng
    Cui, Ruikang
    COMPUTERS & GEOSCIENCES, 2024, 184
  • [37] MFDiff: multiscale feature diffusion model for segmentation of 3D intracranial aneurysm from CT images
    Pei, Xinyu
    Ren, Yande
    Tang, Yueshan
    Wang, Yuanquan
    Zhang, Lei
    Wei, Jin
    Zhao, Di
    PATTERN ANALYSIS AND APPLICATIONS, 2024, 27 (02)
  • [38] 3D Contour Generation based on Diffusion Probabilistic Models
    Wu, Yiqi
    Huang, Xuan
    Song, Kelin
    He, Fazhi
    Zhang, Dejun
    PROCEEDINGS OF THE 2024 27 TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, CSCWD 2024, 2024, : 1992 - 1997
  • [39] Neural Rendering-Based 3D Scene Style Transfer Method via Semantic Understanding Using a Single Style Image
    Park, Jisun
    Cho, Kyungeun
    MATHEMATICS, 2023, 11 (14)
  • [40] AnimatableDreamer: Text-Guided Non-rigid 3D Model Generation and Reconstruction with Canonical Score Distillation
    Wang, Xinzhou
    Wang, Yikai
    Yee, Junliang
    Sung, Fuchun
    Wang, Zhengyi
    Wang, Ling
    Liu, Pengkun
    Sung, Kai
    Wan, Xintong
    Xie, Wende
    Liu, Fangfu
    He, Bin
    COMPUTER VISION - ECCV 2024, PT XXV, 2025, 15083 : 321 - 339