ControlNeRF: Text-Driven 3D Scene Stylization via Diffusion Model

被引:0
|
作者
Chen, Jiahui [1 ]
Yang, Chuanfeng [1 ]
Li, Kaiheng [1 ]
Wu, Qingqiang [1 ]
Hong, Qingqi [1 ]
机构
[1] Xiamen Univ, Dept Digital Media Technol, Xiamen, Peoples R China
来源
ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING-ICANN 2024, PT II | 2024年 / 15017卷
关键词
Stylization; Neural Radiance Fields; Diffusion Model; View Synthesis;
D O I
10.1007/978-3-031-72335-3_27
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
3D scene stylization aims to generate artistically rendered images from various viewpoints within a 3D space while ensuring style consistency regardless of the viewing angle. Traditional 2D methods usually used in this field struggle with maintaining this consistency when applied to 3D environments. To address this issue, we propose a novel approach named ControlNeRF, which employs a customized conditional diffusion model, ControlNet, and introduces latent variables, obtaining a stylized appearance throughout the scene solely driven by text. Specifically, this text-driven approach effectively overcomes the inconveniences associated with using images as style cues, and it not only achieves a high degree of stylistic consistency across various viewpoints but also produces high-quality images. We have conducted rigorous testing on ControlNeRF with diverse styles, which has confirmed these outcomes. Our approach not only advances the field of 3D scene stylization but also opens new possibilities for artistic expression and digital imaging.
引用
收藏
页码:395 / 406
页数:12
相关论文
共 50 条
  • [1] 3DStyle-Diffusion: Pursuing Fine-grained Text-driven 3D Stylization with 2D Diffusion Models
    Yang, Haibo
    Chen, Yang
    Pan, Yingwei
    Yao, Ting
    Chen, Zhineng
    Mei, Tao
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 6860 - 6868
  • [2] Free-Editor: Zero-Shot Text-Driven 3D Scene Editing
    Karim, Nazmul
    Igbal, Hasan
    Khalid, Umar
    Chen, Chen
    Hua, Jing
    COMPUTER VISION - ECCV 2024, PT LXXX, 2025, 15138 : 436 - 453
  • [3] ConIS: controllable text-driven image stylization with semantic intensity
    Yang, Gaoming
    Li, Changgeng
    Zhang, Ji
    MULTIMEDIA SYSTEMS, 2024, 30 (04)
  • [4] NeRF-Art: Text-Driven Neural Radiance Fields Stylization
    Wang, Can
    Jiang, Ruixiang
    Chai, Menglei
    He, Mingming
    Chen, Dongdong
    Liao, Jing
    IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2024, 30 (08) : 4983 - 4996
  • [5] MotionDiffuse: Text-Driven Human Motion Generation With Diffusion Model
    Zhang, Mingyuan
    Cai, Zhongang
    Pan, Liang
    Hong, Fangzhou
    Guo, Xinying
    Yang, Lei
    Liu, Ziwei
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (06) : 4115 - 4128
  • [6] AvatarStudio: Text-driven Editing of 3D Dynamic Human Head Avatars
    Mendiratta, Mohit
    Pan, Xingang
    Elgharib, Mohamed
    Teotia, Kartik
    Mallikarjun, B. R.
    Tewari, Ayush
    Golyanik, Vladislav
    Kortylewski, Adam
    Theobalt, Christian
    ACM TRANSACTIONS ON GRAPHICS, 2023, 42 (06):
  • [7] LGTM: Local-to-Global Text-Driven Human Motion Diffusion Model
    Sun, Haowen
    Zheng, Ruikun
    Huang, Haibin
    Ma, Chongyang
    Huang, Hui
    Hu, Ruizhen
    PROCEEDINGS OF SIGGRAPH 2024 CONFERENCE PAPERS, 2024,
  • [8] 3D Multi-scene Stylization Based on Conditional Neural Radiance Fields
    Zhang, Sijia
    Liu, Ting
    Li, Zhuoyuan
    Sun, Yi
    ADVANCES IN NEURAL NETWORKS-ISNN 2024, 2024, 14827 : 103 - 112
  • [9] Multi-channel correlated diffusion for text-driven artistic style transfer
    Jiang, Guoquan
    Wang, Canyu
    Huo, Zhanqiang
    Xu, Huan
    VISUAL COMPUTER, 2025,
  • [10] DIFFUSIONSTR: DIFFUSION MODEL FOR SCENE TEXT RECOGNITION
    Fujitake, Masato
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 1585 - 1589