Rapid 3D Model Generation with Intuitive 3D Input

被引:3
作者
Chen, Tianrun [1 ,4 ]
Ding, Chaotao [2 ]
Zhang, Shangzhan [1 ]
Yu, Chunan [2 ]
Zang, Ying [2 ]
Li, Zejian [3 ]
Peng, Sida [3 ]
Sun, Lingyun [1 ]
机构
[1] Zhejiang Univ, Coll Comp Sci & Technol, Hangzhou, Peoples R China
[2] Huzhou Univ, Sch Informat Engn, Huzhou, Peoples R China
[3] Zhejiang Univ, Sch Software Technol, Hangzhou, Peoples R China
[4] Moxin Huzhou Technol Co Ltd, KOKONI3D, Huzhou, Peoples R China
来源
2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR) | 2024年
基金
国家重点研发计划;
关键词
D O I
10.1109/CVPR52733.2024.01193
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the emergence of AR/VR, 3D models are in tremendous demand. However, conventional 3D modeling with Computer-Aided Design software requires much expertise and is difficult for novice users. We find that AR/VR devices, in addition to serving as effective display mediums, can offer a promising potential as an intuitive 3D model creation tool, especially with the assistance of AI generative models. Here, we propose Deep3DVRSketch, the first 3D model generation network that inputs 3D VR sketches from novice users and generates highly consistent 3D models in multiple categories within seconds, irrespective of the users' drawing abilities. We also contribute KO3D+, the largest 3D sketch-shape dataset. Our method pre-trains a conditional diffusion model on quality 3D data, then fine-tunes an encoder to map 3D sketches onto the generator's manifold using an adaptive curriculum strategy for limited ground truths. In our experiment, our approach achieves state-of- the-art performance in both model quality and fidelity with real-world input from novice users, and users can even draw and obtain very detailed geometric structures. In our user study, users were able to complete the 3D modeling tasks over 10 times faster using our approach compared to conventional CAD software tools. We believe that our Deep3DVRSketch and KO3D+ dataset can offer a promising solution for future 3D modeling in metaverse era. Check the project page at http://research.kokoni3d.com/Deep3DVRSketch.
引用
收藏
页码:12554 / 12564
页数:11
相关论文
共 71 条
[1]  
[Anonymous], 2018, PMLR
[2]  
[Anonymous], 2022, EUR C COMP VIS, DOI DOI 10.1007/978-3-031-20062-518
[3]  
[Anonymous], 2022, EUR C COMP VIS, DOI DOI 10.1007/978-3-031-19769-727
[4]   SymbiosisSketch: Combining 20 & 3D Sketching for Designing Detailed 3D Objects in Situ [J].
Arora, Rahul ;
Kazi, Rubaiat Habib ;
Grossman, Tovi ;
Fitzmaurice, George ;
Singh, Karan .
PROCEEDINGS OF THE 2018 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS (CHI 2018), 2018,
[5]  
Bhavnani S. K., 1999, P SIGCHI C HUM FACT, P183
[6]   Sketch-based interaction and modeling: where do we stand? [J].
Bonnici, Alexandra ;
Akman, Alican ;
Calleja, Gabriel ;
Camilleri, Kenneth P. ;
Fehling, Patrick ;
Ferreira, Alfredo ;
Hermuth, Florian ;
Israel, Johann Habakuk ;
Landwehr, Tom ;
Liu, Juncheng ;
Padfield, Natasha M. J. ;
Sezgin, T. Metin ;
Rosin, Paul L. .
AI EDAM-ARTIFICIAL INTELLIGENCE FOR ENGINEERING DESIGN ANALYSIS AND MANUFACTURING, 2019, 33 (04) :370-388
[7]   A Unified 3D Human Motion Synthesis Model via Conditional Variational Auto-Encoder [J].
Cai, Yujun ;
Wang, Yiwei ;
Zhu, Yiheng ;
Cham, Tat-Jen ;
Cai, Jianfei ;
Yuan, Junsong ;
Liu, Jun ;
Zheng, Chuanxia ;
Yan, Sijie ;
Ding, Henghui ;
Shen, Xiaohui ;
Liu, Ding ;
Thalmann, Nadia Magnenat .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :11625-11635
[8]   Multimedia Research Toward the Metaverse [J].
Chen, Shu-Ching .
IEEE MULTIMEDIA, 2022, 29 (01) :125-127
[9]  
Chen T., 2023, arXiv
[10]   Gastroesophageal Reflux Disease and Rhinosinusitis: A Bidirectional Mendelian Randomization Study [J].
Chen, Teng-yu ;
Lv, Ming-hui ;
Lai, Ren-jie ;
Hu, Yu-qi ;
Ye, Zhong-kang ;
Lv, Qi-liang ;
Chen, Bao-yi ;
Wu, Jun-tao ;
Hong, Haiyu ;
Li, Shuo .
INTERNATIONAL ARCHIVES OF ALLERGY AND IMMUNOLOGY, 2024, 185 (02) :182-189