Text-Guided Synthesis of Crowd Animation

被引:1
作者
Ji, Xuebo [1 ]
Pan, Zherong [2 ]
Gao, Xifeng [2 ]
Pan, Jia [1 ]
机构
[1] Univ Hong Kong, Ctr Transformat Garment Prod TransGP, Hong Kong, Peoples R China
[2] LightSpeed Studios, Seattle, WA USA
来源
PROCEEDINGS OF SIGGRAPH 2024 CONFERENCE PAPERS | 2024年
关键词
Diffusion Model; Multi-Agent Navigation; Collision Avoidance; Crowd Simulation;
D O I
10.1145/3641519.3657516
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Creating vivid crowd animations is core to immersive virtual environments in digital games. This work focuses on tackling the challenges of the crowd behavior generation problem. Existing approaches are labor-intensive, relying on practitioners to manually craft the complex behavior systems. We propose a machine learning approach to synthesize diversified dynamic crowd animation scenarios for a given environment based on a text description input. We first train two conditional diffusion models that generate text-guided agent distribution fields and velocity fields. Assisted by local navigation algorithms, the fields are then used to control multiple groups of agents. We further employ Large-Language Model (LLM) to canonicalize the general script into a structured sentence for more stable training and better scalability. To train our diffusion models, we devise a constructive method to generate random environments and crowd animations. We show that our trained diffusion models can generate crowd animations for both unseen environments and novel scenario descriptions. Our method paves the way towards automatic generating of crowd behaviors for virtual environments. Code and data for this paper are available at: https://github.com/MLZG/Text-Crowd.git.
引用
收藏
页数:11
相关论文
共 43 条
  • [41] Semantic-Layout-Guided Image Synthesis for High-Quality Synthetic-Aperature Radar Detection Sample Generation
    Kuang, Yi
    Ma, Fei
    Li, Fangfang
    Liu, Yingbing
    Zhang, Fan
    Gleich, Dusan
    REMOTE SENSING, 2023, 15 (24)
  • [42] PanGu-Draw: Advancing Resource-Efficient Text-to-Image Synthesis with Time-Decoupled Training and Reusable Coop-Diffusion
    Lu, Guansong
    Guo, Yuanfan
    Han, Jianhua
    Niu, Minzhe
    Zhang, Yihan
    Xu, Songcen
    Huang, Zeyi
    Zhong, Zhao
    Zhang, Wei
    Xu, Hang
    COMPUTER VISION - ECCV 2024, PT XLV, 2025, 15103 : 159 - 176
  • [43] FgC2F-UDiff: Frequency-Guided and Coarse-to-Fine Unified Diffusion Model for Multi-Modality Missing MRI Synthesis
    Xiao, Xiaojiao
    Hu, Qinmin Vivian
    Wang, Guanghui
    IEEE TRANSACTIONS ON COMPUTATIONAL IMAGING, 2024, 10 : 1815 - 1828