SACANet: end-to-end self-attention-based network for 3D clothing animation

被引:0
|
作者
Chen, Yunxi [1 ]
Cao, Yuanjie [1 ]
Fang, Fei [1 ,2 ]
Huang, Jin [1 ,2 ]
Hu, Xinrong [1 ]
He, Ruhan [1 ,2 ]
Zhang, Junjie [1 ]
机构
[1] Wuhan Text Univ, Sch Comp Sci & Artificial Intelligence, Wuhan, Peoples R China
[2] Wuhan Text Univ, Hubei Prov Engn Res Ctr Intelligent Text & Fash, Wuhan, Peoples R China
关键词
3D clothing animation; Clothing wrinkles; Body-garment interpenetrations; Self-attention; Physics-constrained loss function; ALGORITHM;
D O I
10.1007/s00371-024-03633-7
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
In real-time applications such as virtual try-on and video games, achieving realistic character clothing animation is an active research area. Existing learning-based methods aim to simulate realistic clothing effects while meeting real-time requirements. However, due to the diverse sources of clothing deformation, factors such as human body shape, pose, and clothing style all influence the realism of clothing deformation. Existing clothing animation methods often encounter issues such as body-garment interpenetrations and insufficient clothing wrinkles. In this paper, we propose a new end-to-end self-attention-based two-level network architecture for clothing animation, called SACANet, to address these issues. Specifically, we employ a physics-constrained loss function to enable our model to automatically resolve body-garment interpenetrations without the need for post-processing steps. Additionally, within our two-level network, we use an effective self-attention mechanism, namely cross-covariance attention, to enhance our model's ability to learn geometric features and to capture more realistic clothing deformations. The base deformation network learns the global deformation of the clothing by extracting global geometric features from the input data. The detail deformation network learns the local wrinkle details of the clothing by fusing features extracted by specific pose models related to human body shape and clothing style. Through a series of experiments, we comprehensively evaluate and compare the results. These experiments demonstrate the superiority of our method in simulating realistic clothing animation. https://github.com/JUNOHINATA/SACANet.
引用
收藏
页码:3829 / 3842
页数:14
相关论文
共 50 条
  • [21] End-to-End Full-Waveform Echo Decomposition Based on Self-Attention Classification and U-Net Decomposition
    Liu, Gangping
    Ke, Jun
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2022, 15 : 7978 - 7987
  • [22] EXPLORATION OF LANGUAGE-SPECIFIC SELF-ATTENTION PARAMETERS FOR MULTILINGUAL END-TO-END SPEECH RECOGNITION
    Houston, Brady
    Kirchhoff, Katrin
    2022 IEEE SPOKEN LANGUAGE TECHNOLOGY WORKSHOP, SLT, 2022, : 755 - 762
  • [23] End-to-End Pixel-Wisely Detection of Oceanic Eddy on SAR Images With Stacked Attention Network
    Xu, Ming
    Li, Hongping
    Yun, Yuying
    Yang, Fan
    Li, Cuishu
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2023, 16 : 10138 - 10151
  • [24] Improving Hybrid CTC/Attention Architecture with Time-Restricted Self-Attention CTC for End-to-End Speech Recognition
    Wu, Long
    Li, Ta
    Wang, Li
    Yan, Yonghong
    APPLIED SCIENCES-BASEL, 2019, 9 (21):
  • [25] E2DA: Energy Efficient Data Aggregation and End-to-End Security in 3D Reconfigurable WSN
    Ramasamy, Karthick
    Anisi, Mohammad Hossein
    Jindal, Anish
    IEEE TRANSACTIONS ON GREEN COMMUNICATIONS AND NETWORKING, 2022, 6 (02): : 787 - 798
  • [26] CAA-Net: End-to-End Two-Branch Feature Attention Network for Single Image Dehazing
    Jin, Gang
    Zhai, Jingsheng
    Wei, Jianguo
    IEICE TRANSACTIONS ON FUNDAMENTALS OF ELECTRONICS COMMUNICATIONS AND COMPUTER SCIENCES, 2023, E106A (01) : 1 - 10
  • [27] Self-Attention Channel Combinator Frontend for End-to-End Multichannel Far-field Speech Recognition
    Gong, Rong
    Quillen, Carl
    Sharma, Dushyant
    Goderre, Andrew
    Lainez, Jose
    Milanovic, Ljubomir
    INTERSPEECH 2021, 2021, : 3840 - 3844
  • [28] End-to-end 3D CNN for plot-scale soybean yield prediction using multitemporal UAV-based RGB images
    Bhadra, Sourav
    Sagan, Vasit
    Skobalski, Juan
    Grignola, Fernando
    Sarkar, Supria
    Vilbig, Justin
    PRECISION AGRICULTURE, 2024, 25 (02) : 1014 - 1037
  • [29] Multi-Scale Visual Semantics Aggregation with Self-Attention for End-to-End Image-Text Matching
    Zheng, Zhuobin
    Ben, Youcheng
    Yuan, Chun
    ASIAN CONFERENCE ON MACHINE LEARNING, VOL 101, 2019, 101 : 940 - 955
  • [30] Multiscale global and local self-attention-based network for remaining useful life prediction
    Zhang, Zhizheng
    Song, Wen
    Li, Qiqiang
    Gao, Hui
    MEASUREMENT SCIENCE AND TECHNOLOGY, 2023, 34 (12)