Dense Convolutional Recurrent Neural Network for Generalized Speech Animation

被引:0
|
作者
Xiao, Lei [1 ]
Wang, Zengfu
机构
[1] Chinese Acad Sci, Inst Intelligent Machines, Hefei, Anhui, Peoples R China
基金
中国国家自然科学基金;
关键词
LIPS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper presents a novel automated speech animation approach named Dense Convolutional Recurrent Neural Network (DenseCRNN). The approach learns a non-linear mapping from acoustic speech to multiple articulator movements in a unified framework to which feature extraction, context encoding and multi-parameter decoding are integrated. We propose DenseCRNN based on three insights: (1) One can use a convolutional neural network incorporated with dense connectivity to extract speaker-independent features from arbitrary spoken audio effectively. (2) A bidirectional long short-term memory neural network is able to model the context information with respect to the phoneme coarticulation. (3) Multi-domain learning can be implemented to achieve better performance on account of the implicit correlation and explicit difference among outputs, where each domain is responsible for a single visual parameter. Experiments on MNGU0 dataset demonstrate our approach achieves significant improvements over state-of-the-art methods. Moreover, the proposed approach generalizes over different gender or accent, and has the capability of deploying on various character models.
引用
收藏
页码:633 / 638
页数:6
相关论文
共 50 条
  • [1] RECURRENT CONVOLUTIONAL NEURAL NETWORK FOR SPEECH PROCESSING
    Zhao, Yue
    Jin, Xingyu
    Hu, Xiaolin
    2017 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2017, : 5300 - 5304
  • [2] Automatic Speech Recognition trained with Convolutional Neural Network and predicted with Recurrent Neural Network
    Soundarya, M.
    Karthikeyan, P. R.
    Thangarasu, Gunasekar
    2023 9TH INTERNATIONAL CONFERENCE ON ELECTRICAL ENERGY SYSTEMS, ICEES, 2023, : 41 - 45
  • [3] Dilated convolutional recurrent neural network for monaural speech enhancement
    Pirhosseinloo, Shadi
    Brumberg, Jonathan S.
    CONFERENCE RECORD OF THE 2019 FIFTY-THIRD ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS, 2019, : 158 - 162
  • [4] Speech enhancement method based on convolutional gated recurrent neural network
    Yuan W.
    Lou Y.
    Xia B.
    Sun W.
    Huazhong Keji Daxue Xuebao (Ziran Kexue Ban)/Journal of Huazhong University of Science and Technology (Natural Science Edition), 2019, 47 (04): : 13 - 18
  • [5] Speech Separation Using an Asynchronous Fully Recurrent Convolutional Neural Network
    Hu, Xiaolin
    Li, Kai
    Zhang, Weiyi
    Luo, Yi
    Lemercier, Jean-Marie
    Gerkmann, Timo
    Advances in Neural Information Processing Systems, 2021, 27 : 22509 - 22522
  • [6] A Convolutional Recurrent Neural Network for Real-Time Speech Enhancement
    Tan, Ke
    Wang, DeLiang
    19TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2018), VOLS 1-6: SPEECH RESEARCH FOR EMERGING MARKETS IN MULTILINGUAL SOCIETIES, 2018, : 3229 - 3233
  • [7] Speech Separation Using an Asynchronous Fully Recurrent Convolutional Neural Network
    Hu, Xiaolin
    Li, Kai
    Zhang, Weiyi
    Luo, Yi
    Lemercier, Jean-Marie
    Gerkmann, Timo
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [8] Speech separation using an asynchronous Fully Recurrent Convolutional Neural Network
    Department of Computer Science and Technology, Tsinghua Laboratory of Brain and Intelligence , IDG/McGovern Institute of Brain Research, Tsinghua University, Beijing, China
    不详
    不详
    arXiv, 2021,
  • [9] Real-Time Speech Enhancement Based on Convolutional Recurrent Neural Network
    Girirajan, S.
    Pandian, A.
    INTELLIGENT AUTOMATION AND SOFT COMPUTING, 2023, 35 (02): : 1987 - 2001
  • [10] A convolutional recurrent neural network with attention framework for speech separation in monaural recordings
    Chao Sun
    Min Zhang
    Ruijuan Wu
    Junhong Lu
    Guo Xian
    Qin Yu
    Xiaofeng Gong
    Ruisen Luo
    Scientific Reports, 11