MAMGAN: Multiscale attention metric GAN for monaural speech enhancement in the time domain

被引:13
作者
Guo, Huimin [1 ,2 ]
Jian, Haifang [1 ]
Wang, Yequan [3 ]
Wang, Hongchang [1 ,2 ]
Zhao, Xiaofan [3 ]
Zhu, Wenqi [4 ]
Cheng, Qinghua [1 ,2 ]
机构
[1] Chinese Acad Sci, Inst Semicond, Lab Solid State Optoelect Informat Technol, Beijing 100083, Peoples R China
[2] Univ Chinese Acad Sci, Beijing 100049, Peoples R China
[3] Beijing Acad Artificial Intelligence, Beijing 100089, Peoples R China
[4] North China Univ Technol, Sch Informat Sci & Technol, Beijing 100144, Peoples R China
关键词
Speech enhancement; Time domain; Multiscale attention; Attention metric discriminator; RECURRENT NEURAL-NETWORK; SELF-ATTENTION; U-NET; NOISE;
D O I
10.1016/j.apacoust.2023.109385
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
In the speech enhancement (SE) task, the mismatch between the objective function used to train the SE model, and the evaluation metric will lead to the low quality of the generated speech. Although existing studies have attempted to use the metric discriminator to learn the alternative function of evaluation metric from data to guide generator updates, the metric discriminator's simple structure cannot better approximate the function of the evaluation metric, thus limiting the performance of SE. This paper proposes a multiscale attention metric generative adversarial network (MAMGAN) to resolve this problem. In the metric discriminator, the attention mechanism is introduced to emphasize the meaningful features of spatial direction and channel direction to avoid the feature loss caused by direct average pooling to better approximate the calculation of the evaluation metric and further improve SE's performance. In addition, driven by the effectiveness of the self-attention mechanism in capturing long-term dependence, we construct a multiscale attention module (MSAM). It fully considers the multiple representations of signals, which can better model the features of long sequences. The ablation experiment verifies the effectiveness of the attention metric discriminator and the MSAM. Quantitative analysis on the Voice Bank + DEMAND dataset shows that MAMGAN outperforms various time-domain SE methods with a 3.30 perceptual evaluation of speech quality score.
引用
收藏
页数:11
相关论文
共 55 条
[1]   A model distance maximizing framework for speech recognizer-based speech enhancement [J].
BabaAli, Bagher ;
Sameti, Hossein ;
Falk, Tiago H. .
AEU-INTERNATIONAL JOURNAL OF ELECTRONICS AND COMMUNICATIONS, 2011, 65 (02) :99-106
[2]  
Baby D, 2019, INT CONF ACOUST SPEE, P106, DOI [10.1109/ICASSP.2019.8683799, 10.1109/icassp.2019.8683799]
[3]  
Berouti M., 1979, ICASSP 79. 1979 IEEE International Conference on Acoustics, Speech and Signal Processing, P208
[4]  
Botinhao C. V., 2016, SSW, P159
[5]   Speech enhancement using a mixture-maximum model [J].
Burshtein, D ;
Gannot, S .
IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, 2002, 10 (06) :341-351
[6]  
Cao RZ, 2024, Arxiv, DOI arXiv:2203.15149
[7]   Noise spectrum estimation in adverse environments: Improved minima controlled recursive averaging [J].
Cohen, I .
IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, 2003, 11 (05) :466-475
[8]   Speech spectral modeling and enhancement based on autoregressive conditional heteroscedasticity models [J].
Cohen, I .
SIGNAL PROCESSING, 2006, 86 (04) :698-709
[9]  
Defossez A., 2020, arXiv
[10]  
Desplanques B, 2020, Arxiv, DOI arXiv:2005.07143