Unconstrained vocal pattern recognition algorithm based on attention mechanism

被引:1
|
作者
Li, Yaqian [1 ]
Zhang, Xiaolong [2 ]
Zhang, Xuyao [3 ]
Li, Haibin [1 ]
Zhang, Wenming [4 ]
机构
[1] Yanshan Univ, Pattern Recognized, Elect Engn, Qinhuangdao, Hebei, Peoples R China
[2] Yanshan Univ, Speaker Diarizat, Elect Engn, Qinhuangdao, Hebei, Peoples R China
[3] Yanshan Univ, Speaker Verificat, Elect Engn, Qinhuangdao, Hebei, Peoples R China
[4] Yanshan Univ, Camera Calibrat, Elect Engn, Qinhuangdao, Hebei, Peoples R China
基金
中国国家自然科学基金;
关键词
Voiceprint recognition; Unconstrained datasets; Attention mechanism; Feature fusion;
D O I
10.1016/j.dsp.2023.103973
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Deep learning-based voiceprint recognition methods rely heavily on adequate datasets, especially those closer to the natural environment and more complex under unconstrained conditions. Yet, the data types of open-source speech datasets are too homogeneous nowadays, and there are some differences with the address collected in natural application environments. For few Chinese datasets used, this paper proposes and produces an unconstrained Chinese speech dataset with richer data types closer to those collected in a natural environment. To address the inadequate extraction of acoustic features in the unconstrained speech dataset, a new two-dimensional convolutional residual network structure based on the attention mechanism is designed and applied to acoustic feature extraction. The residual block structure in the residual network is improved by the SE module and the CBAM module to obtain the SE-Cov2d and CSA-Cov2d models respectively. Finally, it is experimentally demonstrated that the attention mechanism can help the network focus on more critical feature information and fuse more differentiated features in feature extraction. (c) 2023 Elsevier Inc. All rights reserved.
引用
收藏
页数:8
相关论文
共 50 条
  • [21] An Illegal Dress Recognition Algorithm Based on Key-point Detection and Attention Mechanism and Its Application
    He G.
    Qi D.
    Yan Y.
    Zhongguo Dianji Gongcheng Xuebao/Proceedings of the Chinese Society of Electrical Engineering, 2022, 42 (05): : 1826 - 1836
  • [22] Improved U-Net robust weld seam recognition algorithm based on integrating attention mechanism
    Zhou, Siyu
    Liu, Shuaishi
    Yang, Hongtao
    Song, Yihu
    Jisuanji Jicheng Zhizao Xitong/Computer Integrated Manufacturing Systems, CIMS, 2025, 31 (01): : 135 - 146
  • [23] Adaptive Graph Convolutional 3D Point Cloud Recognition Algorithm Based on Attention Mechanism
    Ma, Yuan
    She, Li-Huang
    Li, Jia-Wei
    Bao, Xi-Rong
    Dongbei Daxue Xuebao/Journal of Northeastern University, 2024, 45 (06): : 786 - 792
  • [24] An incomplete fingerprint recognition algorithm based on pattern entropy
    Zhang, Jie
    Jing, Xiao-Jun
    Liu, Xin-Jing
    Li, Shuai
    Dianzi Yu Xinxi Xuebao/Journal of Electronics and Information Technology, 2012, 34 (12): : 3040 - 3045
  • [25] Single Image Defogging Algorithm Based on Attention Mechanism
    Cao Ruihu
    Zhang Pengchao
    Wang Lei
    Zhang Fan
    Kang Jie
    LASER & OPTOELECTRONICS PROGRESS, 2023, 60 (02)
  • [26] Deep Collaborative Recommendation Algorithm Based on Attention Mechanism
    Cui, Can
    Qin, Jiwei
    Ren, Qiulin
    APPLIED SCIENCES-BASEL, 2022, 12 (20):
  • [27] Occluded Pedestrian Detection Algorithm Based on Attention Mechanism
    Zou Ziyin
    Gai Shaoyan
    Da Feipeng
    Li Yu
    ACTA OPTICA SINICA, 2021, 41 (15)
  • [28] An Image Captioning Algorithm Based on Combination Attention Mechanism
    Liu, Jinlong
    Cheng, Kangda
    Jin, Haiyan
    Wu, Zhilu
    ELECTRONICS, 2022, 11 (09)
  • [29] A Semantic Segmentation Algorithm Based on Improved Attention Mechanism
    Chen, Chunyu
    Wu, Xinsheng
    Chen, An
    2020 INTERNATIONAL SYMPOSIUM ON AUTONOMOUS SYSTEMS (ISAS), 2020, : 244 - 248
  • [30] Algorithm for Skeleton Action Recognition by Integrating Attention Mechanism and Convolutional Neural Networks
    Liu, Jianhua
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2023, 14 (08) : 604 - 613