Efficient Integrated Features Based on Pre-trained Models for Speaker Verification

被引:0
|
作者
Li, Yishuang [1 ,2 ]
Guan, Wenhao [3 ]
Huang, Hukai [3 ]
Miao, Shiyu [2 ]
Su, Qi [2 ]
Li, Lin [1 ,2 ]
Hong, Qingyang [3 ]
机构
[1] Xiamen Univ, Inst Artificial Intelligence, Xian, Peoples R China
[2] Xiamen Univ, Sch Elect Sci & Engn, Xian, Peoples R China
[3] Xiamen Univ, Sch Informat, Xian, Peoples R China
来源
INTERSPEECH 2024 | 2024年
基金
中国国家自然科学基金;
关键词
speaker verification; pre-trained models; feature integration; t-SNE; SPEECH;
D O I
10.21437/Interspeech.2024-1889
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Previous work has explored the application of pre-trained models (PTMs) in speaker verification(SV). Most researchers directly replaced handcrafted features with the universal representations of the PTMs, and jointly fine-tuned PTMs with the downstream SV networks, which undoubtedly discarded important spectral information contained in handcrafted features and also increased the training cost. In this paper, we proposed an efficient feature integration method that utilized a Fine-grained Fusion Module to fuse the multi-layer representations of the PTMs adaptively. Then we integrated the fused representations with handcrafted features to obtain the integrated features, which were subsequently fed into the SV network. The experimental results demonstrated that using the integrated features effectively enhanced the performance of the SV systems, and yielded decent results with no need to fine-tune the PTMs. Moreover, employing full-parameter fine-tuning led to the current optimal results.
引用
收藏
页码:2140 / 2144
页数:5
相关论文
共 50 条
  • [21] BARTpho: Pre-trained Sequence-to-Sequence Models for Vietnamese
    Nguyen Luong Tran
    Duong Minh Le
    Dat Quoc Nguyen
    INTERSPEECH 2022, 2022, : 1751 - 1755
  • [22] Performance Evaluation of Pre-trained Models in Sarcasm Detection Task
    Wang, Haiyang
    Song, Xin
    Zhou, Bin
    Wang, Ye
    Gao, Liqun
    Jia, Yan
    WEB INFORMATION SYSTEMS ENGINEERING - WISE 2021, PT II, 2021, 13081 : 67 - 75
  • [23] GAIT analysis based on GENDER detection using pre-trained models and tune parameters
    Vora C.
    Katkar V.
    Lunagaria M.
    Discover Artificial Intelligence, 2024, 4 (01):
  • [24] Multi-Label Conditional Generation From Pre-Trained Models
    Proszewska, Magdalena
    Wolczyk, Maciej
    Zieba, Maciej
    Wielopolski, Patryk
    Maziarka, Lukasz
    Smieja, Marek
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (09) : 6185 - 6198
  • [25] Automatic Detection of Liver Cancer Using Hybrid Pre-Trained Models
    Othman, Esam
    Mahmoud, Muhammad
    Dhahri, Habib
    Abdulkader, Hatem
    Mahmood, Awais
    Ibrahim, Mina
    SENSORS, 2022, 22 (14)
  • [26] Diet Code Is Healthy: Simplifying Programs for Pre-trained Models of Code
    Zhang, Zhaowei
    Zhang, Hongyu
    Shen, Beijun
    Gu, Xiaodong
    PROCEEDINGS OF THE 30TH ACM JOINT MEETING EUROPEAN SOFTWARE ENGINEERING CONFERENCE AND SYMPOSIUM ON THE FOUNDATIONS OF SOFTWARE ENGINEERING, ESEC/FSE 2022, 2022, : 1073 - 1084
  • [27] SecretGen: Privacy Recovery on Pre-trained Models via Distribution Discrimination
    Yuan, Zhuowen
    Wu, Fan
    Long, Yunhui
    Xiao, Chaowei
    Li, Bo
    COMPUTER VISION - ECCV 2022, PT V, 2022, 13665 : 139 - 155
  • [28] An analysis of pre-trained stable diffusion models through a semantic lens
    Bonechi, Simone
    Andreini, Paolo
    Corradini, Barbara Toniella
    Scarselli, Franco
    NEUROCOMPUTING, 2025, 614
  • [29] Detecting COVID-19 in chest CT images based on several pre-trained models
    Hassan, Esraa
    Shams, Mahmoud Y.
    Hikal, Noha A.
    Elmougy, Samir
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (24) : 65267 - 65287
  • [30] Graph-Based Audio Classification Using Pre-Trained Models and Graph Neural Networks
    Castro-Ospina, Andres Eduardo
    Solarte-Sanchez, Miguel Angel
    Vega-Escobar, Laura Stella
    Isaza, Claudia
    Martinez-Vargas, Juan David
    SENSORS, 2024, 24 (07)