Voice Deepfake Detection Using the Self-Supervised Pre-Training Model HuBERT

被引:3
|
作者
Li, Lanting [1 ]
Lu, Tianliang [1 ]
Ma, Xingbang [1 ]
Yuan, Mengjiao [1 ]
Wan, Da [1 ]
机构
[1] Peoples Publ Secur Univ China, Coll Informat & Cyber Secur, Beijing 100038, Peoples R China
来源
APPLIED SCIENCES-BASEL | 2023年 / 13卷 / 14期
关键词
voice deepfake detection; self-supervised learning; pre-training; feature map scaling; anti-spoofing;
D O I
10.3390/app13148488
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
In recent years, voice deepfake technology has developed rapidly, but current detection methods have the problems of insufficient detection generalization and insufficient feature extraction for unknown attacks. This paper presents a forged speech detection method (HuRawNet2_modified) based on a self-supervised pre-trained model (HuBERT) to improve detection (and address the above problems). A combination of impulsive signal-dependent additive noise and additive white Gaussian noise was adopted for data boosting and augmentation, and the HuBERT model was fine-tuned on different language databases. On this basis, the size of the extracted feature maps was modified independently by the & alpha;-feature map scaling (& alpha;-FMS) method, with a modified end-to-end method using the RawNet2 model as the backbone structure. The results showed that the HuBERT model could extract features more comprehensively and accurately. The best evaluation indicators were an equal error rate (EER) of 2.89% and a minimum tandem detection cost function (min t-DCF) of 0.2182 on the database of the ASVspoof2021 LA challenge, which verified the effectiveness of the detection method proposed in this paper. Compared with the baseline systems in databases of the ASVspoof 2021 LA challenge and the FMFCC-A, the values of EER and min t-DCF decreased. The results also showed that the self-supervised pre-trained model with fine-tuning can extract acoustic features across languages. And the detection can be slightly improved when the languages of the pre-trained database, and the fine-tuned and tested database are the same.
引用
收藏
页数:15
相关论文
共 50 条
  • [31] Self-supervised pre-training improves fundus image classification for diabetic retinopathy
    Lee, Joohyung
    Lee, Eung-Joo
    REAL-TIME IMAGE PROCESSING AND DEEP LEARNING 2022, 2022, 12102
  • [32] Mutual information-driven self-supervised point cloud pre-training
    Xu, Weichen
    Fu, Tianhao
    Cao, Jian
    Zhao, Xinyu
    Xu, Xinxin
    Cao, Xixin
    Zhang, Xing
    KNOWLEDGE-BASED SYSTEMS, 2025, 307
  • [33] A SELF-SUPERVISED PRE-TRAINING FRAMEWORK FOR VISION-BASED SEIZURE CLASSIFICATION
    Hou, Jen-Cheng
    McGonigal, Aileen
    Bartolomei, Fabrice
    Thonnat, Monique
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 1151 - 1155
  • [34] SDCluster: A clustering based self-supervised pre-training method for semantic segmentation of remote sensing images
    Xu, Hanwen
    Zhang, Chenxiao
    Yue, Peng
    Wang, Kaixuan
    ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2025, 223 : 1 - 14
  • [35] Abdominal Organs and Pan-Cancer Segmentation Based on Self-supervised Pre-training and Self-training
    Li, He
    Han, Meng
    Wang, Guotai
    FAST, LOW-RESOURCE, AND ACCURATE ORGAN AND PAN-CANCER SEGMENTATION IN ABDOMEN CT, FLARE 2023, 2024, 14544 : 130 - 142
  • [36] Self-Supervised Pre-Training with Bridge Neural Network for SAR-Optical Matching
    Qian, Lixin
    Liu, Xiaochun
    Huang, Meiyu
    Xiang, Xueshuang
    REMOTE SENSING, 2022, 14 (12)
  • [37] Single-atom catalysts property prediction via Supervised and Self-Supervised pre-training models
    Wang, Lanjing
    Chen, Honghao
    Yang, Longqi
    Li, Jiali
    Li, Yong
    Wang, Xiaonan
    CHEMICAL ENGINEERING JOURNAL, 2024, 487
  • [38] Self-Supervised Pre-Training for Deep Image Prior-Based Robust PET Image Denoising
    Onishi, Yuya
    Hashimoto, Fumio
    Ote, Kibo
    Matsubara, Keisuke
    Ibaraki, Masanobu
    IEEE TRANSACTIONS ON RADIATION AND PLASMA MEDICAL SCIENCES, 2024, 8 (04) : 348 - 356
  • [39] PerFedRec plus plus : Enhancing Personalized Federated Recommendation with Self-Supervised Pre-Training
    Luo, Sichun
    Xiao, Yuanzhang
    Zhang, Xinyi
    Liu, Yang
    Ding, Wenbo
    Song, Linqi
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2024, 15 (05)
  • [40] Multi-modal cross-domain self-supervised pre-training for fMRI and EEG fusion
    Wei, Xinxu
    Zhao, Kanhao
    Jiao, Yong
    Carlisle, Nancy B.
    Xie, Hua
    Fonzo, Gregory A.
    Zhang, Yu
    NEURAL NETWORKS, 2025, 184