ADAPTABLE MULTI-DOMAIN LANGUAGE MODEL FOR TRANSFORMER ASR

被引:4
|
作者
Lee, Taewoo [1 ]
Lee, Min-Joong [2 ]
Kang, Tae Gyoon [2 ]
Jung, Seokyeoung [1 ]
Kwon, Minseok [1 ]
Hong, Yeona [1 ]
Lee, Jungin [1 ]
Woo, Kyoung-Gu [1 ]
Kim, Ho-Gyeong [2 ]
Jeong, Jiseung [2 ]
Lee, Jihyun [2 ]
Lee, Hosik [2 ]
Choi, Young Sang [2 ]
机构
[1] Samsung Elect, AI R&D Grp, Suwon Shi, South Korea
[2] Samsung Elect, Samsung Adv Inst Technol, Suwon Shi, South Korea
来源
2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021) | 2021年
关键词
End-to-end (E2E) automatic speech recognition (ASR); language model (LM); multi-domain adaptation;
D O I
10.1109/ICASSP39728.2021.9413475
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
We propose an adapter based multi-domain Transformer based language model (LM) for Transformer ASR. The model consists of a big size common LM and small size adapters. The model can perform multi-domain adaptation with only the small size adapters and its related layers. The proposed model can reuse the full fine-tuned LM which is fine-tuned using all layers of an original model. The proposed LM can be expanded to new domains by adding about 2% of parameters for a first domain and 13% parameters for after second domain. The proposed model is also effective in reducing the model maintenance cost because it is possible to omit the costly and time-consuming common LM pre-training process. Using proposed adapter based approach, we observed that a general LM with adapter can outperform a dedicated music domain LM in terms of word error rate (WER).
引用
收藏
页码:7358 / 7362
页数:5
相关论文
共 10 条
  • [1] Adversarial Training for Multi-domain Speaker Recognition
    Wang, Qing
    Rao, Wei
    Guo, Pengcheng
    Xie, Lei
    2021 12TH INTERNATIONAL SYMPOSIUM ON CHINESE SPOKEN LANGUAGE PROCESSING (ISCSLP), 2021,
  • [2] Multi-domain adaptation for regression under conditional distribution shift
    Taghiyarrenani, Zahra
    Nowaczyk, Slawomir
    Pashami, Sepideh
    Bouguelia, Mohamed-Rafik
    EXPERT SYSTEMS WITH APPLICATIONS, 2023, 224
  • [3] A multi-domain adaptive neural machine translation method based on domain data balancer
    Xu, Jinlei
    Wen, Yonghua
    Huang, Shuanghong
    Yu, Zhengtao
    INTELLIGENT DATA ANALYSIS, 2024, 28 (03) : 685 - 698
  • [4] Multi-domain adaptation for named entity recognition with multi-aspect relevance learning
    Li, Jiarui
    Liu, Jian
    Chen, Yufeng
    Xu, Jinan
    LANGUAGE RESOURCES AND EVALUATION, 2023, 57 (02) : 803 - 818
  • [5] Multi-domain adaptation for named entity recognition with multi-aspect relevance learning
    Jiarui Li
    Jian Liu
    Yufeng Chen
    Jinan Xu
    Language Resources and Evaluation, 2023, 57 : 803 - 818
  • [6] A Multi-domain Adaptation for Sentiment Classification Algorithm Based on Class Distribution
    Hu, Kongbing
    Zhang, Yuhong
    Hu, Xuegang
    2012 IEEE INTERNATIONAL CONFERENCE ON GRANULAR COMPUTING (GRC 2012), 2012, : 179 - 184
  • [7] Effective domain awareness and adaptation approach via mask substructure for multi-domain neural machine translation
    Huang, Shuanghong
    Guo, Junjun
    Yu, Zhengtao
    Wen, Yonghua
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (19): : 14047 - 14060
  • [8] Effective domain awareness and adaptation approach via mask substructure for multi-domain neural machine translation
    Shuanghong Huang
    Junjun Guo
    Zhengtao Yu
    Yonghua Wen
    Neural Computing and Applications, 2023, 35 : 14047 - 14060
  • [9] Self-attention StarGAN for Multi-domain Image-to-Image Translation
    He, Ziliang
    Yang, Zhenguo
    Mao, Xudong
    Lv, Jianming
    Li, Qing
    Liu, Wenyin
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2019: IMAGE PROCESSING, PT III, 2019, 11729 : 537 - 549
  • [10] Learning Domain Specific Sub-layer Latent Variable for Multi-domain Adaptation Neural Machine Translation
    Huang, Shuanghong
    Feng, Chong
    Shi, Ge
    Li, Zhengjun
    Zhao, Xuan
    Li, Xinyan
    Wang, Xiaomei
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2024, 23 (06)