Cross-lingual distillation for domain knowledge transfer with sentence transformers

被引:0
|
作者
Piperno, Ruben [1 ,2 ]
Bacco, Luca [1 ,3 ]
Dell'Orletta, Felice [1 ]
Merone, Mario [2 ]
Pecchia, Leandro [2 ,4 ]
机构
[1] Inst Computat Linguist Antonio Zampolli, Natl Res Council, ItaliaNLP Lab, Via Giuseppe Moruzzi, 1, I-56124 Pisa, Italy
[2] Univ Campus Biomed Roma, Dept Engn, Res Unit Intelligent Technol Hlth & Wellbeing, Via Alvaro Portillo 21, I-00128 Rome, Italy
[3] Univ Campus Biomed Roma, Dept Engn, Res Unit Comp Syst & Bioinformat, Via Alvaro Portillo 21, I-00128 Rome, Italy
[4] Fdn Policlin Univ Campus Biomed Roma, Via Alvaro del Portillo 200, I-00128 Rome, Italy
关键词
Cross-lingual learning; Knowledge distillation; Sentence transformers; Biomedical domain; Domain adaptation;
D O I
10.1016/j.knosys.2025.113079
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent advancements in Natural Language Processing (NLP) have substantially enhanced language understanding. However, non-English languages, especially in specialized and low-resource domains like biomedicine, remain largely underrepresented. Bridging this gap is essential for promoting inclusivity and expanding the global applicability of NLP technologies. This study presents a cross-lingual knowledge distillation framework that utilizes sentence transformers to improve domain-specific NLP capabilities in non-English languages. Specifically, the framework focuses on biomedical text classification tasks. By aligning sentence embeddings between a teacher model trained on English biomedical corpora and a multilingual student model, the proposed method effectively transfers both domain-specific and task-specific knowledge. This alignment allows the student model to efficiently process and adapt to biomedical texts in Spanish, French, and German, particularly in low-resource settings with limited tuning data. Extensive experiments with domain-adapted models like BioBERT and multilingual BERT with machine-translated text pairs demonstrate substantial performance improvements in downstream biomedical NLP tasks. The proposed framework proves highly effective in scenarios characterized by limited training data availability. The results highlight the scalability and effectiveness of this approach, facilitating the development of robust multilingual models tailored to the biomedical domain, thus advancing global accessibility and impact in biomedical NLP applications.
引用
收藏
页数:10
相关论文
共 50 条
  • [21] Cross-lingual Adaptation Using Universal Dependencies
    Taghizadeh, Nasrin
    Faili, Heshaam
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2021, 20 (04)
  • [22] Cross-domain recommendation via knowledge distillation
    Li, Xiuze
    Huang, Zhenhua
    Wu, Zhengyang
    Wang, Changdong
    Chen, Yunwen
    KNOWLEDGE-BASED SYSTEMS, 2025, 311
  • [23] Cross-lingual Adaptation for Recipe Retrieval with Mixup
    Zhu, Bin
    Ngo, Chong-Wah
    Chen, Jingjing
    Chan, Wing-Kwong
    PROCEEDINGS OF THE 2022 INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2022, 2022, : 258 - 267
  • [24] Cross-lingual learning for text processing: A survey
    Pikuliak, Matus
    Simko, Marian
    Bielikova, Maria
    EXPERT SYSTEMS WITH APPLICATIONS, 2021, 165
  • [25] I2KD-SLU: An Intra-Inter Knowledge Distillation Framework for Zero-Shot Cross-Lingual Spoken Language Understanding
    Mao, Tianjun
    Zhang, Chenghong
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT VIII, 2023, 14261 : 345 - 356
  • [26] Exploiting Cross-Domain And Cross-Lingual Ultrasound Tongue Imaging Features For Elderly And Dysarthric Speech Recognition
    Hu, Shujie
    Xie, Xurong
    Geng, Mengzhe
    Cui, Mingyu
    Deng, Jiajun
    Li, Guinan
    Wang, Tianzi
    Meng, Helen
    Liu, Xunying
    INTERSPEECH 2023, 2023, : 2313 - 2317
  • [27] Cross-Domain Deepfake Detection Based on Latent Domain Knowledge Distillation
    Wang, Chunpeng
    Meng, Lingshan
    Xia, Zhiqiu
    Ren, Na
    Ma, Bin
    IEEE SIGNAL PROCESSING LETTERS, 2025, 32 : 896 - 900
  • [28] Transductive Representation Learning for Cross-Lingual Text Classification
    Guo, Yuhong
    Xiao, Min
    12TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2012), 2012, : 888 - 893
  • [29] DISENTANGLED SPEAKER AND LANGUAGE REPRESENTATIONS USING MUTUAL INFORMATION MINIMIZATION AND DOMAIN ADAPTATION FOR CROSS-LINGUAL TTS
    Xin, Detai
    Komatsu, Tatsuya
    Takamichi, Shinnosuke
    Saruwatari, Hiroshi
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 6608 - 6612
  • [30] CROSS LINGUAL TRANSFER LEARNING FOR ZERO-RESOURCE DOMAIN ADAPTATION
    Abad, Alberto
    Bell, Peter
    Carmantini, Andrea
    Renals, Steve
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 6909 - 6913