Multilingual Pre-training Model-Assisted Contrastive Learning Neural Machine Translation

被引:0
|
作者
Sun, Shuo [1 ]
Hou, Hong-xu [1 ]
Yang, Zong-heng [1 ]
Wang, Yi-song [1 ]
机构
[1] Inner Mongolia Univ, Coll Comp Sci, Natl & Local Joint Engn Res Ctr Intelligent Infor, Inner Mongolia Key Lab Mongolian Informat Proc Te, Hohhot, Peoples R China
来源
2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN | 2023年
关键词
Low-Resource NMT; Pre-training Model; Contrastive Learning; Dynamic Training;
D O I
10.1109/IJCNN54540.2023.10191766
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Since pre-training and fine-tuning have been a successful paradigm in Natural Language Processing (NLP), this paper adopts the SOTA pre-training model-CeMAT as a strong assistant for low-resource ethnic language translation tasks. Aiming at the exposure bias problem in the fine-tuning process, we use the contrastive learning framework and propose a new contrastive examples generation method, which uses self-generated predictions as contrastive examples to expose the model to errors during inference. Moreover, in order to effectively utilize the limited bilingual data in low-resource tasks, this paper proposes a dynamic training strategy to fine-tune the model, and refines the model step by step by taking word embedding norm and uncertainty as the criteria of evaluate data and model respectively. Experimental results demonstrate that our method significantly improves the quality compared to the baselines, which fully verifies the effectiveness.
引用
收藏
页数:7
相关论文
共 50 条
  • [41] Contrastive Pre-training and Representation Distillation for Medical Visual Question Answering Based on Radiology Images
    Liu, Bo
    Zhan, Li-Ming
    Wu, Xiao-Ming
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2021, PT II, 2021, 12902 : 210 - 220
  • [42] Contrastive Pre-training with Multi-level Alignment for Grounded Multimodal Named Entity Recognition
    Bao, Xigang
    Tian, Mengyuan
    Wang, Luyao
    Zha, Zhiyuan
    Qin, Biao
    PROCEEDINGS OF THE 4TH ANNUAL ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2024, 2024, : 795 - 803
  • [43] Multimodal Hate Speech Detection in Memes Using Contrastive Language-Image Pre-Training
    Arya, Greeshma
    Hasan, Mohammad Kamrul
    Bagwari, Ashish
    Safie, Nurhizam
    Islam, Shayla
    Ahmed, Fatima Rayan Awad
    De, Aaishani
    Khan, Muhammad Attique
    Ghazal, Taher M.
    IEEE ACCESS, 2024, 12 : 22359 - 22375
  • [44] RaxCS: Towards cross-language code summarization with contrastive pre-training and retrieval augmentation
    Yang, Kaiyuan
    Wang, Junfeng
    Song, Zihua
    INFORMATION AND SOFTWARE TECHNOLOGY, 2025, 183
  • [45] Multi-Modality Multi-Attribute Contrastive Pre-Training for Image Aesthetics Computing
    Huang, Yipo
    Li, Leida
    Chen, Pengfei
    Wu, Haoning
    Lin, Weisi
    Shi, Guangming
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2025, 47 (02) : 1205 - 1218
  • [46] Cross-Modal Contrastive Pre-Training for Few-Shot Skeleton Action Recognition
    Lu, Mingqi
    Yang, Siyuan
    Lu, Xiaobo
    Liu, Jun
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (10) : 9798 - 9807
  • [47] Neighborhood-enhanced contrast for pre-training graph neural networks
    Yichun Li
    Jin Huang
    Weihao Yu
    Tinghua Zhang
    Neural Computing and Applications, 2024, 36 : 4195 - 4205
  • [48] Contrastive learning-based joint pre-training for unsupervised domain adaptive person re-identificationContrastive learning-based joint pre-training for unsupervised domain adaptive person re-identificationJ. Wang et al.
    Jing Wang
    Xiaohong Li
    Xuesong Dai
    Shuo Zhuang
    Meibin Qi
    Multimedia Systems, 2025, 31 (3)
  • [49] Neighborhood-enhanced contrast for pre-training graph neural networks
    Li, Yichun
    Huang, Jin
    Yu, Weihao
    Zhang, Tinghua
    NEURAL COMPUTING & APPLICATIONS, 2024, 36 (08) : 4195 - 4205
  • [50] Contrastive Learning Based Visual Representation Enhancement for Multimodal Machine Translation
    Wang, Shike
    Zhang, Wen
    Guo, Wenyu
    Yu, Dong
    Liu, Pengyuan
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,