Converting Continuous-Space Language Models into N-gram Language Models with Efficient Bilingual Pruning for Statistical Machine Translation

被引:4
|
作者
Wang, Rui [1 ]
Utiyama, Masao [2 ]
Goto, Isao [3 ,4 ]
Sumita, Eiichiro [2 ]
Zhao, Hai [1 ,5 ]
Lu, Bao-Liang [1 ,5 ]
机构
[1] Shanghai Jiao Tong Univ, Ctr Brain Like Comp & Machine Intelligence, Dept Comp Sci & Engn, 800 Dongchuan Rd, Shanghai 200240, Peoples R China
[2] Natl Inst Informat & Commun Technol, Multilingual Translat Lab, 3-5 Hikaridai, Kyoto 6190289, Japan
[3] NHK Japan Broadcasting Corp, Sci & Technol Res Labs, Setagaya Ku, 1-10-11 Kinuta, Tokyo 1578510, Japan
[4] Natl Inst Informat & Commun Technol, Kyoto 6190289, Japan
[5] Shanghai Jiao Tong Univ, Key Lab Shanghai Educ Commiss Intelligent Interac, 800 Dongchuan Rd, Shanghai 200240, Peoples R China
基金
中国国家自然科学基金;
关键词
Machine translation; continuous-space language model; neural network language model; language model pruning;
D O I
10.1145/2843942
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The Language Model (LM) is an essential component of Statistical Machine Translation (SMT). In this article, we focus on developing efficient methods for LM construction. Our main contribution is that we propose a Natural N-grams based Converting (NNGC) method for transforming a Continuous-Space Language Model (CSLM) to a Back-off N-gram Language Model (BNLM). Furthermore, a Bilingual LM Pruning (BLMP) approach is developed for enhancing LMs in SMT decoding and speeding up CSLM converting. The proposed pruning and converting methods can convert a large LM efficiently by working jointly. That is, a LM can be effectively pruned before it is converted from CSLM without sacrificing performance, and further improved if an additional corpus contains out-of-domain information. For different SMT tasks, our experimental results indicate that the proposed NNGC and BLMP methods outperform the existing counterpart approaches significantly in BLEU and computational cost.
引用
收藏
页数:26
相关论文
共 16 条
  • [1] Bilingual Continuous-Space Language Model Growing for Statistical Machine Translation
    Wang, Rui
    Zhao, Hai
    Lu, Bao-Liang
    Utiyama, Masao
    Sumita, Eiichiro
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2015, 23 (07) : 1209 - 1220
  • [2] Discovering Finance Keywords via Continuous-Space Language Models
    Tsai, Ming-Feng
    Wang, Chuan-Ju
    Chien, Po-Chuan
    ACM TRANSACTIONS ON MANAGEMENT INFORMATION SYSTEMS, 2016, 7 (03)
  • [3] Evaluating n-gram Models for a Bilingual Word Sense Disambiguation Task
    Pinto, David
    Vilarino, Darnes
    Balderas, Carlos
    Tovar, Mireya
    Beltran, Beatriz
    COMPUTACION Y SISTEMAS, 2011, 15 (02): : 209 - 220
  • [4] Efficient Embedded Decoding of Neural Network Language Models in a Machine Translation System
    Zamora-Martinez, Francisco
    Jose Castro-Bleda, Maria
    INTERNATIONAL JOURNAL OF NEURAL SYSTEMS, 2018, 28 (09)
  • [5] Machine Translation based on Neural Network Language Models
    Zamora-Martinez, Francisco
    Jose Castro-Bleda, Maria
    PROCESAMIENTO DEL LENGUAJE NATURAL, 2010, (45): : 221 - 228
  • [6] Recent Advances in Interactive Machine Translation With Large Language Models
    Wang, Yanshu
    Zhang, Jinyi
    Shi, Tianrong
    Deng, Dashuai
    Tian, Ye
    Matsumoto, Tadahiro
    IEEE ACCESS, 2024, 12 : 179353 - 179382
  • [7] Interactive Multi-System Machine Translation with Neural Language Models
    Rikters, Matiss
    DATABASES AND INFORMATION SYSTEMS IX, 2016, 291 : 269 - 280
  • [8] Joint pairwise learning and masked language models for neural machine translation of English
    Yang, Shuhan
    Yang, Qun
    ARTIFICIAL LIFE AND ROBOTICS, 2025,
  • [9] Overcoming language barriers via machine translation with sparse Mixture-of-Experts fusion of large language models
    Zhu, Shaolin
    Jian, Dong
    Xiong, Deyi
    INFORMATION PROCESSING & MANAGEMENT, 2025, 62 (03)
  • [10] LARGE LANGUAGE MODELS "AD REFERENDUM": HOW GOOD ARE THEY AT MACHINE TRANSLATION IN THE LEGAL DOMAIN?
    Briva-Iglesias, Vicent
    Camargo, Joao Lucas Cavalheiro
    Dogru, Gokhan
    MONTI, 2024, 16 : 75 - 107