Fully Quantized Transformer for Machine Translation

被引:0
|
作者
Prato, Gabriele [1 ]
Charlaix, Ella [2 ]
Rezagholizadeh, Mehdi [2 ]
机构
[1] Univ Montreal, Mila, Montreal, PQ, Canada
[2] Huawei Noahs Ark Lab, Montreal, PQ, Canada
来源
FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020 | 2020年
关键词
NEURAL-NETWORKS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
State-of-the-art neural machine translation methods employ massive amounts of parameters. Drastically reducing computational costs of such methods without affecting performance has been up to this point unsuccessful. To this end, we propose FullyQT: an allinclusive quantization strategy for the Transformer. To the best of our knowledge, we are the first to show that it is possible to avoid any loss in translation quality with a fully quantized Transformer. Indeed, compared to fullprecision, our 8-bit models score greater or equal BLEU on most tasks. Comparing ourselves to all previously proposed methods, we achieve state-of-the-art quantization results.
引用
收藏
页码:1 / 14
页数:14
相关论文
共 50 条
  • [31] Clinical translation of nanomedicine with integrated digital medicine and machine learning interventions
    Ahmad, Farooq
    Muhmood, Tahir
    COLLOIDS AND SURFACES B-BIOINTERFACES, 2024, 241
  • [32] A Comparative Study of Power Transformer Winding Fault Diagnosis Using Machine Learning Algorithms
    Dlamini, G. A. Z.
    Thango, B. A.
    Bokoro, P. N.
    2024 32ND SOUTHERN AFRICAN UNIVERSITIES POWER ENGINEERING CONFERENCE, SAUPEC, 2024, : 26 - 30
  • [33] Named Entity Correction in Neural Machine Translation Using the Attention Alignment Map
    Lee, Jangwon
    Lee, Jungi
    Lee, Minho
    Jang, Gil-Jin
    APPLIED SCIENCES-BASEL, 2021, 11 (15):
  • [34] Context-Aware Linguistic Steganography Model Based on Neural Machine Translation
    Ding, Changhao
    Fu, Zhangjie
    Yang, Zhongliang
    Yu, Qi
    Li, Daqiu
    Huang, Yongfeng
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 868 - 878
  • [35] Overcoming Language Barriers in Academia: Machine Translation Tools and a Vision for a Multilingual Future
    Steigerwald, Emma
    Ramirez-Castaneda, Valeria
    Brandt, Debora Y. C.
    Baldi, Andras
    Shapiro, Julie Teresa
    Bowker, Lynne
    Tarvin, Rebecca D.
    BIOSCIENCE, 2022, 72 (10) : 988 - 998
  • [36] Language Model Method for Collocation Rules of Parts of Speech in Machine Translation System
    Liu, Jinhui
    Zhang, Feng
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2024, 23 (08) : 18 - 18
  • [37] Improving Neural Machine Translation with Linear Interpolation of a Short-Path Unit
    Li, Yachao
    Li, Junhui
    Zhang, Min
    Li, Yixin
    Zou, Peng
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2020, 19 (03)
  • [38] Efficient Embedded Decoding of Neural Network Language Models in a Machine Translation System
    Zamora-Martinez, Francisco
    Jose Castro-Bleda, Maria
    INTERNATIONAL JOURNAL OF NEURAL SYSTEMS, 2018, 28 (09)
  • [39] Coarse-to-Fine Output Predictions for Efficient Decoding in Neural Machine Translation
    Chen, Qi
    Kwong, Oi Yee
    Li, Yinqiao
    Xiao, Tong
    Zhu, Jingbo
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2022, 21 (06)
  • [40] MaxSD: A Neural Machine Translation Evaluation Metric Optimized by Maximizing Similarity Distance
    Ma, Qingsong
    Meng, Fandong
    Zheng, Daqi
    Wang, Mingxuan
    Graham, Yvette
    Jiang, Wenbin
    Liu, Qun
    NATURAL LANGUAGE UNDERSTANDING AND INTELLIGENT APPLICATIONS (NLPCC 2016), 2016, 10102 : 153 - 161