Fully Quantized Transformer for Machine Translation

被引:0
|
作者
Prato, Gabriele [1 ]
Charlaix, Ella [2 ]
Rezagholizadeh, Mehdi [2 ]
机构
[1] Univ Montreal, Mila, Montreal, PQ, Canada
[2] Huawei Noahs Ark Lab, Montreal, PQ, Canada
来源
FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020 | 2020年
关键词
NEURAL-NETWORKS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
State-of-the-art neural machine translation methods employ massive amounts of parameters. Drastically reducing computational costs of such methods without affecting performance has been up to this point unsuccessful. To this end, we propose FullyQT: an allinclusive quantization strategy for the Transformer. To the best of our knowledge, we are the first to show that it is possible to avoid any loss in translation quality with a fully quantized Transformer. Indeed, compared to fullprecision, our 8-bit models score greater or equal BLEU on most tasks. Comparing ourselves to all previously proposed methods, we achieve state-of-the-art quantization results.
引用
收藏
页码:1 / 14
页数:14
相关论文
共 50 条
  • [41] The fully visible Boltzmann machine and the Senate of the 45th Australian Parliament in 2016
    Bagnall, Jessica J.
    Jones, Andrew T.
    Karavarsamis, Natalie
    Nguyen, Hien D.
    JOURNAL OF COMPUTATIONAL SOCIAL SCIENCE, 2020, 3 (01): : 55 - 81
  • [42] The fully visible Boltzmann machine and the Senate of the 45th Australian Parliament in 2016
    Jessica J. Bagnall
    Andrew T. Jones
    Natalie Karavarsamis
    Hien D. Nguyen
    Journal of Computational Social Science, 2020, 3 : 55 - 81
  • [43] Evolving Fully Automated Machine Learning via Life-Long Knowledge Anchors
    Zheng, Xiawu
    Zhang, Yang
    Hong, Sirui
    Li, Huixia
    Tang, Lang
    Xiong, Youcheng
    Zhou, Jin
    Wang, Yan
    Sun, Xiaoshuai
    Zhu, Pengfei
    Wu, Chenglin
    Ji, Rongrong
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2021, 43 (09) : 3091 - 3107
  • [44] A Hybrid machine-learning method for oil-immersed power transformer fault diagnosis
    Yang, Xiaohui
    Chen, Wenkai
    Li, Anyi
    Yang, Chunsheng
    IEEJ TRANSACTIONS ON ELECTRICAL AND ELECTRONIC ENGINEERING, 2020, 15 (04) : 501 - 507
  • [45] Using artificial intelligence to develop a machine translation system and teaching resources in the Tuvan language
    Novikova, Marina L.
    Novikov, Philipp N.
    NOVYE ISSLEDOVANIYA TUVY-NEW RESEARCH OF TUVA, 2024, (01): : 6 - 17
  • [46] Towards Explainable Formal Methods: From LTL to Natural Language with Neural Machine Translation
    Cherukuri, Himaja
    Ferrari, Alessio
    Spoletini, Paola
    REQUIREMENTS ENGINEERING: FOUNDATION FOR SOFTWARE QUALITY, REFSQ 2022, 2022, 13216 : 79 - 86
  • [47] A Survey of Deep Learning Architectures for Privacy-Preserving Machine Learning With Fully Homomorphic Encryption
    Podschwadt, Robert
    Takabi, Daniel
    Hu, Peizhao
    Rafiei, Mohammad H. H.
    Cai, Zhipeng
    IEEE ACCESS, 2022, 10 : 117477 - 117500
  • [48] Predictor-Estimator: Neural Quality Estimation Based on Target Word Prediction for Machine Translation
    Kim, Hyun
    Jung, Hun-Young
    Kwon, Hongseok
    Lee, Jong-Hyeok
    Na, Seung-Hoon
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2017, 17 (01)
  • [49] Feature Extraction Methods for Binary Code Similarity Detection Using Neural Machine Translation Models
    Ito, Norimitsu
    Hashimoto, Masaki
    Otsuka, Akira
    IEEE ACCESS, 2023, 11 : 102796 - 102805
  • [50] Trace2trace-A Feasibility Study on Neural Machine Translation Applied to Human Motion Trajectories
    Crivellari, Alessandro
    Beinat, Euro
    SENSORS, 2020, 20 (12) : 1 - 15