Improving Text Generation with Dynamic Masking and Recovering

被引:0
|
作者
Liu, Zhidong [1 ]
Li, Junhui [1 ]
Zhu, Muhua [2 ]
机构
[1] Soochow Univ, Sch Comp Sci & Technol, Suzhou, Peoples R China
[2] Tencent News, Beijing, Peoples R China
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Due to different types of inputs, diverse text generation tasks may adopt different encoder-decoder frameworks. Thus most existing approaches that aim to improve the robustness of certain generation tasks are input-relevant, and may not work well for other generation tasks. Alternatively, in this paper we present a universal approach to enhance the language representation for text generation on the base of generic encoder-decoder frameworks. This is done from two levels. First, we introduce randomness by randomly masking some percentage of tokens on the decoder side when training the models. In this way, instead of using ground truth history context, we use its corrupted version to predict the next token. Then we propose an auxiliary task to properly recover those masked tokens. Experimental results on several text generation tasks including machine translation (MT), AMR-to-text generation, and image captioning show that the proposed approach can significantly improve over competitive baselines without using any task-specific techniques. This suggests the effectiveness and generality of our proposed approach.
引用
收藏
页码:3878 / 3884
页数:7
相关论文
共 50 条
  • [1] On Improving Text Generation Via Integrating Text Coherence
    Ai, Lisi
    Gao, Baoli
    Zheng, Jianbing
    Gao, Ming
    PROCEEDINGS OF 2019 6TH IEEE INTERNATIONAL CONFERENCE ON CLOUD COMPUTING AND INTELLIGENCE SYSTEMS (CCIS), 2019, : 6 - 10
  • [2] Effective Unsupervised Constrained Text Generation based on Perturbed Masking
    Fu, Yingwen
    Ou, Wenjie
    Yu, Zhou
    Lin, Yue
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), 2022, : 1417 - 1427
  • [3] Improving semantic coverage of data-to-text generation model using dynamic memory networks
    Seifossadat, Elham
    Sameti, Hossein
    NATURAL LANGUAGE ENGINEERING, 2024, 30 (03) : 454 - 479
  • [4] XFBoost: Improving Text Generation with Controllable Decoders
    Peng, Xiangyu
    Sollami, Michael
    arXiv, 2022,
  • [5] Improving Text Generation Through Introducing Coherence Metrics
    O. O. Marchenko
    O. S. Radyvonenko
    T. S. Ignatova
    P. V. Titarchuk
    D. V. Zhelezniakov
    Cybernetics and Systems Analysis, 2020, 56 : 13 - 21
  • [6] Improving Adversarial Text Generation by Modeling the Distant Future
    Zhang, Ruiyi
    Chen, Changyou
    Gan, Zhe
    Wang, Wenlin
    Shen, Dinghan
    Wang, Guoyin
    Wen, Zheng
    Carin, Lawrence
    58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), 2020, : 2516 - 2531
  • [7] Improving Text Generation via Neural Discourse Planning
    Chernyavskiy, Alexander
    WSDM'22: PROCEEDINGS OF THE FIFTEENTH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING, 2022, : 1543 - 1544
  • [8] Improving Text Generation Through Introducing Coherence Metrics
    Marchenko, O. O.
    Radyvonenko, O. S.
    Ignatova, T. S.
    Titarchuk, P. V.
    Zhelezniakov, D. V.
    CYBERNETICS AND SYSTEMS ANALYSIS, 2020, 56 (01) : 13 - 21
  • [9] Text Generation From Data With Dynamic Planning
    Yang, Sen
    Liu, Yang
    Feng, Dawei
    Li, Dongsheng
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2022, 30 : 26 - 34
  • [10] MaskGEC: Improving Neural Grammatical Error Correction via Dynamic Masking
    Zhao, Zewei
    Wang, Houfeng
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 1226 - 1233