Improving Text Generation with Dynamic Masking and Recovering

被引:0
|
作者
Liu, Zhidong [1 ]
Li, Junhui [1 ]
Zhu, Muhua [2 ]
机构
[1] Soochow Univ, Sch Comp Sci & Technol, Suzhou, Peoples R China
[2] Tencent News, Beijing, Peoples R China
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Due to different types of inputs, diverse text generation tasks may adopt different encoder-decoder frameworks. Thus most existing approaches that aim to improve the robustness of certain generation tasks are input-relevant, and may not work well for other generation tasks. Alternatively, in this paper we present a universal approach to enhance the language representation for text generation on the base of generic encoder-decoder frameworks. This is done from two levels. First, we introduce randomness by randomly masking some percentage of tokens on the decoder side when training the models. In this way, instead of using ground truth history context, we use its corrupted version to predict the next token. Then we propose an auxiliary task to properly recover those masked tokens. Experimental results on several text generation tasks including machine translation (MT), AMR-to-text generation, and image captioning show that the proposed approach can significantly improve over competitive baselines without using any task-specific techniques. This suggests the effectiveness and generality of our proposed approach.
引用
收藏
页码:3878 / 3884
页数:7
相关论文
共 50 条
  • [21] Disentangling Representations of Text by Masking Transformers
    Zhang, Xiongyi
    van de Meent, Jan-Willem
    Wallace, Byron C.
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 778 - 791
  • [22] Improving vision by pupil masking
    Bonaque-Gonzalez, Sergio
    Rios-Rodriguez, Susana
    Lopez-Gil, Norberto
    BIOMEDICAL OPTICS EXPRESS, 2016, 7 (07): : 2538 - 2550
  • [23] Improving PLMs for Graph-to-Text Generation by Relational Orientation Attention
    Wang, Tao
    Shen, Bo
    Zhang, Jinglin
    Zhong, Yu
    NEURAL PROCESSING LETTERS, 2023, 55 (06) : 7967 - 7983
  • [24] Improving Maximum Likelihood Training for Text Generation with Density Ratio Estimation
    Song, Yuxuan
    Miao, Ning
    Zhou, Hao
    Yu, Lantao
    Wang, Mingxuan
    Li, Lei
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108
  • [25] Improving DRS-to-Text Generation Through Delexicalization and Data Augmentation
    Amin, Muhammad Saad
    Anselma, Luca
    Mazzei, Alessandro
    NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS, PT I, NLDB 2024, 2024, 14762 : 121 - 136
  • [26] Search and Learn: Improving Semantic Coverage for Data-to-Text Generation
    Jolly, Shailza
    Zhang, Zi Xuan
    Dengel, Andreas
    Mou, Lili
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 10858 - 10866
  • [27] Implicit Unlikelihood Training: Improving Neural Text Generation with Reinforcement Learning
    Lagutin, Evgeny
    Gavrilov, Daniil
    Kalaidin, Pavel
    16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 1432 - 1441
  • [28] Improving Controllable Text Generation with Position-Aware Weighted Decoding
    Gu, Yuxuan
    Feng, Xiaocheng
    Ma, Sicheng
    Wu, Jiaming
    Gong, Heng
    Qin, Bing
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), 2022, : 3449 - 3467
  • [29] Improving PLMs for Graph-to-Text Generation by Relational Orientation Attention
    Tao Wang
    Bo Shen
    Jinglin Zhang
    Yu Zhong
    Neural Processing Letters, 2023, 55 : 7967 - 7983
  • [30] Dynamic pattern masking
    Bowen, RW
    deRidder, H
    INVESTIGATIVE OPHTHALMOLOGY & VISUAL SCIENCE, 1996, 37 (03) : 2389 - 2389