Improving Text Generation with Dynamic Masking and Recovering

被引:0
|
作者
Liu, Zhidong [1 ]
Li, Junhui [1 ]
Zhu, Muhua [2 ]
机构
[1] Soochow Univ, Sch Comp Sci & Technol, Suzhou, Peoples R China
[2] Tencent News, Beijing, Peoples R China
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Due to different types of inputs, diverse text generation tasks may adopt different encoder-decoder frameworks. Thus most existing approaches that aim to improve the robustness of certain generation tasks are input-relevant, and may not work well for other generation tasks. Alternatively, in this paper we present a universal approach to enhance the language representation for text generation on the base of generic encoder-decoder frameworks. This is done from two levels. First, we introduce randomness by randomly masking some percentage of tokens on the decoder side when training the models. In this way, instead of using ground truth history context, we use its corrupted version to predict the next token. Then we propose an auxiliary task to properly recover those masked tokens. Experimental results on several text generation tasks including machine translation (MT), AMR-to-text generation, and image captioning show that the proposed approach can significantly improve over competitive baselines without using any task-specific techniques. This suggests the effectiveness and generality of our proposed approach.
引用
收藏
页码:3878 / 3884
页数:7
相关论文
共 50 条
  • [41] Improving Compositional Generalization with Self-Training for Data-to-Text Generation
    Mehta, Sanket Vaibhav
    Rao, Jinfeng
    Tay, Yi
    Kale, Mihir
    Parikh, Ankur P.
    Strubell, Emma
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 4205 - 4219
  • [42] Improving Efficiency of Natural-Language Text Generation for Automatic Pedagogical Questions
    Gomazkova, Yulia
    Sychev, Oleg
    Gumerov, Marat
    COMPUTATIONAL SCIENCE AND ITS APPLICATIONS-ICCSA 2024 WORKSHOPS, PT II, 2024, 14816 : 37 - 50
  • [43] Improving Factual Completeness and Consistency of Image-to-Text Radiology Report Generation
    Miura, Yasuhide
    Zhang, Yuhao
    Tsai, Emily Bao
    Langlotz, Curtis P.
    Jurafsky, Dan
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 5288 - 5304
  • [44] ScenarioDiff: Text-to-video Generation with Dynamic Transformations of Scene Conditions
    Zhang, Yipeng
    Wang, Xin
    Chen, Hong
    Qin, Chenyang
    Hao, Yibo
    Mei, Hong
    Zhu, Wenwu
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2025,
  • [45] Controlled Text Generation for Large Language Model with Dynamic Attribute Graphs
    Liang, Xun
    Wang, Hanyu
    Song, Shichao
    Hu, Mengting
    Wang, Xunzhi
    Li, Zhiyu
    Xiong, Feiyu
    Tang, Bo
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 5797 - 5814
  • [46] Lightweight, Dynamic Graph Convolutional Networks for AMR-to-Text Generation
    Zhang, Yan
    Guo, Zhijiang
    Teng, Zhiyang
    Lu, Wei
    Cohen, Shay B.
    Liu, Zuozhu
    Bing, Lidong
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 2162 - 2172
  • [47] CopyCAT: Masking Strategy Conscious Augmented Text for Machine Generated Text Detection
    Liu, Chien-Liang
    Kao, Hung-Yu
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2023, PT I, 2023, 13935 : 367 - 379
  • [48] Visual design guidelines for improving learning from dynamic and interactive digital text
    Jin, Sung-Hee
    COMPUTERS & EDUCATION, 2013, 63 : 248 - 258
  • [49] TEXT GENERATION
    KUNST, AE
    SUB-STANCE, 1977, (16) : 159 - 171
  • [50] Improving Streaming Speech Recognition With Time-Shifted Contextual Attention And Dynamic Right Context Masking
    Khanh Le
    Duc Chau
    INTERSPEECH 2024, 2024, : 4478 - 4482