An End-to-End Contrastive Self-Supervised Learning Framework for Language Understanding

被引:6
|
作者
Fang, Hongchao [1 ]
Xie, Pengtao [1 ]
机构
[1] Univ Calif San Diego, La Jolla, CA 92093 USA
关键词
D O I
10.1162/tacl_a_00521
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Self-supervised learning (SSL) methods such as Word2vec, BERT, and GPT have shown great effectiveness in language understanding. Contrastive learning, as a recent SSL approach, has attracted increasing attention in NLP. Contrastive learning learns data representations by predicting whether two augmented data instances are generated from the same original data example. Previous contrastive learning methods perform data augmentation and contrastive learning separately. As a result, the augmented data may not be optimal for contrastive learning. To address this problem, we propose a four-level optimization framework that performs data augmentation and contrastive learning end-to-end, to enable the augmented data to be tailored to the contrastive learning task. This framework consists of four learning stages, including training machine translation models for sentence augmentation, pretraining a text encoder using contrastive learning, finetuning a text classification model, and updating weights of translation data by minimizing the validation loss of the classification model, which are performed in a unified way. Experiments on datasets in the GLUE benchmark (Wang et al., 2018a) and on datasets used in Gururangan et al. (2020) demonstrate the effectiveness of our method.
引用
收藏
页码:1324 / 1340
页数:17
相关论文
共 50 条
  • [1] End-to-end Jordanian dialect speech-to-text self-supervised learning framework
    Safieh, Ali A.
    Abu Alhaol, Ibrahim
    Ghnemat, Rawan
    FRONTIERS IN ROBOTICS AND AI, 2022, 9
  • [2] ActiveStereoNet: End-to-End Self-supervised Learning for Active Stereo Systems
    Zhang, Yinda
    Khamis, Sameh
    Rhemann, Christoph
    Valentin, Julien
    Kowdle, Adarsh
    Tankovich, Vladimir
    Schoenberg, Michael
    Izadi, Shahram
    Funkhouser, Thomas
    Fanello, Sean
    COMPUTER VISION - ECCV 2018, PT VIII, 2018, 11212 : 802 - 819
  • [3] END-TO-END SPOKEN LANGUAGE UNDERSTANDING USING TRANSFORMER NETWORKS AND SELF-SUPERVISED PRE-TRAINED FEATURES
    Morais, Edmilson
    Kuo, Hong-Kwang J.
    Thomas, Samuel
    Tuske, Zoltan
    Kingsbury, Brian
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 7483 - 7487
  • [4] Self-supervised end-to-end graph local clustering
    Zhe Yuan
    World Wide Web, 2023, 26 : 1157 - 1179
  • [5] Self-supervised end-to-end graph local clustering
    Yuan, Zhe
    WORLD WIDE WEB-INTERNET AND WEB INFORMATION SYSTEMS, 2023, 26 (03): : 1157 - 1179
  • [6] SCaLa: Supervised Contrastive Learning for End-to-End Speech Recognition
    Fu, Li
    Li, Xiaoxiao
    Wang, Runyu
    Fan, Lu
    Zhang, Zhengchen
    Chen, Meng
    Wu, Youzheng
    He, Xiaodong
    INTERSPEECH 2022, 2022, : 1006 - 1010
  • [7] End-to-end learning of self-rectification and self-supervised disparity prediction for stereo vision
    Zhang, Xuchong
    Zhao, Yongli
    Wang, Hang
    Zhai, Han
    Sun, Hongbin
    Zheng, Nanning
    NEUROCOMPUTING, 2022, 494 : 308 - 319
  • [8] Self-Supervised Representations Improve End-to-End Speech Translation
    Wu, Anne
    Wang, Changhan
    Pino, Juan
    Gu, Jiatao
    INTERSPEECH 2020, 2020, : 1491 - 1495
  • [9] Geometric Consistency for Self-Supervised End-to-End Visual Odometry
    Iyer, Ganesh
    Murthy, J. Krishna
    Gupta, Gunshi
    Krishna, K. Madhava
    Paull, Liam
    PROCEEDINGS 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2018, : 380 - 388
  • [10] End-to-End Integration of Speech Recognition, Speech Enhancement, and Self-Supervised Learning Representation
    Chang, Xuankai
    Maekaku, Takashi
    Fujita, Yuya
    Watanabe, Shinji
    INTERSPEECH 2022, 2022, : 3819 - 3823