An End-to-End Contrastive Self-Supervised Learning Framework for Language Understanding

被引:6
|
作者
Fang, Hongchao [1 ]
Xie, Pengtao [1 ]
机构
[1] Univ Calif San Diego, La Jolla, CA 92093 USA
关键词
D O I
10.1162/tacl_a_00521
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Self-supervised learning (SSL) methods such as Word2vec, BERT, and GPT have shown great effectiveness in language understanding. Contrastive learning, as a recent SSL approach, has attracted increasing attention in NLP. Contrastive learning learns data representations by predicting whether two augmented data instances are generated from the same original data example. Previous contrastive learning methods perform data augmentation and contrastive learning separately. As a result, the augmented data may not be optimal for contrastive learning. To address this problem, we propose a four-level optimization framework that performs data augmentation and contrastive learning end-to-end, to enable the augmented data to be tailored to the contrastive learning task. This framework consists of four learning stages, including training machine translation models for sentence augmentation, pretraining a text encoder using contrastive learning, finetuning a text classification model, and updating weights of translation data by minimizing the validation loss of the classification model, which are performed in a unified way. Experiments on datasets in the GLUE benchmark (Wang et al., 2018a) and on datasets used in Gururangan et al. (2020) demonstrate the effectiveness of our method.
引用
收藏
页码:1324 / 1340
页数:17
相关论文
共 50 条
  • [21] Investigating Self-supervised Pre-training for End-to-end Speech Translation
    Ha Nguyen
    Bougares, Fethi
    Tomashenko, Natalia
    Esteve, Yannick
    Besacier, Laurent
    INTERSPEECH 2020, 2020, : 1466 - 1470
  • [22] PVStereo: Pyramid Voting Module for End-to-End Self-Supervised Stereo Matching
    Wang, Hengli
    Fan, Rui
    Cai, Peide
    Liu, Ming
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (03) : 4353 - 4360
  • [23] Self-supervised Variational Contrastive Learning with Applications to Face Understanding
    Yavuz, Mehmet Can
    Yanikoglu, Berrin
    2024 IEEE 18TH INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION, FG 2024, 2024,
  • [24] Toward Understanding the Feature Learning Process of Self-supervised Contrastive Learning
    Wen, Zixin
    Li, Yuanzhi
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [25] Understanding Self-Supervised Learning Dynamics without Contrastive Pairs
    Tian, Yuandong
    Chen, Xinlei
    Ganguli, Surya
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139 : 7279 - 7289
  • [26] FundusNet, A self-supervised contrastive learning framework for Fundus Feature Learning
    Mojab, Nooshin
    Alam, Minhaj
    Hallak, Joelle
    INVESTIGATIVE OPHTHALMOLOGY & VISUAL SCIENCE, 2022, 63 (07)
  • [27] A NOVEL CONTRASTIVE LEARNING FRAMEWORK FOR SELF-SUPERVISED ANOMALY DETECTION
    Li, Jingze
    Lian, Zhichao
    Li, Min
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 3366 - 3370
  • [28] Image classification framework based on contrastive self-supervised learning
    Zhao H.-W.
    Zhang J.-R.
    Zhu J.-P.
    Li H.
    Jilin Daxue Xuebao (Gongxueban)/Journal of Jilin University (Engineering and Technology Edition), 2022, 52 (08): : 1850 - 1856
  • [29] Exploring Transfer Learning For End-to-End Spoken Language Understanding
    Rongali, Subendhu
    Liu, Beiye
    Cai, Liwei
    Arkoudas, Konstantine
    Su, Chengwei
    Hamza, Wael
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 13754 - 13761
  • [30] MAS-CL: An End-to-End Multi-Atlas Supervised Contrastive Learning Framework for Brain ROI Segmentation
    Sun, Liang
    Fu, Yanling
    Zhao, Junyong
    Shao, Wei
    Zhu, Qi
    Zhang, Daoqiang
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 4319 - 4333