SupMPN: Supervised Multiple Positives and Negatives Contrastive Learning Model for Semantic Textual Similarity

被引:5
|
作者
Dehghan, Somaiyeh [1 ]
Amasyali, Mehmet Fatih [1 ]
机构
[1] Yildiz Tech Univ, Dept Comp Engn, TR-34220 Istanbul, Turkey
来源
APPLIED SCIENCES-BASEL | 2022年 / 12卷 / 19期
关键词
Natural Language Processing; sentence embedding; Semantic Textual Similarity; BERT; contrastive learning; deep learning; ACTIVITY RECOGNITION;
D O I
10.3390/app12199659
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Semantic Textual Similarity (STS) is an important task in the area of Natural Language Processing (NLP) that measures the similarity of the underlying semantics of two texts. Although pre-trained contextual embedding models such as Bidirectional Encoder Representations from Transformers (BERT) have achieved state-of-the-art performance on several NLP tasks, BERT-derived sentence embeddings have been proven to collapse in some way, i.e., sentence embeddings generated by BERT depend on the frequency of words. Therefore, almost all BERT-derived sentence embeddings are mapped into a small area and have a high cosine similarity. Hence, sentence embeddings generated by BERT are not so robust in the STS task as they cannot capture the full semantic meaning of the sentences. In this paper, we propose SupMPN: A Supervised Multiple Positives and Negatives Contrastive Learning Model, which accepts multiple hard-positive sentences and multiple hard-negative sentences simultaneously and then tries to bring hard-positive sentences closer, while pushing hard-negative sentences away from them. In other words, SupMPN brings similar sentences closer together in the representation space by discrimination among multiple similar and dissimilar sentences. In this way, SupMPN can learn the semantic meanings of sentences by contrasting among multiple similar and dissimilar sentences and can generate sentence embeddings based on the semantic meaning instead of the frequency of the words. We evaluate our model on standard STS and transfer-learning tasks. The results reveal that SupMPN outperforms state-of-the-art SimCSE and all other previous supervised and unsupervised models.
引用
收藏
页数:20
相关论文
共 50 条
  • [1] Contrastive Meta-Learner for Automatic Text Labeling and Semantic Textual Similarity
    Cooper, Ryan
    Kliesner, Kenneth W.
    Zenker, Stephen
    IEEE ACCESS, 2024, 12 : 166792 - 166799
  • [2] SuperConText: Supervised Contrastive Learning Framework for Textual Representations
    Moukafih, Youness
    Sbihi, Nada
    Ghogho, Mounir
    Smaili, Kamel
    IEEE ACCESS, 2023, 11 : 16820 - 16830
  • [3] Improved contrastive learning model via identification of false-negatives in self-supervised learning
    Auh, Joonsun
    Cho, Changsik
    Kim, Seon-tae
    ETRI JOURNAL, 2024, 46 (06) : 1020 - 1029
  • [4] SEMANTIC-ENHANCED SUPERVISED CONTRASTIVE LEARNING
    Zhang, Pingyue
    Wu, Mengyue
    Yu, Kai
    2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024, 2024, : 6030 - 6034
  • [5] Self-supervised contrastive representation learning for semantic segmentation
    Liu B.
    Cai H.
    Wang Y.
    Chen X.
    Xi'an Dianzi Keji Daxue Xuebao/Journal of Xidian University, 2024, 51 (01): : 125 - 134
  • [6] Predicting Semantic Textual Similarity of Arabic Question Pairs using Deep Learning
    Einea, Omar
    Elnagar, Ashraf
    2019 IEEE/ACS 16TH INTERNATIONAL CONFERENCE ON COMPUTER SYSTEMS AND APPLICATIONS (AICCSA 2019), 2019,
  • [7] Spectral Learning of Semantic Units in a Sentence Pair to Evaluate Semantic Textual Similarity
    Mehndiratta, Akanksha
    Asawa, Krishna
    8TH INTERNATIONAL CONFERENCE ON BIG DATA ANALYTICS, BDA 2020, 2020, 12581 : 49 - 59
  • [8] Evaluating Semantic Textual Similarity in Clinical Sentences Using Deep Learning and Sentence Embeddings
    Antunes, Rui
    Silva, Joao Figueira
    Matos, Sergio
    PROCEEDINGS OF THE 35TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING (SAC'20), 2020, : 662 - 669
  • [9] Contrastive semantic similarity learning for image captioning evaluation
    Zeng, Chao
    Kwong, Sam
    Zhao, Tiesong
    Wang, Hanli
    INFORMATION SCIENCES, 2022, 609 : 913 - 930
  • [10] A Semantic Textual Similarity Calculation Model Based on Pre-training Model
    Ding, Zhaoyun
    Liu, Kai
    Wang, Wenhao
    Liu, Bin
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, KSEM 2021, PT II, 2021, 12816 : 3 - 15