SupMPN: Supervised Multiple Positives and Negatives Contrastive Learning Model for Semantic Textual Similarity

被引:5
|
作者
Dehghan, Somaiyeh [1 ]
Amasyali, Mehmet Fatih [1 ]
机构
[1] Yildiz Tech Univ, Dept Comp Engn, TR-34220 Istanbul, Turkey
来源
APPLIED SCIENCES-BASEL | 2022年 / 12卷 / 19期
关键词
Natural Language Processing; sentence embedding; Semantic Textual Similarity; BERT; contrastive learning; deep learning; ACTIVITY RECOGNITION;
D O I
10.3390/app12199659
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Semantic Textual Similarity (STS) is an important task in the area of Natural Language Processing (NLP) that measures the similarity of the underlying semantics of two texts. Although pre-trained contextual embedding models such as Bidirectional Encoder Representations from Transformers (BERT) have achieved state-of-the-art performance on several NLP tasks, BERT-derived sentence embeddings have been proven to collapse in some way, i.e., sentence embeddings generated by BERT depend on the frequency of words. Therefore, almost all BERT-derived sentence embeddings are mapped into a small area and have a high cosine similarity. Hence, sentence embeddings generated by BERT are not so robust in the STS task as they cannot capture the full semantic meaning of the sentences. In this paper, we propose SupMPN: A Supervised Multiple Positives and Negatives Contrastive Learning Model, which accepts multiple hard-positive sentences and multiple hard-negative sentences simultaneously and then tries to bring hard-positive sentences closer, while pushing hard-negative sentences away from them. In other words, SupMPN brings similar sentences closer together in the representation space by discrimination among multiple similar and dissimilar sentences. In this way, SupMPN can learn the semantic meanings of sentences by contrasting among multiple similar and dissimilar sentences and can generate sentence embeddings based on the semantic meaning instead of the frequency of the words. We evaluate our model on standard STS and transfer-learning tasks. The results reveal that SupMPN outperforms state-of-the-art SimCSE and all other previous supervised and unsupervised models.
引用
收藏
页数:20
相关论文
共 50 条
  • [21] A Semantic Similarity Distance-Aware Contrastive Learning for Abstractive Summarization
    Huang, Ying
    Li, Zhixin
    PRICAI 2023: TRENDS IN ARTIFICIAL INTELLIGENCE, PT I, 2024, 14325 : 173 - 185
  • [22] Learning to Rank Hypernyms of Financial Terms Using Semantic Textual Similarity
    Ghosh S.
    Chopra A.
    Naskar S.K.
    SN Computer Science, 4 (5)
  • [23] CLSSATP: Contrastive learning and self-supervised learning model for aquatic toxicity prediction
    Lin, Ye
    Yang, Xin
    Zhang, Mingxuan
    Cheng, Jinyan
    Lin, Hai
    Zhao, Qi
    AQUATIC TOXICOLOGY, 2025, 279
  • [24] Learning Semantic Textual Similarity via Multi-Teacher Knowledge Distillation: A Multiple Data Augmentation method
    Lu, Zhikun
    Zhao, Ying
    Li, Jinnan
    Tian, Yuan
    2024 9TH INTERNATIONAL CONFERENCE ON COMPUTER AND COMMUNICATION SYSTEMS, ICCCS 2024, 2024, : 1197 - 1203
  • [25] Contrastive and consistent feature learning for weakly supervised object localization and semantic segmentation
    Ki, Minsong
    Uh, Youngjung
    Lee, Wonyoung
    Byun, Hyeran
    NEUROCOMPUTING, 2021, 445 : 244 - 254
  • [26] A multi-strategy contrastive learning framework for weakly supervised semantic segmentation
    Yuan, Kunhao
    Schaefer, Gerald
    Lai, Yu-Kun
    Wang, Yifan
    Liu, Xiyao
    Guan, Lin
    Fang, Hui
    PATTERN RECOGNITION, 2023, 137
  • [27] PRCL: Probabilistic Representation Contrastive Learning for Semi-Supervised Semantic Segmentation
    Xie, Haoyu
    Wang, Changqi
    Zhao, Jian
    Liu, Yang
    Dan, Jun
    Fu, Chong
    Sun, Baigui
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, 132 (10) : 4343 - 4361
  • [28] DuCL: Dual-stage contrastive learning framework for Chinese semantic textual matching
    Zuo, Youhui
    Lu, Wenpeng
    Peng, Xueping
    Wang, Shoujin
    Zhang, Weiyu
    Qiao, Xinxiao
    COMPUTERS & ELECTRICAL ENGINEERING, 2023, 106
  • [29] Siamese BERT Architecture Model with attention mechanism for Textual Semantic Similarity
    Ruihao Li
    Lianglun Cheng
    Depei Wang
    Junming Tan
    Multimedia Tools and Applications, 2023, 82 : 46673 - 46694
  • [30] Research on Text Sentiment Semantic Optimization Method Based on Supervised Contrastive Learning
    Xiong, Shuchu
    Li, Xuan
    Wu, Jiani
    Zhou, Zhaohong
    Meng, Han
    Data Analysis and Knowledge Discovery, 2024, 8 (06) : 69 - 81