SupMPN: Supervised Multiple Positives and Negatives Contrastive Learning Model for Semantic Textual Similarity

被引:5
|
作者
Dehghan, Somaiyeh [1 ]
Amasyali, Mehmet Fatih [1 ]
机构
[1] Yildiz Tech Univ, Dept Comp Engn, TR-34220 Istanbul, Turkey
来源
APPLIED SCIENCES-BASEL | 2022年 / 12卷 / 19期
关键词
Natural Language Processing; sentence embedding; Semantic Textual Similarity; BERT; contrastive learning; deep learning; ACTIVITY RECOGNITION;
D O I
10.3390/app12199659
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Semantic Textual Similarity (STS) is an important task in the area of Natural Language Processing (NLP) that measures the similarity of the underlying semantics of two texts. Although pre-trained contextual embedding models such as Bidirectional Encoder Representations from Transformers (BERT) have achieved state-of-the-art performance on several NLP tasks, BERT-derived sentence embeddings have been proven to collapse in some way, i.e., sentence embeddings generated by BERT depend on the frequency of words. Therefore, almost all BERT-derived sentence embeddings are mapped into a small area and have a high cosine similarity. Hence, sentence embeddings generated by BERT are not so robust in the STS task as they cannot capture the full semantic meaning of the sentences. In this paper, we propose SupMPN: A Supervised Multiple Positives and Negatives Contrastive Learning Model, which accepts multiple hard-positive sentences and multiple hard-negative sentences simultaneously and then tries to bring hard-positive sentences closer, while pushing hard-negative sentences away from them. In other words, SupMPN brings similar sentences closer together in the representation space by discrimination among multiple similar and dissimilar sentences. In this way, SupMPN can learn the semantic meanings of sentences by contrasting among multiple similar and dissimilar sentences and can generate sentence embeddings based on the semantic meaning instead of the frequency of the words. We evaluate our model on standard STS and transfer-learning tasks. The results reveal that SupMPN outperforms state-of-the-art SimCSE and all other previous supervised and unsupervised models.
引用
收藏
页数:20
相关论文
共 50 条
  • [41] CLGLIAM: contrastive learning model based on global and local semantic interaction for address matching
    Lei, Jianjun
    Wu, Chen
    Wang, Ying
    APPLIED INTELLIGENCE, 2023, 53 (23) : 29267 - 29281
  • [42] Self Supervised Contrastive Learning on Multiple Breast Modalities Boosts Classification Performance
    Perek, Shaked
    Amit, Mika
    Hexter, Efrat
    PREDICTIVE INTELLIGENCE IN MEDICINE, PRIME 2021, 2021, 12928 : 117 - 127
  • [43] Multiple representation contrastive self-supervised learning for pulmonary nodule detection
    Torki, Asghar
    Adibi, Peyman
    Kashani, Hamidreza Baradaran
    KNOWLEDGE-BASED SYSTEMS, 2024, 301
  • [44] Semantic segmentation algorithm for foggy cityscapes images by fusing self-supervised contrastive learning
    Liu, Liwei
    Wang, Rui
    Meng, Xutao
    CHINESE JOURNAL OF LIQUID CRYSTALS AND DISPLAYS, 2024, 39 (07) : 990 - 1000
  • [45] Supervised contrastive learning based dual-mixer model for Remaining Useful Life prediction
    Fu, En
    Hu, Yanyan
    Peng, Kaixiang
    Chu, Yuxin
    RELIABILITY ENGINEERING & SYSTEM SAFETY, 2024, 251
  • [46] Simcryocluster: a semantic similarity clustering method of cryo-EM images by adopting contrastive learning
    Tang, Huanrong
    Wang, Yaowu
    Ouyang, Jianquan
    Wang, Jinlin
    BMC BIOINFORMATICS, 2024, 25 (01)
  • [47] Global and Local Contrastive Self-Supervised Learning for Semantic Segmentation of HR Remote Sensing Images
    Li, Haifeng
    Li, Yi
    Zhang, Guo
    Liu, Ruoyun
    Huang, Haozhe
    Zhu, Qing
    Tao, Chao
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
  • [48] Sentence modeling via multiple word embeddings and multi-level comparison for semantic textual similarity
    Nguyen Huy Tien
    Nguyen Minh Le
    Tomohiro, Yamasaki
    Tatsuya, Izuha
    INFORMATION PROCESSING & MANAGEMENT, 2019, 56 (06)
  • [49] Predicting learning performance using NLP: an exploratory study using two semantic textual similarity methods
    Papadimas, C.
    Ragazou, V.
    Karasavvidis, I.
    Kollias, V.
    KNOWLEDGE AND INFORMATION SYSTEMS, 2025, : 4567 - 4595
  • [50] Min-Max Similarity: A Contrastive Semi-Supervised Deep Learning Network for Surgical Tools Segmentation
    Lou, Ange
    Tawfik, Kareem
    Yao, Xing
    Liu, Ziteng
    Noble, Jack
    IEEE TRANSACTIONS ON MEDICAL IMAGING, 2023, 42 (10) : 2832 - 2841