SupMPN: Supervised Multiple Positives and Negatives Contrastive Learning Model for Semantic Textual Similarity

被引:5
|
作者
Dehghan, Somaiyeh [1 ]
Amasyali, Mehmet Fatih [1 ]
机构
[1] Yildiz Tech Univ, Dept Comp Engn, TR-34220 Istanbul, Turkey
来源
APPLIED SCIENCES-BASEL | 2022年 / 12卷 / 19期
关键词
Natural Language Processing; sentence embedding; Semantic Textual Similarity; BERT; contrastive learning; deep learning; ACTIVITY RECOGNITION;
D O I
10.3390/app12199659
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Semantic Textual Similarity (STS) is an important task in the area of Natural Language Processing (NLP) that measures the similarity of the underlying semantics of two texts. Although pre-trained contextual embedding models such as Bidirectional Encoder Representations from Transformers (BERT) have achieved state-of-the-art performance on several NLP tasks, BERT-derived sentence embeddings have been proven to collapse in some way, i.e., sentence embeddings generated by BERT depend on the frequency of words. Therefore, almost all BERT-derived sentence embeddings are mapped into a small area and have a high cosine similarity. Hence, sentence embeddings generated by BERT are not so robust in the STS task as they cannot capture the full semantic meaning of the sentences. In this paper, we propose SupMPN: A Supervised Multiple Positives and Negatives Contrastive Learning Model, which accepts multiple hard-positive sentences and multiple hard-negative sentences simultaneously and then tries to bring hard-positive sentences closer, while pushing hard-negative sentences away from them. In other words, SupMPN brings similar sentences closer together in the representation space by discrimination among multiple similar and dissimilar sentences. In this way, SupMPN can learn the semantic meanings of sentences by contrasting among multiple similar and dissimilar sentences and can generate sentence embeddings based on the semantic meaning instead of the frequency of the words. We evaluate our model on standard STS and transfer-learning tasks. The results reveal that SupMPN outperforms state-of-the-art SimCSE and all other previous supervised and unsupervised models.
引用
收藏
页数:20
相关论文
共 50 条
  • [31] Siamese BERT Architecture Model with attention mechanism for Textual Semantic Similarity
    Li, Ruihao
    Cheng, Lianglun
    Wang, Depei
    Tan, Junming
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 82 (30) : 46673 - 46694
  • [32] DP-CCL: A Supervised Contrastive Learning Approach Using CodeBERT Model in Software Defect Prediction
    Sahar, Sadia
    Younas, Muhammad
    Khan, Muhammad Murad
    Sarwar, Muhammad Umer
    IEEE ACCESS, 2024, 12 : 22582 - 22594
  • [33] Semantic Textual Similarity of Portuguese-Language Texts: An Approach Based on the Semantic Inferentialism Model
    Pinheiro, Vladia
    Furtado, Vasco
    Albuquerque, Adriano
    COMPUTATIONAL PROCESSING OF THE PORTUGUESE LANGUAGE, 2014, 8775 : 183 - 188
  • [34] Dense Supervised Dual-Aware Contrastive Learning for Airborne Laser Scanning Weakly Supervised Semantic Segmentation
    Luo, Ziwei
    Zeng, Tao
    Jiang, Xinyi
    Peng, Qingyu
    Ma, Ying
    Xie, Zhong
    Pan, Xiong
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2025, 63
  • [35] Remote Sensing Image Semantic Change Detection Boosted by Semi-Supervised Contrastive Learning of Semantic Segmentation
    Zhang, Xiuwei
    Yang, Yizhe
    Ran, Lingyan
    Chen, Liang
    Wang, Kangwei
    Yu, Lei
    Wang, Peng
    Zhang, Yanning
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2024, 62 : 1 - 13
  • [36] Distributional Semantic Model Based on Convolutional Neural Network for Arabic Textual Similarity
    Mahmoud, Adnen
    Zrigui, Mounir
    INTERNATIONAL JOURNAL OF COGNITIVE INFORMATICS AND NATURAL INTELLIGENCE, 2020, 14 (01) : 35 - 50
  • [37] Extending Monolingual Semantic Textual Similarity Task to Multiple Cross-lingual Settings
    Hayashi, Yoshihiko
    Luo, Wentao
    LREC 2016 - TENTH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2016, : 1233 - 1239
  • [38] Source Model Selection for Transfer Learning of Image Classification using Supervised Contrastive Loss
    Cho, Young-Seong
    Kim, Samuel
    Lee, Jee-Hyong
    2021 IEEE INTERNATIONAL CONFERENCE ON BIG DATA AND SMART COMPUTING (BIGCOMP 2021), 2021, : 325 - 329
  • [39] End-to-end model for automatic seizure detection using supervised contrastive learning
    Li, Haotian
    Dong, Xingchen
    Zhong, Xiangwen
    Li, Chuanyu
    Cui, Haozhou
    Zhou, Weidong
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 133
  • [40] CLGLIAM: contrastive learning model based on global and local semantic interaction for address matching
    Jianjun Lei
    Chen Wu
    Ying Wang
    Applied Intelligence, 2023, 53 : 29267 - 29281