Unveiling Key Aspects of Fine-Tuning in Sentence Embeddings: A Representation Rank Analysis

被引:0
|
作者
Jung, Euna [1 ]
Kim, Jaeill [2 ]
Ko, Jungmin [3 ]
Park, Jinwoo [1 ]
Rhee, Wonjong [3 ,4 ,5 ]
机构
[1] Samsung Adv Inst Technol, Suwon 16678, Gyeonggi Do, South Korea
[2] LINE Investment Technol, Seongnam Si 13529, Gyeonggi Do, South Korea
[3] Seoul Natl Univ, Interdisciplinary Program Artificial Intelligence, Seoul 08826, South Korea
[4] Seoul Natl Univ, Dept Intelligence & Informat, Seoul 08826, South Korea
[5] Seoul Natl Univ, RICS, Seoul 08826, South Korea
来源
IEEE ACCESS | 2024年 / 12卷
基金
新加坡国家研究基金会;
关键词
Training; Linguistics; Contrastive learning; Market research; Correlation; Semantics; Visualization; Phase measurement; Natural language processing; Loss measurement; Sentence embedding; self-supervised learning; contrastive learning; fine-tuning; representation rank;
D O I
10.1109/ACCESS.2024.3485705
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The latest advancements in unsupervised learning of sentence embeddings predominantly involve employing contrastive learning-based (CL-based) fine-tuning over pre-trained language models. In this study, we analyze the latest sentence embedding methods by adopting representation rank as the primary tool of analysis. We first define Phase 1 and Phase 2 of fine-tuning based on when representation rank peaks. Utilizing these phases, we conduct a thorough analysis and obtain essential findings across key aspects, including alignment and uniformity, linguistic abilities, and correlation between performance and rank. For instance, we find that the dynamics of the key aspects can undergo significant changes as fine-tuning transitions from Phase 1 to Phase 2. Based on these findings, we experiment with a rank reduction (RR) strategy that facilitates rapid and stable fine-tuning of the latest CL-based methods. Through empirical investigations, we showcase the efficacy of RR in enhancing the performance and stability of five state-of-the-art sentence embedding methods. The code is available at (https://github.com/SNU-DRL/SentenceEmbedding_Rank).
引用
收藏
页码:159877 / 159888
页数:12
相关论文
共 37 条
  • [21] A pool-based pattern generation algorithm for logical analysis of data with automatic fine-tuning
    Caserta, Marco
    Reiners, Torsten
    EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 2016, 248 (02) : 593 - 606
  • [22] Robustness Fine-Tuning Deep Learning Model for Cancers Diagnosis Based on Histopathology Image Analysis
    El-Ghany, Sameh Abd
    Azad, Mohammad
    Elmogy, Mohammed
    DIAGNOSTICS, 2023, 13 (04)
  • [23] Sentiment Analysis Using Pre-Trained Language Model With No Fine-Tuning and Less Resource
    Kit, Yuheng
    Mokji, Musa Mohd
    IEEE ACCESS, 2022, 10 : 107056 - 107065
  • [24] AdaFT: An efficient domain-adaptive fine-tuning framework for sentiment analysis in chinese financial texts
    Guofeng Yan
    Kuashuai Peng
    Yongfeng Wang
    Hengliang Tan
    Jiao Du
    Heng Wu
    Applied Intelligence, 2025, 55 (10)
  • [25] Enhancing the Analysis of Interdisciplinary Learning Quality with GPT Models: Fine-Tuning and Knowledge-Empowered Approaches
    Zhong, Tianlong
    Cai, Chang
    Zhu, Gaoxia
    Ma, Min
    ARTIFICIAL INTELLIGENCE IN EDUCATION: POSTERS AND LATE BREAKING RESULTS, WORKSHOPS AND TUTORIALS, INDUSTRY AND INNOVATION TRACKS, PRACTITIONERS, DOCTORAL CONSORTIUM AND BLUE SKY, AIED 2024, 2024, 2151 : 157 - 165
  • [26] Enhanced Feature Representation for Multimodal Fake News Detection Using Localized Fine-Tuning of Improved BERT and VGG-19 Models
    Hamed, Suhaib Kh.
    Ab Aziz, Mohd Juzaiddin
    Yaakub, Mohd Ridzwan
    ARABIAN JOURNAL FOR SCIENCE AND ENGINEERING, 2024, : 7423 - 7439
  • [27] Leveraging Low-Rank Adaptation for Parameter-Efficient Fine-Tuning in Multi-Speaker Adaptive Text-to-Speech Synthesis
    Hong, Changi
    Lee, Jung Hyuk
    Kim, Hong Kook
    IEEE ACCESS, 2024, 12 : 190711 - 190727
  • [28] Parameter-Efficient Fine-Tuning of Pre-trained Large Language Models for Financial Text Analysis
    Langa, Kelly
    Wang, Hairong
    Okuboyejo, Olaperi
    ARTIFICIAL INTELLIGENCE RESEARCH, SACAIR 2024, 2025, 2326 : 3 - 20
  • [29] Lexicon-based fine-tuning of multilingual language models for low-resource language sentiment analysis
    Dhananjaya, Vinura
    Ranathunga, Surangika
    Jayasena, Sanath
    CAAI TRANSACTIONS ON INTELLIGENCE TECHNOLOGY, 2024, 9 (05) : 1116 - 1125
  • [30] Optimizing Airline Review Sentiment Analysis: A Comparative Analysis of LLaMA and BERT Models through Fine-Tuning and Few-Shot Learning
    Roumeliotis, Konstantinos I.
    Tselikas, Nikolaos D.
    Nasiopoulos, Dimitrios K.
    CMC-COMPUTERS MATERIALS & CONTINUA, 2025, 82 (02): : 2769 - 2792