Unveiling Key Aspects of Fine-Tuning in Sentence Embeddings: A Representation Rank Analysis

被引:0
|
作者
Jung, Euna [1 ]
Kim, Jaeill [2 ]
Ko, Jungmin [3 ]
Park, Jinwoo [1 ]
Rhee, Wonjong [3 ,4 ,5 ]
机构
[1] Samsung Adv Inst Technol, Suwon 16678, Gyeonggi Do, South Korea
[2] LINE Investment Technol, Seongnam Si 13529, Gyeonggi Do, South Korea
[3] Seoul Natl Univ, Interdisciplinary Program Artificial Intelligence, Seoul 08826, South Korea
[4] Seoul Natl Univ, Dept Intelligence & Informat, Seoul 08826, South Korea
[5] Seoul Natl Univ, RICS, Seoul 08826, South Korea
来源
IEEE ACCESS | 2024年 / 12卷
基金
新加坡国家研究基金会;
关键词
Training; Linguistics; Contrastive learning; Market research; Correlation; Semantics; Visualization; Phase measurement; Natural language processing; Loss measurement; Sentence embedding; self-supervised learning; contrastive learning; fine-tuning; representation rank;
D O I
10.1109/ACCESS.2024.3485705
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The latest advancements in unsupervised learning of sentence embeddings predominantly involve employing contrastive learning-based (CL-based) fine-tuning over pre-trained language models. In this study, we analyze the latest sentence embedding methods by adopting representation rank as the primary tool of analysis. We first define Phase 1 and Phase 2 of fine-tuning based on when representation rank peaks. Utilizing these phases, we conduct a thorough analysis and obtain essential findings across key aspects, including alignment and uniformity, linguistic abilities, and correlation between performance and rank. For instance, we find that the dynamics of the key aspects can undergo significant changes as fine-tuning transitions from Phase 1 to Phase 2. Based on these findings, we experiment with a rank reduction (RR) strategy that facilitates rapid and stable fine-tuning of the latest CL-based methods. Through empirical investigations, we showcase the efficacy of RR in enhancing the performance and stability of five state-of-the-art sentence embedding methods. The code is available at (https://github.com/SNU-DRL/SentenceEmbedding_Rank).
引用
收藏
页码:159877 / 159888
页数:12
相关论文
共 37 条
  • [11] Fine-Tuning of Distil-BERT for Continual Learning in Text Classification: An Experimental Analysis
    Shah, Sahar
    Manzoni, Sara Lucia
    Zaman, Farooq
    Es Sabery, Fatima
    Epifania, Francesco
    Zoppis, Italo Francesco
    IEEE ACCESS, 2024, 12 : 104964 - 104982
  • [12] Efficient Neural Network Fine-Tuning via Layer Contribution Analysis
    Liu, Zhizhuo
    Zhou, Nanjian
    Liu, Min
    Liu, Zhibang
    Xu, Chaonong
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT IV, ICIC 2024, 2024, 14865 : 350 - 361
  • [13] Empirical analysis of the fine-tuning for Unsupervised Anomaly Detection in the ICT system
    Matsuo, Yoichi
    2023 19TH INTERNATIONAL CONFERENCE ON NETWORK AND SERVICE MANAGEMENT, CNSM, 2023,
  • [14] Unveiling the Power of Large Language Models: A Comparative Study of Retrieval-Augmented Generation, Fine-Tuning, and Their Synergistic Fusion for Enhanced Performance
    Budakoglu, Gulsum
    Emekci, Hakan
    IEEE ACCESS, 2025, 13 : 30936 - 30951
  • [15] Fine-tuning language model embeddings to reveal domain knowledge: An explainable artificial intelligence perspective on medical decision making
    Kraisnikovic, Ceca
    Harb, Robert
    Plass, Markus
    Al Zoughbi, Wael
    Holzinger, Andreas
    Mueller, Heimo
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2025, 139
  • [16] Low-Rank Fine-Tuning Meets Cross-modal Analysis: A Robust Framework for Age-Related Macular Degeneration Categorization
    Zhen, Baochen
    Qi, Yongbin
    Tang, Zizhen
    Liu, Chaoyong
    Zhao, Shilin
    Yu, Yansuo
    Liu, Qiang
    JOURNAL OF IMAGING INFORMATICS IN MEDICINE, 2025,
  • [17] Classification and Magnitude Estimation of Global and Local Seismic Events Using Conformer and Low-Rank Adaptation Fine-Tuning
    Jin, Yooseok
    Kim, Gwantae
    Ko, Hanseok
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2024, 21 : 1 - 5
  • [18] Fine-Tuning Llama 3 for Sentiment Analysis: Leveraging AWS Cloud for Enhanced Performance
    Shantanu Kumar
    Shruti Singh
    SN Computer Science, 5 (8)
  • [19] Analysis of Bias in GPT Language Models through Fine-tuning Containing Divergent Data
    Turi, Leandro Furlam
    Cavalini, Athus
    Comarela, Giovanni
    Oliveira-Santos, Thiago
    Badue, Claudine
    De Souza, Alberto F.
    2024 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN 2024, 2024,