Unveiling Key Aspects of Fine-Tuning in Sentence Embeddings: A Representation Rank Analysis

被引:0
作者
Jung, Euna [1 ]
Kim, Jaeill [2 ]
Ko, Jungmin [3 ]
Park, Jinwoo [1 ]
Rhee, Wonjong [3 ,4 ,5 ]
机构
[1] Samsung Adv Inst Technol, Suwon 16678, Gyeonggi Do, South Korea
[2] LINE Investment Technol, Seongnam Si 13529, Gyeonggi Do, South Korea
[3] Seoul Natl Univ, Interdisciplinary Program Artificial Intelligence, Seoul 08826, South Korea
[4] Seoul Natl Univ, Dept Intelligence & Informat, Seoul 08826, South Korea
[5] Seoul Natl Univ, RICS, Seoul 08826, South Korea
来源
IEEE ACCESS | 2024年 / 12卷
基金
新加坡国家研究基金会;
关键词
Training; Linguistics; Contrastive learning; Market research; Correlation; Semantics; Visualization; Phase measurement; Natural language processing; Loss measurement; Sentence embedding; self-supervised learning; contrastive learning; fine-tuning; representation rank;
D O I
10.1109/ACCESS.2024.3485705
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The latest advancements in unsupervised learning of sentence embeddings predominantly involve employing contrastive learning-based (CL-based) fine-tuning over pre-trained language models. In this study, we analyze the latest sentence embedding methods by adopting representation rank as the primary tool of analysis. We first define Phase 1 and Phase 2 of fine-tuning based on when representation rank peaks. Utilizing these phases, we conduct a thorough analysis and obtain essential findings across key aspects, including alignment and uniformity, linguistic abilities, and correlation between performance and rank. For instance, we find that the dynamics of the key aspects can undergo significant changes as fine-tuning transitions from Phase 1 to Phase 2. Based on these findings, we experiment with a rank reduction (RR) strategy that facilitates rapid and stable fine-tuning of the latest CL-based methods. Through empirical investigations, we showcase the efficacy of RR in enhancing the performance and stability of five state-of-the-art sentence embedding methods. The code is available at (https://github.com/SNU-DRL/SentenceEmbedding_Rank).
引用
收藏
页码:159877 / 159888
页数:12
相关论文
共 37 条
  • [31] Adaptive Pre-Training and Collaborative Fine-Tuning: A Win-Win Strategy to Improve Review Analysis Tasks
    Mao, Qianren
    Li, Jianxin
    Lin, Chenghua
    Chen, Congwen
    Peng, Hao
    Wang, Lihong
    Yu, Philip S.
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2022, 30 : 622 - 634
  • [32] Exploiting Fine-tuning of Self-supervised Learning Models for Improving Bi-modal Sentiment Analysis and Emotion Recognition
    Yang, Wei
    Fukayama, Satoru
    Heracleous, Panikos
    Ogata, Jun
    INTERSPEECH 2022, 2022, : 1998 - 2002
  • [33] Exploring the potential of using ChatGPT for rhetorical move-step analysis: The impact of prompt refinement, few-shot learning, and fine-tuning
    Kim, Minjin
    Lu, Xiaofei
    JOURNAL OF ENGLISH FOR ACADEMIC PURPOSES, 2024, 71
  • [34] Implementation of a Whisper Architecture-Based Turkish Automatic Speech Recognition (ASR) System and Evaluation of the Effect of Fine-Tuning with a Low-Rank Adaptation (LoRA) Adapter on Its Performance
    Polat, Hueseyin
    Turan, Alp Kaan
    Kocak, Cemal
    Ulas, Hasan Basri
    ELECTRONICS, 2024, 13 (21)
  • [35] What did the occupant say? Fine-tuning and evaluating a large language model for efficient analysis of multi-domain indoor environmental quality feedback
    Sadick, Abdul-Manan
    Chinazzo, Giorgia
    BUILDING AND ENVIRONMENT, 2025, 274
  • [36] An Empirical Evaluation of the Zero-Shot, Few-Shot, and Traditional Fine-Tuning Based Pretrained Language Models for Sentiment Analysis in Software Engineering
    Shafikuzzaman, Md
    Islam, Md Rakibul
    Rolli, Alex C.
    Akhter, Sharmin
    Seliya, Naeem
    IEEE ACCESS, 2024, 12 : 109714 - 109734
  • [37] Fine-tuning adaptive stochastic optimizers: determining the optimal hyperparameter ϵ\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\epsilon$$\end{document} via gradient magnitude histogram analysis
    Gustavo Silva
    Paul Rodriguez
    Neural Computing and Applications, 2024, 36 (35) : 22223 - 22243