Contrastive Meta-Learner for Automatic Text Labeling and Semantic Textual Similarity

被引:0
作者
Cooper, Ryan [1 ]
Kliesner, Kenneth W. [1 ]
Zenker, Stephen [1 ]
机构
[1] Sandia Natl Labs, Albuquerque, NM 87185 USA
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Labeling; Semantics; Task analysis; Encoding; Bidirectional control; Metalearning; Data models; Deep learning; Few shot learning; Semantic textual similarity; constrastive meta-learning; deep metric learning; few-shot learning; triplet loss; transformers; embedding space; natural language processing; named entity recognition;
D O I
10.1109/ACCESS.2024.3424401
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Generating large labeled datasets is a common barrier in machine learning efforts, with frequent challenges in both labeling and creating useful models for these datasets. We introduce a new approach to the automatic text labeling and semantic textual similarity tasks, which utilizes an encoder layer that is fine-tuned using triplet loss. This approach, contrastive meta-learning (CML), is specifically designed to create a naturally separable embedding space, based on minimal a priori examples. We find that through the use of CML, we are able to perform up to state-of-the-art performance on similar few-shot learning automatic labeling methodologies. For the semantic textual similarity task, CML creates a close approximation to a model trained with the full dataset, with as little as 8 training examples, whereas other common approaches require outside datasets.
引用
收藏
页码:166792 / 166799
页数:8
相关论文
共 16 条
  • [1] Bao Y., 2020, P INT C LEARN REPR, DOI [10.48550/arXiv.1908.06039, DOI 10.48550/ARXIV.1908.06039]
  • [2] Cer Daniel, 2017, 11 INT WORKSH SEM EV, P1, DOI [10.18653/v1/S17-2001, DOI 10.18653/V1/S17-2001]
  • [3] Chen JF, 2022, AAAI CONF ARTIF INTE, P10492
  • [4] Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
  • [5] Finn C, 2017, PR MACH LEARN RES, V70
  • [6] Gao TY, 2019, AAAI CONF ARTIF INTE, P6407
  • [7] Geng RY, 2019, 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019), P3904
  • [8] Han CC, 2021, FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, P1664
  • [9] Misra R, 2022, Arxiv, DOI arXiv:2209.11429
  • [10] Prasath N.., 2020, Tech. Rep. Reuters- 21578