Leveraging Pretrained Language Models for Enhanced Entity Matching: A Comprehensive Study of Fine-Tuning and Prompt Learning Paradigms
被引:0
作者:
Wang, Yu
论文数: 0引用数: 0
h-index: 0
机构:
Anhui Med Univ, Sch Biomed Engn, Hefei 230001, Peoples R ChinaAnhui Med Univ, Sch Biomed Engn, Hefei 230001, Peoples R China
Wang, Yu
[1
]
Zhou, Luyao
论文数: 0引用数: 0
h-index: 0
机构:
Anhui Med Univ, Sch Biomed Engn, Hefei 230001, Peoples R ChinaAnhui Med Univ, Sch Biomed Engn, Hefei 230001, Peoples R China
Zhou, Luyao
[1
]
Wang, Yuan
论文数: 0引用数: 0
h-index: 0
机构:
Univ Sci & Technol China, Inst Adv Technol, Hefei 230001, Peoples R China
Anhui HYJK Med Technol Co Ltd, Hefei 230001, Peoples R ChinaAnhui Med Univ, Sch Biomed Engn, Hefei 230001, Peoples R China
Wang, Yuan
[2
,3
]
Peng, Zhenwan
论文数: 0引用数: 0
h-index: 0
机构:
Anhui Med Univ, Sch Biomed Engn, Hefei 230001, Peoples R ChinaAnhui Med Univ, Sch Biomed Engn, Hefei 230001, Peoples R China
Peng, Zhenwan
[1
]
机构:
[1] Anhui Med Univ, Sch Biomed Engn, Hefei 230001, Peoples R China
[2] Univ Sci & Technol China, Inst Adv Technol, Hefei 230001, Peoples R China
[3] Anhui HYJK Med Technol Co Ltd, Hefei 230001, Peoples R China
Computational linguistics - Learning systems - Natural language processing systems - Transfer learning - Zero-shot learning;
D O I:
10.1155/2024/1941221
中图分类号:
TP18 [人工智能理论];
学科分类号:
081104 ;
0812 ;
0835 ;
1405 ;
摘要:
Pretrained Language Models (PLMs) acquire rich prior semantic knowledge during the pretraining phase and utilize it to enhance downstream Natural Language Processing (NLP) tasks. Entity Matching (EM), a fundamental NLP task, aims to determine whether two entity records from different knowledge bases refer to the same real-world entity. This study, for the first time, explores the potential of using a PLM to boost the EM task through two transfer learning techniques, namely, fine-tuning and prompt learning. Our work also represents the first application of the soft prompt in an EM task. Experimental results across eleven EM datasets show that the soft prompt consistently outperforms other methods in terms of F1 scores across all datasets. Additionally, this study also investigates the capability of prompt learning in few-shot learning and observes that the hard prompt achieves the highest F1 scores in both zero-shot and one-shot context. These findings underscore the effectiveness of prompt learning paradigms in tackling challenging EM tasks.