PAG-LLM: Paraphrase and Aggregate with Large Language Models for Minimizing Intent Classification Errors

被引:3
|
作者
Yadav, Vikas [1 ,2 ]
Tang, Zheng [3 ]
Srinivasan, Vijay [3 ]
机构
[1] Univ Arizona, Tucson, AZ 85721 USA
[2] ServiceNow, Santa Clara, CA 95054 USA
[3] Samsung Res Amer, Mountain View, CA USA
来源
PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024 | 2024年
关键词
Large Langauge Model; Intent Classification; Paraphrasing; Generation confidence; Aggregation;
D O I
10.1145/3626772.3657959
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large language models (LLM) have achieved remarkable success in natural language generation but lesser focus has been given to their applicability in key tasks such as intent-classification. We show that LLMs like LLaMa can achieve high performance on intent classification tasks with large number of classes but still make classification errors and worse, generate out-of-vocabulary intent labels. To address these critical issues, we introduce Paraphrase and AGgregate (PAG)-LLM approach wherein an LLM generates multiple paraphrases of the input query (parallel queries), performs intent classification for the original query and each paraphrase, and at the end aggregate all the predicted intent labels based on their confidence scores. We evaluate PAG-LLM on two large intent classification datasets: CLINC, and Banking and show 22.7% and 15.1% error reduction. We showthat PAG-LLM is especially effective for hard examples where LLM is uncertain, and reduces the critical misclassification and hallucinated label generation errors.
引用
收藏
页码:2569 / 2573
页数:5
相关论文
empty
未找到相关数据