PAG-LLM: Paraphrase and Aggregate with Large Language Models for Minimizing Intent Classification Errors
被引:3
|
作者:
论文数: 引用数:
h-index:
机构:
Yadav, Vikas
[1
,2
]
论文数: 引用数:
h-index:
机构:
Tang, Zheng
[3
]
论文数: 引用数:
h-index:
机构:
Srinivasan, Vijay
[3
]
机构:
[1] Univ Arizona, Tucson, AZ 85721 USA
[2] ServiceNow, Santa Clara, CA 95054 USA
[3] Samsung Res Amer, Mountain View, CA USA
来源:
PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024
|
2024年
关键词:
Large Langauge Model;
Intent Classification;
Paraphrasing;
Generation confidence;
Aggregation;
D O I:
10.1145/3626772.3657959
中图分类号:
TP18 [人工智能理论];
学科分类号:
081104 ;
0812 ;
0835 ;
1405 ;
摘要:
Large language models (LLM) have achieved remarkable success in natural language generation but lesser focus has been given to their applicability in key tasks such as intent-classification. We show that LLMs like LLaMa can achieve high performance on intent classification tasks with large number of classes but still make classification errors and worse, generate out-of-vocabulary intent labels. To address these critical issues, we introduce Paraphrase and AGgregate (PAG)-LLM approach wherein an LLM generates multiple paraphrases of the input query (parallel queries), performs intent classification for the original query and each paraphrase, and at the end aggregate all the predicted intent labels based on their confidence scores. We evaluate PAG-LLM on two large intent classification datasets: CLINC, and Banking and show 22.7% and 15.1% error reduction. We showthat PAG-LLM is especially effective for hard examples where LLM is uncertain, and reduces the critical misclassification and hallucinated label generation errors.