Prompt Programming for Large Language Models: Beyond the Few-Shot Paradigm

被引:359
作者
Reynolds, Laria [1 ]
McDonell, Kyle [1 ]
机构
[1] KNC Ai, Atwater, CA 95301 USA
来源
EXTENDED ABSTRACTS OF THE 2021 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS (CHI'21) | 2021年
关键词
language models; transformers; GPT-3; few-shot learning; prompt programming; metaprompts; serial reasoning; semiotics;
D O I
10.1145/3411763.3451760
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Prevailing methods for mapping large generative language models to supervised tasks may fail to sufficiently probe models' novel capabilities. Using GPT-3 as a case study, we show that 0-shot prompts can significantly outperform few-shot prompts. We suggest that the function of few-shot examples in these cases is better described as locating an already learned task rather than meta-learning. This analysis motivates rethinking the role of prompts in controlling and evaluating powerful language models. We discuss methods of prompt programming, emphasizing the usefulness of considering prompts through the lens of natural language. We explore techniques for exploiting the capacity of narratives and cultural anchors to encode nuanced intentions and techniques for encouraging deconstruction of a problem into components before producing a verdict. Informed by this more encompassing theory of prompt programming, we also introduce the idea of a metaprompt that seeds the model to generate its own natural language prompts for a range of tasks. Finally, we discuss how these more general methods of interacting with language models can be incorporated into existing and future benchmarks and practical applications.
引用
收藏
页数:7
相关论文
共 27 条
[1]  
Bojar O., 2014, P 9 WORKSHOP STAT MA, DOI DOI 10.3115/V1/W14-3302
[2]  
Brown TB, 2020, ADV NEUR IN, V33
[3]  
Cote Marc-Alexandre, 2019, ARXIV180611532CSLG
[4]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[5]  
Fan Angela, 2018, ARXIV180504833CSCL
[6]  
Gan Zhe, 2019, MULTISTEP REASONING
[7]  
Gao Leo, 2020, LEOGAO DEV
[8]  
Gao Tianyu, 2020, ARXIV201215723CSCL
[9]  
Gwern Branwen, 2020, Gwern. GPT-3 Creative Fiction
[10]  
Hendrycks Dan, 2021, INT C LEARNING REPRE