Using large language models in psychology

被引:97
作者
Demszky, Dorottya [1 ]
Yang, Diyi [2 ]
Yeager, David [3 ,4 ]
Bryan, Christopher [3 ,5 ]
Clapper, Margarett [3 ,4 ]
Chandhok, Susannah [6 ]
Eichstaedt, Johannes [7 ,8 ]
Hecht, Cameron [3 ,4 ]
Jamieson, Jeremy [9 ]
Johnson, Meghann [3 ]
Jones, Michaela [3 ]
Krettek-Cobb, Danielle [6 ]
Lai, Leslie [6 ]
Jonesmitchell, Nirel [3 ]
Ong, Desmond [3 ,4 ]
Dweck, Carol [7 ]
Gross, James [7 ]
Pennebaker, James [4 ]
机构
[1] Stanford Univ, Grad Sch Educ, Stanford, CA 94305 USA
[2] Stanford Univ, Dept Comp Sci, Stanford, CA 94305 USA
[3] Univ Texas Austin, Texas Behav Sci & Policy Inst, Austin, TX 78712 USA
[4] Univ Texas Austin, Dept Psychol, Austin, TX 78712 USA
[5] Univ Texas Austin, Dept Business Govt & Soc, Austin, TX 78712 USA
[6] Google LLC, Mountain View, CA USA
[7] Stanford Univ, Dept Psychol, Stanford, CA USA
[8] Stanford Univ, Inst Human Ctr AI, Stanford, CA USA
[9] Univ Rochester, Dept Psychol, Rochester, NY USA
来源
NATURE REVIEWS PSYCHOLOGY | 2023年 / 2卷 / 11期
基金
美国国家卫生研究院; 美国国家科学基金会;
关键词
RESPONSES; MINDSET; STRESS;
D O I
10.1038/s44159-023-00241-5
中图分类号
B84 [心理学];
学科分类号
04 ; 0402 ;
摘要
Large language models (LLMs), such as OpenAI's GPT-4, Google's Bard or Meta's LLaMa, have created unprecedented opportunities for analysing and generating language data on a massive scale. Because language data have a central role in all areas of psychology, this new technology has the potential to transform the field. In this Perspective, we review the foundations of LLMs. We then explain how the way that LLMs are constructed enables them to effectively generate human-like linguistic output without the ability to think or feel like a human. We argue that although LLMs have the potential to advance psychological measurement, experimentation and practice, they are not yet ready for many of the most transformative psychological applications - but further research and development may enable such use. Next, we examine four major concerns about the application of LLMs to psychology, and how each might be overcome. Finally, we conclude with recommendations for investments that could help to address these concerns: field-initiated 'keystone' datasets; increased standardization of performance benchmarks; and shared computing and analysis infrastructure to ensure that the future of LLM-powered research is equitable. Large language models (LLMs), which can generate and score text in human-like ways, have the potential to advance psychological measurement, experimentation and practice. In this Perspective, Demszky and colleagues describe how LLMs work, concerns about using them for psychological purposes, and how these concerns might be addressed.
引用
收藏
页码:688 / 701
页数:14
相关论文
共 118 条
  • [51] Promoting rich discussions in mathematics classrooms: Using personalized, automated feedback to support reflection and instructional change
    Jacobs, Jennifer
    Scornavacco, Karla
    Harty, Charis
    Suresh, Abhijit
    Lai, Vivian
    Sumner, Tamara
    [J]. TEACHING AND TEACHER EDUCATION, 2022, 112
  • [52] Treating Stimuli as a Random Factor in Social Psychology: A New and Comprehensive Solution to a Pervasive but Largely Ignored Problem
    Judd, Charles M.
    Westfall, Jacob
    Kenny, David A.
    [J]. JOURNAL OF PERSONALITY AND SOCIAL PSYCHOLOGY, 2012, 103 (01) : 54 - 69
  • [53] Working With AI to Persuade: Examining a Large Language Model's Ability to Generate Pro-Vaccination Messages
    Karinshak E.
    Liu S.X.
    Park J.S.
    Hancock J.T.
    [J]. Proceedings of the ACM on Human-Computer Interaction, 2023, 7 (CSCW1)
  • [54] Lai CK, 2014, J EXP PSYCHOL GEN, V143, P1765, DOI 10.1037/a0036260
  • [55] Lai VV, 2023, Arxiv, DOI arXiv:2301.09656
  • [56] An introduction to latent semantic analysis
    Landauer, TK
    Foltz, PW
    Laham, D
    [J]. DISCOURSE PROCESSES, 1998, 25 (2-3) : 259 - 284
  • [57] A solution to Plato's problem: The latent semantic analysis theory of acquisition, induction, and representation of knowledge
    Landauer, TK
    Dumais, ST
    [J]. PSYCHOLOGICAL REVIEW, 1997, 104 (02) : 211 - 240
  • [58] Landi H., 2023, Fierce Healthcare
  • [59] Pre-train, Prompt, and Predict: A Systematic Survey of Prompting Methods in Natural Language Processing
    Liu, Pengfei
    Yuan, Weizhe
    Fu, Jinlan
    Jiang, Zhengbao
    Hayashi, Hiroaki
    Neubig, Graham
    [J]. ACM COMPUTING SURVEYS, 2023, 55 (09)
  • [60] Liu X, 2022, PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022): (SHORT PAPERS), VOL 2, P61