Using large language models in psychology

被引:97
作者
Demszky, Dorottya [1 ]
Yang, Diyi [2 ]
Yeager, David [3 ,4 ]
Bryan, Christopher [3 ,5 ]
Clapper, Margarett [3 ,4 ]
Chandhok, Susannah [6 ]
Eichstaedt, Johannes [7 ,8 ]
Hecht, Cameron [3 ,4 ]
Jamieson, Jeremy [9 ]
Johnson, Meghann [3 ]
Jones, Michaela [3 ]
Krettek-Cobb, Danielle [6 ]
Lai, Leslie [6 ]
Jonesmitchell, Nirel [3 ]
Ong, Desmond [3 ,4 ]
Dweck, Carol [7 ]
Gross, James [7 ]
Pennebaker, James [4 ]
机构
[1] Stanford Univ, Grad Sch Educ, Stanford, CA 94305 USA
[2] Stanford Univ, Dept Comp Sci, Stanford, CA 94305 USA
[3] Univ Texas Austin, Texas Behav Sci & Policy Inst, Austin, TX 78712 USA
[4] Univ Texas Austin, Dept Psychol, Austin, TX 78712 USA
[5] Univ Texas Austin, Dept Business Govt & Soc, Austin, TX 78712 USA
[6] Google LLC, Mountain View, CA USA
[7] Stanford Univ, Dept Psychol, Stanford, CA USA
[8] Stanford Univ, Inst Human Ctr AI, Stanford, CA USA
[9] Univ Rochester, Dept Psychol, Rochester, NY USA
来源
NATURE REVIEWS PSYCHOLOGY | 2023年 / 2卷 / 11期
基金
美国国家卫生研究院; 美国国家科学基金会;
关键词
RESPONSES; MINDSET; STRESS;
D O I
10.1038/s44159-023-00241-5
中图分类号
B84 [心理学];
学科分类号
04 ; 0402 ;
摘要
Large language models (LLMs), such as OpenAI's GPT-4, Google's Bard or Meta's LLaMa, have created unprecedented opportunities for analysing and generating language data on a massive scale. Because language data have a central role in all areas of psychology, this new technology has the potential to transform the field. In this Perspective, we review the foundations of LLMs. We then explain how the way that LLMs are constructed enables them to effectively generate human-like linguistic output without the ability to think or feel like a human. We argue that although LLMs have the potential to advance psychological measurement, experimentation and practice, they are not yet ready for many of the most transformative psychological applications - but further research and development may enable such use. Next, we examine four major concerns about the application of LLMs to psychology, and how each might be overcome. Finally, we conclude with recommendations for investments that could help to address these concerns: field-initiated 'keystone' datasets; increased standardization of performance benchmarks; and shared computing and analysis infrastructure to ensure that the future of LLM-powered research is equitable. Large language models (LLMs), which can generate and score text in human-like ways, have the potential to advance psychological measurement, experimentation and practice. In this Perspective, Demszky and colleagues describe how LLMs work, concerns about using them for psychological purposes, and how these concerns might be addressed.
引用
收藏
页码:688 / 701
页数:14
相关论文
共 118 条
  • [1] Achiam J., 2023, Gpt-4 technical report
  • [2] Achiam OJ, 2023, Arxiv, DOI [arXiv:2303.08774, DOI 10.48550/ARXIV.2303.08774]
  • [3] Aher G, 2023, PR MACH LEARN RES, V202, P337
  • [4] Aher G, 2022, Arxiv, DOI [arXiv:2208.10264, 10.48550/arXiv.2208.10264, DOI 10.48550/ARXIV.2208.10264]
  • [5] Short-term prediction of suicidal thoughts and behaviors in adolescents: Can recent developments in technology and computational science provide a breakthrough?
    Allen, Nicholas B.
    Nelson, Benjamin W.
    Brent, David
    Auerbach, Randy P.
    [J]. JOURNAL OF AFFECTIVE DISORDERS, 2019, 250 : 163 - 169
  • [6] [Anonymous], 2023, ChatGPT used by teachers more than students, new survey from Walton Family Foundation Finds
  • [7] Out of One, Many: Using Language Models to Simulate Human Samples
    Argyle, Lisa P. P.
    Busby, Ethan C. C.
    Fulda, Nancy
    Gubler, Joshua R. R.
    Rytting, Christopher
    Wingate, David
    [J]. POLITICAL ANALYSIS, 2023, 31 (03) : 337 - 351
  • [8] The power of self-persuasion
    Aronson, E
    [J]. AMERICAN PSYCHOLOGIST, 1999, 54 (11) : 875 - 884
  • [9] Social media conversations reveal large psychological shifts caused by COVID-19's onset across US cities
    Ashokkumar, Ashwini
    Pennebaker, James W.
    [J]. SCIENCE ADVANCES, 2021, 7 (39):
  • [10] Bai YT, 2022, Arxiv, DOI [arXiv:2212.08073, DOI 10.48550/ARXIV.2212.08073]