Using large language models in psychology

被引:97
作者
Demszky, Dorottya [1 ]
Yang, Diyi [2 ]
Yeager, David [3 ,4 ]
Bryan, Christopher [3 ,5 ]
Clapper, Margarett [3 ,4 ]
Chandhok, Susannah [6 ]
Eichstaedt, Johannes [7 ,8 ]
Hecht, Cameron [3 ,4 ]
Jamieson, Jeremy [9 ]
Johnson, Meghann [3 ]
Jones, Michaela [3 ]
Krettek-Cobb, Danielle [6 ]
Lai, Leslie [6 ]
Jonesmitchell, Nirel [3 ]
Ong, Desmond [3 ,4 ]
Dweck, Carol [7 ]
Gross, James [7 ]
Pennebaker, James [4 ]
机构
[1] Stanford Univ, Grad Sch Educ, Stanford, CA 94305 USA
[2] Stanford Univ, Dept Comp Sci, Stanford, CA 94305 USA
[3] Univ Texas Austin, Texas Behav Sci & Policy Inst, Austin, TX 78712 USA
[4] Univ Texas Austin, Dept Psychol, Austin, TX 78712 USA
[5] Univ Texas Austin, Dept Business Govt & Soc, Austin, TX 78712 USA
[6] Google LLC, Mountain View, CA USA
[7] Stanford Univ, Dept Psychol, Stanford, CA USA
[8] Stanford Univ, Inst Human Ctr AI, Stanford, CA USA
[9] Univ Rochester, Dept Psychol, Rochester, NY USA
来源
NATURE REVIEWS PSYCHOLOGY | 2023年 / 2卷 / 11期
基金
美国国家卫生研究院; 美国国家科学基金会;
关键词
RESPONSES; MINDSET; STRESS;
D O I
10.1038/s44159-023-00241-5
中图分类号
B84 [心理学];
学科分类号
04 ; 0402 ;
摘要
Large language models (LLMs), such as OpenAI's GPT-4, Google's Bard or Meta's LLaMa, have created unprecedented opportunities for analysing and generating language data on a massive scale. Because language data have a central role in all areas of psychology, this new technology has the potential to transform the field. In this Perspective, we review the foundations of LLMs. We then explain how the way that LLMs are constructed enables them to effectively generate human-like linguistic output without the ability to think or feel like a human. We argue that although LLMs have the potential to advance psychological measurement, experimentation and practice, they are not yet ready for many of the most transformative psychological applications - but further research and development may enable such use. Next, we examine four major concerns about the application of LLMs to psychology, and how each might be overcome. Finally, we conclude with recommendations for investments that could help to address these concerns: field-initiated 'keystone' datasets; increased standardization of performance benchmarks; and shared computing and analysis infrastructure to ensure that the future of LLM-powered research is equitable. Large language models (LLMs), which can generate and score text in human-like ways, have the potential to advance psychological measurement, experimentation and practice. In this Perspective, Demszky and colleagues describe how LLMs work, concerns about using them for psychological purposes, and how these concerns might be addressed.
引用
收藏
页码:688 / 701
页数:14
相关论文
共 118 条
  • [71] Metz Cade., 2023, New York Times
  • [72] Michelmann S, 2023, Arxiv, DOI [arXiv:2301.10297, DOI 10.48550/ARXIV.2301.10297, 10.48550/arXiv.2301.10297]
  • [73] Megastudies improve the impact of applied behavioural science
    Milkman, Katherine L.
    Gromet, Dena
    Ho, Hung
    Kay, Joseph S.
    Lee, Timothy W.
    Pandiloski, Pepi
    Park, Yeji
    Rai, Aneesh
    Bazerman, Max
    Beshears, John
    Bonacorsi, Lauri
    Camerer, Colin
    Chang, Edward
    Chapman, Gretchen
    Cialdini, Robert
    Dai, Hengchen
    Eskreis-Winkler, Lauren
    Fishbach, Ayelet
    Gross, James J.
    Horn, Samantha
    Hubbard, Alexa
    Jones, Steven J.
    Karlan, Dean
    Kautz, Tim
    Kirgios, Erika
    Klusowski, Joowon
    Kristal, Ariella
    Ladhania, Rahul
    Loewenstein, George
    Ludwig, Jens
    Mellers, Barbara
    Mullainathan, Sendhil
    Saccardo, Silvia
    Spiess, Jann
    Suri, Gaurav
    Talloen, Joachim H.
    Taxer, Jamie
    Trope, Yaacov
    Ungar, Lyle
    Volpp, Kevin G.
    Whillans, Ashley
    Zinman, Jonathan
    Duckworth, Angela L.
    [J]. NATURE, 2021, 600 (7889) : 478 - +
  • [74] Monarch RM., 2021, Human-In-the-Loop Machine Learning: Active Learning and Annotation for HumanCentered AI
  • [75] Linguistic measures of psychological distance track symptom levels and treatment outcomes in a large set of psychotherapy transcripts
    Nook, Erik C.
    Hull, Thomas D.
    Nock, Matthew K.
    Somerville, Leah H.
    [J]. PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2022, 119 (13)
  • [76] Novikova Jekaterina., 2017, P 2017 C EMP METH NA, P2241, DOI [DOI 10.18653/V1/D17-1238, 10.18653/v1/D17-1238]
  • [77] Deep neural networks detect suicide risk from textual facebook posts
    Ophir, Yaakov
    Tikochinski, Refael
    Asterhan, Christa S. C.
    Sisso, Itay
    Reichart, Roi
    [J]. SCIENTIFIC REPORTS, 2020, 10 (01)
  • [78] Ouyang L, 2022, ADV NEUR IN
  • [79] Social Simulacra: Creating Populated Prototypes for Social Computing Systems
    Park, Joon Sung
    Popowski, Lindsay
    Cai, Carrie J.
    Morris, Meredith Ringel
    Liang, Percy
    Bernstein, Michael S.
    [J]. PROCEEDINGS OF THE 35TH ANNUAL ACM SYMPOSIUM ON USER INTERFACE SOFTWARE AND TECHNOLOGY, UIST 2022, 2022,
  • [80] Data and its (dis)contents: A survey of dataset development and use in machine learning research
    Paullada, Amandalynne
    Raji, Inioluwa Deborah
    Bender, Emily M.
    Denton, Emily
    Hanna, Alex
    [J]. PATTERNS, 2021, 2 (11):