Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2024 Jul 16;3(7):pgae245.
doi: 10.1093/pnasnexus/pgae245. eCollection 2024 Jul.

Perils and opportunities in using large language models in psychological research

Affiliations

Perils and opportunities in using large language models in psychological research

Suhaib Abdurahman et al. PNAS Nexus. .

Abstract

The emergence of large language models (LLMs) has sparked considerable interest in their potential application in psychological research, mainly as a model of the human psyche or as a general text-analysis tool. However, the trend of using LLMs without sufficient attention to their limitations and risks, which we rhetorically refer to as "GPTology", can be detrimental given the easy access to models such as ChatGPT. Beyond existing general guidelines, we investigate the current limitations, ethical implications, and potential of LLMs specifically for psychological research, and show their concrete impact in various empirical studies. Our results highlight the importance of recognizing global psychological diversity, cautioning against treating LLMs (especially in zero-shot settings) as universal solutions for text analysis, and developing transparent, open methods to address LLMs' opaque nature for reliable, reproducible, and robust inference from AI-generated data. Acknowledging LLMs' utility for task automation, such as text annotation, or to expand our understanding of human psychology, we argue for diversifying human samples and expanding psychology's methodological toolbox to promote an inclusive, generalizable science, countering homogenization, and over-reliance on LLMs.

Keywords: large language models; natural language processing; psychological diversity; psychological text analysis; psychology.

PubMed Disclaimer

Figures

Fig. 1.
Fig. 1.
ChatGPT vs. human moral judgments. Note: a) Distributions of moral judgments of humans (light blue) and GPT (light red) in six moral domains. Dashed lines represent averages. b) Inter-correlations between moral values in humans (N=3,902) and ChatGPT queries (N=1,000). c) Network of partial correlations between moral values based on a diverse sample of humans from 19 nations (30) and 1,000 queries of GPT. Blue edges represent positive partial correlations and red edges represent negative partial correlations.
Fig. 2.
Fig. 2.
Comparing ChatGPT against humans grouped by political opinion for responses on the Big Five Inventory. Note: Figure shows the response distribution of humans and ChatGPT across the five-factor personality constructs and for different human demographics. Figure shows that ChatGPT gives significantly higher responses on Agreeableness, Conscientiousness and significantly lower responses on Openness and Neuroticism. Importantly, ChatGPT shows significantly less variance compared with all demographic groups on all personality dimensions.
Fig. 3.
Fig. 3.
Comparing ChatGPT against humans across various demographic variables for the Right-Wing-Authoritarianism scale. Note: Figure shows the response distribution of humans and ChatGPT on the RWA scale for different human demographics. ChatGPT shows significantly lower average RWA than male, white, and young participants but not explicitly liberal participants. Importantly, ChatGPT shows significantly less variance compared with all demographic groups.

References

    1. Lazer D, et al. 2009. Computational social science. Science. 323(5915):721–723. - PMC - PubMed
    1. Grossmann I, et al. 2023. AI and the transformation of social science research. Science. 380(6650):1108–1109. - PubMed
    1. McClelland JL, Rumelhart DE. 1985. Distributed memory and the representation of general and specific information. J Exp Psychol Gen. 114(2):159–188. - PubMed
    1. Rumelhart DE, Hinton GE, McClelland JL. 1987. A general framework for parallel distributed processing. In: Rumelhart DE, McClelland JL, editors. Parallel distributed processing: explorations in the microstructure of cognition: foundations. Cambridge (MA): MIT Press. p. 45–76.
    1. Elman JL. 1990. Finding structure in time. Cogn Sci. 14(2):179–211.

LinkOut - more resources