Large language models and humans converge in judging public figures' personalities
- PMID: 39359393
- PMCID: PMC11443023
- DOI: 10.1093/pnasnexus/pgae418
Large language models and humans converge in judging public figures' personalities
Abstract
ChatGPT-4 and 600 human raters evaluated 226 public figures' personalities using the Ten-Item Personality Inventory. The correlation between ChatGPT-4 and aggregate human ratings ranged from r = 0.76 to 0.87, outperforming the models specifically trained to make such predictions. Notably, the model was not provided with any training data or feedback on its performance. We discuss the potential explanations and practical implications of ChatGPT-4's ability to mimic human responses accurately.
Keywords: AI; large language models; personality perception; zero-shot predictions.
© The Author(s) 2024. Published by Oxford University Press on behalf of National Academy of Sciences.
Figures


References
-
- Brown TB, et al. 2020. Language models are few-shot learners. arXiv, arXiv:2005.14165v4, preprint: not peer reviewed.
-
- Wei J, et al. 2022. Emergent abilities of large language models. arXiv, arXiv:2206.07682v2, preprint: not peer reviewed.
LinkOut - more resources
Full Text Sources