Empirical Examination of Large Language Models in Regional Psychological Structures Simulation: Personality and Well-being

Read the full article See related articles

Discuss this preprint

Start a discussion What are Sciety discussions?

Listed in

This article is not in any list yet, why not save it to one of your lists.
Log in to save this article

Abstract

The rapid advancement of large language models (LLMs) had opened new avenues for simulating psychological structures at the population level. This study compared the performance of Kimi-Chat v1.5 (a China-trained model) and GPT-4 (a globally trained model) in reproducing regional psychological profiles in China, focusing on the Big-Five personality traits and subjective well-being. By using the 2018 China Family Panel Studies (CFPS 2018) as the benchmark for real human data, we assessed the fidelity of both LLMs in capturing regional variations across seven major Chinese regions. Results indicated that Kimi-Chat v1.5 more accurately replicated human responses, particularly in regions with distinct cultural characteristics, while GPT-4 showed significant discrepancies, particularly in well-being and openness. Our findings emphasized the importance of training-corpus lineage and suggested that culturally adapted LLMs could be a useful tool in regional psychological research. We discussed the implications of these findings for future applications and highlighted the limitations of current LLM capabilities in simulating human psychological complexity.

Article activity feed