Simulating human well-being with large language models: Systematic validation and misestimation across 64,000 individuals from 64 countries
Pat Pataranutaporn,
Nattavudh Powdthavee (),
Chayapatr Archiwaranguprok and
Pattie Maes
Additional contact information
Pat Pataranutaporn: a Media Lab, Massachusetts Institute of Technology , Cambridge , MA 02139-4307
Nattavudh Powdthavee: b Division of Economics, Nanyang Technological University , Singapore 639818 , Singapore
Chayapatr Archiwaranguprok: a Media Lab, Massachusetts Institute of Technology , Cambridge , MA 02139-4307
Pattie Maes: a Media Lab, Massachusetts Institute of Technology , Cambridge , MA 02139-4307
Proceedings of the National Academy of Sciences, 2025, vol. 122, issue 48, e2519394122
Abstract:
Subjective well-being is central to economic, medical, and policy decision-making. We evaluate whether large language models (LLMs) can provide valid predictions of well-being across global populations. Using natural-language profiles from 64,000 individuals in 64 countries, we benchmark four leading LLMs against self-reports and statistical models. Unlike regressions, which estimate relationships from survey data, LLMs draw only on individual characteristics (e.g., sociodemographic, attitudinal, and psychological factors) together with associations encoded during pretraining, rather than from the survey’s subjective well-being responses. They produced plausible patterns consistent with known correlates such as income and health, but systematically underperformed relative to regressions and showed the largest errors in underrepresented countries, reflecting biases rooted in global digital and economic inequality. A preregistered experiment revealed that LLMs rely on surface-level linguistic associations rather than conceptual understanding, leading to predictable distortions in unfamiliar contexts. Injecting contextual information partly reduced—but did not remove—these biases. These findings demonstrate that while LLMs can simulate broad correlates of life satisfaction, they fail to capture its experiential and cultural depth. Accordingly, they should not be used as substitutes for human self-reports of well-being; doing so would risk reinforcing inequality and undermining human agency.
Keywords: subjective well-being; large language models; global inequality; artificial intelligence; life satisfaction (search for similar items in EconPapers)
Date: 2025
References: Add references at CitEc
Citations:
Downloads: (external link)
https://doi.org/10.1073/pnas.2519394122 (application/pdf)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:nas:journl:v:122:y:2025:p:e2519394122
Access Statistics for this article
More articles in Proceedings of the National Academy of Sciences from Proceedings of the National Academy of Sciences
Bibliographic data for series maintained by PNAS Product Team ().