A computational theory of learning moral weights
Discuss this preprint
Start a discussion What are Sciety discussions?Listed in
This article is not in any list yet, why not save it to one of your lists.Abstract
What determines whose welfare people consider in their moral decisions? We propose that people learn entity-specific moral weights through reinforcement learning (RL), where decision outcomes provide the learning signal. We formalize this in a computational model in which agents update their moral weights for different stakeholders based on whether considering those stakeholders’ welfare led to better- or worse-than-expected outcomes. To test this model, we simulate agents learning in market environments (which reward cooperation with strangers) versus non-market environments (which reward exploitation). We show that this mechanism is sufficient to explain two empirical phenomena linking market integration to prosociality: (1) cross-cultural variation in dictator game offers across small-scale societies, and (2) within-cultural variation in lost-letter return rates across 188 Italian municipalities. Together, our simulation results suggest that updating moral weights via RL may be an important mechanism of moral change at both the individual and the societal level.