Disentangling the Effects of Counterfactual Feedback on Maximization and Risk Preference across Gains and Losses
Discuss this preprint
Start a discussion What are Sciety discussions?Listed in
This article is not in any list yet, why not save it to one of your lists.Abstract
Decisions under risk can be characterized by two complementary but distinct behavioral variables: the propensity to choose the expected value-maximizing option and the tendency to select the risky option. Human reinforcement learning (RL) research has traditionally focused more on the former (maximization), providing many insights into how fundamental features of the learning environment influence behavior. For instance, previous studies have shown that counterfactual feedback about unchosen options exerts a strong influence on maximization. However, comparatively little is known about scenarios in which maximization and risk propensity are assessed orthogonally and simultaneously. To address this gap, we designed a series of RL experiments in which we manipulated key factors, including the feedback information regime (partial vs. complete) and the relative expected values of risky and safe options. Contrary to the conventional assumption that “more information is better,” our results clearly show that, in such contexts, the effect of foregone outcomes on risk preference outweighs its influence on maximization. We further compared RL choices with description-based versions of the same decision problems. While replicating established findings on the description–experience gap, we found that complete information had little impact on its expression, suggesting that this gap does not arise primarily from instant feedback or insufficient outcome sampling.