» Articles » PMID: 33811071

Two Sides of the Same Coin: Beneficial and Detrimental Consequences of Range Adaptation in Human Reinforcement Learning

Overview
Journal Sci Adv
Specialties Biology
Science
Date 2021 Apr 3
PMID 33811071
Citations 13
Authors
Affiliations
Soon will be listed here.
Abstract

Evidence suggests that economic values are rescaled as a function of the range of the available options. Although locally adaptive, range adaptation has been shown to lead to suboptimal choices, particularly notable in reinforcement learning (RL) situations when options are extrapolated from their original context to a new one. Range adaptation can be seen as the result of an adaptive coding process aiming at increasing the signal-to-noise ratio. However, this hypothesis leads to a counterintuitive prediction: Decreasing task difficulty should increase range adaptation and, consequently, extrapolation errors. Here, we tested the paradoxical relation between range adaptation and performance in a large sample of participants performing variants of an RL task, where we manipulated task difficulty. Results confirmed that range adaptation induces systematic extrapolation errors and is stronger when decreasing task difficulty. Last, we propose a range-adapting model and show that it is able to parsimoniously capture all the behavioral results.

Citing Articles

Comparing experience- and description-based economic preferences across 11 countries.

Anllo H, Bavard S, Benmarrakchi F, Bonagura D, Cerrotti F, Cicue M Nat Hum Behav. 2024; 8(8):1554-1567.

PMID: 38877287 DOI: 10.1038/s41562-024-01894-9.


Foraging in a non-foraging task: Fitness maximization explains human risk preference dynamics under changing environment.

Mochizuki Y, Harasawa N, Aggarwal M, Chen C, Fukuda H PLoS Comput Biol. 2024; 20(5):e1012080.

PMID: 38739672 PMC: 11115364. DOI: 10.1371/journal.pcbi.1012080.


Recent Opioid Use Impedes Range Adaptation in Reinforcement Learning in Human Addiction.

Gueguen M, Anllo H, Bonagura D, Kong J, Hafezi S, Palminteri S Biol Psychiatry. 2023; 95(10):974-984.

PMID: 38101503 PMC: 11065633. DOI: 10.1016/j.biopsych.2023.12.005.


Intrinsic rewards explain context-sensitive valuation in reinforcement learning.

Molinaro G, Collins A PLoS Biol. 2023; 21(7):e3002201.

PMID: 37459394 PMC: 10374061. DOI: 10.1371/journal.pbio.3002201.


The functional form of value normalization in human reinforcement learning.

Bavard S, Palminteri S Elife. 2023; 12.

PMID: 37428155 PMC: 10393293. DOI: 10.7554/eLife.83891.


References
1.
Akaishi R, Umeda K, Nagase A, Sakai K . Autonomous mechanism of internal choice estimate underlies decision inertia. Neuron. 2013; 81(1):195-206. DOI: 10.1016/j.neuron.2013.10.018. View

2.
Thrailkill E, Trask S, Vidal P, Alcala J, Bouton M . Stimulus control of actions and habits: A role for reinforcer predictability and attention in the development of habitual behavior. J Exp Psychol Anim Learn Cogn. 2018; 44(4):370-384. PMC: 6233324. DOI: 10.1037/xan0000188. View

3.
Dumbalska T, Li V, Tsetsos K, Summerfield C . A map of decoy influence in human multialternative choice. Proc Natl Acad Sci U S A. 2020; 117(40):25169-25178. PMC: 7547229. DOI: 10.1073/pnas.2005058117. View

4.
Elliott R, Agnew Z, Deakin J . Medial orbitofrontal cortex codes relative rather than absolute value of financial rewards in humans. Eur J Neurosci. 2008; 27(9):2213-8. DOI: 10.1111/j.1460-9568.2008.06202.x. View

5.
Bavard S, Lebreton M, Khamassi M, Coricelli G, Palminteri S . Reference-point centering and range-adaptation enhance human reinforcement learning at the cost of irrational preferences. Nat Commun. 2018; 9(1):4503. PMC: 6206161. DOI: 10.1038/s41467-018-06781-2. View