» Articles » PMID: 33501916

A New Model of Decision Processing in Instrumental Learning Tasks

Overview
Journal Elife
Specialty Biology
Date 2021 Jan 27
PMID 33501916
Citations 17
Authors
Affiliations
Soon will be listed here.
Abstract

Learning and decision-making are interactive processes, yet cognitive modeling of error-driven learning and decision-making have largely evolved separately. Recently, evidence accumulation models (EAMs) of decision-making and reinforcement learning (RL) models of error-driven learning have been combined into joint RL-EAMs that can in principle address these interactions. However, we show that the most commonly used combination, based on the diffusion decision model (DDM) for binary choice, consistently fails to capture crucial aspects of response times observed during reinforcement learning. We propose a new RL-EAM based on an advantage racing diffusion (ARD) framework for choices among two or more options that not only addresses this problem but captures stimulus difficulty, speed-accuracy trade-off, and stimulus-response-mapping reversal effects. The RL-ARD avoids fundamental limitations imposed by the DDM on addressing effects of absolute values of choices, as well as extensions beyond binary choice, and provides a computationally tractable basis for wider applications.

Citing Articles

Joint modeling of choices and reaction times based on Bayesian contextual behavioral control.

Schwobel S, Markovic D, Smolka M, Kiebel S PLoS Comput Biol. 2024; 20(7):e1012228.

PMID: 38968304 PMC: 11290629. DOI: 10.1371/journal.pcbi.1012228.


How cortico-basal ganglia-thalamic subnetworks can shift decision policies to maximize reward rate.

Bahuguna J, Verstynen T, Rubin J bioRxiv. 2024; .

PMID: 38826315 PMC: 11142098. DOI: 10.1101/2024.05.21.595174.


EZ-CDM: Fast, simple, robust, and accurate estimation of circular diffusion model parameters.

Qarehdaghi H, Rad J Psychon Bull Rev. 2024; 31(5):2058-2091.

PMID: 38587755 DOI: 10.3758/s13423-024-02483-7.


The effects of non-diagnostic information on confidence and decision making.

Kohl A, Sauer J, Palmer M, Brooks J, Heathcote A Mem Cognit. 2024; 52(5):1182-1194.

PMID: 38489145 PMC: 11315710. DOI: 10.3758/s13421-024-01535-6.


Joint Modelling of Latent Cognitive Mechanisms Shared Across Decision-Making Domains.

Stevenson N, Innes R, Boag R, Miletic S, Isherwood S, Trutti A Comput Brain Behav. 2024; 7(1):1-22.

PMID: 38425991 PMC: 10899373. DOI: 10.1007/s42113-023-00192-3.


References
1.
Turner B, Sederberg P . A generalized, likelihood-free method for posterior estimation. Psychon Bull Rev. 2013; 21(2):227-50. PMC: 4143986. DOI: 10.3758/s13423-013-0530-0. View

2.
Trueblood J, Heathcote A, Evans N, Holmes W . Urgency, leakage, and the relative nature of information processing in decision-making. Psychol Rev. 2020; 128(1):160-186. DOI: 10.1037/rev0000255. View

3.
Palminteri S, Wyart V, Koechlin E . The Importance of Falsification in Computational Cognitive Modeling. Trends Cogn Sci. 2017; 21(6):425-433. DOI: 10.1016/j.tics.2017.03.011. View

4.
Collins A, Frank M . Within- and across-trial dynamics of human EEG reveal cooperative interplay between reinforcement learning and working memory. Proc Natl Acad Sci U S A. 2018; 115(10):2502-2507. PMC: 5877949. DOI: 10.1073/pnas.1720963115. View

5.
Pedersen M, Frank M, Biele G . The drift diffusion model as the choice rule in reinforcement learning. Psychon Bull Rev. 2016; 24(4):1234-1251. PMC: 5487295. DOI: 10.3758/s13423-016-1199-y. View