» Articles » PMID: 31185008

Models That Learn How Humans Learn: The Case of Decision-making and Its Disorders

Overview
Specialty Biology
Date 2019 Jun 12
PMID 31185008
Citations 20
Authors
Affiliations
Soon will be listed here.
Abstract

Popular computational models of decision-making make specific assumptions about learning processes that may cause them to underfit observed behaviours. Here we suggest an alternative method using recurrent neural networks (RNNs) to generate a flexible family of models that have sufficient capacity to represent the complex learning and decision- making strategies used by humans. In this approach, an RNN is trained to predict the next action that a subject will take in a decision-making task and, in this way, learns to imitate the processes underlying subjects' choices and their learning abilities. We demonstrate the benefits of this approach using a new dataset drawn from patients with either unipolar (n = 34) or bipolar (n = 33) depression and matched healthy controls (n = 34) making decisions on a two-armed bandit task. The results indicate that this new approach is better than baseline reinforcement-learning methods in terms of overall performance and its capacity to predict subjects' choices. We show that the model can be interpreted using off-policy simulations and thereby provides a novel clustering of subjects' learning processes-something that often eludes traditional approaches to modelling and behavioural analysis.

Citing Articles

An image-computable model of speeded decision-making.

Jaffe P, Santiago-Reyes G, Schafer R, Bissett P, Poldrack R Elife. 2025; 13.

PMID: 40019474 PMC: 11870652. DOI: 10.7554/eLife.98351.


Using recurrent neural network to estimate irreducible stochasticity in human choice behavior.

Ger Y, Shahar M, Shahar N Elife. 2024; 13.

PMID: 39240757 PMC: 11379453. DOI: 10.7554/eLife.90082.


Interaction between Functional Connectivity and Neural Excitability in Autism: A Novel Framework for Computational Modeling and Application to Biological Data.

Takahashi Y, Murata S, Ueki M, Tomita H, Yamashita Y Comput Psychiatr. 2024; 7(1):14-29.

PMID: 38774640 PMC: 11104370. DOI: 10.5334/cpsy.93.


Sensitivity to intrinsic rewards is domain general and related to mental health.

Blain B, Pinhorn I, Sharot T Nat Ment Health. 2024; 1(9):679-691.

PMID: 38665692 PMC: 11041740. DOI: 10.1038/s44220-023-00116-x.


Reviewing explore/exploit decision-making as a transdiagnostic target for psychosis, depression, and anxiety.

Lloyd A, Roiser J, Skeen S, Freeman Z, Badalova A, Agunbiade A Cogn Affect Behav Neurosci. 2024; 24(5):793-815.

PMID: 38653937 PMC: 11390819. DOI: 10.3758/s13415-024-01186-9.


References
1.
Yu A, Cohen J . Sequential effects: Superstition or rational behavior?. Adv Neural Inf Process Syst. 2015; 21:1873-1880. PMC: 4580342. View

2.
Song H, Yang G, Wang X . Reward-based training of recurrent neural networks for cognitive and value-based tasks. Elife. 2017; 6. PMC: 5293493. DOI: 10.7554/eLife.21492. View

3.
Montague P, Dolan R, Friston K, Dayan P . Computational psychiatry. Trends Cogn Sci. 2011; 16(1):72-80. PMC: 3556822. DOI: 10.1016/j.tics.2011.11.018. View

4.
ODoherty J, Hampton A, Kim H . Model-based fMRI and its application to reward learning and decision making. Ann N Y Acad Sci. 2007; 1104:35-53. DOI: 10.1196/annals.1390.022. View

5.
LeCun Y, Bengio Y, Hinton G . Deep learning. Nature. 2015; 521(7553):436-44. DOI: 10.1038/nature14539. View