» Articles » PMID: 33141824

A Recurrent Neural Network Framework for Flexible and Adaptive Decision Making Based on Sequence Learning

Overview
Specialty Biology
Date 2020 Nov 3
PMID 33141824
Citations 5
Authors
Affiliations
Soon will be listed here.
Abstract

The brain makes flexible and adaptive responses in a complicated and ever-changing environment for an organism's survival. To achieve this, the brain needs to understand the contingencies between its sensory inputs, actions, and rewards. This is analogous to the statistical inference that has been extensively studied in the natural language processing field, where recent developments of recurrent neural networks have found many successes. We wonder whether these neural networks, the gated recurrent unit (GRU) networks in particular, reflect how the brain solves the contingency problem. Therefore, we build a GRU network framework inspired by the statistical learning approach of NLP and test it with four exemplar behavior tasks previously used in empirical studies. The network models are trained to predict future events based on past events, both comprising sensory, action, and reward events. We show the networks can successfully reproduce animal and human behavior. The networks generalize the training, perform Bayesian inference in novel conditions, and adapt their choices when event contingencies vary. Importantly, units in the network encode task variables and exhibit activity patterns that match previous neurophysiology findings. Our results suggest that the neural network approach based on statistical sequence learning may reflect the brain's computational principle underlying flexible and adaptive behaviors and serve as a useful approach to understand the brain.

Citing Articles

Sequential neuronal processing of number values, abstract decision, and action in the primate prefrontal cortex.

Viswanathan P, Stein A, Nieder A PLoS Biol. 2024; 22(2):e3002520.

PMID: 38364194 PMC: 10871863. DOI: 10.1371/journal.pbio.3002520.


Increasing comprehensiveness and reducing workload in a systematic review of complex interventions using automated machine learning.

Uthman O, Court R, Enderby J, Al-Khudairy L, Nduka C, Mistry H Health Technol Assess. 2022; .

PMID: 36562494 PMC: 10068584. DOI: 10.3310/UDIR6682.


Category learning in a recurrent neural network with reinforcement learning.

Zhang Y, Pan X, Wang Y Front Psychiatry. 2022; 13:1008011.

PMID: 36387007 PMC: 9640766. DOI: 10.3389/fpsyt.2022.1008011.


An Evaluation of 3D-Printed Materials' Structural Properties Using Active Infrared Thermography and Deep Neural Networks Trained on the Numerical Data.

Szymanik B Materials (Basel). 2022; 15(10).

PMID: 35629753 PMC: 9146560. DOI: 10.3390/ma15103727.


Gated recurrence enables simple and accurate sequence prediction in stochastic, changing, and structured environments.

Foucault C, Meyniel F Elife. 2021; 10.

PMID: 34854377 PMC: 8735865. DOI: 10.7554/eLife.71801.

References
1.
Ding L, Gold J . Caudate encodes multiple computations for perceptual decisions. J Neurosci. 2010; 30(47):15747-59. PMC: 3005761. DOI: 10.1523/JNEUROSCI.2894-10.2010. View

2.
Zhang Z, Cheng Z, Lin Z, Nie C, Yang T . A neural network model for the orbitofrontal cortex and task space acquisition during reinforcement learning. PLoS Comput Biol. 2018; 14(1):e1005925. PMC: 5771635. DOI: 10.1371/journal.pcbi.1005925. View

3.
Florian R . Reinforcement learning through modulation of spike-timing-dependent synaptic plasticity. Neural Comput. 2007; 19(6):1468-502. DOI: 10.1162/neco.2007.19.6.1468. View

4.
Kiani R, Shadlen M . Representation of confidence associated with a decision by neurons in the parietal cortex. Science. 2009; 324(5928):759-64. PMC: 2738936. DOI: 10.1126/science.1169405. View

5.
Yang G, Joglekar M, Song H, Newsome W, Wang X . Task representations in neural networks trained to perform many cognitive tasks. Nat Neurosci. 2019; 22(2):297-306. PMC: 11549734. DOI: 10.1038/s41593-018-0310-2. View