» Articles » PMID: 16596980

Dynamic Response-by-response Models of Matching Behavior in Rhesus Monkeys

Overview
Date 2006 Apr 7
PMID 16596980
Citations 223
Authors
Affiliations
Soon will be listed here.
Abstract

We studied the choice behavior of 2 monkeys in a discrete-trial task with reinforcement contingencies similar to those Herrnstein (1961) used when he described the matching law. In each session, the monkeys experienced blocks of discrete trials at different relative-reinforcer frequencies or magnitudes with unsignalled transitions between the blocks. Steady-state data following adjustment to each transition were well characterized by the generalized matching law; response ratios undermatched reinforcer frequency ratios but matched reinforcer magnitude ratios. We modelled response-by-response behavior with linear models that used past reinforcers as well as past choices to predict the monkeys' choices on each trial. We found that more recently obtained reinforcers more strongly influenced choice behavior. Perhaps surprisingly, we also found that the monkeys' actions were influenced by the pattern of their own past choices. It was necessary to incorporate both past reinforcers and past choices in order to accurately capture steady-state behavior as well as the fluctuations during block transitions and the response-by-response patterns of behavior. Our results suggest that simple reinforcement learning models must account for the effects of past choices to accurately characterize behavior in this task, and that models with these properties provide a conceptual tool for studying how both past reinforcers and past choices are integrated by the neural systems that generate behavior.

Citing Articles

Striatal arbitration between choice strategies guides few-shot adaptation.

Yang M, Jung M, Lee S Nat Commun. 2025; 16(1):1811.

PMID: 39979316 PMC: 11842591. DOI: 10.1038/s41467-025-57049-5.


Signatures of Perseveration and Heuristic-Based Directed Exploration in Two-Step Sequential Decision Task Behaviour.

Brands A, Mathar D, Peters J Comput Psychiatr. 2025; 9(1):39-62.

PMID: 39959565 PMC: 11827566. DOI: 10.5334/cpsy.101.


Contributions of Attention to Learning in Multidimensional Reward Environments.

Wang M, Soltani A J Neurosci. 2024; 45(7).

PMID: 39681464 PMC: 11823339. DOI: 10.1523/JNEUROSCI.2300-23.2024.


Computational and Neural Evidence for Altered Fast and Slow Learning from Losses in Problem Gambling.

Iigaya K, Larsen T, Fong T, ODoherty J J Neurosci. 2024; 45(1).

PMID: 39557579 PMC: 11694394. DOI: 10.1523/JNEUROSCI.0080-24.2024.


Pupillary responses to directional uncertainty while intercepting a moving target.

Marquez I, Trevino M R Soc Open Sci. 2024; 11(10):240606.

PMID: 39359460 PMC: 11444787. DOI: 10.1098/rsos.240606.


References
1.
Baum W . Matching, undermatching, and overmatching in studies of choice. J Exp Anal Behav. 1979; 32(2):269-81. PMC: 1332902. DOI: 10.1901/jeab.1979.32-269. View

2.
Sugrue L, Corrado G, Newsome W . Matching behavior and the representation of value in the parietal cortex. Science. 2004; 304(5678):1782-7. DOI: 10.1126/science.1094765. View

3.
Palya W, Walter D, Kessel R, Lucke R . Investigating Behavioral Dynamics With A Fixed-time Extinction Schedule And Linear Analysis. J Exp Anal Behav. 1996; 66(3):391-409. PMC: 1284580. DOI: 10.1901/jeab.1996.66-391. View

4.
Tanji J, Hoshi E . Behavioral planning in the prefrontal cortex. Curr Opin Neurobiol. 2001; 11(2):164-70. DOI: 10.1016/s0959-4388(00)00192-6. View

5.
Shadlen M, Newsome W . Neural basis of a perceptual decision in the parietal cortex (area LIP) of the rhesus monkey. J Neurophysiol. 2001; 86(4):1916-36. DOI: 10.1152/jn.2001.86.4.1916. View