Evolution of Cooperation Facilitated by Reinforcement Learning with Adaptive Aspiration Levels
Overview
Authors
Affiliations
Repeated interaction between individuals is the main mechanism for maintaining cooperation in social dilemma situations. Variants of tit-for-tat (repeating the previous action of the opponent) and the win-stay lose-shift strategy are known as strong competitors in iterated social dilemma games. On the other hand, real repeated interaction generally allows plasticity (i.e., learning) of individuals based on the experience of the past. Although plasticity is relevant to various biological phenomena, its role in repeated social dilemma games is relatively unexplored. In particular, if experience-based learning plays a key role in promotion and maintenance of cooperation, learners should evolve in the contest with nonlearners under selection pressure. By modeling players using a simple reinforcement learning model, we numerically show that learning enables the evolution of cooperation. We also show that numerically estimated adaptive dynamics appositely predict the outcome of evolutionary simulations. The analysis of the adaptive dynamics enables us to capture the obtained results as an affirmative example of the Baldwin effect, where learning accelerates the evolution to optimality.
Punish the thief-coevolution of defense and cautiousness stabilizes ownership.
Hinsch M, Komdeur J Behav Ecol Sociobiol. 2017; 71(7):102.
PMID: 28706342 PMC: 5486837. DOI: 10.1007/s00265-017-2330-4.
A novel framework of classical and quantum prisoner's dilemma games on coupled networks.
Deng X, Zhang Q, Deng Y, Wang Z Sci Rep. 2016; 6:23024.
PMID: 26975447 PMC: 4792163. DOI: 10.1038/srep23024.
Moderate intra-group bias maximizes cooperation on interdependent populations.
Tang C, Wang Z, Li X PLoS One. 2014; 9(2):e88412.
PMID: 24533084 PMC: 3922813. DOI: 10.1371/journal.pone.0088412.
Shigaki K, Wang Z, Tanimoto J, Fukuda E PLoS One. 2013; 8(11):e76942.
PMID: 24244270 PMC: 3820665. DOI: 10.1371/journal.pone.0076942.