Reinforcement learning accounts for moody conditional cooperation behavior: experimental results

Yutaka Horita, Masanori Takezawa, Keigo Inukai, Toshimasa Kita, Naoki Masuda

Research output: Contribution to journalArticle (Academic Journal)peer-review

27 Citations (Scopus)
306 Downloads (Pure)


In social dilemma games, human participants often show conditional cooperation (CC) behavior or its variant called moody conditional cooperation (MCC), with which they basically tend to cooperate when many other peers have previously cooperated. Recent computational studies showed that CC and MCC behavioral patterns could be explained by reinforcement learning. In the present study, we use a repeated multiplayer prisoner’s dilemma game and the repeated public goods game played by human participants to examine whether MCC is observed across different types of game and the possibility that reinforcement learning explains observed behavior. We observed MCC behavior in both games, but the MCC that we observed was different from that observed in the past experiments. In the present study, whether or not a focal participant cooperated previously affected the overall level of cooperation, instead of changing the tendency of cooperation in response to cooperation of other participants in the previous time step. We found that, across different conditions, reinforcement learning models were approximately as accurate as a MCC model in describing the experimental results. Consistent with the previous computational studies, the present results suggest that reinforcement learning may be a major proximate mechanism governing MCC behavior.
Original languageEnglish
Article number39275
Number of pages10
JournalScientific Reports
Early online date10 Jan 2017
Publication statusPublished - 10 Jan 2017


  • human behaviour
  • Social evolution


Dive into the research topics of 'Reinforcement learning accounts for moody conditional cooperation behavior: experimental results'. Together they form a unique fingerprint.

Cite this