Reinforcement learning accounts for moody conditional cooperation behavior: experimental results

Yutaka Horita, Masanori Takezawa, Keigo Inukai, Toshimasa Kita, Naoki Masuda*


研究成果: Article査読

25 被引用数 (Scopus)


In social dilemma games, human participants often show conditional cooperation (CC) behavior or its variant called moody conditional cooperation (MCC), with which they basically tend to cooperate when many other peers have previously cooperated. Recent computational studies showed that CC and MCC behavioral patterns could be explained by reinforcement learning. In the present study, we use a repeated multiplayer prisoner's dilemma game and the repeated public goods game played by human participants to examine whether MCC is observed across different types of game and the possibility that reinforcement learning explains observed behavior. We observed MCC behavior in both games, but the MCC that we observed was different from that observed in the past experiments. In the present study, whether or not a focal participant cooperated previously affected the overall level of cooperation, instead of changing the tendency of cooperation in response to cooperation of other participants in the previous time step. We found that, across different conditions, reinforcement learning models were approximately as accurate as a MCC model in describing the experimental results. Consistent with the previous computational studies, the present results suggest that reinforcement learning may be a major proximate mechanism governing MCC behavior.

ジャーナルScientific reports
出版ステータスPublished - 2017 1月 10

ASJC Scopus subject areas

  • 一般


「Reinforcement learning accounts for moody conditional cooperation behavior: experimental results」の研究トピックを掘り下げます。これらがまとまってユニークなフィンガープリントを構成します。