Abstract

One of the prototypical mechanisms in understanding the ubiquitous cooperation in socialdilemma situations is the win–stay, lose–shift rule. In this work, a generalized win–stay,lose–shift learning model—a reinforcement learning model with dynamic aspirationlevel—is proposed to describe how humans adapt their social behaviors based on theirsocial experiences. In the model, the players incorporate the information of the outcomes inprevious rounds with time-dependent aspiration payoffs to regulate the probability ofchoosing cooperation. By investigating such a reinforcement learning rule in the spatialprisoner’s dilemma game and public goods game, a most noteworthy viewpoint is thatmoderate greediness (i.e. moderate aspiration level) favors best the development andorganization of collective cooperation. The generality of this observation is tested againstdifferent regulation strengths and different types of network of interaction as well. We alsomake comparisons with two recently proposed models to highlight the importance of themechanism of adaptive aspiration level in supporting cooperation in structuredpopulations.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.