Cooperation has been recognized as an evolutionary puzzle since Darwin, and remains identified as one of the biggest challenges of the XXIst century. Indirect Reciprocity (IR), a key mechanism that humans employ to cooperate with each other, establishes that individual behaviour depends on reputations, which in turn evolve depending on social norms that classify behaviours as good or bad. While it is well known that different social norms give rise to distinct cooperation levels, it remains unclear how the performance of each norm is influenced by the random exploration of new behaviours, often a key component of social dynamics where a plethora of stimuli may compel individuals to deviate from pre-defined behaviours. Here we study, for the first time, the impact of varying degrees of exploration rates – the likelihood of spontaneously adopting another strategy, akin to a mutation probability in evolutionary dynamics – in the emergence of cooperation under IR. We show that high exploration rates may either improve or harm cooperation, depending on the underlying social norm at work. Regarding some of the most popular social norms studied to date, we find that cooperation under Simple-standing and Image-score is enhanced by high exploration rates, whereas the opposite occurs for Stern-judging and Shunning.
The act of cooperation is generally framed quantitatively as an interaction in which an individual provides a benefit b to another at a cost c to himself1,2,3. Traditionally, one assumes that the benefit exceeds the cost (b > c). This means that whenever two individuals are given the option to cooperate or not (that is, to defect) with each other, the joint social optimum is achieved when both cooperate; yet, for each one individually, the fact that cooperation is costly configures defection as the preferred option4. In this sense, cooperation embodies a fascinating social dilemma within societies. When played bilaterally and simultaneously, this donation game turns into a Prisoner’s Dilemma (PD), which provides a convenient abstraction of a wide range of human interactions where the pursuit of self-interest leads to poor collective outcomes. In this context, two fundamental questions have puzzled scientists since Darwin: How is cooperation so widespread in human societies? How can cooperation emerge where it is absent?
The human capacity to establish and use reputation systems suggests some answers. Indeed, humans developed a huge machinery used at profit to share information about others5,6; they evolved to shape decision-making based on the reputation of those they interact with; and they act influenced by what they want others to know about them. All together, reputations work as a mechanism of social control and a lever for cooperation, altruism and collective action7,8,9,10.
Cooperation and reputation have been mathematically linked in models of Indirect Reciprocity (IR)11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42. IR models comprise individuals who adopt heuristics for decision-making based on reputations. In general, the complexity of IR models is limitless. Let us consider the simplest scenario where reputations can either be good (G) or bad (B), and individuals can adopt one of two possible actions: to cooperate (C) or to defect (D). As before, if they cooperate they loose c and the opponent earns b. Otherwise, they pay no cost and confer no benefit. The decision to opt between C and D is not arbitrary. It is encoded in an action rule that prescribes what to do against a G or a B opponent. Naturally, four strategies emerge: always cooperate (AllC); always defect (AllD); discriminate between G and B, only cooperating with G (Disc); or cooperate uniquely with a B (pDisc). Here we assume that the adoption of each strategy follows a process of social learning43 i.e., at each time step, one individual is picked and is given the opportunity to imitate the strategy of a model agent depending on its (better) performance.
Within this environment, reputations are dynamic. After each interaction, the reputation of the individual who decided between C or D (called the donor) when facing another individual (called the recipient) will eventually change. The rule that settles the new reputation to the donor given his actions and the characteristics of himself and the recipient is commonly called a social norm14,20,21. The present work involves the so-called 2nd-order social norms21,23,31. Mathematically, such a norm can be represented as a vector of the type [w, x, y, z], where each position provides the information regarding the new reputation of the donor (G/B), given the reputation of the recipient (G/B) and the action of the donor (C/D): w is the new reputation of an individual that chose C when facing a G opponent; x is the new reputation of an individual that chose D when facing a G opponent; y is the new reputation of whoever chose C when facing a B opponent; and finally z is the new reputation of an individual that opted for D when facing a B opponent. This implies the existence of 16 social norms. Out of these, taking symmetries into account (G and B can be swapped19 as, a priori, they have no meaning and thus constitute pure labels), only 10 social norms are truly distinct. For similar pairs of norms, we discuss the norm that attributes a positive valuation to G, given its relation with C. Popular examples studied in the past are18,21,31: Image-Score (IS) [G, B, G, B]16,17; Simple-Standing (SS) [G, B, G, G]23,35,44; Shunning (SH) [G, B, B, B]28,45; Stern-Judging (SJ) or Kandori [G, B, B, G]30. Interestingly, each social norm defines the dynamics of reputation assessment, which in turn impacts the payoff obtained by each behavioural strategy and consequently their representation in the population. This very simple setting defines a model of IR (detailed in Methods) whose fingerprint is present in a vast set of IR works13,16,19,20,21,22,31. In Fig. 1, in order to provide an intuitive visualisation, we use a ring construction30 to depict some of the most popular social norms studied to date.
So far, IR models neglected the strategic ambiguity that characterizes human interactions by emphasizing strategy adoption through fitness driven mechanisms only (e.g., social learning, cultural imitation, genetic inheritance, etc.) and disregarding the spontaneous adoption of new behaviours. This assumption is questionable, however. The emotional nature of many processes of individual decision-making, together with the creative urge to try new behaviours and the inability to assess accurately the reputation or success of others, all add up to augment the ambiguity associated with the process of strategy adoption. This behavioural ambiguity has been modelled employing a changeable exploration rate, that is, a varying probability that a new strategy is adopted without any sort of individual or social influence46,47,48. This process interestingly resembles (and may be formulated in a mathematically identical way) biological mutations in genetic settings. However, whereas in genetics random mutations are often rare, in social evolution this is not necessarily the case46: On the contrary, high exploration rates may turn out to be the norm rather than the exception, which may strongly affect the evolutionary dynamics of populations facing cooperation47,49 and fairness dilemmas50. In this work we address this issue in the context of IR. We resort to the toolkit of evolutionary game theory (EGT)43 and numerically explore the dynamics of strategy adoption, when social norms govern the co-evolving dynamics of reputation assignment and when individuals may spontaneously adopt (with arbitrary exploration rates) any strategy47. We compute the so-called stationary distribution of strategies in finite populations of size Z, and the population-wide gradients of selection, which allow us to characterize, in detail, the dynamics of strategy adoption. We find that the strategy ambiguity stemming from high exploration rates favours cooperation under the norms Image-Score (IS) and Simple-Standing (SS), whereas it inhibits cooperation under Shunning (SH) and Stern-Judging (SJ).
In Fig. 2 we depict, for a wide interval of exploration rates (from 10−3/Z to 1) and Z = 50, the cooperation levels associated with each social norm. We use specific colours to represent the behaviour associated with the 4 most popular 2nd-order social norms, defined in Fig. 1. As shown, whenever μ > 10−1/Z (indicated by the leftmost vertical dashed line) the cooperation level associated with each social norm is clearly affected by high exploration rates, with the ranking of each norm even changing. While SJ and SS preserve the status of norms that noticeably sustain more cooperation, there are important effects worth pointing for large values of μ: (1) IS and SS benefit, in most cases, from higher exploration rates; (2) the cooperation rate under SH and SJ slightly decreases for high exploration rates and (3) as μ approaches 1, cooperation in all the remaining social norms (drawn with a black colour) generally increases with μ, while cooperation rates under all norms approach 0.5.
Let us start by clarifying point 3): Increasing the exploration rate implies that social learning plays a decreasing role in the overall evolution of decision making. In particular, for very large mutation rates, since we have a significant fraction of the population (μ) just exploring the strategy space, the system typically does not access the grey shaded areas pictured in Fig. 2 (see Methods for details). This way, those social norms that are unable to promote cooperation under social learning benefit from the co-existence of all strategies. Naturally, the converse happens for those norms that already amplify the levels of cooperation under a regime of social learning, as is the case of SJ and SS. In the following discussion, we focus on points (1) and (2), i.e., on the interval of μ in which fitness dependent social learning — and thereby reputations and social norms —play the steering role in strategy adoption.
To clarify the points 1) and 2) and to further understand the effects of high exploration rates on norms SH, SJ, IS, and SS, we show in Figs 3 and 4 i) the gradients of selection (Γx), together with ii) the reputation distribution per state (γx) and iii) the stationary distribution (λx) of strategies, for the 4 most popular social norms (see section Methods for a detailed account of these metrics). The tetrahedrons represent the entire state space (simplex) defined by this 4-strategy dynamics, where each corner defines a monomorphic state in which the entire population adopts the same strategy. For simplicity and visualization purposes we provide details of the evolutionary dynamics inside the triangular slices of the complete 3D dynamics. Inside these triangular slices, each arrow corresponds to Γx and represents the most probable direction of evolution. The prevalence in each state (λx) is associated with the background colour intensity whereas blue/red tones translate into an increased number of G/B individuals (γx).
Figure 3 shows that the harshness of SH becomes more evident for high μ (Fig. 3-B), and cooperation is thereby precluded. High μ means that B labels are even easier to be attributed due to the presence of individuals that spontaneously adopt AllD and pDisc, strategies that in general are labelled B under SH. The increase in B individuals makes Disc and AllD almost indistinguishable and turns the coordination between these strategies (Fig. 3-A) into a co-existence dynamics (Fig. 3-B). In the case of SJ we observe that, for low μ, most of the time is spent in the highly cooperative monomorphic states Disc and pDisc (Fig. 3-C,D). When μ increases (Fig. 3-E) cooperation declines as the prevalence under monomorphic states Disc/pDisc decreases, with individuals exploring the other strategies (AllD and AllC).
Figure 4 reveals why cooperation increases at high μ, when the prevailing norm is either IS or SS. For IS, high exploration rates have the ability to move the system away from states where AllD is the prevalent strategy (Fig. 4-B). Indeed, high values of μ increase the number of individuals that spontaneously adopt Disc or AllC, placing and stabilizing the dynamics in the interior of the simplex. In the case of SS, high values of μ allow the population to overcome the “coordination barrier” between AllD and Disc (Fig. 4-E), thus making it easier to achieve the minimum number of Discs that renders advantageous to have a G reputation; this way, the population spends less time in AllD and more time near the edge where AllC and Disc co-exist, which naturally benefits cooperation.
Besides the particular effect of μ in each social norm, it is noteworthy that cooperation levels remain qualitatively unchanged (apart from the slow monotonic increase of cooperation under SS) for a wide range of values of the exploration rate μ (Fig. 2). These results confirm, for the first time, that for a wide interval of values of μ the Small-Mutation Approximation31,51 (SMA) proves accurate in the context of IR. With SMA, one assumes that μ is small enough so that the system will spent a negligible fraction of time in polymorphic states. This way, SMA allows a convenient characterization of evolutionary processes, formally in the limit when μ ≪ 1/Z, through a reduced (embedded) Markov chain involving monomorphic states only. Yet, its range of applicability is often unclear48, requiring the use of large-scale computer simulations as the ones performed here.
In this work we investigate the effect of an arbitrary exploration rate on the evolutionary dynamics of cooperation under indirect reciprocity. This is an analysis of general interest given the ubiquity of creative and emotional traits that characterize human behaviour, urging them to explore new strategies47. The reputation systems studied here were based on the action of the donor and the reputation of the recipient, a feature associated with so-called 2nd-order norms21,31. We show that random exploration of the strategy space has a non-trivial effect in the behaviour dynamics of populations under IR: when high, it increases cooperation under IS (interestingly, a norm just demanding 1st order information) and SS, while it decreases cooperation under SH and SJ. Despite these key observations, our results indicate that the so-called “leading two” norms21 — SJ and SS — remain the most effective in promoting cooperation, a result that may not hold in the space of high-order norms. Overall, our results suggest that a general heuristics can, in fact, be intuitively derived: when the social norm is able to provide high cooperation rates by relying on the stability of some cooperative monomorphic states (e.g., Disc in SJ and SH), exploration is pernicious; when the social norm is unable to provide that stability, allowing instead for the occurrence of some uncooperative monomorphic state (e.g., AllD in IS and in SS) high exploration rates favour overall cooperation by allowing the population to move away from this undesirable state.
Despite these results, we confirm that SJ remains the social norm that, overall, promotes the highest levels of cooperation. The strength of this norm relies on the efficiency to promote a strong stability of two highly cooperative monomorphic states. No other polymorphic state (i.e., in which more than one strategy co-exists) is able to reach those levels of cooperation; hence, any ambiguity source that moves the population to the interior of the simplex will be detrimental to cooperation. This, however, is not the whole story, since the reference levels of cooperation attained under SJ are so high that the disadvantageous effect induced by large exploration rates does not prevent SJ to remain the leading norm in what concerns the promotion of cooperation.
The dynamics under SJ also puts in evidence a remarkable fact. Because of the inherent symmetry of SJ with respect to the G and B labels, the net effect of SJ resembles some sort of a divide-and-conquer procedure. Indeed, under SJ the whole state space is divided into two smaller basins of attraction that push the population to full cooperation in both cases: if there is a majority of individuals that cooperate with B and defect with G (pDisc), the population will move into a state where everyone is B; if there is a majority of Disc, the population will move to a state where everyone is G; in both limits, cooperation prevails, and that ensures the collective optimum (Fig. 3-D). Ultimately, the notions of Good and Bad, as we normally use them, are defined by the actions and not by the labels attributed to specific reputations. Clearly, in the case of SJ, the meaning of the signals G(ood) or B(ad) may emerge from a simple convention52,53. This means, in turn, that if we would consider different populations evolving independently and under the assessment of this norm, full cooperation might still be achieved under different emergent conventions for what is Good or Bad. Given the conceptual simplicity of the underlying IR model employed, this constitutes a remarkable feature of SJ, related also to complex topics such as conflicting moral systems30,54 or the appearance of in/out groups together with their inherent normative values. These observations can only be made, however, to the extent that (like we have done in this work), one provides the global dynamics considering the 4 possible action rules (AllD, Disc, pDisc, AllC), instead of carrying out the analysis including only 3, as often happens. This feature, together with the systematic investigation of high exploration rates in IR, were carried out here for the first time.
It is also noteworthy the way SJ implies a moral judgement that, besides justifying the defection against undesirable opponents, also condemns cooperating with those. Indeed, under SJ, whoever cooperates with an opponent carrying a Bad reputation gets himself a Bad reputation. Interestingly, this judgement is somehow verified in the behaviour of toddlers, who prefer those that mistreat (rather than help) opponents that misbehaved in the past55,56,57,58.
Finally, given the current importance of reputation-based systems, indirect reciprocity models emerge nowadays as relevant toolkits for artificial intelligence applications, web platforms and systems supporting sharing economies. Online communities are nowadays pervasive and most of them profit from the readiness of its users to cooperate, which is often supported by reputation59. Overall, reputation mechanisms are considered a key element in the design of multiagent systems60,61,62. The simulation technique that we present here, together with the proposed metrics to visualize the emergent dynamics, can provide the appropriate basis in which to study (and evaluate ref. 63) other challenges related with strategic dynamics and reputation systems, such as the effect of considering different reputation management schemes64,65,66,67, the design of new underlying structures of interaction36,68,69,70,71,72,73,74 or the formalization of bottom-up artificial morality and machine ethics75.
Strategies and reputations
We model a population of Z individuals that interact with each other and may change their strategy over time (in the way described below). Individuals play with each other a donation game that reproduces, in a simple way, the mathematics of cooperation43. Each individual may choose one of two actions: to cooperate (C), meaning that the donor incurs a cost c to confer a benefit b to the opponent; to defect (D), whereby neither player earns or looses anything. Each individual follows a strategy (action rule) that defines how to act against a Good (G) or Bad (B) opponent. This rule is encoded in a vector [x, y] where the first position dictates what to do (C/D) against a G and the second position what to do against a B. There are four possible strategies: AllC [C, C], Disc [C, D], pDisc [D, C] and AllD [D, D]. Initially, each individual has one reputation (G or B) and strategy; these are attributed randomly with uniform probability.
Update of strategies and reputations
At each (discrete) time step (t) one individual X is randomly selected to update its strategy. With probability μ (so called exploration rate43,46,47,48) X adopts a random strategy within the full space of possible strategies. With probability 1-μ strategy change may take place through social learning; in this case, X compares its fitness (fX) with that (fY) of another individual Y randomly selected, changing its strategy to that of Y with a probability that increases with the fitness difference, given by , with β = 1, ensuring a significant selection strength. This imitation process and the associated probability function are well documented47,76 and known as pairwise comparison rule. Both fitnesses fX and fY are associated with the average payoff obtained in 2Z donation games, always played against random opponents. In each of these games, both individuals play once as donor and once as recipient. After each donation game, with a probability τ, a new reputation is attributed to the individual acting as donor, in accordance with the social norm fixed in the population. With probability 1 − τ, the donor keeps the same reputation. For simplicity, we assume a public reputations scheme16,19,20,31 such that, through gossip or rumours, new reputations spread to everyone, a simplification that can naturally be relaxed in future works based on different reputation database schemes64. As already described in the Introduction and depicted in Fig. 1, here we consider that a social norm is a vector of the type [w, x, y, z], where each position provides the information regarding the new reputation of the donor (G/B), given the reputation of the recipient (G/B) and the action of the donor (C/D).
We allow for the inclusion of errors of three different types: execution errors – with probability ε, there is a failure to cooperate when the action rule dictates so34; assignment errors – with probability α, the assigned reputation is the opposite of the one prescribed by the social norm; and private assessment errors – with probability χ, when deciding about what action to employ or when deciding the next reputation of an opponent, the retrieved reputation of an individual is the opposite of the one actually owned21. The effect of each kind of error in the cooperation levels supported by each specific social norm was already discussed31. While Figs 1, 2 and 3 report the results for ε = α = χ = 0.01, we confirm that the effect of exploration rate discussed in the main text remains unchanged if (all else constant) ε, α or χ go up to 0.1 and τ down to 0.01.
Tracing the dynamics
Each possible state of this population can be enumerated and arranged spatially, so that its dynamics becomes straightforward to visualize. As there are Z individuals and 4 different strategies, each state of the population is identified by a tuple x = (k0, k1, k2) meaning that, in state x, there are k0 individuals adopting strategy AllD, k1 adopting strategy pDisc, k2 adopting strategy Disc and k3 = Z-k0-k1-k2 adopting strategy AllC. In total, there are (Z + 1) (Z + 2) (Z + 3)/6 states. These states can intuitively be assembled in a 3D simplex (a tetrahedron as in Figs 3 and 4), where the vertices represent states in which all the agents are adopting the same one unique strategy (monomorphic or pure states). In Figs 3 and 4, to allow for an easier interpretation of the results, we present the dynamics along cross sections of this space (2D simplexes, triangles of variable size depending on the specific cross section position). Provided this description, we retrieve information in four different forms discussed below: average cooperation rates, average time in each state, average fraction of good and bad reputations, and gradient of selection.
Average cooperation rate
We use the general result that one generation corresponds to Z discrete time steps where a strategy update may occur, and we simulate the system for G = 107 generations. We repeat each simulation R = 100 times (runs), such that in each run the pseudo-random number sequence will be different. The average cooperation rate (ηi) in run i is computed by dividing the total number of cooperative acts (Ci) by the total number of donation games (Ki):
As μ → 1, exploration increasingly dominates selection (in the form of social learning), implying that reputations and social norms are subjected to minor perturbations with respect to a purely random choice process. Particularly, since there is always the probability μ that an individual explores any other strategy, the space of accessible configurations is effectively reduced with the increase of μ47, such that . The minimum fraction of strategy i is given by one of the roots of , where and are the probabilities of increasing and decreasing (respectively) the number of individuals of strategy i, for a general state x. In the most unfavourable scenario for strategy i (i.e., strong selection (), and ), can be calculated by solving , which results in . This quantity is close to μ/d, as employed in ref. 47. In general η = k3/Z + γk2/Z + (1 − γ)k1/Z - with γ standing for the fraction of G individuals (see below) — which is minimized for k1/Z = k2/Z = k3/Z = x, such that, for d = 4, ηmin = x + γx + (1 − γ)x = 2x. A similar argument can be used to compute the minimum defection rate, leading to a ηmax = 1 − ηmin. The inaccessible cooperation rates are represented in Fig. 1 by a grey background.
Average time in each state
We keep a counter that totalizes the number of time steps a given state of the population is reached. After the proper normalization (the total number of steps is GRZ), we collect information on the average fraction of time (λx) that the population spends in each state x = (k0, k1, k2). This information is conveyed in the triangles of Figs 3 and 4 by means of colour intensity. In the tetrahedrons, we place a sphere in the states where the population spends more time. The area covered by the spheres accounts for 80% of the total simulation time, i.e., assuming that state x0 is the most visited state, , we place spheres up to state .
Average fraction of Good and Bad reputations
By directly placing the system in each possible state x = (k0, k1, k2) and simulating one time step per run (total number of runs R, each starting in the same state x), we can draw a picture of the average number of G/B. After each time step we save the resulting fraction of G individuals, Gt,x/(Gt,x + Bt,x). After all runs, we have the average fraction of G individuals accruing from a specific state x,
Gradient of selection
Following a similar procedure as the one described in the previous section, we can compute the transition probability between each pair of neighbour states x and y. For each state x = (k0, k1, k2), we count the number of times that, after each time step, the system moves into each of the (maximum) 12 adjacent states y = (l0, l1, l2). For that, we keep the quantities βx,i and δx,i, i.e., the number of times that a strategy i is born (li = ki + 1) or dies (li = ki − 1) in state x. We keep the number of runs that we start a time step in each state in x, Rx. This provides the required information to represent the so-called gradient of selection Γx, i.e., a vector field that provides an approximation for the most probable direction that the system will follow once located in state x, given the update of strategies (imitation and exploration) by each agent (described above),
How to cite this article: Santos, F. P. et al. Evolution of cooperation under indirect reciprocity and arbitrary exploration rates. Sci. Rep. 6, 37517; doi: 10.1038/srep37517 (2016).
Publisher’s note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Axelrod, R. & Hamilton, W. D. The evolution of cooperation. Science 211, 1390–1396 (1981).
Rand, D. G. & Nowak, M. A. Human cooperation. Trends. Cogn. Sci. 17, 413–425 (2013).
Nowak, M. A. Five rules for the evolution of cooperation. Science 314, 1560–1563 (2006).
Dawes, R. M. Social dilemmas. Annu. Rev. Psychol. 31, 169–193 (1980).
Dunbar, R. Grooming, Gossip, and the Evolution of Language (Harvard University Press, 1998).
Sommerfeld, R. D., Krambeck, H.-J., Semmann, D. & Milinski, M. Gossip as an alternative for direct observation in games of indirect reciprocity. Proc. Natl. Acad. Sci. USA 104, 17435–17440 (2007).
Milinski, M., Semmann, D. & Krambeck, H.-J. Reputation helps solve the ‘tragedy of the commons’. Nature 415, 424–426 (2002).
Wedekind, C. & Milinski, M. Cooperation through image scoring in humans. Science 288, 850–852 (2000).
Bolton, G. E., Katok, E. & Ockenfels, A. Cooperation among strangers with limited information about reputation. J. Public Econ. 89, 1457–1468 (2005).
Seinen, I. & Schram, A. Social status and group norms: Indirect reciprocity in a repeated helping experiment. Eur. Econ. Rev. 50, 581–602 (2006).
Boyd, R. & Richerson, P. J. The evolution of indirect reciprocity. Soc. Networks 11, 213–236 (1989).
Brandt, H. & Sigmund, K. Indirect reciprocity, image scoring, and moral hazard. Proc. Natl. Acad. Sci. USA 102, 2666–2670 (2005).
Brandt, H. & Sigmund, K. The good, the bad and the discriminator—errors in direct and indirect reciprocity. J. Theor. Biol. 239, 183–194 (2006).
Chalub, F. A. C. C., Santos, F. C. & Pacheco, J. M. The evolution of norms. J. Theor. Biol. 241, 233–240 (2006).
Ghang, W. & Nowak, M. A. Indirect reciprocity with optional interactions. J. Theor. Biol. 365, 1–11 (2015).
Nowak, M. A. & Sigmund, K. Evolution of indirect reciprocity by image scoring. Nature 393, 573–577 (1998).
Nowak, M. A. & Sigmund, K. The dynamics of indirect reciprocity. J. Theor. Biol. 194, 561–574 (1998).
Nowak, M. A. & Sigmund, K. Evolution of indirect reciprocity. Nature 437, 1291–1298 (2005).
Ohtsuki, H. & Iwasa, Y. How should we define goodness? —reputation dynamics in indirect reciprocity. J. Theor. Biol. 231, 107–120 (2004).
Ohtsuki, H. & Iwasa, Y. The leading eight: social norms that can maintain cooperation by indirect reciprocity. J. Theor. Biol. 239, 435–444 (2006).
Ohtsuki, H. & Iwasa, Y. Global analyses of evolutionary dynamics and exhaustive search for social norms that maintain cooperation by reputation. J. Theor. Biol. 244, 518–531 (2007).
Ohtsuki, H., Iwasa, Y. & Nowak, M. A. Reputation Effects in Public and Private Interactions. PLoS Comput. Biol. 11, e1004527 (2015).
Panchanathan, K. & Boyd, R. A tale of two defectors: the importance of standing for evolution of indirect reciprocity. J. Theor. Biol. 224, 115–126 (2003).
Panchanathan, K. & Boyd, R. Indirect reciprocity can stabilize cooperation without the second-order free rider problem. Nature 432, 499–502 (2004).
Roberts, G. Evolution of direct and indirect reciprocity. Proc. R. Soc. London B 275, 173–179 (2008).
Sigmund, K. Moral assessment in indirect reciprocity. J. Theor. Biol. 299, 25–30 (2012).
Suzuki, S. & Akiyama, E. Evolution of indirect reciprocity in groups of various sizes and comparison with direct reciprocity. J. Theor. Biol. 245, 539–552 (2007).
Takahashi, N. & Mashima, R. The emergence of indirect reciprocity: Is the standing strategy the answer. Center for the study of cultural and ecological foundations of the mind, Hokkaido University, Japan, Working paper series 29 (2003).
Olejarz, J., Ghang, W. & Nowak, M. A. Indirect Reciprocity with Optional Interactions and Private Information. Games 6, 438–457 (2015).
Pacheco, J. M., Santos, F. C. & Chalub, F. A. C. Stern-judging: A simple, successful norm which promotes cooperation under indirect reciprocity. PLoS Comput. Biol. 2, e178 (2006).
Santos, F. P., Santos, F. C. & Pacheco, J. M. Social Norms of Cooperation in Small-Scale Societies. PLoS Comput. Biol. 12, e1004709 (2016).
Ohtsuki, H., Iwasa, Y. & Nowak, M. A. Indirect reciprocity provides only a narrow margin of efficiency for costly punishment. Nature 457, 79–82 (2009).
Nakamura, M. & Ohtsuki, H. Indirect reciprocity in three types of social dilemmas. J. Theor. Biol. 355, 117–127 (2014).
Fishman, M. A. Indirect reciprocity among imperfect individuals. J. Theor. Biol. 225, 285–292 (2003).
Leimar, O. & Hammerstein, P. Evolution of cooperation through indirect reciprocity. Proc. R. Soc. London B 268, 745–753 (2001).
Peleteiro, A., Burguillo, J. C. & Chong, S. Y. In Proceedings of the 2014 International Conference on Autonomous Agents and Multi-Agent Systems 669–676 (International Foundation for Autonomous Agents and Multiagent Systems, 2014).
Uchida, S. & Sigmund, K. The competition of assessment rules for indirect reciprocity. J Theor Biol 263, 13–19 (2010).
Nakamura, M. & Masuda, N. Indirect reciprocity under incomplete observation. PLOS Comput Biol 7, e1002113 (2011).
Fu, F., Hauert, C., Nowak, M. A. & Wang, L. Reputation-based partner choice promotes cooperation in social networks. Phys. Rev. E 78, 026117 (2008).
Tanaka, H., Ohtsuki, H. & Ohtsubo, Y. The price of being seen to be just: an intention signalling strategy for indirect reciprocity. Proc. R. Soc. London B 283, 20160694 (2016).
Nax, H. H., Perc, M., Szolnoki, A. & Helbing, D. Stability of cooperation under image scoring in group interactions. Sci. Rep. 5 (2015).
Sasaki, T., Okada, I. & Nakai, Y. Indirect reciprocity can overcome free-rider problems on costly moral assessment. Biol. Lett. 12 (2016).
Sigmund, K. The Calculus of Selfishness (Princeton University Press, 2010).
Sugden, R. The Economics of Rights, Co-operation and Welfare (Basil Blackwell, Oxford, 1986).
Takahashi, N. & Mashima, R. The importance of subjectivity in perceptual errors on the emergence of indirect reciprocity. J. Theor. Biol. 243, 418–436 (2006).
Traulsen, A., Semmann, D., Sommerfeld, R. D., Krambeck, H.-J. & Milinski, M. Human strategy updating in evolutionary games. Proc. Natl. Acad. Sci. USA 107, 2962–2966, doi: 10.1073/pnas.0912515107 (2010).
Traulsen, A., Hauert, C., De Silva, H., Nowak, M. A. & Sigmund, K. Exploration dynamics in evolutionary games. Proc. Natl. Acad. Sci. USA 106, 709–712 (2009).
Wu, B., Gokhale, C. S., Wang, L. & Traulsen, A. How small are small mutation rates? J. Math. Biol. 64, 803–827 (2012).
Hauser, O. P., Nowak, M. A. & Rand, D. G. Punishment does not promote cooperation under exploration dynamics when anti-social punishment is possible. J. Theor. Biol. 360, 163–171 (2014).
Santos, F. P., Santos, F. C., Paiva, A. & Pacheco, J. M. Evolutionary dynamics of group fairness. J. Theor. Biol. 378, 96–102 (2015).
Fudenberg, D. & Imhof, L. Imitation Processes with Small Mutations. J. Econ. Theory 131, 251–262 (2005).
Lewis, D. Convention: A Philosophical Study (John Wiley & Sons, 2008).
Skyrms, B. Evolution of the Social Contract (Cambridge University Press, 2014).
Greene, J. Moral Tribes: Emotion, Reason and the Gap Between us and Them. (Atlantic Books Ltd., 2014).
Tasimi, A. & Wynn, K. Costly rejection of wrongdoers by infants and children. Cognition 151, 76–79 (2016).
Hamlin, J. K. & Wynn, K. Young infants prefer prosocial to antisocial others. Cognitive Dev. 26, 30–39 (2011).
Hamlin, J. K., Wynn, K. & Bloom, P. Social evaluation by preverbal infants. Nature 450, 557–559 (2007).
Hamlin, J. K., Wynn, K., Bloom, P. & Mahajan, N. How infants and toddlers react to antisocial others. Proc. Natl. Acad. Sci. USA 108, 19931–19936 (2011).
Jøsang, A., Ismail, R. & Boyd, C. A survey of trust and reputation systems for online service provision. Decis. Support. Syst. 43, 618–644 (2007).
Pinyol, I. & Sabater-Mir, J. Computational trust and reputation models for open multi-agent systems: a review. Artif. Intell. Rev. 40, 1–25 (2013).
Conte, R. & Paolucci, M. Reputation in Artificial Societies: Social Beliefs for Social Order Vol. 6 (Springer Science & Business Media, 2002).
Castelfranchi, C. & Falcone, R. In Proceedings of the 1998 International Conference on Multi Agent Systems 72–79 (IEEE, 1998).
Hazard, C. J. & Singh, M. P. In Proceedings of the International Joint Conference in Artificial Intelligence, IJCAI 2013 191–197 (AAAI Press, 2013).
Bachrach, Y., Parnes, A., Procaccia, A. D. & Rosenschein, J. S. Gossip-based aggregation of trust in decentralized reputation systems. Auton. Agent Multi Agent Syst. 19, 153–172 (2009).
Uchida, S. Effect of private information on indirect reciprocity. Phys Rev E 82, 036111 (2010).
Tanabe, S., Suzuki, H. & Masuda, N. Indirect reciprocity with trinary reputations. J Theor Biol 317, 338–347 (2013).
Sasaki, T., Okada, I. & Nakai, Y. The evolution of conditional moral assessment in indirect reciprocity. arXiv preprint arXiv:1605.02166 (2016).
Sabater, J. & Sierra, C. Review on computational trust and reputation models. Artif. Intell. Rev. 24, 33–60 (2005).
Pinheiro, F. L., Santos, M. D., Santos, F. C. & Pacheco, J. M. Origin of peer influence in social networks. Phys. Rev. Lett. 112, 098702 (2014).
Santos, F. C. & Pacheco, J. M. Scale-free networks provide a unifying framework for the emergence of cooperation. Phys. Rev. Lett. 95, 098104 (2005).
Santos, F. C., Santos, M. D. & Pacheco, J. M. Social diversity promotes the emergence of cooperation in public goods games. Nature 454, 213–216 (2008).
Szabó, G. & Fáth, G. Evolutionary games on graphs. Phys Rep 446, 97–216 (2007).
Perc, M. & Szolnoki, A. Coevolutionary games–a mini review. BioSystems 99, 109–125 (2010).
Wang, Z., Wang, L., Szolnoki, A. & Perc, M. Evolutionary games on multilayer networks: a colloquium. Eur. Phys. J. B 88, 1–15 (2015).
Pereira, L. M. & Saptawijaya, A. Bridging two realms of machine ethics. Rethinking Machine Ethics in the Age of Ubiquitous Technology 197–224 (2015).
Szabó, G. & Tőke, C. Evolutionary prisoner’s dilemma game on a square lattice. Phys. Rev. E 58, 69 (1998).
The authors thank Vítor V. Vasconcelos for fruitful discussions. This research was supported by Fundação para a Ciência e Tecnologia (FCT) through grants SFRH/BD/94736/2013, PTDC/EEI-SII/5081/2014, PTDC/MAT/STA/3358/2014 and by multiannual funding of CBMA and INESC-ID (under the projects UID/BIA/04050/2013 and UID/CEC/50021/2013) provided by FCT.
The authors declare no competing financial interests.
About this article
Cite this article
Santos, F., Pacheco, J. & Santos, F. Evolution of cooperation under indirect reciprocity and arbitrary exploration rates. Sci Rep 6, 37517 (2016). https://doi.org/10.1038/srep37517
Scientific Reports (2021)
Autonomous Agents and Multi-Agent Systems (2020)
Scientific Reports (2018)
Scientific Reports (2018)