Introduction

Honesty is the best policy; but he who is governed by that maxim is not an honest man.

- Archbishop Richard Whately

In 2007, Dana et al. conducted a novel version of the two-player Dictator Game, where one player is given an endowment, and decides whether to give the second player any of it. Previous studies found that, against the purely economic expectation that self-interested first players should keep the whole endowment, players were likely to give a substantial portion of the endowment away, suggesting a widespread human preference for fairness and prosociality. Dana et al. (2007) found, however, that when dictator-players were given the option to pay 10% of the endowment to quietly exit the game without the second player knowing about it, just under half of players did so, kee** 90% of the endowment for themselves. Paying for the opportunity to keep more of the endowment quietly, they found, motivated a greater number of players than expected.

Several hypotheses, which have received both empirical and computational attention over the last several decades, aim to explain widespread cooperation in humans at the proximate and ultimate levels (Hamilton 1964; Trivers 1971; Alexander 1987; Frank 1988; Axelrod and Hamilton 1981; Frank 1998; Nowak and Sigmund 1998; Boyd et al. 2003; Boyd and Richerson 1992; Fehr and Fischbacher 2003; Axelrod et al. 2004; Roberts 2005; Henrich and Henrich 2007; Boehm 2017; Wrangham 2019; also see West et al. 2007); see Table 1. Prominent among these, at least at the proximate level, is strong reciprocity (Gintis 2000; Fehr et al. 2002; Bowles and Gintis 2004), which relies on partner choice and partner control (for a discussion of this distinction, see Baumard et al. 2013) mechanisms. Given that there is almost always some economic motivation to cheat in social relationships, individuals, according to these theorists, must both choose potential partners and control existing partners with the threat of punishment (Barclay 2016; Raihani 2021).

Table 1 Overview of well-known proximate and ultimate hypotheses that aim to explain widespread human cooperation

The force of both strategies has been shown in numerous conceptual and empirical studies (for example, Gintis 2000; Fehr and Gachter 2002; Panchanathan and Boyd 2004; Wiessner 2005, although see Singh and Garfield 2022), which also tie into foundational concepts in cooperation studies, such as reputation, emotion, intelligence, and honest signalling (McElreath et al. 2003a, b; McNally et al. 2012; Számadó et al. 2021; Giardini et al. 2022). Yet a common feature of many computational experiments is that cooperation is equated with number of cooperators: in a given population, insofar as individuals cooperate in given dyadic or group-level interactions, they are considered cooperators. Research, in these cases, implicitly assumes that as the fraction of cooperative actions increases, so does the fraction of cooperators (some have also argued that “types” are stable across time, see Kurzban and Houser 2005).

The logic of cultural group selection (Boyd and Richerson 1988; Henrich 2004, 2020) goes further. Given that, as conceptual research continues to robustly show, partner control and choice strategies are effective at maintaining cooperation in human populations, it is probable that successful groups, over our evolutionary history, developed and transmitted cultural traits that promote conditional cooperation and punishment (Gintis et al. 2005; Henrich and Muthukrishna 2021). Cultural groups that effectively reduced within-group behavioural heterogeneity with relation to cooperation, according to this body of work, maximized cooperation and out-competed those with a higher proportion of free-riders in the population—those groups without effective partner choice and control mechanisms. Where cooperation flourishes, so does the cultural group, according to cultural group selection theorists.

Some authors note, however, that there remains an issue of framing: conditional cooperators are also conditional defectors (Hagen and Hammerstein 2006; Bernhard and Cushman 2022; Ibrahim 2022). In Prisoner’s Dilemmas (PDs), for example, a recently noted set of tactics, known as zero-determinant strategies (Press and Dyson 2012; Hilbe et al. 2013; Stewart and Plotkin 2013), may cooperate with high frequency—but yet always ensure that the agent’s payoff is at least as great as the relevant partner’s. In real-world interactions, it is vague whether equivalent interactions should be labelled as prosocial, where “extortion” might be an equally apt label (Bernhard and Cushman 2022).

The ambiguity of social descriptions in agent interactions, therefore, leaves open the question of whether a prosocial strategy reflects prosocial preferences, though widespread empirical evidence suggests this assumption is nonetheless supported (Fehr and Gächter 2002; Rand et al. 2012; Henrich 2020; although see Mulder et al. 2006). Explaining the greater-than-expected degree of prosociality in public goods games and PDs across cultures has, instead, been the focus of an ever-increasing number of models in the evolutionary sciences (Henrich et al. 2005).

This focus, and related attachment to real-world findings in economic experiments, suggests that the ascriptions given to social interactions are no less important to the evolution of cooperation than are the strategies that yield the optimal payoff outcomes (Delton 2022). Coupled with the idea that individuals have more than economic reasons to cooperate, the notion that humans are instilled, genetically or culturally speaking, with prosocial preferences requires further elucidation. To determine whether the free-rider problem has been solved by cultural transmission in human groups, in other words, it is critical to understand what a free rider is—other than, in the simplest terms, an individual who defects in economic games.

The aim of this paper is to explore the notion of prosocial preferences, and their attendant consequences for cultural group selection, through an agent-based modelling paradigm that separates individual behaviours, types, and appearances.Footnote 1 My goal is to ask whether framing interactions with more specific social descriptions than those seen in many simple PD models, which explore the success of strategies only (such as in the classic work by Axelrod and Hamilton 1981), affects the interpretation of whether a high degree of cooperation is equivalent to a high degree of prosociality—that is, whether individuals who cooperate have necessarily internalized the social norms in a given culture (Sperber and Baumard 2012). High rates of cooperation, both in economic games and in wider society, may have more to do with the lack of opportunity to defect than with universal human preferences—a problem exemplified by Dana et al.’s (2007) findings.

I develop three agent-based models with random dyadic pairing that explore the success of strong reciprocity in PD interactions (see Gintis 2000). The models account for behaviour, individual type, and individual appearance as variables, which aims to represent not more complex interactions, but more complex descriptions of the interactions being modelled. I aim to capture this complexity using mimicry (see Nettle and Dunbar 1997) as a tactic for evading partner control strategies, although I do not take partner choice or reputation into account. The models progress in complexity, both in terms of strategies available and the linguistic descriptions attributable to the dyadic interactions. For example, while the strong reciprocity model can give only the simple formulation, player x punishes player y or player x cooperates with player y, the mimicry model qualifies some instances of cooperation as mimic player x cooperates with cooperative player y. While researchers designing these models sometimes make claims about the agents’ motivations or preferences, alternate interpretations of these interactions are possible.

I discuss the distinction between cooperation and cooperators, that is cultural processes and cultural agents, in relation to previous findings in theoretical models, as well as ethnographic evidence showing higher-than-expected rates of prosociality across cultures (Henrich et al. 2005). My aim is to show that strategies that prevent defection are not unbreakable barriers to free riding, and further to suggest that, as societies grew larger, humans used more complex strategies—founded in increasingly complex behaviours built on the use of language—to subvert cooperative norms for personal benefit. Acting honestly—or in the case of economic games, prosocially—should not, I argue, be equated with holding prosocial preferences.

Methods

General model description

(For a full model description according to the ODD standard protocol for agent-based models [Grimm et al. 2006], see the Supplement §1. Visit github.com/jonathanrgoodman/Opportunity for all source code and simulated data and the Supplement).

In the present set of models, I explored the implications of mimicry on random-pairing–based PDs. The models were designed with stepwise complexity, starting with standard strong reciprocity and graduating to mimicry-mediated covert strategies (for models of covert signalling, see Robson 1990; Smaldino et al. 2018). See Tables 2 and S1 for an overview of the evaluated parameters.

Table 2 Global and individual variables in the present set of models. For further details, see Supplement §1

For simplicity (also see discussion), I chose to ignore elements important in previous agent-based models, including partner choice (West-Eberhard 1983; Nöe and Hammerstein 1995; Aktipis 2004; Bshary and Bergmüller 2008; Nesse 2016) reputation (Milinski et al. 2002; Nowak and Sigmund 2005; Barclay 2013), or memory (Nettle and Dunbar 1997), and instead relied entirely on partner control (sensu Gintis 2000). I also made no assumptions, except in the final model, about the degree of heterogeneity in a given population at baseline, though I do assume that behavioural mutations, both within and between generations, occur through cultural learning. This is because the first two models explore the distinction between action and preference, which I discuss further below.

In all models, at generation 1, individuals started with a random potential to reproduce (PTR) score ranging from 0–1 from a normal distribution (with a mean of 0.5 and a standard deviation of 0.2, rounded to the nearest hundredth). Individuals also started with a default action from a set of possible conditional action expressions (CAEs); these included cooperate and defect in the present models. I do not assume these variables are encoded culturally or genetically, but that a range of social and environmental factors determine these qualities. All individuals across the range of models I discuss here have both PTR and a CAE, but individuals in models 2 and 3 also have other qualities (see Table 2 and model descriptions below).

After random pairings, individuals can either cooperate (c) or defect (d), with a range of possible combinations given in Table 3; this followed standard PD models (following Axelrod and Hamilton 1981). All cooperators are assumed to be strong reciprocators who both defect against and punish defectors. The parameter values for punishing and being punished were set at − 0.1 and − 0.4, respectively, across the 3 models; these values were directly applied to the relevant individual’s PTR (Table 4). Table 4 gives the payoff structure across models; these conformed to the standard PD game rules in the literature, where dc > cc > dd > cd and 2cc > dc + cd (Press and Dyson 2012).

Table 3 Possible action combinations across models
Table 4 Payoff structures to focal individual (ego) across models

At the end of each round of dyadic interactions and once payoffs are distributed, individuals reproduce with probability PTR and individuals in the new generation inherit parent PTR and CAE. After reproduction, PTR fluctuates by a random number in the f sequence (which I held steady across models at − 0.4 to 0.4, in increments of 0.1), and individuals adopt a new default action with probability µ. The schematic for Model 1 is depicted in Fig. 1 using the DiagrammeR package for R (Iannone 2022).

Fig. 1
figure 1

Schematic of Model 1’s schedule. After random dyads form, the bubbles describe the possible actions that may follow; these do not necessarily represent the types and appearances of individuals as described in Models 2 and 3. All 3 models follow this schedule

Model 1: strong reciprocity

Model 1 is a simple version of the strong reciprocity model first developed by Gintis (2000). In this model, cooperators defect against defectors and punish them (the punishment is assumed to be costly, see Frank, 1995), directly affecting defector PTR. I assume that individuals have perfect knowledge of each other’s actions, both during and after given interactions. Note that in this model, cooperators never cooperate with defectors, so the only possible payoffs in Table 4 are 0 and 0.1. In this model, type is not distinguished from strategy, so the only possible linguistic descriptions relate to individual actions, for example player x punishes player y. The possible interactions from Model 1 are given in Table S2.

Model 2: mimicry

In this model, I introduce two new individual-level qualities: mimicry and sensitivity, numeric variables that ranged from 0–1, where mimicry allowed individuals to attempt to deceive partners into cooperating while defecting, and sensitivity determined whether a partner could detect mimicry. To reflect this addition, I introduced another individual-level quality: type, which could be mimic or honest [signaller]. Individuals thus had both strategy (CAE; cooperate/defect) and type.

Individuals each had random mimicry and sensitivity scores at generation 0 from a normal distribution; as with PTR, mimicry and sensitivity fluctuated by f after dyadic interactions. Mimics defaulted to defect and attempted to trick honest signallers into cooperating, and defected when mimicry was successful; accordingly, type allows for a more complex linguistic description than in Model 1, for example mimic x defects against honest signaller y.

The mimicry model was identical to the strong reciprocity model in all respects except attempts at mimicry. Honest cooperators and defectors did not make use of mimicry, but only sensitivity; mimics used both mimicry and sensitivity. On random pairing, a defector switches to cooperate if the relevant partner has at least twice the focal individual’s mimicry score (mimicryfocal ≤ 2 × sensitivitypartner). This behaviour was stored as an instance of cooperation in the analysis; after the interaction, the mimic switched back to their default defect strategy.

Otherwise, the focal individual defects and the payoff structure in Table 4 holds, differing from Model 1 in that mimics with mimicryfocal > 2 × sensitivitypartner deceive partners into cooperating, receiving the maximum payoff from Table 3, while the cheated cooperator receives the sucker’s payoff. Honest defectors could not receive the maximum payoff, as in Model 1.

Finally, as with Model 1, individuals switched strategy with probability µ. In addition, however, individuals also switched type with probability µ using an unrelated probabilistic function; switching type did not affect individual strategy. (For example, an honest cooperator switching type did not necessarily switch strategy. It was therefore possible to switch from an honest cooperator to a mimic cooperator in order that mutation of type did not add significantly to the change in frequency of cooperators). The possible interactions from Model 2 are given in Table S3.

Model 3: silent defectors

Finally, in the silent defector model, I introduce a third individual-level variable: appearance. The default appearance for all individuals was overt, meaning that any strategy-type combination (honest cooperator, honest defector, mimic cooperator, mimic defector) was overt, which in these models means “visible to all partners.”

Again, the model was identical to Model 2, except that, at each reproduction phase, mimics mutated to the silent appearance with probability µs (and those with the silent appearance in generations after 1 had the same probability of mutating back to overt appearance). Mutation of appearance did not affect other individual-level attributes, such as strategy. Appearance made possible linguistic descriptions still more complex than in Model 2, for example mimic x silently defects against honest signaller y.

Mimics who “discovered” or mutated the silent signalling strategy faked the cooperation strategy unless mimicryfocal ≤ sensitivitypartner, in which case they, as in Model 2, switched to cooperate for the duration of the dyadic pairing. As with Model 2, individuals who switched to cooperate were recorded as cooperating in the relevant generation.

Where mimicryfocal > sensitivitypartner among those with the silent mimicry appearance, ego defected and gained the maximum payoff where partners cooperated. The defection was, however, not detectable, which I propose was due to the mimic using a novel strategy to avoid the partner having knowledge of the mimic’s identity—a quality I discuss further below. Mimics with high mimicry scores who mutated to the silent appearance, therefore, defected without incurring punishment. Notably, furthermore, these individuals punished defectors in dyadic interactions, in kee** with behaviours proposed in the literature around the second-order free-rider problem (Heckathorn 1989; Panchanathan and Boyd 2004; see also Matthew 2017). The possible interactions in Model 3 are given in Table S4.

Analyses

I ran a proof-of-concept initial PD model without strong reciprocity to show the model’s general functionality. I then ran each model for 100,000 generations (except for the basic PD model and for Model 1, which each ran for ~ 3000 generations, with a probability function for randomly ending the model loop) with set global parameters, and calculated the frequency of cooperation as a percentage over the generations, as well as fluctuations around individual strategies (cooperate vs defect), types (mimic vs honest), and appearance (overt vs silent). All scripts were written using basic R functions in R Studio (R Core Team 2022); results were represented using the ggplot2 (Wickham et al. 2021) package for R.

Results

Strong reciprocity model (Model 1)

I initially ran a proof-of-concept PD model without conditional cooperation or punishment; the results are in Supplement §2 figures S1 and S2. As expected, and following early research into the evolution of cooperation (see Axelrod and Hamilton 1981), defection was consistently the best strategy assuming no memory of repeated interaction, with defectors quickly replacing cooperators, even with a mutation rate of 0.1.

Figure 2 gives the results from the strong reciprocity model designed to follow the basic principles as set out by Gintis (2000). With identical parameters to those in the initial PD model without strong reciprocity (N = 100; generations = 10e4; µ = 0.1; ∂ = 0.0001), cooperation quickly rose to fixation, and held near 100% despite a high mutation rate.

Fig. 2
figure 2

Percentage of cooperation in the strong reciprocity model with the above parameters; the model ran for 10,000 generations (values taken every 100 generations for viewability). Cooperation remained nearly 100% despite a high mutation rate of 0.1. Parameters: N = 100; generations = 10e4; µ = 0.1; ∂ = 0.0001

I note in particular that the relative frequency of cooperators in the population is almost identical to the percentage of cooperation (Fig. 3). There is, otherwise put, no distinction in this model between individuals who cooperate and cooperators, and individuals who defect and defectors.

Fig. 3
figure 3

Frequency of cooperators (green) and defectors (red) in the strong reciprocity model with conditional cooperation and punishment over 10,000 generations (values taken once every 100 generations for viewability). The number of cooperators in the population corresponds to the percentage of cooperative actions in the population. Parameters: N = 100; generations = 10e4; µ = 0.1; ∂ = 0.0001

Mimicry model (Model 2)

The mimicry model was identical to the strong reciprocity model, except that agents could differ by type (honest, mimic) as well as by action (cooperate, defect). Individuals also had sensitivity (0–1) and mimicry (0–1) scores to simulate the competition between disguise as a cooperator and detection of mimicry (see Methods).

Figures 4 and 5 give contrasting results to those in the strong reciprocity model, with nearly identical parameters (N = 100; generations = 10e4; µ = 0.01; ∂ = 0). Over the 100,000 generations, the percentage of cooperative actions is almost always higher than 50% (median, 82% [range 44–99%]; Fig. 4), though the frequency of defectors (honest or mimic) is always higher than the frequency of cooperators (Fig. 5).

Fig. 4
figure 4

Percentage of cooperation over 100,000 generations (values taken every 1000 generations for viewability) in Model 2. Median cooperation over the whole period is 82% (range 44–99%). Parameters: N = 100; generations = 10e4; µ = 0.01; ∂ = 0

Fig. 5
figure 5

Frequency of cooperators (green) and defectors (red) over 100,000 generations (values taken every 1000 generations) in Model 2. Defectors (honest and mimic, grouped together) generally have a higher frequency in the population throughout the run. Parameters: N = 100; generations = 10e4; µ = 0.01; ∂ = 0

Figure 6, lastly, gives the frequency of mimics and honest signallers over the 100,000 generations. Neither mimics nor cooperators consistently remain at a higher frequency where µ = 0.01).

Fig. 6
figure 6

Frequency of honest signallers (blue) and mimics (orange) over 100,000 generations (values taken every 1000 generations) in Model 2. Honest signallers may be defectors or cooperators by default; mimics default to defect. Neither type is consistently more frequent over the model’s run. Parameters: N = 100; generations = 10e4; µ = 0.01; ∂ = 0

Covert mimicry (Model 3)

The covert mimicry model is identical to the mimicry model, except that individuals have default behaviours (cooperate, defect), types (honest, mimic) and appearances (overt, covert; see Methods). Unlike with default behaviours and types, however, the population is not initially mixed, and all agents have the overt appearance at baseline. I also added a new variable, µs, the probability that an overt appearance mutates to covert, and vice versa.

Figure 7 gives the percentage of cooperation vs the number of cooperators over 100,000 generations with identical parameters to the mimicry model (N = 100; generations = 10e4; µ = 0.01; ∂ = 0); I also set µs at 0.0001). On first mutation to covert appearance, the percentage of cooperation drops from centering above 75% to centering above 50%; this trend continues throughout the remaining generations (median cooperation, 60% [range 27–97%]).

Fig. 7
figure 7

Percentage of cooperation vs cooperators over 100,000 generations with the above parameters in the covert mimicry model (values taken once every 1000 generations) in Model 3. Cooperation drops from centering above 75% to centering above 50% after first mutation to the covert appearance (median cooperation, 60% [range 27–97%]); the number of cooperators, however, is consistently lower. Parameters: N = 100; generations = 10e4; µ = 0.01; µs = 0.0001; ∂ = 0

Once the first mutation to the covert appearance occurs, furthermore, individuals with that appearance quickly go into fixation, with occasional mutations back to the overt appearance (Fig. 8). This has meaningful impacts on the default strategies and types of individuals in the population, even though cooperation remains largely the majority behaviour. Figure 8 includes the frequencies of defector strategy, mimic type, and silent appearance over 100,000 generations in Model 3. See the online supplement for further plots contrasting strategies, types, and appearances.

Fig. 8
figure 8

The frequencies of defector strategy, mimic type, and silent appearance (grouped as individual “qualities”) in the population over 100,000 generations (values taken once every 1000 generations) in Model 3. After initial mutation to the covert appearance, this trait largely remains in fixation, and the defector strategy and mimic type also increase to nearly 100% after silent mimics enter the population, though there is some fluctuation. This finding is of particular note given the high rate of cooperation still noted in the population (Fig. 7). Parameters: N = 100; generations = 10e4; µ = 0.01; µs = 0.0001; ∂ = 0

Discussion

The aims of this paper are two-fold: to show, in line with previous suggestions in theoretical discussions (Trivers 1971), that the percentage of cooperative behaviours and the number of cooperators in a population are not identical, and that strategies that allow agents to evade detection do not markedly decrease the percentage of cooperative behaviours, but regardless spread in the population. I also suggest that these findings call into question the force of model-derived hypotheses from strong reciprocity theorists. The possibility of invasion by unknown strategies, such as silent defection, which, by definition, simple models cannot account for, allows for the illusion of high rates of cooperation without a high frequency of cooperators. Cooperation is phenotypic, and while two strategies may be phenotypically identical, we cannot speculate about the underlying motivations driving them. In this section, I discuss the implications of each of these points within the context of previous work.

Cooperation versus cooperators

In his classic paper on reciprocity, Trivers (1971) noted that individuals are unlikely to be cooperators or free riders, but to switch between strategies depending on the context. The question of whether individuals belong to the class “cooperator” or “free rider” has been put in previous work (for example, McElreath et al. 2003a, b), though it remains largely unexplored in models simulating the circumstances under which cooperation is likely to evolve.

Classic strategies (Tit-for-Tat, see Axelrod and Hamilton 1981; “forgiving”, see Fudenberg et al. 2012; GRIM, Friedman 1971; Axelrod 2000) and newer strategies, such as zero-determinant strategies (Press and Dyson 2012), cooperate or defect conditionally, though in the case of Tit-for-Tat, the default behaviour on first interaction is cooperate—earning this strategy the label “cooperator.” My aim with the paradigm presented in the above set of models is, firstly, to question the justification of this label.

The initial model gives us only the information that agent x did or did not cooperate with y, and vice versa. This leaves open the possibility that x would have behaved differently, had the circumstances been different, and moreover equates the agent’s action with any underlying motivations that led to the action (see Hagen and Hammerstein 2006; Sperber and Baumard 2012; Casey et al. 2021). We should not, in other words, assume that a high frequency of cooperation implies a high number of cooperators in a given population of agents—the circumstances are too vague to tell us more than that, under the circumstances given, cooperation is likely to evolve and become common in a population (as, for example, is seen in many models with Tit-for-Tat and punishment strategies).

Evading detection

The design of the above models aims to make clear that “mimic x cooperates with honest signaller y” appears to be the same interaction as “player x cooperates with player y.” Yet the extra linguistic information given in the former formulation identifies the mimic as an individual who cooperates only because of the circumstances, not because of its status as a cooperator. Thus individuals whose default behaviour would be to defect cooperate not necessarily because of prosocial preferences, but possibly because they perceive that the threat of detection is too high to risk exposure to punishment.

When, as in this model, we introduce the possibility of hiding defection, these individuals continue to cooperate at a high rate, though strong reciprocators—at least, in their simplest form—are almost entirely replaced in the population, and the covert strategy goes into fixation shortly after first mutation to it. This emulates the results of Nettle and Dunbar (1997), which explored the impact of a mimic strategy in a PD, though these authors found that an increasingly complex signal, coupled with high memory capacity, can prevent mimics from overwhelming a population (though also see Wiseman and Yilankaya 2001; this may relate to the importance of trust based on signals in determining whether to cooperate with a partner [Han et al. 2021]). In the present model, however, the covert mimicry strategy may prevent anti-defection strategies altogether, either through cooperation with highly sensitive individuals or through defection without detection.

An interesting question will be the degree to which opportunistic individuals can avoid detection in their relevant cultures, and whether their strategies reflect an arms race between mimicry and mimicry sensitivity over cultural evolutionary time. While Sperber and Baumard (2012) suggest that Machiavellian-like social qualities (following Humphrey 1976; Byrne and Whiten 1988) may be frequency-dependent in a given society—otherwise free-riding would quickly go into fixation and be open to invasion by cooperators—it is unlikely that detection evasion would be an effective long-term strategy in small-scale societies (for a contemporary example, see Wiessner 2005). As humans probably evolved living in groups where individuals had an average of no more than 150 close connections (Dunbar 1993; but see Lindenfors et al. 2021, for a recent critique of Dunbar’s calculations), the circumstances under which one could mask one’s identity, or to otherwise defect without detection, would have been limited, save for some of the circumstances Trivers (1971) notes, such as refusing help to someone likely to die.

As groups grew in size, however, the opportunities for defection without detection would have been likely to increase (Moreau 2020). Human groups began and continue to rely on concentric circles of connections; these wider “cognitive groups” allow for greater opportunities for subtle cases of deceit and defection by outsiders. While many researchers accept the notion that punishment and sanctions against norm violators kept defection in check (Fehr and Gachter 2002; Panchanathan and Boyd 2004), it is possible that, following Sperber and Baumard’s (2012) prediction, a small number of such violators might be able to successfully endure. Strategies might include masquerading as an in-group member by faking tags (as distinguished from signals, see Axelrod et al. 2004) associated with membership (for a general model of tag-based cooperation, see Riolo et al. 2001; see also McElreath et al. 2003a, b; Traulsen and Nowak 2007; Cohen 2012; Cohen and Haun 2013; Moya and Boyd 2016; Bell and Paegle 2021; see also Goodman et al. 2023), or moving quickly between inhabited areas, placing costs on locals and leaving before others notice one’s defections. Moreover, most previous models do not explore the cost of checking whether one has been cheated (Han et al. 2021). Furthermore, there is nothing to suggest, either in models or in studies of small-scale cultures, that should circumstances change, new behavioural mutations would necessarily not invade. Wrangham’s (2021; see also Hare 2017) position that ancient humans selected against reactive aggression as part of the self-domestication process does not, as an example, suggest that, as societies grew larger, proactive aggression and Machiavellian strategies could not emerge that subverted within-group cooperation.

To counter these risks, growing communities were likely to invest in mechanisms such as policing to enforce social norms around cooperation (Tullock 2004, inter alia). These investments and their consequent mechanisms reflect an arms race between individuals who aim to subvert cooperative systems through mimicry and those who cooperate honestly. Yet as social systems grew (and continue to grow) in complexity, improved mechanisms for detecting cheaters were (and are) required, in an analogous way to how immune systems evolve to thwart disease (sensu Aktipis 2020; see also Goodman and Ewald 2021). The consequence is that, across large-scale societies, we do not see the virtual ubiquity of cooperation predicted by strong reciprocity models. Novel methods for defection and exploitation also emerge regularly. One recent example in Western societies is affinity fraud, where individuals use signals reflecting shared social identity to garner trust and then exploit victims, usually for financial reasons (Blois and Ryan 2013). These and analogous dishonest signals of cooperative intent are, however, likely to become costlier to develop and enact as cultural groups improve detection strategies.Footnote 2

Empirical tests may explore the validity of this set of points. In a public goods game, for example, it may be that individuals who appear to be following norms—giving some, but not too much (see Barclay and Willer 2007)—may end with a greater number of resources insofar as they defect only in the final round. Analogous points might be true in ultimatum, dictator and PD games—and may help to explain the finding that individuals who score high psychopathy tests tend to do well in laboratory experiments of PD games (Mokros et al. 2008). Moreover, frequency-dependent opportunism may explain why, despite evidence for hypotheses around self-domestication and cultural group selection, psychopathy continues to pervade society.

Moreover, while researchers, particularly in evolutionary psychology, suggest that humans evolved a cheater detection module (Cosmides and Tooby 1992; see also Verplaetse et al. 2007), the models presented here suggest it is unlikely that any module is effective enough to thwart all free riders in a given culture. Previous research suggests, for example, that mimicry of honest human social signals is detected in about two-thirds of cases (laughter, Bryant and Aktipis 2014; lies, Fonseca and Peters 2021; cooperative intent, Verplaetse et al. 2007; accents of language, Tate, 1979, Goodman et al. 2021). Yet where honest signallers and mimics (and overt and silent signallers) interact over generations, a two-thirds average detection rate suggests an arms race between these types and appearances, while maintaining cooperation at a high level (see Supplement §3).

Cultural evolution and internalization

One possible objection to the claims I make here is that punishment strategies evolved into norms, which spread across societies in our evolutionary past (Boyd and Richerson 2002; Boyd et al. 2011; for a study of relevance to religion see, for example, Atran and Henrich 2010). Cultural groups in which individuals behaved similarly—cooperating within the culture and punishing sanction violators—were likely to outcompete groups without effective cooperation-enforcing norms (Henrich and Henrich 2007). Furthermore, to the extent that individuals within successful cultures internalized the norms enforcing cooperation, which may entail forming prosocial preferences (Henrich and Muthukrishna 2021), group success–maximizing behavioural trends were likely to spread.

I am not claiming, however, that groups with opportunistic free riders who mimic cooperators are likely to be less successful. It may even be that, following some claims given in the self-domestication hypothesis (e.g., Wrangham 2021, inter alia), ancient humans selected for, not just cooperative tendencies, but intelligence for navigating complex social situations (the social brain hypothesis, Humphrey 1976; Gavrilets and Vose 2006; Shultz and Dunbar 2007; see also Dunbar and Shultz 2017). If true, it is possible that successful cultural groups do not only consist of cooperative individuals, but individuals capable of steering society—and consequently, themselves—to success through effective strategizing that may involve Machiavellian strategizing (Humphrey 1976). Internalization is not a valid objection, as the models presented here make no claim about beliefs, but rather the possibility of disguised defection. Previous work shows, for example, that individuals are effective at justifying transgressions (see Mazar et al. 2008), and self-deception may help reduce the cognitive load of free riding when opportunities are presented (e.g., Trivers 2000).

A second objection is that disguised mimics are likely to overwhelm a population of free riders, leaving them open to invasion by strong reciprocators through cultural evolutionary processes. For example, given that humans have an evolved bias to copy successful individuals, or to conform to behaviours around them (Boyd and Richerson 2002; Henrich and Henrich 2007), free riders are likely to be copied until the behaviour reaches fixation. While this is a likely outcome where appearances are overt, the silent mimicry strategy cannot necessarily be copied—the successful individuals will appear to be cooperators even if they defect. This prevents the behaviour spreading through a conformity or prestige bias, and maintains silent defection as a frequency-dependent strategy. Insofar as the silent, opportunistic defector does not teach its behaviour to others in the next generation (or in a genetic model, does not reproduce more than others in the population), the strategy will not overwhelm a population of cooperators—a simulation worth exploring in future models.

Finally, one may object that these findings relate only to cooperation in dyads, while major issues in human cooperation are more likely to concern large-scale social dilemmas. This is an invalid objection for two reasons. First, dyadic meetings between strangers—that is, where agents have little to no information about their potential cooperative partners—remain an evolutionary problem from a game-theoretical point of view. Second, and more importantly, covert mimicry is an equally concerning strategy in large-scale social dilemmas, where free riders may cheat, potentially with less risk of detection than in dyadic pairings. Further research, both modelling and empirical, may confirm this suggestion.

Language and the Ring of Gyges effect

A further strategy for masking uncooperative behaviour is language: insofar as individuals can effectively justify norm transgressions, they can avoid sanctions (relying on the distinction between a rule’s wording and judgments about it, see Rawls 1955). Language is therefore an effective tool for masking opportunistic defection, and presents strategic individuals with opportunities for defection in complex social scenarios.Footnote 3 This may involve lying—or just effective rhetoric—but there is no established reason to suppose that either strategy is likely to be out-competed by individuals who internalize norms of strong reciprocity. The linguistic descriptions attending the present set of models attest to this; the simple descriptions of strategies in Model 1 give far less information than do the types and appearances of Models 2 and 3, respectively. There is, however, no reason to assume that, in any real-world scenario, attributing adjectives and adverbs to social interactions will not colour the actions of individuals in ways that elicit moral judgment. Saying Donald cooperated with Boris does not imply that Donald is a cooperative person, or that he does not also silent defect against Boris, should the opportunity arise.

A further but related problem may be found in Plato’s Republic, where Gyges of Lydia, a shepherd, found a ring in a cave that allowed him to become invisible. Gyges used his newfound power to murder the king, marry the queen, and install himself as ruler. While invisibility is not possible, language allows individuals to hide their defections from others, saying, as in Model 3, that a player silently defects assumes perfect knowledge. And yet if Donald were to silently defect against Boris, by definition, no one would know about it. Language creates, in other words, a Ring of Gyges effect through which strategic humans hide their exploitative actions.

Consequences and future directions

Taken together, the findings from the models given here, coupled with the human ability to navigate social circumstances effectively using language, suggest that models in which individuals are presented only with a few options cannot account for strategies for defection while avoiding punishment. Following Dennett (1988), who notes, following the philosopher Karl Popper, we should assume that the human ability to innovate—creating previously unknown strategies in social competitions—is problematic for models that account only for a set of established strategies in games like the PD. Future work should, following the results of Dana et al. (2007), try to address this by giving, in laboratory conditions, only a subset of individuals an opportunity for hidden defection, and establishing whether individuals are likely to take the opportunity. Similarly, these findings should make ethnographic researchers more cautious when making claims about human prosocial preferences, such as inequity aversion, following experiments using economic games (Fehr and Schmidt 1999). Humans across cultures may not act in these games entirely in their economic self-interest, but it does not follow that these behaviours reflect prosociality.

The problem of opportunity is that individuals may not defect in cooperation-related games not because of prosocial preferences, but because they do not have the opportunity to do so. This does not undermine the cultural group selection arguments based on internalized norms and prosocial preferences (e.g., Henrich 2020), but rather suggests that such internalization and prosociality may not be as widespread as supposed in culturally successful groups (see Burton-Chellew and West 2013). Some individuals may only mimic the behaviours of devout actors (Singh and Hoffman 2021; see also Atran and Henrich 2010, for a discussion of the concept of devout actors), and defect when they have the chance, or think of a new way to do so.

The models presented may be explored with other factors in mind. First, unlike with other PD models representing dyadic interactions, I do not account for genetic (or cultural) flow; individuals cannot move between islands (as, for example, in Leimar and Hammerstein 2001). This is an unrealistic assumption, but for simplicity I do not consider the alternative here. Similarly, and more importantly, I ignore the influence of indirect reciprocity, reputation, and biological markets. While I recognize these are foundational elements in non-kinship-based cooperation models, I again do not consider them here for the sake of simplicity. In both cases, furthermore, I suggest that hidden strategies are no less likely to evolve if accounted for in the models, but rather the manifestation of hidden defection will be different. Future work should evaluate this assumption.

Conclusion

In this paper, I have suggested that the possibility of hidden mimicry presents a problem for theories of cultural group selection relying on strong reciprocity. While prosociality is essential for success in between-group competition and to a functional society, individuals who cooperate only insofar as it benefits them—and moreover, who defect when presented with the opportunity—may thrive within a population without reducing cooperation below 50% in a simulation of agents. Individuals may therefore behave prosocially without being prosocial—an element of human psychology that researchers cannot account for in agent-based models. While cooperation, like honesty, may be the best policy, we should not assume that individuals who cooperate are themselves cooperators. The complexity of language, and the uniquely human ability to represent oneself differently in different social situations, make the possibility of undetected defection real in everyday scenarios. With growing civil unrest across the world, and with a growing number of political leaders who continue to undermine cooperative norms, future work in the social sciences should look to ways to promote cooperation, rather than to assume its pervasiveness.