Hostname: page-component-cd9895bd7-fscjk Total loading time: 0 Render date: 2024-12-26T05:58:23.712Z Has data issue: false hasContentIssue false

Self-interested learning is more important than fair-minded conditional cooperation in public-goods games

Published online by Cambridge University Press:  17 October 2022

Maxwell N. Burton-Chellew*
Affiliation:
Department of Economics, HEC-University of Lausanne, 1015 Lausanne, Switzerland Department of Ecology and Evolution, Biophore, University of Lausanne, 1015 Lausanne, Switzerland.
Claire Guérin
Affiliation:
Department of Ecology and Evolution, Biophore, University of Lausanne, 1015 Lausanne, Switzerland.
*
*Corresponding author. E-mail: maxwell.burton@unil.ch

Abstract

Why does human cooperation often unravel in economic experiments despite a promising start? Previous studies have interpreted the decline as the reaction of disappointed altruists retaliating in response to non-altruists (Conditional Cooperators hypothesis). This interpretation has been considered evidence of a uniquely human form of cooperation, motivated by an altruistic concern for equality (‘fairness’) and requiring special evolutionary explanations. However, experiments have typically shown individuals not only information about the decisions of their groupmates (social information) but also information about their own payoffs. Showing both confounds explanations based on conditional cooperation with explanations based on confused individuals learning how to better play the game (Confused Learners hypothesis). Here we experimentally decouple these two forms of information, and thus these two hypotheses, in a repeated public-goods game. Analysing 616 Swiss university participants, we find that payoff information leads to a greater decline, supporting the Confused Learners hypothesis. In contrast, social information has a small or negligible effect, contradicting the Conditional Cooperators hypothesis. We also find widespread evidence of both confusion and selfish motives, suggesting that human cooperation is maybe not so unique after all.

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
Copyright © The Author(s), 2022. Published by Cambridge University Press

Social media summary: Do people strive for equality or personal success? Decoupling conditional cooperation and learning in public-goods games

Introduction

Sound knowledge about the specific motives behind altruistic acts predominantly stems from laboratory experiments. (Fehr & Fischbacher, Reference Fehr and Fischbacher2003)

We need to ask what the totality of [experimental] evidence tells us about the maintained assumptions of theory and its test framework. (Smith, Reference Smith2010)

Economic experiments using the public-goods game have shown that many people initially make costly contributions towards a group beneficial account (cooperate) but that this cooperation is fragile and typically declines over time (Burton-Chellew & West, Reference Burton-Chellew and West2021; Chaudhuri, Reference Chaudhuri2011; Jouxtel, Reference Jouxtel2019; Ledyard, Reference Ledyard, Kagel and Roth1995; Neugebauer, Perote, Schmidt, & Loos, Reference Neugebauer, Perote, Schmidt and Loos2009; Thielmann, Böhm, Ott, & Hilbig, Reference Thielmann, Böhm, Ott and Hilbig2021; Zelmer, Reference Zelmer2003). There are two major competing explanations for this decline in contributions that are both based on different forms of learning (Andreoni, Reference Andreoni1995; Burton-Chellew & West, Reference Burton-Chellew and West2021; Cooper & Stockman, Reference Cooper and Stockman2002; Houser & Kurzban, Reference Houser and Kurzban2002). One hypothesis is that cooperation unravels owing to disappointment among cooperative individuals as they learn that people are not as cooperative as they anticipated (Fischbacher & Gachter, Reference Fischbacher and Gachter2010; Fischbacher, Gachter, & Fehr, Reference Fischbacher, Gachter and Fehr2001). This ‘Conditional Cooperators’ hypothesis posits that people are mostly motivated by an altruistic concern for equality (or ‘fairness’), and that human cooperation requires a unique evolutionary explanation (Fehr & Fischbacher, Reference Fehr and Fischbacher2003; Fehr & Schurtenberger, Reference Fehr and Schurtenberger2018; Henrich & Muthukrishna, Reference Henrich and Muthukrishna2021). The competing hypothesis is that cooperation decreases as confused individuals learn that contributing is not as profitable as they thought. This ‘Confused Learners’ hypothesis posits that people are mostly motivated by self-interest but are not well adapted to laboratory games and thus need experience to recalibrate their behaviour (Barclay, Reference Barclay2012; Barclay & Daly, Reference Barclay and Daly2003; Bateson, Nettle, & Roberts, Reference Bateson, Nettle and Roberts2006; Burnham & Hare, Reference Burnham and Hare2007; Burton-Chellew, El Mouden, & West, Reference Burton-Chellew, El Mouden and West2016; Burton-Chellew, Nax, & West, Reference Burton-Chellew, Nax and West2015; Burton-Chellew & West, Reference Burton-Chellew and West2013, Reference Burton-Chellew and West2021; Hagen & Hammerstein, Reference Hagen and Hammerstein2006; Haley & Fessler, Reference Haley and Fessler2005; Krupp et al., Reference Krupp, Barclay, Daly, Kiyonari, Dingle and Wilson2005; McAuliffe, Burton-Chellew, & McCullough, Reference McAuliffe, Burton-Chellew and McCullough2019; Pedersen, Kurzban, & McCullough, Reference Pedersen, Kurzban and McCullough2013; Raihani & Bshary, Reference Raihani and Bshary2015; Silva & Mace, Reference Silva and Mace2014).

Despite decades of study there is still much debate and controversy over these competing explanations (Burton-Chellew & West, Reference Burton-Chellew and West2021; Camerer, Reference Camerer2013; Fehr & Schurtenberger, Reference Fehr and Schurtenberger2018). One reason for this is because the standard experimental design typically confounds these two explanations, by coupling together information on earnings from the game (payoff information) with information on groupmates’ decisions (social information). If individuals can observe both payoff and social information, it is unclear which information is more psychologically salient and motivating behaviour. This is a problem because the two forms of information are often highly correlated: more cooperative groupmates lead to higher payoffs and vice versa. Consequently, both the Conditional Cooperators and the Confused Learners hypotheses can predict an increase in cooperation in response to more cooperation by others and vice versa (Burton-Chellew, El Mouden, & West, Reference Burton-Chellew, El Mouden and West2017b; Burton-Chellew et al., Reference Burton-Chellew, Nax and West2015; Nax, Burton-Chellew, West, & Young, Reference Nax, Burton-Chellew, West and Young2016).

One solution to the problem of confounding is to engineer modified versions of the public-goods game that test for either Conditional Cooperators or Confused Learners (Andreoni, Reference Andreoni1995; Andreozzi, Ploner, & Saral, Reference Andreozzi, Ploner and Saral2020; Angelovski, Di Cagno, Güth, Marazzi, & Panaccione, Reference Angelovski, Di Cagno, Güth, Marazzi and Panaccione2018; Bayer, Renner, & Sausgruber, Reference Bayer, Renner and Sausgruber2013; Burton-Chellew et al., Reference Burton-Chellew, El Mouden and West2016; Burton-Chellew & West, Reference Burton-Chellew and West2013; Croson, Fatas, & Neugebauer, Reference Croson, Fatas and Neugebauer2005; Di Cagno, Galliera, Güth, & Panaccione, Reference Di Cagno, Galliera, Güth and Panaccione2016; Ferraro & Vossler, Reference Ferraro and Vossler2010; Fischbacher & Gachter, Reference Fischbacher and Gachter2010; Fischbacher et al., Reference Fischbacher, Gachter and Fehr2001; Gunnthorsdottir, Houser, & McCabe, Reference Gunnthorsdottir, Houser and McCabe2007; Houser & Kurzban, Reference Houser and Kurzban2002; Kocher, Martinsson, Persson, & Wang, Reference Kocher, Martinsson, Persson and Wang2016; Kuemmerli, Burton-Chellew, Ross-Gillespie, & West, Reference Kuemmerli, Burton-Chellew, Ross-Gillespie and West2010; Kurzban & Houser, Reference Kurzban and Houser2005; Shapiro, Reference Shapiro2009; Thoni & Volk, Reference Thoni and Volk2018). For example, the strategy method, which forces individuals to specify in advance how they will respond to their groupmates’ contributions, was used to show that many individuals will, if so prompted, condition their contributions on the contribution of their groupmates, consistent with a concern for equality (Fischbacher & Gachter, Reference Fischbacher and Gachter2010; Fischbacher et al., Reference Fischbacher, Gachter and Fehr2001; Kocher, Cherry, Kroll, Netzer, & Sutter, Reference Kocher, Cherry, Kroll, Netzer and Sutter2008). Alternatively, games played with computerised groupmates have also shown conditional behaviour, symptomatic of ‘confusion’, and games with a ‘black box’ have shown similar declines in contributions that can only be explained by payoff-based learning (Burton-Chellew et al., Reference Burton-Chellew, El Mouden and West2016; Burton-Chellew & West, Reference Burton-Chellew and West2013, 2021; Ferraro & Vossler, Reference Ferraro and Vossler2010; Houser & Kurzban, Reference Houser and Kurzban2002; Shapiro, Reference Shapiro2009). However, while these experiments are informative, it could be argued these modified public-goods games are problematic because they are either not the same game or not the same decision mechanism, and thus behaviour may change in unknown ways (Camerer, Reference Camerer2013; Fischbacher, Gächter, & Quercia, Reference Fischbacher, Gächter and Quercia2012).

Here we avoided this potential problem by retaining the basic game and decision process of the standard public-goods game. We used standard instructions but merely varied what information participants received as feedback after each round of decision making (Fischbacher & Gachter, Reference Fischbacher and Gachter2010). This way we could decouple payoff information from social information during the learning process (Figure 1). Our primary method was to provide individuals with either just their personal payoff information (Payoff treatment), or just social information on the decisions of their groupmates (Social treatment, individual decisions were shown in either full, or just the group average decision). This decoupling allowed us to evaluate the relative importance of payoff-based learning and conditional cooperation for causing the decline in contributions. If the decline is only due to payoff-based learning, then the decline will not occur unless payoff information is shown i.e., when participants are only shown social information. In contrast, if the decline is only due to conditional cooperation, then adding payoff information will not increase the rate of decline.

Figure 1. Experimental design. We decoupled two typical forms of information, about either the contribution decisions of the focal player's groupmates (social information) or the focal individual's own earnings (yellow/top individual, payoff information). This made four general treatments: one treatment with neither form of information (‘No-info’, N = 21 groups); one with just social information (‘Social’, N = 53 groups; 20 groups observed all individual decisions whereas 33 groups observed the group average only); one with just payoff information (‘Payoff’, N = 40 groups); and one with both forms of information (‘Combined’, N = 40 groups; 20/20 groups with either full/limited social information). We never showed the earnings (payoff) of groupmates. Experimental timeline: participants first had one round of decision making with computerised groupmates in a public-goods game (income maximisation test, IMT1). They were then randomly assigned to one of four information treatments for nine rounds in a constant group of four real people. They then repeated the income maximisation test (IMT2) before two survey questions (Qs) regarding their motivation and their understanding of the payoffs.

We also included a control treatment to provide a baseline measure for how behaviour may change over time when individuals have neither payoff nor social information, but still have time to reflect and re-read the payoff instructions (No-info treatment, Figure 1; Burton-Chellew et al., Reference Burton-Chellew, El Mouden and West2017b; Neugebauer et al., Reference Neugebauer, Perote, Schmidt and Loos2009). We also included a fourth treatment that presented the usual, combined, information, replicating the typical design in the literature (Combined treatment). Crucially, this combined treatment allowed us to evaluate the effect of adding payoff information to social information, which is technically redundant if individuals fully understand the game (a key assumption of the conditional cooperation interpretation). This redundancy is a strength of our design, because if individuals perfectly understand the game, then behaviour in the Combined treatment should not differ from the Social treatment. If, on the other hand, individuals are confused learners, their behaviour should differ across treatments, with a faster decline in contributions when payoffs are shown. We also included a range of alternative decision tasks both before and after the main game to test for confusion and motivations more directly (Figure 1).

Methods

Data collection

We collected data two times, a year apart. The collection bouts were identical in design except that degree of social information showed either all individual contributions (full social information), or just the group average contribution (limited social information). In both cases adding payoff information is technically redundant because both forms can be used to calculate personal payoffs, but only the limited social information can be calculated from personal payoffs. The first collection (‘study 1’) used detailed social information and 280 participants (70 groups) in the Autumn semester of 2018, from 21 to 30 November. The second bout of data collection (‘study 2’), with limited social information, used 336 participants (84 groups) in the Autumn semester of 2019, from 7 to 11 October. All data were anonymous and collected electronically.

Participants and location

In total we ran 40 sessions with 616 participants (154 groups). Each session involved 12–20 participants (three to five groups). All sessions were conducted at the Faculty of Business and Economics (HEC), University of Lausanne (UNIL), Switzerland, in the HEC-LABEX facility. LABEX forbids deceiving participants and requires that all experimental designs obtain prior ethical approval from the LABEX ethics committee. HEC-LABEX used the ORSEE software to recruit participants and excluded all participants from previous experiments by the same authors (Greiner, Reference Greiner2015). Participants were mostly students enrolled at either UNIL or the Swiss Federal Polytechnic School, and they can be from diverse ethnic, socio-economic and cultural backgrounds (these variables were not recorded). According to the questionnaires we had a near equal gender ratio (306 female, 304 male, 2 other, and 4 declined to answer) and most of our participants were under 26 years of age (277 aged under 20, 309 aged 20–25, 25 aged 26–30, two aged 30–35, two over 35 and one declined to answer, responses to the age question were categorical to increase anonymity). We wanted to avoid participants that had studied the game theory of social dilemmas so in study 1 we excluded HEC students. In study 2 we were advised to allow first year students as it was early in the academic year.

Financial incentives

We paid all participants a show-up fee of 10 Swiss Francs (CHF). For each of the 11 incentivised decisions, we endowed participants with 20 Monetary Units (MU) and each MU was worth 0.025 CHF, so 20 MU were worth 0.5 CHF. While these stakes are not large, they are comparable with related literature, and generally speaking, studies have found little effect of using larger stakes (Kocher, Martinsson, & Visser, Reference Kocher, Martinsson and Visser2008; Larney, Rotella, & Barclay, Reference Larney, Rotella and Barclay2019; Yamagishi, Li, Matsumoto, & Kiyonari, Reference Yamagishi, Li, Matsumoto and Kiyonari2016).

Experimental procedure

The experiment had several stages (Figure 1). We provide a full copy of the instructions in English in our Supplementary Methods. For each stage participants had to answer questions regarding their understanding of the instructions (‘control questions’). For all questions, participants were allowed unlimited time and two attempts before we showed them the correct answers.

Procedures were standard for a behavioural economics experiment. Participants were randomly assigned to a partitioned computer terminal. We announced that communication was forbidden, and that the HEC-LABEX forbids deception. Participants then progressed, at their own pace, through on-screen instructions and control questions based on hypothetical scenarios detailing the general public-goods game decision and payoffs. We used publicly available instructions, control questions, and structural parameters and translated them into French (Fischbacher & Gachter, Reference Fischbacher and Gachter2010). The specific parameters were group sizes of four participants, an endowment of 20 MU, and a multiplier/efficiency Factor of 1.6, which meant all contributions were multiplied by 1.6 before being shared out equally, to give a marginal per capita return of 0.4. As this is less than 1, the rational selfish income maximising decision in a single round is to contribute 0 MU. At this stage the instructions have only described the general decision task and the payoffs involved. The instructions so far have not mentioned game length, type of groupmates or information feedback.

We then used an asocial control treatment to test if individuals maximised their income when there were no social concerns (pre-game income maximisation test, IMT1). We informed participants that they would face the same decision but in ‘a special case’ with computerised groupmates programmed to play randomly. We assured them that this decision would not be seen by other participants, would have no future consequences, and no financial consequences for other participants but did have financial consequences for them. We made them answer five true or false statements about the situation they faced and provided them with the correct answers afterwards (Supplementary Methods). After their decision, we did not provide them with any immediate feedback about their earnings or the computer decisions, to prevent learning about the game before the information treatments. This income maximisation test allowed us to estimate how much individuals still contribute even when they can have no rational concern for fairness or helping others.

Next, participants learned that they would now face the same decision but with real people, as outlined in the original instructions. We informed them that they had been randomly grouped with three other participants for nine rounds of decision making. Although repeated games can favour some strategic cooperation, previous studies using constant groups have still observed the typical decline in contributions (Andreoni & Croson, Reference Andreoni, Croson, Plott and Smitt2008; Burton-Chellew & West, Reference Burton-Chellew and West2021), and backwards induction would mean that the selfish income maximising decision in a finite game of known length among rational individuals is still to contribute 0 MU (Ambrus & Pathak, Reference Ambrus and Pathak2011; Burton-Chellew, El Mouden, & West, Reference Burton-Chellew, El Mouden and West2017a; Dijkstra & van Assen, Reference Dijkstra and van Assen2017; Kreps, Milgrom, Roberts, & Wilson, Reference Kreps, Milgrom, Roberts and Wilson1982; Krockow, Colman, & Pulford, Reference Krockow, Colman and Pulford2016). We then explained to each group, according to its randomly assigned treatment, what type of information they would all be shown after each round. We assured them that all groupmembers would receive the same type of information. Again, we also made them answer five true or false statements about the situation they faced.

Specifically, we told participants, in the:

  • No-info treatment, ‘You and everyone else in the group will not receive any information after each round. No participants will be able to know your investments at any time. Your earnings will not be shown to you each round, but you will receive the money at the end of the experiment’.

  • Social treatment, ‘The information that each person will receive will only be the decision of each group member/average decision of the group [study 1/study 2]. Your earnings will not be shown to you each round, but you will receive the money at the end of the experiment’.

  • Payoff treatment, ‘The information that each person will receive will only be their own earnings in each round. No participants will be able to know your investments at any time’.

  • Combined treatment, ‘The information that each person will receive will only be their own earnings, and the decision of each group member/and the average decision of the group [study 1/ study 2]’.

We varied the detail of the social information because showing the individual decisions, as we did in study 1, could either impede the decline via individuals attempting to strategically signal their cooperative behaviour, or could, conversely, accelerate the decline if individuals responded more strongly to the presence of low/zero contributors than high/full contributors (Burton-Chellew et al., Reference Burton-Chellew, El Mouden and West2017a; Carpenter, Reference Carpenter2004). Therefore study 2 used the simpler information of just showing the group average to avoid these effects.

We used fewer groups in our baseline measure, the No-info treatment (N = 21), because we wanted to devote more resources to the other treatments. This was also justified because the individuals are independent data points in the No-info treatment (because they cannot affect the behaviour of their groupmates). We used more groups in the Social treatment (N = 53) because we reasoned it was the treatment that was least well studied in the literature (studies normally show participants their payoffs). Therefore, whenever we had enough participants to form more than four groups in a session (in ‘study 2’), we allocated the extra group to the Social treatment. Our sample sizes, of 20–53 groups per treatment, are comparable with if not larger than related literature (Andreoni, Reference Andreoni1995; Bayer et al., Reference Bayer, Renner and Sausgruber2013; Chaudhuri, Paichayontvijit, & Smith, Reference Chaudhuri, Paichayontvijit and Smith2017; Fischbacher & Gachter, Reference Fischbacher and Gachter2010).

After the repeated public-goods game with information treatments the participants were then told that they would again face the special case with computers, for just one round. This post-game income maximisation test (IMT2) allowed us to measure how the different information treatments had affected learning about how to maximise income.

At the end of the experiment, we asked participants two unincentivised questions regarding their understanding of the game's payoffs, and their motivation when grouped with humans. Specifically, to test understanding of the game's payoffs, we asked them, ‘In the basic decision situation, played for one round only, if a player wants to maximise his or her earnings, should they decide their contribution depending on what the other people in their group contribute?’ Participants had a choice of four answers: the correct answer, ‘No’; or two incorrect answers, ‘Yes’ or ‘Sometimes’; or they could respond ‘Do not know’ (Burton-Chellew et al., Reference Burton-Chellew, El Mouden and West2016). Asking participants about a strictly one-shot game allowed us to more cleanly measure their understanding of the game's payoffs and the social dilemma involved than if we had asked them about a repeated game. If a participant does not know the correct answer to this question, then they do not understand the game's payoffs.

To measure motivations we asked, ‘Which of these descriptions best describes your motivation during the rounds with humans?’ and offered participants four possible responses, that corresponded to selfishness (‘Making myself as much money as possible’), competitiveness (‘Making myself more money than other people’), a desire for fairness, as is often assumed to motivate conditional cooperation (‘Avoiding unequal outcomes so that I make neither more, nor less, than other people’) (Fehr & Fischbacher, Reference Fehr and Fischbacher2003; Fehr & Schmidt, Reference Fehr and Schmidt1999; Fehr & Schurtenberger, Reference Fehr and Schurtenberger2018) or a desire to help the group (‘Making the group as much money as possible even if it meant making myself less money’). For both questions we reversed the on-screen order of responses for half of the participants (two within each group).

Statistical analyses

Our analysis script is available on the Open Science Framework. We analysed the data using R-Studio version 1.3.1093 (R Team, 2020). Data were imported with the zTree package (Kirchkamp, Reference Kirchkamp2019). All statistical tests were two-tailed. Significance values in linear mixed models were estimated by the lmerTest package, which uses the Satterthwaite approximation for estimating degrees of freedom (Kuznetsova, Brockhoff, & Christensen, Reference Kuznetsova, Brockhoff and Christensen2017). Data figures were made with the ggplot2 package (Wickham, Reference Wickham2009).

When analysing the rate of decline in contributions we did not use individual contributions because responses within constant groups are not independent data. Instead, we analysed mean group contributions, which also have the benefit of tending towards a normal distribution (central limit theorem, Crawley, Reference Crawley2007). This allowed us to use a linear mixed model (LMM), with more interpretable coefficients on the rate of decline. We used maximum likelihood rather than restricted maximum likelihood that specified group identity as a random effect (random intercept and slope for each group). The game round variable was modified by −1 for each round to set the intercept to equal round one. We compared a series of models using a likelihood ratio test to settle on an optimal model for the main analyses (Supplementary Table S1). The optimal model controlled for differences in average levels of contribution in study 1 and study 2 but did not model any interactions between year of study and treatment effects.

Results

Payoff information led to a greater decline than social information

In support of the Confused Learners hypothesis, contributions declined faster when we showed individuals their payoff information instead of social information (Figure 2). We found that the rate of decline was significantly faster among those shown only their payoff compared to those shown just social information (LMM, controlling for study: estimated difference in rate of decline between Payoff and Social treatment = −1.7 percentage points faster per round in the Payoff treatment, 95% CI = [−0.60, −2.73], t 1,154.0 = −3.1, p = 0.002, Table 1). Specifically, when individuals saw their payoffs, contributions declined by −3.6 percentage points per round, decreasing from 49 to 18% over nine rounds, whereas among those shown only social information, contributions only declined at a rate of −1.9 percentage points per round, from 47 to 29% (Table 1). These rates are comparable with those found in a large comparative study of 237 games, which found an overall average rate of decline of 2.4 percentage points per round (although note that the rate depends on group size and marginal per capita return; Burton-Chellew & West, Reference Burton-Chellew and West2021).

Figure 2. Declining cooperation. Depending on randomly assigned treatment, groups of individuals were shown no information (grey, 21 groups); or social information (magenta, 53 groups); or payoff information (green, 40 groups); or both social and payoff information combined (blue, 40 groups). Dashed vertical lines represent 95% confidence intervals for each round. Payoff information led to a greater decline than social information. Supplementary Figure S1 also shows the same data depending on the level of social information shown (either all individual decisions or just the group average).

Table 1. Declining cooperation. Results from a linear mixed model fit by maximum likelihood on the percentage contribution by each group per round depending on information treatment (reference treatment: social)

Number of observations = 1386.

Number of independent groups = 154.

a t-Tests use Satterthwaite's method for estimating degrees of freedom.

Social information had a negligible effect

In contrast, the effect of showing individuals social information had a negligible effect upon the rate of decline, despite being central to the Conditional Cooperators hypothesis. Compared with those individuals who saw no information (No-info treatment), the rate of decline was not significantly faster when individuals saw only information on the decisions of their groupmates, strongly contradicting the Conditional Cooperators hypothesis (Social treatment) (LMM: estimated difference in rate of decline between No-info and Social treatments = −0.9 percentage points per round, 95% CI = [ 2.17, −0.45], t 1,154.0 = −1.3, p = 0.197, Figure 2, Table 1). These results were robust to an alternative analysis using a generalised linear mixed model to control for a potential lack of normality in the distribution of the response variable (Supplementary Results, Supplementary Table S2).

Repeating the above analyses separately for each study, which had different levels of social information, did not qualitatively change the conclusion that payoff information is what drives the decline (Supplementary Results, Supplementary Figure S1, Supplementary Tables S3 and S4). There was also no overall significant three-way interaction between study, game round and information treatment (LMM: Study × Game round × Information treatment, F 3,154 = 0.7, p = 0.575, Supplementary Table S1).

Redundant payoff information accelerated the decline

We confirmed our main result, that the rate of decline was faster when payoff information was available, with alternative analyses. First, we found that the addition of social information to payoff information did not lead to a significantly faster decline in the Combined information treatment compared to the Payoff treatment (LMM: estimated difference in rate of decline between Payoff and Combined treatments = 0.2 percentage points per round, 95% CI = [1.33, −0.95], t 1,154.0 = 0.3, p = 0.744). This was either because individuals were learning from their payoffs in both cases or because individuals without social information were using their payoffs to calculate their groupmates’ decisions before responding conditionally.

However, if individuals were using their payoffs to calculate their groupmates’ decisions, then the rate of decline should not have been faster when the technically redundant payoff information was added to social information, but this is what we found. Specifically, the rate of decline was significantly faster among those shown both social and payoff information (Combined information treatment), than among those shown just social information (Social treatment, LMM: estimated difference in rate of decline between Social and Combined treatments = −1.5 percentage points faster per round in the Combined treatment, 95% CI = [−0.41, −2.54], t 1,154.0 = −2.7, p = 0.007). This means that the addition of payoff information to social information led to a faster decline, despite this information being technically redundant, in support of the Confused Learners hypothesis and contradicting the Conditional Cooperators hypothesis.

Second, for descriptive purposes, we analysed all the groups separately (Table 2), and found that groups which saw payoff information were almost twice as likely to develop a negative correlation between their contributions and round of the game (percentage of groups with a significant negative Pearson correlation, Payoff treatment = 70%; Combined treatment = 65%; Social treatment = 36%, No-info treatment = 24%; total number of groups = 40/40/53/21 respectively). Among those groups in the Social treatment, 17% even finished at a higher level than their starting level (N = 9/53). In contrast, only 2% of groups in the Payoff treatment finished at a higher level (N = 1/40) (Supplementary Figures S2–S5 show the time profiles for each group). These comparisons would suggest that the Social treatment, even though it implicitly contained payoff information, was behaviourally more akin to the No-info treatment than to the two treatments showing explicit payoff information (Payoff and Combined treatments).

Table 2. Group summaries. Shown per treatment are the percentage of groups that finished the final round either lower or higher than their first round, and the percentage that had a significant Pearson correlation between contributions and all nine rounds of the game

Supplementary tasks support the Confused Learners hypothesis

The results from the dynamic behaviour in the repeated public-goods game found that payoff information, and not social information, drove the decline in contributions. Next, we analysed the results from a range of supplementary decision tasks to test for understanding and motivations more directly (Figure 1). Our results from these three tasks confirmed that most participants are self-interested but confused.

First, most participants failed to maximise their income in the asocial control

In support of the Confused Learners hypothesis, we found (1) that in the pre-game income maximisation test played with computers (IMT1), most participants (82%) contributed and thus behaved as if confused. This was irrational because they merely ‘burnt’ money for no social benefit (82% contributed more than 0 MU, N = 504/616; mean average contribution = 41%, or 8.2/20 MU, median = 8 MU); and (2) that the greatest improvement in the post-game test (IMT2) occurred among those who had experienced the treatments containing payoff information (Payoff and Combined treatments vs Social and No-info treatments, generalised linear quasibinomial model controlling for individual's contribution in the pre-game test: t 1,612 = −2.6, p = 0.009, Figure 3, Supplementary Figure S6, Supplementary Table S5).

Figure 3. Income maximisation tests. Most individuals failed to maximise their income even when grouped with computers. Violin plots show distribution of contributions. Box plots show the median contribution (horizontal bar) and interquartile range. The mean contributions are shown by the empty black circles. We tested all individuals twice, one before and once after the repeated public-goods game. The least improvement was among individuals shown only social information (pink). Number of individuals: No-info = 84; Social = 212; Payoff =160; Combined = 160.

If the decline in contributions in the repeated games was caused by Conditional Cooperators changing their beliefs about their groupmates, then behaviour towards computers programmed to play randomly should not have changed between the pre- and post-game income maximisation tests (IMT1 vs IMT2), but that is not what we found. Instead, we found that improvement in the income maximisation tests, where contributions decreased from 41% in IMT1 to 27–35% in IMT2 depending on treatment, tracked changes in the repeated game with humans (where contributions decreased from 47 to 19–34% depending on treatment), suggesting learning about the game's payoffs was responsible. The surprising result was that there was a significant improvement in the No-info condition (paired t-test, mean of difference = 1.75 MU (9%), 95% CI = [0.41 MU, 3.09 MU], t = 2.6, d.f. = 83, p = 0.011), perhaps because the participants used their time to reflect and re-read the instructions which were always available in a help box.

Second, most participants appeared confused about the nature of the payoffs

Participants appeared to believe the game's payoffs were interdependent, meaning one needs to know the behaviour of others to calculate the best income maximising response, which is not true in the linear public-goods game. Specifically, when we asked participants if the income-maximising strategy depended on the actions of their groupmates or not, only 25% of participants answered correctly (‘No’) (N = 153/616). In contrast, 72% of participant answered incorrectly (N = 444), indicating that they had false beliefs about the game's Payoff, with 53% responding that ‘Yes’, an income maximising individual should decide their contribution depending on what others contribute (N = 325) and another 19% responding ‘Sometimes’ (N = 119). The remaining participants, 3%, indicated that they did not know (‘Do not know’, N = 19) (Table 3). Incorrect individuals were significantly more likely to have cooperated with computers, consistent with confused conditional cooperation (Burton-Chellew et al., Reference Burton-Chellew, El Mouden and West2016; numbers that contributed more than 0 MU towards computers: in IMT1, N = 399 of 463 non-correct responses vs 105 of 153 correct responses; Fisher's exact test, p < 0.001; in IMT2, N = 358/463 vs 75/153; Fisher's exact test, p < 0.001).

Table 3. Confusion about payoffs. Summary of participant responses to: ‘In the basic decision situation, played for one round only, if a player wants to maximise his or her earnings, should they decide their contribution depending on what the other people in their group contribute?’a

a We reversed the order of responses for half of the participants (2 within each group).

b This is the correct response.

Finally, most participants reported selfish motives

Clearly one important aim of laboratory experiments is to gain knowledge about the specific motives behind social behaviours (Fehr & Fischbacher, Reference Fehr and Fischbacher2003). This is traditionally done with incentivised decisions, in the hope that the costs of decisions will inhibit non-genuine responses, and consequently reveal participants’ true motivations (Andreoni & Miller, Reference Andreoni and Miller2002). However, such an economical/mathematical approach risks being confounded by greater misunderstanding. Therefore, we simply asked participants to identify with one of four possible motivations, which should provide a lower bound estimate on the level of non-socially describable motivations (Methods, Table 4). The results contradicted the idea that most people are fair-minded conditional cooperators. Instead, most participants, 54%, responded that they were best described by the selfish motivation (N = 305/568 surveyed). A further 6% responded that they had a competitive motivation (N = 35), meaning that 60% of participants freely admitted to a socially undesirable response (N = 340). In contrast, only 28% responded that they were best described by the fair or inequity-averse motivation (N = 160) and only 12% that they were motivated by maximising the group's success even at personal cost (pro-group motivation) (N = 68) (Table 4).

Table 4. Selfish motivations. Summary of participant responses to: which of these descriptions best describes your motivation during the rounds with humans?a

a We reversed the order of responses for half of the participants (two within each group).

b ‘Making myself as much money as possible’.

c ‘Avoiding unequal outcomes so that I make neither more, nor less, than other people’.

d ‘Making the group as much money as possible even if it meant making myself less money’.

e ‘Making myself more money than other people.’

f Forty-eight participants, 12 in each treatment, were not presented with this question.

Discussion

We tested competing explanations for behaviour in public-goods games by controlling whether individuals could learn from their own payoffs or not (Confused Learner's hypothesis), and whether they could respond conditionally to their groupmates’ contributions or not (Conditional Cooperators hypothesis; Figure 1). We found substantial support for the Confused Learner's hypothesis across multiple results. Specifically, we found that: (1) payoff information generated the greatest decline in contributions, regardless of whether social information was present or not (Figure 2, Tables 1 and 2); (2) most individuals demonstrated confusion about the game's payoffs when playing with computerised groupmates (Figure 3) or when directly asked (Table 3); and (3) that most individuals admitted to being selfishly motivated to make themselves as much money as possible (Table 4). If more than 50% of participants freely admit to being selfishly motivated, it seems unjustified to conclude that humans are especially altruistic.

Overall, our results suggest that apparent altruism in public-goods games mostly arises from confused but self-interested individuals trying to learn how to play the game, and not from fair-minded altruistic individuals trying to help their group or equalise payoffs. If individuals are confused, we cannot be sure what game they think they are playing (Chou, McConnell, Nagel, & Plott, Reference Chou, McConnell, Nagel and Plott2009; Columbus, Munich, & Gerpott, Reference Columbus, Munich and Gerpott2020; Ferraro & Vossler, Reference Ferraro and Vossler2010; Plott & Zeiler, Reference Plott and Zeiler2005). For example, if individuals erroneously think the best action depends on what others do, like a ‘stag-hunt’ game or ‘threshold’ public-goods game, perhaps because these games are more common in daily life, then it makes sense for them to act conditionally on social information even if they are self-interested (Croson & Marks, Reference Croson and Marks2000; Henrich et al., Reference Henrich, Boyd, Bowles, Camerer, Fehr, Gintis and Tracer2005; Rondeau, Poe, & Schulze, Reference Rondeau, Poe and Schulze2005). This could explain why individuals appear to conditionally cooperate with computers, that cannot benefit, and why responses to our question on how to maximise incomes closely resembled the frequencies of different ‘social types’ typically reported in other experiments (~50–55% = ‘Yes’ = ‘Conditional Cooperator’; ~20–30% = ‘No’ = ‘Free Rider’, etc.; Burton-Chellew et al., Reference Burton-Chellew, El Mouden and West2016; Fischbacher & Gachter, Reference Fischbacher and Gachter2010; Fischbacher et al., Reference Fischbacher, Gachter and Fehr2001; Thoni & Volk, Reference Thoni and Volk2018). In this case social types would really be artefacts of variation in levels of understanding (Burton-Chellew et al., Reference Burton-Chellew, El Mouden and West2016) and repeated public-goods games with payoff information would be measuring rates of learning rather than social preferences (Burton-Chellew & West, Reference Burton-Chellew and West2021). This is consistent with the recent finding that the apparent preference for conditional cooperation decreases with experience, contradicting a key assumption of the Conditional Cooperators hypothesis (Andreozzi et al., Reference Andreozzi, Ploner and Saral2020).

Our study shows the value of using a range of treatments and methods. For example, our No-info treatment provided an informative baseline treatment when comparing rates of decline (Supplementary Tables S3 and S4). Our simple, direct, surveys of understanding and motivation complimented our use of the more traditional methods of incentivised decision making (‘behavioural economics’). The latter is less susceptible to socially desirable responses, but the former is simpler to understand and the results easier to interpret. Our income maximisation tests with computerised groupmates provided quick and simple diagnostic tests of ‘rational’ behaviour, a commonly assumed hypothesis in economic experiments aiming to measure social behaviours.

Future directions

It could be useful to expand our experiment. One could test which information is of more interest to participants by allowing them to choose (Burton-Chellew & D'Amico, Reference Burton-Chellew and D'Amico2021), or to use eye or mouse tracking software (Geran & Weixing, Reference Geran and Weixing2020; Jiang, Potters, & Funaki, Reference Jiang, Potters and Funaki2016; Lahey & Oxley, Reference Lahey and Oxley2016). One could vary how groups are formed, group size, or the length of the game, to investigate if the salience of social information depends on the probability of future social interactions (Burton-Chellew et al., Reference Burton-Chellew, El Mouden and West2017a; Fiala & Suetens, Reference Fiala and Suetens2017; Reuben & Suetens, Reference Reuben and Suetens2012; Trivers, Reference Trivers, Kappeler and van Schaik2006). One could ask participants about their expectations about their groupmates’ contributions to gain insights into their thought processes (Chaudhuri et al., Reference Chaudhuri, Paichayontvijit and Smith2017; Fischbacher & Gachter, Reference Fischbacher and Gachter2010). We did not do so because asking individuals to guess the contributions of their groupmates may distort behaviour by stimulating conditional cooperation in individuals that would otherwise not have thought about their groupmates. One could use a range of different instructions to test if our results generalise to different set-ups or different cultures (Li, Reference Li2017). Here we simply replicated common instructions and used a common participant pool, to enable comparisons with many prior key studies (Fehr & Gachter, Reference Fehr and Gachter2002; Fischbacher & Gachter, Reference Fischbacher and Gachter2010; Fischbacher et al., Reference Fischbacher, Gachter and Fehr2001). We replicated the standard result, but our expanded design and varied approach allowed us to identify that most participants were in fact self-interested, not altruistic, but needed to learn about the game's payoffs.

Acknowledgements

We are grateful to Tiago Fernandes for assistance with translations and conducting the first study, Luis Santos-Pinto and Laurent Lehmann for helpful discussion, and the audience of the Behavioural Economics and Experimental Research group at UNIL for feedback. Funding was provided by Professor Laurent Lehmann's independent research budget.

Author contributions

MNBC conceived and designed the study. Both authors worked together to collect the data. MNBC analysed the data and wrote the first draft. Both authors worked on the final draft.

Financial support

This research received no specific grant from any funding agency, commercial or not-for-profit sectors.

Conflicts of interest

Author MNBC and Author CG declare none

Data availability

The data and analyses that support the findings of this study are openly available in the Open Science Framework at https://osf.io/t4smj

Supplementary material

To view supplementary material for this article, please visit https://doi.org/10.1017/ehs.2022.45

References

Ambrus, A., & Pathak, P. A. (2011). Cooperation over finite horizons: A theory and experiments. Journal of Public Economics, 95(7–8), 500512. https://doi.org/10.1016/J.Jpubeco.2010.11.016CrossRefGoogle Scholar
Andreoni, J. (1995). Cooperation in public-goods experiments – Kindness or confusion. American Economic Review, 85(4), 891904.Google Scholar
Andreoni, J., & Croson, R. (2008). Partners versus strangers: Random rematching in public goods experiments. In Plott, C. R. & Smitt, V. L. (Eds.), Handbook of experimental economics results (Vol. 1, pp. 776). North-Holland.Google Scholar
Andreoni, J., & Miller, J. (2002). Giving according to garp: An experimental test of the consistency of preferences for altruism. Econometrica, 70(2), 737753.CrossRefGoogle Scholar
Andreozzi, L., Ploner, M., & Saral, A. S. (2020). The stability of conditional cooperation: Beliefs alone cannot explain the decline of cooperation in social dilemmas. Scientific Reports, 10(1), 13610. https://doi.org/10.1038/s41598-020-70681-zCrossRefGoogle ScholarPubMed
Angelovski, A., Di Cagno, D., Güth, W., Marazzi, F., & Panaccione, L. (2018). Behavioral spillovers in local public good provision: An experimental study. Journal of Economic Psychology, 67, 116134. https://doi.org/10.1016/j.joep.2018.05.003CrossRefGoogle Scholar
Barclay, P. (2012). Proximate and ultimate causes of punishment and strong reciprocity. Behavioral and Brain Sciences, 35(1). https://doi.org/10.1017/S0140525×11001154CrossRefGoogle ScholarPubMed
Barclay, P., & Daly, M. (2003). Humans should be individualistic and utility-maximizing, but not necessarily ‘rational’. Behavioral and Brain Sciences, 26(2), 154155.CrossRefGoogle Scholar
Bateson, M., Nettle, D., & Roberts, G. (2006). Cues of being watched enhance cooperation in a real-world setting. Biology Letters, 2(3), 412414. https://doi.org/10.1098/Rsbl.2006.0509Google ScholarPubMed
Bayer, R. C., Renner, E., & Sausgruber, R. (2013). Confusion and learning in the voluntary contributions game. Experimental Economics, 16(4), 478496. https://doi.org/10.1007/S10683-012-9348-2Google Scholar
Burnham, T. C., & Hare, B. (2007). Engineering human cooperation: Does involuntary neural activation increase public goods contributions? Human Nature – An Interdisciplinary Biosocial Perspective, 18(2), 88108. https://doi.org/10.1007/S12110-007-9012-2Google ScholarPubMed
Burton-Chellew, M. N., & D'Amico, V. (2021). A preference to learn from successful rather than common behaviours in human social dilemmas. Proceedings of the Royal Society B – Biological Sciences, 288(1965), artn 20211590. https://doi.org/10.1098/rspb.2021.1590CrossRefGoogle ScholarPubMed
Burton-Chellew, M. N., El Mouden, C., & West, S. A. (2016). Conditional cooperation and confusion in public-goods experiments. Proceedings of the National Academy of Sciences of the United States of America, 113(5), 12911296. https://doi.org/10.1073/pnas.1509740113CrossRefGoogle ScholarPubMed
Burton-Chellew, M. N., El Mouden, C., & West, S. A. (2017a). Evidence for strategic cooperation in humans. Proceedings of the Royal Society B – Biological Sciences, 284(1856), artn 20170689. https://doi.org/10.1098/rspb.2017.0689CrossRefGoogle Scholar
Burton-Chellew, M. N., El Mouden, C., & West, S. A. (2017b). Social learning and the demise of costly cooperation in humans. Proceedings of the Royal Society B – Biological Sciences, 284(1853), artn 20170067. https://doi.org/10.1098/rspb.2017.0067Google Scholar
Burton-Chellew, M. N., Nax, H. H., & West, S. A. (2015). Payoff based learning explains the decline in cooperation in human public goods games. Proceedings of the Royal Society B – Biological Sciences, 282, 20142678. http://dx.doi.org/10.1098/rspb.2014.2678CrossRefGoogle ScholarPubMed
Burton-Chellew, M. N., & West, S. A. (2013). Prosocial preferences do not explain human cooperation in public-goods games. Proceedings of the National Academy of Sciences of the United States of America, 110(1), 216221. https://doi.org/10.1073/pnas.1210960110CrossRefGoogle Scholar
Burton-Chellew, M. N., & West, S. A. (2021). Payoff-based learning best explains the rate of decline in cooperation across 237 public-goods games. Nature Human Behaviour, 5(10), 13301377. https://doi.org/10.1038/s41562-021-01107-7CrossRefGoogle ScholarPubMed
Camerer, C. F. (2013). Experimental, cultural, and neural evidence of deliberate prosociality. Trends in Cognitive Sciences, 17(3), 106108. https://doi.org/10.1016/J.Tics.2013.01.009CrossRefGoogle ScholarPubMed
Carpenter, J. P. (2004). When in Rome: conformity and the provision of public goods. The journal of Socio-Economics, 33(4), 395408.CrossRefGoogle Scholar
Chaudhuri, A. (2011). Sustaining cooperation in laboratory public goods experiments: A selective survey of the literature. Experimental Economics, 14(1), 4783. https://doi.org/10.1007/s10683-010-9257-1Google Scholar
Chaudhuri, A., Paichayontvijit, T., & Smith, A. (2017). Belief heterogeneity and contributions decay among conditional cooperators in public goods games. Journal of Economic Psychology, 58, 1530.Google Scholar
Chou, E., McConnell, M., Nagel, R., & Plott, C. R. (2009). The control of game form recognition in experiments: Understanding dominant strategy failures in a simple two person ‘guessing’ game. Experimental Economics, 12(2), 159179.CrossRefGoogle Scholar
Columbus, S., Munich, J., & Gerpott, F. H. (2020). Playing a different game: Situation perception mediates framing effects on cooperative behaviour. Journal of Experimental Social Psychology, 90, artn 104006. https://doi.org/10.1016/j.jesp.2020.104006CrossRefGoogle Scholar
Cooper, D. J., & Stockman, C. K. (2002). Fairness and learning: an experimental examination. Games and Economic Behavior, 41(1), 2645.CrossRefGoogle Scholar
Crawley, M. J. (2007). The R book. Wiley.Google Scholar
Croson, R., Fatas, E., & Neugebauer, T. (2005). Reciprocity, matching and conditional cooperation in two public goods games. Economics Letters, 87(1), 95101. https://doi.org/10.1016/J.Econlet.2004.10.007CrossRefGoogle Scholar
Croson, R., & Marks, M. (2000). Step returns in threshold public goods: A meta- and experimental analysis. Experimental Economics, 2(3), 239259.CrossRefGoogle Scholar
Di Cagno, D., Galliera, A., Güth, W., & Panaccione, L. (2016). A hybrid public good experiment eliciting multi-dimensional choice data. Journal of Economic Psychology, 56, 2038. https://doi.org/10.1016/j.joep.2016.05.001Google Scholar
Dijkstra, J., & van Assen, M. A. L. M. (2017). Explaining cooperation in the finitely repeated simultaneous and sequential prisoner's dilemma game under incomplete and complete information. Journal of Mathematical Sociology, 41(1), 125.CrossRefGoogle Scholar
Fehr, E., & Fischbacher, U. (2003). The nature of human altruism. Nature, 425(6960), 785791.CrossRefGoogle ScholarPubMed
Fehr, E., & Gachter, S. (2002). Altruistic punishment in humans. Nature, 415(6868), 137140.CrossRefGoogle ScholarPubMed
Fehr, E., & Schmidt, K. M. (1999). A theory of fairness, competition, and cooperation. Quarterly Journal of Economics, 114(3), 817868.CrossRefGoogle Scholar
Fehr, E., & Schurtenberger, I. (2018). Normative foundations of human cooperation. Nature Human Behaviour, 2, 458468. https://doi.org/10.1038/s41562-018-0385-5CrossRefGoogle ScholarPubMed
Ferraro, P. J., & Vossler, C. A. (2010). The source and significance of confusion in public goods experiments. Journal of Economic Analysis & Policy, 10(1). https://doi.org/10.2202/1935-1682.2006Google Scholar
Fiala, L., & Suetens, S. (2017). Transparency and cooperation in repeated dilemma games: A meta study. Experimental Economics, 20(4), 755771.CrossRefGoogle ScholarPubMed
Fischbacher, U., & Gachter, S. (2010). Social preferences, beliefs, and the dynamics of free riding in public goods experiments. American Economic Review, 100(1), 541556. https://doi.org/10.1257/aer.100.1.541CrossRefGoogle Scholar
Fischbacher, U., Gachter, S., & Fehr, E. (2001). Are people conditionally cooperative? Evidence from a public goods experiment. Economics Letters, 71(3), 397404.CrossRefGoogle Scholar
Fischbacher, U., Gächter, S., & Quercia, S. (2012). The behavioral validity of the strategy method in public good experiments. Journal of Economic Psychology, 33(4), 897913. https://doi.org/10.1016/j.joep.2012.04.002Google Scholar
Geran, T., & Weixing, W. (2020). A review of mouse-tracking applications in economic studies. Journal of Economics and Behavioral Studies, 11(6(J)). https://doi.org/10.22610/jebs.v11i6(J).3000Google Scholar
Greiner, B. (2015). Subject pool recruitment procedures: Organizing experiments with ORSEE. Journal of the Economic Science Association, 1(1), 114125. https://doi.org/10.1007/s40881-015-0004-4CrossRefGoogle Scholar
Gunnthorsdottir, A., Houser, D., & McCabe, K. (2007). Disposition, history and contributions in public goods experiments. Journal of Economic Behavior & Organization, 62(2), 304315. https://doi.org/10.1016/j.bebo.2005.03.008CrossRefGoogle Scholar
Hagen, E. H., & Hammerstein, P. (2006). Game theory and human evolution: A critique of some recent interpretations of experimental games. Theoretical Population Biology, 69(3), 339348. https://doi.org/10.1016/J.Tpb.2005.09.005CrossRefGoogle ScholarPubMed
Haley, K. J., & Fessler, D. M. T. (2005). Nobody's watching? Subtle cues affect generosity in an anonymous economic game. Evolution and Human Behavior, 26(3), 245256.CrossRefGoogle Scholar
Henrich, J., Boyd, R., Bowles, S., Camerer, C., Fehr, E., Gintis, H., … Tracer, D. (2005). ‘Economic man’ in cross-cultural perspective: behavioral experiments in 15 small-scale societies. Behavioral and Brain Sciences, 28(6), 795815; discussion 815–755. http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=16372952CrossRefGoogle Scholar
Henrich, J., & Muthukrishna, M. (2021). The origins and psychology of human cooperation. Annual Review of Psychology, Vol 72, 72, 207240.CrossRefGoogle ScholarPubMed
Houser, D., & Kurzban, R. (2002). Revisiting kindness and confusion in public goods experiments. American Economic Review, 92(4), 10621069.CrossRefGoogle Scholar
Jiang, T., Potters, J., & Funaki, Y. (2016). Eye-tracking social preferences. Journal of Behavioral Decision Making, 29(2–3), 157168. https://doi.org/10.1002/bdm.1899CrossRefGoogle Scholar
Jouxtel, J. (2019). Voluntary contributions of time: Time-based incentives in a linear public goods game. Journal of Economic Psychology, 75, 102139. https://doi.org/10.1016/j.joep.2019.01.002CrossRefGoogle Scholar
Kirchkamp, O. (2019). Importing z-tree data into R. Journal of Behavioral and Experimental Finance, 22, 12. https://doi.org/10.1016/j.jbef.2018.11.008CrossRefGoogle Scholar
Kocher, M. G., Cherry, T., Kroll, S., Netzer, R. J., & Sutter, M. (2008). Conditional cooperation on three continents. Economics Letters, 101(3), 175178. https://doi.org/10.1016/J.Econlet.2008.07.015CrossRefGoogle Scholar
Kocher, M. G., Martinsson, P., Persson, E., & Wang, X. (2016). Is there a hidden cost of imposing a minimum contribution level for public good contributions? Journal of Economic Psychology, 56, 7484. https://doi.org/10.1016/j.joep.2016.05.007CrossRefGoogle Scholar
Kocher, M. G., Martinsson, P., & Visser, M. (2008). Does stake size matter for cooperation and punishment? Economics Letters, 99(3), 508511. https://doi.org/10.1016/J.Econlet.2007.09.048CrossRefGoogle Scholar
Kreps, D. M., Milgrom, P., Roberts, J., & Wilson, R. (1982). Rational cooperation in the finitely repeated prisoners-dilemma. Journal of Economic Theory, 27(2), 245252.CrossRefGoogle Scholar
Krockow, E. M., Colman, A. M., & Pulford, B. D. (2016). Cooperation in repeated interactions: A systematic review of Centipede game experiments, 1992–2016. European Review of Social Psychology, 27(1), 231282.CrossRefGoogle Scholar
Krupp, D. B., Barclay, P., Daly, M., Kiyonari, T., Dingle, G., & Wilson, M. (2005). Let's add some psychology (and maybe even some evolution) to the mix. Behavioral and Brain Sciences, 28(6), 828829. https://doi.org/10.1017/S0140525×0535014xCrossRefGoogle Scholar
Kuemmerli, R., Burton-Chellew, M. N., Ross-Gillespie, A., & West, S. A. (2010). Resistance to extreme strategies, rather than prosocial preferences, can explain human cooperation in public goods games. Proceedings of the National Academy of Sciences of the United States of America, 107(22), 1012510130. https://doi.org/10.1073/pnas.1000829107CrossRefGoogle Scholar
Kurzban, R., & Houser, D. (2005). Experiments investigating cooperative types in humans: A complement to evolutionary theory and simulations. Proceedings of the National Academy of Sciences of the United States of America, 102(5), 18031807. https://doi.org/10.1073/pnas.0408759102CrossRefGoogle ScholarPubMed
Kuznetsova, A., Brockhoff, P. B., & Christensen, R. H. B. (2017). lmerTest Package: Tests in linear mixed effects models. Journal of Statistical Software, 82(13), 126. https://doi.org/10.18637/jss.v082.i13CrossRefGoogle Scholar
Lahey, J. N., & Oxley, D. (2016). The power of eye tracking in economics experiments. American Economic Review, 106(5), 309313. https://doi.org/10.1257/aer.p20161009Google Scholar
Larney, A., Rotella, A., & Barclay, P. (2019). Stake size effects in ultimatum game and dictator game offers: A meta-analysis. Organizational Behavior and Human Decision Processes, 151, 6172.CrossRefGoogle Scholar
Ledyard, J. (1995). Public goods: A survey of experimental research. In Kagel, J. H. & Roth, A. E. (Eds.), Handbook of experimental economics (pp. 253279). Princeton University Press.Google Scholar
Li, K. K. (2017). How does language affect decision-making in social interactions and decision biases? Journal of Economic Psychology, 61, 1528. https://doi.org/10.1016/j.joep.2017.03.003CrossRefGoogle Scholar
McAuliffe, W. H. B., Burton-Chellew, M. N., & McCullough, M. E. (2019). Cooperation and learning in unfamiliar situations. Current Directions in Psychological Science, 28(5), 436440. https://doi.org/10.1177/0963721419848673CrossRefGoogle Scholar
Nax, H. H., Burton-Chellew, M. N., West, S. A., & Young, H. P. (2016). Learning in a black box. Journal of Economic Behavior & Organization, 127, 115. http://dx.doi.org/10.1016/j.jebo.2016.04.006CrossRefGoogle Scholar
Neugebauer, T., Perote, J., Schmidt, U., & Loos, M. (2009). Selfish-biased conditional cooperation: On the decline of contributions in repeated public goods experiments. Journal of Economic Psychology, 30(1), 5260. https://doi.org/10.1016/J.Joep.2008.04.005CrossRefGoogle Scholar
Pedersen, E. J., Kurzban, R., & McCullough, M. E. (2013). Do humans really punish altruistically? A closer look. Proceedings of the Royal Society B – Biological Sciences, 280(1758). https://doi.org/10.1098/Rspb.2012.2723CrossRefGoogle ScholarPubMed
Plott, C. R., & Zeiler, K. (2005). The willingness to pay-willingness to accept gap, the ‘endowment effect,’ subject misconceptions, and experimental procedures for eliciting valuations. American Economic Review, 95(3), 530545.CrossRefGoogle Scholar
Raihani, N. J., & Bshary, R. (2015). Why humans might help strangers. Frontiers in Behavioral Neuroscience, 9, artn 39. https://doi.org/10.3389/Fnbeh.2015.00039Google ScholarPubMed
Reuben, E., & Suetens, S. (2012). Revisiting strategic versus non-strategic cooperation. Experimental Economics, 15(1), 2443.CrossRefGoogle Scholar
Rondeau, D., Poe, G. L., & Schulze, W. D. (2005). VCM or PPM? A comparison of the performance of two voluntary public goods mechanisms. Journal of Public Economics, 89(8), 15811592. https://doi.org/10.1016/j.jpubeco.2004.06.014Google Scholar
R Team (2020). RStudio: Integrated development environment for R. http://www.rstudio.comGoogle Scholar
Shapiro, D. A. (2009). The role of utility interdependence in public good experiments. International Journal of Game Theory, 38(1), 81106. https://doi.org/10.1007/s00182-008-0141-6CrossRefGoogle Scholar
Silva, A. S., & Mace, R. (2014). Cooperation and conflict: Field experiments in Northern Ireland. Proceedings of the Royal Society B – Biological Sciences, 281(1792). https://doi.org/10.1098/rspb.2014.1435CrossRefGoogle ScholarPubMed
Smith, V. L. (2010). Theory and experiment: What are the questions? Journal of Economic Behavior & Organization, 73(1), 315. https://doi.org/10.1016/j.jebo.2009.02.008CrossRefGoogle Scholar
Thielmann, I., Böhm, R., Ott, M., & Hilbig, B. E. (2021). Economic games: An introduction and guide for research. Collabra: Psychology, 7(1). https://doi.org/10.1525/collabra.19004Google Scholar
Thoni, C., & Volk, S. (2018). Conditional cooperation: Review and refinement. Economic Letters, 171, 3740.CrossRefGoogle Scholar
Trivers, R. L. (2006). Reciprocal altruism: 30 years later. In Kappeler, P. M. & van Schaik, C. P. (Eds.), Cooperation in primates and humans: Mechanisms and evolution. Springer.Google Scholar
Wickham, H. (2009). ggplot2: Elegant graphics for data analysis. Springer.CrossRefGoogle Scholar
Yamagishi, T., Li, Y., Matsumoto, Y., & Kiyonari, T. (2016). Moral bargain hunters purchase moral righteousness when it is cheap: Within-individual effect of stake size in economic games. Science Reports, 6, 27824. https://doi.org/10.1038/srep27824CrossRefGoogle ScholarPubMed
Zelmer, J. (2003). Linear public goods experiments: A meta-analysis. Experimental Economics, 6, 299310.CrossRefGoogle Scholar
Figure 0

Figure 1. Experimental design. We decoupled two typical forms of information, about either the contribution decisions of the focal player's groupmates (social information) or the focal individual's own earnings (yellow/top individual, payoff information). This made four general treatments: one treatment with neither form of information (‘No-info’, N = 21 groups); one with just social information (‘Social’, N = 53 groups; 20 groups observed all individual decisions whereas 33 groups observed the group average only); one with just payoff information (‘Payoff’, N = 40 groups); and one with both forms of information (‘Combined’, N = 40 groups; 20/20 groups with either full/limited social information). We never showed the earnings (payoff) of groupmates. Experimental timeline: participants first had one round of decision making with computerised groupmates in a public-goods game (income maximisation test, IMT1). They were then randomly assigned to one of four information treatments for nine rounds in a constant group of four real people. They then repeated the income maximisation test (IMT2) before two survey questions (Qs) regarding their motivation and their understanding of the payoffs.

Figure 1

Figure 2. Declining cooperation. Depending on randomly assigned treatment, groups of individuals were shown no information (grey, 21 groups); or social information (magenta, 53 groups); or payoff information (green, 40 groups); or both social and payoff information combined (blue, 40 groups). Dashed vertical lines represent 95% confidence intervals for each round. Payoff information led to a greater decline than social information. Supplementary Figure S1 also shows the same data depending on the level of social information shown (either all individual decisions or just the group average).

Figure 2

Table 1. Declining cooperation. Results from a linear mixed model fit by maximum likelihood on the percentage contribution by each group per round depending on information treatment (reference treatment: social)

Figure 3

Table 2. Group summaries. Shown per treatment are the percentage of groups that finished the final round either lower or higher than their first round, and the percentage that had a significant Pearson correlation between contributions and all nine rounds of the game

Figure 4

Figure 3. Income maximisation tests. Most individuals failed to maximise their income even when grouped with computers. Violin plots show distribution of contributions. Box plots show the median contribution (horizontal bar) and interquartile range. The mean contributions are shown by the empty black circles. We tested all individuals twice, one before and once after the repeated public-goods game. The least improvement was among individuals shown only social information (pink). Number of individuals: No-info = 84; Social = 212; Payoff =160; Combined = 160.

Figure 5

Table 3. Confusion about payoffs. Summary of participant responses to: ‘In the basic decision situation, played for one round only, if a player wants to maximise his or her earnings, should they decide their contribution depending on what the other people in their group contribute?’a

Figure 6

Table 4. Selfish motivations. Summary of participant responses to: which of these descriptions best describes your motivation during the rounds with humans?a

Supplementary material: PDF

Burton-Chellew and Guérin supplementary material

Burton-Chellew and Guérin supplementary material

Download Burton-Chellew and Guérin supplementary material(PDF)
PDF 647.4 KB