Main

What motivates people to work harder? Money is a logical starting place. Many modern jobs pay people for their time and effort. Some workers earn money contingent on how many cars they sell, how much data they enter or how many apples they pick. In such jobs, the implicit belief is that the more the employer pays, the more the employee works. The idea is an old one. Max Weber wrote that piece-rates are “one of the technical means which the modern employer uses to secure the greatest amount of work from his men”1.

Yet money is not the only source of motivation2. In their work, people also respond to psychological motivators, such as social norms, praise, self-actualization, desire to reciprocate and fear of social rejection3. A large body of research has shown that both monetary and non-monetary incentives can motivate people in a wide variety of settings4,5,6,7.

Understanding the (cost-) effectiveness of different incentives has important practical implications. Employers are constantly searching for ways to motivate their workers and improve their performance. Governments and nonprofits face a similar problem. They spend billions of dollars on public campaigns to encourage socially beneficial behaviours such as getting vaccinated or wearing seatbelts. The better we can motivate effortful behaviours, the better our interventions, workplaces and economies will be.

Researchers have been working on this problem recently. Large-scale experiments directly contrasted the effectiveness of monetary and non-monetary incentives in motivating effort8,9,10. While both monetary and non-monetary incentives were found to be motivating, money had a large advantage over psychological motivators.

This line of work is important because it uses strong methods. The studies clearly compared the effectiveness of different incentives, with tight controls and random assignment. However, the samples in both experiments consisted almost entirely of people from the United States and did not consider cultural differences. Can managers and policymakers in India, Indonesia or Nigeria apply scientific insights from these studies to issues facing their own countries?

Researchers have warned against indiscriminately exporting interventions that have demonstrated effectiveness in North America or Western Europe to other world regions11,12,13,14,15,16,17. Yet so far, most interventions have rarely been tested outside of Western, educated, industrialized, rich and democratic (WEIRD)18 cultures19,20,21. For example, a 2018 meta-review looked at the effectiveness of ‘green nudges’— non-monetary incentives to use less electricity and switch to more environmentally friendly energy sources21. All 40 studies were conducted in high-income counties22, and 35 were in the United States or Western Europe.

This extreme sampling bias is not unique to intervention studies. Studies asking general questions about human nature face a similar problem18. Over 90% of all papers published in top psychological journals use samples drawn exclusively from WEIRD cultures23,24. Despite this sampling bias, researchers then use these basic insights to develop interventions assuming that the findings are universal24,25.

Although rare, psychological interventions designed with a particular cultural context in mind have shown promise in addressing a range of pressing issues26,27,28,29,30,31. For example, a study in China tried to encourage workers in a textile factory to throw their waste in trash cans instead of on the production floor28. The researchers found that placing printed images of golden coins (symbols of luck and good fortune in China) on the floor reduced littering, while fining people real money failed. Similarly, an intervention in Niger tried to encourage women to start their own businesses by paying them money or by changing community norms around female entrepreneurship26. The norms worked better per dollar spent than the cash transfers.

These studies highlight that economic activities are embedded within systems of local norms, prevalent morality and social networks32,33,34,35. Numerous studies across disciplines have demonstrated how socio-cultural factors interact with formal institutions to shape economic judgements and behaviour36,37,38,39,40,41,42,43,44,45,46.

In work contexts, cultural embeddedness is crucial for understanding the nature of obligations that exist between employers and their employees47,48,49, particularly because employment is both an economic and a social relationship50. Labour participation in the market economies of WEIRD cultures is said to be enacted in an impersonal and transactional manner, in accordance with the norms of market exchange51,52,53,54. Employees must work a specified amount of time or produce a specified amount of output to receive a specified amount of money from the employer. The scope of the mutual obligations is explicitly outlined in a formal contract55,56.

Outside of WEIRD cultures, work usually operates under norms of reciprocity exchange, which are marked by ‘contractual incompleteness’51,57. In such ‘relational contracts’55,58, people usually do not rely on the strict meanings of sentences specified in contracts. Instead, they interpret obligations on the basis of existing relationships and norms59. Even when there is a formal agreement, people express some expectations tacitly and obligations often go beyond the terms specified in the contract58.

Why might certain exchange norms and psychological contracts become more widespread in a society? According to some scholars, their prevalence is best explained by institutional factors60. For instance, in non-WEIRD cultures, legal mechanisms such as courts tend to be less effective. This makes formal enforcement of contracts difficult and encourages people to prioritize established and trusted economic partners61,62,63.

Other researchers emphasize the role of cultural dimensions, such as the degree of risk aversion64, interpersonal trust65 and particularly collectivism66,67,68. Collectivism is more common in non-WEIRD cultures and refers to the tendency to value duties and responsibilities to one’s in-group41,69,70. In collectivist cultures, people might by default assume reciprocal obligations and see formal contracts as at best nuisances and at worst threats to social harmony and harbingers of conflict54,71.

Of course, not all employment relations are governed solely by norms of market exchange in WEIRD cultures and norms of reciprocity exchange in non-WEIRD cultures. Rather, people from all cultures can apply different mental models of exchange norms and psychological contracts on the basis of the features of a situation, including in economic transactions60,72,73,74,75,76. Nevertheless, cultural and institutional influences can make people more likely to interpret a situation as either transactional or relational by increasing the salience of a mental frame that is ‘consistent with what is likely or what ought to happen between employee and employer’56.

We argue that different exchange norms and psychological contracts influence the marginal advantage of monetary relative to non-monetary incentives, which we call the ‘money advantage.’ Workers from WEIRD cultures, operating under market exchange norms, should prioritize explicit quid pro quo arrangements—establishing what needs to be done to receive a tangible benefit56,77—and feel less obligated to reciprocate beyond what is contractually prescribed and formally enforced78,79.

In contrast, workers from non-WEIRD cultures, guided by reciprocity exchange norms, might not stop at the contractual minimum, even if they receive no extra money for doing so. They might intuit that this minimum does not represent their employer’s real expectation80. Or they might reciprocate above the minimum prescribed by the contract to establish trust and earn loyalty for future interactions77,81,82,83. Therefore, monetary incentives might have a smaller marginal effect on workers using reciprocity exchange norms compared with those using market exchange norms.

There is little direct evidence about the relative influence of monetary and non-monetary incentives across cultures. Surveys have found that people from WEIRD cultures report valuing money less as a source of job motivation84,85,86,87. For instance, responding to a questionnaire about the desirability of different job characteristics, new IT recruits from the United States placed less emphasis on receiving monetary bonuses for completing projects than did their Chinese counterparts85.

However, these studies did not actually test the effectiveness of money. Explicit attitudes predict economic behaviour in some domains but not in others, and this attitude–behaviour link often depends on whether personal attitudes correspond to prevalent cultural norms and shared belief systems88,89,90. Norms shared by a network, culture or institution can direct individual economic behaviour regardless of personal attitudes through mechanisms such as internalization or sanctioning91,92.

People in the United States and other WEIRD cultures overemphasize the importance of extrinsic rewards, such as money, in their lay theories of what motivates others93,94,95,96,97. Some evidence suggests that people from WEIRD cultures agree more that self-interest is the primary determinant of others’ behaviour67,98,99.

Institutional designs often reflect such lay beliefs about what motivates others85,91. For example, a survey asked Citibank managers around the world about their employees’ primary motivation99. Compared with their colleagues in Asia and Latin America, managers in the United States were more likely to think that money and other extrinsic motivators were their employees’ primary motive. Given this belief among managers, it is perhaps unsurprising that across industries, incentive pay is more prevalent in WEIRD cultures100,101,102.

Similarly, the few experiments that directly compared the effectiveness of incentives across countries seem to suggest that money advantage is higher in WEIRD cultures101,103,104,105. For instance, in an experiment where the researchers tried paying students in the United States and China to get better test scores, money improved performance in the United States, but not in China104. Another study hired data-entry workers in three developing countries using contracts with a fixed salary or an incentive contingent on performance101. Performance pay increased effort the least in the country highest on collectivism.

In this paper, we systematically compared the effectiveness of money in the United States and the United Kingdom (two WEIRD cultures) vs China, India, Mexico and South Africa (four non-WEIRD cultures). In our experiments, workers received a fixed salary, a fixed salary plus a psychological intervention or a fixed salary plus an additional monetary incentive. We tested whether money is equally effective in WEIRD and non-WEIRD cultures. We measured effectiveness as effort and as cost-effectiveness of different incentives. In our last study, we randomly assigned people in India to take the study in English or Hindi to see whether this shifted people’s use of exchange norms and the motivational advantage of money.

Study 1. Re-analysis of a large-scale experiment

In our first study, we re-analysed data from ref. 8, which gave Amazon Mechanical Turk (MTurk) workers a variety of incentives to perform a simple task. The task required participants to interchangeably click buttons ‘a’ and ‘b’ on their keyboards as many times as possible for 10 min. The researchers tested how much different incentives would increase the number of button presses.

All participants received a US$1.00 fee for participating in the experiment (all $ amounts henceforth are in US$). On top of that, some people received monetary incentives, such as an extra cent for every 100 clicks. The researchers compared money to many incentives from social psychology. For example, in one condition, the researchers gave participants a social reference point: “most participants perform well on the task, pressing over 2,000 times”. The results showed that (1) both pay-for-effort (monetary) and non-monetary incentives improved effort over the flat-fee condition, which included no additional incentives and (2) money outperformed non-monetary interventions by a large margin.

Fortunately for our purposes, many of the recruited MTurk workers were in India. This gave us a convenient opportunity to test cultural differences, namely, whether the money advantage was higher in the United States than in India. We excluded conditions with both a monetary and a psychological component (for example, loss aversion; more information is available in Methods, and Supplementary Table 9a,b). This left us with a sample of 5,526 participants from the United States (3,196 female, MeanAgeCategory = 2.62, s.d.AgeCategory = 1.29) and 768 participants in India (247 female, MeanAgeCategory = 2.41, s.d.AgeCategory = 1.02). Using this combined sample, we compared the effectiveness of pooled monetary incentives vs pooled psychological interventions.

Results

The difference in effort between monetary incentives and psychological interventions was larger in the United States than in India, as evidenced by an interaction between country and incentive type (unstandardized regression coefficient b = 170.56, t(6,287) = 2.92, P = 0.004, 95% confidence interval (CI) 55.91–285.21) (Fig. 1 and Supplementary Table 1). In these analyses, we controlled for age, gender and education. Main effects of country and incentive type for this and the following studies are reported in section ‘Main effects for reported regressions’ in Supplementary Information. Extended Data Fig. 1 shows effort in the individual incentive treatments by country.

Fig. 1: Study 1, pooled monetary vs pooled non-monetary conditions in the United States and India.
figure 1

Effects of pooled monetary incentives (green) and pooled non-monetary treatments (blue) in the United States (N = 5,526 participants on MTurk) and India (N = 768 participants on MTurk) from a previous study8. a, The central tendency and distribution of effort by incentive type and country. The black line within each box represents the median; the red dot shows the mean; upper and lower bounds show the third and first quartiles, respectively; whiskers represent 1.5× the interquartile range, with black dots showing observations outside of this range. The width of each violin corresponds to the frequency of observations at any given number of images rated on the y axis. The interaction between country and incentive type in a multiple linear regression model is statistically significant (b = 170.56, t(6,287) = 2.92, P = 0.004, 95% CI 55.91–285.21). b, The money advantage, that is, how much more effective money is than the pooled non-monetary treatments in each country. In b, error bars are bootstrapped 95% CIs for the mean relative difference in the number of button presses in the pooled monetary vs non-monetary treatments.

The results from Study 1 suggest that the money advantage differs across cultures. Specifically, monetary incentives outperformed psychological interventions more in the United States than in India.

However, differences in technology could explain the results. For example, MTurk participants from the two countries could have understood the instructions differently. Furthermore, more participants in India could have completed the study on their phones or had slower internet connections, which could have limited how much they could ramp up their button-pressing, particularly in response to the monetary incentives.

Studies 2a–c. The effectiveness of incentives in 5 cultures

In Study 2, we improved on the weaknesses of Study 1 by designing a new task. In our new task, participants saw images and assessed whether each image showed a building. Participants rated images one by one for a maximum of 10 min. In the monetary conditions, we paid participants more for rating more images. The bonus across Studies 2a–c ranged from 5 cents to 9 cents for every 10 images. In the non-monetary conditions, participants received the same pay regardless of how many images they rated. We asked comprehension questions to ensure that participants understood whether they would receive extra pay for rating more images.

We explicitly informed participants that they could quit the task without losing their base pay after rating 10 images. This gave a clear contractual minimal.

We also wanted to change the explicitly meaningless button-pressing task in Study 1 to be more like real-world work, which serves a purpose5,106. To this end, we told participants that we needed their help in training a machine-learning image-classification algorithm.

Since the non-monetary conditions had the same payout structure, in what follows, we first report the results comparing monetary vs pooled non-monetary conditions, including the fixed-salary (or flat-fee) condition. Then we compare the monetary incentive with each non-monetary treatment individually. In all regressions in Studies 2a–c, we control for gender, age and education. Regressions with additional controls, including Internet connection, are reported in ‘Additional controls’ in Supplementary Information.

Study 2a. The United Kingdom and China

To test whether the results are generalizable to other cultures, we ran Study 2a in two new countries, the United Kingdom and China. To this end, we recruited 1,067 participants on Prolific in the United Kingdom (544 female, 12 non-binary, MeanAge = 40.04, s.d.Age = 13.60) and 1,086 participants on social media through student networks at Hubei University in Wuhan, China (626 female, MeanAge = 23.31, s.d.Age = 5.78). We compared the monetary condition (5 cents per 10 images) with the social norm and flat-fee conditions.

Results

The difference in the effectiveness between monetary and non-monetary incentives was larger in the United Kingdom than in China, as evidenced by the interaction between incentive and culture (b = 38.40, t(2,145) = 9.65, P < 0.001, 95% CI 30.59–46.20) (Fig. 2a and Supplementary Table 2). In China, extra pay increased effort by 19.9% over the two non-monetary conditions. In the United Kingdom, money increased effort by 109.5% (Fig. 2b).

Fig. 2: Study 2a, monetary vs pooled non-monetary conditions in the United Kingdom and China.
figure 2

Effects of a monetary incentive (green) and pooled non-monetary treatments (flat fee and social norm; blue) in the United Kingdom (N = 1,067 participants recruited on Prolific) and China (N = 1,086 participants recruited on social media). a, The central tendency and distribution of effort by incentive type and country. The black line within each box represents the median and the red dot shows the mean; upper and lower bounds show the third and first quartiles, respectively; whiskers represent 1.5× the interquartile range, with black dots showing observations outside of this range. The width of each violin corresponds to the frequency of observations at any given number of images rated on the y axis. The interaction between country and incentive type in a multiple linear regression model is statistically significant (b = 38.40, t(2,145) = 9.65, P < 0.001, 95% CI 30.59–46.20). b, The money advantage, that is, how much more effective money is than the pooled non-monetary treatments in each country. In b, error bars are bootstrapped 95% CIs for the mean relative difference in the number of images rated in the monetary vs pooled non-monetary conditions.

The money advantage over each of the two non-monetary conditions (flat fee and norm) was significantly larger in the United Kingdom than in China (b = −44.88, t(2,143) = −9.82, P < 0.001, 95% CI −53.84 to −35.91 for the flat-fee condition; and b = −31.55, t(2,143) = −6.96, P < 0.001, 95% CI −40.44 to −22.67 for the norm condition) (Extended Data Fig. 2a and Supplementary Table 7).

In China, the monetary condition was significantly less cost-effective than the norm condition (two-sided Welch’s t(681.11) = −2.54, P = 0.011, PBonf = 0.045, Meandifference = −4.28, Cohen’s d = −0.19, 95% CI −7.58 to −0.97) and did not significantly differ in cost-effectiveness from the flat-fee condition (two-sided Welch’s t(658.54) = 0.45, P = 0.653, PBonf = 1.000, Meandifference = 0.73, d = 0.03, 95% CI −2.45 to 3.90) (Extended Data Fig. 2c). In the United Kingdom, the monetary condition was significantly more cost-effective than each of the two non-monetary conditions: monetary and norm, two-sided Welch’s t(574.32) = 2.70, P = 0.007, PBonf = 0.029, Meandifference = 6.84, d = 0.20, 95% CI 1.86 to 11.82; monetary and flat fee, two-sided Welch’s t(692.87) = 10.88, P < 0.001, PBonf < 0.001, Meandifference = 22.11, d = 0.81, 95% CI 18.12–26.10.

Study 2b. Adjusting the pay to reflect the real value of money in Mexico and the United States

We followed up on Study 2a to make up for two shortcomings of the design. For one, Study 2a compared people on two different platforms (Prolific in the United Kingdom and social media in China). Because Prolific is a work platform, it could explain the difference in the money advantage without reference to culture. Second, we paid participants in the two countries the same amount of money. However, a dollar goes further in China than in the United Kingdom. To compensate for these shortcomings, in Studies 2b and 2c, we recruited workers only on crowdsourcing platforms and adjusted for local purchasing power.

In Study 2b, we extended the sample of non-WEIRD cultures to Mexico and again compared the monetary condition with the social norm and flat-fee conditions. We recruited a Prolific sample in Mexico (N = 1,053; 536 female, 26 non-binary, MeanAge = 24.51, s.d.Age = 5.43) and two Prolific samples in the United States. In one of these US samples (N = 1,098; 652 female, 24 non-binary, MeanAge = 37.28, s.d.Age = 13.77), we paid participants the same amount of money as in Mexico. In the other US sample (N = 1,122; 674 female, 25 non-binary, MeanAge = 36.13, s.d.Age = 13.31), we increased the pay so that it became subjectively equivalent to the amount received by the participants in Mexico. To this end, we used a common economics procedure to establish subjective pay equivalence107 (see Methods). Participants in the increased-pay sample in the United States received $1.56 as base pay for their participation (compared with $1.30 in the other two samples) and a piece-rate of 6 cents (compared with 5 cents).

Results

The money advantage was larger in the United States than in Mexico, regardless of whether we compared Mexico to the US sample with the same nominal pay (b = 13.93, t(2,143) = 3.42, P < 0.001, 95% CI 5.94–21.92), or that with the same subjective pay (b = 19.28, t(2,167) = 4.71, P < 0.001, 95% CI 11.26–27.30) (Fig. 3a and Supplementary Table 3). Stated differently, the effectiveness of monetary and non-monetary motivators was closer to one another in Mexico than in the United States. In the United States, money increased effort by 142.9% over the two pooled non-monetary conditions in the sample with the same nominal pay and by 165.6% in the sample with the same subjective pay (Fig. 3b). In Mexico, the difference was 41.6%.

Fig. 3: Study 2b, monetary vs pooled non-monetary conditions in the United States and Mexico.
figure 3

Effects of a monetary incentive (green) and pooled non-monetary treatments (flat fee and social norm; blue) in Mexico (N = 1,053 participants recruited on Prolific) and two samples in the United States: one with the same nominal pay (N = 1,098 on Prolific) as in Mexico and one with the same subjective107 pay (N = 1,122 participants recruited on Prolific) as in Mexico. a, The central tendency and distribution of effort by incentive type and country. The black line within each box represents the median and the red dot shows the mean; upper and lower bounds show the third and first quartiles, respectively; whiskers represent 1.5× the interquartile range, with black dots showing observations outside of this range. The width of each violin corresponds to the frequency of observations at any given number of images rated on the y axis. The interaction between country and incentive type in a multiple linear regression model is statistically significant (b = 13.93, t(2,143) = 3.42, P < 0.001, 95% CI 5.94–21.92 for the comparison between Mexico and the US sample with the same nominal pay; b = 19.28, t(2,167) = 4.71, P < 0.001, 95% CI 11.26–27.30 for the comparison between Mexico and the US sample with the same subjective pay). b, The money advantage, that is, how much more effective money is than the pooled non-monetary treatments in each sample. In b, error bars are bootstrapped 95% CIs for the mean relative difference in the number of images rated in the monetary vs pooled non-monetary conditions.

Next, we analysed individual conditions. The interaction between culture and incentive was statistically significant. The money advantage was larger in the United States than in Mexico for the flat-fee condition (b = −10.32, t(2,141) = −2.24, P = 0.025, 95% CI −19.37 to −1.27) and for the norm condition (b = −17.69, t(2,141) = −3.81, P < 0.001, 95% CI −26.80 to −8.58) (Extended Data Fig. 3a and Supplementary Table 8). We found the same pattern in our comparison of Mexico and the US sample with the same subjective pay: money was relatively more effective in the United States than in Mexico when compared with the flat-fee condition (b = −13.41, t(2,165) = −2.88, P = 0.004, 95% CI −22.55 to −4.27) and the norm condition (b = −25.55, t(2,165) = −5.47, P < 0.001, 95% CI −34.72 to −16.38).

Finally, we analysed the amount of effort per dollar spent for each of the individual incentive conditions. In Mexico, the monetary condition was significantly less cost-effective than the norm condition (two-sided Welch’s t(527.85) = −3.90, P < 0.001, PBonf < 0.001, Meandifference = −9.03, d = −0.30, 95% CI −13.58 to −4.49) but more cost-effective than the flat-fee condition (two-sided Welch’s t(605.56) = 4.00, P < 0.001, PBonf < 0.001, Meandifference = 8.07, d = 0.30, 95% CI 4.11–12.03) (Extended Data Fig. 3c).

In both samples in the United States, money was more cost-effective than the norm condition (same nominal pay: two-sided Welch’s t(671.05) = 5.09, P < 0.001, PBonf < 0.001, Meandifference = 10.97, d = 0.38, 95% CI 6.74–15.20; increased pay: two-sided Welch’s t(721.26) = 7.76, P < 0.001, PBonf < 0.001, Meandifference = 13.13, d = 0.57, 95% CI 9.81–16.46) and the flat-fee condition (same nominal pay: two-sided Welch’s t(665.00) = 13.70, P < 0.001, PBonf < 0.001, Meandifference = 21.85, d = 1.01, 95% CI 18.72–24.99; increased pay: two-sided Welch’s t(742.17) = 12.96, P < 0.001, PBonf < 0.001, Meandifference = 18.97, d = 0.95, 95% CI 16.10–21.85). Thus, consistent with the findings in Study 2a, we found that a monetary incentive is less cost-effective than a non-monetary intervention (social norm) in a non-WEIRD culture (Mexico), but not in a WEIRD culture (the United States).

Study 2c. Pre-registered replication in the United States and South Africa

In Study 2c, we made three improvements over Study 2b. First, we pre-registered the hypothesis and analysis. Second, we extended the sample to another non-WEIRD culture. We recruited Prolific workers in the United States (two samples, N = 662 each; 318 female, 20 non-binary, MeanAge = 36.91, s.d.Age = 12.62 in the sample with the same nominal pay; 323 female, 17 non-binary, MeanAge = 36.27, s.d.Age = 12.52 in the sample with the same subjective pay) and South Africa (N = 649; 316 female, 6 non-binary, MeanAge = 28.29, s.d.Age = 7.45). South Africa is an interesting test case because, although it scores lower on individualism than the United States, it is more individualistic than China or India108. Individualism is often contrasted with collectivism and refers to the tendency to prioritize individual goals and achievement41,69,70. South Africa’s English-speaking population tends to be particularly individualist109,110. All participants in South Africa completed the study in English and reported speaking it fluently.

Third, we tested a wider range of psychological interventions. We tried to ‘stress test’ our findings by choosing non-monetary interventions we suspected might be particularly effective in WEIRD cultures. In one condition, we told participants they were competing with other participants, since some researchers see competition as a facet of individualism111,112. In another condition, we incentivized people with donations to organized charities, which receive more contributions in WEIRD compared with non-WEIRD cultures (even controlling for income)113,114. Given these stress tests, we anticipated smaller effect sizes than in Studies 2a and 2b. We compared these two conditions (competition and charity) with the monetary condition.

As in Study 2b, we ran some US participants with the same pay as offered to the participants in South Africa. Other US participants worked for pay adjusted to reflect local differences in purchasing power, using a procedure analogous to the one in Study 2b107. In the US sample with the same subjective pay, we increased the base pay from $1.30 to $2.25 in all conditions, as well as the piece-rate from 5 to 9 cents per 10 images in the monetary condition.

We pre-registered pooled analyses because we did not have sufficient power to compare individual conditions. We report the comparisons between the monetary and each of the non-monetary conditions in ‘Mean effort and cost-effectiveness in individual conditions in Study 2c’ in Supplementary Information, but readers should be aware that the statistical power for these comparisons is lower.

Results

The difference in the effectiveness between monetary and non-monetary conditions was larger in the United States than in South Africa, both when we compared the sample in South Africa to the US sample with the same nominal pay (b = 7.98, t(1,303) = 1.97, P = 0.049, 95% CI 0.03–15.92) and to the US sample with the same subjective pay (b = 15.40, t(1,303) = 3.74, P < 0.001, 95% CI 7.33–23.48) (Fig. 4a and Supplementary Table 4). Additional money increased effort by 66.7% in South Africa, but by 126.6% in the US sample with the same nominal pay and by 155.2% in the US sample with increased pay (Fig. 4b).

Fig. 4: Study 2c, monetary vs pooled non-monetary conditions in the United States and South Africa.
figure 4

Effects of a monetary incentive (green) and pooled non-monetary treatments (competition and charity; blue) in South Africa (N = 649 participants on Prolific) and two samples in the United States: one with the same nominal pay (N = 662 on Prolific) as in South Africa and one with the same subjective107 pay (N = 662 participants on Prolific) as in South Africa. a, The central tendency and distribution of effort by incentive type and country. The black line within each box represents the median and the red dot shows the mean; upper and lower bounds show the third and first quartiles, respectively; whiskers represent 1.5× the interquartile range, with black dots showing observations outside of this range. The width of each violin corresponds to the frequency of observations at any given number of images rated on the y axis. The interaction between country and incentive type in a multiple linear regression model is statistically significant (b = 7.98, t(1,303) = 1.97, P = 0.049, 95% CI 0.03–15.92 for the comparison between South Africa and the US sample with the same nominal pay; b = 15.40, t(1,303) = 3.74, P < 0.001, 95% CI 7.33–23.48 for the comparison between South Africa and the US sample with the same subjective pay). b, The money advantage, that is, how much more effective money is than the pooled non-monetary treatments in each sample. In b, error bars are bootstrapped 95% CIs for the mean relative difference in the number of images rated in the monetary vs pooled non-monetary conditions.

Studies 3a,b. Comparing psychological incentives and minimal pay

Studies 2a–c extended the findings from Study 1 to four other cultures and to a more meaningful task. The money advantage was higher in WEIRD than in non-WEIRD cultures both when the contractual minimum was left ambiguous (Study 1) and when it was made explicit (Studies 2a–c). Cultural differences persisted after we adjusted the pay to reflect local purchasing power (Studies 2b and 2c).

In Study 3a, we pushed the boundaries of just how little extra money beyond the base pay it would take to spur extra effort. The bonuses we paid in Studies 2a–c were not large, but they were above average for crowdsourcing sites. Surveys have found that crowdsourcing workers average $3.31 per hour, and the pay is even lower for non-Western workers115,116. Our studies maxed out at 10 min, which would work out to 55 cents at these sites’ average wage. In the monetary conditions in Studies 2a–c, participants could earn more than that from the piece-rates alone (and the base pay made the earnings even better). Since the piece-rates were above average, it may not be surprising that monetary incentives would be more motivating than psychological interventions.

This made us wonder whether it was the actual pragmatic value of the money that motivated people or whether money holds a symbolic value beyond the actual usefulness of the amount. To test this, we changed the piece-rates in Study 3a to negligible amounts.

Study 3a. Lowering the piece-rate in India and the United States

In Study 3a, we lowered the bonus to a penny for every 20 images completed. We compared this minimal pay condition to a social norm intervention in samples in India (N = 352 recruited on MTurk; 83 female, MeanAge = 35.79, s.d.Age = 8.73) and the United States (N = 382 recruited on Prolific; 197 female, 14 non-binary, MeanAge = 37.76, s.d.Age = 13.99). We pre-registered the hypothesis, methods and analysis. As in the previous studies, we analysed the data using regressions controlling for age, gender and education.

Results

The money advantage was larger in the United States than in India, as evidenced by a statistically significant interaction between cultures (United States, India) and incentives (norm, monetary) (b = 13.77, t(726) = 2.31, P = 0.021, 95% CI 2.09–25.45) (Fig. 5a and Supplementary Table 5). A tiny monetary incentive, when compared to the social norm condition, increased effort by just 1.6% in India but by 48.1% in the United States (Fig. 5b).

Fig. 5: Study 3a, norm vs minimal pay condition in India and the United States.
figure 5

Effects of a minimal monetary incentive (of 1 cent per 20 image ratings; green) and a social norm condition (blue) in the United States (N = 382 participants recruited on Prolific) and India (N = 352 participants recruited on MTurk). a, The central tendency and distribution of effort by incentive and country. The black line within each box represents the median and the red dot shows the mean; upper and lower bounds show the third and first quartiles, respectively; whiskers represent 1.5× the interquartile range, with black dots showing observations outside of this range. The width of each violin corresponds to the frequency of observations at any given number of images rated on the y axis. The interaction between country and incentive in a multiple linear regression model is statistically significant (b = 13.77, t(726) = 2.31, P = 0.021, 95% CI 2.09–25.45). b. The money advantage, that is, how much more effective the minimal incentive is compared to the social norm condition in each country. c, The central tendency and distribution of cost-effectiveness (effort per dollar spent) of each incentive by country. Graph elements are analogous to those in a, with the width of each violin corresponding to the frequency of observations at any given level of cost-effectiveness (effort per dollar spent) rated on the y axis. Minimal monetary incentive is significantly more cost-effective than the social norm condition in the United States (two-sided Welch’s t(365.20) = 3.14, P = 0.002, PBonf = 0.004, Meandifference = 13.00, d = 0.32, 95% CI 4.86–21.15) but not in India (two-sided Welch’s t(345.72) = −0.27, P = 0.785, PBonf = 1.000, Meandifference = −1.06, d = −0.03, 95% CI −8.73 to 6.60). In b, error bars are bootstrapped 95% CIs for the mean relative difference in the number of images rated in the minimal-monetary-incentive vs social norm condition.

In India, paying 1 additional cent per 20 images rated was not significantly more cost-effective than emphasizing a descriptive norm (two-sided Welch’s t(345.72) = −0.27, P = 0.785, PBonf = 1.000, Meandifference = −1.06, d = −0.03, 95% CI −8.73 to 6.60) (Fig. 5c). In the United States, minimal pay was significantly more cost-effective than the social norm condition (two-sided Welch’s t(365.20) = 3.14, P = 0.002, PBonf = 0.004, Meandifference = 13.00, d = 0.32, 95% CI 4.86–21.15).

Therefore, Study 3a showed that relative to a psychological intervention, even a minimal monetary incentive was more motivating in a WEIRD compared with a non-WEIRD culture. In the pre-registered Supplementary Study 3b, we further found that gamification117,118,119—triggering the sense of competitive games or accumulating non-financial rewards such as points in a video game—could not explain the effectiveness of minimal pay in the United States (Extended Data Fig. 4).

Study 4. Randomly assigning cultural frames through language

In Study 4, we conducted a lab-in-the-field experiment to overcome the problem of causality. In interpreting the results of our previous studies, we attributed the differences to culture, but we did not randomly assign culture. To get at this issue, we randomly assigned bilingual people in India (N = 2,065; 286 female, 2 non-binary, MeanAge = 24.87, s.d.Age = 4.90) to complete the study in Hindi or in English.

Previous studies have found that having bilingual people switch languages activates a discrete set of social and moral norms in the culture associated with each language120,121,122,123,124. For example, one study surveyed managers in Hong Kong in Chinese or English122. When surveyed in English, the surveyed managers endorsed values more similar to those espoused by American managers. For instance, they rated conformity and tradition as less important, while they rated achievement and hedonism as more important.

We recruited participants through a post on the Facebook group QMaths. This group has over 280,000 members interested in preparing for competitive exams for jobs in sectors ranging from banking to railways.

We compared how many images participants completed when given a monetary reward vs a psychological intervention (social norm). In addition, we asked questions to explore the mechanism behind the cultural differences. The seven exploratory variables asked about participants’ motivation for and perception of the task, such as whether they enjoyed completing the task and whether they completed it only for money.

Results

The money advantage was larger when participants took the study in English than in Hindi, as evidenced by an interaction between language and incentive type (b = 10.69, t(2,061) = 3.31, P = 0.001, 95% CI 4.35–17.02) (Fig. 6a and Supplementary Table 6). The difference in effort between the monetary and the norm condition was 51.8% in English and 27.0% in Hindi (Fig. 6b).

Fig. 6: Study 4, norm vs monetary condition in India, by language prime (English or Hindi).
figure 6

Effects of a monetary incentive (green) and a social norm treatment (blue) in India (N = 2,065 participants recruited on Facebook), by assigned language (English or Hindi). a, The central tendency and distribution of effort by language and incentive conditions. The black line within each box represents the median and the red dot shows the mean; upper and lower bounds show the third and first quartiles, respectively; whiskers represent 1.5× the interquartile range, with black dots showing observations outside of this range. The width of each violin corresponds to the frequency of observations at any given number of images rated on the y axis. The interaction between language and incentive in a multiple linear regression model is statistically significant (b = 10.69, t(2,061) = 3.31, P = 0.001, 95% CI 4.35–17.02). b, The money advantage, that is, how much more effective the monetary condition is compared to the social norm condition. c, The central tendency and distribution of the cost-effectiveness (effort per dollar spent) by language and incentive. Graph elements are analogous to those in a, with the width of each violin corresponding to the frequency of observations at any given level of cost-effectiveness (effort per dollar spent) rated on the y axis. The monetary incentive is more cost-effective than the social norm condition in English (two-sided Welch’s t(919.21) = 4.37, P < 0.001, PBonf < 0.001, Meandifference = 4.34, d = 0.27, 95% CI 2.39–6.28), but the two incentives do not significantly differ in their cost-effectiveness in Hindi (two-sided Welch’s t(915.62) = 0.30, P = 0.761, PBonf = 1.000, Meandifference = 0.30, d = 0.02, 95% CI −1.64 to 2.24). In b, error bars are bootstrapped 95% CIs for the mean relative difference in the number of images rated in the norm vs monetary condition.

In our analyses on cost-effectiveness, we found that the monetary incentive was more cost-effective than the social norm in English (two-sided Welch’s t(919.21) = 4.37, P < 0.001, PBonf < 0.001, Meandifference = 4.34, d = 0.27, 95% CI 2.39–6.28) (Fig. 6c). By contrast, in Hindi, the two conditions did not significantly differ in their cost-effectiveness (two-sided Welch’s t(915.62) = 0.30, P = 0.761, PBonf = 1.000, Meandifference = 0.30, d = 0.02, 95% CI −1.64 to 2.24).

Next, we analysed whether the cultural prime influenced motivations. Language produced a statistically significant difference in one of seven items (with Bonferroni corrections). Participants were more likely to say that they completed the task only for money in English than in Hindi (b = 0.40, t(2,017) = 3.49, P = 0.001, PBonf = 0.004, 95% CI 0.18–0.63, 99.3% CI 0.09–0.71), which suggests that English activated a more transactional mental frame across incentive treatments. We report further details on the motivation questions in Supplementary Tables 14, 15a,b and 16, and Supplementary Fig. 3.

Thus, in Study 4 we found that English increased the money advantage. However, we recognize that using language to switch cultural frames is not reducible to the simple claim that the Hindi condition ‘randomly assigns’ Indian culture, while the English condition ‘randomly assigns’ United States (or United Kingdom, or WEIRD) culture. The differential set of associations activated by the two languages is likely to be considerably more complex120,125.

For instance, English might activate concepts related to work, professional life or achievement. This may be particularly true given that English is the primary language used for higher education in India and is often associated with status and prestige126,127. Similarly, people who completed the study in English may have implicitly compared themselves to native English speakers, which could have led to more upward social comparisons. These are empirically testable conjectures worthy of unpacking in future studies. However, it is worth remembering that professionalism, drive for prestige and social comparison are all part of culture too.

Discussion

How effective are different incentives in motivating effort? The data here suggest that the answer depends on culture. The motivating power of money over psychological interventions was stronger in the United States and the United Kingdom than in China, India, Mexico and South Africa. In China and Mexico, a social norm intervention was more cost-effective than a monetary incentive. In our last study, we found that randomly assigning people in India to take the study in English (compared with in Hindi) increased the money advantage. Participants who took the study in English were also more likely to report that they completed the task ‘only for money’, which suggests that they thought of their work more in accordance with the transactional norms of market exchange.

We interpret our results as reflecting cultural differences in exchange norms and psychological contracts. However, one can argue that our findings show that people in WEIRD cultures are closer to the model of Homo economicus, particularly in its narrow reading that includes the ‘selfishness axiom’128, or the assumption that ‘individuals seek to maximize their own material gains…and expect others to do the same’40. If maximizing material gains is the sole goal, our participants should work only when it pays. Many Americans did just that, living up to Benjamin Franklin’s dictum that “time is money.”

The most marked difference was observed in Study 2b. There, over half of the American participants quit the task as soon as they could do so without losing the base pay when there was no monetary incentive to continue. In Mexico, over 90% of participants continued even when they knew that extra effort would not result in more pay (see ‘Probability of quitting the task at the first opportunity in Studies 2a–4’ in Supplementary Information).

We do not see the Homo economicus interpretation as contradictory to our focus on transactional contracts and exchange norms. Both make the same prediction regarding effort in the non-monetary conditions67,129. However, Homo economicus has become somewhat of a strawman130, particularly since most economists agree that rational agents can derive utility not solely from material gains, but also from moral and reputational considerations38,131. Both working above the contractual minimum and quitting in the absence of monetary incentives can be seen as rational utility-maximizing actions. Therefore, we neither interpret the behaviour of non-WEIRD participants in our studies as ‘irrational’, nor do we evaluate the behaviour of WEIRD participants as demonstrating an obsession with money. Instead, we theorize that transactional mentality is a continuum, whereby people in some cultures are more likely to perceive work as governed by explicit contracts and inherently involving monetary remuneration132,133.

Our experiments have limitations because they stripped out several elements of work in the real world. We can only theorize whether the same cultural differences in the money advantage would arise in different tasks. For one, our tasks gave all workers a base salary, meaning that no one worked solely for commission. In addition, everyone worked alone, without the potential influence of coworkers. We also randomly assigned incentives to workers, but some workers in the real world can choose between jobs on the basis of how much those jobs pay and how much fulfilment they give134. Moreover, while our experiments focused on the effectiveness of incentives in terms of motivating effort, it is not the sole metric relevant to evaluating their overall value135. For instance, prioritizing performance pay can lead to negative consequences such as higher levels of stress136 and lower quality of non-work relationships137. Therefore, in future research, it is crucial to assess various incentives on other factors, such as people’s sense of overall well-being.

The findings of our study also raise thorny questions about pay. People could use our data to justify paying people in non-WEIRD cultures less. The extent to which employers across different countries already take advantage of these tendencies deserves further research. While performance pay has been found to be more frequent in WEIRD countries, employees in non-WEIRD countries might receive more of other types of compensation, such flexible benefits or group-based performance rewards100,138. This having been said, we do not believe that our findings justify exploitation: non-monetary incentives cannot and should not be seen as a substitute for paying a fair wage. In our design, everyone received a base pay (a ‘salary’), which was probably the first-order motivator. On MTurk and Prolific, participants would probably not start tasks in the first place if they were not paid84. Instead, we interpret our results as suggesting that, given an adequate salary, workers from non-WEIRD cultures are less incentivized by further pay-for-effort incentives as compared with free psychological motivators. We discuss this further in the ‘Additional discussion’ section in Supplementary Information.

Furthermore, although we use the terms ‘WEIRD’ and ‘non-WEIRD’ throughout the paper, we agree with the researchers who coined this term that the underlying cultural differences are not binary18,139,140,141. One way to move from the WEIRD–non-WEIRD dichotomy into a more nuanced continuum is the calculation of cultural distance. Researchers used responses to the World Values Survey to calculate different cultures’ ‘distance’ from the United States140. Interestingly, these cultural distance scores follow the strength of the money advantage in our studies (Extended Data Fig. 5). Future studies with more cultures and identical conditions can more robustly test the predictive power of cultural distance.

The findings in this paper are a reminder that we should avoid extrapolating the conclusions from studies based on the 12% of the world population that lives in WEIRD cultures to the remaining 88% (ref. 18). Instead, monetary—relative to psychological—incentives may increase effort less in non-WEIRD cultures. Across the cultures we have studied, the money advantage was largest in the United States and smallest in China. Accordingly, when designing interventions and reward schemes, particularly under limited resources, the relative benefit of adding a pay-per-effort incentive might be attenuated in non-WEIRD cultures.

Methods

All studies were carried out in accordance with all the ethical regulations and were approved by the Institutional Review Board (IRB15-1623 and IRB20-1056) at the University of Chicago. Informed consent was obtained from study participants consistent with the IRB protocol.

Study 1

Categorizing treatments

The original ref. 8 experiment included a total of 18 incentive treatments, to which participants from both India and the United States were randomly assigned. Out of those 18 treatments, we first selected all pay-for-effort treatments with linear, immediate and guaranteed piece-rates (for example, “You will be paid an extra 1 cent for every 100 points”) and categorized them as ‘monetary.’ These monetary treatments included a range of pay-for-effort incentives with varying linear piece-rates, ranging from 1 cent per 1,000 button presses to 10 cents per 100 button presses.

Using the criteria of linear, immediate and guaranteed piece-rates, we excluded those conditions that, for instance, tested for loss aversion (“You will be paid an extra 40 cents if you score at least 2,000 points”) or delay discounting (“You will be paid an extra 1 cent for every 100 points” with a 4-week delay). The reasons for these exclusions were that the former conditions only offered additional (lump) payment upon reaching a high number of presses, while the latter offered extra payment that was not immediate.

To create a list of non-monetary conditions, we selected treatments where the participants could not earn any additional money for themselves, neither immediately nor at some point in the future, and categorized them as ‘non-monetary’. The non-monetary treatments included conditions with incentives labelled by ref. 8 as ‘social psychological’ (such as the social-norm treatment “Many participants scored more than 2,000”), the two charity conditions, where the participants could earn money for the Red Cross but not for themselves, as well as the flat-fee ‘control’ condition (“Your score will not affect your payment”).

Supplementary Table 9a summarizes all the treatments we included on the basis of these criteria and Supplementary Table 9b lists all the treatments we excluded, together with the corresponding reasons for exclusion. Our final sample consisted of 6,294 participants: 5,526 in the United States and 768 in India. Demographic information is available in Supplementary Table 11a.

Study 2a–c

Establishing the minimal amount of effort

Instead of the button-pressing task, we asked participants to classify whether images contained a building or not. We kept the original 10-min limit but made it explicit to participants (including a comprehension check) that they were free to quit the task after every 10 image ratings without losing their base pay. We did this to remove any ambiguity regarding whether participants would be punished for not exerting a lot of effort, particularly in the non-monetary conditions, and because we were worried that non-WEIRD participants might be more inclined to think they might be penalized for doing so.

Internet connectivity

We also assumed that people might be faster at rating images on their computers than on their phones or tablets and that people in different countries might have different Internet speeds. While the main effect of culture on effort was not our primary variable of interest, we still chose to limit the participants to laptop or desktop computers only. We also asked participants to self-report how long it took for images to load. This allowed us to rule out technological differences across countries as an explanation for the pattern of findings. Analyses controlling for Internet connectivity are presented in Supplementary Table 13a.

Work meaning

Because the meaning behind work matters for the effectiveness of incentives106,142, we wanted to replicate the findings from Study 1, in which the work was explicitly meaningless, on a task where participants would be provided with some purpose behind the monotonous work. Therefore, we told participants that image classifications would help the researchers with ‘training a machine-learning algorithm’. The main rationale behind this change was the assumption that most real-life work tasks are not as devoid of meaning as pressing two buttons interchangeably for no apparent reason.

Base pay and piece-rates

Our goal was to select a piece-rate that would provide a sufficiently strong incentive for the participants to exert effort. However, we did not want the piece-rate to be so high as to remove any meaningful variation. We agreed to set the base pay across studies to $1.30 and the piece-rate in the monetary condition to 5 cents per 10 images (unless otherwise noted). Participants did not receive partial pay for rating fewer than 10 images within each increment. That is, a participant who rated 11 images and a participant who rated 19 images would both receive a bonus of 5 cents.

On the basis of a pilot, we expected the participants in the monetary condition to earn, on average, an additional 30 to 50 cents, a nominal amount and proportion of final earnings (relative to base pay for taking the study) between those of the ‘1 cent per 100 presses’ and ‘4 cents per 100 presses’ conditions in ref. 8. However, we estimated that participants working their hardest would be able to complete at least 150 images within the span of 10 min and thus earn 75 cents.

Incentive treatments

We wanted to have the minimal number of conditions to meaningfully test our hypothesis with a sufficiently large sample. Across Studies 2a–c, we kept one pay-for-effort incentive condition and two non-monetary conditions.

In Studies 2a and b, the non-monetary conditions included a social norm condition, where the participants were told that many other participants tried hard on the task and assessed 160 images, and a flat-fee condition, where no additional instructions were given. In Study 2c, we changed the non-monetary treatments. In one of them, participants read that the task was a competition and that they would see how they did relative to others upon completion (competition condition). In the other, participants read that they would not receive any additional pay; however, the researchers would transfer 5 cents for every 10 images participants rated to the Red Cross (charity condition).

In all the non-monetary conditions, the amount of money one earned was not contingent on how much effort one exerted. Every participant received the same salary regardless of how many images they rated. Yet, receiving a fixed salary might invoke norms with regards to obligations and reciprocity, which can differ across cultures143,144. Since the pay-off structure was the same across all the non-monetary conditions, we first pooled across the non-monetary conditions in each study and compared them to the monetary condition.

Procedure

Participants first read the consent form. Then they read about the nature of the task and were told that they would receive their base pay if they completed at least 10 image ratings. Next, participants had to pass an attention check that asked them about the purpose of the task mentioned on the previous page (to train a machine-learning image-classification database). After that, they learned about the structure of the task and their condition assignment. They had to pass two comprehension checks: one asking them about the maximum duration of the task (10 min) and the other asking them whether they would receive additional pay on the basis of how many images they rated. Further detail and exact wordings are provided in ‘Attention and comprehension checks’ in Supplementary Information.

Only those participants who passed the attention check and both comprehension checks proceeded to the image-rating task that contained the dependent variable. The participants saw several images one by one, with a 10-min timer visible. Images included pictures of plants, flowers, urban landscapes and buildings (for examples of images used in the picture-rating task, see Supplementary Images 1 and 2). For each image, the participants had to answer whether it contained a building.

After rating the first 10 images, the participants saw a screen that asked them whether they wanted to quit or continue with the task. If they chose to quit, they proceeded with the other questions in the survey. Participants who continued with the image-rating task saw a similar screen after rating every 10 images.

Cost-effectiveness

To calculate the amount of effort per dollar spent, we divided the number of images each participant rated by the pay they received. Confidence intervals for the amount per dollar spent are 95% confidence intervals of the mean effort divided by the cost in dollar amount adjusted for each level of effort. This cost would be the same for the non-monetary conditions for all amounts of effort within the confidence interval (except for the charity condition, which entailed an additional cost to implement contingent on participants’ effort). However, it would vary in the monetary conditions (that is, would be higher for higher values within the confidence interval).

Study 2a

Participants

In China, we recruited participants (N = 1,086) who were either students at Hubei University (N = 188) or users of two major social media in China: WeChat Moments (N = 79) or QQ Zones (N = 819). In the United Kingdom, we recruited participants on Prolific (N = 1,067). A colleague from Hubei University completed the translation. Demographic information is available in Supplementary Table 11b.

Incentive treatments

Participants were randomly assigned to one of the three conditions: monetary, flat fee and norm. The base pay was ¥8.25 ($1.30 at the time of the experiment) in China and $1.30 in the United Kingdom. In the monetary condition, the bonus for every additional 10 images in the monetary condition was ¥0.30 ($0.05) in China and $0.05 in the United Kingdom.

Study 2b

Participants

Participants for all three samples were recruited from Prolific: N = 1,053 in Mexico; N = 1,098 in the US sample with the same nominal pay; N = 1,122 in the US sample with increased pay pre-tested to be subjectively equivalent107 to that in Mexico. The participants in Mexico completed the study in Spanish, while those in the United States did so in English. A professional translator and a Spanish-native research assistant completed the translation. Demographic information is available in Supplementary Table 11b.

Incentive treatments

Participants were randomly assigned to one of the three conditions: monetary, flat fee and social norm. The incentive treatments were identical to those in Study 2a, except for the increase in pay in one of the US samples, as described below. Participants in Mexico received a $1.30 base pay for completing the study and, in the monetary condition, a $0.05 bonus per 10 image ratings. The two US samples (described in the next section) differed in both the flat fee that the participants received ($1.30 vs $1.56) and in the piece-rate ($0.05 vs $0.06 per 10 images) in the monetary condition.

Becker–DeGroot–Marshak procedure

To address the limitation of Study 1 and 2a, namely, the potentially different subjective values of the pay amounts between the samples: for Study 2b, we collected three samples of participants on Prolific: one from Mexico (N = 1,053) and two from the United States, with one US sample having the same nominal pay as in Mexico (N = 1,098) and the other having the same subjective pay as in Mexico (N = 1,122). Participants in the United States completed the study in English, while those in Mexico did so in Spanish.

To determine subjectively equivalent pay amounts for Prolific participants from Mexico and from the United States, we followed the Becker–DeGroot–Marschak (BDM) procedure107. A separate group of participants got briefly acquainted with the task and were then asked how much remuneration they would need to complete its full version. A detailed description of the procedure and the results is provided in ‘Becker–Degroot–Marshak (BDM) procedure for establishing pay equivalence’ in Supplementary Information.

Study 2c

Pre-registration

The study was pre-registered on 3 February 2023. The pre-registration is available on AsPredicted: https://aspredicted.org/dm562.pdf. The study did not deviate from the pre-registration. We pre-registered a total of 665 participants per sample to have enough power to detect an effect size of f = 0.14 for the interaction between incentive type and culture.

Participants

As in Study 2b, we recruited three samples of participants on Prolific: one from South Africa (N = 649) and two from the United States, with one US sample having the same nominal pay as in South Africa (N = 662) and the other having the same subjective pay as in South Africa (N = 662). All participants completed the survey in English. We pre-screened participants in South Africa who were fluent in English. Demographic information is available in Supplementary Table 11b.

Incentive treatments

Unlike in Studies 2a and b, participants in this study were first randomly assigned to types of incentives (monetary or non-monetary) and not individual incentive treatments. Those assigned to the non-monetary incentive type were then randomly assigned to either the competition or the charity condition. Thus, incentive types have roughly identical numbers of people per cell (330), but each individual non-monetary treatment had half as many people (165) as the monetary condition. In our previous studies, the number of participants was the same in each individual condition, not in each type of incentive condition (monetary or non-monetary). The pay structure in the monetary condition is described in the next section.

BDM procedure

To determine subjectively equivalent pay amounts between Prolific participants from South Africa and from the United States, we similarly followed the BDM procedure107. A separate group of participants got briefly acquainted with the task and were then asked how much remuneration they would need to complete the full version of the survey.

The two US samples differed in both the flat fee that participants received ($1.30 vs $2.25) for participation and in the pay-for-effort rate ($0.05 vs $0.09 per 10 images) in the monetary incentive condition. Participants in South Africa received a $1.30 base pay for completing the study in all three conditions and a $0.05 bonus per 10 image ratings in the monetary incentive condition. A detailed description of the procedure and the results are available in ‘Becker–Degroot–Marshak (BDM) procedure for establishing pay equivalence’ in Supplementary Information.

Study 3a

Pre-registration

The study was pre-registered on 25 August 2022, although we deviated from the pre-registration as described below. The pre-registration is available on AsPredicted: https://aspredicted.org/uz5gh.pdf.

The main deviation concerned the analysis plan. We pre-registered data analysis from both countries separately using t-tests, with a prediction that, for people in the United States, a small monetary incentive would result in higher effort than emphasizing the social norm; our second prediction was that the two incentives would be statistically indistinguishable from each other in India. While these hypotheses were supported by the data, we believe that we did not have sufficient power to detect the smallest meaningful effect size in India, hence the pre-registered analysis was inappropriate. Instead, we ran a multiple linear regression model, consistent with other studies in the paper, to probe for the interaction between incentive and country. We present the results of the pre-registered analyses in ‘Additional analyses for Study 3a’ in Supplementary Information.

Second, we recruited participants in the United States on Prolific instead of MTurk. The pre-registration stated that recruitment would take place on MTurk. We made this change because we noticed that some US MTurkers had posted reviews of the previous reiterations of the study, informing other MTurkers that it was sufficient to complete only 10 images to receive the pay, which we thought might compromise the quality of the data.

Third, we pre-registered 352 participants per country after exclusions (that is, people who did not rate a single image or those who did not complete the full study and receive payment) to have 80% power to observe a small to medium-sized effect (d = 0.30). We sampled more people to allow for exclusions and repeated submissions (exclusion details for all studies are available in ‘Exclusion data and criteria for Studies 2–4’ in Supplementary Information). Fewer people than we had estimated did not meet the inclusion criteria in the United States, hence the final sample included 382 people in the United States and 352 people in India. Excluding the last 30 participants in the United States did not significantly change any of the main results, and these analyses are reported in ‘Additional analyses for Study 3a’ in Supplementary Information.

Participants

In India (N = 352), we recruited people on MTurk. In the United States (N = 382), we recruited people on Prolific. Participants in both countries completed the survey in English. Demographic information is available in Supplementary Table 11c.

Incentive treatments

Participants were randomly assigned to one of the three conditions: minimal pay or norm. Procedure was identical to the one used in Studies 2a–c, except for the base pay that was $1.00 (as opposed to $1.30 in previous studies) in both India and the United States across conditions. The norm condition was identical to the one in Studies 2a and b. In the minimal-pay condition, the participants could earn an extra cent for completing every 20 image ratings.

Study 4

Participants

We recruited participants through the Center for Social and Behavior Change at Ashoka University, a private university in Haryana, India. The final sample consisted of 2,065 participants recruited through advertisements on the Facebook group ‘QMaths’. This group has over 280,000 members interested in preparing for competitive exams for jobs in sectors ranging from banking to railways. We selected this Facebook group because (1) Ashoka had an ongoing relationship with one of the moderators of the group and (2) the members would generally be proficient in English and Hindi. Demographic information is available in Supplementary Table 11d.

Language

Participants were randomly assigned to complete the survey in Hindi or in English. Participants in the Hindi condition completed the whole survey, starting with the consent form, in Hindi. Participants in the English condition completed the entire survey in English. Two research assistants from Ashoka University completed the translation from English into Hindi, occasionally changing the original English wordings to ensure compatibility between the two languages.

We included checks to ensure that participants were proficient in both languages. One week before taking the survey for Study 4, participants completed another survey for which they had to report being ‘very good’ or ‘fluent’ speakers of both English and Hindi.

Incentive treatments

Participants were randomly assigned to incentive and language conditions. To ensure a sufficiently high number of participants per condition, we kept two conditions: one monetary condition and one non-monetary condition (social norm). Therefore, the study followed a 2 language (English, Hindi) × 2 incentive (monetary, non-monetary) between-subjects design. The social norm condition was the same as in the previous studies. In the monetary condition, participants received a monetary bonus of ₹5 rupees ($0.0665) for every 10 images they rated. Everyone received ₹150 ($1.995) for completing the main experimental task. To calculate cost-efficiency, we used the exchange rate on the day the last response was collected (7 December 2021; ₹1 = $0.0133).

Additional variables

We included seven exploratory variables to measure participants’ perceptions of the task and motivations for completing it. These variables were all measured on a 7-point Likert scale (1 = strongly disagree to 7 = strongly agree). Participants reported their agreement with the following statements: “I enjoyed completing the task’; “I am satisfied with how well I did on the task”; “I believe that I helped others by completing the task”; “Completing the task was boring”; “I only completed the task for money”; “I could have assessed more pictures if I'd tried harder”; I am satisfied with the amount of pay I received”. Details are available in ‘Exploratory variables’ in Supplementary Information.

Reporting summary

Further information on research design is available in the Nature Portfolio Reporting Summary linked to this article.