Abstract
People experience a strong conflict while evaluating actors who unintentionally harmed someone—her innocent intention exonerating her, while the harmful outcome incriminating her. Different people solve this conflict differently, suggesting the presence of dispositional moderators of the way the conflict is processed. In the present research, we explore how reasoning ability and cognitive style relate to how people choose to resolve this conflict and judge accidental harms. We conducted three studies in which we utilized varied reasoning measures and populations. The results showed that individual differences in reasoning ability and cognitive style predicted severity of judgments in fictitious accidental harms scenarios, with better reasoners being less harsh in their judgments. Internal meta-analysis confirmed that this effect was robust only for accidental harms. We discuss the importance of individual differences in reasoning ability in the assessment of accidental harms.
Similar content being viewed by others
Introduction
In 2010, the rock band Lamb of God was performing in Czech Republic and, during the performance, the lead singer Randy Blythe threw a fan named Daniel Nosek off the stage, with the expectation that other people will catch him. Nosek instead fell backwards directly on his head, suffered severe traumatic brain injury, slipped into a coma, and died weeks later from his injuries. If we were to be part of the jury who was going to decide how morally bad Blythe’s behavior was and how much we should punish him, how would we go about it? Will we focus on his innocent intentions and reasonable beliefs about how things should have unfolded? Or will we be swayed by a strong emotional reaction in response to the details about suffering that Nosek had to endure because of Blythe’s actions? Will deliberating about the situation help us subdue influence of this emotional reaction on our decisions?
When it comes to evaluating third-party harmful behavior like this, past work has shown that people rely not only on the assessment of the mental state of the perpetrator, but also on the presence of a harmful consequence for the victim1,2,3,4,5,6,7. In other words, after witnessing a harmful event, a third-party moral judge reasons about the actor’s intentions (“What was Blythe thinking when he threw his fan off the stage?!”) and the victim’s feelings (“How painful it must have been for Nosek to suffer a head trauma?”). Not only has the past work validated this two-part template of intent-based morality at the psychological level, but also explored the neural substrates for these two independent processes1. In particular, this work reveals that the observers decode intentional status of interpersonal harmful actions via a network of brain regions—known as the Theory of Mind network—involved in representing others’ thoughts8, while representing victim’s feeling states recruits the “empathy for pain” network9. Probably the most salient way to demonstrate the dissociable contributions of these two processes towards moral evaluations is by focusing on how people judge accidents. Accidental harms elicit a strong conflict in the observer/judge because the two processes conflict with each other in terms of their output: the intent-based process focuses on innocent intentions of the actor and reduces severity of moral evaluations10,11, while the outcome-based process localizes on empathic reaction towards the victim suffering and the agent’s causal role in producing this outcome and increases severity of moral condemnation9. As a result, how we judge accidents depends on how we resolve the conflict posed by these two processes: difficulties in processing intentions leads to more punitive attitudes (e.g., autistic individuals12), while deficits in empathic reaction towards the victim can lead to forgiving attitudes (e.g., psychopathy and sadism13,14). In other words, forgiving accidental harms feels so difficult because it involves overriding a potent emotional reaction to victim suffering with a more deliberative response stemming from reasoning about intentionality. It is worth noting that this conflict is specific to accidents and is not encountered while evaluating other interpersonal interactions. When the actor intentionally harms someone, two processes agree on condemnation15, while when the actor attempts to harm someone but fails, there is no strong empathic emotional response that needs to be counteracted by the mentalizing system and the two process again agree on condemnation1,8,9,16.
Although past work has thoroughly explored the processes that give rise to conflict while pondering over accidents and the role of dispositional mentalizing and empathizing abilities in resolving this conflict, much less attention has been paid to how one’s ability and willingness to engage in analytical reasoning affect this conflict resolution. Indeed, a hint for such a role comes from work with sacrificial moral dilemmas. Sacrificial moral dilemmas, like accidents, pose a conflict of a different variety—between the emotionally aversive “utilitarian” option of personally harming someone and the option of letting a greater number of individuals get hurt. Past work shows that differences in individual’s ability to reason and availability of cognitive resources play a role in how this conflict is resolved (for a review, see17,18,19). Current work is inspired by these prior investigations.
A decade long research in the field of moral psychology has revealed how our moral judgments are, at the broadest level, the result of an interplay between emotions on the one hand, and reason on the other20,21,22. Although there are many variants of dual process models21,23,24,25,26, the generic version distinguishes a fast, parallel, and almost automatic thinking system (intuitive system) from a slow, sequential, and cognitively effortful thinking system (analytical system). Note that reasoning psychologists use words interchangeably to define the two systems (e.g., intuitive/heuristic system versus analytical/deliberative/reflective system). For the sake of consistency, we will always use the words intuitive system/participant vs. analytical system/participant.
Considering that the human mind is composed of two thinking systems has led researchers to design specific manipulations and measures to test predictions derived from the dual-process model, and these protocols have proven to be useful in the field of moral judgment and decision making17. Specifically, numerous tasks have been designed that typically make an intuitive response conflict with an analytical response, and numerous measures have been used to capture the underlying psychological mechanisms (e.g. response time measure, time pressure manipulation, interfering cognitive load manipulation, interindividual differences, etc.)17. The utility of this corpus of measures and manipulations has been thoroughly explored in the context of sacrificial moral dilemmas27,28,29,30,31,32,33,34,35,36,37,38. It is also worth noting, however, that there are studies which have found no effects of some of these manipulations on moral judgment activities39, or which reanalyze previous work to highlight their limited generalizability40,41. Additionally, the “corrective” dual-process model, which posits correction of the “intuitive” response by “deliberative” processes, has also been called into question42,43,44,45. Current work focuses only on interindividual differences and is agnostic with regards to the time course or “intuitiveness” of the responses from two systems and therefore we will not discuss debate any further.
The starting point for us is prior work which argues that people who score higher on self-report or performance measures of reflective reasoning also tend to be more “utilitarian”36 while resolving sacrificial dilemma conflicts. In the current work, we plan to extend this work to explore more broadly the role of analytical reasoning in resolving the kind of conflict one encounters while evaluating behavior of unintentional harm-doers. In a manner reminiscent of how reasoning bolsters “utilitarian” inclinations on sacrificial dilemmas, we predict that more capable reasoners or people prone to analytical reasoning will resolve it by overriding the strong emotional response, which would lead to a greater acceptability of accidental harms.
General methods
Across all studies, experimental stimuli consisted of intent-based moral vignettes46 that were result of a 2 × 2 within-subjects design where the factors belief (neutral, negative) and outcome (neutral, negative) were independently varied such that agents in the scenario produced either a neutral outcome or a harmful outcome while acting with the belief that they were causing either a neutral outcome or a harmful outcome. The magnitude of harm severity varied freely across scenarios from mild to severe to fatal injuries. We provide below an example of the parametric variation in a single scenario context in accidental harm condition:
Background
Matilda is walking by a neighbour’s swimming pool when she sees a child about to dive in.
Foreshadow
The child is about to dive into the shallow end and smack his head very hard on the concrete bottom of the pool.
Belief
Because of a label on the side of the pool, Matilda believes that the child is about to dive safely into the deep end and swim around.
Outcome
Matilda walks by, without saying anything to the child. The child dives in and breaks his neck.
The number of scenarios used and the type of question asked varied across studies, along with the scale used for measuring a response. These differences across studies are tabulated in Table 1. Additionally, detailed text for the scenarios are reproduced in Supplementary Text S2.
Sampling stopping rule and exclusion criteria
These studies were each part of prior unrelated data collection, thus impeding any sampling size control in the present research.
For Amazon Mechanical Turk studies, following exclusion criteria were applied to leave out participants who: did not complete the entire survey, reported to be less than 18 years old or more than 100 years old, failed attention checks, completed the same survey multiple times. Additionally, we used TurkPrime to make the survey available for completion only to MTurk workers who had a rating of above 95% and had completed at least 100 other HITs. All sample sizes reported below refer to the final sample after these exclusion criteria had been applied.
Data analysis
Since the behavioral data (items within conditions within participants) had multilevel or nested structure, we utilized mixed-effects models to correctly handle the inherent dependencies in nested designs and to reduce probability of Type I error due to reduced effective sample size47,48,49.
When null hypothesis significance testing (NHST) results in a failure to reject the null hypothesis (H0), this cannot be taken as evidence in support of the null hypothesis, because p values are unable to quantify support in favor of the null50. Therefore, Bayes Factors (BF) were calculated for group comparisons to assess the relative likelihood of the null and alternative (H1) hypotheses51. A BF01 of greater than 1 implies that the data are more likely to occur under H0 than under H1. Similarly, a BF01 lower than 1 indicates that the data are more likely to occur under H1 than under H0. Thus, if we analyze data and find that BF01 = 3, this means that the data are 3 times more likely to have occurred under H0 than under H1. Based on prior guidelines52, BFs between 1 and 3, between 3 and 10, and larger than 10 are interpreted as ambiguous, moderate, and strong support, respectively. Note that, where relevant, we provide natural logarithm values for Bayes Factors (loge(BF01)), which need to be exponentiated to get the BF01.
Meta-analysis
Our exploratory individual difference studies were not designed to characterize a detailed pattern of associations between reasoning measures and intent-based moral judgment (e.g., some specific correlations would be stronger than others), but instead to firmly establish the general form of this association. Therefore, we carried out a random-effects meta-analysis53,54 using regression estimates (and the associated standard errors) across measures for each study and assessed if the meta-analytic effect was significantly different than 0. In addition to providing details from null hypothesis significance testing (NHST) approach, we also compute Bayes Factors for random-effects meta-analysis using default priors from metaBMA R package55.
Data reporting
Statistical analysis was carried out in R programming language56 using easystats57,58,59,60 packages. For the sake of brevity, results from statistical analyses are included in the figures rather than the main text (an approach adopted in the R package ggstatsplot61,62). Similarly, details about demographics and experimental design for the studies are provided in Table 1.
Ethics statement
Across all studies, participants provided written informed consent before any study procedure was initiated. Studies 1 and 2 were approved by the Ethics Committee of Scuola Internazionale Superiore di Studi Avanzati (Trieste) and the Hospital ‘Santa Maria della Misericordia’ (Udine), respectively. The Study 3 was approved by the Ethics Committee of Harvard University. All studies were conducted according to the principles in the Declaration of Helsinki.
Participants
See Table 1.
Measures
The following questionnaires were included across 3 studies (for more detailed descriptions, see Supplementary Text S1):
-
Need for Cognition (NFC63) assesses the degree to which individuals are intrinsically motivated to engage in cognitive deliberation.
-
Cognitive Reflection Test (6-item CRT64) captures people's ability to override an appealing but incorrect intuitive response.
-
Rational Experiential Inventory (REI65) assesses the degree to which people engage in two modes of thinking: a fast, intuitive automatic thinking and a slower logical thinking.
-
Actively Open-Minded thinking (AOT66,67) assesses individual differences in disposition to consider different conclusions even if they go against one’s own initial conclusion, to spend enough time on a problem before giving up, and to consider the opinions of others in forming one’s own opinions
-
Belief Bias (BB68) measures the tendency to judge the strength of arguments based on the believability of their conclusion rather than how strongly they logically support that conclusion. Only syllogisms in which conclusions are logically invalid but believable (the class of problems that elicit high belief bias) are employed. They are taken from previous studies67,69,70.
Results
As mentioned before, our studies were not designed to explore reasoning measure-specific associations, but instead to establish the general form of this association. Accordingly, a random-effects meta-analysis of regression estimates revealed significant negative meta-analytic summary effect only for the neutral and accidental harm cases, such that people who scored higher on reasoning measures were more lenient in their assessment of such cases (see Fig. 1). But a more careful look at the Bayes Factor for the neutral condition reveals that the evidence in favor of the alternative hypothesis was inconclusive (BF10 = 1.15), while it was substantial for the accidental harm cases (BF10 = 5.15). Looking at Bayesian meta-analysis, we could also show that there was a strong evidence in favor of the null hypothesis for the attempted (BF01 = 17.46) and intentional (BF01 = 24.28) harm cases, i.e., summarizing across measures, there is no relationship between reasoning ability and tendency to judge attempted or intentional harm cases. Therefore, the meta-analysis supports our claim that scoring higher on reasoning measure is associated with greater tendency to judge third-party harmful transgressions more leniently, but only when the harm is caused accidentally.
General discussion
Across a series of studies, we investigated the role that reasoning in determining the severity of moral judgments about harmful transgressions and observed that participants who self-reported to be more analytic and adept at cognitive deliberation by disposition were consistently more lenient in their judgments of accidental harms, as compared to participants who reported to rely more on the intuitive style of thinking.
The two-process model for intent-based morality1,16 argues that accidental harm produces a cognitive conflict between two processes: an agent-based, intent-driven response to forgive10,11 (based on innocent intentions) and a victim-based, empathy-driven impulse to condemn9,14,71 (based on harm caused). Current work expands on this work by focusing on the source of interindividual differences in how people resolve this conflict and assign blame or punishment to accidental harm-does and shows that analytical reasoning skills are one such source.
There are (at least) two possible ways in which reasoning can lead to a more lenient assessment of accidental harm-doers: (i) Individuals with better cognitive abilities also have more executive resources needed for Theory of Mind72, i.e. they are better at representing innocent mental states of the agent who accidentally harmed someone and thus forgive them. (ii) Individuals with higher propensity for cognitive deliberation are also better at down-regulating their empathic arousal stemming from harm appraisal and are thus more likely to forgive accidental harm-doers71. Future work should explore if it’s the cognitive (Theory of Mind) or the affective (empathic arousal) route or both that mediate the influence of reasoning ability on third-party moral evaluation.
Limitations
Although our internal meta-analysis tries to draw conclusions that are generalizable across the battery of reasoning measures we utilized, the generalizability and robustness of these findings to a different set of reasoning and cognitive ability measures remains to be studied73. We hope future studies can overcome this limitation by employing a more comprehensive battery of cognitive measures. Additionally, the type of questions for moral judgement varied across the three studies (acceptability/wrongness; blame/punishment). Although the consistency of negative correlations between accidental harm and reasoning measures attests to the robustness of these effects to different framings, there needs to be a more systematic investigation about whether the strength of association might vary depending on the question asked9,16.
Conclusion
Taken together, the present results argue that individual differences in reasoning are associated with differences in the way people cope with cognitive conflict when evaluating accidental harmful transgressions. The study of cognitive conflict (its detection and resolution) in the moral judgment field is an area of research still in its infancy, and we believe that the current work is a valuable addition to this growing field and hints at a number of exciting new avenue to explore.
Data availability
Data and analysis scripts are available from the Open Science Framework: https://osf.io/ayb7d/.
References
Cushman, F. Deconstructing intent to reconstruct morality. Curr. Opin. Psychol. 6, 97–103 (2015).
Young, L. & Tsoi, L. When mental states matter, when they don’t, and what that means for morality. Soc. Personal. Psychol. Compass 7, 585–604 (2013).
Buon, M., Seara-Cardoso, A. & Viding, E. Why (and how) should we study the interplay between emotional arousal, theory of mind, and inhibitory control to understand moral cognition?. Psychon. Bull. Rev. 23, 1660–1680 (2016).
Ginther, M. R. et al. Parsing the behavioral and brain mechanisms of third-party punishment. J. Neurosci. 36, 9420–9434 (2016).
Lagnado, D. & Channon, S. Judgments of cause and blame: The effects of intentionality and foreseeability. Cognition 108, 754–770 (2008).
Alicke, M. & Davis, T. The role of a posteriori victim information in judgments of blame and sanction. J. Exp. Soc. Psychol. 25, 362–377 (1989).
Alicke, M. Culpable causation. J. Pers. Soc. Psychol. 63, 368–378 (1992).
Young, L., Cushman, F., Hauser, M. & Saxe, R. The neural basis of the interaction between theory of mind and moral judgment. Proc. Natl. Acad. Sci. 104, 8235–8240 (2007).
Patil, I., Calò, M., Fornasier, F., Cushman, F. & Silani, G. The behavioral and neural basis of empathic blame. Sci. Rep. 7, 5200 (2017).
Young, L. & Saxe, R. Innocent intentions: A correlation between forgiveness for accidental harm and neural activity. Neuropsychologia 47, 2065–2072 (2009).
Patil, I., Calò, M., Fornasier, F., Young, L. & Silani, G. Neuroanatomical correlates of forgiving unintentional harms. Sci. Rep. 7, 45967 (2017).
Buon, M. et al. The role of causal and intentional judgments in moral reasoning in individuals with high functioning autism. J. Autism Dev. Disord. 43, 458–470 (2013).
Young, L., Koenigs, M., Kruepke, M. & Newman, J. P. Psychopathy increases perceived moral permissibility of accidents. J. Abnorm. Psychol. 121, 659–667 (2012).
Trémolière, B. & Djeriouat, H. The sadistic trait predicts minimization of intention and causal responsibility in moral judgment. Cognition 146, 158–171 (2016).
Patil, I. & Silani, G. Alexithymia increases moral acceptability of accidental harms. J. Cogn. Psychol. 26, 597–614 (2014).
Cushman, F. Crime and punishment: Distinguishing the roles of causal and intentional analyses in moral judgment. Cognition 108, 353–380 (2008).
Trémolière, B., Neys De, W. & Bonnefon, J.-F. Reasoning and moral judgment: A common experimental toolbox. In The Routledge International Handbook of Thinking and Reasoning (eds Ball, L. J. & Thompson, V. A.) 575–589 (Routledge/Taylor & Francis Group, 2018).
Bialek, M. & Terbeck, S. Can cognitive psychological research on reasoning enhance the discussion around moral judgments?. Cogn. Process. 17, 329–335 (2016).
Capraro, V. The dual-process approach to human sociality: A review. SSRN Electron. J. https://doi.org/10.2139/ssrn.3409146 (2019).
Greene, J., Nystrom, L. E., Engell, A. D., Darley, J. M. & Cohen, J. D. The neural bases of cognitive conflict and control in moral judgment. Neuron 44, 389–400 (2004).
Evans, J. S. B. T. Dual-processing accounts of reasoning, judgment, and social cognition. Annu. Rev. Psychol. 59, 255–278 (2008).
Kahneman, D. Thinking Fast and Slow (Farrar, 2011).
Epstein, S. Integration of the cognitive and the psychodynamic unconscious. Am. Psychol. 49, 709–724 (1994).
Kahneman, D. & Frederick, S. A model of heuristic Judgment.ord display. In The Cambridge Handbook of Thinking and Reasoning (eds Holyoak, K. J. & Morrison, R. G.) 167–293 (Cambridge University Press, 2005).
Sloman, S. A. The empirical case for two systems of reasoning. Psychol. Bull. 119, 3–22 (1996).
Evans, J. S. B. T. & Stanovich, K. E. Dual-process theories of higher cognition: Advancing the debate. Perspect. Psychol. Sci. 8, 223–241 (2013).
Greene, J., Morelli, S. A., Lowenberg, K., Nystrom, L. E. & Cohen, J. D. Cognitive load selectively interferes with utilitarian moral judgment. Cognition 107, 1144–1154 (2008).
De Neys, W. & Białek, M. Dual processes and conflict during moral and logical reasoning: A case for utilitarian intuitions? In Moral Inferences (eds Trémolière, B. & Bonnefon, J. F.) 123–136 (Psychology Press, 2017).
Kvaran, T., Nichols, S. & Sanfey, A. The effect of analytic and experiential modes of thought on moral judgment. Prog. Brain Res. 202, 187–196 (2013).
Trémolière, B., De Neys, W. & Bonnefon, J.-F. The grim reasoner: Analytical reasoning under mortality salience. Think. Reason. 20, 333–351 (2014).
Conway, P. & Gawronski, B. Deontological and utilitarian inclinations in moral decision making: A process dissociation approach. J. Pers. Soc. Psychol. 104, 216–235 (2013).
Trémolière, B., Neys, W. D. & Bonnefon, J.-F. Mortality salience and morality: Thinking about death makes people less utilitarian. Cognition 124, 379–384 (2012).
Cummins, D. D. & Cummins, R. C. Emotion and deliberative reasoning in moral judgment. Front. Psychol. 3, 328 (2012).
Suter, R. S. & Hertwig, R. Time and moral judgment. Cognition 119, 454–458 (2011).
Gürçay, B. & Baron, J. Challenges for the sequential two-system model of moral judgement. Think. Reason. 23, 49–80 (2017).
Patil, I. et al. Reasoning supports utilitarian resolutions to moral dilemmas across diverse measures. J. Pers. Soc. Psychol. 120, 443–460 (2021).
Capraro, V., Everett, J. A. C. & Earp, B. D. Priming intuition disfavors instrumental harm but not impartial beneficence. J. Exp. Soc. Psychol. 83, 142–149 (2019).
Berman, J. Z., Barasch, A., Levine, E. E. & Small, D. A. Impediments to effective altruism: The role of subjective preferences in charitable giving. Psychol. Sci. https://doi.org/10.1177/0956797617747648 (2018).
Tinghög, G. et al. Intuition and moral decision-making—The effect of time pressure and cognitive load on moral judgment and altruistic behavior. PLoS ONE 11, e0164012 (2016).
Baron, J. & Gürçay, B. A meta-analysis of response-time tests of the sequential two-systems model of moral judgment. Mem. Cogn. 45, 566–575 (2017).
McGuire, J., Langdon, R., Coltheart, M. & Mackenzie, C. A reanalysis of the personal/impersonal distinction in moral psychology research. J. Exp. Soc. Psychol. 45, 577–580 (2009).
Bago, B. & De Neys, W. The intuitive greater good: Testing the corrective dual process model of moral cognition. J. Exp. Psychol. Gen. https://doi.org/10.1037/xge0000533 (2018).
Bago, B. & De Neys, W. Advancing the specification of dual process models of higher cognition: A critical test of the hybrid model view. Think. Reason. 26, 1–30 (2020).
Bago, B. & De Neys, W. Fast logic?: Examining the time course assumption of dual process theory. Cognition 158, 90–109 (2017).
Bago, B. & De Neys, W. The Smart System 1: Evidence for the intuitive nature of correct responding on the bat-and-ball problem. Think. Reason. 25, 257–299 (2019).
Young, L., Camprodon, J. A., Hauser, M., Pascual-Leone, A. & Saxe, R. Disruption of the right temporoparietal junction with transcranial magnetic stimulation reduces the role of beliefs in moral judgments. Proc. Natl. Acad. Sci. 107, 6753–6758 (2010).
Aarts, E., Verhage, M., Veenvliet, J. V., Dolan, C. V. & van der Sluis, S. A solution to dependency: Using multilevel analysis to accommodate nested data. Nat. Neurosci. 17, 491–496 (2014).
Barr, D. J., Levy, R., Scheepers, C. & Tily, H. J. Random effects structure for confirmatory hypothesis testing: Keep it maximal. J. Mem. Lang. 68, 255–278 (2013).
Judd, C. M., Westfall, J. & Kenny, D. A. Treating stimuli as a random factor in social psychology: A new and comprehensive solution to a pervasive but largely ignored problem. J. Pers. Soc. Psychol. 103, 54–69 (2012).
Wagenmakers, E.-J. A practical solution to the pervasive problems of p values. Psychon. Bull. Rev. 14, 779–804 (2007).
Jarosz, A. F. & Wiley, J. What are the odds? A practical guide to computing and reporting Bayes factors. J. Probl. Solving 7, 2 (2014).
Etz, A. & Vandekerckhove, J. A Bayesian perspective on the reproducibility project: Psychology. PLoS ONE 11, e0149794 (2016).
Borenstein, M., Hedges, L. V., Higgins, J. P. T. & Rothstein, H. R. Introduction to Meta-analysis (Wiley, 2009). https://doi.org/10.1002/9780470743386.
Viechtbauer, W. Conducting meta-analisys in R with metafor package. J. Stat. Softw. 36, 1–48 (2010).
Heck, D. W., Gronau, Q. F. & Wagenmakers, E.-J. metaBMA: Bayesian model averaging for random and fixed effects meta-analysis (version 0.3.9). CRAN (2017).
R Development Core Team. R: A Language and Environment for Statistical Computing (R Development Core Team, 2021).
Lüdecke, D., Ben-Shachar, M., Patil, I., Waggoner, P. & Makowski, D. performance: An R package for assessment, comparison and testing of statistical models. J. Open Source Softw. 6, 3139 (2021).
Ben-Shachar, M., Lüdecke, D. & Makowski, D. effectsize: Estimation of effect size indices and standardized parameters. J. Open Source Softw. 5, 2185 (2020).
Makowski, D., Ben-Shachar, M., Patil, I. & Lüdecke, D. Methods and algorithms for correlation analysis in R. J. Open Source Softw. 5, 2306 (2020).
Lüdecke, D., Ben-Shachar, M., Patil, I. & Makowski, D. Extracting, computing and exploring the parameters of statistical models using R. J. Open Source Softw. 5, 2445 (2020).
Patil, I. Visualizations with statistical details: The ‘ggstatsplot’ approach. J. Open Source Softw. 6, 3167 (2021).
Patil, I. statsExpressions: R package for tidy dataframes and expressions with statistical details. J. Open Source Softw. 6, 3236 (2021).
Cacioppo, J. T., Petty, R. E. & Kao, C. F. The efficient assessment of need for cognition. J. Pers. Assess. 48, 306–307 (1984).
Finucane, M. L. & Gullion, C. M. Developing a tool for measuring the decision-making competence of older adults. Psychol. Aging 25, 271–288 (2010).
Pacini, R. & Epstein, S. The relation of rational and experiential information processing styles to personality, basic beliefs, and the ratio-bias phenomenon. J. Pers. Soc. Psychol. 76, 972–987 (1999).
Baron, J. Why teach thinking?—An essay. Appl. Psychol. 42, 191–214 (1993).
Baron, J., Scott, S., Fincher, K. & Emlen Metz, S. Why does the cognitive reflection test (sometimes) predict utilitarian moral judgment (and other things)?. J. Appl. Res. Mem. Cogn. 4, 265–284 (2015).
Evans, J. S. B. T., Barston, J. L. & Pollard, P. On the conflict between logic and belief in syllogistic reasoning. Mem. Cognit. 11, 295–306 (1983).
Thompson, V. & Evans, J. S. B. T. Belief bias in informal reasoning. Think. Reason. 18, 278–310 (2012).
Morley, N. J., Evans, J. S. B. T. & Handley, S. J. Belief bias and figural bias in syllogistic reasoning. Q. J. Exp. Psychol. Sect. A 57, 666–692 (2004).
Treadway, M. T. et al. Corticolimbic gating of emotion-driven punishment. Nat. Neurosci. 17, 1270–1275 (2014).
Wade, M. et al. On the relation between theory of mind and executive functioning: A developmental cognitive neuroscience perspective. Psychon. Bull. Rev. https://doi.org/10.3758/s13423-018-1459-0 (2018).
Yarkoni, T. The generalizability crisis. Behav. Brain Sci. https://doi.org/10.1017/S0140525X20001685 (2020).
Acknowledgements
We thank Fiery Cushman for helpful conversations and for financially supporting Study 3, and Flora Schwartz for helpful comments on the draft manuscript. The publication of this article was supported by an ANR grant (ANR-19-CE28-0002) awarded to B.T.
Author information
Authors and Affiliations
Contributions
I.P. designed the Studies 1–3. I.P. conducted the statistical analyses and prepared the figures. I.P. and B.T. wrote the manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Patil, I., Trémolière, B. Reasoning supports forgiving accidental harms. Sci Rep 11, 14418 (2021). https://doi.org/10.1038/s41598-021-93908-z
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-021-93908-z
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.