Most normative ethical theories and religious traditions offer general moral principles that dictate how people ought to act across different situations. For example, Kantian deontologists argue that people should always respect the dignity of all persons and always treat people as an end, never merely as a means (Kant, 1785/1959). Some utilitarians argue that people should always act in a way that achieves the greatest good for the greatest number (Mill, 1861/1998). The Ten Commandments in Christian religious traditions offer a set of moral principles for the faithful to follow. In making everyday judgments and decisions, people appeal to a variety of different moral principles (Albertzart, 2013; Väyrynen, 2008): One might assert that people should always be honest, treat others fairly, show respect to authority figures, or help others in need. Such moral principles are meant to be stable guides for making future decisions and for justifying evaluations of past actions. That is, they are meant to be fixed and insensitive to any individual’s immediate circumstances or needs.

But do moral principles stably guide and justify people’s actions and judgments? Some cognitive science research suggests that people may use general moral principles to make and change their judgments and decisions with some reliability (Cushman et al., 2006; Horne et al., 2015). For example, people commonly appeal to the action principle (i.e., that it is worse to cause harm by direct action than to cause equivalent harms by omission) to justify their moral judgments across diverse scenarios (Cushman & Young, 2011; Cushman et al., 2006). Furthermore, whether people agree with a utilitarian moral principle reliably predicts the judgments they make in moral dilemmas (Lombrozo, 2009), and reminding people of an endorsed utilitarian principle induces judgment change in moral dilemmas in a way that accords with the principle (Horne et al., 2015). More anecdotally, many of us can recall situations in which we chose not to perform an action that would benefit us in order to abide by a particular moral principle, such as “Keep your promises.” Despite some evidence that people use moral principles in making and justifying their judgments and decisions, they still might flexibly shift their agreement with moral principles depending upon the events that are accessed in memory. In particular, we suggest that people might hold moral principles with less conviction after recalling their own violations of those principles yet maintain their level of conviction after recalling others’ violations of those same principles.

Why might people reduce agreement with moral principles after recollecting their own violations of those principles, but not the violations of others? One possible answer to this question is that people tend to excuse their own moral transgressions, but not the moral transgressions of others. Selectively excusing their own past transgressions more than the transgressions of others may motivate them to reduce the extent to which they agree with the violated moral principles. That is, excusing our own improprieties may highlight that the principle should be held less rigidly or with less conviction. Lending credence to this possible explanation, converging lines of research suggest that people evaluate themselves more favorably than others (Alicke & Sedikides, 2009; Taylor & Brown, 1988, 1994). For example, people believe they are more virtuous, intelligent, talented, and compassionate than the average person (Alicke & Govorun, 2005; Batson & Collins, 2011). In fact, people exhibit a pronounced sense of moral superiority obtained by evaluating themselves more favorably than others (Batson & Collins, 2011; Tappin & McKay, 2017). And people use memory search, reconstruction, and interpretation to support more favorable evaluations of themselves than of others (D’Argembeau & Van der Linden, 2008; Demiray & Janssen, 2015; Kunda, 1990; Stanley & De Brigard, 2019; Stanley et al., 2017, 2019a). For example, people are more likely to recall and vividly reexperience positive information about themselves than they are to recall such information about others (D’Argembeau & Van der Linden, 2008). If people evaluate themselves more favorably than others, then they may excuse their own remembered transgressions more than similar kinds of remembered transgressions committed by others. In this way, moral principles would not always actively regulate moral judgments and decisions; instead, the accessibility of certain events in memory might dictate the endorsement of moral principles moment to moment.

In two studies, we investigate whether the accessibility of relevant events in memory shifts agreement with common moral principles. Using complementary experimental designs, Studies 1 and 2 test the central hypothesis that, after recalling an event in which they personally violated a moral principle, participants would agree less strongly with the violated principle, relative to participants who recall others’ violations. Study 2 addresses the proposed mechanism with a mediation design. That is, Study 2 tests whether the shift in moral principle agreement after recalling a personal past violation of that principle is explained, in part, by participant’s willingness to selectively excuse their own immoral actions.

Both studies presented herein were formally preregistered. For both studies, we report all exclusion criteria, all materials and conditions included, and all independent and dependent measures. Deidentified data are publicly available on OSF (https://osf.io/nqjf7/).

Study 1

In Study 1, participants recalled an event in which they violated a moral principle, or an event in which another person violated a moral principle. After recalling the event in accordance with the cue, they then rated their agreement with the violated moral principle. We assume that, because of random assignment to conditions, participants across conditions exhibit the same level of agreement, on average, with the moral principles prior to recalling a violation committed by themselves or by others. Assuming no baseline differences in moral principle agreement, any differences in agreement observed after recalling different kinds of past events is likely produced by change in principle agreement following the experimental manipulation. The preregistration for Study 1 is available on OSF (https://osf.io/cq32m). Study 2 then uses a pre–post experimental design to more directly measure change in moral principle agreement.

Materials and methods

Participants

A total of 1,205 individuals completed this study via Amazon’s Mechanical Turk for monetary compensation. Participant recruitment was restricted to individuals in the United States who had completed at least 50 HITs and had a prior approval rating of at least 90%. Ninety-three participants were excluded for failing a check question. As such, we were left with 1,112 participants (Mage = 36 years, SD = 11, age range: 18–77, 502 females, 606 males). We conducted a power analysis using G*Power for a one-way analysis of variance (ANOVA) with power (1 − β) set at .85 and α = .05 (two-tailed). We aimed to recruit a total of 1,205 to detect small-to-moderate effect sizes for each of the three principles (after expected exclusions). This target sample size would ensure that we would have at least 100 participants per cell in our experimental design. All participants reported being fluent English speakers. Informed consent was obtained from each participant in accordance with protocol approved by the Duke University Campus Institutional Review Board.

Materials

Three moral principles were in this study: (1) People should be honest with other people; (2) people should treat other people fairly; and (3) people should not harm other people. Recent research has successfully used cues involving dishonesty, unfairness, and harm to elicit memories of morally wrong actions committed by oneself and by others (Stanley et al., 2019a). Therefore, we expected that most participants would be able to recall a specific memory from the personal past in which they violated each of these principles.

Procedure

This study was self-paced and consisted of a single session. Participants were randomly presented with one of the three moral principles and randomly assigned to the actor condition, the recipient condition, or the observer condition. In the actor condition, participants were asked to recall an event from their personal past in which they themselves violated the principle (i.e., dishonesty, unfairness, harm). In the recipient condition, participants were asked to recall an event from their personal past in which someone else violated the principle and the participant was the recipient of the transgression (e.g., a memory in which someone else was dishonest with the participant). In the observer condition, participants were asked to recall an event from their personal past in which they witnessed someone else violating the principle and the participant was not the recipient of the transgression. We included both recipient and observer conditions as comparisons to comprehensively account for the possible ways in which people actually encounter moral transgressions in the world. Moreover, the inclusion of both recipient and observer conditions also helps to rule out the possibility that the hypothesized self–other effects are simply the product of having been a direct participant in the remembered event (either as an actor or a as victim), as opposed to a mere witness.

Regardless of condition, participants were instructed to recall only actions that they believed to be morally wrong. All participants were also reminded that all of their responses were completely confidential. All participants described the remembered event in 3–5 sentences and provided the month and year that it occurred. This strategy for eliciting memories of moral transgressions was adapted from Stanley et al. (2019a). On the next page, participants were asked whether they were able to generate a memory of an action that they believe to be morally wrong in accordance with the cue provided. Participants who reported that they did not generate a memory in accordance with the cue were excluded from our analyses (a total of 19 participants reported that they did not generate a memory in accordance with the cue). Specifically, 98%, 98%, and 99% of participants reported being able to recall a morally wrong violation of the principle in the actor, recipient, and observer conditions, respectively.

Participants then completed an unrelated 1-minute categorization distractor task, after which they were presented with the same moral principle again. Participants were instructed to make a judgment about how much they agreed or disagreed with the principle on a 7-point scale (1 = strongly disagree, 7 = strongly agree).

At the end, participants were asked the following: Do you feel that you paid attention, avoided distractions, and took the survey seriously? They responded by selecting one of the following: (1) no, I was distracted; (2) no, I had trouble paying attention; (3) no, I did not take the study seriously; (4) no, something else affected my participation negatively; or (5) yes. Participants were ensured that their responses would not affect their payment or their eligibility for future studies. Only those participants who selected ‘5’ were included in the analyses (see exclusions, above). This same attention check question has been used in recent published research (e.g., Henne et al., 2019; Stanley et al., 2019b, 2020, 2021). Upon completion, participants were monetarily compensated for their time.

Statistical analyses

Data were analyzed using R with the ‘lme4’ software package (Bates et al., 2015) and the ‘lmerTest’ software package (Kuznetsova et al., 2017). Data were fit to linear mixed-effects models (LMEM). Significance for fixed effects was assessed using Satterthwaite approximations to degrees of freedom, and 95% confidence intervals around beta values were computed using parametric bootstrapping (on our view, 95% CIs around beta-values offer the best available indication of effect size for LMEMs). We present the results of LMEMs in the main text, but separate analyses for each individual principle are available in the supplement.

Results

We tested our central hypothesis that, after recalling an event in which they personally violated a moral principle, participants would agree less strongly with the violated principle—relative to participants who recalled moral principle violations committed by others. To this end, we computed a LMEM with condition (actor, recipient, observer) on moral principle agreement ratings. The moral principle was the random effect in the model (random intercepts only). Corroborating our hypothesis, participants in the actor condition agreed less strongly with the moral principles than participants in the recipient (b = .31, SE = .05, t = 6.21, p < .001, 95% CI [.21, .41]) and observer conditions (b = .28, SE = .05, t = 5.57, p < .001, 95% CI [.18, .38]); however, there was no difference in the change in agreement ratings between recipient and observer conditions (p = .58). Figure 1 graphically depicts agreement ratings for each individual moral principle as a function of experimental condition. See Supplemental Information for descriptive statistics and additional analyses for each principle taken separately.

Fig. 1
figure 1

Mean agreement with the moral principles. Error bars indicate 95% confidence intervals. Light-grey points represent individual participant responses evenly jittered

Study 2

Study 2 builds on Study 1 by implementing a pre–post experimental design to directly measure change in moral principle agreement after recalling a moral principle violation committed by oneself or by others. Study 2 also tests whether the selective willingness for people to excuse their own immoral actions, but not the immoral actions of others, explains the shift in agreement with the moral principles. The preregistration for Study 2 is available on OSF (https://osf.io/sb96tf).

Materials and methods

Participants

A total of 1,510 individuals completed this study via Prolific for monetary compensation. Participant recruitment was restricted to individuals in the United States and the United Kingdom. One hundred and ninety-three participants were excluded for failing a check question (see below for details). As such, we were left with 1,317 participants (Mage = 35 years, SD = 14, age range: 18–80, 794 females, 499 males). We determined the overall sample size based on the sample size (per cell) in Study 1. We aimed to recruit at least 1,500 to ensure that we would have at least 100 participants per cell in our experimental design (after expected exclusions). All participants reported being fluent English speakers. Informed consent was obtained from each participant in accordance with protocol approved by the Duke University Campus Institutional Review Board.

Materials

Four moral principles were in this study: (1) People should be honest with other people; (2) people should not harm other people; (3) people should be loyal to other people; and (4) people should show respect to other people. Note that principles (1) and (2) were used in Study 1, but principles (3) and (4) are new to Study 2. We included these two new moral principles to obtain more variety in our set of moral principles, as we expect our hypothesized effects to generalize across all moral principles.

Procedure

This study was self-paced and consisted of a single session. Participants were randomly presented with one of the four moral principles, and they were asked to make an initial judgment about how much they agreed with the principle (1 = strongly disagree, 7 = strongly agree). Participants then completed an unrelated 1-min categorization distractor task, after which they were randomly assigned to the actor condition, the recipient condition, or the observer condition. The instructions for each condition were the same as in Study 1. As before, participants described the memory in 3–5 sentences and then reported the month and year that the event occurred. Using a 7-point scale (1 = definitely no, 7 = definitely yes), participants then indicated whether they believed that the circumstances surrounding the remembered event excused the moral principle violator’s action.

On the next page, participants were asked whether they were able to generate a memory of an action that they believe to be morally wrong in accordance with the cue provided. Participants who reported that they did not generate a memory in accordance with the cue were excluded from our analyses (a total of 171 participants reported that they did not generate a memory in accordance with the cue). Specifically, 82%, 91%, and 97% of participants reported being able to recall a morally wrong violation of the principle in the actor, recipient, and observer conditions, respectively.

Participants then completed an unrelated 1-minute categorization distractor task, after which they were presented with the same moral principle again. Participants were instructed to make a final judgment about how much they agreed or disagreed with the principle on the same 7-point scale (1 = strongly disagree, 7 = strongly agree). At the end, participants answered the same attention check question as in Study 1. Upon completion, participants were monetarily compensated for their time.

Statistical analyses

The statistical software and methods used in Study 1 were also used in Study 2. To address possible mediation effects, using the ‘mediation’ package in R (Tingley et al., 2014), we also computed the average causal mediation effect (ACME), or the indirect effect, and the proportion of the effect mediated.

Results

We tested our central hypothesis that, after recalling an event in which they personally violated a moral principle, participants would agree less strongly with the violated principle—relative to participants who recalled moral principle violations committed by others. To this end, we computed a LMEM with condition (actor, recipient, observer) on the change in agreement ratings (final rating minus initial rating). The moral principle was the random effect in the model (random intercepts only). Participants in the actor condition shifted principle agreement ratings (toward disagreeing from agreeing) more than participants in the recipient (b = .20, SE = .03, t = 5.72, p < .001, 95% CI [.13, .26]) and observer conditions (b = .22, SE = .03, t = 6.28, p < .001, 95% CI [.14, .29]); however, there was no difference in the change in agreement ratings between recipient and observer conditions (p = .55). Figure 2 graphically depicts change in agreement ratings for each individual moral principle as a function of experimental condition. See Supplemental Information for descriptive statistics and analyses with each principle taken separately.

Fig. 2
figure 2

Mean excused judgments (a) and change in moral principle agreement (final judgment minus initial judgment) (b) for each moral principle as a function of condition. Error bars indicate 95% confidence intervals. Light-grey points represent individual participant responses evenly jittered

Next, we investigated whether participants are more likely to excuse their own moral principle violations, relative to the moral principle violations of others. To this end, we computed a LMEM with condition (actor, recipient, observer) on excused judgments. The moral principle was the random effect in the model. Participants in the actor condition were more likely to report that their moral principle violations were excused than participants in the recipient (b = 1.31, SE = .10, t = 12.95, p < .001, 95% CI [1.12, 1.50]) and observer conditions (b = 1.43, SE = .10, t = 14.08, p < .001, 95% CI [1.24, 1.63]); however, there was no difference in excused judgments between the recipient and observer conditions (p = .22). Figure 2 graphically depicts excused judgments for each individual principle as a function of experimental condition.

Finally, we tested whether the effect of condition on the change in moral principle agreement judgments is mediated by beliefs about the transgressor being excused. To this end, we computed two mediation models within a LMEM framework. In the first model, only actor and recipient conditions were included. Beliefs about the transgressor being excused mediated the effect of condition (actor vs. recipient) on the change in agreement judgments (ACME = .07, p < .001, 95% CI [.03, .10]; Prop. Mediated = .31, p < .001, 95% CI [.16, .58]). In the second model, only actor and observer conditions were included. Beliefs about the transgressor being excused mediated the effect of condition (actor vs. observer) on the change in agreement ratings (ACME = .07, p < .001, 95% CI [.04, .10]; Prop. Mediated = .34, p < .001, 95% CI [.20, .56]).

General discussion

Across two studies, we investigated whether recalling certain past events situationally shifts agreement with commonly endorsed moral principles. Both studies corroborate our central hypothesis: relative to participants who recalled an instance in which another person violated a moral principle, when participants recalled an instance in which they personally violated a moral principle, they agreed less strongly, on average, with the very principle they violated. Interestingly, most participants continued to agree with the moral principles after recalling their own violations, but they agreed less strongly than participants who recalled violations committed by others, regardless of whether the participants themselves were the victims of such violations. In other words, participants did not tend to reject the moral principles after recalling their own violations, but they did hold those principles with less conviction than other participants who recalled other people’s moral violations. In Study 2, we then found that participants were more willing to excuse their own remembered violations than others’ violations. This selective willingness for participants to excuse their own violations, but not the violations of others (whether or not they were the victims), statistically accounted for the observed shifts in moral principle agreement.

Our findings are compatible with the literature on moral disengagement, but they do diverge from this past research in important ways. Moral disengagement mechanisms (e.g., distorting the consequences of actions, dehumanizing victims) help people to convince themselves that their actions are permissible and that their ethical standards need not apply in certain contexts (Bandura, 1999; Bandura et al., 1996; Detert et al., 2008). These disengagement mechanisms are thought to help people to protect their favorable views of themselves. Note that convincing oneself that a particular action is morally acceptable in a particular context via moral disengagement entails maintaining the same level of agreement with the overarching moral principles; the principle just does not apply in some particular context. In contrast, our findings suggest that by reflecting on their own morally objectionable actions, people’s agreement with the overarching, guiding principles changes. It is not that the principle does not apply; it is that the principle is held with less conviction.

Empirical research on human morality has largely focused on how and why individuals make certain moral judgments and decisions using hypothetical vignettes (e.g., Cushman et al., 2006; Greene et al., 2001; Nichols & Mallon, 2006; Henne et al., 2016; Henne & Sinnott-Armstrong, 2018; Stanley et al., 2018; Valdesolo & DeSteno, 2006). These studies have produced valuable insights into how people make moral judgments and decisions. However, people rarely, if ever, experience events that are identical, or even similar, to the contrived dilemmas that have dominated the moral psychology literature (FeldmanHall et al., 2012; Hofmann et al., 2014). In fact, recent research has suggested that judgments and decisions made in hypothetical dilemmas do not predict how people make real-life moral judgments and decisions (Bostyn et al., 2018). A key advantage of the current research is the use of ecologically valid and personally relevant memories of real-life transgressions committed by the participants and by others.

Normative ethical theories and religious traditions that offer general moral principles are meant to help us to understand aspects of ourselves and our world in ways that offer insights and guidance for living a moral life (Albertzart, 2013; Väyrynen, 2008). Our findings introduce some cause for doubt about the stability of moral principles over time, and therefore, their reliability as accurate indicators of moral judgments and actions in the real world. Future research should investigate whether and to what extent the effects identified in our studies generalize to other kinds of moral violations. For example, while our effects are clear and consistent for harm, honesty, and loyalty violations, people may be less willing to excuse violations that elicit strong affective responses like disgust (violations of purity and sanctity norms)—regardless of who commits the transgression.