Why Doing Nothing Feels Less Wrong Than Doing Something
"Subjects judged the active poisoning of a tennis opponent to be more immoral than simply failing to warn him of a suspect dish โ even with identical outcomes."
โ Spranca, Minsk & Baron (1991), Journal of Experimental Social Psychology
Imagine two people, both of whom cause the same harm. One actively does something to cause it. The other simply fails to prevent it. Who is more morally wrong?
Logically, if the outcomes are identical, the moral weight should be the same. But our minds don't work that way. We have a deep-seated omission bias: harmful actions feel more blameworthy than equally harmful inactions.
This bias affects real decisions: parents refuse vaccines (inaction) even when the disease risk is far higher than the vaccine risk (action). We'd rather let five people die than push one person to save them. Doing nothing feels safer โ even when it isn't.
You'll read 4 scenarios. Each has two versions with identical outcomes.
Rate how morally wrong each version is.
Omission bias was first systematically studied by Spranca, Minsk, and Baron (1991), who found that people consistently judge harmful actions as more immoral than equally harmful omissions.
Subjects read about John, who wants his tennis opponent to lose. In one version, John poisons his opponent's drink. In another, John simply fails to warn his opponent about a dish he notices looks suspicious. Both result in the opponent getting sick. Subjects rated the active poisoning as significantly more immoral โ even though John's intent and the outcome were identical.
Actions feel more directly causal. "I did this" carries more weight than "I let this happen" โ even when the causal chain is equally strong.
Inaction preserves the status quo. Interfering with the natural course of events feels like a greater transgression than allowing it to unfold.
It's harder to blame someone for something they didn't do. Our legal and moral systems focus on actions, making omissions feel less punishable.
Actions have known consequences we're responsible for. Inaction lets us hide behind "I couldn't have known" โ even when we could have.
Omission bias explains vaccine hesitancy: parents fear the action of vaccinating (1 in a million risk) more than the omission of not vaccinating (1 in 10,000 disease risk). The action risk feels "on their hands" while the omission risk feels like "nature taking its course."
Parents refuse vaccines (action) even when disease risk (from omission) is 100x higher. "If something goes wrong from the vaccine, it's my fault."
Most people won't push one person to save five, but would pull a lever. Pushing is direct action; pulling feels more like redirecting fate.
Murder is a crime; failing to save someone usually isn't. Our laws encode omission bias โ punishing commission far more than omission.
Doctors struggle more with actively ending life than with withdrawing treatment โ even when the patient outcome is identical.
Investors regret bad investments (action) more than missed opportunities (omission), leading to excessive caution and underperformance.
"Omission bias may reflect a general tendency to judge harmful actions as worse than harmful omissions, regardless of consequences."
โ Baron & Ritov (2004), Handbook of Judgment and Decision Making