This Paradox Splits Smart People 50/50
Based on Veritasium's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.
Newcomb’s paradox splits people because evidential decision theory treats the predictor’s accuracy as decision-relevant evidence, while causal decision theory treats only causal influence as decision-relevant.
Briefing
Newcomb’s paradox—where a near-perfect predictor offers a choice between taking one “mystery” box or taking both a mystery box plus $1,000—splits people almost evenly because two reasonable ways of calculating “what you should do” treat correlation and causation differently. The setup is simple: the predictor has already placed $1,000,000 in the mystery box if it expects you to take only that box; if it expects you to take both, it leaves the mystery box empty. Since the prediction happens before you enter and your decision can’t change what’s already been arranged, the puzzle becomes a test of how people reason under uncertainty.
One camp—often called “one-boxers”—leans on evidential decision theory: the predictor’s past accuracy is evidence about what’s in the mystery box. If the predictor is highly reliable, then choosing the one-box option is strongly correlated with walking away with $1,000,000. In that view, expected value calculations favor one-boxing once the predictor’s accuracy is even slightly above random. The argument is essentially: “My choice is a signal of what the predictor already expects, and the predictor’s track record makes that signal meaningful.”
The other camp—“two-boxers”—leans on causal decision theory: the decision you make now can’t cause the contents of the already-set boxes to change. With that assumption, the mystery box’s contents are fixed, so taking both boxes guarantees an extra $1,000 on top of whatever the mystery box contains. In expected-utility terms, two-boxing dominates one-boxing because it adds $1,000 without reducing the chance of getting the $1,000,000. The paradox persists because both camps treat different probability models as the “right” one: one-boxers use probabilities tied to evidence about the predictor’s accuracy; two-boxers use probabilities tied to what the choice can causally influence.
The disagreement isn’t just math—it points to deeper questions about rationality and free will. If a predictor is perfect, then the only way to get the $1,001,000 outcome would be to be the kind of person who one-boxes in advance, then switch at the last second. That raises whether free will exists in a way that can matter once prediction is already locked in. The discussion connects this to “Why Ain’tcha Rich?”—the observation that if one-boxing is the rational strategy for maximizing money, then one-boxers should systematically win more—and to philosopher arguments (including Gibbard and Harper) that sometimes “irrationality” can be rewarded when prediction is strong.
Finally, the conversation reframes the paradox as a problem about rules, not moment-to-moment choices. If people could pre-commit—through reputation across repeated trials, through mechanisms that let choices affect the past, or through explicit commitments—then the “worse” option (one-boxing) could become the best rule to live by. The same logic is compared to deterrence and game-theory commitments in Cold War “assured destruction” (MAD): stability can depend on committing to an outcome that you hope never occurs. The core takeaway is that Newcomb’s paradox isn’t meaningless; it forces clarity about whether decisions should track evidence or causation, and what “rational” behavior means when your action is entangled with prediction.
Cornell Notes
Newcomb’s paradox asks whether to take only a “mystery” box or to take both the mystery box and a guaranteed $1,000 when a supercomputer has already predicted the choice. One-boxers use evidential decision theory: the predictor’s accuracy is evidence that the mystery box contains $1,000,000 if they choose one-box. Two-boxers use causal decision theory: the choice can’t change what was already placed in the boxes, so taking both always adds $1,000 without reducing the mystery-box payoff. The split reveals a deeper conflict between treating correlation as decision-relevant evidence versus treating only causal influence as decision-relevant. It also links to free will, rationality, and the value of pre-commitment.
Why do one-boxers think one-boxing can be optimal even though the boxes are already set?
Why do two-boxers say two-boxing dominates one-boxing?
What hidden assumption divides the camps: evidence or causation?
How does the paradox connect to free will?
Why does pre-commitment matter, and how does it change the “best rule”?
Review Questions
- In Newcomb’s paradox, what changes in the expected-utility calculation when switching from evidential decision theory to causal decision theory?
- What does it mean to say that two-boxing “dominates” one-boxing in the causal framework, and why doesn’t that settle the paradox for one-boxers?
- How do pre-commitment, repeated trials, and reputation potentially reconcile the “best rule” with the last-moment choice?
Key Points
- 1
Newcomb’s paradox splits people because evidential decision theory treats the predictor’s accuracy as decision-relevant evidence, while causal decision theory treats only causal influence as decision-relevant.
- 2
One-boxing can win in expected value when the predictor’s correctness probability is slightly above random, because one-boxing is correlated with the $1,000,000 outcome.
- 3
Two-boxing can dominate in the causal framework because the choice can’t change the already-set contents, so taking both adds $1,000 without reducing the mystery-box payoff.
- 4
The paradox highlights a real distinction between correlation and causation: the camps disagree on whether correlation should drive the probability model used for decisions.
- 5
The disagreement connects to free will: if prediction is perfect and happens before the choice, the choice may be unable to affect what was already arranged.
- 6
The discussion reframes the puzzle as a rules problem: pre-commitment (via reputation, repeated trials, or explicit commitments) can make the “worse” last-moment option the best long-run rule.
- 7
Deterrence strategies like MAD illustrate how committing in advance to an outcome can stabilize a system even when the committed action is undesirable.