Newcomb's Paradox: The Problem That Splits Smart People

Quick Summary
Newcomb's paradox divides brilliant minds 50/50. Should you take one box or two? Explore the decision theory, free will, and rationality behind this famous puzzle.
In This Article
The Box Problem That Starts Arguments Every Time
Imagine walking into a room with a supercomputer and two boxes. One box is open, and you can clearly see $1,000 inside. The other is sealed. You're told that the supercomputer — which has a near-perfect track record of predicting human behaviour — has already made a guess about what you'll do. If it predicted you'd take only the sealed box, it placed $1 million inside. If it predicted you'd take both boxes, the sealed one is empty.
Related Post
What do you do?
This is Newcomb's paradox, a thought experiment devised by physicist William Newcomb in 1960 and later popularised by philosopher Robert Nozick. Decades after its invention, it still cleaves intelligent people almost perfectly down the middle. A 2016 Guardian poll of over 31,000 readers found 53.5% would take only the mystery box, while 46.5% would grab both. What's remarkable isn't just that people disagree — it's that each side thinks the other is being obviously foolish.
If you've never encountered this problem before, hold your instinct for a moment. Because whichever answer feels obvious to you, there's a rigorous philosophical tradition backing the other side.
The Case for Taking One Box
One-boxers anchor their argument in outcomes. Look at the evidence: the supercomputer has correctly predicted thousands of people before you. Almost everyone who took only the mystery box walked away with a million dollars. Almost everyone who grabbed both boxes walked away with $1,000.
This reasoning is formalised in what philosophers call evidential decision theory (EDT). EDT says you should choose the action that, given the available evidence, is most strongly associated with a good outcome. The historical data is unambiguous — one-boxing correlates almost perfectly with being rich.
To put numbers to it: if the computer's accuracy rate is anything above roughly 50.05%, the expected monetary value of one-boxing exceeds that of two-boxing. If the computer is correct 99% of the time, a one-boxer can expect around $990,000 on average, while a two-boxer can expect just under $11,000. The math isn't close.
There's also a personality dimension here. The kind of person who genuinely commits to one-boxing — who internalises it as a value rather than a strategy — is precisely the kind of person the predictor would identify as a one-boxer. You don't just choose one box. You are a one-boxer, and the supercomputer already knew that before you walked through the door.
The Case for Taking Two Boxes
Two-boxers aren't being greedy or naive. Their argument is philosophically clean and hard to dismiss: the boxes are already set up. Whatever is in that sealed box is already in there. Your decision right now cannot reach back in time and change it. So why would you leave $1,000 sitting on the table?
This is the logic of causal decision theory (CDT), which says rational agents should choose actions based on what they can cause to happen, not what their choice merely correlates with. You cannot cause the million to appear by wishing hard enough. The causal chain is already closed.
Philosophers call this reasoning strategic dominance. Map out every possible scenario: if the box is empty, two-boxing gets you $1,000 and one-boxing gets you nothing. If the box contains a million, two-boxing gets you $1,001,000 and one-boxing gets you $1,000,000. In every single scenario, two-boxing leaves you $1,000 better off. A dominated strategy — one that is worse across all scenarios — should never be chosen. And one-boxing is the dominated strategy.
Two-boxers would also point out that your deliberation cannot retroactively influence a prediction that was made before you even knew about the problem. Believing otherwise, they argue, is a form of magical thinking — what some philosophers bluntly call "wishful thinking dressed up as probability."
Why Both Sides Are Technically Correct
Here's what makes Newcomb's paradox genuinely interesting rather than just a pub debate: both sides are using internally consistent, well-supported frameworks. EDT and CDT are not fringe theories — they are serious, competing approaches to rational decision-making, each with respected defenders in academic philosophy.
The disagreement isn't really about the boxes. It's about what rationality itself means.
Continue Reading
Related Guides
Keep exploring this topic
EDT asks: given everything I know, which choice is most strongly associated with a good result? CDT asks: which choice can I actually cause to produce a better result? In normal life, these frameworks converge — rational actions tend to cause good outcomes and correlate with them. Newcomb's paradox is specifically engineered to drive them apart, which is why it feels so uncomfortable.
Philosophers Allan Gibbard and William Harper argued in a landmark 1978 paper that causal decision theory is correct and two-boxing is rational — but they candidly acknowledged that two-boxers will consistently end up poorer. They concluded that the game is simply "rigged" to reward what they considered irrational behaviour. Critics, understandably, found this unsatisfying. If your theory of rationality reliably produces worse real-world outcomes, that's a serious strike against the theory.
The "Why Ain'cha Rich?" Problem
This brings us to one of the sharpest challenges in the whole debate, known colloquially as the "Why Ain'cha Rich?" argument. If two-boxing is the rational choice, why do two-boxers consistently walk away with a fraction of what one-boxers earn?
The two-boxer's honest response is that rationality and success are not the same thing. A rational agent playing a rigged game can still lose. The game has been constructed so that predictable rationality is punished and a kind of strategic irrationality — or pre-commitment — is rewarded. That's a feature of the game's design, not a flaw in the reasoning.
But this response has limits. In the real world, we don't usually have the luxury of dismissing bad outcomes as someone else's fault. If a strategy reliably produces worse results across thousands of trials, there's a reasonable argument that it's simply the wrong strategy — regardless of how clean its internal logic looks.
This maps directly onto the prisoner's dilemma, another classic of game theory. In a single-round prisoner's dilemma, the dominant strategy is always to defect — to betray your partner. But in repeated rounds, cooperation consistently outperforms defection. A society of cooperators thrives; a society of rational defectors collapses. Rationality at the individual level and rationality at the collective level can be mutually exclusive. Newcomb's paradox lives in exactly that gap.
What This Tells Us About Free Will and Determinism
Newcomb's paradox has a philosophical sting in its tail that goes beyond money. If the supercomputer is truly near-perfect — and especially if you extend the thought experiment to imagine a 100%-accurate predictor — then you run into a deeply uncomfortable question: does free will exist?
If a predictor can know with certainty what you will choose before you make that choice, in what sense are you choosing freely? Many two-boxers implicitly assume some form of libertarian free will — the belief that your deliberation is genuinely open, that you could choose differently in the moment, and that therefore your choice now cannot have been fully determined in advance.
One-boxers, on the other hand, tend to be more comfortable with the idea that human behaviour is, to a large degree, predictable — that our choices flow from our characters, our histories, our neural architecture. The supercomputer isn't magic; it's just very, very good at reading the kind of person you are.
This isn't merely academic. Debates about free will and determinism have real consequences in law, ethics, and public policy. If behaviour is highly predictable and shaped by factors outside our control, how should we think about moral responsibility? The answer shapes everything from criminal sentencing to how we design incentive structures in organisations.
Of course, even hard determinists tend to agree on a pragmatic point: whether or not free will exists in some ultimate metaphysical sense, we are compelled to act as if it does. Our social systems, our relationships, and our sense of self are built on that assumption. You still have to choose a box.
How to Actually Win Newcomb's Game
There is, in theory, one way to beat the system and walk away with $1,001,000: convince the predictor you're a one-boxer, then grab both boxes at the last second. But here's the catch — if the predictor is good enough, it already knows you're the kind of person who would try that. The deception is itself predicted.
This leads to three legitimate scenarios where even committed two-boxers might rationally switch:
Recommended
Top picks for this topic
reMarkable 2 Paper Tablet
The closest thing to writing on paper. Zero distractions. The best tool for thinkers and note-takers.
Logitech MX Keys S Keyboard
Smart illuminated keyboard for multi-device power users. Backlit, comfortable, works across Mac/PC/mobile.
As an affiliate, Zeebrain may earn a commission on qualifying purchases at no extra cost to you. We only recommend products we genuinely stand behind.
Free Weekly Newsletter
Enjoying this guide?
Get the best articles like this one delivered to your inbox every week. No spam.
Pre-commitment: If you could make a binding, credible promise to the predictor before it makes its prediction — and if your word is reliably bankable — then one-boxing becomes the rational choice under CDT as well, because you are now causally influencing the predictor's decision.
Repeated trials: Over multiple rounds, your past choices build a track record. A two-boxer who switches to one-boxing establishes a new pattern the predictor will eventually recognise. One-boxing becomes instrumentally rational across the long run.
Retrocausality: This is the wild card. If the predictor works by observing the future — through some hypothetical mechanism like a wormhole — then your present choice does causally influence what was placed in the box. In that case, CDT itself recommends one-boxing.
None of these loopholes are available in the standard formulation of the paradox. But they're instructive, because they reveal exactly which assumptions are doing the heavy lifting in each position.
Conclusion: The Paradox Is the Point
Newcomb's paradox isn't a trick question with a hidden correct answer waiting to be revealed. It's a carefully constructed lens that exposes the fault lines in how we think about rationality, causation, and agency. The fact that it splits intelligent, thoughtful people almost evenly isn't a sign that one side hasn't thought hard enough. It's a sign that we hold genuinely different intuitions about what it means to make a good decision.
If you're a one-boxer, you probably trust evidence, embrace probabilistic thinking, and are comfortable with the idea that your character shapes your choices in ways you don't fully control. If you're a two-boxer, you likely prize causal clarity, resist letting correlation override causation, and hold onto a robust sense of deliberative agency.
Both of those orientations have strengths. Both have costs. The boxes are already set up — but the choice you make says something real about how you navigate a world that is simultaneously driven by causes and saturated with meaning.
Choose wisely. Or don't. The supercomputer already knows.
Frequently Asked Questions
What is Newcomb's paradox, in simple terms?
Newcomb's paradox is a decision theory puzzle where you must choose between taking one sealed box or two boxes (one sealed, one containing $1,000). A highly accurate predictor has already placed either $1 million or nothing in the sealed box based on whether it predicted you'd take one box or two. The paradox is that two equally valid lines of reasoning — evidential and causal decision theory — recommend opposite choices.
Is there a definitively correct answer to Newcomb's paradox?
No. Newcomb's paradox is a genuine philosophical dilemma, not a trick question. Evidential decision theory supports taking one box, while causal decision theory supports taking both. Both frameworks are internally consistent and have serious academic defenders. The paradox is valuable precisely because it reveals a deep tension in our understanding of rationality itself.
What is the difference between evidential and causal decision theory?
Evidential decision theory (EDT) says you should choose the action most strongly associated with a good outcome, based on available evidence. Causal decision theory (CDT) says you should choose the action that will most effectively cause a good outcome. In everyday decisions these usually agree, but Newcomb's paradox is engineered to force them apart — EDT recommends one-boxing, CDT recommends two-boxing.
Does Newcomb's paradox have anything to do with free will?
Yes, directly. If a predictor can accurately forecast your choice before you make it, this raises serious questions about whether your deliberation is truly free. Two-boxers tend to assume genuine open-ended agency; one-boxers are often more comfortable with the idea that choices are deeply shaped — and therefore predictable — by character and history. The paradox doesn't resolve the free will debate, but it makes the stakes of that debate concrete and financially legible.
How does Newcomb's paradox relate to the prisoner's dilemma?
Both puzzles reveal a conflict between individual rationality and collective or long-term rationality. In the prisoner's dilemma, the individually rational move is always to defect, but mutual defection produces worse outcomes than mutual cooperation. Similarly, in Newcomb's paradox, the causally rational move (two-boxing) consistently produces worse financial outcomes than the evidentially rational move (one-boxing). Both problems show that optimising for local, individual logic can undermine broader or longer-term success.
Frequently Asked Questions
The Box Problem That Starts Arguments Every Time
Imagine walking into a room with a supercomputer and two boxes. One box is open, and you can clearly see $1,000 inside. The other is sealed. You're told that the supercomputer — which has a near-perfect track record of predicting human behaviour — has already made a guess about what you'll do. If it predicted you'd take only the sealed box, it placed $1 million inside. If it predicted you'd take both boxes, the sealed one is empty.
What do you do?
This is Newcomb's paradox, a thought experiment devised by physicist William Newcomb in 1960 and later popularised by philosopher Robert Nozick. Decades after its invention, it still cleaves intelligent people almost perfectly down the middle. A 2016 Guardian poll of over 31,000 readers found 53.5% would take only the mystery box, while 46.5% would grab both. What's remarkable isn't just that people disagree — it's that each side thinks the other is being obviously foolish.
If you've never encountered this problem before, hold your instinct for a moment. Because whichever answer feels obvious to you, there's a rigorous philosophical tradition backing the other side.
The Case for Taking One Box
One-boxers anchor their argument in outcomes. Look at the evidence: the supercomputer has correctly predicted thousands of people before you. Almost everyone who took only the mystery box walked away with a million dollars. Almost everyone who grabbed both boxes walked away with $1,000.
This reasoning is formalised in what philosophers call evidential decision theory (EDT). EDT says you should choose the action that, given the available evidence, is most strongly associated with a good outcome. The historical data is unambiguous — one-boxing correlates almost perfectly with being rich.
To put numbers to it: if the computer's accuracy rate is anything above roughly 50.05%, the expected monetary value of one-boxing exceeds that of two-boxing. If the computer is correct 99% of the time, a one-boxer can expect around $990,000 on average, while a two-boxer can expect just under $11,000. The math isn't close.
There's also a personality dimension here. The kind of person who genuinely commits to one-boxing — who internalises it as a value rather than a strategy — is precisely the kind of person the predictor would identify as a one-boxer. You don't just choose one box. You are a one-boxer, and the supercomputer already knew that before you walked through the door.
The Case for Taking Two Boxes
Two-boxers aren't being greedy or naive. Their argument is philosophically clean and hard to dismiss: the boxes are already set up. Whatever is in that sealed box is already in there. Your decision right now cannot reach back in time and change it. So why would you leave $1,000 sitting on the table?
This is the logic of causal decision theory (CDT), which says rational agents should choose actions based on what they can cause to happen, not what their choice merely correlates with. You cannot cause the million to appear by wishing hard enough. The causal chain is already closed.
Philosophers call this reasoning strategic dominance. Map out every possible scenario: if the box is empty, two-boxing gets you $1,000 and one-boxing gets you nothing. If the box contains a million, two-boxing gets you $1,001,000 and one-boxing gets you $1,000,000. In every single scenario, two-boxing leaves you $1,000 better off. A dominated strategy — one that is worse across all scenarios — should never be chosen. And one-boxing is the dominated strategy.
Two-boxers would also point out that your deliberation cannot retroactively influence a prediction that was made before you even knew about the problem. Believing otherwise, they argue, is a form of magical thinking — what some philosophers bluntly call "wishful thinking dressed up as probability."
Why Both Sides Are Technically Correct
Here's what makes Newcomb's paradox genuinely interesting rather than just a pub debate: both sides are using internally consistent, well-supported frameworks. EDT and CDT are not fringe theories — they are serious, competing approaches to rational decision-making, each with respected defenders in academic philosophy.
The disagreement isn't really about the boxes. It's about what rationality itself means.
EDT asks: given everything I know, which choice is most strongly associated with a good result? CDT asks: which choice can I actually cause to produce a better result? In normal life, these frameworks converge — rational actions tend to cause good outcomes and correlate with them. Newcomb's paradox is specifically engineered to drive them apart, which is why it feels so uncomfortable.
Philosophers Allan Gibbard and William Harper argued in a landmark 1978 paper that causal decision theory is correct and two-boxing is rational — but they candidly acknowledged that two-boxers will consistently end up poorer. They concluded that the game is simply "rigged" to reward what they considered irrational behaviour. Critics, understandably, found this unsatisfying. If your theory of rationality reliably produces worse real-world outcomes, that's a serious strike against the theory.
The "Why Ain'cha Rich?" Problem
This brings us to one of the sharpest challenges in the whole debate, known colloquially as the "Why Ain'cha Rich?" argument. If two-boxing is the rational choice, why do two-boxers consistently walk away with a fraction of what one-boxers earn?
The two-boxer's honest response is that rationality and success are not the same thing. A rational agent playing a rigged game can still lose. The game has been constructed so that predictable rationality is punished and a kind of strategic irrationality — or pre-commitment — is rewarded. That's a feature of the game's design, not a flaw in the reasoning.
But this response has limits. In the real world, we don't usually have the luxury of dismissing bad outcomes as someone else's fault. If a strategy reliably produces worse results across thousands of trials, there's a reasonable argument that it's simply the wrong strategy — regardless of how clean its internal logic looks.
This maps directly onto the prisoner's dilemma, another classic of game theory. In a single-round prisoner's dilemma, the dominant strategy is always to defect — to betray your partner. But in repeated rounds, cooperation consistently outperforms defection. A society of cooperators thrives; a society of rational defectors collapses. Rationality at the individual level and rationality at the collective level can be mutually exclusive. Newcomb's paradox lives in exactly that gap.
What This Tells Us About Free Will and Determinism
Newcomb's paradox has a philosophical sting in its tail that goes beyond money. If the supercomputer is truly near-perfect — and especially if you extend the thought experiment to imagine a 100%-accurate predictor — then you run into a deeply uncomfortable question: does free will exist?
If a predictor can know with certainty what you will choose before you make that choice, in what sense are you choosing freely? Many two-boxers implicitly assume some form of libertarian free will — the belief that your deliberation is genuinely open, that you could choose differently in the moment, and that therefore your choice now cannot have been fully determined in advance.
One-boxers, on the other hand, tend to be more comfortable with the idea that human behaviour is, to a large degree, predictable — that our choices flow from our characters, our histories, our neural architecture. The supercomputer isn't magic; it's just very, very good at reading the kind of person you are.
This isn't merely academic. Debates about free will and determinism have real consequences in law, ethics, and public policy. If behaviour is highly predictable and shaped by factors outside our control, how should we think about moral responsibility? The answer shapes everything from criminal sentencing to how we design incentive structures in organisations.
Of course, even hard determinists tend to agree on a pragmatic point: whether or not free will exists in some ultimate metaphysical sense, we are compelled to act as if it does. Our social systems, our relationships, and our sense of self are built on that assumption. You still have to choose a box.
How to Actually Win Newcomb's Game
There is, in theory, one way to beat the system and walk away with $1,001,000: convince the predictor you're a one-boxer, then grab both boxes at the last second. But here's the catch — if the predictor is good enough, it already knows you're the kind of person who would try that. The deception is itself predicted.
This leads to three legitimate scenarios where even committed two-boxers might rationally switch:
Pre-commitment: If you could make a binding, credible promise to the predictor before it makes its prediction — and if your word is reliably bankable — then one-boxing becomes the rational choice under CDT as well, because you are now causally influencing the predictor's decision.
Repeated trials: Over multiple rounds, your past choices build a track record. A two-boxer who switches to one-boxing establishes a new pattern the predictor will eventually recognise. One-boxing becomes instrumentally rational across the long run.
Retrocausality: This is the wild card. If the predictor works by observing the future — through some hypothetical mechanism like a wormhole — then your present choice does causally influence what was placed in the box. In that case, CDT itself recommends one-boxing.
None of these loopholes are available in the standard formulation of the paradox. But they're instructive, because they reveal exactly which assumptions are doing the heavy lifting in each position.
Conclusion: The Paradox Is the Point
Newcomb's paradox isn't a trick question with a hidden correct answer waiting to be revealed. It's a carefully constructed lens that exposes the fault lines in how we think about rationality, causation, and agency. The fact that it splits intelligent, thoughtful people almost evenly isn't a sign that one side hasn't thought hard enough. It's a sign that we hold genuinely different intuitions about what it means to make a good decision.
If you're a one-boxer, you probably trust evidence, embrace probabilistic thinking, and are comfortable with the idea that your character shapes your choices in ways you don't fully control. If you're a two-boxer, you likely prize causal clarity, resist letting correlation override causation, and hold onto a robust sense of deliberative agency.
Both of those orientations have strengths. Both have costs. The boxes are already set up — but the choice you make says something real about how you navigate a world that is simultaneously driven by causes and saturated with meaning.
Choose wisely. Or don't. The supercomputer already knows.
Frequently Asked Questions
What is Newcomb's paradox, in simple terms?
Newcomb's paradox is a decision theory puzzle where you must choose between taking one sealed box or two boxes (one sealed, one containing $1,000). A highly accurate predictor has already placed either $1 million or nothing in the sealed box based on whether it predicted you'd take one box or two. The paradox is that two equally valid lines of reasoning — evidential and causal decision theory — recommend opposite choices.
Is there a definitively correct answer to Newcomb's paradox?
No. Newcomb's paradox is a genuine philosophical dilemma, not a trick question. Evidential decision theory supports taking one box, while causal decision theory supports taking both. Both frameworks are internally consistent and have serious academic defenders. The paradox is valuable precisely because it reveals a deep tension in our understanding of rationality itself.
What is the difference between evidential and causal decision theory?
Evidential decision theory (EDT) says you should choose the action most strongly associated with a good outcome, based on available evidence. Causal decision theory (CDT) says you should choose the action that will most effectively cause a good outcome. In everyday decisions these usually agree, but Newcomb's paradox is engineered to force them apart — EDT recommends one-boxing, CDT recommends two-boxing.
Does Newcomb's paradox have anything to do with free will?
Yes, directly. If a predictor can accurately forecast your choice before you make it, this raises serious questions about whether your deliberation is truly free. Two-boxers tend to assume genuine open-ended agency; one-boxers are often more comfortable with the idea that choices are deeply shaped — and therefore predictable — by character and history. The paradox doesn't resolve the free will debate, but it makes the stakes of that debate concrete and financially legible.
How does Newcomb's paradox relate to the prisoner's dilemma?
Both puzzles reveal a conflict between individual rationality and collective or long-term rationality. In the prisoner's dilemma, the individually rational move is always to defect, but mutual defection produces worse outcomes than mutual cooperation. Similarly, in Newcomb's paradox, the causally rational move (two-boxing) consistently produces worse financial outcomes than the evidentially rational move (one-boxing). Both problems show that optimising for local, individual logic can undermine broader or longer-term success.
About Zeebrain Editorial
Our editorial team is dedicated to providing clear, well-researched, and high-utility content for the modern digital landscape. We focus on accuracy, practicality, and insights that matter.
More from Science & Tech
Explore More Categories
Keep browsing by topic and build depth around the subjects you care about most.


