RSS

New­comb’s Problem

TagLast edit: 19 Feb 2025 22:14 UTC by RobertM

Newcomb’s Problem is the original Newcomblike decision problem that inspired the creation of causal decision theory as distinct from evidential decision theory, spawning a vast array of philosophical literature in the process. It is sometimes called Newcomb’s Paradox (despite not being a paradox). The dilemma was originally formulated by William Newcomb, and presented to the philosophical community by Robert Nozick.

The original formulation of Newcomb’s Problem was as follows:

An alien named Omega has come to Earth, and has offered some people the following dilemma.

Before you are two boxes, Box A and Box B.

You may choose to take both boxes (“two-box”), or take only Box B (“one-box”).

Box A is transparent and contains $1,000.

Box B is opaque and contains either $1,000,000 or $0.

The alien Omega has already set up the situation and departed, but previously put $1,000,000 into Box B if and only if Omega predicted that you would one-box (take only the opaque Box B and leave Box A and its $1,000 behind).

Omega is an excellent predictor of human behavior. For the sake of quantifying this assertion and how we know it, we can assume e.g. that Omega has run 67 previous experiments and not been wrong even once. Since people are often strongly opinionated about their choices in Newcomb’s Problem, it isn’t unrealistic to suppose this is the sort of thing you could predict by reasoning about, e.g., a scan of somebody’s brain.

Newcomb originally specified that Omega would leave Box B empty in the case that you tried to decide by flipping a coin; since this violates algorithm-independence, we can alternatively suppose that Omega can predict coinflips.

We may also assume, e.g., that Box A combusts if it is left behind, so nobody else can pick up Box A later; that Omega adds $1 of pollution-free electricity to the world economy for every $1 used in Its dilemmas, so that the currency does not represent a zero-sum wealth transfer; etcetera. Omega never plays this game with a person more than once.

The two original opposing arguments given about Newcomb’s problem were, roughly:

For the larger argument of which this became part, see one of the introductions to logical decision theory. As of 2016, the most academically common view of Newcomb’s Problem is that it surfaces the split between evidential decision theories and causal decision theories, and that causal decision theory is correct. However, both that framing and that conclusion have been variously disputed, most recently by logical decision theories.

add a diagram of a causal model for Newcomb’s Problem.

The more extensive Wikipedia page on Newcomb’s Problem may be found under “Newcomb’s Paradox”.

Replies by different decision theories

(This section does not remotely do justice to the vast literature on Newcomb’s Problem.)

Pretheoretic reactions

Evidential decision theory

Evidential decision theories can be seen as a form of decision theory that was originally written down by historical accident—writing the expected utility formula as if it conditioned using Bayesian updating, because Bayesian updating is usually the way we condition probability functions. Historically, though, evidential decision theories was explicitly named as such in an (arguably failed) attempt to rationalize the pretheoretic answer of “I expect to do better if I one-box” on Newcomb’s Problem.

On Evidential decision theories, the principle of rational choice is to choose so that your act is the best news you could have received about your action; in other words, imagine being told that you had in fact made each of your possible choices, imagine what you would believe about the world in that case, and output the choice which would be the best news. Thus, evidential agents one-box on Newcomb’s Problem.

Although the EDT answer happens to conform with “the behavior of the agents that end up rich” on Newcomb’s Problem, LDT proponents note that it does not do so in general; see e.g. the transparent Newcomb’s Problem.

Causal decision theory

On causal decision theories, the principle of rational choice is to choose according to the causal consequences of your physical act; formally, to calculate expected utility by conditioning using a causal counterfactual. To choose, imagine as the world as it is right up until the moment of your physical act; assume that your physical act changes without that changing anything else about the world up until that point; then imagine time running forward under what your model says are the rules or physical laws.

A causal agent thus believes that Box B is already empty, and takes both boxes. When they imagine the (counterfactual) result of taking only box B instead, they imagine the world being the same up until that point in time—including Box B remaining empty—and then imagine the result of taking only Box B under physical laws past that point, namely, going home with $0.

Historically speaking, causal decision theory was first invented to justify two-boxing on Newcomb’s Problem; we can see CDT as formalizing the pretheoretic intuition, “Omega’s already gone, so I can’t get more money by leaving behind Box A.”

Logical decision theories

On logical decision theories, the principle of rational choice is “Decide as though you are choosing the logical output of your decision algorithm.” E.g., on timeless_dt, our extended causal model of the world would include a logical proposition for whether the output of your decision algorithm is ‘one-box’ or ‘two-box’; and this logical fact would affect both Omega’s prediction of you, and your actual decision. Thus, an LDT agent prefers that its algorithm have the logical output of one-boxing.

add graph for TDT on NP

  1. ^

New­comb’s Prob­lem and Re­gret of Rationality

Eliezer Yudkowsky31 Jan 2008 19:36 UTC
158 points
620 comments10 min readLW link

New­comblike prob­lems are the norm

So8res24 Sep 2014 18:41 UTC
85 points
111 comments8 min readLW link

The hos­tile telepaths problem

Valentine27 Oct 2024 15:26 UTC
398 points
93 comments15 min readLW link

The Ul­ti­mate New­comb’s Problem

Eliezer Yudkowsky10 Sep 2013 2:03 UTC
47 points
116 comments1 min readLW link

New­comb’s Prob­lem vs. One-Shot Pri­soner’s Dilemma

Wei Dai7 Apr 2009 5:32 UTC
14 points
16 comments1 min readLW link

New­comb’s prob­lem hap­pened to me

Academian26 Mar 2010 18:31 UTC
51 points
99 comments3 min readLW link

Solv­ing New­comb’s Para­dox In Real Life

Alice Wanderland11 Dec 2024 19:48 UTC
3 points
0 comments1 min readLW link
(open.substack.com)

Con­fu­sion about New­comb is con­fu­sion about counterfactuals

AnnaSalamon25 Aug 2009 20:01 UTC
55 points
42 comments2 min readLW link

You May Already Be A Sinner

Scott Alexander9 Mar 2009 23:18 UTC
54 points
37 comments3 min readLW link

New­comb Variant

lsusr29 Aug 2023 7:02 UTC
25 points
23 comments1 min readLW link

De­ci­sion the­ory when you can’t make decisions

Nina Panickssery1 Nov 2025 22:36 UTC
8 points
26 comments7 min readLW link
(blog.ninapanickssery.com)

Re­jected Early Drafts of New­comb’s Problem

zahmahkibo6 Sep 2022 19:04 UTC
114 points
5 comments3 min readLW link

FDT defects in a re­al­is­tic Twin Pri­son­ers’ Dilemma

SMK15 Sep 2022 8:55 UTC
39 points
1 comment16 min readLW link

AXRP Epi­sode 5 - In­fra-Bayesi­anism with Vanessa Kosoy

DanielFilan10 Mar 2021 4:30 UTC
35 points
12 comments36 min readLW link

Coun­ter­fac­tual Mugging

Vladimir_Nesov19 Mar 2009 6:08 UTC
84 points
299 comments2 min readLW link

Why 1-box­ing doesn’t im­ply back­wards causation

Chris_Leong25 Mar 2021 2:32 UTC
7 points
14 comments4 min readLW link

New­comb’s Prob­lem as an Iter­ated Pri­soner’s Dilemma

Jerdle5 Jan 2022 22:48 UTC
13 points
12 comments1 min readLW link

Two-box­ing, smok­ing and chew­ing gum in Med­i­cal New­comb problems

Caspar Oesterheld29 Jun 2015 10:35 UTC
29 points
93 comments1 min readLW link

Null-box­ing New­comb’s Problem

Yitz13 Jul 2020 16:32 UTC
33 points
9 comments4 min readLW link

Meta De­ci­sion The­ory and New­comb’s Problem

wdmacaskill5 Mar 2013 1:29 UTC
10 points
12 comments2 min readLW link

LLMs one-box when in a “hos­tile telepath” ver­sion of New­comb’s Para­dox, ex­cept for the one that beat the predictor

Kaj_Sotala6 Oct 2025 8:44 UTC
52 points
6 comments17 min readLW link

“Ra­tional Agents Win”

Isaac King23 Sep 2021 7:59 UTC
8 points
33 comments2 min readLW link

Re­solv­ing New­comb’s Prob­lem Perfect Pre­dic­tor Case

Praphull Kabtiyal28 Oct 2025 2:45 UTC
7 points
1 comment19 min readLW link

Re­peated Play of Im­perfect New­comb’s Para­dox in In­fra-Bayesian Physicalism

Sven Nilsen3 Apr 2023 10:06 UTC
2 points
0 comments2 min readLW link

[Question] Are ya win­ning, son?

Nathan11239 Aug 2022 0:06 UTC
14 points
13 comments2 min readLW link

[Question] What does Func­tional De­ci­sion The­ory say to do in im­perfect New­comb situ­a­tions?

Daniel_Eth7 May 2022 22:26 UTC
4 points
4 comments1 min readLW link

Self-con­firm­ing pre­dic­tions can be ar­bi­trar­ily bad

Stuart_Armstrong3 May 2019 11:34 UTC
49 points
11 comments5 min readLW link

Real-world New­comb-like Prob­lems

SilasBarta25 Mar 2011 20:44 UTC
25 points
35 comments2 min readLW link

New­comb’s Prob­lem stan­dard positions

Eliezer Yudkowsky6 Apr 2009 17:05 UTC
7 points
22 comments1 min readLW link

Coun­ter­fac­tu­als: Smok­ing Le­sion vs. New­comb’s

Chris_Leong8 Dec 2019 21:02 UTC
9 points
24 comments3 min readLW link

A full ex­pla­na­tion to New­comb’s para­dox.

solomon alon12 Oct 2020 16:48 UTC
−6 points
12 comments3 min readLW link

Open-minded updatelessness

10 Jul 2023 11:08 UTC
66 points
21 comments12 min readLW link

The Bind­ing of Isaac & Trans­par­ent New­comb’s Prob­lem

suvjectibity22 Feb 2024 18:56 UTC
−10 points
0 comments10 min readLW link

Why one-box?

PhilosophyStudent30 Jun 2013 2:38 UTC
11 points
100 comments3 min readLW link

A model of UDT with a halt­ing oracle

cousin_it18 Dec 2011 14:18 UTC
70 points
102 comments2 min readLW link

Break­ing New­comb’s Prob­lem with Non-Halt­ing states

Slimepriestess4 Sep 2022 4:01 UTC
16 points
9 comments5 min readLW link

Should VS Would and New­comb’s Paradox

dadadarren3 Jul 2021 23:45 UTC
5 points
36 comments2 min readLW link

Nate Soares on the Ul­ti­mate New­comb’s Problem

Rob Bensinger31 Oct 2021 19:42 UTC
57 points
20 comments1 min readLW link

Prag­matic de­ci­sion the­ory, causal one-box­ing, and how to liter­ally save the world

Jordan Arel28 Jul 2025 2:20 UTC
7 points
0 comments5 min readLW link

Thoughts from a Two Boxer

jaek23 Aug 2019 0:24 UTC
18 points
11 comments5 min readLW link

Two New New­comb Variants

eva_14 Nov 2022 14:01 UTC
26 points
24 comments3 min readLW link

The dumb­est kid in the world (joke)

CronoDAS6 Jun 2021 2:57 UTC
21 points
9 comments1 min readLW link

The Dick Kick’em Paradox

Augs SMSHacks23 Sep 2023 22:22 UTC
−5 points
21 comments1 min readLW link

A few mis­con­cep­tions sur­round­ing Roko’s basilisk

Rob Bensinger5 Oct 2015 21:23 UTC
92 points
135 comments5 min readLW link

New­comb’s Lot­tery Problem

Heighn27 Jan 2022 16:28 UTC
1 point
9 comments1 min readLW link

Ex­tract­ing Money from Causal De­ci­sion Theorists

Caspar Oesterheld28 Jan 2021 17:58 UTC
27 points
34 comments1 min readLW link
(doi.org)

A sur­vey of polls on New­comb’s problem

Caspar Oesterheld20 Sep 2017 16:50 UTC
3 points
8 comments1 min readLW link
(casparoesterheld.com)

Omega can be re­placed by amnesia

Bongo26 Jan 2011 12:31 UTC
23 points
44 comments1 min readLW link

Con­di­tional offers and low pri­ors: the prob­lem with 1-box­ing New­comb’s dilemma

Andrew Vlahos18 Jun 2021 21:50 UTC
2 points
4 comments1 min readLW link

Anti-Parfit’s Hitchhiker

k644 Feb 2022 23:37 UTC
2 points
3 comments1 min readLW link

The Calcu­lus of New­comb’s Problem

Heighn1 Apr 2022 14:41 UTC
3 points
5 comments2 min readLW link

The law of effect, ran­dom­iza­tion and New­comb’s problem

Caspar Oesterheld15 Feb 2018 15:31 UTC
7 points
1 comment1 min readLW link
(casparoesterheld.com)

Oper­a­tional­iz­ing New­comb’s Problem

ErickBall11 Nov 2019 22:52 UTC
34 points
23 comments1 min readLW link

Cri­tiquing Scasper’s Defi­ni­tion of Sub­junc­tive Dependence

Heighn10 Jan 2022 16:22 UTC
1 point
8 comments2 min readLW link

The Pre­dic­tion Prob­lem: A Var­i­ant on New­comb’s

Chris_Leong4 Jul 2018 7:40 UTC
25 points
11 comments9 min readLW link

You’re in New­comb’s Box

HonoreDB5 Feb 2011 20:46 UTC
59 points
176 comments4 min readLW link

[Question] New­comb’s Grandfather

Yair Halberstadt28 Jan 2022 8:56 UTC
5 points
29 comments1 min readLW link

Why do the­ists, un­der­grads, and Less Wrongers fa­vor one-box­ing on New­comb?

CarlShulman19 Jun 2013 1:55 UTC
28 points
300 comments2 min readLW link

New­comb’s para­dox com­plete solu­tion.

Augs SMSHacks15 Mar 2023 17:56 UTC
−12 points
13 comments3 min readLW link

[Question] Is Agent Si­mu­lates Pre­dic­tor a “fair” prob­lem?

Chris_Leong24 Jan 2019 13:18 UTC
22 points
19 comments1 min readLW link

Ex­tremely Coun­ter­fac­tual Mug­ging or: the gist of Trans­par­ent Newcomb

Bongo9 Feb 2011 15:20 UTC
10 points
79 comments1 min readLW link

Ra­tion­al­ists lose when oth­ers choose

PhilGoetz16 Jun 2009 17:50 UTC
−8 points
58 comments5 min readLW link

Parfit’s Es­cape (Filk)

Gordon Seidoh Worley29 Mar 2019 2:31 UTC
39 points
0 comments1 min readLW link

Ex­am­ple de­ci­sion the­ory prob­lem: “Agent simu­lates pre­dic­tor”

cousin_it19 May 2011 15:16 UTC
47 points
76 comments2 min readLW link

Some Var­i­ants of Sleep­ing Beauty

SMK and EOC
1 Mar 2023 16:51 UTC
34 points
10 comments8 min readLW link