Newcomb's paradox
   HOME

TheInfoList



OR:

In philosophy and mathematics, Newcomb's paradox, also known as Newcomb's problem, is a
thought experiment A thought experiment is a hypothetical situation in which a hypothesis, theory, or principle is laid out for the purpose of thinking through its consequences. History The ancient Greek ''deiknymi'' (), or thought experiment, "was the most anc ...
involving a game between two players, one of whom is able to predict the future. Newcomb's paradox was created by William Newcomb of the
University of California The University of California (UC) is a public land-grant research university system in the U.S. state of California. The system is composed of the campuses at Berkeley, Davis, Irvine, Los Angeles, Merced, Riverside, San Diego, San Franci ...
's
Lawrence Livermore Laboratory Lawrence Livermore National Laboratory (LLNL) is a federal research facility in Livermore, California, United States. The lab was originally established as the University of California Radiation Laboratory, Livermore Branch in 1952 in respons ...
. However, it was first analyzed in a philosophy paper by Robert Nozick in 1969 and appeared in the March 1973 issue of ''
Scientific American ''Scientific American'', informally abbreviated ''SciAm'' or sometimes ''SA'', is an American popular science magazine. Many famous scientists, including Albert Einstein and Nikola Tesla, have contributed articles to it. In print since 1845, it ...
'', in
Martin Gardner Martin Gardner (October 21, 1914May 22, 2010) was an American popular mathematics and popular science writer with interests also encompassing scientific skepticism, micromagic, philosophy, religion, and literatureespecially the writings of Lew ...
's "
Mathematical Games A mathematical game is a game whose rules, strategies, and outcomes are defined by clear mathematical parameters. Often, such games have simple rules and match procedures, such as Tic-tac-toe and Dots and Boxes. Generally, mathematical games ne ...
". Reprinted with an addendum and annotated bibliography in his book ''The Colossal Book of Mathematics'' (). Today it is a much debated problem in the philosophical branch of
decision theory Decision theory (or the theory of choice; not to be confused with choice theory) is a branch of applied probability theory concerned with the theory of making decisions based on assigning probabilities to various factors and assigning numerical ...
.


The problem

There is a reliable predictor, another player, and two boxes designated A and B. The player is given a choice between taking only box B or taking both boxes A and B. The player knows the following: * Box A is transparent and always contains a visible $1,000. * Box B is opaque, and its content has already been set by the predictor: ** If the predictor has predicted that the player will take both boxes A and B, then box B contains nothing. ** If the predictor has predicted that the player will take only box B, then box B contains $1,000,000. The player does not know what the predictor predicted or what box B contains while making the choice.


Game-theory strategies

In his 1969 article, Nozick noted that "To almost everyone, it is perfectly clear and obvious what should be done. The difficulty is that these people seem to divide almost evenly on the problem, with large numbers thinking that the opposing half is just being silly." The problem continues to divide philosophers today. In a 2020 survey, a modest plurality of professional philosophers chose to two-box (39.0% versus 31.2%). Game theory offers two strategies for this game that rely on different principles: the
expected utility The expected utility hypothesis is a popular concept in economics that serves as a reference guide for decisions when the payoff is uncertain. The theory recommends which option rational individuals should choose in a complex situation, based on the ...
principle and the
strategic dominance In game theory, strategic dominance (commonly called simply dominance) occurs when one strategy is better than another strategy for one player, no matter how that player's opponents may play. Many simple games can be solved using dominance. The o ...
principle. The problem is called a ''paradox'' because two analyses that both sound intuitively logical give conflicting answers to the question of what choice maximizes the player's payout. * Considering the expected utility when the probability of the predictor being right is almost certain or certain, the player should choose box B. This choice statistically maximizes the player's winnings, setting them at about $1,000,000 per game. * Under the dominance principle, the player should choose the strategy that is ''always'' better; choosing both boxes A and B will ''always'' yield $1,000 more than only choosing B. However, the expected utility of "always $1,000 more than B" depends on the statistical payout of the game; when the predictor's prediction is almost certain or certain, choosing both A and B sets player's winnings at about $1,000 per game. David Wolpert and
Gregory Benford Gregory Benford (born January 30, 1941) is an American science fiction author and astrophysicist who is professor emeritus at the Department of Physics and Astronomy at the University of California, Irvine. He is a contributing editor of ''Reas ...
point out that paradoxes arise when not all relevant details of a problem are specified, and there is more than one "intuitively obvious" way to fill in those missing details. They suggest that in the case of Newcomb's paradox, the conflict over which of the two strategies is "obviously correct" reflects the fact that filling in the details in Newcomb's problem can result in two different noncooperative games, and each of the strategies is reasonable for one game but not the other. They then derive the optimal strategies for both of the games, which turn out to be independent of the predictor's infallibility, questions of causality, determinism, and free will.


Causality and free will

Causality issues arise when the predictor is posited as infallible and incapable of error; Nozick avoids this issue by positing that the predictor's predictions are "''almost'' certainly" correct, thus sidestepping any issues of infallibility and causality. Nozick also stipulates that if the predictor predicts that the player will choose randomly, then box B will contain nothing. This assumes that inherently random or unpredictable events would not come into play anyway during the process of making the choice, such as
free will Free will is the capacity of agents to choose between different possible courses of action unimpeded. Free will is closely linked to the concepts of moral responsibility, praise, culpability, sin, and other judgements which apply only to ac ...
or quantum mind processes. However, these issues can still be explored in the case of an infallible predictor. Under this condition, it seems that taking only B is the correct option. This analysis argues that we can ignore the possibilities that return $0 and $1,001,000, as they both require that the predictor has made an incorrect prediction, and the problem states that the predictor is never wrong. Thus, the choice becomes whether to take both boxes with $1,000 or to take only box B with $1,000,000 so taking only box B is always better.
William Lane Craig William Lane Craig (born August 23, 1949) is an American analytic philosopher, Christian apologist, author and Wesleyan theologian who upholds the view of Molinism and neo-Apollinarianism. He is Professor of Philosophy at Houston Baptist ...
has suggested that, in a world with perfect predictors (or time machines, because a time machine could be used as a mechanism for making a prediction),
retrocausality Retrocausality, or backwards causation, is a concept of cause and effect in which an effect precedes its cause in time and so a later event affects an earlier one. In quantum physics, the distinction between cause and effect is not made at the most ...
can occur. The chooser's choice can be said to have ''caused'' the predictor's prediction. Some have concluded that if time machines or perfect predictors can exist, then there can be no
free will Free will is the capacity of agents to choose between different possible courses of action unimpeded. Free will is closely linked to the concepts of moral responsibility, praise, culpability, sin, and other judgements which apply only to ac ...
and choosers will do whatever they are fated to do. Taken together, the paradox is a restatement of the old contention that free will and determinism are incompatible, since determinism enables the existence of perfect predictors. Put another way, this paradox can be equivalent to the
grandfather paradox A temporal paradox, time paradox, or time travel paradox is a paradox, an apparent contradiction, or logical contradiction associated with the idea of time and time travel. The notion of time travel to the future complies with current understanding ...
; the paradox presupposes a perfect predictor, implying the "chooser" is not free to choose, yet simultaneously presumes a choice can be debated and decided. This suggests to some that the paradox is an artifact of these contradictory assumptions.
Gary Drescher Gary L. Drescher is a scientist in the field of artificial intelligence (AI), and author of multiple books on AI, including ''Made-Up Minds: A Constructivist Approach to Artificial Intelligence''. His book describes a theory of how a computer p ...
argues in his book ''Good and Real'' that the correct decision is to take only box B, by appealing to a situation he argues is analogous a rational agent in a deterministic universe deciding whether or not to cross a potentially busy street. Andrew Irvine argues that the problem is structurally isomorphic to
Braess's paradox Braess's paradox is the observation that adding one or more roads to a road network can slow down overall traffic flow through it. The paradox was discovered by the German mathematician Dietrich Braess in 1968. The paradox may have analogies in ...
, a non-intuitive but ultimately non-paradoxical result concerning equilibrium points in physical systems of various kinds. Simon Burgess has argued that the problem can be divided into two stages: the stage before the predictor has gained all the information on which the prediction will be based and the stage after it. While the player is still in the first stage, they are presumably able to influence the predictor's prediction, for example, by committing to taking only one box. So players who are still in the first stage should simply commit themselves to one-boxing. Burgess readily acknowledges that those who are in the second stage should take both boxes. As he emphasises, however, for all practical purposes that is beside the point; the decisions "that determine what happens to the vast bulk of the money on offer all occur in the first tage. So players who find themselves in the second stage without having already committed to one-boxing will invariably end up without the riches and without anyone else to blame. In Burgess's words: "you've been a bad boy scout"; "the riches are reserved for those who are prepared". Burgess has stressed that ''pace'' certain critics (e.g., Peter Slezak) he does not recommend that players try to trick the predictor. Nor does he assume that the predictor is unable to predict the player's thought process in the second stage. Quite to the contrary, Burgess analyses Newcomb's paradox as a common cause problem, and he pays special attention to the importance of adopting a set of unconditional probability values whether implicitly or explicitly that are entirely consistent at all times. To treat the paradox as a common cause problem is simply to assume that the player's decision and the predictor's prediction have a common cause. (That common cause may be, for example, the player's brain state at some particular time before the second stage begins.) It is also notable that Burgess highlights a similarity between Newcomb's paradox and the
Kavka's toxin puzzle Kavka's toxin puzzle is a thought experiment about the possibility of forming an intention to perform an act which, following from reason, is an action one would not actually perform. It was presented by moral and political philosopher Gregory S. ...
. In both problems one can have a reason to intend to do something without having a reason to actually do it. Recognition of that similarity, however, is something that Burgess actually credits to Andy Egan.


Consciousness and simulation

Newcomb's paradox can also be related to the question of machine consciousness, specifically if a perfect
simulation A simulation is the imitation of the operation of a real-world process or system over time. Simulations require the use of models; the model represents the key characteristics or behaviors of the selected system or process, whereas the s ...
of a person's brain will generate the consciousness of that person. Suppose we take the predictor to be a machine that arrives at its prediction by simulating the brain of the chooser when confronted with the problem of which box to choose. If that simulation generates the consciousness of the chooser, then the chooser cannot tell whether they are standing in front of the boxes in the real world or in the virtual world generated by the simulation in the past. The "virtual" chooser would thus tell the predictor which choice the "real" chooser is going to make, and the chooser, not knowing whether they are the real chooser or the simulation, should take only the second box.


Fatalism

Newcomb's paradox is related to logical fatalism in that they both suppose absolute certainty of the future. In logical fatalism, this assumption of certainty creates circular reasoning ("a future event is certain to happen, therefore it is certain to happen"), while Newcomb's paradox considers whether the participants of its game are able to affect a predestined outcome..


Extensions to Newcomb's problem

Many thought experiments similar to or based on Newcomb's problem have been discussed in the literature. For example, a quantum-theoretical version of Newcomb's problem in which box B is entangled with box A has been proposed.


The meta-Newcomb problem

Another related problem is the meta-Newcomb problem. The setup of this problem is similar to the original Newcomb problem. However, the twist here is that the predictor may elect to decide whether to fill box B after the player has made a choice, and the player does not know whether box B has already been filled. There is also another predictor: a "meta-predictor" who has reliably predicted both the players and the predictor in the past, and who predicts the following: "Either you will choose both boxes, and the predictor will make its decision after you, or you will choose only box B, and the predictor will already have made its decision." In this situation, a proponent of choosing both boxes is faced with the following dilemma: if the player chooses both boxes, the predictor will not yet have made its decision, and therefore a more rational choice would be for the player to choose box B only. But if the player so chooses, the predictor will already have made its decision, making it impossible for the player's decision to affect the predictor's decision.


See also

*
Decision theory Decision theory (or the theory of choice; not to be confused with choice theory) is a branch of applied probability theory concerned with the theory of making decisions based on assigning probabilities to various factors and assigning numerical ...
**
Causal decision theory Causal decision theory (CDT) is a school of thought within decision theory which states that, when a rational agent is confronted with a set of possible actions, one should select the action which causes the best outcome in expectation. CDT contras ...
**
Evidential decision theory Evidential decision theory (EDT) is a school of thought within decision theory which states that, when a rational agent is confronted with a set of possible actions, one should select the action with the highest ''news value'', that is, the action w ...
** Alternatives to causal and evidential decision theory * Roko's basilisk


Notes


References

* * Campbell, Richmond and Sowden, Lanning, ed. (1985), ''Paradoxes of Rationality and Cooperation: Prisoners' Dilemma and Newcomb's Problem'', Vancouver: University of British Columbia Press. (an anthology discussing Newcomb's Problem, with an extensive bibliography). * Collins, John
"Newcomb's Problem"
International Encyclopedia of the Social and Behavioral Sciences, Neil Smelser and Paul Baltes (eds.), Elsevier Science (2001). * * (An article discussing the popularity of Newcomb's problem.) {{Paradoxes Philosophical paradoxes Thought experiments Mathematical paradoxes Prediction 1969 introductions Decision-making paradoxes