Different interpretations as to
why the predictor is so accurate yield different rational solutions to the problem. Some such interpretations are illustrated in this section with diagrams in a convention similar to Mackie's 1977 paper "Newcomb's Paradox and the Direction of Causation", using white squares for events whose outcome is "open" to the player's rational deliberation, black squares for events whose outcome is not open to deliberation, and arrows for causal connections. In the diagrams, references to a choice being "free" refer to the choice being open to influence by decision-theoretical reasoning about the rational choice in the decision problem, not necessarily in some broader "
free will" sense,
William Lane Craig has suggested that, in a world with perfect predictors (or
time machines, because a time machine could be used as a mechanism for making a prediction),
retrocausality can occur. The chooser's choice can be said to have
caused the predictor's prediction. Some have concluded that if time machines or perfect predictors can exist, then there can be no
free will and choosers will do whatever they are fated to do. Taken together, the paradox is a restatement of the old contention that free will and
determinism are incompatible, since determinism enables the existence of perfect predictors. Put another way, this paradox can be equivalent to the
grandfather paradox; the paradox presupposes a perfect predictor, implying the "chooser" is not free to choose, yet simultaneously presumes a choice can be debated and decided. This suggests to some that the paradox is an artifact of these contradictory assumptions. If the predictor is posited as
infallible and incapable of error, it can seem that it is impossible for there to be a true answer as to what it is most reasonable to
choose, as there is no true choice, and the player's decision is predetermined by the psychological facts surveyed by the predictor. For instance, if the predictor has such an accurate track record because (f) he is a capable hypnotist who manipulates all choosers to one-box or two-box in accordance with whether he chose to put the money in the opaque box, then it would be reasonable to two-box if one could, but by construction, one is not free to. Similarly, if the predictor has an infallibly correct psychology (g), then the player's choice cannot be determined by a strategist's recommendation of what is most reasonable to do, as it is fixed in advance by his psychological character. Any solution then becomes moot. This interpretation of Newcomb's paradox is related to
logical fatalism in that they both suppose absolute certainty of the future. In logical fatalism, this assumption of certainty creates
circular reasoning ("a future event is certain to happen, therefore it is certain to happen"), while Newcomb's paradox considers whether the participants of its game are able to affect a predestined outcome.
Character formation However, the problem can be thought of as a situation (d) where, although the predictor is infallible, the player is free to develop their own character at some point ahead of the problem situation so as to determine what the predictor will predict. Under this condition, it seems that taking only B is the correct option. This analysis argues that we can ignore the possibilities that return $0 and $1,001,000, as they both require that the predictor has made an incorrect prediction, and the problem states that the predictor is never wrong. Thus, the choice becomes whether to take both boxes with $1,000 or to take only box B with $1,000,000 so taking only box B is always better. An example of this interpretation is
Gary Drescher, who argues in his book
Good and Real that the correct decision is to take only box B, by appealing to a situation he argues is analogous a rational agent in a deterministic universe deciding whether or not to cross a potentially busy street. Such a rational agent chooses to cross only if their reasoning process indicates it is safe, because the kind of decision procedure they run tends to correlate with situations where crossing is safe—even in a deterministic universe where their choice does not causally change the traffic. Alternatively, however, if (e) the predictor can be systematically fooled, then the player should determine their character so as to always fool the predictor. As Mackie says, "the best character of all to develop, if it were possible, is one which would fool these psychologist-seers — the player should appear to be a closed-box-only-taker, and should perhaps start by intending to take the closed box only, but then change his mind at the last minute and take both boxes; if the game is to be played repeatedly, he must appear at the start of each new game to be a reformed character who, despite his former lapses, will take only the closed box this time, and yet he must in the end yield to temptation again and take both boxes after all." So players who find themselves in the second stage without having already committed to one-boxing will invariably end up without the riches and without anyone else to blame. In Burgess's words: "you've been a bad boy scout"; "the riches are reserved for those who are prepared". Burgess has stressed that
pace certain critics (e.g., Peter Slezak) he does not recommend that players try to trick the predictor. Nor does he assume that the predictor is unable to predict the player's thought process in the second stage. Quite to the contrary, Burgess analyses Newcomb's paradox as a common cause problem, and he pays special attention to the importance of adopting a set of unconditional probability values whether implicitly or explicitly that are entirely consistent at all times. To treat the paradox as a common cause problem is simply to assume that the player's decision and the predictor's prediction have a common cause. (That common cause may be, for example, the player's brain state at some particular time before the second stage begins.) Burgess highlights a similarity between Newcomb's paradox and the
Kavka's toxin puzzle. Recognition of the similarity, however, is something that Burgess credits to Andy Egan. In both problems one can have a reason to intend to do something without having a reason to actually do it. Similarly, both problems may ask whether it is possible to truly intend or commit to do something and later on still choose not to do so. An answer to this by Gibbard & Harper is that, if given the opportunity beforehand, the rational agent should adopt the disposition to take only the opaque box, and this disposition must be stable, that is, the agent should actually one-box at the moment of choice, not merely beforehand.
No connection at all Finally, it is possible to suppose that the predictor's past accuracy came from other factors which are now causally unconnected with the player's move, such as (h) very lucky sheer coincidence, or (i) a form of hypnosis which the player is able, in some rare cases, to break free from, or (j) a form of highly correct psychology which nevertheless it is possible, in some rare cases, to fool. The indeterminacy of such a causal connection is signified by a dashed arrow. In those situations, it is reasonable to make the two-boxing choice, but simply much more difficult to do so when there is money in the opaque box. == Analogies and extensions ==