A brief note on Newcomb's problem
I’ve seen many claims lately that one-boxing is obviously correct in a way that impugns the skill of any philosopher who would doubt it. Although I am a one-boxer, I think treating two-boxers as defective is unfair. Much better philosophers than I have been two boxers. Because I have been irritated by this discourse, I wanted to briefly review the case for two boxing to show it’s actually pretty strong.
Here's a quick refresher on the problem: In front of you are two boxes: a red box and a blue box. Either the red box contains ten thousand dollars, and the blue box contains one million dollars, or the red box contains ten thousand dollars and the blue box contains nothing. You have to choose whether to take the blue box and the red box or just the blue box.
An oracle who can see the future placed the money in the boxes. If the oracle sees that you will pick both boxes, it will put only 10,000 dollars in the red box. If it sees that you will only take the blue box, it will put 10,000 dollars in the red box and one million dollars in the blue box.
Should you take both boxes or just the blue box? Personally, I would take just the blue box. I would then open it to find it contains a million dollars. Great! Surely, anyone who took both boxes is stupid.
No, I don’t think so, though they will end up less rich. Here’s why it’s not stupid to take both boxes. Either the money is already in the blue box and the red box, or it is only in the red box. Your picking which box(es) to take has no effect on this. It’s already done.
And so comes the fundamental argument for taking both boxes. Either the money is in the blue box and the red box, or it is just in the red box. If it is in both boxes, picking both boxes will get you more money. If it is only in one box, the only way you’ll get any money at all is to pick both boxes. So, whatever the state of the world, picking both boxes results in a better outcome. The principle that a choice is correct if, however, things turn out to be, it yields the best outcome is called choice dominance. It is intuitively a very persuasive principle. Not picking both boxes violates choice dominance.
Consider a common variant. Suppose the boxes were transparent. Suppose that you could already see if each contained money. It was evident that no magic was happening at the time of picking the boxes. You could literally see that the money was already in the red box alone, or both boxes. Would you then not choose to take the red box as well? Why deny yourself 10,000 dollars when you can see the matter is already resolved? This is, I think, where much of the “easy” case for one boxing comes from- the false sense that because the boxes are opaque, things “haven’t happened” until the boxes are open.
Consider another variant. Suppose instead of a supernatural oracle, the money-placer is just a really good psychologist who gets it right 99% of the time. Nothing supernatural is involved. Here again, now that we have removed certainty and the supernatural element, many people are more inclined to take both boxes.
Now personally I am a one boxer, but I don’t think it’s an easy matter. In fact, my own tentative guess is some kind of Humean relativism about decision theory is right. Reason only tells you what will happen as a result of various actions, the causal effects your actions will have, etc. Whether you ‘should’ follow CDT, FDT, EDT, etc. is set by what strategy appeals to you, perhaps by the precise formulation of your desires, etc because reason really only gives us the positive facts and how our actions will alter them and not what we should do.
Deriving categorical imperatives from reason is impossible, but even a practical reason of hypothetical imperatives arising from the facts and our desires is ultimately just a way of talking. Most of the time the idea of practical reason seems to make sense because what it means to desire X constrains what a thinking agent who desires X must do in certain circumstances- if: (1) you desire X, (2) there is a strategy that will get you X with no downsides and (3) You’re aware of this, but you do not do employ the strategy then you do not really desire X.
However, these definitional constraints on what it means to desire something are not enough to decide between one boxing and two boxing. Reason can only tell us what is, what will, and the causal relations between them, and in a case like Newcomb’s problem, this, combined with a desire for money, doesn’t constrain us to act in a certain way. Seeking X, all else being equal is written into the concept of wanting X, but the exact resolution of what “seeking” in a context like this means- where evidence, causation, etc. come apart— is not part of our concept of wanting. What matters is if we want the money in a one-boxer or a two-boxer way, and all reason can do is tell us what will happen and why if we take either option.



The problem with choice dominance is that it implicitly assumes that the CDT method of assessing counterfactuals is correct. If you instead use the FDT method of assessing counterfactuals, for example, then choice dominance would favor one-boxing in both Newcomb's problem and the Transparent Newcomb Problem. So really, a better name for choice dominance would be CDT choice dominance, or causal choice dominance. Since FDT/functional choice dominance also exists, you do not need to give up on choice dominance in order to support one-boxing; instead, you can simply change which version of choice dominance you use. Thus, this post's defense of two-boxers is much weaker than it first appears.
Note that the above is essentially just a restatement/summary of an argument from the paper Functional Decision Theory: A New Theory of Instrumental Rationality rather than something original to me.
https://arxiv.org/pdf/1710.05060#page=19
Maybe I'm honest, or maybe I'm simple, but I think people *correctly* predict each other all the time in everyday life. Label the boxes "excellent employment" and "*not* embezzling everything in front of you" and I think the common-sense case for one-boxing in *transparent* Newcomb's becomes much clearer.
The big prize box is visibly full because you were predicted to be the kind of person who could consistently leave the small prize alone. The big prize is visibly *empty* because -- well, as one of the MIRI folks put it (Soares, I think?), decisions are for making bad outcomes inconsistent. It *isn't* empty, not for me. And when it is, despite that? I'll take the 1% failure rate and go home empty-handed, over turning around and becoming a two-boxer *in every situation like this*.
Evidence and causation are right there, you described the entire setup in a single paragraph. It just is not that complicated. You do need something that can look spooky to sophilists -- a *logical* (not causal) connection between your actions and other people's models of your actions -- but it's not hard to define. If CDT ignores that kind of relation and therefore predictably loses money, so much the worse for CDT. I think this is only complicated to philosophers who've tied themselves up in knots over it. Given the setup there is an action that reliably and predictably wins, and if you can't philosophically justify taking it, that is a failure in your philosophy. *Justifying* the tendency to lose, or waffling between them -- that's just sad. You can do better. Your is/ought intervention just denies the setup to sidestep the problem, and the setup can be patched until it doesn't (assume you want money, assume linear utility in dollars, assume prizes other than money, etc etc etc until you *have to* engage with the question, Least Convenient Possible World style.)
I dunno, you're right that the dominance principle is not to be discarded lightly. But I think I can safely do so, given that maximizing the outcomes for the situation I'm in *affects the situations I end up in* when other people can *see you doing it* and contribute to your situation. I don't think it has to be complex.
("But there's no way people could" go play smash or street fighter or competitive pokémon with your local ranked champions, you will get your *soul read* like your innermost plans are a cheap billboard ad. Everyone who believes they have a libertarian / unpredictable free will should try this some time. Mixing it up is a skill that takes time and practice and can't be perfect.)
("But I don't *want* to embezzle/steal/exploit" yes evolution and culture have had you internalize this, in everyday situations far away from abstract philosophy.)