My entire understanding of FDT and how it varies from CDT or EDT is based on Eliezer and Nate’s paper that I linked to, so I will only comment on what I think they mean. If someone was complaining that “Omega favors irrationality” and expanded that statement to “The choice they reward favors irrational persons”, I would make the argument that rewarding people based on the choice/action they take is not an “unfair” scenario.
Some excerpts from FDT:
The standard defense of two-boxing is that Newcomb’s problem rewards irrationality. Indeed, it is always possible to construct dilemmas that reward bad decision theories. As an example, we can imagine a decision rule that says to pick the action that comes earliest in the alphabet (under some canonical encoding). In most dilemmas, this rule does poorly; but the rule fares well in scenarios where a reliable predictor rewards exactly the agents that follow this rule, and punishes everyone else. The causal decision theorist can argue that Newcomb’s problem is similarly constructed to reward EDT and FDT agents. [...] Newcomb’s predictor is not filling boxes according to how the agent arrives at a decision; she is only basing her action on a prediction of the decision itself. While it is appropriate to call dilemmas unfair when they directly reward or punish agents for their decision procedures, we deny that there is anything unfair about rewarding or punishing agents for predictions about their actions. It is one thing to argue that agents have no say in what decision procedure they implement, and quite another thing to argue that agents have no say in what action they output.
So that’s the “cannon” take on “Omega favors irrationality”.
On the second half of your comment:
I’d agree that you one-boxing would be evidence that Omega had given you the Million $. If you went through Newcomb’s problem using “Of the actions I could take, which one would provide the most Bayesian evidence that I was going to get lots of utility?” you would one-box. However, that process is not the decision process specified by CDT. We’re now talking about EDT (which as this comment points out, is sorta wonky).
If that doesn’t feel satisfactory, and it still feels like CDT would one-box, it might be useful to explicitly go through the exact decision process that you think CDT prescribes.
My entire understanding of FDT and how it varies from CDT or EDT is based on Eliezer and Nate’s paper that I linked to, so I will only comment on what I think they mean. If someone was complaining that “Omega favors irrationality” and expanded that statement to “The choice they reward favors irrational persons”, I would make the argument that rewarding people based on the choice/action they take is not an “unfair” scenario.
Some excerpts from FDT:
So that’s the “cannon” take on “Omega favors irrationality”.
On the second half of your comment:
I’d agree that you one-boxing would be evidence that Omega had given you the Million $. If you went through Newcomb’s problem using “Of the actions I could take, which one would provide the most Bayesian evidence that I was going to get lots of utility?” you would one-box. However, that process is not the decision process specified by CDT. We’re now talking about EDT (which as this comment points out, is sorta wonky).
If that doesn’t feel satisfactory, and it still feels like CDT would one-box, it might be useful to explicitly go through the exact decision process that you think CDT prescribes.