The thrust of your argument is that an agent that uses causal decision theory will defect in a one-shot Prisoner’s Dilemma.
You specify CDT when you say that
No matter what Agent_02 does, actually implementing Action_X would bear no additional value
because this implies Agent_01 looks at the causal effects of do(Action_X) and decides what to do based solely on them. Prisoner’s Dilemma because Action_X corresponds to Cooperate, and not(Action_X) to Defect, with an implied Action_Y that Agent_02 could perform that is of positive utility to Agent_01 (hence, ‘trade’). One-shot because without causal interaction between the agents, they can’t update their beliefs.
That CDT using agents unconditionally defect in the one-shot PD is old news. That you should defect against CDT using agents in the one-shot PD is also old news. So your post rather gives the impression that you haven’t done the research on the decision theories that make acausal trade interesting as a concept.
The thrust of your argument is that an agent that uses causal decision theory will defect in a one-shot Prisoner’s Dilemma.
You specify CDT when you say that
because this implies Agent_01 looks at the causal effects of do(Action_X) and decides what to do based solely on them. Prisoner’s Dilemma because Action_X corresponds to Cooperate, and not(Action_X) to Defect, with an implied Action_Y that Agent_02 could perform that is of positive utility to Agent_01 (hence, ‘trade’). One-shot because without causal interaction between the agents, they can’t update their beliefs.
That CDT using agents unconditionally defect in the one-shot PD is old news. That you should defect against CDT using agents in the one-shot PD is also old news. So your post rather gives the impression that you haven’t done the research on the decision theories that make acausal trade interesting as a concept.