you might accidentally realize that such-and-such type of agent will threaten you regardless of what you commit to and then if you are a coward you will “give in” by making an exception for that agent.
this seems like a problem for humans and badly-built AIs. Nothing that reliably one-boxes should ever do this.
Or do you mean one-boxing in Transparent Newcomb? Then your claim might be true, but even then it depends on how seriously we take the “regardless of what you commit to” clause.
True, sorry, I forgot the whole set of paradoxes that led up to FDT/UDT. I mean something like… “this is equivalent to the problem that FDT/UDT already has to solve anyways.” Allowing you to make exceptions doesn’t make your job harder.
I concur in general, but:
this seems like a problem for humans and badly-built AIs. Nothing that reliably one-boxes should ever do this.
EDT reliably one-boxes, but EDT would do this.
Or do you mean one-boxing in Transparent Newcomb? Then your claim might be true, but even then it depends on how seriously we take the “regardless of what you commit to” clause.
True, sorry, I forgot the whole set of paradoxes that led up to FDT/UDT. I mean something like… “this is equivalent to the problem that FDT/UDT already has to solve anyways.” Allowing you to make exceptions doesn’t make your job harder.