To me, a “vanilla” Prisoner’s Dilemma involves actual human prisoners who may reason about their partners. I don’t mean to imply that I think “standard” PD involves credible pre-commitments nor perfect knowledge of the opponent. While I agree that in standard PD there’s no causal connection between actions, there can be logical connections between actions that make for interesting strategies (eg if you expect them to use TDT).
On this point, I’m inclined to think that we agree and are debating terminology.
“Aggravating” may have been too strong a word; “disappointed” might have been better
That’s even worse! :-)
I readily admit that my presentation is tailored to my personality, and I understand how others may find it grating.
That said, a secondary goal of this post was to instill doubt in concepts that look sacred (terminal goals, epistemic rationality) and encourage people to consider that even these may be sacrificed for instrumental gains.
It seems you already grasp the tradeoffs between epistemic and instrumental rationality and that you can consistently reach mental states that are elusive to naive epistemically rational agents, and that you’ve come to these conclusions by a different means than I. By my analysis, there are many others who need a push before they are willing to even consider “terminal goals” and “false beliefs” as strategic tools. This post caters more to them.
I’d be very interested to hear more about how you’ve achieved similar results with different techniques!
To me, a “vanilla” Prisoner’s Dilemma involves actual human prisoners who may reason about their partners. I don’t mean to imply that I think “standard” PD involves credible pre-commitments nor perfect knowledge of the opponent. While I agree that in standard PD there’s no causal connection between actions, there can be logical connections between actions that make for interesting strategies (eg if you expect them to use TDT).
On this point, I’m inclined to think that we agree and are debating terminology.
That’s even worse! :-)
I readily admit that my presentation is tailored to my personality, and I understand how others may find it grating.
That said, a secondary goal of this post was to instill doubt in concepts that look sacred (terminal goals, epistemic rationality) and encourage people to consider that even these may be sacrificed for instrumental gains.
It seems you already grasp the tradeoffs between epistemic and instrumental rationality and that you can consistently reach mental states that are elusive to naive epistemically rational agents, and that you’ve come to these conclusions by a different means than I. By my analysis, there are many others who need a push before they are willing to even consider “terminal goals” and “false beliefs” as strategic tools. This post caters more to them.
I’d be very interested to hear more about how you’ve achieved similar results with different techniques!