I don’t think
This gets at the deeper issue: how do we represent an action having a consequence?
I don’t think it’s
In the Prisoner’s Dilemma tournament, you don’t respond to my action, you respond to what you can prove about my action, so it’s really
But I think that even in single player games, we should think of the environment as containing an implementation of us which takes an action when it can prove it’s the “right” action, so even then it’s provability that has consequences.
This may seem unnatural, but that’s just because considered decisions are unnatural, and mostly we just do stuff. But I think it’s a good model of thinking with a notebook open and plenty of time. That is, our model is that the consequences in the environment trace back to the existence of a proof in Peano Arithmetic.
My only reservation really is that I think maybe it should be
I wanted to share this post on social media about a half hour ago (so, like 7 or 7:30 am in California?), but couldn’t load the LessWrong frontpage:
Just curious now what the issue was; it would be kind of funny, although I suppose only to me, if it was an issue introduced by the AI workflow you describe here.