I agree that AI deterrence will necessarily fail if:
All AI’s modify themselves to ignore threats from all agents (including ones it considers irrational), and
any deterrence simulation counts as a threat.
Why do you believe that both or either of these statements are true? Do you have some concrete definition of ‘threat’ in mind?
I don’t believe statement 1 and don’t see why it’s required. After all, we are quite rational, and so is our future FAI.
I agree that AI deterrence will necessarily fail if:
All AI’s modify themselves to ignore threats from all agents (including ones it considers irrational), and
any deterrence simulation counts as a threat.
Why do you believe that both or either of these statements are true? Do you have some concrete definition of ‘threat’ in mind?
I don’t believe statement 1 and don’t see why it’s required. After all, we are quite rational, and so is our future FAI.