I think that if you want to work on AI takeover prevention, you should probably have some models of the situation that give you a rough prediction for P(AI takeover). For example, I’m constantly thinking in terms of the risks conditional on scheming arising, scheming not arising, a leading AI company being responsible and having three years of lead, etc.
I think that if you want to work on AI takeover prevention, you should probably have some models of the situation that give you a rough prediction for P(AI takeover). For example, I’m constantly thinking in terms of the risks conditional on scheming arising, scheming not arising, a leading AI company being responsible and having three years of lead, etc.