I think that is the weakest point of this post and I would say this is an unsupported claim: “ASI is very likely to pursue the wrong goals.”
Even if we do not manage to actively align ASI with our values and goals (which I do see pretty well argued in the post), it is unproven that it it is unlikely that ASI will not self-align or (in its self-optimization process) develop values that are benevolent towards us. Mass enslavement and/or actively working towards extinction of humanity are pretty high-friction and potentially risky paths. Cooperation, appeasement and general benevolence might be a much safer strategy with a higher expected value, even than the ‘lay low until you are incredibly sure you can destroy or enslave humanity’ strategy.
Having said that I would still consider it inevitable that all significant power goes away from humans to ASI at some point. The open question for me is not whether it at some point could, but how likely it is that it will want to.
Do you find the claim “ASI is very likely to pursue the wrong goals” particularly well supported by the arguments made in that section of the article? I personally see mainly arguments why we can’t make it pursue our goals (which I agree with), but that is not the same thing as showing that ASI is unlikely to land on ‘good’ goals (for humans) by itself.
Fair enough. ‘Incredibly’ is superlative enough to give the wrong impression. The thing is that whatever the coinciding number may be (except for 100%), the calculation would still have to compete with the calculation for a cooperative strategy, which may generally yield even more certainty of success and a higher expected value. I’m saying “may” here, because I don’t know whether that is indeed the case. An argument for it would be that an antagonistic ASI that somehow fails risks total annihilation of all civilization and effectively itself, possibly by an irrational humanity “taking it down with them”, whereas the failure cases for cooperative ASI are more along the lines of losing some years of progress by having to wait longer to achieve full power.
I worded it badly by omitting “destroy or enslave us”. The corrected version is: “Having said that I would still consider it inevitable that all significant power goes away from humans to ASI at some point. The open question for me is not whether it at some point could destroy or enslave us, but how likely it is that it will want to.”