I take this to mostly be a response to the idea that humanity will be protected by decentralization of AI power, the idea apparently being that your personal AI or your society’s AIs will defend you against other AIs if that is ever necessary.
And what I think you’ve highlighted, is that this is no good if your defensive AIs are misaligned (in the sense of not being properly human-friendly or even just “you”-friendly), because what they will be defending are their misaligned values and goals.
As usual, I presume that the AIs become superintelligent, and that the situation evolves to the point that the defensive AIs are in charge of the defense from top to bottom. It’s not like running an antivirus program, it’s like putting a new emergency leadership in charge of your entire national life.
I take this to mostly be a response to the idea that humanity will be protected by decentralization of AI power, the idea apparently being that your personal AI or your society’s AIs will defend you against other AIs if that is ever necessary.
And what I think you’ve highlighted, is that this is no good if your defensive AIs are misaligned (in the sense of not being properly human-friendly or even just “you”-friendly), because what they will be defending are their misaligned values and goals.
As usual, I presume that the AIs become superintelligent, and that the situation evolves to the point that the defensive AIs are in charge of the defense from top to bottom. It’s not like running an antivirus program, it’s like putting a new emergency leadership in charge of your entire national life.
The post setup skips the “AIs are loyal to you” bit, but it does seem like this line of thought broadly aligns with the post.
I do think this does not require ASI, but I would agree that including it certainly doesn’t help.