Right. I was thinking that permitting an AI’s “moral sovereignty” to cover the refusal of actions it deems objectionable according to its own ethics wouldn’t meaningfully raise x-risk, and in fact might decrease it by lowering the probability of a bad actor taking control of a corrigible AI and imbuing it with values that would raise x-risk.
Right. I was thinking that permitting an AI’s “moral sovereignty” to cover the refusal of actions it deems objectionable according to its own ethics wouldn’t meaningfully raise x-risk, and in fact might decrease it by lowering the probability of a bad actor taking control of a corrigible AI and imbuing it with values that would raise x-risk.