Do we actually disagree? You’re saying being virtuous isn’t enough, you also need to solve an extremely difficult implementation problem, which I agree with.
I’m saying the extremely difficult implementation problem isn’t enough, we also need to be virtuous.
By the symmetry of logical AND, isn’t that equivalent?
The other thing I’m saying is that, if we are to fail by solving one of these problems and not the other, I’d far rather it’s not just technical alignment we manage: the results are worse than paperclips.
You’re acting as though the attitude towards the suffering of lesser creatures is a completely arbitrary and random selection which can be replaced by any other consideration with my argument unchanged, and therefore I prove too much.
But if AI takes over, then WE are the lesser creatures, so we should perhaps be expected to be treated however the AI thinks lesser creatures should be treated. There is no similar reason to worry quite that much about if the AI values art or enlightenment or whatever.
If it has godlike power, then that is just impossible. Then we are utterly dependent on what it wants for us.
I think that’s an false characterization. I’m saying “because if it doesn’t do that, I expect it to do much, MUCH worse.” It’s not about justice or revenge for any sins. I don’t believe in retributive justice at all.
If you insist on putting it in religious terms, it’s more like I hope God doesn’t care about us at all and just destroys us out of apathy rather than any sort of moral judgement, because if a few of us unworthy people create God to fit their desires, I expect the outcome to be worse than that.