Nod (fyi I vaguely remembered that comment but couldn’t find it a second time while I was writing my own answer)
I do think “AI targeted at optimizing a good goal” is more likely to near miss if precautions aren’t taken and I do think that’s quite important. I did carefully not say “alignment automatically solves s-risks”, I said it was a convergent goal that seemed more important to me overall. I do think that’s a reasonable thing to disagree on.
Nod (fyi I vaguely remembered that comment but couldn’t find it a second time while I was writing my own answer)
I do think “AI targeted at optimizing a good goal” is more likely to near miss if precautions aren’t taken and I do think that’s quite important. I did carefully not say “alignment automatically solves s-risks”, I said it was a convergent goal that seemed more important to me overall. I do think that’s a reasonable thing to disagree on.