They very briefly discuss automated AI alignment research as a proposal for mitigating AI risk, but their arguments against that plan do not respond to the most thoughtful versions of these plans. (In their defense, the most thoughtful versions of these plans basically haven’t been published, though Ryan Greenblatt is going to publish a detailed version of this plan soon. And I think that there are several people who have pretty thoughtful versions of these plans, haven’t written them up (at least publicly), but do discuss them in person.)
Am a bit confused by this section—did you think that part 3 was awful because it didn’t respond to (as yet unpublished) plans, or for some other reason?
Am a bit confused by this section—did you think that part 3 was awful because it didn’t respond to (as yet unpublished) plans, or for some other reason?