For readers, I want to connect Legg’s vision (and your agenda) with OAA: https://www.lesswrong.com/posts/jRf4WENQnhssCb6mJ/davidad-s-bold-plan-for-alignment-an-in-depth-explanation, which recognises “deliberative dialogues”/LM agent alignment as not robustly safe enough for superintelligence capabilities and deep and very consequential (pivotal action-level) plans, but perhaps good enough to task thus-aligned human-level LM agents with accelerating the progress on the OAA agenda itself.
For readers, I want to connect Legg’s vision (and your agenda) with OAA: https://www.lesswrong.com/posts/jRf4WENQnhssCb6mJ/davidad-s-bold-plan-for-alignment-an-in-depth-explanation, which recognises “deliberative dialogues”/LM agent alignment as not robustly safe enough for superintelligence capabilities and deep and very consequential (pivotal action-level) plans, but perhaps good enough to task thus-aligned human-level LM agents with accelerating the progress on the OAA agenda itself.