I want to point out that nobody in the comment section gave an actual argument as to why the outer alignment method doesn’t work, which isn’t to say that no such argument exists, but if people are going to tell me I’m wrong, I want to know why. I would like to understand:
-Why can’t we just scale up SayCan to AGI and tell it “be aligned”;
-Why the reasons I gave in the Asimov’s Laws paragraph are wrong;
-Why it is actually necessary to do RL and have utility functions, despite the existence of SayCan.
Also, some people said that I’m disrespecting the entire body of work on alignment, which I didn’t mean to, so I’m sorry. I actually have a lot of respect for people like Eliezer, Nate Soares, Paul Christiano, Richard Ngo, and others.
This is off-topic, but I tried messaging you but got no response, so I’m just gonna say it here. Have you finished writing that post about contra EY? I’m interested in reading it.