Folks sometimes talk about the human alignment problem, which I think is what you’re getting at. I think the earliest instance of it can be found in this post. Searching for “human alignment problem” on LW will turn up more stuff, although I don’t think anyone has done an exhaustive post summarizing what we really mean by this, though it’s generally one of two things:
humans aligning themselves to goals (getting yourself to do something)
aligning multiple humans to a goal (like in an organization)
Folks sometimes talk about the human alignment problem, which I think is what you’re getting at. I think the earliest instance of it can be found in this post. Searching for “human alignment problem” on LW will turn up more stuff, although I don’t think anyone has done an exhaustive post summarizing what we really mean by this, though it’s generally one of two things:
humans aligning themselves to goals (getting yourself to do something)
aligning multiple humans to a goal (like in an organization)