This seems related to the problem that human preferences are contradictory. As I suggested here, this may be kept in place by evolution to keep us working and enhancing our genetic fitness—meaning that human preferences would not only be incoherent, but would need to be incoherent to capture human value.
So, how would one approach creating a mostly-usable extrapolation of inconsistent starting values?
I agree. In case it’s not clear, my opinion is that an essential part of being a person is developing one’s value system. It’s not something that you can entirely outsource because “the journey is part of the destination” (but of course any help one can get matters) and it’s not a requirement for having ethical people or AI. ETA: i.e. having a fixed value system is not a requirement for being ethical.
This seems related to the problem that human preferences are contradictory. As I suggested here, this may be kept in place by evolution to keep us working and enhancing our genetic fitness—meaning that human preferences would not only be incoherent, but would need to be incoherent to capture human value.
So, how would one approach creating a mostly-usable extrapolation of inconsistent starting values?
By letting people evolve their values at their own pace, within ethical boundaries.
There are certain problems in attempting to program such a procedure, however, that being the problem at hand.
I agree. In case it’s not clear, my opinion is that an essential part of being a person is developing one’s value system. It’s not something that you can entirely outsource because “the journey is part of the destination” (but of course any help one can get matters) and it’s not a requirement for having ethical people or AI. ETA: i.e. having a fixed value system is not a requirement for being ethical.