So we want FAI’s values to be conducive with human values. But how are people proposing FAI will deal with the fact that humans have conflicting and inconsistent values, both within individual humans and between different humans? Do we just hope that our values when extrapolated and reasoned far enough are consistent enough to be feasibly satisfied? I feel like I might have missed the LW post where this kind of thing was discussed, or maybe I read it and was left unsatisfied with the reasoning.
What happens, for example, to the person who takes their religion very seriously and wants everyone to convert to their religion? Or just people whose values are heavily tied to their religious beliefs, of which there probably are a few, considering there are billions of religious people around.
I feel like I might have missed the LW post where this kind of thing was discussed, or maybe I read it and was left unsatisfied with the reasoning.
I do not think anyone who’s thought about it deeply thinks that they have the best possible solution to this problem, and most proposed solutions are preliminary at best.
How would you as an individual deal with having inconsistent values? Let’s assume that you become super intelligent, so you can rather reliably see the possible consequences of any changes you would try to do.
The person who takes their religion very seriously probably believes that the religion is factually true. Or possibly they believe that the religion is false, but that it is a “Noble Lie”, i.e. that the belief in the religion is a net benefit. Either way, there is an empirical claim they have related to the religion. They would probably be interested in knowing whether this claim is true.
These are not full answers, just hints towards possible answers.
So we want FAI’s values to be conducive with human values. But how are people proposing FAI will deal with the fact that humans have conflicting and inconsistent values, both within individual humans and between different humans? Do we just hope that our values when extrapolated and reasoned far enough are consistent enough to be feasibly satisfied? I feel like I might have missed the LW post where this kind of thing was discussed, or maybe I read it and was left unsatisfied with the reasoning.
What happens, for example, to the person who takes their religion very seriously and wants everyone to convert to their religion? Or just people whose values are heavily tied to their religious beliefs, of which there probably are a few, considering there are billions of religious people around.
I do not think anyone who’s thought about it deeply thinks that they have the best possible solution to this problem, and most proposed solutions are preliminary at best.
There’s a pdf discussing the matter linked here—though as you can see, a lot of people failed to read it (or “seer” has a lot of sockpuppets).
How would you as an individual deal with having inconsistent values? Let’s assume that you become super intelligent, so you can rather reliably see the possible consequences of any changes you would try to do.
The person who takes their religion very seriously probably believes that the religion is factually true. Or possibly they believe that the religion is false, but that it is a “Noble Lie”, i.e. that the belief in the religion is a net benefit. Either way, there is an empirical claim they have related to the religion. They would probably be interested in knowing whether this claim is true.
These are not full answers, just hints towards possible answers.