a specific algorithm/dynamic for judging values, or
a complicated blob of values like happiness, love, creativity, novelty, self determination, fairness, life (as in protecting theirof), etc.?
If it’s 1, can we say something interesting and non-trivial about the algorithm, besides the fact that it’s an algorithm? In other words, everything can be viewed as an algorithm, but what’s the point of viewing morality as an algorithm?
If it’s 2, why do we think that two people on opposite sides of the Earth are referring to the same complicated blob of values when they say “morality”? I know the argument about the psychological unity of humankind (not enough time for significant genetic divergence), but what about cultural/memetic evolution?
I’m guessing the answer to my first question is something like, morality is an algorithm whose current “state” is a complicated blob of values like happiness, love, … so both of my other questions ought to apply.
If it’s 2, why do we think that two people on opposite sides of the Earth are referring to the same complicated blob of values when they say “morality”? I know the argument about the psychological unity of humankind (not enough time for significant genetic divergence), but what about cultural/memetic evolution?
You don’t even have to do any cross-cultural comparisons to make such an argument. Considering the insights from modern behavioral genetics, individual differences within any single culture will suffice.
There is no reason to be at all tentative about this. There’s tons of cog sci data about what people mean when they talk about morality. It varies hugely (but predictably) across cultures.
Why are you using algorithm/dynamic here instead of function or partial function? (On what space, I will ignore that issue, just as you have...) Is it supposed to be stateful? I’m not even clear what that would mean. Or is function what you mean by #2? I’m not even really clear on how these differ.
You might have gotten confused because I quoted Psy-Kosh’s phrase “specific algorithm/dynamic for judging values” whereas Eliezer’s original idea I think was more like an algorithm for changing one’s values in response to moral arguments. Here are Eliezer’s own words:
I would say, by the way, that the huge blob of a computation is not just my present terminal values (which I don’t really have - I am not a consistent expected utility maximizers); the huge blob of a computation includes the specification of those moral arguments, those justifications, that would sway me if I heard them.
Others have pointed out that this definition is actually quite unlikely to be coherent: people would be likely to be ultimately persuaded by different moral arguments and justifications if they had different experiences and heard arguments in different orders etc.
Others have pointed out that this definition is actually quite unlikely to be coherent
Yes, see here for an argument to that effect by Marcello and subsequent discussion about it between Eliezer and myself.
I think the metaethics sequence is probably the weakest of Eliezer’s sequences on LW. I wonder if he agrees with that, and if so, what he plans to do about this subject for his rationality book.
I think the metaethics sequence is probably the weakest of Eliezer’s sequences on LW. I wonder if he agrees with that, and if so, what he plans to do about this subject for his rationality book.
This is somewhat of a concern given Eliezer’s interest in Friendliness!
As far as I can understand, Eliezer has promoted two separate ideas about ethics: defining personal morality as a computation in the person’s brain rather than something mysterious and external, and extrapolating that computation into smarter creatures. The former idea is self-evident, but the latter (and, by extension, CEV) has received a number of very serious blows recently. IMO it’s time to go back to the drawing board. We must find some attack on the problem of preference, latch onto some small corner, that will allow us to make precise statements. Then build from there.
defining personal morality as a computation in the person’s brain rather than something mysterious and external
But I don’t see how that, by itself, is a significant advance. Suppose I tell you, “mathematics is a computation in a person’s brain rather than something mysterious and external”, or “philosophy is a computation in a person’s brain rather than something mysterious and external”, or “decision making is a computation in a person’s brain rather than something mysterious and external” how much have I actually told you about the nature of math, or philosophy, or decision making?
Is morality actually:
a specific algorithm/dynamic for judging values, or
a complicated blob of values like happiness, love, creativity, novelty, self determination, fairness, life (as in protecting theirof), etc.?
If it’s 1, can we say something interesting and non-trivial about the algorithm, besides the fact that it’s an algorithm? In other words, everything can be viewed as an algorithm, but what’s the point of viewing morality as an algorithm?
If it’s 2, why do we think that two people on opposite sides of the Earth are referring to the same complicated blob of values when they say “morality”? I know the argument about the psychological unity of humankind (not enough time for significant genetic divergence), but what about cultural/memetic evolution?
I’m guessing the answer to my first question is something like, morality is an algorithm whose current “state” is a complicated blob of values like happiness, love, … so both of my other questions ought to apply.
Wei_Dai:
You don’t even have to do any cross-cultural comparisons to make such an argument. Considering the insights from modern behavioral genetics, individual differences within any single culture will suffice.
There is no reason to be at all tentative about this. There’s tons of cog sci data about what people mean when they talk about morality. It varies hugely (but predictably) across cultures.
Why are you using algorithm/dynamic here instead of function or partial function? (On what space, I will ignore that issue, just as you have...) Is it supposed to be stateful? I’m not even clear what that would mean. Or is function what you mean by #2? I’m not even really clear on how these differ.
You might have gotten confused because I quoted Psy-Kosh’s phrase “specific algorithm/dynamic for judging values” whereas Eliezer’s original idea I think was more like an algorithm for changing one’s values in response to moral arguments. Here are Eliezer’s own words:
Others have pointed out that this definition is actually quite unlikely to be coherent: people would be likely to be ultimately persuaded by different moral arguments and justifications if they had different experiences and heard arguments in different orders etc.
Yes, see here for an argument to that effect by Marcello and subsequent discussion about it between Eliezer and myself.
I think the metaethics sequence is probably the weakest of Eliezer’s sequences on LW. I wonder if he agrees with that, and if so, what he plans to do about this subject for his rationality book.
This is somewhat of a concern given Eliezer’s interest in Friendliness!
As far as I can understand, Eliezer has promoted two separate ideas about ethics: defining personal morality as a computation in the person’s brain rather than something mysterious and external, and extrapolating that computation into smarter creatures. The former idea is self-evident, but the latter (and, by extension, CEV) has received a number of very serious blows recently. IMO it’s time to go back to the drawing board. We must find some attack on the problem of preference, latch onto some small corner, that will allow us to make precise statements. Then build from there.
But I don’t see how that, by itself, is a significant advance. Suppose I tell you, “mathematics is a computation in a person’s brain rather than something mysterious and external”, or “philosophy is a computation in a person’s brain rather than something mysterious and external”, or “decision making is a computation in a person’s brain rather than something mysterious and external” how much have I actually told you about the nature of math, or philosophy, or decision making?
The linked discussion is very nice.