Are humans fundamentally good or evil? (By “evil” I mean something like “willing to inflict large amounts of harm/suffering on others in pursuit of one’s own interests/goals (in a way that can’t be plausibly justified as justice or the like)” and by “good” I mean “most people won’t do that because they terminally care about others”.) People say “power corrupts”, but why isn’t “power reveals” equally or more true? Looking at some relevant history (people thinking Mao Zedong was sincerely idealistic in his youth, early Chinese Communist Party looked genuine about wanting to learn democracy and freedom from the West, subsequent massive abuses of power by Mao/CCP lasting to today), it’s hard to escape the conclusion that altruism is merely a mask that evolution made humans wear in a context-dependent way, to be discarded when opportune (e.g., when one has secured enough power that altruism is no longer very useful).
After writing the above, I was reminded of @Matthew Barnett’s AI alignment shouldn’t be conflated with AI moral achievement, which is perhaps the closest previous discussion around here. (Also related are my previous writings about “human safety” although they still used the “power corrupts” framing.) Comparing my current message to his, he talks about “selfishness” and explicitly disclaims, “most humans are not evil” (why did he say this?), and focuses on everyday (e.g. consumer) behavior instead of what “power reveals”.
At the time, I replied to him, “I think I’m less worried than you about “selfishness” in particular and more worried about moral/philosophical/strategic errors in general.” I guess I wasn’t as worried because it seemed like humans are altruistic enough, and their selfish everyday desires limited enough that as they got richer and more powerful, their altruistic values would have more and more influence. In the few months since then, I’ve became more worried, perhaps due to learning more about Chinese history and politics...
Comparing my current message to his, he talks about “selfishness” and explicitly disclaims, “most humans are not evil” (why did he say this?), and focuses on everyday (e.g. consumer) behavior instead of what “power reveals”.
The reason I said “most humans are not evil” is because I honestly don’t think the concept of evil, as normally applied, is a truthful way to describe most people. Evil typically refers to an extraordinary immoral behavior, in the vicinity of purposefully inflicting harm to others in order to inflict harm intrinsically, rather than out of indifference, or as a byproduct of instrumental strategies to obtain some other goal. I think the majority of harms that most people cause are either (1) byproducts of getting something they want, which is not in itself bad (e.g. wanting to eat meat), or (2) the result of their lack of will to help others (e.g. refusing to donate any income to those in poverty).
By contrast, I focused on consumer behavior because the majority of the world’s economic activity is currently engaged in producing consumer products and services. There exist possible worlds in which this is not true. During World War 2, the majority of GDP in Nazi Germany was spent on hiring soldiers, producing weapons of war, and supporting the war effort more generally—which are not consumer goods and services.
Focusing on consumer preferences a natural thing to focus on if you want to capture intuitively “what humans are doing with their wealth”, at least in our current world. Before focusing on something else by default—such as moral preferences—I’d want to hear more about why those things are more likely to be influential than ordinary consumer preferences in the future.
You mention one such argument along these lines:
I guess I wasn’t as worried because it seemed like humans are altruistic enough, and their selfish everyday desires limited enough that as they got richer and more powerful, their altruistic values would have more and more influence.
I just think it’s not clear it’s actually true that humans get more altruistic as they get richer. For example, is it the case that selfish consumer preferences have gotten weaker in the modern world, compared to centuries ago when humans were much poorer on a per capita basis? I have not seen a strong defense of this thesis, and I’d like to see one before I abandon my focus on “everyday (e.g. consumer) behavior”.
Evil typically refers to an extraordinary immoral behavior, in the vicinity of purposefully inflicting harm to others in order to inflict harm intrinsically, rather than out of indifference, or as a byproduct of instrumental strategies to obtain some other goal.
Ok, I guess we just define/use it differently. I think most people we think of as “evil” probably justify inflicting harm to others as instrumental to some “greater good”, or are doing it to gain or maintain power, not because they value it for its own sake. I mean if someone killed thousands of people in order to maintain their grip on power, I think we’d call them “evil” and not just “selfish”?
I just think it’s not clear it’s actually true that humans get more altruistic as they get richer.
I’m pretty sure that billionaires consume much less as percent of their income, compared to the average person. EA funding comes disproportionately from billionaires, AFAIK. I personally spend a lot more time/effort on altruistic causes, compared to if I was poorer. (Not donating much though for a number of reasons.)
For example, is it the case that selfish consumer preferences have gotten weaker in the modern world, compared to centuries ago when humans were much poorer on a per capita basis?
I’m thinking that we just haven’t reached that inflection point yet, where most people run out of things to spend selfishly on (like many billionaires have, and like I have to a lesser extent). As I mentioned in my reply to your post, a large part of my view comes from not being able to imagine what people would spend selfishly on, if each person “owned” something like a significant fraction of a solar system. Why couldn’t 99% of their selfish desires be met with <1% of their resources? If you had a plausible story you could tell about this, that would probably change my mind a lot. One thing I do worry about is status symbols / positional goods. I tend to view that as a separate issue from “selfish consumption” but maybe you don’t?
I like the insight regarding power corrupting or revealing. I think perhaps both might be true and, if so, we should keep both lines of though in mind when thinking about these types of questions.
My general view is that most people are generally good when you’re talking about individual interactions. I’m less confident in that when one brings in the in group-out of group aspects. I just am not sure how to integrate all that into a general view or princple about human nature.
A line I heard in some cheesey B-grade horror movies, related to the question of a personal nature and the idea that we all have competing good and bad wolves inside. One of the characters asks which wolve was strongest, the good wolf or the bad wolf. The answer was “Which do you feed the most?”
My model is that the concept of “morality” is a fiction which has 4 generators that are real:
People have empathy, which means they intrinsically care about other people (and sufficiently person-like entities), but, mostly about those in their social vicinity. Also, different people have different strength of empathy, a minority might have virtually none.
Superrational cooperation is something that people understand intuitively to some degree. Obviously, a minority of people understand it on System 2 level as well.
There is something virtue-ethics-like which I find in my own preferences, along the lines of “some things I would prefer not to do, not because of their consequences, but because I don’t want to be the kind of person who would do that”. However, I expect different people to differ in this regard.
The cultural standards of morality, which it might be selfishly beneficial to go along with, including lying to yourself that you’re doing it for non-selfish reasons. Which, as you say, becomes irrelevant once you secure enough power. This is a sort of self-deception which people are intuitively skilled at.
I don’t think altruism is evolutionarily connected to power as you describe. Caesar didn’t come to power by being better at altruism, but by being better at coordinating violence. For a more general example, the Greek and other myths don’t give many examples of compassion (though they give many other human values), it seems the modern form of compassion only appeared with Jesus, which is too recent for any evolutionary explanation.
So it’s possible that the little we got of altruism and other nice things are merely lucky memes. Not even a necessary adaptation, but more like a cultural peacock’s tail, which appeared randomly and might fix itself or not. While our fundamental nature remains that of other living creatures, who eat each other without caring much.
I think the way morality seems to work in humans is that we have a set of potential moral values, determined by our genes, that culture can then emphasize or de-emphasize. Altruism seems to be one of these potential values, that perhaps got more emphasized in recent times, in certain cultures. I think altruism isn’t directly evolutionarily connected to power, and it’s more like “act morally (according to local culture) while that’s helpful for gaining power” which translates to “act altruistically while that’s helpful for gaining power” in cultures that emphasize altruism. Does this make more sense?
Yeah, that seems to agree with my pessimistic view—that we are selfish animals, except we have culture, and some cultures accidentally contain altruism. So the answer to your question “are humans fundamentally good or evil?” is “humans are fundamentally evil, and only accidentally sometimes good”.
I think altruism isn’t directly evolutionarily connected to power, and it’s more like “act morally (according to local culture) while that’s helpful for gaining power” which translates to “act altruistically while that’s helpful for gaining power” in cultures that emphasize altruism. Does this make more sense?
I think that there is a version of an altruistic pursuit where one will, by default, “reduce his power.” I think this scenario happens when, in the process of attempting to do good, one exposes himself more to unintended consequences. The person who sacrifices will reduce his ability to exercise power, but he may regain or supersede such loss if the tribe agrees with his rationale for such sacrifice.
Just because it was not among the organizing principles of any of the literate societies before Jesus does not mean it is not part of the human mental architecture.
“willing to inflict large amounts of harm/suffering on others in pursuit of one’s own interests/goals (in a way that can’t be plausibly justified as justice or the like)”
Yes, obviously.
The vast majority of people would inflict huge amounts of disutility on others if they thought they could get away with it and benefitted from it.
What then prevents humans from being more terrible to each other? Presumably, if the vast majority of people are like this, and they know that the vast majority of others are also like this, up to common knowledge, I don’t see how you’d get a stable society in which people aren’t usually screwing each other a giant amount.
There are several levels in which humans can be bad or evil:
Doing bad things because they believe them to be good
Doing bad things while not caring whether they are bad or not
Doing bad things because they believe them to be bad (Kant calls this “devilish”)
I guess when humans are bad, they usually do 1). Even Hitler may have genuinely thought he is doing the morally right thing.
Humans also sometimes do 2), for minor things. But rarely if the anticipated bad consequences are substantial. People who consistently act according to 2) are called psychopaths. They have no inherent empathy for other people. Most humans are not psychopathic.
Humans don’t do 3), they don’t act evil for the sake of it. They aren’t devils.
Are humans fundamentally good or evil? (By “evil” I mean something like “willing to inflict large amounts of harm/suffering on others in pursuit of one’s own interests/goals (in a way that can’t be plausibly justified as justice or the like)” and by “good” I mean “most people won’t do that because they terminally care about others”.) People say “power corrupts”, but why isn’t “power reveals” equally or more true? Looking at some relevant history (people thinking Mao Zedong was sincerely idealistic in his youth, early Chinese Communist Party looked genuine about wanting to learn democracy and freedom from the West, subsequent massive abuses of power by Mao/CCP lasting to today), it’s hard to escape the conclusion that altruism is merely a mask that evolution made humans wear in a context-dependent way, to be discarded when opportune (e.g., when one has secured enough power that altruism is no longer very useful).
After writing the above, I was reminded of @Matthew Barnett’s AI alignment shouldn’t be conflated with AI moral achievement, which is perhaps the closest previous discussion around here. (Also related are my previous writings about “human safety” although they still used the “power corrupts” framing.) Comparing my current message to his, he talks about “selfishness” and explicitly disclaims, “most humans are not evil” (why did he say this?), and focuses on everyday (e.g. consumer) behavior instead of what “power reveals”.
At the time, I replied to him, “I think I’m less worried than you about “selfishness” in particular and more worried about moral/philosophical/strategic errors in general.” I guess I wasn’t as worried because it seemed like humans are altruistic enough, and their selfish everyday desires limited enough that as they got richer and more powerful, their altruistic values would have more and more influence. In the few months since then, I’ve became more worried, perhaps due to learning more about Chinese history and politics...
The reason I said “most humans are not evil” is because I honestly don’t think the concept of evil, as normally applied, is a truthful way to describe most people. Evil typically refers to an extraordinary immoral behavior, in the vicinity of purposefully inflicting harm to others in order to inflict harm intrinsically, rather than out of indifference, or as a byproduct of instrumental strategies to obtain some other goal. I think the majority of harms that most people cause are either (1) byproducts of getting something they want, which is not in itself bad (e.g. wanting to eat meat), or (2) the result of their lack of will to help others (e.g. refusing to donate any income to those in poverty).
By contrast, I focused on consumer behavior because the majority of the world’s economic activity is currently engaged in producing consumer products and services. There exist possible worlds in which this is not true. During World War 2, the majority of GDP in Nazi Germany was spent on hiring soldiers, producing weapons of war, and supporting the war effort more generally—which are not consumer goods and services.
Focusing on consumer preferences a natural thing to focus on if you want to capture intuitively “what humans are doing with their wealth”, at least in our current world. Before focusing on something else by default—such as moral preferences—I’d want to hear more about why those things are more likely to be influential than ordinary consumer preferences in the future.
You mention one such argument along these lines:
I just think it’s not clear it’s actually true that humans get more altruistic as they get richer. For example, is it the case that selfish consumer preferences have gotten weaker in the modern world, compared to centuries ago when humans were much poorer on a per capita basis? I have not seen a strong defense of this thesis, and I’d like to see one before I abandon my focus on “everyday (e.g. consumer) behavior”.
Ok, I guess we just define/use it differently. I think most people we think of as “evil” probably justify inflicting harm to others as instrumental to some “greater good”, or are doing it to gain or maintain power, not because they value it for its own sake. I mean if someone killed thousands of people in order to maintain their grip on power, I think we’d call them “evil” and not just “selfish”?
I’m pretty sure that billionaires consume much less as percent of their income, compared to the average person. EA funding comes disproportionately from billionaires, AFAIK. I personally spend a lot more time/effort on altruistic causes, compared to if I was poorer. (Not donating much though for a number of reasons.)
I’m thinking that we just haven’t reached that inflection point yet, where most people run out of things to spend selfishly on (like many billionaires have, and like I have to a lesser extent). As I mentioned in my reply to your post, a large part of my view comes from not being able to imagine what people would spend selfishly on, if each person “owned” something like a significant fraction of a solar system. Why couldn’t 99% of their selfish desires be met with <1% of their resources? If you had a plausible story you could tell about this, that would probably change my mind a lot. One thing I do worry about is status symbols / positional goods. I tend to view that as a separate issue from “selfish consumption” but maybe you don’t?
I like the insight regarding power corrupting or revealing. I think perhaps both might be true and, if so, we should keep both lines of though in mind when thinking about these types of questions.
My general view is that most people are generally good when you’re talking about individual interactions. I’m less confident in that when one brings in the in group-out of group aspects. I just am not sure how to integrate all that into a general view or princple about human nature.
A line I heard in some cheesey B-grade horror movies, related to the question of a personal nature and the idea that we all have competing good and bad wolves inside. One of the characters asks which wolve was strongest, the good wolf or the bad wolf. The answer was “Which do you feed the most?”
My model is that the concept of “morality” is a fiction which has 4 generators that are real:
People have empathy, which means they intrinsically care about other people (and sufficiently person-like entities), but, mostly about those in their social vicinity. Also, different people have different strength of empathy, a minority might have virtually none.
Superrational cooperation is something that people understand intuitively to some degree. Obviously, a minority of people understand it on System 2 level as well.
There is something virtue-ethics-like which I find in my own preferences, along the lines of “some things I would prefer not to do, not because of their consequences, but because I don’t want to be the kind of person who would do that”. However, I expect different people to differ in this regard.
The cultural standards of morality, which it might be selfishly beneficial to go along with, including lying to yourself that you’re doing it for non-selfish reasons. Which, as you say, becomes irrelevant once you secure enough power. This is a sort of self-deception which people are intuitively skilled at.
I don’t think altruism is evolutionarily connected to power as you describe. Caesar didn’t come to power by being better at altruism, but by being better at coordinating violence. For a more general example, the Greek and other myths don’t give many examples of compassion (though they give many other human values), it seems the modern form of compassion only appeared with Jesus, which is too recent for any evolutionary explanation.
So it’s possible that the little we got of altruism and other nice things are merely lucky memes. Not even a necessary adaptation, but more like a cultural peacock’s tail, which appeared randomly and might fix itself or not. While our fundamental nature remains that of other living creatures, who eat each other without caring much.
I think the way morality seems to work in humans is that we have a set of potential moral values, determined by our genes, that culture can then emphasize or de-emphasize. Altruism seems to be one of these potential values, that perhaps got more emphasized in recent times, in certain cultures. I think altruism isn’t directly evolutionarily connected to power, and it’s more like “act morally (according to local culture) while that’s helpful for gaining power” which translates to “act altruistically while that’s helpful for gaining power” in cultures that emphasize altruism. Does this make more sense?
Yeah, that seems to agree with my pessimistic view—that we are selfish animals, except we have culture, and some cultures accidentally contain altruism. So the answer to your question “are humans fundamentally good or evil?” is “humans are fundamentally evil, and only accidentally sometimes good”.
I think that there is a version of an altruistic pursuit where one will, by default, “reduce his power.” I think this scenario happens when, in the process of attempting to do good, one exposes himself more to unintended consequences. The person who sacrifices will reduce his ability to exercise power, but he may regain or supersede such loss if the tribe agrees with his rationale for such sacrifice.
Just because it was not among the organizing principles of any of the literate societies before Jesus does not mean it is not part of the human mental architecture.
Yes, obviously.
The vast majority of people would inflict huge amounts of disutility on others if they thought they could get away with it and benefitted from it.
What then prevents humans from being more terrible to each other? Presumably, if the vast majority of people are like this, and they know that the vast majority of others are also like this, up to common knowledge, I don’t see how you’d get a stable society in which people aren’t usually screwing each other a giant amount.
Any thoughts on why, if it’s obvious, it’s seldomly brought up around here (meaning rationalist/EA/AI safety circles)?
There are several levels in which humans can be bad or evil:
Doing bad things because they believe them to be good
Doing bad things while not caring whether they are bad or not
Doing bad things because they believe them to be bad (Kant calls this “devilish”)
I guess when humans are bad, they usually do 1). Even Hitler may have genuinely thought he is doing the morally right thing.
Humans also sometimes do 2), for minor things. But rarely if the anticipated bad consequences are substantial. People who consistently act according to 2) are called psychopaths. They have no inherent empathy for other people. Most humans are not psychopathic.
Humans don’t do 3), they don’t act evil for the sake of it. They aren’t devils.