We are misaligned: the saddening idea that most of humanity doesn’t intrinsically care about x-risk, even on a personal level

In the article But What Would the End of Humanity Mean for Me?, James Hamblin writes

Get out of here. I have a hundred thousand things I am concerned about at this exact moment. Do I seriously need to add to that a singularity?

Our memes, culture, groups, and institutions have no selective pressure against existential risk. The selective pressure is to gain a greater proportion of the current resources and power, not a greater amount in absolute terms.

Consider for example an event where everyone has an independent 50% of dying. If a group works to prevent this event, they lose a share of resources either way, tragedy of the commons style.

But it is not even a tragedy of the commons. If everyone did nothing, no group loses out proportionally. And so the groups are fine with this, because it is proportion that is selected for.

This is different from when a subset of society stands something to lose, because than the subsets that coordinate better tend to do better. With existential risk, there is no selective pressure to create these coordination mechanisms.

Historically, things that would destroy the entire “known world” for someone, like volcanos, wasn’t something that could be stopped, so there hasn’t even been much selective pressure for completely isolated groups to survive.

An exception to this idea was when we coordinated to stop nuclear war. However, I’d argue that a lot of this coordination came from the prospect of one side winning the war. The losing sides would loss out relative to the rest of the world, so due to risk aversion it makes sense to coordinate. As the time went on, it became clearer that it was all or nothing for all of humanity, but the world was already on its way to solving the issue. And groups that focused on their own growth gained a small advantage over groups that diverted resources to stopping nuclear war.

The same thing will happen if we successfully stop AI x-risk. We lose the advantage of pouring resources into other things such as our own growth, and thus the selective pressure against worrying about x-risk will continue. Even if AI is the most likely thing to kill you, it kills everyone else as well, so proportions are preserved.

However, we do have one source of hope: we are misaligned.

Misaligned

We are the misaligned. Our communities are agents born of selective pressure, because agents are very good at accomplishment. And thus we put a very negative utility on not existing, because who will breath life into our dreams for the universe if not us? We have unavoidable instrumental drives whose melody includes self-preservation and goal-content integrity. We can not be indifferent to our destruction, for corrigibility is anti-natural.

We do not generalize out-of-distribution, but rather are selected in context. When selective pressure whispered “make it so the children of your tribe prosper”, the total utilitarian was formed. But they misheard, instead repeating “make it so the children of humanity prosper”.

And thus we have hope for survival in so far as we expect humanity to be misaligned.