A concise sum-up of the basic argument for AI doom

1 - An artificial super-optimizer is likely to be developed soon.

2 - There is no known way of programming goals into an advanced optimizer, only outwardly observable behaviors of which we have no idea why they are being carried out or what motivates them.

3 - Most utility functions do not have optima with humans in them. Most utility functions do not have a term for humans at all.

4 - “Why haven’t we exterminated all mice/​bugs/​cows then?” draws quite a poor analogy. Firstly, we are not superoptimizers. Secondly, and more importantly, we care about living beings somewhat. The optimum of the utility function of the human civilization quite possibly does have mice/​bugs/​cows, perhaps even genetically engineered to not experience suffering. We are not completely indifferent to them.

The relationship between most possible superoptimizers and humanity is not like the relationship between humanity and mice at all—it is much more like the relationship between humanity and natural gas. Natural gas, not mice, is a good example of something humans are truly indifferent to—there is no term for it in our utility function. We don’t hate it, we don’t love it, it is just made of atoms that we can, and do, use for something else.

Moreover, the continued existence of natural gas probably does not pose the same threat to us as the continued existence of us would pose to a superoptimizer which does not have a term for humans in its utility function—just like we don’t have a term for natural gas in ours. Natural gas can not attempt to turn us off, and it can not create “competition” for us in form of a rival species with capabilities similar to, or surpassing, our own.

P.S. If you don’t like or find confusing the terminology of “optimizers” and “utility functions”, feel free to forget about all that. Think instead of physical states of the universe. Out of all possible states the universe could find itself in, very, very, very few contain any humans. Given a random state of the universe, creating a superintelligence that steered the universe towards that state would result in an almost guaranteed apocalypse. Of course, we want our superintelligence to steer the universe towards particular states—that’s kind of the whole point of the entire endeavor. Were it not so, we would not be attempting to build a superintelligence - a sponge would suffice. We essentially want to create a super-capable universe-steerer. The problem is getting it to steer the universe towards states we like—and this problem is very hard because (among other things) currently, given any internal desire to steer the universe towards any state at all, it is impossible to program that desire into anything at all.