Overwhelming Superintelligence

There’s many debates about “what counts as AGI” or “what counts as superintelligence?”.

Some people might consider those arguments “goalpost moving.” Some people were using “superintelligence” to mean “overwhelmingly smarter than humanity”. So, it may feel to them like it’s watering it down if you use it to mean “spikily good at some coding tasks while still not really successfully generalizing or maintaining focus.”

I think there’s just actually a wide range of concepts that need to get talked about. And, right now, most of the AIs that people will wanna talk about are kinda general and kinda superintelligent and kinda aligned.

If you have an specific concept you wanna protect, I think it’s better to just give it a clunky name that people don’t want to use in casual conversation,[1] rather than pumping against entropy to defend a simple term that could be defined to mean other things.

Previously OpenPhil had used “Transformative AI” to mean “AI that is, you know, powerful enough to radically transform society, somehow.” I think that’s a useful term. But, it’s not exactly what If Anyone Builds It is cautioning about.

The type of AI I’m most directly worried about is “overwhelmingly superhuman compared to humanity.” (And, AIs that might quickly bootstrap to become overwhelmingly superhuman).

I’ve been lately calling that Overwhelming Superintelligence.

Overwhelming Superintelligence is scary both because it’s capable of strategically outthinking humanity, and, because any subtle flaws or incompatibilities between what it wants, and what humans want, will get driven to extreme levels.

I think if anyone builds Overwhelmed Superintelligence without hitting a pretty narrow alignment target, everyone probably dies. (And, if not, the future is probably quite bad).


Appendix: Lots of “Careful Moderate Superintelligence”

I am separately worried about “Carefully Controlled Moderate Superintelligences that we’re running at scale, each instance of which is not threatening, but, we’re running a lot of them, giving them lots of room to maneuver.”

This is threatening partly because at some point that they may give rise to Overwhelming Superintelligence, but, also because sharing the planet with a slightly smarter species still doesn’t seem like it bodes well. (See humans, neanderthals, chimpanzees). They don’t have to do anything directly threatening, just keep being very useful while subtly steering things such that they get more power in the future.

  1. ^

    I actually think AIdon’tkilleveryoneism is pretty good.