All images from the WaitButWhy sequence on AI

Lots of people seem to like visual learning. I don’t see much of an issue with that. People who have fun with thinking tend to get more bang for their buck.

It seems reasonable to think that that Janus’s image of neural network shoggoths makes it substantially easier for a lot of people to fully operationalize the concept that RLHF could steer humanity off of a cliff:

Lots of people I’ve met said that they were really glad that they encountered Tim Urban’s WaitButWhy blog post on AI back in 2015, which was largely just a really good distillation of Nick Bostrom’s Superintelligence (2014). It’s a rather long (but well-written) post, so what impressed me was not the distillation, but the images.

The images in the post were very vivid, especially in the middle. It seems to me like images can work as a significant thought aid, by leaning on visual memory to aid recall, and/​or to make core concepts more cognitively available during the thought process in general.

But also, almost by themselves, the images do a pretty great job describing the core concepts of AI risk, as well as the general gist of the entirety of Tim Urban’s sequence. Considering that he managed to get that result, even though the post itself is >22,000 words (around as long as the entire CFAR handbook), maybe Tim Urban was simultaneously doing something very wrong and very right with writing the distillation; could he have turned a 2-hour post into a 2-minute post by just doubling the number of images?

If there was a true-optimal blog post to explain AI safety for the first time, to an otherwise-uninterested layperson (a very serious matter in AI governance), it wouldn’t be surprising to me if that true-optimal blog post contained a lot of images. Walls of text are inevitable at some point or another, but there’s the old saying that a picture’s worth a thousand words. Under current circumstances, it makes sense for critical AI safety concepts to be easier and less burdensome to think and learn about for the first time, rather than harder and more burdensome to think about for the first time.

Edge1
Edge
Projections
S-Curves
PPTExponentialGrowthof_Computing-1
Intelligence
Intelligence2
Train1
Train2
Train3
Train4
staircase
staircase2

I’ve found the above two pictures particularly helpful, for doing an excellent job depicting the scope. Out of all the images that could be used to help describe AGI to someone for the first time, I would pick those two.

Tripwire
beam2
Howard Graph
Square1
Square2

These two images are pretty good as well, as a primer for the current state of affairs. Obviously, they were drawn in 2015 and need to be redrawn.

only-humans-cartoon
Outcome Spectrum

This image puts the significance of the situation into context.