AI safety should be made more accessible using non text-based media

I’ve been doing some thinking on AI safety’s awareness problem, after a quick search I found that this post summarizes my thoughts pretty well. In short, AI safety has an awareness problem in a way that other major crises do not (I’ll draw parallels specifically with climate change in my analysis). Most ordinary people have not even heard of the problem. Of those that have, most do not understand the potential risks. They cannot concretely imagine the ways that things could go horribly wrong. I’ll outline a few reasons I think this is an undesirable state of affairs, but on a surface level I feel it should be obvious to most people convinced of the severity of the issue why the alignment problem should be garnering at least as much attention as climate change, if not more.

The reason I’m writing this up in the first place though is to point out what I see as a critical obstacle for raising awareness that I feel is somewhat overlooked, namely, that virtually all of the best entry-level material on the alignment problem is text-based. It is sadly the case that many many people are simply unwilling to read anything longer than a short blog post or article, ever. Of those that are, getting them to consume lengthy non-fiction on what appears at first glance to be a fairly dry technical topic is still an almost insurmountable challenge. It seems to simply be that for people like that there’s currently no easy way to get them invested in the problem, but it really doesn’t have to be that way.

Climate change’s media landscape

In sharp contrast with AI safety, google brings up a number of non text-based material on climate change aimed at a general audience:

AI safety’s media landscape

AI safety does, at least, have a small amount of decent non text-based materials. However, each is somewhat flawed to a greater or lesser extent when it comes to reaching a wide audience:

  • Robert Miles has done some fantastic work making video explainers on AI safety, but these mostly consist of him talking into a camera with the occasional medium quality graphic or animation. This is quite impressive for someone who spends most of their time (I assume) doing actual work in the field, but a far-cry from the condensed high-quality climate change explainers produced by experienced animation teams seen before. Plus, a lot of the time these are fairly academic explainers that fail to provide a strong narrative to latch on to.

  • Universal paperclips is a short browser-based idle game where you play as a paperclip maximizer attempting to… do its thing. This game provides a pretty good depiction of a relatively realistic fast takeoff scenario, but without context it does not seem to provide any strong indications that the events of the game are a real thing that people should be worried could actually happen. I know I wouldn’t be particularly convinced if this was the only thing someone had shown me on the topic. It’s reach is also probably not helped by the fact that it looks mostly like a colorless simple HTML page where you just watch numbers go up.

  • Sentient AI is a trope in a wide variety of sci-fi media, but even remotely realistic depictions of an unfriendly superintelligence rising to power are sorely lacking, with the actual state of things having probably done more harm than good when it comes to AI safety’s credibility as a field.

Further suggestions for media to include in this section are welcome.

Example project idea

While the examples I’ve posted for climate might give anyone reading this ideas for potential projects it may be worth pursuing or funding, I did have a concrete idea for a project that may succeed at grabbing the attention of demographics who may be valuable to the field but wouldn’t find out about it otherwise. I think someone should adapt Gwern’s recent short story, It Looks Like You’re Trying to Take Over the World, into a graphic novel. Disclaimer: I am not an illustrator, I know next to nothing about that industry, I don’t even know what the copyright situation would be, this is just an idea I think someone who would be equipped to pursue it might not have thought of. The main reason I’ve picked out this story in particular is that it does its best to “imagine a hard takeoff scenario using solely known sorts of NN & scaling effects…” I envision a main storyline punctuated by short, illustrated explainers for the concepts involved (this might be harder than I’m imagining, I don’t have the expertise to judge), meant to provide a basic understanding for the layperson or, failing that, to at least confirm to them that at every step of the way to doom there are concrete, known, and potentially intractable failure modes. I don’t expect a project like this to bring AI safety into the mainstream, but I feel like it would be a massive help in allowing someone worried about the issue to express to someone unfamiliar with the problem why they’re concerned about things in an entertaining and accessible way. And it might even be profitable!

Why bother?

In closing this post, I want to verbalize some of my intuitions on why it’s even worth bringing these issues to more people’s attention (this is by no means a comprehensive list):

  • Policy is hard enough to push for these days, but I imagine it would be even harder if almost none of the officials or voters have even heard of or understand the issues underlying a certain proposal.

  • Working on a well-known crisis generally confers more status than working on a relatively obscure academic topic. Climate science as a career path has a status incentive that AI safety does not, whether you’re directly working on technical solutions to the problems or not.

  • Can’t choose a career in a field you’ve never even heard of.

  • While I’m under the impression AI safety as a field is not doing *too* poorly in terms of funding, I still wonder how much more could be done with the vast resources governments could bring to bear on a major issue with vast popular support.