I think this particular issue has less to do with public sentiment & more to do with problems that require solutions which would inconvenience you today for a better tomorrow.
Like climate change: it is an issue everyone recognizes will massively impact the future negatively (to the point where multiple forecasts suggest trillions of dollars of losses). Still, since fixing this issue will cause prices of everyday goods to rise significantly and force people into switching to green alternatives en masse, no one advocates for solutions. News articles get released each year stating record high temperatures & natural disaster rates; people complain that the seasons have been getting more extreme each passing year (for example, the monsoon in Northern India has been really inconsistent for several years now—talking from personal experience). Yet, changes are gradual (carbon tax is still wildly unpopular, and even though the public sentiment around electric cars has been getting tame, people still don’t advocate for anti-car measures). Compare this to Y2K; it was known that the “bug” would be massively catastrophic, and even though it was massively expensive to fix, they did fix it. Why? Because the issue didn’t affect the lives of common folks substantively.
Though my model is certainly not all-encompassing, like how the problem of CFCs causing Ozone depletion in the upper atmosphere was largely solved, even though it very much did impact people’s everyday lives & did cost a lot of money & even required global co-operation. I guess there is a tipping point on the inconvenience caused v/s perceived threat graph when people start mobilizing for the issue.
PS: It’s funny how I ended at the crux of the original article, you should now be able to apply threshold modelling for this issue, since perceived threat largely depends on how many other people (in your local sphere) are shouting that this issue needs everyone’s attention.
You’re very right, in addition to people not working on AI risk because they don’t see others working on it, you also have the problem that people aren’t interested in working on future risk to begin with.
I do think that military spending is a form of addressing future risks, and people are capable of spending a lot on it.
I guess military spending doesn’t inconvenience you today, because there already are a lot of people working in the military, who would lose their jobs if you reduced military spending. So politicians will actually make more people lose their jobs if they reduced military spending.
Hmm. But now that you bring up climate change, I do think that there is hope… because some countries do regulate a lot and spend a lot on climate change, at least in Europe. And it is a complex scientific topic which started with the experts.
Maybe the AI Notkilleveryone movement should study what worked and failed for the green movement.
That’s an interesting idea. The military would undoubtedly care about AI alignment — they’d want their systems to operate strictly within set parameters. But the more important question is: do we even want the military to be investing in AI at all? Because that path likely leads to AI-driven warfare. Personally, I’d rather live in a world without autonomous robotic combat or AI-based cyberwarfare.
But as always, I will pray that some institution (like the EU) leads the charge & start instilling it into people’s heads that this is a problem we must solve.
Oops, I didn’t mean we should involve the military in AI alignment. I meant the military is an example of something working on future threats, suggesting that humans are capable of working on future threats.
I think the main thing holding back institutions is that public opinion does not believe in AI risk. I’m not sure how to change that.
To sway public opinion about AI safety, let us consider the case of nuclear warfare—a domain where long-term safety became a serious institutional concern. Nuclear technology wasn’t always surrounded by protocols, safeguards, and watchdogs. In the early days, it was a raw demonstration of power: the bombs dropped on Hiroshima and Nagasaki were enough to show the sheer magnitude of destruction possible. That spectacle shocked the global conscience. It didn’t take long before nation after nation realized that this wasn’t just a powerful new toy, but an existential threat. As more countries acquired nuclear capabilities, the world recognized the urgent need for checks, treaties, and oversight. What began as an arms race slowly transformed into a field of serious, respected research and diplomacy—nuclear safety became a field in its own right.
The point is: public concern only follows recognition of risk. AI safety, like nuclear safety, will only be taken seriously when people see it as more than sci-fi paranoia. For that shift to happen, we need respected institutions to champion the threat. Right now, it’s mostly academics raising the alarm. But the public—especially the media and politicians—won’t engage until the danger is demonstrated or convincingly explained. Unfortunately for the AI safety issue, evidence of AI misalignment causing significant trouble will probably mean it’s too late. Adding fuel to this fire is the fact that politicians aren’t gonna campaign about AI safety if the corpos in your country don’t want to & your enemies are already neck-to-neck in AI dev.
In my subjective opinion, we need the AI variant of Hiroshima. But I’m not too keen on this idea, for it is a rather dreadful thought.
Edit: I should clarify what I mean by “the AI variant of Hiroshima.” I don’t think a large-scale inhuman military operation is necessary (as I already said, I don’t want AI warfare). What I mean instead is something that causes significant damage & makes it to newspaper headlines worldwide. Examples: strong evidence that AI swayed the presidential election one way; a gigantic economic crash caused because of a rogue AI (not the AI bubble bursting); millions of jobs being lost in a short timeframe because of one revolutionary model, which then snaps because of misalignment; etc. There are still dreadful, but at least no human lives are lost & it gets the point across that AI safety is an existential issue.
I think different kinds of risks have different “distributions” of how much damage they do. For example, the majority of car crashes causes no injuries (but damage to the cars), a smaller number causes injuries and some causes fatalities, and the worst ones can cause multiple fatalities.
For other risks like structural failures (of buildings, dams, etc.) the distribution has a longer tail: in the worst case very many people can die. But the distribution still tapers off towards greater number of fatalities, and people sort have have a good idea of how bad it can get before the worst version happens.
For risks like war, the distribution has an even longer tail, and people are often caught by surprise how bad they can get.
But for AI risk, the distribution of damage caused is very weird. You have one distribution for AI causing harm due to its lack of common sense, where it might harm a few people, or possibly cause one death. Yet you have another distribution for AI taking over the world, with a high probability of killing everyone, a high probability of failing (and doing zero damage), and only a tiny bit of probability in between.
It’s very very hard to learn from experience in this case. Even the biggest wars tend to surprise everyone (despite having a relatively more predictable distribution).
That’s a cool way to frame damage risks, but I think your distribution for AI damage is for ASI, not AGI. I think it’s very reasonable that an AGI-based system may cause the type of damage that I am talking about.
Even if you believe that as soon as we achieve AGI, we’ll accelerate to ASI because AGI by definition is self-improving, it still takes time to train a model, and research is slow. I hope that the window b/w AGI & ASI is large enough for such a “Hiroshima event” to occur, so humanity wakes up to the risks of maligned AI systems.
PS: Sorry for the late response, I was offline for a couple of days
No need to say sorry for that! On a forum, there is no expectation to receive a reply. If every reply obligated the recipient to make another reply, comment chains will drag on forever.
You can freely wait a year before replying.
I’m worried that once a “Hiroshima event” occurs, humanity won’t have another chance. If the damage is caused by the AGI/ASI taking over places, then the more power it obtains, the more it can obtain even more power, so it won’t stop at any scale.
If the damage is caused by bad actors using an AGI to invent a very deadly technology, there is a decent chance humanity can survive, but it’s very uncertain. A technology can never be uninvented, and more and more people will know about it.
I think this particular issue has less to do with public sentiment & more to do with problems that require solutions which would inconvenience you today for a better tomorrow.
Like climate change: it is an issue everyone recognizes will massively impact the future negatively (to the point where multiple forecasts suggest trillions of dollars of losses). Still, since fixing this issue will cause prices of everyday goods to rise significantly and force people into switching to green alternatives en masse, no one advocates for solutions. News articles get released each year stating record high temperatures & natural disaster rates; people complain that the seasons have been getting more extreme each passing year (for example, the monsoon in Northern India has been really inconsistent for several years now—talking from personal experience). Yet, changes are gradual (carbon tax is still wildly unpopular, and even though the public sentiment around electric cars has been getting tame, people still don’t advocate for anti-car measures).
Compare this to Y2K; it was known that the “bug” would be massively catastrophic, and even though it was massively expensive to fix, they did fix it. Why? Because the issue didn’t affect the lives of common folks substantively.
Though my model is certainly not all-encompassing, like how the problem of CFCs causing Ozone depletion in the upper atmosphere was largely solved, even though it very much did impact people’s everyday lives & did cost a lot of money & even required global co-operation. I guess there is a tipping point on the inconvenience caused v/s perceived threat graph when people start mobilizing for the issue.
PS: It’s funny how I ended at the crux of the original article, you should now be able to apply threshold modelling for this issue, since perceived threat largely depends on how many other people (in your local sphere) are shouting that this issue needs everyone’s attention.
You’re very right, in addition to people not working on AI risk because they don’t see others working on it, you also have the problem that people aren’t interested in working on future risk to begin with.
I do think that military spending is a form of addressing future risks, and people are capable of spending a lot on it.
I guess military spending doesn’t inconvenience you today, because there already are a lot of people working in the military, who would lose their jobs if you reduced military spending. So politicians will actually make more people lose their jobs if they reduced military spending.
Hmm. But now that you bring up climate change, I do think that there is hope… because some countries do regulate a lot and spend a lot on climate change, at least in Europe. And it is a complex scientific topic which started with the experts.
Maybe the AI Notkilleveryone movement should study what worked and failed for the green movement.
That’s an interesting idea. The military would undoubtedly care about AI alignment — they’d want their systems to operate strictly within set parameters. But the more important question is: do we even want the military to be investing in AI at all? Because that path likely leads to AI-driven warfare. Personally, I’d rather live in a world without autonomous robotic combat or AI-based cyberwarfare.
But as always, I will pray that some institution (like the EU) leads the charge & start instilling it into people’s heads that this is a problem we must solve.
Oops, I didn’t mean we should involve the military in AI alignment. I meant the military is an example of something working on future threats, suggesting that humans are capable of working on future threats.
I think the main thing holding back institutions is that public opinion does not believe in AI risk. I’m not sure how to change that.
To sway public opinion about AI safety, let us consider the case of nuclear warfare—a domain where long-term safety became a serious institutional concern. Nuclear technology wasn’t always surrounded by protocols, safeguards, and watchdogs. In the early days, it was a raw demonstration of power: the bombs dropped on Hiroshima and Nagasaki were enough to show the sheer magnitude of destruction possible. That spectacle shocked the global conscience. It didn’t take long before nation after nation realized that this wasn’t just a powerful new toy, but an existential threat. As more countries acquired nuclear capabilities, the world recognized the urgent need for checks, treaties, and oversight. What began as an arms race slowly transformed into a field of serious, respected research and diplomacy—nuclear safety became a field in its own right.
The point is: public concern only follows recognition of risk. AI safety, like nuclear safety, will only be taken seriously when people see it as more than sci-fi paranoia. For that shift to happen, we need respected institutions to champion the threat. Right now, it’s mostly academics raising the alarm. But the public—especially the media and politicians—won’t engage until the danger is demonstrated or convincingly explained. Unfortunately for the AI safety issue, evidence of AI misalignment causing significant trouble will probably mean it’s too late.
Adding fuel to this fire is the fact that politicians aren’t gonna campaign about AI safety if the corpos in your country don’t want to & your enemies are already neck-to-neck in AI dev.
In my subjective opinion, we need the AI variant of Hiroshima. But I’m not too keen on this idea, for it is a rather dreadful thought.
Edit: I should clarify what I mean by “the AI variant of Hiroshima.” I don’t think a large-scale inhuman military operation is necessary (as I already said, I don’t want AI warfare). What I mean instead is something that causes significant damage & makes it to newspaper headlines worldwide. Examples: strong evidence that AI swayed the presidential election one way; a gigantic economic crash caused because of a rogue AI (not the AI bubble bursting); millions of jobs being lost in a short timeframe because of one revolutionary model, which then snaps because of misalignment; etc. There are still dreadful, but at least no human lives are lost & it gets the point across that AI safety is an existential issue.
I think different kinds of risks have different “distributions” of how much damage they do. For example, the majority of car crashes causes no injuries (but damage to the cars), a smaller number causes injuries and some causes fatalities, and the worst ones can cause multiple fatalities.
For other risks like structural failures (of buildings, dams, etc.) the distribution has a longer tail: in the worst case very many people can die. But the distribution still tapers off towards greater number of fatalities, and people sort have have a good idea of how bad it can get before the worst version happens.
For risks like war, the distribution has an even longer tail, and people are often caught by surprise how bad they can get.
But for AI risk, the distribution of damage caused is very weird. You have one distribution for AI causing harm due to its lack of common sense, where it might harm a few people, or possibly cause one death. Yet you have another distribution for AI taking over the world, with a high probability of killing everyone, a high probability of failing (and doing zero damage), and only a tiny bit of probability in between.
It’s very very hard to learn from experience in this case. Even the biggest wars tend to surprise everyone (despite having a relatively more predictable distribution).
That’s a cool way to frame damage risks, but I think your distribution for AI damage is for ASI, not AGI. I think it’s very reasonable that an AGI-based system may cause the type of damage that I am talking about.
Even if you believe that as soon as we achieve AGI, we’ll accelerate to ASI because AGI by definition is self-improving, it still takes time to train a model, and research is slow. I hope that the window b/w AGI & ASI is large enough for such a “Hiroshima event” to occur, so humanity wakes up to the risks of maligned AI systems.
PS: Sorry for the late response, I was offline for a couple of days
No need to say sorry for that! On a forum, there is no expectation to receive a reply. If every reply obligated the recipient to make another reply, comment chains will drag on forever.
You can freely wait a year before replying.
I’m worried that once a “Hiroshima event” occurs, humanity won’t have another chance. If the damage is caused by the AGI/ASI taking over places, then the more power it obtains, the more it can obtain even more power, so it won’t stop at any scale.
If the damage is caused by bad actors using an AGI to invent a very deadly technology, there is a decent chance humanity can survive, but it’s very uncertain. A technology can never be uninvented, and more and more people will know about it.
Or more! (I was delighted to receive this reply.)