The elephant in the room: even if current major AI companies will align their AIs, there will be hackers (can create viruses with agentic AI component to steal money), rogue states (can decide to use AI agents to spread propaganda and to spy) and military (AI agents in drones and to hack infrastructure). So we need to align the world, not just the models:
Imagine a agentic AI botnet starts to spread on user computers and GPUs. I call it the agentic explosion, it’s probably going to happen before the “intelligence-agency” explosion (intelligence on its own cannot explode, an LLM is a static geometric shape—a bunch of vectors—without GPUs). Right now we are hopelessly unprepared. We won’t have time to create “agentic AI antiviruses”.
To force GPU and OS providers to update their firmware and software to at least have robust updatable blacklists of bad AI (agentic?) models. And to have robust whitelists, in case there will be so many unaligned models, blacklists will become useless.
We can force NVIDIA to replace agentic GPUs with non-agentic ones. Ideally those non-agentic GPUs are like sandboxes that run an LLM internally and can only spit out the text or image as safe output. They probably shouldn’t connect to the Internet, use tools, or we should be able to limit that in case we’ll need to.
This way NVIDIA will have the skin in the game and be directly responsible for the safety of AI models that run on its GPUs.
The same way Apple feels responsible for the App Store and the apps in it, doesn’t let viruses happen.
NVIDIA will want it because it can potentially like App Store take 15-30% cut from OpenAI and other commercial models, while free models will remain free (like the free apps in the App Store).
Replacing GPUs can double NVIDIA’s business, so they can even lobby themselves to have those things. All companies and CEOs want money, have obligations to shareholders to increase company’s market capitalization. We must make AI safety something that is profitable. Those companies that don’t promote AI safety should go bankrupt or be outlawed.
Some ideas, please steelman them:
The elephant in the room: even if current major AI companies will align their AIs, there will be hackers (can create viruses with agentic AI component to steal money), rogue states (can decide to use AI agents to spread propaganda and to spy) and military (AI agents in drones and to hack infrastructure). So we need to align the world, not just the models:
Imagine a agentic AI botnet starts to spread on user computers and GPUs. I call it the agentic explosion, it’s probably going to happen before the “intelligence-agency” explosion (intelligence on its own cannot explode, an LLM is a static geometric shape—a bunch of vectors—without GPUs). Right now we are hopelessly unprepared. We won’t have time to create “agentic AI antiviruses”.
To force GPU and OS providers to update their firmware and software to at least have robust updatable blacklists of bad AI (agentic?) models. And to have robust whitelists, in case there will be so many unaligned models, blacklists will become useless.
We can force NVIDIA to replace agentic GPUs with non-agentic ones. Ideally those non-agentic GPUs are like sandboxes that run an LLM internally and can only spit out the text or image as safe output. They probably shouldn’t connect to the Internet, use tools, or we should be able to limit that in case we’ll need to.
This way NVIDIA will have the skin in the game and be directly responsible for the safety of AI models that run on its GPUs.
The same way Apple feels responsible for the App Store and the apps in it, doesn’t let viruses happen.
NVIDIA will want it because it can potentially like App Store take 15-30% cut from OpenAI and other commercial models, while free models will remain free (like the free apps in the App Store).
Replacing GPUs can double NVIDIA’s business, so they can even lobby themselves to have those things. All companies and CEOs want money, have obligations to shareholders to increase company’s market capitalization. We must make AI safety something that is profitable. Those companies that don’t promote AI safety should go bankrupt or be outlawed.