Agent-foundations researcher. Working on Synthesizing Standalone World-Models, aiming at a timely technical solution to the AGI risk fit for worlds where alignment is punishingly hard and we only get one try.
Currently looking for additional funders ($1k+, details). Consider reaching out if you’re interested, or donating directly.
Or get me to pay you money ($5-$100) by spotting holes in my agenda or providing other useful information.
On balance, I support banning/obstructing datacenter buildout. That said, I’m not actually sure whether that impacts the omnicide risk positively or negatively.
I don’t think the LLM paradigm is AGI-complete. I don’t have utter confidence in that, but I think it’s more likely than not. And in worlds in which it is non-AGI-complete, the labs’ obsession with them is very helpful. They’re wasting macroeconomic amounts of money on it, pouring most of our generation’s AI-researcher talent into it, distracting funding from other AGI approaches. If LLMs then fail to usher in the Singularity, if it does turn out to be a bubble that pops (e. g., in 2030-2032, when the ability to aggressively scale compute is supposed to run out), this should cause another AI winter. AGI would become a decidedly unsexy thing to work on once more, in industry and probably in academia both.
What would obstructing the LLM paradigm (via various compute limits) do? Well, it may cause the AGI megacorps to start looking into other directions now, while they still have massive amount of unspent manpower and capital. Which may lead to someone “succeeding” sooner.[1]
Perhaps one shouldn’t interrupt their enemy while they’re making a mistake.
Or perhaps that’s not how the story goes. Perhaps: LLMs are a strong signal that AI is a massively powerful technology, a signal legible to many more people than theoretical arguments. They’re attracting macroeconomic amounts of funding to AI, funneling a large fraction of our generation’s talent towards working on AI, fueling inter-company and geopolitical AI-race dynamics. And while they cause other AGI approaches to be relatively neglected, they cause so much more absolute attention to be pointed at the AI industry that the amounts of funding/talent going into non-LLM AGI routes is still much greater than in the no-LLMs counterfactual. On top of that, even if LLMs aren’t AGI-complete in themselves, they may still be useful enough to speed up SWE/research along other AGI paradigms as well.
And perhaps effectively banning LLMs would cause a chilling effect much greater than if they were merely a research bet that didn’t work out. It would be a signal that AGI research is something the world does not want to see in any form, turning funding and talent away from the whole endeavor.
I’m not really sure which model is correct. Which AI Winter would be colder and broader: one caused by a proactive LLM ban, or one caused by LLMs’ dramatic betrayal of investor hopes? Do LLMs currently cause net increase or decrease of attention to non-LLM AGI research? Can non-AGI LLMs significantly speed up non-LLM AGI research if allowed to continue scaling? And, of course: is the premise that LLMs are non-AGI-complete even correct?
I’m on balance in favor of obstructing LLMs and upper-bounding the available compute. “We must not get in the way of the current paradigm because they are totally making a mistake so all this investment is actually delaying AGI!” has the flavor of some galaxy-brained 4D chess nonsense. But I’m not fully sure.
(Also, worlds in which politicians/the public are concerned about AI enough to ban datacenter buildout is of course a more promising world than the one in which they’re not. This is true regardless of whether the direct effects of banning datacenter construction are positive or negative. In such worlds, they’d probably be willing to engage in additional anti-AGI interventions. So perhaps datacenter-impeding is worth supporting just to put us in those worlds, shift the Overton window that way? Not sure whether it’s causal or evidential, though.)
There are also arguments that this would be bad even in the LLMs-are-AGI-complete worlds, because – the argument would go – while LLMs may be nontrivial to align, they’re easier to align than some paradigms that may replace them. I don’t put much stock into this argument, it usually relies on assuming the LLM masks/personas is the entity of interest that we need to align. I don’t put zero stock in it, though, I guess.