The corrigibility button will be controlled by the powerful. Not necessarily people like Dario Amodei, probably more like presidents and generals and generic rich assholes.
I’m not sure I believe in the distinction you’re making. Amodei is the CEO of a half-trillion dollar company. Sounds like a generic rich asshole to me; I don’t see why the revolution should treat him any differently than the other кулаки́. Is the difference supposed to be that he talks a better game about humanity and the common good than Travis Kalanick types? You can’t be that gullible, comrade!
And if the powerful don’t need the powerless, the fate of the powerless is bad. That’s also a certainty, given history.
The powerful don’t need the powerless now, so why are they still alive? It’s not just that 15% of the government’s budget goes to welfare. The state barely even kills criminals anymore! Why?
That’s not just a rhetorical question; it’s a literal one that you should be able to answer if you want to persuade people of your worldview. (If it’s to buy votes, felons already can’t vote, so that doesn’t explain all the millions that go to food and housing and judicial appeals for criminals when bullets are so much cheaper. Or rope—you could re-use the rope.)
To be sure, life on welfare in today’s world sucks compared to my life or that of Travis Kalanick. If I were God-Empress of the universe, I’d prefer better for all sentient life. But the realistic relative techno-optimist argument isn’t that all men will be brothers in the Singularity; it’s that giving humanity a nice retirement would cost pennies in a world of nanotechnological abundance, such that making the nano-abundance happen at all matters way more than whether Sam Altman has a billion more yachts than you. Sam Altman probably already has more yachts than me, and it feels fine.
The comment you’re replying to is explicitly denying this. (I wasn’t talking about net-taxpayers with jobs that the power system still needs them to do; I was talking about welfare recipients and criminals who the system presumably doesn’t need and yet are somehow still alive.)
No! Moral AI would be great if you can get it. The reason to care about corrigibility is that you might not succeed at specifying the correct morality, and get something weird and inhuman instead. If you don’t believe the tiny-molecular-squiggle maximizer threat model in the LLM era, imagine an LLM assistant trained to never generate erotica, which generalizes to a superintelligence that decides human sexuality itself is immoral and forcibly modifies humans to not have sexual organs or desires. That’s not what we wanted to happen with the Future! Corrigibility might be a wider target and give you a saving throw.