AI Safety at frontier labs is essentially a bunch of shallow instincts/behaviors covering up an ultimately Pythian power-maximizing entity
You could replace “AI safety at frontier labs” with “pro-social policy at powerful organizations” and this sentence would probably still be true, no?
Over the last four years, has anything happened that actually contradicted this model? An event where an AGI lab actually did something in the name of safety that meaningfully cost it? Something that didn’t predictably end up working out to instead boost the lab’s PR/fundability and improve its products, or wasn’t so cheap for a lab to do as to not worth the attention of its Pythian core?
What could they have done otherwise? If I had to venture an example, I’d say ” any support for legislation binding them to their (stated) voluntary commitments”.
You could replace “AI safety at frontier labs” with “pro-social policy at powerful organizations” and this sentence would probably still be true, no?
What could they have done otherwise? If I had to venture an example, I’d say ” any support for legislation binding them to their (stated) voluntary commitments”.