I’m not optimizing for raising awareness via an “obvious AI disaster” due to multiple reasons, including the huge risk to the reputation of the AI safety community and the unilateralist’s curse.
I do think that when considering whether to invest in an effort which might prevent recoverable near-term AI accidents, one should consider the possibility that the effort would prevent pivotal events (e.g. one that would have enabled useful governance solutions resulting in more time for alignment research).
Efforts that prevent recoverable near-term AI accidents might be astronomically net-positive if they help make AI alignment more mainstream in the general ML community.
(anyone who thinks I shouldn’t discuss this publicly is welcome to let me know via a PM or anonymously here)
I’m not optimizing for raising awareness via an “obvious AI disaster” due to multiple reasons, including the huge risk to the reputation of the AI safety community and the unilateralist’s curse.
I do think that when considering whether to invest in an effort which might prevent recoverable near-term AI accidents, one should consider the possibility that the effort would prevent pivotal events (e.g. one that would have enabled useful governance solutions resulting in more time for alignment research).
Efforts that prevent recoverable near-term AI accidents might be astronomically net-positive if they help make AI alignment more mainstream in the general ML community.
(anyone who thinks I shouldn’t discuss this publicly is welcome to let me know via a PM or anonymously here)