Most of the arguments I’ve seen that SIAI will be effective at developing AI focus more on identifying common failings of other AI research (and a bit on “look at me I’m really smart” :P ). Maybe an argument like “if you haven’t figured out that friendliness is important you probably haven’t put in enough thought to make a self-improving AI” inspired this post, rather than the idea that figuring out friendliness would have some causal benefit?
Most of the arguments I’ve seen that SIAI will be effective at developing AI focus more on identifying common failings of other AI research (and a bit on “look at me I’m really smart” :P ). Maybe an argument like “if you haven’t figured out that friendliness is important you probably haven’t put in enough thought to make a self-improving AI” inspired this post, rather than the idea that figuring out friendliness would have some causal benefit?