Regarding no Safe Natural Intelligence: I agree that there is no such thing, but this is not really a strong argument against?
I think it’s a pretty strong argument. There are no humans I’d trust with the massively expanded capabilities that AI will bring, so I have to believe that the training methods for humans are insufficient.
We WANT divergence from “business as usual” human beliefs and actions, and one of the ways to get there is by different specifications and training mechanisms. The hard part is we don’t yet know how to specify precisely how we want it to differ.
I think it’s a pretty strong argument. There are no humans I’d trust with the massively expanded capabilities that AI will bring, so I have to believe that the training methods for humans are insufficient.
We WANT divergence from “business as usual” human beliefs and actions, and one of the ways to get there is by different specifications and training mechanisms. The hard part is we don’t yet know how to specify precisely how we want it to differ.