evals for things like automated ML R&D are only worrying for people who already believe in AI xrisk
I don’t think this is true – or, more specifically, I think there are a lot of people who will start to worry about AI xrisk if things like automated ML R&D pick up. Most people who dismiss AI xrisk I don’t think do so because they think intelligence is inherently good, but instead because AI xrisk just seems too “scifi.” But if AI is automating ML R&D, then the idea of things getting out of hand won’t feel as scifi. In principle, people should be able to separate the question of “will AI soon be able to automate ML R&D” from the question of “if AI could automate ML R&D, would it pose an xrisk”, but I think most low-decouplers struggle to make this separation. For the kind of reaction that a “normal” person will have to automated ML R&D, I think this reaction from a CBS host interviewing Hinton is informative.
(I agree with your general point that it’s better to focus on worrying capabilities, and also I agree with some of your other points, such as how demos might be more useful than evals.)
I think this is a good point. At the same time, I suspect the main reason we’re likely to be wiser a year from now is that we’ll have done stuff over the coming year that we’ll learn from. And the more we spend over the next year, the more we’ll be able to do, leading to more learning. In some ways this feels like “yes, maybe from an individual level it’ll feel better to wait and learn more, but your spending now not only lets you learn better but also lets others learn better.” I think the factor I’m pointing to is actually substantial, in particular if you’re funding highly promising areas that are relatively new and that others are skeptical of or feel insufficiently knowledgeable about.