I continue to be deeply, deeply skeptical of the whole ‘we need exactly the best possible frontier models to do good alignment work’ argument, that whatever the current best option is must be both necessary and sufficient.
I know asking other people to do work for you is generally, unfair, but would it be too much to ask for you to include a section along the lines of “AI alignment breakthroughs this week”?
There’s a new innovation at least as significant as this or this almost every week, and it would be a bit heartening to see that you were actually tracking progress towards solving the alignment problem instead of just focusing on pausing AI progress.
If your claim is that we can make progress on alignment without working on cutting-edge models, then you should actually be tracking/encouraging such research.
I know asking other people to do work for you is generally, unfair, but would it be too much to ask for you to include a section along the lines of “AI alignment breakthroughs this week”?
There’s a new innovation at least as significant as this or this almost every week, and it would be a bit heartening to see that you were actually tracking progress towards solving the alignment problem instead of just focusing on pausing AI progress.
If your claim is that we can make progress on alignment without working on cutting-edge models, then you should actually be tracking/encouraging such research.
Okay, I’m going to try and do this once a week for a month to see if people find it useful.
Here is the post for the first week.