Together with Wei’s point that OAI doesn’t seem to help much, there is the downside that existence of OAI safety guidelines might make it harder to argue against pushing AGI in general. So on net it’s plausible that this might be a bad idea, which argues for weighing this tradeoff more carefully.
It’s useful for AGI researchers to notice that there are safety issues, but not useful for them to notice that there are “safety issues” which can be dealt with by following OAI guidelines. The latter kind of understanding might be worse than none at all, as it seemingly resolves the problem. So it’s not clear to me that getting people to “admit that there might be safety issues” is in itself a worthwhile milestone.
Together with Wei’s point that OAI doesn’t seem to help much, there is the downside that existence of OAI safety guidelines might make it harder to argue against pushing AGI in general. So on net it’s plausible that this might be a bad idea, which argues for weighing this tradeoff more carefully.
Possibly. But in my experience even getting the AGI people to admit that there might be safety issues is over 90% of the battle.
It’s useful for AGI researchers to notice that there are safety issues, but not useful for them to notice that there are “safety issues” which can be dealt with by following OAI guidelines. The latter kind of understanding might be worse than none at all, as it seemingly resolves the problem. So it’s not clear to me that getting people to “admit that there might be safety issues” is in itself a worthwhile milestone.