I feel like this argument breaks down unless leaders are actually waiting for legible problems to be solved before releasing their next updates. So far, this isn’t the vibe I’m getting from players like OpenAI and xAI. It seems like they are releasing updates irrespective of most alignment concerns (except perhaps the superficial ones that are bad for PR). Making illegible problems legible is good either way, but not necessarily as good as solving the most critical problems regardless of their legibility.
I agree there’s a lot of bad signs, but, I think it is kind of the case that their current releases just aren’t that dangerous and if I never thought they were going to be come more dangerous, I don’t know that I’d be that worked up about the current thing.
I feel like this argument breaks down unless leaders are actually waiting for legible problems to be solved before releasing their next updates. So far, this isn’t the vibe I’m getting from players like OpenAI and xAI. It seems like they are releasing updates irrespective of most alignment concerns (except perhaps the superficial ones that are bad for PR). Making illegible problems legible is good either way, but not necessarily as good as solving the most critical problems regardless of their legibility.
I agree there’s a lot of bad signs, but, I think it is kind of the case that their current releases just aren’t that dangerous and if I never thought they were going to be come more dangerous, I don’t know that I’d be that worked up about the current thing.