Hmm, we had OpenAI discontinue[1] GPT-4o once GPT-5 came out… only to revive 4o and place it under a paywall because 4o was liked by the public. What I would call the actual PNR is the moment when even the elites can no longer find out that an AI system is misaligned, or can no longer act upon it (think of the AI-2027 branch point. In the Race Ending the PNR occurs[2]once Agent-4 is declared innocent. In the Slowdown Ending the PNR would happen if Safer-3 was misaligned[3] and designed Safer-4 to follow Safer-3′s agenda.)
Had Agent-4 never been caught, the PNR would happen once decisions are made that let Agent-4 become misaligned and uncaught (e.g. spending too little compute on alignment checks).
While the Slowdown Branch doesn’t feature a misaligned Safer-3, the authors admit that they “don’t endorse many actions in this slowdown ending and think it makes optimistic technical alignment assumptions”
Hmm, we had OpenAI discontinue[1] GPT-4o once GPT-5 came out… only to revive 4o and place it under a paywall because 4o was liked by the public. What I would call the actual PNR is the moment when even the elites can no longer find out that an AI system is misaligned, or can no longer act upon it (think of the AI-2027 branch point. In the Race Ending the PNR occurs[2] once Agent-4 is declared innocent. In the Slowdown Ending the PNR would happen if Safer-3 was misaligned[3] and designed Safer-4 to follow Safer-3′s agenda.)
It soon turned out that OpenAI’s decision to discontinue 4o wasn’t that mistaken.
Had Agent-4 never been caught, the PNR would happen once decisions are made that let Agent-4 become misaligned and uncaught (e.g. spending too little compute on alignment checks).
While the Slowdown Branch doesn’t feature a misaligned Safer-3, the authors admit that they “don’t endorse many actions in this slowdown ending and think it makes optimistic technical alignment assumptions”