Your original objection was to Gyges saying alignment was solveable, at all, not to a claim that it was solvable soon.
. I have yet to understand what prevents the AI-2027-like scenario where misalignment creeps into becoming increasingly difficult to notice
I dont see why that would lead to extinction. Gyges says they are only concerned with doom.
It is sometimes claimed that
sufficiently advanced AI will almost certainly (“inevitably”) kill everyone (“doom”), and therefore
So do I.
I’m specifically addressing the argument for a high probability of near extinction (doom) from AI....not whether it is barely possible, or whether other, less bad outcomes (dystopias) are probable. I’m coming from the centre, not the other extreme
Your original objection was to Gyges saying alignment was solveable, at all, not to a claim that it was solvable soon.
I dont see why that would lead to extinction. Gyges says they are only concerned with doom.
So do I.
I don’t see how that’s a phase change.