I think that you also haven’t assessed the Rogue Replication Timeline, nor, well, my take where the AI is unalignable to the Spec because the Spec and/or the training data[1] are biased. It also seems to imply that Agent-3 or Agent-2 might actively collude with Agent-4 instead of simply failing to catch it.
Which is most Western sources. The bias could be so great that a recent post mentions “Zack Davis documenting endorsement of anti-epistemology (see Where to Draw the Boundaries? and A Hill of Validity in Defense of Meaning) to placate trans ideology even many important transgender Rationality community members overtly reject.”
I think that you also haven’t assessed the Rogue Replication Timeline, nor, well, my take where the AI is unalignable to the Spec because the Spec and/or the training data[1] are biased. It also seems to imply that Agent-3 or Agent-2 might actively collude with Agent-4 instead of simply failing to catch it.
P.S. Shanzon might have used the fact that Narrow Misalignment is Hard, Emergent Misalignment is Easy as a reference.
Which is most Western sources. The bias could be so great that a recent post mentions “Zack Davis documenting endorsement of anti-epistemology (see Where to Draw the Boundaries? and A Hill of Validity in Defense of Meaning) to placate trans ideology even many important transgender Rationality community members overtly reject.”
Yeah I’ve read most of the submissions but still haven’t gotten around to finishing them & writing up the results, sorry!