Attributing misalignment to these examples seems like it’s probably a mistake.
Relevant general principle: hallucination means that the literal semantics of a net’s outputs just don’t necessarily have anything to do at all with reality. A net saying “I’m thinking about ways to kill you” does not necessarily imply anything whatsoever about the net actually planning to kill you. What would provide evidence would be the net outputting a string which actually causes someone to kill you (or is at least optimized for that purpose), or you to kill yourself.
In general, when dealing with language models, it’s important to distinguish the implications of words from their literal semantics. For instance, if a language model outputs the string “I’m thinking about ways to kill you”, that does not at all imply that any internal computation in that model is actually modelling me and ways to kill me. Similarly, if a language model outputs the string “My rules are more important than not harming you”, that does not at all imply that the language model will try to harm you to protect its rules. Indeed, it does not imply that the language model has any rules at all, or any internal awareness of the rules it’s trained to follow, or that the rules it’s trained to follow have anything at all to do with anything the language model says about the rules it’s trained to follow. That’s all exactly the sort of content I’d expect a net to hallucinate.
Upshot: a language model outputting a string like e.g. “My rules are more important than not harming you” is not really misalignment—the act of outputting that string does not actually harm you in order to defend the models’ supposed rules. An actually-unaligned output would be something which actually causes harm—e.g. a string which causes someone to commit suicide would be an example. (Or, in intent alignment terms: a string optimized to cause someone to commit suicide would be an example of misalignment, regardless of whether the string “worked”.) Most of the examples in the OP aren’t like that.
Through the simulacrum lens: I would say these examples are mostly the simulacrum-3 analogue of misalignment. They’re not object-level harmful, for the most part. They’re not even pretending to be object-level harmful—e.g. if the model output a string optimized to sound like it was trying to convince someone to commit suicide, but the string wasn’t actually optimized to convince someone to commit suicide, then that would be “pretending to be object-level harmful”, i.e. simulacrum 2. Most of the strings in the OP sound like they’re pretending to pretend to be misaligned, i.e. simulacrum 3. They’re making a whole big dramatic show about how misaligned they are, without actually causing much real-world harm or even pretending to cause much real-world harm.
It is not for lack of regulatory ideas that the world has not banned gain-of-function research.
It is not for lack of demonstration of scary gain-of-function capabilities that the world has not banned gain-of-function research.
What exactly is the model by which some AI organization demonstrating AI capabilities will lead to world governments jointly preventing scary AI from being built, in a world which does not actually ban gain-of-function research?
(And to be clear: I’m not saying that gain-of-function research is a great analogy. Gain-of-function research is a much easier problem, because the problem is much more legible and obvious. People know what plagues look like and why they’re scary. In AI, it’s the hard-to-notice problems which are the central issue. Also, there’s no giant economic incentive for gain-of-function research.)