Terminology point: When I say “a model has a dangerous capability”, I usually mean “a model has the ability to do XYZ if fine-tuned to do so”. You seem to be using this term somewhat differently as model organisms like the ones you discuss are often (though not always) looking at questions related to inductive biases and generalization (e.g. if you train a model to have a backdoor and then train it in XYZ way does this backdoor get removed).
Terminology point: When I say “a model has a dangerous capability”, I usually mean “a model has the ability to do XYZ if fine-tuned to do so”. You seem to be using this term somewhat differently as model organisms like the ones you discuss are often (though not always) looking at questions related to inductive biases and generalization (e.g. if you train a model to have a backdoor and then train it in XYZ way does this backdoor get removed).