[Question] Where are the AI safety replications?

A popular topic on LessWrong is that much of science fails to replicate, because of bad incentives and increasingly complex statistics. I think most of us want more replication studies to be published, but it’s of course very difficult to effect that change in mainstream science.

For AI safety specifically, though, it seems like there’s more opportunity to change these norms. So what’s missing exactly? External funding incentives? Better scientific rigor making it less important? People still consider it low-status? A simple lack of coordinated effort to do it? Something else entirely?