In Emergence of Simulators and Agents, my AISC collaborators and I suggested that whether consequentialist or simulator-like cognition (which one could describe as a subcategory of process-based reasoning) emerges depends critically on environmental and training conditions, particularly the “feedback gap” (the delay, uncertainty, or inference depth between action and feedback). Large feedback gaps select for instrumental reasoning and power-seeking; small feedback gaps select for imitation and compression. As examples, LLMs are trained primarily via SSL (minimal feedback gap) and display predominantly simulator-like behavior, whereas RL-trained AlphaZero is clearly agentic.
The dynamic you describe of patterns steering toward states where they have more steering capacity outcompeting other patterns is real, but may be context dependent. If so, CCCT requires both: (1) the conditions for consequentialist reasoning being advantageous being inevitable and (2) consequentialism being inevitable given those conditions.
Claim 1, regarding conditions, is the part that needs defending. The “consequentialism is inevitable” argument requires showing either:
Market/competitive forces will inevitably push toward large-feedback-gap deployments (agentic AI doing long-horizon tasks), or
Even in small-feedback-gap contexts, consequentialist subpatterns will somehow emerge and take over.
Without establishing one of these (1 seems plausible to me, but that’s an intuitive claim), the convergence thesis describes a risk contingent on our choices, not an inevitability. Of course, process-based reasoning is not the same as “safe” by any means, but that shifts the terrain of the argument.
Fast-follow on frontier capabilities; compete on useful applications and convenient integration. This works all the way up to recursive self improvement or superintelligence, at which point none of this matters.
Lobby for regulation that slows down everyone. Justify this practice to your stockholders by pointing out that your competitive advantage is in applications and so fair regulation hits your competitors harder.
Publish scary demos and evals. Show them in the context of your competitors models first, with an acknowledgement that the issues also apply to your products. This won’t change anything directly, but it will be useful ammunition for external activists who can push for things you can’t.
Encourage (or fail to effectively discourage) union organizing among your employees so that you are not bound to hyper-optimizing for stockholder short term ROI. Maintain a collaborative relationship with union leadership, since you both have an interest in the company’s success. There is no guarantee that union leadership will be as safety conscious as you are, but it is impossible for them to be less safety conscious than stockholders, so this is a net win regardless.
Write a list of voluntary commitments, not just for slowing down but also regarding lobbying, based on what you would like to see if you had full consensus with your significant competitors. Include effective measures for monitoring and enforcement. Don’t worry about competitiveness, that doesn’t matter because none of this is binding until a critical mass has signed on. When it looks good, circulate the agreement to your competitors and invite their feedback, as long as that feedback remains within the spirit of the agreement. Let them actually reject the agreement rather than just assuming that they will. Maintain a paper trail. If possible, publicize any rejection and use it against them.