Why Alignment Fails Without a Functional Model of Intelligence

My recent paper, Why AI Alignment Is Not Reliably Achievable Without a Functional Model of Intelligence: A Model-Theoretic Proof argues that alignment under conceptual novelty isn’t something you can train into a system or enforce externally. It must be structurally guaranteed from within—and that guarantee depends on the presence of specific internal functions. These functions, taken together, define what I refer to as the Functional Model of Intelligence (FMI). The paper shows that even capable systems, and even those that generalize well under most conditions, will eventually fail to preserve alignment unless they possess a specific internal structure. Without it, misalignment isn’t a matter of insufficient training or poor objective design. It’s baked into the architecture. Even if this particular FM turns out to be incomplete, the core result still holds: alignment requires a complete and recursively evaluable functional model of intelligence. That is, some FMI must exist for alignment to be achievable under novelty. The proof shows why.

Visual Explanation + Free Workshop

If you’re curious about this result but prefer to think in diagrams rather than equations, I’m running a workshop on Visualizing AI Alignment on August 10, 2025.

The workshop will walk through the main ideas of the paper—like recursive coherence, conceptual drift, and structural completeness—using fully visual reasoning tools. The goal is to make the core constraints of alignment immediately visible, even to those without a formal logic background.

Virtual attendance is free, and we’re accepting brief submissions through July 24, 2025.

Call for Participation and submission info here