Unprecedented Catastrophes Have Non-Canonical Probabilities
The chance of a bridge failing, of an asteroid striking the earth, of whether your child will get into Harvard for a special reason only you know, and of whether AI will kill everyone are all things that can be expressed with probability, but they are not all the same type of probability. There is a structural difference in the “probability” of a bridge collapsing being a (
I know what you are probably thinking: “we’ve heard this before...you are going to go down some rabbithole about how P(doom) is tribally divisive, or that it distracts from gears-level models, or maybe another Popperian argument that it is philosophically meaningless to assign probabilities to unprecedented events.”
Don’t worry! This post is none of those things. (If you want to double-check, feel free to skip down to the Why Is This Different section.)
And to be clear at the outset: I am not anti-Bayesian, and I definitely think quantifying uncertainty is generally a good idea. However; I am going to argue that probability estimates for unprecedented catastrophes are non-canonical[1], and I am going to do so within the confines of Bayesian machinery itself, using likelihood ratios, posterior convergence, and algorithmic information theory.
First, what do I mean by a canonical probability? A canonical probability is stable across reasonable scientific frameworks given the available evidence. For example, a
However, genuinely unprecedented catastrophic risks are non-canonical because the available non-catastrophic data structurally refuses to wash out your priors. Without past events to sharply discriminate between competing models the probability remains only partially identified and your resulting estimate is only a measurement of your chosen ontology rather than a measurement of the world.[2]
The illusion of precise catastrophic probabilities can be traced to two distinct mathematical failures: one based on how we process evidence, and the other in how we define the event itself. Below, I will break down these two grounds of failure, using Nick Bostrom’s recent paper “Optimal Timing for Superintelligence” and Eliezer Yudkowsky’s asteroid rebuttal parable as an example.[3] Finally, I’ll offer a concrete suggestion of how to escape them and restore empirical rigor.
Data Doesn’t Resolve Ontology
Before showing where probabilities of unprecedented catastrophes break, let’s briefly return to the asteroid impact example to show when they do not break. The key idea with an asteroid is that we can gather non-event data that directly updates the impact parameter and we can washout whatever prior we started with. Econometrics has a great way of expressing this concept: it is called being point-identified.[4] The data has forced the parameter to a specific, narrow point.
Now let’s turn to catastrophic probabilities that do break. The battle between Bostrom and Yudkowsky is a great example of genuinely different ontological starting points, but I don’t mean to make overarching claims about their complete intellectual positions. They draw on multiple frameworks and are not formal description languages themselves.
Bostrom recently wrote a paper that framed building AGI as a “risky surgery for a condition that will otherwise prove fatal”. The crux of his argument is that if we want to maximize QALY, high probabilities of catastrophe from AGI are worth accepting because everyone alive today will die anyway if we do nothing. The surgery-framing model is structurally native and simple to state.[5]
In response, Yudkowsky, wrote a mocking parable on X called the “Asteroid of Immortality”. In Yudkowsky’s framework, AGI is like an incoming asteroid, and Bostrom’s reasoning for eternal life amounts to logic of a cult rather than a surgeon. In Yud’s alignment-centric framework, “doom via misaligned optimization” is the natively simple, default prediction.
To understand where probabilities associated with these views will come from, we first need to formalize what a framework is. Think of a framework as the programming language of their worldviews: it determines which explanations are naturally simple to express, and which are unnatural and super complex. Imagine two ideal rational agents using two different Universal Turing Machines as their formal description languages (their “programming languages”). Machine
There is a fundamental problem, however: the data that we are gathering right now cannot help us choose which of these frameworks is likely to be right. This is quite counterintuitive because when we Bayesian update we expect that new data will wash out our subjective priors. But unprecedented catastrophes hit a concept I call The Evidential Screening Property. Another year without world-ending events gives us more data, but this data fits equally well into a model that predicts it’s all over and a model that predicts it’s all ok. Because both models perfectly predict the prefix of history (the track record of AI development leading up to the current moment), the likelihood ratio between them is approximately 1. The catastrophe parameter is entirely screened off from the evidence, leaving our prior beliefs untouched.
To put this in more precise terms, we are not getting more useful data over time, just more noise that fits both theories perfectly. Formally, the log-likelihood ratio is bounded by a tiny number of bits,
This takes us back to econometrics and a concept called Partial Identification. I briefly mentioned point identification at the start of this section, that is, we can use data we gather from something like an extinction-asteroid to force a sharp single risk estimate. In contrast, the non-catastrophic prefix of AI history structurally fails to do this and the data only constrains the parameter to a hugely ambiguous region. This is what Manski calls being partially identified: the hugely ambiguous region of data is a prior-free identification region
The Complexity Inversion and the Failure of Model Averaging
If you are familiar with Bayesian math, as many of you probably are, you may now be thinking something like this:
Ok, the log-odds between Bostrom and Yud’s specific and extreme models might shift wildly based on your ontology, but with Bayes we averaging over the mixture of all possible models, not just one. When we have a rich hypothesis space a diffuse and large number of middle-mass moderate models will anchor the probability and reduce the extreme swings. In other words, something like the aggregate of P(doom) will remain stable.
This is a smart defense, but I think it empirically fails for models used today under a specific condition I will call the Complexity Inversion: switching frameworks flips which type of futures are considered simple and which are considered complex.
Recall earlier that I likened frameworks to programming languages. In algorithmic information theory (“AIT”), Kolmogorov complexity is the rule that penalizes a theory based on how many lines of code it takes to express it. The longer the code, the lower the prior probability. When we switch frameworks, like switching between Yud’s and Bostrom’s, we aren’t just changing weights or adding noise, we are swapping which entire clusters of models (high-risk versus low-risk) get taxed by this complexity penalty.
Perhaps under Yudkowsky’s Machine
In AIT the Invariance Theorem tells us that translating concepts between two Turing machines incurs a fixed overhead cost, which here I will denote as
So let
where
The interesting thing about this inequality is that as the translation advantage
I should briefly pause again here: I love postmodern fiction, but I am not a postmodernist or a nihilist. I don’t mean to imply that math or data are fake or anything like that (which you could accidentally conclude if you think the implication is “you’re just saying assign random P(doom) based on your feelings and nothing matters.) I’m definitely not saying this, and the divergence is mathematically capped by the translation cost between frameworks. By the Invariance Theorem, the parameter
The point isn’t to say that P(doom) can be anything, it is to say that it is underdetermined within a specific range, proportional to how different the competing scientific ontologies really are.
Differential Screening: The Random Walk of Doom
Let’s give room to the skeptic again: “Ok, I see the points above, but eventually, as clear and unambiguous evidence accumulates, the data force convergence, right?”
Unfortunately not. This assumes a monotone accumulation of evidence, that is, the assumption that every new datapoint pushes the probability in the exact same direction. But this is not correct, when frameworks carve up the hypothesis space using different causal joints, the same piece of evidence can cause them to update in opposite directions. I call this divergence of uniform agreement on what an observation means Differential Screening.
Returning to Yud and Bostrom (again these are just hypothetical examples, I am not saying I speak for them and I don’t know all of their positions) say an AI company deploys a new model that happens to find a ton of zero-day exploits.
Under Machine
(Yud’s alignment framework): this event is negative. It is understood as representing a causal model where the system’s capabilities are demonstrably outpacing our ability to control them. The Bayesian update is gap-widening and risk goes up.Under Machine
(Bostrom’s surgery-framing framework): views this exact same event differently, say as a causal model where this is the warning shot that wakes up government regulation and boosts safety ahead of AGI. The update is gap widening in the opposite direction of Yud and risk goes down.
The log-odds gap between these two frameworks over time does not steadily shrink because a substantial fraction of new observations are gap-widening in opposite directions. The cumulative differential accumulates mixed-sign increments rather than declining monotonically. The trajectory of our collective belief doesn’t end up as a converging asymptote, instead it acts like a bounded-increment random walk.[6]
As a result, the expected time to canonical agreement scales non-linearly with the fraction of disputed evidence. It pushes the convergence timescale far, far beyond any policy-relevant horizon.
The Halting Problem of Permanent Loss
In the sections above we showed that even if we perfectly agreed on what “doom” meant, the evidence could not pin down the number. However; there is an even more pernicious and problematic formulation of catastrophic event prediction: ones involving expansive formulations that end up with ill-defined definitions of the event itself.
Think about a common definition of existential risk as something that leads to the “permanent loss of humanity’s potential.” If we want to calculate a real probability for this, we have to be precise about expressing it as a coherent, evaluatable mathematical object in code. We need to formulate a computable classifier because for computationally universal generative models (complex agent-based simulations or LLM rollouts over millions of tokens) the measure
A computational agent can only estimate the probability of doom by essentially running Monte Carlo rollouts of finite-time trajectories and scoring them. To accurately score a rollout, the mushy philosophical concept of “permanent loss” must be operationalized into a computable, halting indicator function
Even a set-theoretic Bayesian who doesn’t care about running code or doing Monte Carlo simulations, and just wants to define their event
Defining this event feels sort of structurally cursed, and the reason why is that it comes down to the infinite nature of the environment we are trying to predict. Recent papers have suggested that the universe, or the post-AGI environment, is computationally universal[7] so “permanent loss of potential” is not a localized event, but an asymptotic property of trajectories. Formally, it has the logical structure of a
The mathematical classification of unsolvable problems is known as the arithmetical hierarchy and the number of alternating quantifiers determines just how deeply uncomputable a problem is. A standard Halting Problem only has one: does there exist a step where this program stops? Sometimes this can be solved by just running the code and waiting. But the notion of permanent loss stacks two quantifiers: it requires that there exists a threshold followed by a forever state. Because you can never confirm a forever state just by watching a finite simulation (because humanity could theoretically recover at step
This matters because any practical probability estimate requires a total computable classifier
The problem is that membership in a nontrivial tail set cannot be determined by a finite prefix. Therefore, any computable classifier must systematically misclassify some trajectories.
To solve this you might think you could write a more complex and better-describe classifier to reduce the error rate, but it is possible to strictly prove that reducing this classification error requires algorithmically incompressible bits of information. You run into a Precision-Robustness Tradeoff. Working in the Solomonoff mixture over Cantor space, the product of your classifier’s error
This equation establishes a strict mathematical balancing act: to drive the error down, your code’s complexity must go up. But the trap is much worse than just having to write a longer program, correctly classifying the low-complexity boundary cases requires deciding the halting problem. As you may expect this is worrisome because no general algorithm can solve the Halting Problem. Your classifier literally cannot compute the answers to these boundary cases, it must have the answers hardcoded into it by you, the programmer. To reduce your error, your classifier must physically encode initial segments of Chaitin’s constant relative to the halting oracle (
Slightly imprecisely: Chaitin’s constant is the solution to the Halting Problem expressed in binary digits. However; the sequence contains answers to uncomputable problems so it has a property called “2 randomness” which makes the sequence indistinguishable from computationally irreducible static. So since these bits are 2-random they cannot be compressed and for every factor-of-two reduction in predicate error, it is required to have one additional bit of perfectly random, incompressible specification.
Now if you wanted those bits, where could they come from?
Computation can’t generate them because they are uncomputable.
Data can’t supply them because data only tells you which finite prefix you are on, not how to classify its asymptotic tail.
Cross-agent consensus won’t sync them because the Invariance Theorem doesn’t force two different frameworks to choose extensionally identical classifiers.
It turns out you are entirely leaning on unforced subjective taxonomy which traps you in an Impossibility Triangle. You must either
Accept Ambiguity: and use a simple predicate with low algorithmic complexity. This will cause you to mathematically misclassify a vast number of simple, computable futures. You may end up calculating the probability of the wrong event.
Accept Fragility: by using a highly complex event predicate to reduce misclassification. However; the bits required to define it are incompressible and must come from your specific ontology and your estimate becomes fiercely encoding-dependent.
Accept Both: and pick a middle ground. You suffer from moderate ambiguity and moderate fragility at the same time.
The overall issue is this: if you reduce the predicate ambiguity, as we just discussed, you are forced deeper into framework-dependence, with all of the issues we previously highlighted. We proved the data cannot correct your priors and now we proved you must inject uncomputable bias just to define an event. Together these guarantee that any precise percentage of these types of risks are a structural illusion. You are no longer measuring the “probability” of the event, you are only measuring the algorithmic cost of your own worldview.
Why This is Different
If you came from the introduction: welcome! And if you just read through the essay above, you might be tempted to map this argument onto existing debates. Especially because one type of unprecedented risk includes P(doom), and I used it as an example, it is critical for me to explain why the formalism above puts us into a different bucket.
Usually critiques of P(doom) fall into one of three buckets:
A Psychological Critique: something like “humans are bad at forecasting, poorly calibrated, and overly influenced by tribal vibes or sci-fi tropes.”
A Complexity Critique: someone says that the world is too messy, the causal graphs are too dense, and our current models are simply not gears-level enough to produce a reliable number.
An Epistemological Critique: usually a standard frequentist or Popperian objection that genuinely unprecedented events cannot be assigned probabilities because they lack reference classes, making the exercise philosophically moot.
This post is none of those because I am assuming the perspective of an ideal Bayesian agent. In doing this I am granting the agent infinite compute, perfect logical omniscience, and flawless Solomonoff induction. Even after doing all of this, the math shows that canonical point-estimates for unprecedented catastrophes are structurally blocked.
Specifically, here is what separates this essay from the usual discourse:
It is not just about “Priors matter”, it is about Evidential Screening. The standard Bayesian defense is priors are subjective for small samples, but enough data washes them out. However; the Evidential Screening Property proves that for unprecedented risks, the data structurally cannot do this because all available non-catastrophic data perfectly fits both the “safe”-ontology and the “doom”-ontology. This makes the catastrophe parameter mathematically screened off. We could observe a thousand years of safe AI prefix data and the priors would never wash out.
It proves that more data won’t save us. The intuitive hope is that as AI capabilities scale we will approach a sort of asymptote of consensus. But Differential Screening proves that because different ontologies have misaligned causal joints, they interpret the exact same capabilities jump as evidence in opposite directions. This demonstrates that we don’t tend to a narrowing asymptote, but rather it leads to a bounded-increment random walk and more data can easily result in permanent polarization.
It proves that better definitions demand uncomputable magic. When faced with the imprecise nuance of expanded definitions of P(doom), a rationalist impulse is to operationalize the definition. The Precision-Robustness Tradeoff proves that this instinct is a trap. For asymptotic tail events like “permanent loss of potential”, driving down classification error mathematically requires incompressible, 2-random bits of Chaitin’s constant. This is an insurmountable problem: you cannot compute them and the data cannot supply them. The ambiguity isn’t a linguistic failure, but instead a strict boundary in the arithmetical hierarchy.
It is not epistemic nihilism. As I mentioned above, I am not throwing my hands up and saying, “We can’t know anything, so assign whatever number feels right.” The math strictly bounds the divergence by the translation cost between frameworks (
). I am tearing down the illusion of single point-estimates specifically to replace them with mathematically rigorous alternatives.
Most critiques of P(doom) attack the agent making the estimate. This framework attacks the information geometry of the problem itself. And it is a problem for P(utopia) as well!
How to Restore Canonical Risk
The probabilities of unprecedented catastrophes are non-canonical because their precise numerical value is a syntax error generated by compilation costs, which is trapped by evidential screening, only partially identified by data, and then choked by uncomputable predicates. So what are we supposed to do? Wait for the end while wringing our hands? No, but we have to be clearer about what survives the mathematics:
Find the Holy Grail: Agreed Precursors With Aligned Causal Joints
You can break the Evidential Screening Property if rival frameworks agree ex ante on an observable intermediate causal node. In other words, if Yudkowsky and Bostrom can agree that “doom is strictly preceded by an AI system autonomously acquiring $100M of illicit computing resources” or public health experts agree that “a novel pandemic is strictly preceded by sustained human-to-human transmission”, then we have an observable precursor.
There is a catch: it is not enough to simply agree that the precursor is necessary. Both frameworks must explicitly agree ex ante on the directional update that observation triggers, that is, precursors where the causal joints align. This is not a new concept , but it is my top and best guess at a constructive recommendation. The screening property is what makes P(doom) non-canonical; agreed precursors with aligned derivatives are what break the screening property. When they do align, non-event data discriminates between models that predict different precursor rates, likelihood ratios can grow, prior gaps get washed out, and estimates converge! Before we can solve AI alignment, we should solve epistemic alignment among ourselves.
This points to a clear shift: progress does not come from debating unobservable asymptotic mechanisms (“is it a surgery or an asteroid?”) or refining subjective point-estimates. It comes from doing the difficult work of building cross-framework consensus on the observable causal nodes, and their directional updates, that necessarily precede the catastrophe. As treaties and agreements are of high interest to AI safety groups, this seems like a tractable area to focus on, and one that does not require nationstate agreements. It only requires rival labs and pundits to sit down and agree on what a specific test result will actually mean before the test is run.
The allure of a single, objective probability estimate is essentially a desire to outsource our existential fear to the comfort of a single number. It is unfortunate that math refuses to cooperate for this purpose. It is the case that when dealing with unprecedented catastrophes your framework is your fate. Until we find agreed precursors with aligned derivatives, we aren’t doing Bayesian updating, we are just arguing about which programming language is prettier.
Appendix: All of The Mathematics
This section goes deep into the math, with minimal explanation, into the concepts above. I am working on a full paper that melds the narrative essay with the precise mathematics below, please let me know if you would like to see this. I don’t think it would have fared well on LW for clarity. There are, of course, numerous sources that support the bulk of the mathematics utilized because I am not proving any new computability theorems or doing anything very special. The small additions of my own (at least to my knowledge) are: the Mixture Probability Dispersion Theorem (Theorem 1) and its composite-cluster proof, the Precision-Robustness Tradeoff (Theorem 2) and the direct routing through
Formal Setting and Objects
Spaces and measures. The sample space is Cantor space
Description languages. Let
Mixtures. Solomonoff’s uncomputable prior assigns weight
Classifiers and events. A total computable classifier
Ground 1: Canonicality Failure via Evidential Screening
Proposition 1 (Encoding Swing). For any models
Proof. The posterior odds decompose as
Definition 1 (Evidential Screening & Partial Identification). An estimation problem satisfies the evidential screening property with respect to a target event
for all achievable
The prior-free identification region for the parameter of interest
Denote this region
Theorem 1 (Mixture Probability Dispersion). Let
Proof. The total mixture probability under
By the odds decomposition,
Subtracting the symmetric expansion for
Corollary (Translation Barrier Cap). By the Invariance Theorem,
Definition 2 & Proposition (Differential Screening & Gap Dynamics). The log-odds gap is
(a) If
(b) If
Ground 2: The Event Predicate
Identity 1 (Additive Decomposition). The total cross-framework discrepancy structurally decomposes into Predicate Divergence and Model Divergence:
Assumption 1. Target catastrophe
Hypothesis (Computable Witnesses).
Lemma 1. For any computable sequence
Proof. A prefix-free program computing the Dirac measure
Theorem 2 (Precision-Robustness Tradeoff). Let disagreement region be
(a) Topological Lower Bound: For any total computable classifier
(b) Incompressibility: Specifying a classifier that correctly categorizes a sequence family up to complexity
Proof of (a).
Proof of (b). Let
By Assumption 1,
Suppose a classifier
Conjecture 1 (Extensional Divergence). For any
Why I want it, why I can’t get it. The objective function
I can not figure this out because correctly classifying
- ^
I am not talking about “non-canonical” probability in a Boltzmann sense. Only as defined here.
- ^
A quick disclaimer: in this essay I formalize a number of concepts into mathematical terms. I want to be clear that I am not claiming your brain is literally running Solomonoff induction over Cantor space. But it is very useful to establish formal upper bounds on inference using algorithmic information theory. The argument is this: by showing the structural limits of the problem for a mathematically perfect infinite-compute superintelligence that cannot wash out its priors with non-catastrophic data, our limited human heuristics wouldn’t be able to do it either. This is not an argument about the literal mechanics of your cognition.
- ^
To be clear: Bostrom and Yudkowsky aren’t running competing simulations of AI trajectories based on these works, but their frameworks (the QALY-maximizing surgery ontology versus the alignment-failure asteroid ontology) are good examples of the kind of deep ontological divergence that when formalized into actual generative models produces the encoding dependence I talk about in this essay.
- ^
See Charles Manski’s work.
- ^
Interestingly, Bostrom’s own analysis doesn’t make the mistake I mention here. He takes P(doom) as an input parameter and optimizes across the full range from 1% to 99%, rather than claiming to know the exact number. The divergence between his framework and Yudkowsky’s is not primarily about what P(doom) is. The difference is about what the right decision-theoretic framework is and for reasoning about it and this is actually an encoding dependence itself. It is just one operating at a meta-level: the choice of whether to frame the problem as “risky surgery” or “incoming asteroid” then determines which decision-theoretic apparatus seems natural, which then determines what actions seem rational across the identification region.
- ^
To get really rigorous: in the limiting case where computable priors natively exclude mechanisms outside their ontological primitives, they assign literally zero probability to each other’s core catastrophic mechanisms. Nielsen & Stewart proved that rational agents whose measures fail mutual absolute continuity don’t merely practically fail to converge, they can permanently, rationally polarize on the exact same stream of infinite evidence.
- ^
Memory-augmented LLMs are likely Turing-complete.
Is it your position that, if an Earth-sized planet had never been hit by an meteor before, but now the smart alien inhabitants see in their telescopes that a Moon-sized meteor is heading at high speed straight for the planet, they should say P(extinction) is unknowable?
If so, that’s kinda crazy right? They have (let us assume) excellent knowledge of physics, geology, and so on. They can apply that scientific knowledge to figure out what happens when giant high-speed projectiles hit planets, even if they haven’t seen that happen before. Yes it’s hard, and yes they might get it wrong in practice, but you seem to be making a stronger statement that it’s unknowable in principle, right?
Back to the AI case, I would say strong evidence is common, and we can just look around us and learn an enormous amount about algorithms, about geopolitics, about economics, about everything else. There is way more than enough information in the everyday world today there for an ideal reasoner to wash away all of their priors with any bearing on the question of AI doom. AI doom is not an isolated question. I mean, just go actually look at the people arguing for and against AI doom. What do they say? You’ll notice that they are making arguments based on underlying models of how the world works (algorithms, economics, etc.), and then the other person questions those models based on other aspects of the world, and so on. I think it’s highly implausible to say that a theoretical ideal reasoner with infinite patience and care would never be able to get to the bottom of this discourse, based only on presently-available information. I mean, as an existence proof, a theoretical ideal reasoner with infinite patience and care would be able to do an atom-by-atom simulation “in their head” of a million different civilizations building superintelligence, right?
Hm, no, I wouldn’t say P(extinction) is unknowable in that case, and I explicitly agree with this notion in the essay. It’s the first example I use when discussing canonical probability. The smart aliens might be facing an unprecedented event, but not an unidentified generative mechanism. Meteor mechanisms would be observed, whether by smart aliens or humans by physics and gravity observations from other things. So when they get new data, the data is cleanly discriminated between models and the impact parameter is point identified.
AI doom/utopia is not like that, it is structurally different. Disagreements aren’t about parameter values of agreed upon mechanism, it’s a battle of different causal ontologies. The generative mechanisms are the thing in dispute, which leads to your second point.
You say “There is way more than enough information in the everyday world today there for an ideal reasoner to wash away all of their priors with any bearing on the question of AI doom.” But Bayesianism requires more than just lots of data, it requires discriminating data. If hypothesis A (doom via deceptive alignment) and hypothesis B (safety via institutional equilibrium) both perfectly predict the exact same present-day prefix of algorithms and geopolitics, then the likelihood ratio between them is exactly 1. While frameworks predict the prefix with different accuracy on some dimensions, successful mechanistic models are modular. Once both composite models predict the observed prefix identically, the likelihood ratio on the catastrophic parameter will be effectively 1 and Bayes’ theorem says your posterior is identical to the prior.
I agree that people debate about every day data to argue for their models, but in the essay this is exactly what I formalize as Differential Screening. When new data causes frameworks to update in different directions, they have misaligned causal joints and everyday data does not force an asymptote of convergence. That is the point I was making in showing that it forces a bounded-increment random walk. To be clear: more data doesn’t automatically wash away priors when the directional derivative of that data is exactly what is under dispute.
Finally, the notion of an atom-by-atom simulator conflates infinite compute with infinite discriminatory data. The whole scaffolding of the essay already assumes an ideal reasoner with infinite compute.
An ideal Bayesian agent (like Solomonoff induction) doesn’t know the true causal transition matrix of the universe. Instead, it weights a mixture of all possible computable universes that perfectly output the data prefix observed so far. Infinite compute won’t let you magically manufacture missing discriminating data, it will allow the reasoner to calculate its prior with infinite precision.
Even if we grant a miracle: say your reasoner is literally Laplace’s Demon. It perfectly simulates the exact position of all of the atoms at some year in the future. Now what? Is that arrangement of atoms “doom”?
Again, my point is that AI risk and other unprecedented catastrophes like that are trapped by the information geometry problem itself. The best solution I can think of is to establish agreed precursors with aligned causal joints. Otherwise, the ideal reasoner will eventually realize the event predicate requires solving the Halting problem and output what the essay predicts: a syntax error.