Why I’m building Noetica

TL;DR: I’m exploring whether a single protocol that combines calibration practice, debiasing, and metacognitive monitoring can outperform those components in isolation. This post is a research proposal: I’m looking for critique on the architecture, novelty claim, measurement strategy, and what a sensible pilot could actually show.

Human beings have developed powerful sciences for understanding and manipulating the external world.

By comparison, we seem to have much weaker public methods for improving the quality of everyday reasoning under uncertainty, disagreement, and motivational distortion.

This post is not a claim that I have founded a new field, and it is not a claim that disagreement can be engineered away. It is a proposal: perhaps there is room for a more structured, measurable training protocol than the partial tools we currently have.

I’m calling that tentative proposal Noetica.

The hypothesis is modest:

Integrating calibration practice, debiasing, and metacognitive monitoring into a single protocol might produce benefits beyond those components used separately.

If not, the framework should be treated as redundant.


1. The problem this tries to solve

The tools for thinking better already exist in partial form.

Kahneman and Tversky mapped systematic biases.
Tetlock and the Good Judgment Project showed that calibration can improve.
Gigerenzer demonstrated that fast heuristics can outperform deliberation in some high-validity environments.
CFAR developed hands-on rationality training.
The Scout Mindset helped articulate the motivational side of epistemic posture.
Mindfulness traditions train attention to internal distortion in real time.

What I do not know of is a unified, accessible, operational framework that answers all four of these questions at once:

How do I detect, in real time, that my thinking is being distorted?
Which reasoning tool should I use in this specific context?
How do I measure whether I am actually improving?
How do I iterate based on outcomes rather than introspective feelings?

I may well be missing prior work; if so, I want to know about it. One reason for posting this is precisely to find out whether something close already exists.

Still, my current impression is that existing approaches each handle some of these questions well, but rarely all four as a single structured and measurable process.

Noetica is an attempt to build that integration layer.


2. Why “Noetica”

The name is just a placeholder for a line of inquiry: the possibility that cognition can be trained not only descriptively, but operationally and measurably.

If the data eventually show that the framework is redundant or ineffective, the name should disappear with the failed hypothesis.


3. Honest positioning: when Noetica is trivial

Before arguing for value, I want to be explicit about when Noetica adds nothing.

Noetica is trivial if, in a controlled comparison, it produces:

  • the same calibration improvement as forecasting training alone, without better transfer or compliance

  • the same debiasing results as checklist-based approaches, without added speed, measurability, or real-time applicability

  • the same outcomes as “read Superforecasting and keep a journal,” without the structured protocol making a meaningful difference

If the data show this, the framework is redundant.

It should still be published transparently, declared non-novel or ineffective in its current form, and redesigned.

This is not false modesty. It is the falsification criterion the project is aiming toward.


4. Where existing approaches stop

Here is my current reading of the landscape: where each approach provides real value, and where a gap may still remain.

Cognitive psychology

Kahneman, Tversky, Stanovich, and others gave us the richest map of cognitive failure ever produced. This work is foundational. But it is primarily descriptive. It tells you which biases exist, under what conditions, and with what consequences. It does not by itself provide a real-time operational protocol for detecting and interrupting them as they occur.

Forecasting training

Tetlock and the Good Judgment Project provide one of the strongest examples we have of measurable epistemic improvement. Calibration training works. The limitation is that forecasting training focuses on probability estimation tasks and does not directly solve the broader problem of motivational contamination, identity threat, or context-dependent operator selection before the estimate is even formed.

Ecological rationality

Gigerenzer’s work is essential because it demonstrates that deliberation is not universally superior. In the right environments, trained heuristics can outperform slower, more elaborate processing. The limitation is that this research does not fully solve the operational problem of determining whether one’s current environment is actually high-validity. That classification problem may itself be one of the central sources of error.

LessWrong /​ the Sequences

The LessWrong community and the Sequences have generated many of the most useful conceptual tools I know in this area. My impression, though, is that these exist mostly as individual techniques, norms, and insights rather than as a single measurable training protocol with predefined validation criteria.

CFAR-style rationality training

This work is valuable because it operationalizes rationality rather than merely theorizing about it. The limitation, as far as I can tell, is accessibility, cost, and limited publicly available pre-registered longitudinal evidence with clearly quantified effect sizes.

Mindfulness and contemplative practice

These traditions can improve noticing, emotional clarity, and the capacity to catch cognitive tension in real time. That matters. But noticing distortion is not the same thing as knowing what cognitive tool to apply next.

The gap I see is not in the absence of tools. It is in the absence of a selection layer plus accountability loop.

What seems missing is a framework that says:

Given this kind of uncertainty, this time pressure, this social context, this information structure, and this risk profile, what kind of thinking should I do now and how will I know whether that choice was good?


5. Five claims about where value might exist

Noetica becomes non-trivial only if it demonstrates at least one measurable delta over existing approaches taken separately.

Delta 1 — Operational integration: the Contextual Selection Matrix

The central possible contribution is a meta-algorithm I call the Contextual Selection Matrix (CSM).

It takes five environmental variables as input:

Environment validity (Ve): does this environment provide regular, rapid, and interpretable feedback?
Available time (Td): seconds/​minutes versus hours/​days
Social context (Cs): solo reasoning versus adversarial disagreement
Data uncertainty (Id): are usable base rates available or not?
Irreversibility risk (Ri): can the error be corrected, or is it costly or catastrophic?

The intuition is simple: many reasoning failures are not failures of intelligence in the abstract. They are failures of tool selection under context.

A quick example:

Suppose I’m under time pressure, defending a position in a heated disagreement, with incomplete information and moderate reputational stakes. That is not the same cognitive situation as making a solo forecast with reliable base rates and low emotional involvement. Yet people often apply the same style of thinking to both.

The CSM is an attempt to make that choice explicit.

In simplified form:

  1. High validity + time-critical
    Use a calibrated ecological heuristic.

  2. Adversarial context
    Use steelmanning plus crux isolation before object-level analysis.

  3. Extreme uncertainty with no usable base rates
    Use Fermi decomposition.

  4. Base rates available, solo analysis
    Start with base-rate anchoring, then update Bayesianly.

  5. Tail risk present
    Require a pre-mortem before final commitment.

Fallback rule:

If environment validity cannot be reliably determined, assume it is low.

Humans are often overconfident about whether they are in a skill domain rather than a noise domain. Assuming validity where there is chaos may be one of the most common and costly epistemic mistakes.

This is meant to reconcile what can otherwise sound like competing recommendations:

Tetlock: calibrate carefully.
Gigerenzer: trust trained intuition in valid environments.

The proposed reconciliation is not universalism, but context-dependent operator selection.

Testable claim: participants trained on CSM-based selection may produce lower Brier scores on held-out prediction tasks than participants trained only in forecasting or only in debiasing checklists.


Delta 2 — Metacognitive monitoring: training the supervisory layer

Most interventions train better reasoning. Noetica attempts to train something slightly different as well:

the ability to observe one’s own reasoning process in real time and intervene when it is being corrupted, before the output solidifies.

Using Stanovich’s tripartite framing loosely:

System 1 generates fast responses.
System 2 can elaborate them algorithmically.
But without reflective supervision, System 2 can become a sophisticated rationalization engine for outputs generated elsewhere.

Noetica is partly an attempt to operationalize and train that supervisory function.

Testable claim: participants using the protocol may show increased frequency of real-time interferent detection, measured through blind-coded daily logs, compared to a control group receiving standard debiasing practice.


Delta 3 — Standardized measurability and protocol accountability

Most people who take a critical thinking course or read rationality books have no rigorous way to know whether they improved. They may feel more reflective or more sophisticated, but that is not enough.

In Noetica, if the relevant metrics do not move, the protocol failed.

The accountability sits with the framework, not with the user’s self-concept.

This is one of the strongest normative commitments in the project: no vague claims of “inner growth,” no epistemic self-congratulation based on introspective feeling alone.

Long-run testable claim: if the framework is worth taking seriously, later controlled studies should show non-trivial effects on at least some predefined outcome measures, rather than only participant self-report.


Delta 4 — Transfer across domains

One of the hardest problems in debiasing research is transfer. Many interventions help on the trained task and nowhere else.

If Noetica shows that improved calibration or operator selection transfers from training exercises to daily decisions with verifiable outcomes, that would be a serious contribution.

Testable claim: improvement on daily prediction logs and real-world decision logs may correlate with improvement on a standardized calibration test, suggesting transfer beyond task-specific practice.


Delta 5 — Compliance and accessibility

Even if theoretical efficacy were only comparable to higher-friction training environments, Noetica could still matter if it produced equivalent outcomes with lower cost, lower dropout, less time demand, and easier dissemination.

A protocol that fits into 15 minutes per day for 4 weeks is obviously less ambitious than immersive workshop formats, but it may be more scalable.

Testable claim: a minimal protocol may achieve reasonably high completion rates over 28 days.


6. Core architecture

Axioms

Epistemic Entropy
Without deliberate metacognitive effort, cognition tends to degrade toward self-serving simplification, rationalization, and energy conservation.

Bounded Rationality
No agent has infinite time, information, or computational capacity. Perfect rationality is impossible; strategic use of appropriate heuristics is necessary.

Error Quantifiability
The distance between belief and reality is, in principle, measurable. It is not merely a philosophical mood. This grounds the framework in probabilistic and empirical accountability.

Algorithmic Plurality
There is no single universally optimal way to think. Accuracy depends on selecting the right reasoning procedure for the context.

These are not meant as metaphysical truths. They are working assumptions that make the framework legible and testable.

Cognitive-Motivational Interferents (CMI): initial taxonomy

The framework currently tracks four main classes of interferents:

Type 1 — Identity Protection
Distorting evidence to defend ego, role, ideology, or tribal membership.

Type 2 — Energy Optimization
Replacing hard questions with easier substitutes in order to reduce cognitive effort.

Type 3 — Affective /​ Sunk Cost Distortion
Letting current emotional valence or irrecoverable prior investment bias current judgment.

Type 4 — Social Cascade
Adopting beliefs because of perceived consensus or network pressure rather than direct evaluation of evidence.

This taxonomy is provisional. It may be incomplete, clumsy, or partially overlapping. But it gives the protocol a way to operationalize what it is trying to detect.

Cognitive operators

Alpha — Bayesian Update
Replace binary belief with probability estimates and revise quantitatively as evidence arrives.

Beta — Fermi Decomposition
Break complex uncertain estimates into smaller independent components to reduce aggregate error.

Gamma — Base Rate Anchoring
Start from the outside view before case-specific interpretation.

Delta — Pre-Mortem
Assume failure occurred; work backward to identify plausible causes before commitment.

Epsilon — Steelmanning + Crux Isolation
Reconstruct the strongest version of the opposing view, then identify the key point on which the disagreement turns.

These are not claimed to be exhaustive or uniquely original. The novelty, if any, is in their integration, routing logic, and measurable deployment within a single protocol.

The 6-phase operational cycle

  1. Noticing
    Detect cognitive friction, emotional charge, confusion, urgency, defensiveness, or other signals of distortion.

  2. Contextualization
    Define the problem and estimate constraints: time, risk, data availability, social structure, environment validity.

  3. Strategy Selection
    Use the CSM to select the appropriate cognitive operator.

  4. Execution
    Apply the operator while maintaining active monitoring.

  5. Disconfirmation
    Search for evidence that would falsify the current model or preferred conclusion.

  6. Update & Log
    Revise probabilities, record what changed, and log the process for later analysis.

The cycle restarts whenever new evidence appears or a new interferent is detected.


7. Validation framework

This post should be read as a proposal for a pilot and a longer-run research program, not as a claim that efficacy has already been established or that the first pilot could by itself justify strong conclusions.

At the moment, I see two distinct validation stages.

Stage 1 — Pilot /​ feasibility study

The first goal is not to prove that Noetica works. It is to find out whether the protocol is usable, measurable, and worth testing more rigorously.

The main pilot questions are:

  • Can participants actually apply the Contextual Selection Matrix with reasonable consistency?

  • Is daily adherence high enough to make the protocol viable?

  • Are the proposed measures sensitive enough to detect signal rather than mostly noise?

  • Can the journal-based measures be coded with acceptable inter-rater reliability?

  • Do the data suggest enough promise to justify a preregistered controlled comparison?

At this stage, outcome measures are exploratory rather than decisive.

Stage 2 — Controlled efficacy testing

Only after a pilot would it make sense to run preregistered controlled studies designed to test whether the integrated protocol produces meaningful gains over existing components taken separately.

The current candidate outcome measures are:

  1. Predictive calibration
    Brier score decomposition, including reliability and resolution, plus Expected Calibration Error where appropriate.

  2. Belief-update quality
    Performance on structured update tasks, potentially compared against Bayesian benchmark solutions where that comparison is well-defined.

  3. Interferent reduction
    Frequency and intensity of post-hoc rationalization, identity defense, or similar distortions, measured through blind-coded journals or related proxy measures.

  4. Belief-system coherence
    Some measure of contradiction or inconsistency across linked propositions, though this metric is currently the least well operationalized and may need substantial revision or removal.

  5. Metacognitive resolution
    Confidence-accuracy correlation and degree of over/​under-confidence.

Long-run success criteria

The stronger version of the framework should be treated as successful only if later controlled studies show something like:

  • Cohen’s d ≥ 0.5 on at least 3 of 5 primary metrics

  • replication in at least 2 independent studies

  • adequate statistical power and transparent analysis plans

  • full pre-registration of hypotheses, outcomes, and analysis decisions

  • public transparency of data where ethically feasible

If those standards are not met, the stronger form of the framework fails.


8. Protocol v1 outline

Duration: 28 days
Time cost: approximately 15 minutes per day

This protocol is meant as a minimal pilotable version, not yet as a validated training regimen.

Daily practice

Morning
Formulate three falsifiable short-horizon predictions about the day or near-term events.
Assign explicit probabilities from 0–100%.
Resolve them within 24–48 hours and log the Brier score.

Evening
Identify at least one moment of cognitive friction from the day.
Log:

  • the trigger

  • the CMI type detected

  • the operator applied, if any

  • the belief before and after

  • the confidence delta

Weekly practice

Once per week, select one complex decision, disagreement, or entrenched belief.
Run the full CSM:

  • estimate the five environmental variables

  • select the operator

  • execute the analysis

  • log the update

Pre/​post measurement

A 50-item calibration test spanning mixed domains such as geography, biology, history, physics, and base rates, scored probabilistically rather than as simple right/​wrong recall.

At this stage, the purpose of the pre/​post measurement is partly exploratory: to estimate signal quality, participant variance, and whether the instrument is good enough for later controlled work.


9. Scope and boundaries

Noetica is strictly about the process of cognition.

It does not claim to address:

  • the substantive truth of specific beliefs

  • ethics or moral philosophy

  • goal selection

  • subjective well-being except indirectly

  • neurobiological mechanisms

  • consensus as an end in itself

It is not a complete philosophy of mind or life.

It is an attempt to improve the quality, traceability, and corrigibility of reasoning.

At most, it tries to improve the engine, not determine the destination.


10. Known limitations and open questions

I want to be explicit about what I do not know.

1. Does the CSM actually outperform unstructured deliberation?

The logic feels plausible to me, but the central question is empirical: does explicit context-sensitive operator selection improve outcomes, or merely create a more elaborate self-description?

2. Can users classify the context well enough for the matrix to help?

This may be the deepest failure mode. If people are bad at estimating environment validity, time structure, risk, or information quality, then a selection framework could give the illusion of rigor while preserving the original error.

3. Will transfer occur?

The debiasing literature is mixed here. This is one of the riskiest claims in the whole project.

4. Is 15 minutes per day enough?

The protocol is intentionally minimal for compliance reasons. But the dosage may simply be too low to matter.

5. Is the CMI taxonomy adequate?

Four classes may be too few, too broad, or incorrectly partitioned.

6. Are the calibration items well constructed?

Item balance, difficulty calibration, and cultural neutrality all need external review.

7. Will daily logs generate evidence or performance theater?

Self-monitoring can improve awareness, but it can also encourage stylized self-description. This needs methodological control.

These are not peripheral details. They are central pressure points.


11. What I’m looking for

I’m posting this because I want the framework exposed to criticism before I get too attached to it.

In particular, I’m looking for:

  • theoretical critique
    Are the assumptions defensible? Is the architecture coherent? Where does the logic break?

  • methodological critique
    Are the proposed measures appropriate? Which parts are underdefined, noisy, or likely to fail in practice?

  • attacks on the novelty claim
    If this is mostly recombination of existing methods with little real integration, I want to know.

  • prior art I may have missed
    If something close already exists and has been tested, that matters more than preserving the appearance of originality.

  • eventual pilot collaborators
    People willing to run Protocol v1 for 28 days and share anonymized data.

Noetica makes no promise of wisdom, happiness, or moral improvement. At most, it is a proposal for a disciplined, measurable attempt to improve reasoning under uncertainty and distortion.

If the idea survives criticism and later data, that would be interesting. If it does not, it should be treated as a failed or redundant proposal.


References

Kahneman, D. (2011). Thinking, Fast and Slow.
Tetlock, P. E., & Gardner, D. (2015). Superforecasting: The Art and Science of Prediction.
Stanovich, K. E. (2011). Rationality and the Reflective Mind.
Gigerenzer, G. (2007). Gut Feelings: The Intelligence of the Unconscious.
Galef, J. (2021). The Scout Mindset.
Tavris, C., & Aronson, E. (2007). Mistakes Were Made (But Not by Me).
Kahan, D. M. (2013). “Ideology, motivated reasoning, and cognitive reflection.” Judgment and Decision Making, 8(4), 407–424.