The Legitimacy Gate
Prediction markets are inputs, not the cognition layer.
A signal is not yet governance until some legitimate actor must respond to it. Prediction markets, AI audits, expert forecasts, and open-source diagnostics are inputs to a cognition layer. AI lowers the cost of producing diagnoses, not the cost of institutional ownership. The bottleneck is the procedural attachment by which a signal becomes decision-relevant, not the quality of the signal itself.
I. The legitimacy gate
You can perfectly compute the failure mode of a government policy, price it in a prediction market, and publish it for free. The policy will still happen.
A prediction market says the policy will fail. An AI audit finds the contradiction. A think tank publishes the mechanism. A spreadsheet proves the fiscal path is impossible. Nothing happens.
This is the standard failure mode for the modern repertoire of governance fixes. The signal is correct. The signal is public. The signal does not bind, does not steer, does not get answered, and within a year it is forgotten. The default diagnosis is either public choice (“politicians are corrupt”) or epistemic inadequacy (“voters are irrational”). Both are sometimes true. Neither names the missing procedural step.
The missing step is what I’ll call the legitimacy gate: the set of questions a cognition artifact must answer before it can attach to a public decision.
- Who authorized this model?
- Who can contest it?
- Which institution owns the response?
- What happens if it is ignored?
- Which legitimate decision-maker must take it into account?
A cognition artifact passes the legitimacy gate when a named actor must respond to it through a visible procedure: endorse the model, contest it, override it, delay the decision, trigger an audit, log it in a public register, or commit to a forecast-to-outcome review. Without that procedural consequence, the artifact remains advice.
Earth’s state machinery is legitimacy-first. Who may decide got centuries of institutional iteration. What will the decision do did not. The asymmetry is the spine of Legitimacy Came Before Cognition; this essay is its applied sibling. A signal can be true, useful, public, and still have nowhere legitimate to attach. When that happens, it becomes advice, journalism, activism, or a spreadsheet that someone politely fails to read.
II. Specimen: every actor was acting within role
The 2022 Statistics Finland reclassification of state-subsidized housing loans (developed in detail in Legitimacy Came Before Cognition) is the canonical case: every actor acted within role, the classification was procedurally legitimate, no institution owned the consequence model. Finland’s headline EDP debt ratio rose by about six percentage points overnight; the classified stock has since grown toward €20 billion; the country’s primary countercyclical housing instrument is now politically harder to use just as construction enters a deep recession.
Every actor was acting within role. That was the failure.
A prediction market could have helped only if someone had formulated the right question and attached it to the decision before the classification locked in. That formulation step — selecting which question becomes institutionally live, who owns the answer, and which decision-maker must take it into account — is the coupling layer. It is what no Finnish institution owned. The same shape recurs across procurement decisions, regulatory classifications, fiscal-rule interactions, and AI deployment thresholds: legitimate procedure, no consequence owner, the cost absorbed downstream.
III. Prediction markets are inputs, not the cognition layer
A prediction market answers priced questions: will X happen? It does not select which questions become institutionally live. It does not specify the mechanism a policy claims will produce X. It does not name the hidden variable that absorbs the cost. It does not assign ownership of the response. It does not impose a duty to act when the price moves. A market is signal infrastructure.
A cognition layer, in the sense I mean, produces a different artifact. For a proposed decision, it outputs:
- The mechanism claim — the causal path that is supposed to produce the stated outcome.
- The carrying assumptions — what must be true for the mechanism to work.
- Forecasts — what we should expect to observe and when, including market and expert signals where available.
- An absorption audit — which unowned variables may carry the cost.
- An owner map — which institution owns which downstream consequence.
- A falsification trigger — what observation would show the mechanism failed.
- A review date — when the prediction is compared to outcome in public.
Prediction markets are useful inputs to (3). They do not by themselves produce (1), (2), (4), (5), (6), or (7). A cognition layer is the procedural tripwire that forces a specific decision-maker either to publicly endorse a falsifiable mechanism model — with its assumptions, its absorption audit, and its review date — or to have their decision delayed, contested, or marked as unmodeled in a place that subsequent voters can see. “Attach to a decision” means that level of teeth, not “publish a PDF appendix.”
A note on the relation to Constructive Diagnosis: that essay’s six-field repair specification is the architectural standard for designing the institution that holds the missing primitive. The seven outputs above are what the institution, once built, produces for a given decision on a given day. Specification and output are different artifacts.
The civic-tech version of this problem. The broader belief that public data, transparency dashboards, prediction markets, AI audits, and forecasting tournaments will collectively translate into better governance runs into the same gate.
Hanson’s futarchy — vote on values, bet on beliefs, where representatives define a welfare metric and markets estimate which policies raise it — is the cleanest existing proposal for making a market mechanism decision-coupled rather than merely advisory. It is more serious than “legalize prediction markets” because it tries to pass the legitimacy gate by constitutionalizing the market’s role inside a values metric set by legitimate principals. Its open problems are all legitimacy-gate problems: who defines the welfare measure, who controls the policy agenda, which forecasts are admissible, how manipulation is audited, how overrides work, why losing factions treat the result as legitimate.
The bounded oracles that did emerge on Earth (CBO, central banks, fiscal councils) accepted a bounded mandate from a legitimate principal rather than trying to make cognition the principal. Futarchy is the cleanest attempt to do otherwise; whether the legitimacy stack accepts that move is the test.
IV. AI eats the technical bottleneck, not the legitimacy bottleneck
AI and open-source tooling dramatically lower the cost of producing cognition artifacts. They do not solve the procedural bottleneck. Producing a true, computable diagnosis outside the state does not make it decision-relevant. AI does not decide which artifacts are legitimate, which must be answered, which decision-maker must change course, or which institution must own the response. It changes the political question from “can the state compute this?” to “which legitimate actors must answer when the computation is public?”
This is also where the framework expects to find the most leverage in the next decade. The diagnostic substrate built outside the state is exactly the right move — LawVM, the open-source corpus-graph tool that builds the in-force statute graph for Finland, Estonia, and the UK, is one such tool. From the Finnish graph, one can compute that 184 mandatory delegations remain unexercised: clauses where the legislature used binding language requiring the government to issue subordinate regulations, and the regulations were never issued. Some of these are cases where a legal entitlement appears to exist before the implementing apparatus has been built. The diagnostic is computable from public data. Whether it changes anything depends on whether some legitimate actor is obligated to respond.
Inadequacy is now a computable object. Repair is not.
V. Design constraints: Hayek, Scott, and the NEPA trap
Cognition layers can fail in two opposite directions. Hayek’s knowledge problem and Scott’s high-modernism critique name the first: the layer mistakes its model for the world, and centralized confidence destroys local information. The second is that the layer survives but becomes ceremonial.
The US National Environmental Policy Act (NEPA) and its environmental impact statement (EIS) regime is the cautionary case for the second. NEPA created a procedural requirement for federal agencies to review significant environmental consequences and inform the public before major decisions — recognizably a consequence-modeling requirement. In practice, final EISs commonly run to several hundred pages and take years. The Council on Environmental Quality’s 2013–2018 sample found a 661-page average and 447-page median final EIS, well above the original 150–300-page expectation; recent CEQ data give a median of around 2.2 years from notice of intent to final EIS.
The result is documents that function partly as litigation defense and compliance documentation rather than as agile model-update infrastructure. Power’s Audit Society generalizes the same failure mode: when verification becomes mandated, capture answers with rituals. Both Hayek/Scott and NEPA are objections from inside the framework. They are design constraints, not refutations.
A cognition layer that survives both must:
- publish assumptions, uncertainty, causal paths, excluded variables, and expected absorption points;
- import local knowledge without pretending local knowledge has been fully centralized;
- compare prediction to outcome in public after the decision;
- carry adversarial review and a right of reply;
- have the contestation route mandated, not just the production route — so that compliance-theatre fictions can be challenged by counter-models with formal standing;
- remain contestable rather than final.
Designing around the NEPA trap also requires hard procedural bounds rather than open-ended compliance: strict page and time limits, automated rejection of boilerplate language, mandatory machine-readable structure, and non-trivial adversarial-review funding so that opposing-model production is not orders of magnitude underresourced relative to the official model.
VI. Build the substrate first
The mistake is to start with “mandate this in legislation.” That asks the equilibrium to have already solved itself: asking the inadequate system to bind itself by passing a Mandatory Cognition Act is asking for the legitimacy gate to already be open. Two paths actually work — build signals that need no permission, and build procedures that ride existing legitimate vehicles.
The honest framing for the unofficial artifacts is that they are catalysts, not the cognition layer itself. A one-shot think-tank PDF gets ignored because it is a one-shot signal with no compounding mechanism and no cheap path for any rival institution to use. An automated, versioned, public diagnostic substrate is different in three specific ways:
- It compounds across years and electoral cycles, accumulating receipts faster than any individual decision-maker can outrun.
- It indexes durably against named decisions and named decision-makers, so the cost of a rival rediscovering the inadequacy collapses to a search query.
- It lowers the activation energy for legitimate actors with grievances — opposition staffers, courts, regulators, journalists, fiscal councils — to wield the diagnosis, the way opposition parties, courts, and journalists have long wielded independent think-tank and academic analysis against incumbents, but at automated cost, indexed durability, and structural specificity that previous artifacts could not match.
The substrate doesn’t bypass the legitimacy gate. It catalyzes the actors who already have keys to it.
This is also where substrate diverges from prediction markets. A prediction market produces a probability: “this policy will fail with X%.” That number has no procedural attachment, no specific contradiction, no citeable structure. A statute-graph diagnostic produces something different — “section A requires the government to issue regulation B; regulation B has not been issued; the legal entitlement under section A is therefore unimplementable in cases C and D” — which is an artifact opposition staffers can drop into a parliamentary motion or that courts can incorporate by reference. Probability and legal weapon are not the same kind of object. Markets price questions; substrates produce structured, citeable findings that already speak the legitimacy stack’s native language.
Outside-the-state moves (no permission needed).
- Open-source legal-state replay. LawVM-style tools that compute stale citations, orphaned decrees, unexercised mandatory delegations, and statute-graph drift per jurisdiction, published before each electoral cycle. The state’s legitimacy gate cannot block what is already public.
- Counter-model registries: public, versioned consequence models for major bills, produced unofficially by think tanks, civic groups, or AI-augmented analysts, attached to the bill’s name in a way that journalists, opposition staffers, and courts can find later.
- Forecast-to-outcome receipts: anyone can publish the receipts when an official forecast is wrong by more than its stated uncertainty band, with the receipts attaching durably to the decision-maker’s name in indexable public records.
Pre-existing-vehicle moves (require legitimate principals, not new agencies).
- Generalize CBO/OBR-style scoring beyond fiscal cost to mechanism claims, inside the existing fiscal-council mandate. The legitimacy substrate already exists.
- Attach mechanism briefs to ministerial impact assessments where existing law already requires impact assessment. Use the existing legal hook; expand the content.
- Use bounded oracles’ existing remit to publish consequence models for specific upstream inputs to their domain (e.g., a central bank’s macroprudential mandate may already justify publishing consequence models for housing-finance classifications that affect debt ratios and construction cycles).
These artifacts do not yet bind decisions. What they change is the next argument: the question shifts from “could anyone have known?” to “why was the known model ignored?” That is the lever by which inadequate equilibria start to move.
VII. A falsifiable prediction: AI governance bodies
The framework outputs a concrete prediction. Of the AI safety institutes, AI Act enforcement bodies, and national AI governance organs emerging this decade, survival will track statutory standing × intra-elite usefulness more strongly than technical evaluation quality. A body with thin legitimacy substrate but excellent technical work that nevertheless survives politically without acquiring an intra-elite principal would falsify the framework. So would the inverse: a body with weak technical work and strong intra-elite anchoring that gets defunded or marginalized despite its institutional patrons. If the bodies that survive are the ones with the strongest intra-elite leverage and statutory authority, technical quality being secondary, the framework is right.
VIII. Close
An adequate civilization is not one where every citizen personally outthinks the state. It is one where the state’s decisions are attached to public, contestable consequence models before damage occurs.
Prediction markets are useful. AI is useful. Expert forecasts are useful. Open-source diagnostic substrate is useful. None of these is the cognition layer. The cognition layer is the institutionally legitimate route by which signals become decision-relevant.
The signal problem is being solved. The coupling problem is the work.
Sources and Notes
The legitimacy gate as primitive. The gate is the procedural attachment a cognition artifact must have to bind, contest, or steer a public decision. The deep historical argument for why the modern state has a mature legitimacy stack and a fragmented cognition stack is in Legitimacy Came Before Cognition; this essay is its applied sibling.
Statistics Finland 2022 case. Yle investigation, May 2026, on the 2022 EDP reclassification of state-subsidized housing loans (yle.fi). Corroborated by Statistics Finland’s 2022 change note (the method change concerned EDP debt reporting and explicitly did not affect general-government net lending or deficit). The BLS commissioner case is from AP/Reuters reporting following the August 2025 jobs report.
Futarchy. Robin Hanson, “Futarchy: Vote Values, But Bet Beliefs” (mason.gmu.edu/~rhanson/futarchy.html). The framing here treats futarchy as the cleanest existing attempt to constitutionalize a market mechanism inside a values metric set by legitimate principals; the legitimacy-gate objections are not an attack on Hanson but a description of why the move stays a thought experiment in current polities.
NEPA. EIS length and timeline data are from the Council on Environmental Quality’s Length of Environmental Impact Statements (2013–2018) and EIS Timeline Report (2010–2024). The “ritual of verification” framing generalizes Michael Power, The Audit Society: Rituals of Verification, Oxford University Press, 1997.
Hayek and Scott as design constraints. F. A. Hayek, “The Use of Knowledge in Society,” American Economic Review 35:4 (1945); James C. Scott, Seeing Like a State: How Certain Schemes to Improve the Human Condition Have Failed, Yale University Press, 1998. Both are treated as design constraints on mechanism cognition, not refutations of it.
CBO. Congressional Budget Office, History (cbo.gov/about/history). The 1974 Congressional Budget and Impoundment Control Act created CBO to provide objective, impartial budget and economic information for Congress.
LawVM. Open-source corpus-graph tool that builds the in-force statute graph for Finland, Estonia, and the UK. The Finnish census numbers (184 mandatory unexercised delegations among them) are from a May 2026 census artifact and should be treated as order-of-magnitude rather than audited statistics; a public methodology note is forthcoming.
Bridges to the rest of the corpus. Legitimacy Came Before Cognition develops the historical asymmetry between legitimacy and cognition stacks. Bad Equilibria Are Not One Thing classifies the failure modes the cognition layer would diagnose. Constructive Diagnosis develops the methodological standard (the six-field repair specification) that this essay’s seven-output cognition layer instantiates. The Fourth Branch develops the institutional seat of the diagnostic operator. The Finnish institutional proposal — the Mechanism Authority (Mekanismivirasto) — is documented at mekanismirealismi.fi/mekanismivirasto.
Related:
- Legitimacy Came Before Cognition — the historical asymmetry between legitimacy and cognition stacks; this essay’s parent
- Constructive Diagnosis — the methodological standard whose six-field repair spec this essay’s seven-output cognition layer instantiates
- Powerless Intelligence — the AI-era downstream variant: evaluators exist, response duty does not
- Bad Equilibria Are Not One Thing — the failure-mode taxonomy the cognition layer would diagnose
- The Fourth Branch — the institutional seat of the diagnostic operator
- Mechanism Realism — the foundational ontology the corpus runs on