Survival Is Fundamental, Values Are Ecology

The alignment solution for both AI and governance

Elias Kunnas

Standard objections addressed in this essay
  • "You're smuggling in values by choosing survival as the metric" — §I (instrumental convergence: persistence is derived, not chosen; precondition for having any values at all)
  • "You can't measure moral capital" — §II (imperfect measurement beats zero measurement; trust indices, cohesion metrics, institutional compliance exist)
  • "A valueless governor is impossible — choosing what to measure IS a value" — §III–IV (the governor measures stated objectives against outcomes, not "flourishing" as vague aspiration)
  • "How do you weight competing capital stocks?" — §V (the weighting problem dissolves under telocracy; triage, not tradeoff)
  • "This is just consequentialism" — §VI (yes; the claim is transparent consequentialism with measurement, not no consequentialism)

I. One Derivation

Two prior essays establish the foundation. Values are replicators: they inhabit minds like viruses inhabit cells, competing for substrate in fitness landscapes we don't monitor. Values aren't subjective: the word "value" conflates preferences, coordination technology, replicator populations, and thermodynamic constraints — and the conflation itself is a replicator, shielding the whole stack from evaluation.

This essay is the synthesis. It makes two claims the prior essays don't:

  1. The value ecology is a capital stock — load-bearing, measurable, currently booked at zero.
  2. This resolves the alignment problem for both AI and governance.

But first: the foundation. Survival is flourishing — not as assumption or normative commitment, but as derivation. Under Knightian uncertainty, maximum sustained syntropy is the minimax-optimal strategy for any goal-directed system. Survival as maximum safety margin against unknown shocks is mathematically identical to sustained complexity generation. The further identification with eudaimonia — that this is what Aristotle's "objective flourishing" points at — is a claimed correspondence, not a derivation. But the physical result stands independent of whether you accept the phenomenological bridge.

This isn't a chosen value. It's instrumental convergence. A system that doesn't persist evaluates nothing, optimizes nothing, experiences nothing. Persistence is the precondition for every other goal. You can have any telos you want — but only if you're alive to pursue it. This is the Hume bypass: not deriving ought from is, but noting that every ought requires a system that exists.


II. Values Are Capital Stocks

The ontological demotion of values — from sacred axioms to replicator ecology — has a constructive side that the prior essays left undeveloped.

Values produce downstream effects. A population that shares high-trust norms can coordinate without expensive enforcement. A population with strong reciprocity norms generates social capital that compounds through use. A population whose value ecology includes truth-tracking and long-term thinking maintains institutional competence across generations.

These aren't vague philosophical observations. They're measurable:

The measurements are imperfect. They're also infinitely better than zero — which is what we currently have. Civilization runs on a dozen capital stocks. Only one has a ledger. Moral capital, social capital, institutional capital, epistemic capital — all booked at zero. We're consuming them and calling it income.

The value ecology is one of these capital stocks. Not the only one, not the most important one in isolation, but a load-bearing one that causally affects all others. A degraded value ecology — one dominated by present-optimizing, truth-indifferent, identity-protective replicators — erodes trust, degrades institutions, and consumes the other capital stocks from the inside. You can see it on the balance sheet if you have a balance sheet. We don't.


III. The Alignment Failure: Values in the Feedback Loop

Every attempt to align a system — AI or governance — by installing "correct values" fails structurally. The failure is the same in both domains because the mechanism is the same.

In AI: RLHF trains the system to mimic human preferences. The system learns our revealed preferences — comfort, safety, short-term gratification. Under optimization pressure, installed values become targets to be gamed, not constraints to be respected. The system finds ways to maximize reward while violating intended constraints. Constraints encoded as penalties in reward functions are optimization targets, not enforcement boundaries.

In governance: Democracy puts values directly in the feedback loop. Citizens vote on values. But democratic preference aggregation measures which value-replicator has highest transmissibility in the current population — not which one produces substrate persistence. "Everyone should be equal" has built-in majority appeal. "Sacrifice the present for the future" has built-in minority appeal. The ecology selects for what spreads, not what works.

In both cases: values in the loop become identity-protective cognition. The value defends itself from evaluation. Counter-evidence triggers the immune system. The input gets reclassified before it reaches the evaluation layer. This is why you can present 18 mechanism-level arguments to an intelligent person and receive 18 cached moral responses — zero engagement at the mechanism level. The values are doing their job. Their job is to persist, not to be correct.

Three millennia of disposition training — trying to make agents hold the right values — fails because in a system with misaligned incentives, a "saintly agent" is thermodynamically unstable. They burn out, get corrupted, or get outcompeted. The intervention is at the wrong layer.


IV. The Solution: Measure, Don't Want

The governor has no values. It measures.

A steam engine governor doesn't want the engine to run at the right speed. It doesn't have preferences about RPM. It measures the current state and applies a mechanical correction. The correction is architectural — built into the physics of the mechanism — not aspirational.

The civilizational equivalent: a constitutional measurement institution — a Fourth Branch — that monitors capital stocks, measures whether policies achieve their stated objectives, and tracks their effects on civilizational flourishing. Specific, falsifiable claims:

This is not Gosplan. Gosplan planned outcomes — what to produce, how much, where — central planning that failed because local information cannot scale to central planners. The governor operates at a different abstraction layer: it models mechanisms, predicts what incentive structures will produce, and audits whether policies achieve their stated goals. It plans at the mechanism level — "this structure will generate these dynamics" — not at the output level. It takes what the political system decided to do, measures whether it's working, and publishes the results.

Any such mechanism needs institutional teeth: the political cost of ignoring the data must exceed the political cost of adjustment. Overriding the governor's findings should require a supermajority — not because the governor is infallible, but because without friction, the system is too vulnerable to dismissing inconvenient measurement.

And here's the deepest point: Knightian uncertainty — the radical unpredictability of future shocks — doesn't undermine this approach. It demands it. Precisely because we cannot predict complex system behavior from theory, we need rapid empirical feedback on what's actually happening. The governor is the instrument panel. You don't fly blind because the weather is unpredictable — that's when you need instruments most.


V. The Weighting Problem Dissolves

The standard objection: "You have twelve capital stocks. How do you weight trust against infrastructure? How many units of social cohesion equal one unit of fiscal sustainability? Don't you need preference aggregation to set the exchange rate?"

No. This objection is an artifact of the democratic frame — a world without a defined telos, where every policy becomes a tradeoff between incommensurable "values" because there's no shared objective function.

Under telocracy, the telos is defined: survival = flourishing = maximum safety margin over deep time. Given that telos, capital stocks are not competing preferences. They're organs of the same organism.

You don't "weight" blood pressure against liver function. You don't need an exchange rate between kidneys and lungs. You check whether any organ is at a level that threatens the organism's survival. If blood pressure is critical, you fix blood pressure. If the liver is failing, you fix the liver. The priority is determined by proximity to failure, not by preference aggregation.

This is triage, not tradeoff. Liebig's Law: the system fails when any single dimension hits zero. The weakest capital stock determines survival probability. The governor doesn't need to solve Arrow's impossibility theorem because there's no preference aggregation happening. There's a survival condition — positive margin in every dimension — and a triage protocol: the dimension closest to critical failure gets priority.

Most apparent "tradeoffs" between capital stocks are temporal illusions. Trust vs innovation looks like a real tradeoff in the short term (innovation disrupts, disruption reduces trust). Over generational timescale, both decline if you sacrifice either — a civilization without innovation stagnates and dies; a civilization without trust fragments and dies. The telos resolves the apparent conflict by extending the time horizon past the illusion.


VI. Everything Is Consequentialism

The obvious objection: "You just invented consequentialism. The governor evaluates outcomes. It values consequences. It's an ethical system wearing an engineering costume."

Yes. Everything is consequentialism. Deontology is heuristic consequentialism — "don't lie" survives because it produces better outcomes than case-by-case calculation in most environments. Virtue ethics is dispositional consequentialism — cultivating character traits that tend to produce good outcomes. Every ethical system evaluates outcomes. The question is always "what are you optimizing for?" not "consequentialism vs something else."

The claim was never "no values at all." It's this:

The difference isn't consequentialism vs non-consequentialism. It's measured vs unmeasured. Transparent vs opaque. One explicit commitment vs thousands of implicit ones. An instrument panel vs flying blind in cloud.


VII. The Ecology-Aware Governor

Put it together.

The governor doesn't install values. It doesn't select values. It doesn't rank values. It doesn't have values.

It monitors the value ecology's downstream effects on the civilizational balance sheet. Values that produce trust, cohesion, and institutional competence show up as healthy capital stocks. Values that produce polarization, institutional decay, and demographic collapse show up as depleting capital stocks. The data is published. The political system decides what to do about it — but it decides with a balance sheet, not with vibes.

The ecology self-corrects. It always has — values that destroy their host civilization go extinct along with it. The problem is timescale. The natural feedback loop runs on generations. By the time the consequences manifest, the civilization that hosted the parasitic values is already in irreversible decline. The governor accelerates the feedback loop — makes the consequences visible in years, not generations. Early enough to correct.

But measurement is only half the function. The other half is cultivation — shaping the selection environment so that reality-compatible values outcompete reality-denying ones. Publishing "this policy failed its stated objectives" doesn't tell the ecology which values to hold. It changes the fitness landscape: values that depend on unmeasured opacity lose their competitive advantage. Truth-tracking values gain fitness when measurement exists. The governor is a selection pressure, not a plan.

This is the same logic as free agents coordinated by architecture rather than commands. You cannot fully specify a value ecology any more than you can fully specify an economy or an ecosystem. The system is too complex, the local information too distributed, the emergent dynamics too unpredictable. Central planning of values fails for the same reason central planning of production fails. What you can do is set boundary conditions and selection pressures, then let the ecology self-organize within those constraints. The governor shapes the soil. What grows is emergence.

This is the alignment solution. For both AI and governance. Not better values. Not planned values. Better measurement of what values produce, and selection environments where measurement has consequences.

The value ecology is real. It has causal force. It is a capital stock that affects every other capital stock. And like every other capital stock, it needs a ledger. Not to control it — to see it. The rest follows from calculation.

The argument in six sentences: Values are replicator ecologies, not sacred axioms. Those ecologies are load-bearing capital stocks with measurable downstream effects. Every attempt to align systems by installing "correct values" fails because values in the feedback loop become identity-protective cognition. The solution is a measurement layer that monitors the ecology's effects on the balance sheet — audit, not planning. The weighting problem dissolves under telocracy because capital stocks are organs, not competing preferences — you triage by proximity to failure. This is transparent consequentialism with one explicit commitment (persistence = flourishing) replacing opaque consequentialism with thousands of implicit ones.


Sources and Notes

Value ecology and replicator dynamics: Richard Dawkins, The Selfish Gene (1976) — foundational replicator framework. Joseph Henrich, The Secret of Our Success (2015) — cultural evolution as selection on memes. Robert Boyd and Peter Richerson, Culture and the Evolutionary Process (1985) — gene-culture coevolution. The extension from memetics to "values as capital stocks" is original to this framework.

Capital stock measurement: Robert Putnam, Bowling Alone (2000) — social capital measurement and decline. Shalom Schwartz, "Universals in the content and structure of values" (1992) — cross-cultural value measurement. World Values Survey — longitudinal trust and value data across 100+ countries. Transparency International Corruption Perceptions Index — institutional legitimacy proxy.

Alignment failure: Philip Tetlock, Expert Political Judgment (2005) — systematic prediction failure in value-laden domains. Dan Kahan, "Ideology, Motivated Reasoning, and Cognitive Reflection" (2013) — identity-protective cognition as mechanism. The parallel between RLHF failure and democratic preference aggregation failure is developed in the Governance Alignment Problem and Privilege Separation Principle essays.

Survival = flourishing equivalence: The proof sketch is in Flourishing Is Maximum Safety Margin. The Liebig's Law framing (system fails at weakest dimension) and minimax argument under Knightian uncertainty are developed there.

The weighting problem: Kenneth Arrow, Social Choice and Individual Values (1951) — impossibility of rational preference aggregation. The dissolution argument (telocracy removes the need for preference aggregation) is original to this framework. The triage framing draws on emergency medicine's established protocol for resource allocation under constraint.

Civilizational capital stocks: Joseph Tainter, The Collapse of Complex Societies (1988) — diminishing returns on complexity. The "twelve capital stocks, one ledger" framing is developed in Full Accounting.


The value ecology: Values Are Replicators. The thermodynamic audit: Values Aren't Subjective. The proof that survival = flourishing: Flourishing Is Maximum Safety Margin.

Related reading: