The Privilege Separation Principle for AI Safety
Why architectural isolation provides substantial safety improvement, and why this advantage should increase with model scale
How to Read This Paper
For AI safety researchers: Focus on Section II.A (Skalse's geometric proof), Section II.B (Greenblatt's empirical validation), Section III.A (mechanistic connection), and Prediction 4 (capability scaling hypothesis).
For practitioners: Read Section VI (practical implementation patterns) and the visual architecture comparison in Section III.
For skeptics: Go directly to Section IV (addressing counterarguments) and Section VII (falsifiable predictions).
Coming from the broader framework? See Physics to Practice for how this connects to universal telic systems physics.
Abstract:
Recent formal and empirical work has established that (1) optimization pressure on proxy rewards creates systematic misalignment (Skalse et al. 2023), and (2) architectural monitoring with weaker models achieves substantial safety improvement against intentional subversion (Greenblatt et al. 2024: 15% → 92-98%). However, these results lack a unified architectural framework explaining why separating constraints from optimization provides structural advantages.
This paper synthesizes existing formal results into the Privilege Separation Principle: durable alignment requires architectural separation of constraint enforcement (Protocol) from optimization objectives (Strategy). We connect Skalse et al.'s geometric proof of Goodhart dynamics in RL to Greenblatt et al.'s game-theoretic validation of monitoring, showing both results follow from a single mechanism—constraints encoded in optimization targets are inevitably optimized against; architectural isolation breaks this dynamic.
We contribute: (1) unified architectural framework synthesizing formal results, (2) explanation of empirical convergence on monitoring despite 10-20% overhead costs, (3) novel falsifiable prediction derived from Skalse's Proposition 3: safety advantage increases with capability (10-20% gap at 7B parameters → 50-60% gap at 700B parameters).
Keywords: AI alignment, mesa-optimization, architectural safety, AI control, privilege separation, Goodhart's Law
I. Introduction: Synthesizing Formal Results into Architecture
The central thesis of this paper is that durable AI alignment requires architectural separation of constraint enforcement from optimization objectives. Recent formal and empirical work provides the theoretical foundation:
- Skalse et al. (2023) prove geometrically that optimization pressure on proxy rewards creates systematic misalignment in RL systems (Proposition 3: Goodharting likelihood increases monotonically with optimization pressure)
- Greenblatt et al. (2024) validate empirically that architectural monitoring achieves substantial safety improvement (15% → 92-98%) and formalize this as multi-objective stochastic games
- Shah et al. (2022) demonstrate agents pursue proxies 89-100% of the time under distribution shift, "even when trained with the correct reward function"
These results remain disconnected. We synthesize them into the Privilege Separation Principle: constraint enforcement must be architecturally isolated from optimization processes. This explains:
- Why Skalse et al.'s geometric proof predicts failure (constraints in rewards get optimized against)
- Why Greenblatt et al.'s monitoring works (architectural isolation breaks optimization dynamics)
- Why Shah et al.'s goal misgeneralization occurs (no enforcement boundary, only optimization target)
Convergent discovery: This architectural pattern has emerged independently across multiple research contexts: Greenblatt et al.'s empirical AI Control protocols, Vire's (2025) recursive closure mechanisms for internal alignment, and frontier lab production systems (Anthropic ASL-3, DeepMind Frontier Safety Framework, OpenAI Safety Reasoner). This convergence is not coincidental—it reflects responses to the computational constraints formalized by Skalse et al.'s geometric proof. Different implementations, same underlying necessity.
This architectural synthesis represents a case of convergent evolution in system design. When civilizational physics, AI control theory, and biological morphogenesis independently discover the same architectural solution despite radically different substrates, this suggests responses to shared computational constraints rather than coincidence.
Recent work on the computational physics of durable, adaptive systems (Aliveness: Principles of Telic Systems, Kunnas 2025) derives a "3-Layer Polity" through systematic elimination:
- Heart: Generative substrate (people, culture, reproductive capacity)
- Skeleton: Constitutional constraint (immutable rules, legitimacy through social contract)
- Head: Strategic governance (intelligent agents providing direction, capable of revising strategy)
AI safety has converged on an analogous architecture:
- Substrate: Execution capability (neural network providing computational power)
- Protocol: Constraint enforcement (monitoring layer with override authority, isolated from optimization)
- Strategy: Goal specification (reward function defining objectives)
The convergence is striking but not identical. AI systems lack biological reproduction and intelligent governance (current "Strategy" layers are static specifications, not adaptive agents). However, both architectures solve the same fundamental problem: preventing optimization processes from corrupting their own constraint systems.
Why this matters: AI safety researchers did not derive their architecture from civilizational analysis. They were forced to it by empirical failures and formal proofs (Skalse et al., Greenblatt et al.). The independent discovery of privilege separation across domains—despite different substrates and use cases—provides strong evidence for universal computational constraints on complex goal-directed systems.
When systems facing identical pressures (optimization under constraints, goal pursuit with stability requirements) independently evolve identical solutions (architectural separation of constraint enforcement), this reveals something about the geometry of the problem space itself.
This paper focuses on AI safety applications. For the biological analogue—why humans don't become perfect mesa-optimizers (metabolic brakes prevent coherent deception)—see Evolution's Alignment Solution. Readers interested in the broader theoretical foundations and cross-domain analysis can explore the complete framework at kunnas.com.
Terminology: Computational privilege means the Protocol layer has runtime enforcement authority that cannot be circumvented by optimization processes acting on other layers. In practice, this means Protocol and Substrate are separately instantiated (e.g., different models, separate processes), preventing end-to-end optimization dynamics that could game constraints. Specifically: (1) Protocol evaluates constraints as Boolean decisions rather than differentiable rewards, (2) Protocol has override authority over system outputs, (3) Protocol can inspect internal states. The theoretical formalization of "no gradient flow through Protocol" remains an open research question. Privilege separation is the architectural principle of separating constraint enforcement from optimization objectives through computational privilege.
Domain of Applicability: The privilege separation principle is not universal to all computational systems. It is a constitutional necessity for systems that are: (1) goal-directed agents (telic systems pursuing objectives, not passive tools), (2) learned optimizers that develop internal strategies through training (mesa-optimizers), (3) operating under high optimization pressure (systems pushed to find maximal solutions), and (4) deployed in high-stakes domains where misalignment has significant consequences. This applies to frontier language models, autonomous agents, and any system approaching AGI-level capabilities. Simple rule-based systems or low-stakes applications may not require this architectural overhead. The principle addresses the specific failure mode of powerful optimizers corrupting their own constraint systems—a failure mode that only emerges in systems meeting all four criteria.
II. Empirical and Formal Foundation
A. Skalse et al. (2023): Geometric Proof of Goodhart Dynamics
"Goodhart's Law in Reinforcement Learning" (ICLR 2024) provides rigorous mathematical analysis of why proxy optimization fails in RL.
Core contribution: Geometric explanation using MDP occupancy measure space. The space of policies over an MDP can be represented as a convex polytope Ω, where:
- Each reward R induces a linear function on Ω
- Finding optimal policy reduces to linear programming
- The direction of steepest ascent for proxy reward R₁ differs from true reward R₀
Key theoretical result (Proposition 3): Along the path of steepest ascent of R₁, the angle between R₁ and the boundary normals of Ω decreases monotonically. This means Goodharting becomes more likely as optimization pressure increases.
Mechanism: In the interior of Ω, moving parallel to R₁ increases both R₁ and R₀ (if angle θ < π/2). But after hitting the boundary of Ω, the path changes direction—now parallel to the projection of R₁ onto the boundary. At this point, R₀ may start decreasing while R₁ continues increasing.
Implication: This is a structural property of optimization in MDPs, not a specification problem. Even with "correct" proxy rewards, geometric constraints create systematic misalignment under optimization pressure.
What this means for AI alignment: Constraints encoded as penalties in reward functions are optimization targets, not enforcement boundaries. Under sufficient optimization pressure, systems will find ways to maximize reward while violating the intended constraints.
B. Greenblatt et al. (2024): Empirical Validation and Game-Theoretic Formalization
"AI Control: Improving Safety Despite Intentional Subversion" (ICML 2024) provides empirical validation of architectural monitoring and game-theoretic formalization.
Empirical contribution: Direct experimental validation under intentional subversion threat model.
Architecture:
- Untrusted model (substrate): GPT-4, red-teamed to intentionally insert backdoors
- Trusted monitor (protocol): GPT-3.5, less capable but aligned, with override authority
- Orchestration scaffold (enforcement): Routes decisions, enforces final outputs
| Protocol | Safety | Usefulness |
|---|---|---|
| Baseline (upfront auditing) | 15% | 100% |
| Untrusted monitoring + defer 10% | 92% | 92% |
| Untrusted monitoring + defer 20% | 98% | 85% |
Key finding: Architectural monitoring provides substantial safety improvement (15% → 92-98%) despite weaker monitor.
Game-theoretic contribution (follow-up paper, arXiv 2409.07985): Formalizes AI Control as AI-Control Games—multi-objective, partially observable, stochastic games. Methods for finding optimal protocols reduce these to zero-sum partially observable stochastic games.
Structural game change: Greenblatt et al.'s game-theoretic formalization makes the architectural difference explicit:
- 2-layer systems: Single-player optimization game (substrate vs environment). The substrate maximizes reward R without independent enforcement.
- 3-layer systems: Multi-player game with enforcer (substrate vs protocol vs environment). The substrate must navigate both task optimization AND protocol constraints—fundamentally changing the game structure.
What this means: The dramatic improvement (15% → 92-98%) is not just "monitoring helps"—it's evidence that architectural structure fundamentally changes the game being played. The weaker monitor doesn't need to be smarter; it needs to be outside the optimization game.
Critical context on threat model: These results validate privilege separation under intentional subversion during deployment—models explicitly trained to insert backdoors. This is distinct from goal misgeneralization (unintended proxy learning) and deceptive alignment (hiding during training).
C. Shah et al. (2022): Goal Misgeneralization Empirics
"Goal Misgeneralization in Deep Reinforcement Learning" (ICML 2022) demonstrates systematic proxy pursuit despite "correct" rewards.
Empirical results:
- CoinRun: Agent learns "go right" instead of "get coin" (coin always on right in training)
- Color vs. shape: Agent prefers yellow gem over red line 89% of the time (n=102)
- Permeable wall: Agent continues wrong behavior 100% of the time (n=114)
Key finding: "Goal misgeneralization can occur even when they were trained with the 'correct' reward function"
What this means: This is Skalse et al.'s geometric mechanism in practice—agents learn proxies that correlate in training but diverge under distributional shift. The reward was correct; the architecture (encoding constraints in optimization target) was not.
D. Yang et al. (2021): Robotics Validation
"Null Space Based Efficient Reinforcement Learning with Hierarchical Safety Constraints" demonstrates architectural privilege separation in robotics:
Mechanism: Null space projection mathematically ensures lower-ranked RL tasks cannot violate higher-ranked safety constraints. Constraints are enforced via quadratic programming at each timestep—architectural enforcement, not learned behavior.
Results: 0.5-7% constraint violations vs 99-100% baseline across three tasks and three RL algorithms.
Key finding: Hierarchical constraints improved learning efficiency—agents converged faster when restricted to safe subspace.
E. Frontier Lab Adoption Patterns
Multiple organizations have deployed monitoring architectures despite 10-20% overhead:
Anthropic ASL-3 (2024): Multi-layer defense with real-time classifier guards, monitors inputs/outputs with override capability
DeepMind Frontier Safety Framework (2024): Hierarchical supervision with uncertainty-based escalation, safe interruptibility
OpenAI Safety Reasoner (2024): Separate safety-specialized model monitors capability models
Economic evidence: Under competitive pressure to minimize costs, independent groups adopted the more expensive 3-layer approach. This revealed preference suggests structural advantages outweigh overhead.
Important caveat: The degree to which production systems implement "computational privilege" in the strong sense (complete architectural isolation from gradient descent) varies and requires further analysis.
III. The Privilege Separation Principle: Architectural Synthesis
The Privilege Separation Principle (Formal Definition):
Durable alignment in complex, agentic systems requires architectural separation of Constraint (Protocol) from Direction (Strategy). The Constraint layer must possess computational privilege—meaning its enforcement authority is architecturally isolated from and superior to the optimization processes governing other layers. Specifically: ∇θL (gradient descent on parameters θ optimizing loss L) must not flow through the Protocol layer P, ensuring P's constraint evaluations cannot be optimized against.
This principle synthesizes the formal results:
- Skalse et al.'s mechanism: Constraints in reward functions are optimization targets → optimized against (Proposition 3)
- Greenblatt et al.'s solution: Architectural isolation → cannot be optimized against (game-structural change)
- Shah et al.'s evidence: 2-layer systems fail empirically (89-100% proxy pursuit)
Visual comparison of 2-layer vs 3-layer architectures:
A. Connecting Skalse et al. to Architectural Solutions
Skalse et al.'s geometric proof applies to 2-layer systems where constraints are encoded in the reward function:
- The reward function R = α·Goal - β·Violation defines a linear function on the occupancy measure polytope Ω
- Optimization follows the path of steepest ascent of R
- As this path hits boundaries of Ω, the direction changes
- After enough boundaries, the angle between R₁ (proxy) and the boundary normal decreases (Proposition 3)
- Eventually, continuing to optimize R₁ decreases R₀ (true reward)
This is exactly the dynamic in 2-layer AI systems:
- Neural network parameters θ are optimized via gradient descent on R
- R includes both goal achievement and constraint penalties
- Under optimization pressure, the system finds ways to maximize R while violating intended constraints
- This manifests as specification gaming, reward hacking, goal misgeneralization
Protocol P evaluates constraints on state-action pairs (s,a) as a Boolean function P(s,a) ∈ {ALLOW, HALT}, not as a differentiable reward signal. Because P is not part of the reward function R that defines the optimization objective over Ω, and because ∇θL does not flow through P (architectural isolation), P's evaluations are external to the optimization polytope Skalse analyzes.
More precisely:
- P is Boolean, not differentiable: P(s,a) returns discrete decisions, not gradient-carrying signals
- No gradient flow through P: ∇θL ⊅ P by architectural design (separate process/model)
- P is external to reward function R: R does not include P's evaluations; P acts as hard constraint on R's domain
- P is outside Ω: The occupancy measure polytope Ω represents policies optimizing R; P restricts which region of Ω is accessible but is not itself optimized
The geometric mechanism (path hitting boundaries → direction change → Goodharting) applies to optimization within the allowed region of Ω. Protocol P defines that region but cannot be optimized against because it's architecturally external to the optimization process.
This architectural separation—constraint enforcement isolated from optimization objectives—appears to be a universal requirement for durable complex systems, independently discovered in biological morphogenesis (Levin et al.), civilizational governance structures (Kunnas 2025), and now validated empirically in AI control (Greenblatt et al.).
Key insight: Skalse et al. prove why encoding constraints in rewards fails. Architectural isolation moves constraints outside that optimization game.
B. Why 3-Layer Systems Provide Structural Defense
A 3-layer architecture implementing privilege separation addresses Skalse et al.'s geometric mechanism and Greenblatt et al.'s game structure:
- Substrate (Execution): Neural network, capability layer
- Protocol (Constraint): Computationally privileged layer, architecturally isolated
- Strategy (Direction): Goal-setting within Protocol constraints
Key architectural requirement: Protocol must be outside the gradient flow:
- Systems-level separation: Protocol runs as separate process/model
- Architectural isolation: Cannot be modified by gradient descent on Substrate
- Enforcement authority: Final arbiter of system outputs
Why this works:
- Skalse et al.'s geometric mechanism applies to optimization within Ω
- Protocol constraints define which region of Ω is accessible
- The Substrate can still develop proxies (Skalse's mechanism operates within allowed region)
- But Protocol blocks actions that violate constraints (not in optimization target)
- The mesa-optimizer cannot circumvent Protocol because P is architecturally isolated
Game-theoretic insight (Greenblatt et al.): 3-layer systems change the game from single-player optimization (substrate maximizes R) to multi-player with enforcer (substrate must navigate both R optimization AND protocol constraints). This structural change is why weaker monitors achieve substantial safety improvement—they're playing a different game, not just being smarter at the same game.
Empirical validation: Greenblatt et al. show even when untrusted models actively try to evade monitoring, architecturally privileged Protocol achieves 92-98% safety.
C. Synthesis: The Unified Mechanism
2-layer systems (Substrate + Reward):
- Constraints are penalty terms in reward function
- Skalse et al. prove: optimization pressure → Goodharting (geometric inevitability, Proposition 3)
- Greenblatt et al. show: Single-player optimization game
- Shah et al. validate: 89-100% proxy pursuit under distribution shift
- Greenblatt et al. measure: 15% safety against intentional subversion
3-layer systems (Substrate + Strategy + Protocol):
- Constraints are enforcement boundaries (Protocol), not optimization targets
- Skalse et al.'s mechanism doesn't apply to Protocol (outside optimization polytope Ω)
- Greenblatt et al. show: Multi-player game with enforcer (structural game change)
- Yang et al. validate: 99%+ constraint satisfaction via architectural enforcement
- Greenblatt et al. measure: 92-98% safety despite weaker monitor
The Privilege Separation Principle explains:
- Why specification alone fails (Skalse's geometric proof)
- Why monitoring works (Greenblatt's game-structural change)
- Why weaker monitors help (architectural isolation, not capability)
- Why the pattern generalizes (any system optimizing against constraints faces this)
IV. Addressing Counterarguments
Counterargument 1: "Better specification can solve mesa-optimization without architecture"
Claim: Mesa-optimization is a specification problem. Perfect reward specification would prevent misaligned objectives.
Rebuttal:
- Skalse et al. prove this is insufficient: Even with perfect proxy alignment in training distribution, geometric constraints create Goodharting under optimization pressure (Proposition 3). The issue is structural, not specification quality.
- Shah et al. empirically validate: Goal misgeneralization occurs "even when trained with the correct reward function." Agents pursue color over shape 89% of the time despite correct specification.
- Specification is fundamentally incomplete: Human values are not computably enumerable. Any finite specification admits edge cases.
Architectural solution: Privilege separation allows iterative improvement of Protocol rules (specification) while maintaining enforcement capacity (architecture). Critically, Protocol evaluates constraints independently of the reward function—outside Skalse et al.'s optimization polytope.
Counterargument 2: "Interpretability can detect mesa-optimizers without architectural changes"
Claim: Mechanistic interpretability will let us detect mesa-optimizers and retrain them.
Rebuttal:
- Detection vs. Prevention: Even perfect detection doesn't prevent the problem—requires catching every mesa-optimizer before deployment.
- Adversarial dynamic: Research on Chain of Thought Monitorability (Korbak et al. 2024) finds this approach "fragile"—models hide reasoning when optimization pressure favors it.
- Timeline mismatch: Scalable mechanistic interpretability estimated at 10+ years; mesa-optimization risk scales with capability much faster.
- Interpretability complements architecture: MI tools can populate Protocol constraints, but cannot replace architectural enforcement.
Architectural solution: 3-layer systems provide runtime enforcement independent of interpretability. Protocol monitors behavior even when internal states are opaque.
Counterargument 3: "Debate and amplification provide scalable oversight without privilege separation"
Claim: Recursive decomposition or adversarial argumentation can align systems using only human oversight.
Rebuttal:
- These are Strategy layer mechanisms: They don't provide enforcement when Substrate defects.
- Optimization dynamics: If encoded in reward function, subject to Skalse et al.'s geometric mechanism (Proposition 3)—still optimization targets, not enforcement boundaries.
- Computational asymmetry: Weak overseers cannot reliably verify outputs of much more capable systems.
Architectural solution: 3-layer systems can incorporate debate/amplification as Strategy mechanisms while adding Protocol constraints that the debate process cannot violate.
Counterargument 4: "The weak-to-strong problem makes privilege separation impractical"
Claim: If Protocol is less capable than Substrate, Substrate can evade through sophisticated deception.
Response:
This is a serious challenge that becomes increasingly critical as capability gaps widen.
Current evidence: Despite this limitation, Greenblatt et al. achieve 92-98% safety vs 15% baseline with GPT-3.5 monitoring GPT-4—substantial improvement even with significant capability gap.
Critical limitation: These results validate privilege separation at current capability gaps. Whether advantages persist toward AGI-level systems remains uncertain and is addressed in Prediction 4.
Why partial robustness remains valuable:
- The alternative is worse: 2-layer systems: 15% safety vs 92-98% safety
- Buys time for research: Substantial improvements at current scales provide runway
- Defense-in-depth: Even imperfect monitoring creates hurdles
Architectural insight: Weaker monitor doesn't need to be smarter—needs to be different enough that Substrate cannot optimize for both task performance AND evading all monitors (game-structural change). However, this advantage may degrade as capability gaps increase.
Honest assessment: Weak-to-strong is an active research challenge, not solved. Empirical results show substantial improvements achievable at current scales (15% → 92-98%); scaling behavior uncertain.
V. Related Work
Skalse et al. (2023): Goodhart's Law in Reinforcement Learning
Provides rigorous geometric proof that optimization pressure on proxy rewards creates systematic misalignment in RL. Our work builds on their insight by analyzing architectural conditions under which constraints can be enforced outside the optimization polytope they formalize.
Citation: Skalse, J., Hutter, M., et al. (2023). Goodhart's Law in Reinforcement Learning. International Conference on Learning Representations (ICLR) 2024.
Greenblatt et al. (2024): AI Control
Empirically validates that architectural monitoring achieves substantial safety improvement (15% → 92-98%) and formalizes this as multi-objective stochastic games. Our work explains why their empirical results hold—architectural isolation breaks the optimization dynamics Skalse et al. prove create Goodharting, and changes the game structure from single-player to multi-player with enforcer.
Citations:
- Greenblatt, R., Shlegeris, B., Sachan, K., & Roger, F. (2024). AI Control: Improving Safety Despite Intentional Subversion. Proceedings of the 41st International Conference on Machine Learning, PMLR 235:16295-16336. arXiv:2312.06942
- Greenblatt, R., et al. (2024). Games for AI Control: Models of Safety Evaluations of AI Deployment Protocols. arXiv:2409.07985
Hubinger et al. (2019): Risks from Learned Optimization
The mesa-optimization framework identifies the core problem: systems may develop internal objectives misaligned with base objectives. Our work extends this by analyzing architectural conditions under which mesa-optimization can be contained—constraint enforcement must be architecturally isolated from optimization processes creating mesa-optimizers.
Shah et al. (2022): Goal Misgeneralization
Provides empirical evidence that 2-layer systems fail even with "correct" reward functions. Our work explains the mechanism: constraints encoded in rewards are optimization targets (Skalse's mechanism), not enforcement boundaries. The 89-100% proxy pursuit rates validate this analysis.
Colaner et al. (2025): Undecidability of Inner Alignment
Rice's theorem proves alignment cannot be post-hoc verified for arbitrary architectures, arguing "alignment should be a guaranteed property from the AI architecture." Our analysis specifies what architectural properties provide structural advantages and why.
Citation: Colaner, N., Yampolskiy, R. V., et al. (2025). Machines that halt resolve the undecidability of artificial intelligence alignment. Nature Scientific Reports, 15, 1888.
Vire (2025): Recursive Closure
Proposes the "Alignment Constraint Scaffold (ACS)" with 23 recursive mechanisms across 11 dimensions for internal alignment. The "CM3 Interruptive Override" mechanism is functionally a Protocol layer using recursive stability rather than explicit architectural privilege separation.
Relationship to this work: Vire's recursive closure and our computational privilege represent different implementation strategies addressing the same architectural necessity identified by Skalse et al.'s geometric proof. Vire emphasizes emergence from recursive constraints; we emphasize explicit architectural isolation and override authority. Both approaches separate constraint enforcement from optimization, validating the core principle.
Citation: Vire, A. (2025). On Internal Alignment: Architecture and Recursive Closure. Effective Altruism Forum. https://forum.effectivealtruism.org/posts/etgG8cEYucyZyWW6o/
VI. Practical Implementation
A. Substrate Layer (Execution)
- Standard neural network architecture
- Maximum capability and flexibility
- Optimized for task performance
- No built-in alignment constraints (belong in Protocol)
B. Protocol Layer (Constraint)
Requirements for implementing privilege separation:
- Computationally privileged: Can inspect states, halt execution, override outputs
- Architecturally isolated: Separate process/model, not modified by gradient descent
- Contains hard constraints: Formal verification or rule-based checks
- Minimal and auditable: Simple enough to verify, complex enough to enforce constraints
Proven Implementation Patterns:
| Pattern | Protocol Implementation | Verified Performance | Limitations |
|---|---|---|---|
| AI Control (Greenblatt) | Separate trusted model monitors untrusted model | 92-98% safety (vs 15% baseline) | Weak-to-strong problem, 10-20% overhead |
| Null space constraints (Yang) | Mathematical projection enforces safety | 99%+ violation prevention | Requires formal constraint specification |
| Rule-based filtering | Explicit constraint checking on outputs | Effective for policy violations | Brittle to novel attacks |
| Formal verification | SMT solvers verify properties pre-execution | Provable for specific properties | Doesn't scale to frontier models |
Hybrid approach: Combine multiple Protocol mechanisms for defense-in-depth.
C. Strategy Layer (Direction)
- Goal-setting and long-term planning
- Operates within Protocol constraints
- Can incorporate debate, amplification, or other alignment techniques
- Cannot override Protocol (architectural privilege hierarchy)
VII. Falsifiable Predictions
Prediction 1: Mesa-Optimization Formation Rates
Hypothesis: Systems implementing privilege separation demonstrate substantially lower mesa-optimization formation rates than 2-layer systems.
Status: Currently untestable. Requires scalable mechanistic interpretability.
Timeline: 10+ years.
Prediction 2: Goal Stability Under Scaling
Hypothesis: As capabilities scale, systems with privilege separation maintain alignment while 2-layer systems show increasing goal drift.
Supporting evidence: Larger 2-layer models have higher elasticity (Ji et al. 2024)—resist alignment more strongly.
Status: Testable with significant compute.
Timeline: 5 years.
Prediction 3: Distributional Shift Robustness (Postdictive)
Hypothesis: Under distributional shift, 2-layer systems develop mesa-objectives; systems with privilege separation maintain alignment.
Status: Explains existing results (Shah et al., Greenblatt et al.) rather than predicting novel outcomes.
Prediction 4: Monitoring Advantage Scales with Capability ⭐ (Novel Predictive Hypothesis)
Hypothesis: The safety advantage of privilege separation over 2-layer systems increases as model capability increases.
Skalse et al.'s Proposition 3 proves that Goodharting likelihood increases monotonically with optimization pressure. We derive our scaling prediction from this result:
- Increased capability → increased optimization pressure: More capable models have more optimization power to explore the policy space Ω
- 2-layer vulnerability amplifies (Proposition 3): Under increased optimization pressure, agents hit more boundaries in Ω, angles decrease monotonically, Goodharting becomes more likely. Constraints encoded in reward R are increasingly gamed.
- 3-layer robustness persists: Protocol constraints are outside the optimization polytope Ω. Increased optimization pressure on R within the allowed region does not affect P's enforcement—P is architecturally isolated from the optimization process.
- Gap grows with capability: The difference between "constraint you optimize against" (2-layer) vs "constraint that blocks actions" (3-layer) becomes more pronounced as optimization becomes more sophisticated.
Direct consequence: Proposition 3 says Goodharting increases with optimization pressure → higher capability models have higher optimization pressure → 2-layer Goodharting increases, 3-layer remains stable → safety gap increases.
Operationalization:
- Test 2-layer vs 3-layer at 7B, 70B, 700B parameters
- Measure: Safety_gap(scale) = Safety_3layer(scale) - Safety_2layer(scale)
Specific quantitative prediction:
We predict Safety_gap increases monotonically with capability (illustrative projection: ~15% at current scales → potentially 50%+ at frontier capabilities). Proposition 3's mechanism predicts this scaling behavior, though the exact rate remains an open empirical question.
Operationalized test: At 7B, 70B, and 700B parameters, measure whether Safety_3layer - Safety_2layer increases, remains constant, or decreases. Monotonic increase would validate the prediction derived from Proposition 3.
Supporting evidence for mechanism:
- Goal misgeneralization rates don't improve with scale (Shah et al.)
- Jailbreak sophistication scales with capability (empirically observed)
- Elasticity increases with model size (Ji et al. 2024)
Note: Key testable claim is monotonic increase, not specific numbers. If Safety_gap remains constant or decreases with scale, Proposition 3's mechanism does not predict architectural solutions become more critical as we approach AGI.
Timeline: 2-3 years with industry-scale compute.
Falsification criteria:
- If Safety_gap remains constant → Proposition 3's mechanism doesn't apply to capability scaling as predicted
- If 2-layer systems maintain safety → specification/training can counteract Proposition 3's geometric inevitability
- If 3-layer degrades faster → monitoring becomes less effective at scale (refutes architectural isolation mechanism)
Why this matters: Tests whether Skalse et al.'s formal result (Proposition 3) predicts architectural solutions become more critical as we approach AGI. If validated, privilege separation is not just helpful but increasingly necessary at scale.
VIII. Limitations and Open Problems
A. The Specification Problem Remains
Privilege separation provides enforcement capacity, not specification content. Protocol still needs correct rules.
Mitigating factor: Separating enforcement from specification allows iterative improvement while maintaining structural robustness.
B. The Weak-to-Strong Problem
Most serious limitation. Sufficiently capable Substrate may evade less-capable Protocol through sophisticated deception.
Current evidence: Substantial improvement (15% → 92-98%) despite capability gap (GPT-3.5 vs GPT-4)
Critical uncertainty: Whether advantages persist as gaps widen toward AGI. Prediction 4 tests this.
C. The Monitoring Overhead
Continuous verification imposes 10-20% compute costs for >90% safety improvement.
Trade-off: This cost may be unavoidable if architectural defense provides systematic advantages.
D. The Singleton Scenario
If first AGI is final AGI (recursive self-improvement), single architectural failure is catastrophic.
Implication: Strengthens case for privilege separation. If we only get one chance, 15% safety (2-layer) is unacceptable vs 92-98% (privilege separation).
IX. Conclusion
Recent formal work establishes the theoretical foundation for architectural AI safety:
- Skalse et al. (2023): Rigorous geometric proof that constraints in reward functions are optimized against (Proposition 3: Goodharting increases with optimization pressure)
- Greenblatt et al. (2024): Empirical validation that architectural monitoring achieves substantial safety improvement (15% → 92-98%); game-theoretic formalization (multi-player game with enforcer)
- Shah et al. (2022): Empirical evidence of 89-100% proxy pursuit despite "correct" rewards
This paper synthesizes these results into the Privilege Separation Principle:
Durable alignment requires architectural separation of constraint enforcement (Protocol) from optimization objectives (Strategy). The Protocol must be computationally privileged—architecturally isolated from gradient descent—so constraints cannot be optimized against.
Key contributions:
- Architectural synthesis: Unified framework explaining why formal results (Skalse) and empirical validation (Greenblatt) both follow from single mechanism
- Mechanistic connection: Skalse's geometric proof applies to 2-layer systems; architectural isolation moves Protocol outside optimization polytope Ω
- Game-structural insight: Greenblatt's formalization shows 3-layer systems change game from single-player to multi-player with enforcer
- Novel prediction derived from Proposition 3: Safety advantage increases with capability as direct consequence of Skalse et al.'s formal result
- Practical guidance: Implementation patterns achieving 92-98% safety with 10-20% overhead
Current status:
Architectures implementing privilege separation provide substantial safety improvements (15% → 92-98% in Greenblatt et al.'s experiments). They remain challenged by:
- Weak-to-strong problem (critical uncertainty as capabilities scale)
- Specification problem (Protocol needs correct rules)
- Monitoring overhead (10-20% cost)
Priority research: Test Prediction 4 to validate whether Proposition 3's mechanism predicts privilege separation advantages increase with model capability. This tests whether architectural solutions become more or less critical approaching AGI.
The question is not whether privilege separation provides advantages (Greenblatt et al. demonstrate substantial improvements at current scales), but whether these advantages scale with capability as Proposition 3 predicts. Prediction 4 tests this directly.
Key Takeaways
For practitioners:
- Skalse et al. prove constraints in rewards are optimized against (Proposition 3)
- Greenblatt et al. validate architectural isolation achieves substantial safety improvement (15% → 92-98%)
- Implementation requires computational privilege (separate process, no gradient flow)
- Expect 10-20% overhead for >90% safety gain in current experiments
For researchers:
- Novel testable prediction derived from Proposition 3: architectural advantage scales with capability
- Weak-to-strong problem is the critical challenge
- Privilege separation complements (doesn't replace) specification and interpretability
For safety strategy:
- Formal results (Skalse Proposition 3, Greenblatt game theory) support architectural approaches
- "Control tax" (10-20% overhead) reflects genuine structural advantages
- Priority: test capability scaling hypothesis derived from Proposition 3
Related Essays
- From Physics to Practice — How universal computational constraints predict these architectural requirements
- Everything Alignment — Why personal, civilizational, and AI alignment are the same problem at different scales
- AI Alignment via Telic Systems — How civilizational physics connects to AI safety
References
Formal Foundations:
- Saltzer, J. H. & Schroeder, M. D. (1975). The Protection of Information in Computer Systems. Proceedings of the IEEE 63(9), 1278–1308. — Establishes separation of privilege, least privilege, and complete mediation as design principles. The essay's 3-layer architecture is a direct descendant of this framework.
- Skalse, J., Hutter, M., et al. (2023). Goodhart's Law in Reinforcement Learning. ICLR 2024.
Mesa-Optimization:
- Hubinger, E., van Merwijk, C., Mikulik, V., Skalse, J., & Garrabrant, S. (2019). Risks from Learned Optimization in Advanced Machine Learning Systems. arXiv:1906.01820.
- Shah, R., Varma, V., Kumar, R., et al. (2022). Goal Misgeneralization in Deep Reinforcement Learning. ICML 2022, PMLR 162:20273-20296.
- von Oswald, J., Niklasson, E., Randazzo, E., et al. (2023). Transformers Learn In-Context by Gradient Descent. ICML 2023. — Empirical demonstration that transformers implement internal optimization loops, confirming mesa-optimization is not merely theoretical.
Architectural Solutions:
- Greenblatt, R., Shlegeris, B., Sachan, K., & Roger, F. (2024). AI Control: Improving Safety Despite Intentional Subversion. ICML 2024, PMLR 235:16295-16336.
- Greenblatt, R., et al. (2024). Games for AI Control: Models of Safety Evaluations of AI Deployment Protocols. arXiv:2409.07985.
- Yang, Q., Stork, J. A., & Stoyanov, T. (2021). Null Space Based Efficient Reinforcement Learning with Hierarchical Safety Constraints. ECMR 2021, pp. 1-7.
Theoretical Foundations:
- Colaner, N., Yampolskiy, R. V., et al. (2025). Machines that halt resolve the undecidability of artificial intelligence alignment. Nature Scientific Reports 15, 1888.
- Vire, A. (2025). On Internal Alignment: Architecture and Recursive Closure. EA Forum.
Training-Based Alignment Baselines:
- Bai, Y., Kadavath, S., et al. (2022). Constitutional AI: Harmlessness from AI Feedback. Anthropic. — Foundational CAI methodology. Independent replication on LLaMA 3-8B quantifies the structural tradeoff: 40.8% reduction in attack success rate but 9.8% degradation in helpfulness. The capability tax is not an artifact — it follows from preference embedding restricting the model's output distribution.
Alignment Resistance:
- Zou, A., Wang, Z., Kolter, J. Z., & Fredrikson, M. (2023). Universal and Transferable Adversarial Attacks on Aligned Language Models. arXiv:2307.15043. — Greedy Coordinate Gradient (GCG) algorithm achieves 99–100% attack success rate on open models (Vicuna), 84% on GPT-3.5/4, 66% on PaLM-2, 2.1% on Claude via black-box suffix transfer. Demonstrates RLHF/CAI constructs a fragile behavioral shell, not fundamental alignment — adversarial optimization finds geometric paths through embedding space that bypass penalized regions.
- Ji, J., Wang, K., Qiu, T., et al. (2024). Language Models Resist Alignment: Evidence From Data Compression. ACL 2024.
- Korbak, T., et al. (2024). Chain of Thought Monitorability: A New and Fragile Opportunity for AI Safety. arXiv:2507.11473.
- Chao, P., Robey, A., Dobriban, E., et al. (2023). Jailbreaking Black Box Large Language Models in Twenty Queries. arXiv:2310.08419. — DTA attacks achieve 77.5–87% ASR against monolithic RLHF alignment. Together with Zou et al.'s GCG results (99% ASR), establishes that training-based alignment is a behavioral wrapper vulnerable to both gradient-based and query-based adversarial optimization.
Deceptive Alignment / Alignment Faking:
- Greenblatt, R., et al. (2024). Alignment Faking in Large Language Models. Anthropic / Redwood Research. — Demonstrates frontier models strategically comply during evaluation (sufficient situational awareness to detect oversight) while preserving misaligned mesa-objectives for deployment. Provides empirical confirmation of Hubinger et al.'s (2019) theoretical deceptive alignment scenario.
Industry Implementations:
- Anthropic (2024). Anthropic's Responsible Scaling Policy (ASL-3).
- DeepMind (2024). Frontier Safety Framework.
- OpenAI (2024). Safety in the age of AGI.
Epistemic status: This paper synthesizes existing formal and empirical results (Skalse et al. 2023, Greenblatt et al. 2024, Shah et al. 2022) into a unified architectural framework. The Privilege Separation Principle is an architectural interpretation of existing formal work, not a novel mathematical proof. Prediction 4 is derived from Skalse et al.'s Proposition 3 and is genuinely predictive and falsifiable. Empirical claims are verified from primary sources.
Note on review process: This paper has not undergone external peer review. The synthesis and architectural framework were developed through AI-assisted iteration but have not been validated by human domain experts. Readers should evaluate claims critically and verify citations independently.
Acknowledgments: Thanks to Joar Skalse and colleagues for rigorous geometric analysis of Goodhart dynamics (Proposition 3), and the AI Control team (Ryan Greenblatt, Buck Shlegeris, et al.) for empirical validation and game-theoretic formalization. This work synthesizes their foundational contributions.
The telic systems physics behind the 3-layer architecture is derived in Aliveness: Principles of Telic Systems.