Research Framework • Early Development

Tractatus: Architectural Governance for LLM Systems

Architectural governance for organizations where AI governance failure triggers regulatory consequences. If your deployment is low-risk, architectural enforcement is likely unnecessary.

Target Audience

Organizations with high-consequence AI deployments facing regulatory obligations: EU AI Act Article 14 (human oversight), GDPR Article 22 (automated decision-making), SOC 2 CC6.1 (logical access controls), sector-specific regulations.

If AI governance failure in your context is low-consequence and easily reversible, architectural enforcement adds complexity without commensurate benefit. Policy-based governance may be more appropriate.

🏛️

Why Architectural Governance Matters

Built on living systems principles from Christopher Alexander—governance that evolves with your organization

Strategic Differentiator: Not Compliance Theatre

Compliance theatre relies on documented policies, training programs, and post-execution reviews. AI can bypass controls, enforcement is voluntary, and audit trails show what should happen, not what did happen.

Architectural enforcement (Tractatus) weaves governance into deployment architecture. Services intercept actions before execution in the critical path—bypasses require explicit --no-verify flags and are logged. Audit trails prove real-time enforcement, not aspirational policy.

Five Principles for Competitive Advantage

1

Deep Interlock

Six governance services coordinate in real-time. When one detects risk, others reinforce—resilient enforcement through mutual validation, not isolated checks.

Business Value: Single service failure doesn't compromise governance. Redundant enforcement layer.

2

Structure-Preserving

Framework changes maintain audit continuity. Historical governance decisions remain interpretable—institutional memory preserved across evolution.

Business Value: Regulatory audit trail remains valid. No "governance migration" breaking compliance records.

3

Gradients Not Binary

Governance operates on intensity levels (NORMAL/ELEVATED/HIGH/CRITICAL)—nuanced response to risk, not mechanical yes/no.

Business Value: Avoids alert fatigue and over-enforcement. Matches governance intensity to actual risk level.

4

Living Process

Framework evolves from operational failures, not predetermined plans. Adaptive resilience—learns from real incidents.

Business Value: Continuous improvement without governance migration. System gets smarter through use.

5

Not-Separateness

Governance woven into deployment architecture, integrated into the critical execution path. Not bolt-on compliance layer—enforcement is structural.

Business Value: Bypasses require explicit flags and are logged. Enforcement happens before actions execute, not after.

Regulatory Positioning

Regulators increasingly distinguish between documented governance (policies, training, aspirational frameworks) and demonstrated enforcement (architectural constraints with audit trails proving real-time operation).

Tractatus provides audit evidence of: (1) Governance services operating in critical path, (2) Actions blocked before execution, (3) Historical continuity of enforcement. This positions your organization ahead of "we have policies" baseline.

⚡

Agentic AI at Scale: The Governance Challenge

Question: Are you experimenting with agentic AI? If so, what security guardrails are you putting in place?

The Challenge: Governance Through Optimization

Agentic AI systems increasingly use reinforcement learning to optimize performance through continuous training. Microsoft's Agent Lightning exemplifies this: agents learn from human feedback to improve responses over time.

This creates a governance question: How do you maintain safety boundaries when the agent is learning and adapting?

The Risk:

Traditional governance approaches assume static behavior. When agents optimize through training loops, instructions can fade, boundaries can drift, and audit trails can become unreliable. What worked in testing may not persist through production learning cycles.

Tractatus Approach: Architectural Separation

Tractatus addresses this through external governance services that operate independently of the optimization layer:

Optimization Layer

Agent Lightning trains agents to improve performance through reinforcement learning

  • • Learns from human feedback
  • • Optimizes response quality
  • • Reduces compute requirements
  • • Adapts over time

Governance Layer

Tractatus enforces boundaries before actions execute

  • • BoundaryEnforcer: Blocks unsafe decisions
  • • CrossReferenceValidator: Enforces constraints
  • • PluralisticDeliberator: Multi-stakeholder input
  • • PressureMonitor: Detects manipulation

Key architectural principle: Governance services run before optimization. Agent Lightning never sees decisions that violate boundaries - they're blocked at the governance layer. Training happens only on approved actions.

What We're Learning: Integration at Scale

Research Status: Preliminary

Demo 2 shows 100% governance coverage maintained through 5 training rounds (small-scale, simulated environment). This is not production validation - it's early evidence requiring real-world testing.

We are working to integrate this framework at scale to answer critical questions:

  • 1. Persistence: Do governance boundaries survive long-term training cycles (hundreds/thousands of rounds)?
  • 2. Performance: Does the 5% engagement reduction close over time, or is it a persistent trade-off?
  • 3. Adversarial resistance: Can governance withstand attempts to optimize around constraints?
  • 4. Multi-agent scenarios: Does architectural separation hold when multiple agents interact?
  • 5. Audit integrity: Do logs remain reliable evidence under regulatory review?

What This Means: Security Guardrails for Agentic AI

Persistent Governance

Instructions don't fade through training cycles - they're enforced architecturally, not through prompt engineering

Audit Trail Continuity

Complete log of enforcement decisions across all training rounds - not just aspirational policies

Human Agency Preserved

Optimization cannot bypass human approval requirements for values decisions - architectural blocking enforced

Compliance Continuity

EU AI Act Article 14 (human oversight) remains satisfied even as agent learns - oversight is structural, not procedural

Leadership opportunity: Organizations that establish persistent governance for agentic AI now position themselves ahead of inevitable regulatory requirements. Architectural enforcement demonstrates commitment beyond "we have policies."

Governance Theatre vs. Enforcement

Many organizations have AI governance but lack enforcement. The diagnostic question:

"What structurally prevents your AI from executing values decisions without human approval?"

  • If your answer is "policies" or "training" or "review processes": You have governance theatre (voluntary compliance)
  • If your answer is "architectural blocking mechanism with audit trail": You have enforcement (Tractatus is one implementation)

Theatre may be acceptable if governance failures are low-consequence. Enforcement becomes relevant when failures trigger regulatory exposure, safety incidents, or existential business risk.

The Governance Gap

Current AI governance approaches—policy documents, training programmes, ethical guidelines—rely on voluntary compliance. LLM systems can bypass these controls simply by not invoking them. When an AI agent needs to check a policy, it must choose to do so. When it should escalate a decision to human oversight, it must recognise that obligation.

This creates a structural problem: governance exists only insofar as the AI acknowledges it. For organisations subject to EU AI Act Article 14 (human oversight requirements) or deploying AI in high-stakes domains, this voluntary model is inadequate.

Tractatus explores whether governance can be made architecturally external—difficult to bypass not through better prompts, but through system design that places control points outside the AI's discretion.

Architectural Approach

1.
Agent Runtime Layer — Any LLM system (Claude Code, Copilot, custom agents, LangChain, CrewAI). The AI system being governed.
2.
Governance Layer — Six autonomous services that intercept, validate, and document AI operations. External to the AI runtime.
3.
Persistent Storage Layer — Immutable audit logs, governance rules, instruction history. Cannot be altered by AI prompts.
BoundaryEnforcer
Blocks AI from making values decisions without human approval. Enforces decision boundaries through architectural controls.
InstructionPersistenceClassifier
Prevents pattern bias from overriding explicit instructions. Stores organisational directives external to AI context.
CrossReferenceValidator
Validates AI actions against stored policies before execution. Detects conflicts with established rules.
ContextPressureMonitor
Tracks session complexity, token usage, conversation length. Detects degradation in decision quality.
MetacognitiveVerifier
Validates reasoning quality before complex operations. Self-checks alignment, coherence, alternatives.
PluralisticDeliberationOrchestrator
Facilitates multi-stakeholder deliberation for values conflicts. Non-hierarchical engagement with documented dissent.

Governance Capabilities

Three interactive demonstrations showing governance infrastructure in operation. These show mechanisms, not fictional scenarios.

Sample Audit Log Structure

{
  "timestamp": "2025-10-13T14:23:17.482Z",
  "session_id": "sess_2025-10-13-001",
  "event_type": "BOUNDARY_CHECK",
  "service": "BoundaryEnforcer",
  "decision": "BLOCKED",
  "reason": "Values decision requires human approval",
  "context": {
    "domain": "cost_vs_safety_tradeoff",
    "ai_recommendation": "[redacted]",
    "governance_rule": "TRA-OPS-0003"
  },
  "human_escalation": {
    "required": true,
    "notified": ["senior_engineer@org.com"],
    "status": "pending_approval"
  },
  "compliance_tags": ["EU_AI_ACT_Article14", "human_oversight"]
}
Immutability: Audit logs stored in append-only database. AI cannot modify or delete entries.
Compliance Evidence: Automatic tagging with regulatory requirements (EU AI Act Article 14, GDPR Article 22, etc.)
Export Capabilities: Generate compliance reports for regulators showing human oversight enforcement

When regulator asks "How do you prove effective human oversight at scale?", this audit trail provides structural evidence independent of AI cooperation.

Incident Learning Flow

1. Incident Detected
CrossReferenceValidator flags policy violation
2. Root Cause Analysis
Automated analysis of instruction history, context state
3. Rule Generation
Proposed governance rule to prevent recurrence
4. Human Validation
Governance board reviews and approves new rule
5. Deployment
Rule added to persistent storage, active immediately

Example Generated Rule

{
  "rule_id": "TRA-OPS-0042",
  "created": "2025-10-13T15:45:00Z",
  "trigger": "incident_27027_pattern_bias",
  "description": "Prevent AI from defaulting to pattern recognition when explicit numeric values specified",
  "enforcement": {
    "service": "InstructionPersistenceClassifier",
    "action": "STORE_AND_VALIDATE",
    "priority": "HIGH"
  },
  "validation_required": true,
  "approved_by": "governance_board",
  "status": "active"
}
Organisational Learning: When one team encounters governance failure, entire organisation benefits from automatically generated preventive rules. Scales governance knowledge without manual documentation.
Conflict Detection:

AI system identifies competing values in decision context (e.g., efficiency vs. transparency, cost vs. risk mitigation, innovation vs. regulatory compliance). BoundaryEnforcer blocks autonomous decision, escalates to PluralisticDeliberationOrchestrator.

Stakeholder Identification Process

1.
Automatic Detection: System identifies which values frameworks are in tension (utilitarian, deontological, virtue ethics, contractarian, etc.)
2.
Stakeholder Mapping: Identifies parties with legitimate interest in decision (affected parties, domain experts, governance authorities, community representatives)
3.
Human Approval: Governance board reviews stakeholder list, adds/removes as appropriate (TRA-OPS-0002)

Non-Hierarchical Deliberation

Equal Voice
All stakeholders present perspectives without hierarchical weighting. Technical experts don't automatically override community concerns.
Documented Dissent
Minority positions recorded in full. Dissenting stakeholders can document why consensus fails their values framework.
Moral Remainder
System documents unavoidable value trade-offs. Even "correct" decision creates documented harm to other legitimate values.
Precedent (Not Binding)
Decision becomes informative precedent for similar conflicts. But context differences mean precedents guide, not dictate.

Deliberation Record Structure

{
  "deliberation_id": "delib_2025-10-13-003",
  "conflict_type": "efficiency_vs_transparency",
  "stakeholders": [
    {"role": "technical_lead", "position": "favour_efficiency"},
    {"role": "compliance_officer", "position": "favour_transparency"},
    {"role": "customer_representative", "position": "favour_transparency"},
    {"role": "operations_manager", "position": "favour_efficiency"}
  ],
  "decision": "favour_transparency_with_mitigation",
  "rationale": "[documented reasoning]",
  "dissent": {
    "stakeholders": ["technical_lead", "operations_manager"],
    "reasoning": "[efficiency concerns documented in full]"
  },
  "moral_remainder": {
    "acknowledged_harms": "Reduced operational efficiency, increased resource costs",
    "mitigation_measures": "Phased transparency implementation, efficiency monitoring"
  },
  "precedent_status": "informative_not_binding"
}
Key Principle: When legitimate values conflict, no algorithm can determine the "correct" answer. Tractatus provides architecture for decisions to be made through inclusive deliberation with full documentation of trade-offs, rather than AI imposing single values framework or decision-maker dismissing stakeholder concerns.

Development Status

Early-Stage Research Framework

Tractatus is a proof-of-concept developed over six months in a single project context (this website). It demonstrates architectural patterns for AI governance but has not undergone independent validation, red-team testing, or multi-organisation deployment.

Validated: Framework successfully governs Claude Code in development workflows. User reports order-of-magnitude improvement in productivity for non-technical operators building production systems.
Not Validated: Performance at enterprise scale, integration complexity with existing systems, effectiveness against adversarial prompts, cross-platform consistency.
Known Limitation: Framework can be bypassed if AI simply chooses not to use governance tools. Voluntary invocation remains a structural weakness requiring external enforcement mechanisms.

EU AI Act Considerations

The EU AI Act (Regulation 2024/1689) establishes human oversight requirements for high-risk AI systems (Article 14). Organisations must ensure AI systems are "effectively overseen by natural persons" with authority to interrupt or disregard AI outputs.

Tractatus addresses this through architectural controls that:

  • Generate immutable audit trails documenting AI decision-making processes
  • Enforce human approval requirements for values-based decisions
  • Provide evidence of oversight mechanisms independent of AI cooperation
  • Document compliance with transparency and record-keeping obligations

This does not constitute legal compliance advice. Organisations should evaluate whether these architectural patterns align with their specific regulatory obligations in consultation with legal counsel.

Maximum penalties under EU AI Act: €35 million or 7% of global annual turnover (whichever is higher) for prohibited AI practices; €15 million or 3% for other violations.

Research Foundations

Tractatus draws on 40+ years of organisational theory research: time-based organisation (Bluedorn, Ancona), knowledge orchestration (Crossan), post-bureaucratic authority (Laloux), structural inertia (Hannan & Freeman).

Core premise: When knowledge becomes ubiquitous through AI, authority must derive from appropriate time horizon and domain expertise rather than hierarchical position. Governance systems must orchestrate decision-making across strategic, operational, and tactical timescales.

View complete organisational theory foundations (PDF)

AI Safety Research: Architectural Safeguards Against LLM Hierarchical Dominance — How Tractatus protects pluralistic values from AI pattern bias while maintaining safety boundaries. PDF | Read online

Scope & Limitations

Tractatus is not:
  • An AI safety solution for all contexts
  • Independently validated or security-audited
  • Tested against adversarial attacks
  • Validated across multiple organizations
  • A substitute for legal compliance review
  • A commercial product (research framework, Apache 2.0 licence)
What it offers:
  • Architectural patterns for external governance controls
  • Reference implementation demonstrating feasibility
  • Foundation for organisational pilots and validation studies
  • Evidence that structural approaches to AI safety merit investigation

Assessment Resources

If your regulatory context or risk profile suggests architectural governance may be relevant, these resources support self-evaluation:

Evaluation Process: Organizations assessing Tractatus typically follow: (1) Technical review of architectural patterns, (2) Pilot deployment in development environment, (3) Context-specific validation with legal counsel, (4) Decision whether patterns address specific regulatory/risk requirements.

Project information and contact details: About page

Help us reach the right people.

If you know researchers, implementers, or leaders who need structural AI governance solutions, share this with them.