Constitutional AI under Finite Capacity

A structural constitution for accountable AI systems operating under finite capacity.

This public release proposes a constitutional framework for AI systems that operate under bounded resources, uncertainty, overload, latency constraints, safety pressure, and limited reviewability.

The central claim is simple:

Accountable AI requires more than principles, policies, or output-level constraints. Under finite capacity, an AI system must distinguish admissibility from execution, support explicit boundary acts, and preserve witness sufficient for later review.



Core proposal

A constitutionally governed AI system should minimally preserve:

This proposal is not a legal code, a complete ethics framework, or a mandatory software standard. It is a structural governance proposal intended to sit between high-level AI principles and proprietary implementation details.


Public release package

The release is organised as a structured package rather than a single paper.

For policy and governance readers

Recommended entry points:

  1. Policy Translation Brief
  2. Instrument Index
  3. The Finite-Capacity AI Constitution

For technical safety and evaluation readers

Recommended entry points:

  1. The Finite-Capacity AI Constitution
  2. Minimal Reviewable Evidence Specification
  3. Minimal Application Profile for AI Assistant Systems

Package structure

The release contains six main instruments:

  1. The Finite-Capacity AI Constitution
    The constitutional core.

  2. Constitutional Force and Applicability
    Clarifies universal minimum force, governance-strengthening clauses, and profile-dependent qualifications.

  3. Minimal Application Profile for AI Assistant Systems
    Projects the constitutional grammar into AI assistant systems.

  4. Minimal Reviewable Evidence Specification
    States the minimum evidence surface required for constitutional claims to be inspectable.

  5. Questions, Limits, and Misreadings
    Clarifies what the framework claims, what it does not claim, and how it should not be over-read.

  6. Policy Translation Brief
    Translates the package into public-governance language for policy, standards, compliance, and public-authority audiences.


Review and feedback

This release is intended as a public, citable, reviewable contribution to AI governance, AI safety, and constitutional AI discussions.

Feedback is especially welcome on:

Contact: research@synkyria.uk