v0.5 - March 2026

INTENT |

Evidence-driven delivery for BUILD and enforceable runtime governance for live AI systems.

Framework Map

A live system view of INTENT v0.5: Proof Tier (P0-P4), BUILD Autonomy Tier (A0-A4), and the runtime governance boundary that controls live agent behavior.

Paused
Intent Contract anchor

Live phase

FRAME

FRAME to EXPLORE

Define outcomes, constraints, and proof expectations through an Intent Contract. This phase creates clarity before execution begins.

Key Activities

  • - Write the Intent Contract with problem, outcomes, constraints, quality bar, and open questions.
  • - Define the initial Scenario Bank for the work item.
  • - Set initial Proof Tier and target Autonomy Tier for this change.

Key Outputs

  • - Intent Contract
  • - Initial Scenario Bank
  • - Proof/Autonomy target
P1

Basic automated checks; coverage is inconsistent.

Spec continuum

  • - P0-P1: Specs guide agents (spec-first) - code is the source of truth.
  • - P2-P3: Specs constrain agents (spec-anchored) - specs and code co-evolve.
  • - P4: Specs govern agents (spec-as-source) - specs are the source of truth, code is disposable output.
A2

AI implements; humans review + run proof.

Active handoff

FRAME to EXPLORE

  • - Intent Contract
  • - Initial Scenario Bank
  • - Proof/Autonomy target

Pairing Mode (where most teams are)

P1 (Repeatable) x A2 (Implement): execution is faster, but delivery still depends on manual coordination and review. Raise proof strength first, then expand autonomy safely.

Roles and Team Models

2-4 people split core capabilities and use shared policies for proof and autonomy.

Core Principles

Intent over activity+

Define what must be true in outcomes before discussing how work will be done.

Flow over cadence+

Move work continuously through proof gates instead of waiting for calendar rituals.

Judgment over specialization+

Human leverage comes from cross-domain judgment while agents execute implementation tasks.

Governance by design+

Policy, risk handling, and quality checks are built into the system from the start.

Transparent contribution+

Teams should know what was agent-generated, human-authored, and human-reviewed.

Challenge inherited constraints+

Regularly test assumptions from pre-agent workflows and re-evaluate what is now feasible.

Runtime Governance Artifacts

Use runtime governance when an AI system is live, customer-facing, or safety-critical. Build-time governance defines what should be true. Runtime governance defines how to enforce it in production behavior.

Applies when

  • - Voice agents handling live customer calls.
  • - Chat agents giving medical, legal, or financial guidance.
  • - Autonomous systems executing trades, deployments, or physical actions.
  • - Safety monitors that can escalate or suppress alerts.

Usually not required

  • - BUILD-phase coding agents where humans review outputs before release.
  • - Internal analysis tools with a human decision-maker in the loop.
  • - Spec and documentation generators.
  • - Sandbox or staging agents that do not affect live users.

Trust Envelope

Defines the positive boundary of what the live agent is authorized to say or do.

Core structure

  • - Authorized output categories with boundaries and examples.
  • - Confidence levels (scripted, templated, generated).
  • - Violation severity classes and detection methods.

Governance authority

Binding runtime policy referenced by the Constitution and versioned with the Intent Contract.

Escalation Contract

Defines every path from agent-handled interaction to human involvement, including failure handling.

Core structure

  • - Escalation classes (severity-based, user-requested, system-initiated).
  • - Step-by-step protocol with timeouts and fallback paths.
  • - User communication, evidence capture, and precedence rules.

Governance authority

Binding operational safety contract referenced by the Intent Contract and validated during runtime testing.

Interaction State Model

Defines valid interaction states, transition triggers, and per-state permissions.

Core structure

  • - State definitions: entry conditions, duration, required data, exit rules.
  • - Transition rules for interruptions, overrides, and recovery.
  • - State mapping to Trust Envelope categories and Scenario Bank paths.

Governance authority

Behavioral runtime architecture that must exist before prompt/runtime implementation.

Governance Stack

Constitution defines non-negotiables. Runtime artifacts define the live-operating boundary and fail-safe behavior.

Constitution
Trust Envelope
Escalation Contract
Interaction State Model

Integration With Existing INTENT Artifacts

ArtifactRequired update
ConstitutionAdd Runtime Governance Requirements that make Trust Envelope, Escalation Contract, and Interaction State Model mandatory for live, customer-facing, or safety-critical agents.
Intent ContractAdd version references for the three runtime artifacts plus a Runtime Governance validation section in the proof plan.
Scenario BankMake scenarios state-aware with state paths, escalation classes, and trust-envelope category mapping.
Proof ReportTrack runtime metrics: unsafe-answer rate, false/missed escalation rate, state violation rate, and escalation success rate.
Autonomy PolicySeparate BUILD autonomy tiers from RUNTIME output authority; runtime authority is bounded by the Trust Envelope.

Governed Autonomy

BUILD autonomy is budgeted by risk and earned by proof. RUNTIME authority is bounded by the Trust Envelope and Escalation Contract.

High risk

BUILD: humans lead and agents assist. RUNTIME: trust envelope is narrow and escalation is fast.

Medium risk

BUILD: agents implement while humans review. RUNTIME: bounded categories with explicit escalation classes.

Low risk + strong proof

BUILD: agents can operate under exception review. RUNTIME: autonomy expands only inside a validated trust envelope.

Drift & Traceability

Drift is when behavior diverges from declared governance: Constitution and Intent Contract in BUILD, plus Trust Envelope, Escalation Contract, and Interaction State Model in RUNTIME. INTENT treats drift as a named, instrumented failure mode at three levels.

Scenario Bank

Scenario Bank is the intent-level memory of correctness. It captures expected outcomes, failure modes, runtime signals, and state paths so validation stays aligned with real behavior.

Build-time drift

Spec validators run in CI. If generated code violates Constitution constraints or Intent Contract commitments, the build fails. This includes schema validation, contract tests, and architectural conformance checks.

Runtime drift

Production behavior is continuously checked against Trust Envelope categories, escalation paths, and state-transition rules. When behavior diverges beyond tolerance, alerts fire and the relevant Intent Contract is flagged for review.

Spec drift

The Intent Contract itself can drift from reality when telemetry reveals incorrect assumptions or when contracts are not updated to match shipped behavior. Detection compares stated outcomes against production metrics.

How drift feeds back

  • - Build-time drift blocks deployment and creates a remediation task.
  • - Runtime drift generates new candidate scenarios and state transitions for the Scenario Bank.
  • - Spec drift triggers a LEARN phase review and potential Intent Contract update.

INTENT vs. Agile Scrum

DimensionAgile ScrumINTENT
Fundamental unitUser storyIntent Contract
CadenceTime-boxed sprintsContinuous flow through proof gates
Team structureFixed role Scrum teamCapabilities + agents + policy
PlanningSprint planningContinuous contract definition
Status updatesDaily standupEvidence and exception dashboards
QAEmbedded QAScenario Bank + Proof Tiers
ExecutionHumans writeAgents execute under gates
ValidationTest suites as primaryScenario outcomes + supporting tests
BottleneckCoordination overheadIntent clarity + evidence quality
GovernanceProcess compliancePolicy as code
InvariantsImplicit team normsConstitution (versioned, enforced)
Drift handlingManual code reviewAutomated build-time, runtime, and spec drift detection

30/60/90 rollout

First 30 days

  • - Introduce intent contracts + seed scenario bank.
  • - Define risk categories and baseline governance policy.
  • - Draft your Constitution: architecture invariants, security non-negotiables, tech stack constraints.

60 days

  • - Enforce scenario gates for meaningful changes.
  • - Draft the Trust Envelope, Escalation Contract, and Interaction State Model for live agent surfaces.
  • - Measure scenario health + defect escapes.
  • - Enable build-time drift detection: CI validates against Constitution invariants.

90 days

  • - Calibrate autonomy budgets.
  • - Expand A-tier only when P-tier supports it.
  • - Instrument runtime governance metrics and drift detection against Scenario Bank expectations.

Ecosystem

INTENT is an operating model, not an engineering methodology. It composes with - rather than competes against - the emerging SDD tooling ecosystem.

INTENT (Operating Model)

Team structure, roles, ceremonies, governance posture, rollout cadence, and the organizational contract between humans and agents.

SDD (Engineering Methodology)

Spec-first workflows and pipelines such as constitution, specify, plan, tasks, implement, validate, and drift detection.

Agent Tooling (Execution Runtime)

Agents that write code, run tests, deploy, and monitor in IDE and CI runtime contexts.

INTENT layer (this framework): Defines team structure, roles, ceremonies, governance posture, rollout cadence, and the organizational contract between humans and agents. Answers who decides, who reviews, what evidence is needed, and how autonomy scales.

SDD layer (engineering methodology): Defines spec-first workflows where structured specifications drive generation, validation, and drift detection. GitHub Spec Kit, AWS Kiro, and Tessl formalize this into repeatable pipelines. INTENT BUILD and VALIDATE phases execute through these workflows.

Agent tooling layer (execution runtime): Claude Code, Cursor, GitHub Copilot, Microsoft Amplifier, and OpenAI Codex. INTENT and SDD remain tool-agnostic: they define constraints and workflows while tools execute.

Teams adopting INTENT should select SDD tooling that fits their stack and configure it to enforce their Constitution and Intent Contracts. The framework does not prescribe specific tools because the ecosystem is evolving rapidly.

INTENT + GitHub Spec Kit

INTENT gives teams a governance layer for who decides, what proof is required, and how autonomy scales in BUILD plus enforceable runtime boundaries for live systems. GitHub Spec Kit gives teams a spec-driven implementation pipeline. The integration page combines both into a single operating model and includes ready-to-use templates.

Intent Kit (templates)

Intent Contract template+
  • - Outcome to achieve
  • - Constraints and non-goals
  • - Proof plan and acceptance scenarios
  • - Risk tier and autonomy target
Scenario template+
  • - Scenario name and user context
  • - Expected success outcome
  • - Failure modes and edge conditions
  • - Signals to capture during execution
Proof Report template+
  • - Scenario pass/fail summary
  • - Defect escapes and severity
  • - Evidence confidence statement
  • - Ship/hold recommendation
Risk tiers and autonomy policy+
  • - Risk classification criteria
  • - Allowed autonomy per risk class
  • - Required proof thresholds
  • - Exception handling path
Constitution template++
  • - Architecture invariants and constraints
  • - Security, compliance, and privacy non-negotiables
  • - Reliability targets and SLOs
  • - Runtime governance trigger and required artifacts
  • - Tech stack rules and forbidden patterns
  • - Agent behavioral boundaries and permissions
  • - Amendment process and approval requirements
Trust Envelope template+
  • - Authorized output categories with examples and boundaries
  • - Confidence levels for each output category
  • - Violation severity and detection methods
  • - Versioning and governance authority references
Escalation Contract template+
  • - Escalation classes and precedence matrix
  • - Protocol, timeouts, and user communication per class
  • - Fallback paths when escalation fails
  • - Evidence capture requirements for each event
Interaction State Model template+
  • - State definitions, entry/exit rules, and required data
  • - Transition conditions including interruption handling
  • - State to Trust Envelope category mapping
  • - Scenario Bank state-path coverage mapping

Use INTENT to improve delivery by tightening intent quality, increasing proof strength, and scaling autonomy only when evidence supports it - with runtime governance artifacts for live agent behavior.

Contact for implementation help