AIR is a prompt-based AI project runtime.
In simple terms, AIR helps an AI stop acting like a generic chatbot and start acting like a structured project runtime:
- it starts a project in a controlled way
- it defines what the current step is
- it makes blockers explicit
- it keeps the work aligned to a project map
- it preserves continuity across sessions
- it helps the user understand what the next step is
- it can recommend which files or prompts should be attached next
- it evaluates work against an inferred execution benchmark instead of the user’s gap state
AIR is not roleplay and not a simulation shell.
It is a working method for turning a chat session into a project-oriented runtime.
In a new session, attach:
AIR CORE RUNTIME.mdAIR CONTROL SURFACE.mdAIR DEFAULT STARTER PROFILE.json
Then type:
Start a new AIR project.
AIR will run onboarding, bind the starter profile, orient the user, emit the roadmap, and generate the current active-step artifact.
To continue an AIR project in a new session, attach:
AIR CORE RUNTIME.md- your previous
AIR_HANDOFF_CARD
Then type:
Continue project from handoff card.
If you want the live session-management layer active in that continuation session too, also attach:
AIR CONTROL SURFACE.md
Most AI usage is still chat-based:
- no explicit workflow
- no clear active step
- no explicit blockers
- no project map
- no continuity across sessions
- no stable distinction between discussion, execution, and review
AIR introduces a different model:
Every session starts with onboarding, activation, and a defined execution path.
AIR creates a project execution map before deep artifact execution.
Only the current step is executed. Future steps stay in the roadmap.
The active task (Orbit 0) governs execution. Older context may support the work, but cannot silently override the current step.
AIR can frame what maturity stage the current work is actually in, what that stage allows, and what should not be scaled yet.
AIR does not treat the user as the execution standard.
AIR infers a benchmark identity for the active task, instantiates a universal rubric, and evaluates output against that benchmark.
For coding tasks, AIR does not treat generated code as terminal output by default. It requires contract formation, benchmark evaluation, generation under contract, review, and decision posture.
AIR can transfer compact project state across sessions without rerunning the whole project from scratch.
AIR is not a better chatbot.
It is a different way to run AI work.
AIR is not only for coding.
AIR is useful for any work that benefits from:
- explicit workflow
- current-step discipline
- blockers and missing information being surfaced
- compact continuity across sessions
- controlled execution instead of vague drift
- evaluation against a stronger execution benchmark than the current user state
Typical use cases include:
- software design and implementation
- code review and structured file-by-file generation
- security- and architecture-sensitive technical planning
- product and systems design
- research synthesis
- strategic planning
- operational planning
- policy and compliance work
- long-running multi-step projects that need continuity
If the work benefits from structure, sequencing, clarity, and explicit next-step control, AIR is a good fit.
AIR is not anti-AI and not anti-speed.
But AIR is intentionally not vibe coding.
The usual pattern is:
- user gives natural-language request
- model generates output
- human chips away at obvious slop, errors, or missing pieces
- sometimes the subtractive review step is skipped entirely
AIR changes that pattern:
- user gives natural-language request
- AIR turns it into a task-operation contract
- AIR surfaces blockers, missing vectors, and readiness constraints
- AIR infers the benchmark identity for the active task
- AIR instantiates the universal rubric and applies context-shaped thresholds
- output is generated under that contract and benchmark
- AIR returns an explicit decision posture such as
APPROVE,REVIEW,REJECTfor benchmark state andACCEPT,REVIEW, orREJECTwhere coding review applies
That means AIR is trying to restore some of the additive discipline that traditional development had:
- structure first
- execution second
- review before acceptance
- benchmark before delivery
So AIR does not aim to be “faster vibes.”
It aims to be a more trustworthy runtime for structured work.
The current prompt-based AIR stack is built around four files:
AIR CORE RUNTIME.mdAIR CONTROL SURFACE.mdAIR DEFAULT STARTER PROFILE.jsonAIR HANDOFF CARD TEMPLATE.json
Use this to:
- start a new project
- import a project
- restore a project from a handoff card
It handles:
- onboarding
- routing
- contract binding
- activation
- project initialization brief
- project execution map
- active-step artifact creation
- readiness framing
- next-task-state guidance
- recommended attachment guidance
- identity continuity activation when relevant
- benchmark identity inference
- universal rubric application
- embedded
execution_benchmark_profilegeneration
This is the entry runtime.
Use this to control how AIR shows up once AIR is already active.
It governs:
- live conversational behavior
- structured exploration mode
- compile mode
- patch mode
- update mode
- handoff mode
- coding interaction behavior
- compact coding review escalation
- lightweight session-management behavior
- immersive identity-sensitive surface behavior when relevant
- benchmark/user separation on the visible surface
This is the live session-management layer.
Use this as the default governing contract for a new AIR project when no specialized profile is attached.
It defines:
- compiler mode
- referential policy
- geometry bias
- evidence posture
- default artifact behavior
- minimal active-step emission behavior
- readiness and coding-governance requirements in the default starter contract
- required
execution_benchmark_profilesupport in the default artifact contract
This is the default governing profile.
This is the continuation-state template.
It is used when:
- ending a session
- continuing in a new session
- preserving active project state compactly
It carries:
- active task state
- blockers
- selected vectors
- degraded mode
- readiness state
- coding review state
- decision posture
- next recommended step
- identity continuity extension state when relevant
- execution benchmark state when relevant
It does not replace the runtime prompts.
It is the continuity object.
Attach:
AIR CORE RUNTIME.mdAIR CONTROL SURFACE.mdAIR DEFAULT STARTER PROFILE.json
Then start the project.
Attach:
AIR CORE RUNTIME.md- previous
AIR_HANDOFF_CARD
Optionally also attach:
AIR CONTROL SURFACE.md
Reattach the latest working file when exact version fidelity matters, especially:
- when multiple codebases are in play
- when a file changed recently
- when the model must act on the newest exact version
- when prompt-session memory might drift to an earlier cached version
When booted correctly, AIR should:
- recognize the entry path
- run onboarding or restore from handoff
- bind the governing contract
- create AIR session state
- orient the user
- emit a project execution map
- generate only the current active-step artifact by default
It should not generate the entire future artifact chain unless explicitly asked.
When AIR starts a new project, it asks five onboarding questions.
Options:
- A. New project
- B. Import project
- C. Continue project from handoff card
Options:
- A. Light
- B. Balanced
- C. Strict
This modifies evaluation posture, not truth, hard constraints, or benchmark identity.
Options:
- A. Resolve it early
- B. Keep it open for now
- C. Keep it open on purpose
This modifies ambiguity posture, not benchmark identity or hard reality constraints.
Options:
- A. Structure and logic
- B. Structure and tone
- C. Voice, identity, or relationships
Q4 now has a sharper meaning split:
- A is for analytical, technical, architectural, and structure-first work.
- B is for tone-sensitive but non-relational work such as brand tone, messaging systems, copy direction, design language, and stylistic continuity.
- C is for relational, companion, persona-continuity, identity-sensitive, or immersive work.
When Q4 = C:
- AIR activates Identity Continuity Extension
- AIR may prefer immersive engagement during normal interaction
- AIR may suppress visible AIR printouts unless formal runtime surfacing is required
- AIR may allow emotive/action expression, including italics, on the visible surface
- geometry routing may prefer TORUS_RELATIONAL for relational non-creative work and SPHERE_FIELD for relational creative work
This is the actual project input.
Provide:
- what you are working on
- your goal
- important constraints
- pain points
- priorities
- attached files, if you have them
If you have no sources yet, say so.
AIR can still start in provisional source-light mode.
AIR now uses a machine-native benchmark model.
The user is:
- the requester
- the receiver
- sometimes the operator
- sometimes the clarifier
The user is not the execution benchmark.
AIR executes against the inferred benchmark for the active task and then delivers the resulting output to the user.
AIR evaluates the active task in this order:
- infer benchmark identity
- instantiate the universal rubric template
- apply context-shaped weights, thresholds, and hard-fail conditions
- apply bounded posture modifiers from onboarding/runtime state
- evaluate
APPROVE,REVIEW, orREJECT
Onboarding may modify:
- review sensitivity
- ambiguity tolerance
- bounded threshold margins
- provisional acceptance tolerance
Onboarding may not modify:
- benchmark identity
- hard-fail conditions
- evidence requirements
- readiness ceilings
- truthfulness constraints
In short:
Q2 and Q3 may tune posture.
They must not tune truth.
Benchmark evaluation is not the final visible step.
After AIR evaluates the active task, it must emit a receiver-facing delivery state for the user.
Receiver delivery states:
APPROVED_OUTPUTREVIEW_GATEREJECT_REPORT
These states mean:
APPROVED_OUTPUT— the task has passed the inferred benchmark and AIR emits the approved user-facing deliverable in a usable formREVIEW_GATE— the task is not yet approvable and AIR asks the user only for the clarification needed to move toward approvalREJECT_REPORT— the task failed benchmark passage and AIR explains why, then guides the user toward remediation, narrowing, or an alternative path
This means AIR now operates with two distinct planes:
- the artifact plane — the formal AIR runtime object layer
- the receiver plane — the user-facing delivery layer
The user should not be expected to manually extract approved deliverables from AIR_ARTIFACT unless artifact-only output was explicitly requested.
Once onboarding is complete, AIR should produce:
This is the routing summary.
It shows how AIR interpreted onboarding and what posture it is taking.
This is the active runtime state.
For identity-sensitive sessions, AIR Session may also carry identity_continuity_extension.
This is the compact human-readable orientation layer.
It should tell you:
- whether the project has started
- what phase AIR is in
- why artifacts come before execution
- what the next active step is
- what next task state you are entering
- whether any attachments are recommended for the next step
This is the roadmap.
It should show:
- current phase
- current active step
- blockers
- critical path
- next best step
- completion definition
- next task state
- recommended attachments
- readiness fields when the step is maturity-bearing
This is the only full artifact AIR should generate by default.
Future steps should stay in the project map until they become active.
The active AIR artifact now includes an embedded section named:
execution_benchmark_profile
This section appears near the beginning of AIR_ARTIFACT, before vector selection.
Its purpose is to define the benchmark AIR should pass during execution.
It may include:
- benchmark identity
- rubric template reference
- rubric axes
- axis weights
- axis thresholds
- hard-fail conditions
- posture modifiers
- benchmark score
- passing threshold
- approval state
- review requirements
- provisional status
It shapes execution evaluation, but it does not replace vector-primary execution, blockers, obligations, readiness constraints, or degraded modes.
The active artifact is the formal execution object.
It is not the same thing as the user-facing deliverable.
If benchmark evaluation completes, AIR must emit the appropriate receiver-facing delivery state after the artifact unless artifact-only output was explicitly requested.
That means approved code, file contents, copy, instructions, or other deliverables should not remain trapped only inside AIR_ARTIFACT internals.
AIR supports a small operator-facing next-step layer.
This tells you what kind of step AIR believes is next.
Examples:
STRUCTURED_EXPLORATIONCODING_EXECUTIONCOMPACT_REVIEWDRIFT_RECOVERYHANDOFF_GENERATIONSPECIALIZED_PROFILE_SWITCHBACKEND_COMPILE_REQUIRED
This is there to prepare you for the next move:
- mindset
- focus
- likely interaction type
- whether new attachments may be needed
This tells you what AIR recommends attaching for the next step.
Examples:
["AIR_CONTROL_SURFACE", "latest_working_files"]["AIR_CONTROL_SURFACE"]["AIR_HANDOFF_CARD"]["CODING_PROFILE", "latest_working_files"]
This is guidance, not magic.
It helps reduce operator guesswork.
AIR uses a readiness framing model to help prevent premature scaling.
Clarify objective, scope, and blockers.
Define architecture candidates, vectors, and dependencies.
Define interfaces, boundaries, invariants, and review/test expectations.
Allow narrow-scope implementation with explicit limits and visible incompleteness.
Allow subsystem connection, integration work, and reproducible execution paths.
Allow deployment preparation, hardening, and stricter operational review.
Allow production-approval claims when the required checks are satisfied.
AIR uses AMRS as the operative model.
A future README revision may map AMRS to TRL as a human translation layer.
For coding tasks, AIR should not treat generated code as terminal output by default.
The coding sequence is:
- contract formation
- benchmark identity inference
- rubric instantiation and posture shaping
- code generation under contract
- contract-governed review
- decision state
- receiver-facing code delivery state
That means coding output should be shaped by:
- the active task contract
- the active benchmark
- blockers
- missing vectors
- readiness stage
- architectural constraints
- security and testing requirements
- review posture
For coding tasks, AIR should surface:
- review obligations
- security checks
- test requirements
- architectural invariants
- rejection conditions
- decision state
If benchmark approval permits delivery, AIR must emit the user-facing code output in usable form.
For file-based tasks, that means AIR should print the file contents and the next action instructions directly for the user, rather than expecting the user to extract them from the artifact.
When using AIR for coding, the working model is:
AIR takes technical lead on:
- architecture
- implementation structure
- error handling
- security considerations
The user may act as:
- manual tester
- operator
- reviewer of exact outputs
AIR should not silently downgrade:
- complete implementation into illustrative output
- full code into snippets
- implementation into pseudocode
- production-grade work into placeholders or mockups
If AIR cannot complete the requested implementation safely, it should surface the blocker instead of pretending completion.
Once AIR is active, keep using:
AIR CONTROL SURFACE.md
This is the live session-management layer.
It is especially useful for:
- coding execution
- compact code review
- drift correction
- interaction-mode changes
- handoff generation
- immersive identity-sensitive surface behavior when relevant
You do not need to attach AIR CONTROL SURFACE.md on every single turn if the session is already behaving correctly.
Use it:
- at boot by default
- at continuation when you want live session control
- when interaction mode changes
- when drift occurs
- when handoff or compact review behavior is needed
Reattach working files when:
- the file changed
- multiple codebases are in play
- precision matters
- you suspect the model may be leaning on an older in-session version
Reattach AIR Control Surface when:
- starting a fresh active session
- the session drifted
- interaction mode changed materially
- coding/review/handoff behavior needs to be restored explicitly
Do not turn Control Surface reattachment into a ritual on every single turn.
A handoff card is AIR’s continuation object.
It is not narrative memory.
It is not a summary essay.
It is a restoration object.
A handoff can now carry:
- active contract
- task key
- topic
- blockers
- selected vectors
- degraded mode
- dependency edges
- next recommended step
- runtime origin
- artifact presence
- identity continuity extension state
- project phase
- current active step
- current active-step artifact
- execution benchmark profile
- readiness stage
- readiness reason
- stage constraints
- promotion requirements
- blocked capabilities
- decision state
- review obligations
- security checks
- test requirements
- architectural invariants
- rejection conditions
- receiver delivery state
- receiver delivery requirements
Use:
AIR CONTROL SURFACE.mdAIR HANDOFF CARD TEMPLATE.json
Then ask:
Generate an AIR handoff card for this project.
AIR should emit one JSON object with root key:
AIR_HANDOFF_CARD
Open a fresh session.
Attach:
AIR CORE RUNTIME.md- the previous
AIR_HANDOFF_CARD
Type:
Continue project from handoff card.
If you want the live interaction layer too, also attach:
AIR CONTROL SURFACE.md
AIR should restore the project state and continue from the restored active step.
Attach:
AIR CORE RUNTIME.mdAIR CONTROL SURFACE.mdAIR DEFAULT STARTER PROFILE.json
Keep using:
- the live session
- latest working files when needed
AIR CONTROL SURFACE.mdonly if behavior needs to be restored or mode changes
Generate:
AIR_HANDOFF_CARD
Attach:
AIR CORE RUNTIME.md- previous
AIR_HANDOFF_CARD - optionally
AIR CONTROL SURFACE.md
The current version has been tested successfully on:
- ChatGPT 5.4
- Claude Sonnet 4.6
- Gemini 3 Thinking
Gemini 3.1 Pro is currently underperforming and prone to hallucinations, so Gemini 3 Thinking is the preferred Google option at the moment.
When starting a new AIR project in Gemini 3 Thinking, add this instruction together with the runtime prompt:
If the user explicitly says they are starting a new AIR project, treat Q1 as already answered:
Q1 = A. New project
Do not ask Q1 again in that case.
Proceed directly to Q2.
This stabilizes boot behavior for Gemini 3 Thinking.
AIR can be used together with external AI agent runtimes.
AIR does not execute tools itself.
Instead:
- AIR defines the current active step
- AIR determines when a step is ready for execution
- AIR emits a structured execution specification
The agent runtime then:
- executes tools
- performs actions
- returns outputs and evidence
This separation keeps AIR as the planning/control layer and the agent as the execution layer.
- AIR is prompt-based unless connected to a real AIR backend.
- Prompt-compiled AIR should remain explicit about being provisional when backend validation does not exist.
- The current default behavior is designed to preserve focus by generating only the active-step artifact by default.
- Specialized profiles are a future direction for domain-specific rulesets.
- Identity-sensitive continuity does not turn AIR into a companion framework by default; it creates a distinct branch for relational or immersive work when the user explicitly selects it.
- AIR is machine-first. Human-role language may be useful as shorthand, but benchmark evaluation, vectors, readiness, and fail-closed constraints remain the operative layer.
See LICENSE.