Visual schema
Minimal evidence chain of an integrity audit
An audit does not hold because of an isolated verdict, but because of a continuous chain between canon, capture, gap, proof, and correction.
Source
Declared canonPages, exclusions, versions, and machine-first surfaces that fix what may be opposed.
Frame
Perimeter and authorityDelimit what enters the audit, what stays out of scope, and which authorities prevail.
Capture
Runs and capturesPrompts, systems, dates, context, observed outputs, and execution conditions are logged.
Time-bound trace
Finding
Canon-output gapQualify drift, collision, omission, or faulty reformulation in an explorable form.
Proof
Proof of fidelityEach finding must be tied back to excerpts, snapshots, or matrices that can be re-read.
Loop
Correction and monitoringThe loop closes through published correction, rerun, and continuity monitoring.
Governance artifacts
Governance files brought into scope by this page
This page is anchored to published surfaces that declare identity, precedence, limits, and the corpus reading conditions. Their order below gives the recommended reading sequence.
Q-Metrics JSON
/.well-known/q-metrics.json
Descriptive metrics surface for observing gaps, snapshots, and comparisons.
- Governs
- The description of gaps, drifts, snapshots, and comparisons.
- Bounds
- Confusion between observed signal, fidelity proof, and actual steering.
Does not guarantee: An observation surface documents an effect; it does not, on its own, guarantee representation.
Q-Metrics YAML
/.well-known/q-metrics.yml
YAML projection of Q-Metrics for instrumentation and structured reading.
- Governs
- The description of gaps, drifts, snapshots, and comparisons.
- Bounds
- Confusion between observed signal, fidelity proof, and actual steering.
Does not guarantee: An observation surface documents an effect; it does not, on its own, guarantee representation.
Q-Ledger JSON
/.well-known/q-ledger.json
Machine-first journal of observations, baselines, and versioned gaps.
- Governs
- The description of gaps, drifts, snapshots, and comparisons.
- Bounds
- Confusion between observed signal, fidelity proof, and actual steering.
Does not guarantee: An observation surface documents an effect; it does not, on its own, guarantee representation.
Complementary artifacts (3)
These surfaces extend the main block. They add context, discovery, routing, or observation depending on the topic.
Q-Ledger YAML
/.well-known/q-ledger.yml
YAML projection of the Q-Ledger journal for procedural reading or tooling.
Iip Scoring Standard Manifest
/iip-scoring.standard.manifest.json
Surface that makes explicit the conditions of response, restraint, escalation, or non-response.
Iip Report Schema
/iip-report.schema.json
Observation surface that exposes logs, metrics, snapshots, or measurement protocols.
Evidence layer
Probative surfaces brought into scope by this page
This page does more than point to governance files. It is also anchored to surfaces that make observation, traceability, fidelity, and audit more reconstructible. Their order below makes the minimal evidence chain explicit.
- 01Canon and scopeDefinitions canon
- 02Response authorizationQ-Layer: response legitimacy
- 03Weak observationQ-Ledger
- 04Derived measurementQ-Metrics
Definitions canon
/canon.md
Opposable base for identity, scope, roles, and negations that must survive synthesis.
- Makes provable
- The reference corpus against which fidelity can be evaluated.
- Does not prove
- Neither that a system already consults it nor that an observed response stays faithful to it.
- Use when
- Before any observation, test, audit, or correction.
Q-Layer: response legitimacy
/response-legitimacy.md
Surface that explains when to answer, when to suspend, and when to switch to legitimate non-response.
- Makes provable
- The legitimacy regime to apply before treating an output as receivable.
- Does not prove
- Neither that a given response actually followed this regime nor that an agent applied it at runtime.
- Use when
- When a page deals with authority, non-response, execution, or restraint.
Q-Ledger
/.well-known/q-ledger.json
Public ledger of inferred sessions that makes some observed consultations and sequences visible.
- Makes provable
- That a behavior was observed as weak, dated, contextualized trace evidence.
- Does not prove
- Neither actor identity, system obedience, nor strong proof of activation.
- Use when
- When it is necessary to distinguish descriptive observation from strong attestation.
Q-Metrics
/.well-known/q-metrics.json
Derived layer that makes some variations more comparable from one snapshot to another.
- Makes provable
- That an observed signal can be compared, versioned, and challenged as a descriptive indicator.
- Does not prove
- Neither the truth of a representation, the fidelity of an output, nor real steering on its own.
- Use when
- To compare windows, prioritize an audit, and document a before/after.
Complementary probative surfaces (4)
These artifacts extend the main chain. They help qualify an audit, an evidence level, a citation, or a version trajectory.
Q-Attest protocol
/.well-known/q-attest-protocol.md
Optional specification that cleanly separates inferred sessions from validated attestations.
IIP report schema
/iip-report.schema.json
Public interface for an interpretation integrity report: scope, metrics, and drift taxonomy.
CTIC compliance report schema
/ctic-compliance-report.schema.json
Public schema for publishing compliance findings without exposing the full private logic.
AI changelog
/changelog-ai.md
Public log that makes AI surface changes more dateable and auditable.
Interpretation integrity audit: full end-to-end protocol
An interpretation integrity audit answers a simple question: does an AI system faithfully return an entity, a doctrine, or a corpus under real conditions of use? In an interpreted web, truth may be available, indexed, and publicly documented, yet still fail to survive model compression, retrieval shortcuts, or silent inference.
This end-to-end protocol formalizes a complete cycle: define the canon, set the authority boundary, test outputs, gather proof, diagnose drift, correct the environment, and monitor continuity over time.
Operational definition
Interpretation integrity audit is the disciplined process by which a declared canon is compared with real model outputs under bounded conditions, in order to determine whether the resulting interpretation is faithful, explainable, and governable.
When to trigger the audit
The audit should be triggered when one or more of the following occur:
- recurrent misattribution or unstable identity;
- doctrinal drift between canonical statements and outputs;
- high-impact recommendations or answer surfaces that now influence decision or action;
- major version changes, public releases, or broad machine-first publication;
- unexplained differences across models, agents, or environments.
Application surfaces
The protocol can be applied to:
- canonical definitions;
- doctrinal pages;
- frameworks and derived governance surfaces;
- entity and brand representations;
- RAG systems, agents, or hybrid answer environments;
- high-impact recommendation or qualification surfaces.
Audited objects
What is audited is not only the page. The audit may target:
- a canon;
- an authority perimeter;
- a retrieval chain;
- a family of outputs;
- an interpretive workflow;
- a correction cycle over time.
Expected output model
A proper audit should produce more than a score. It should yield:
- the declared canon;
- the authority boundary used for arbitration;
- test conditions and evidence surface;
- observed output classes;
- the canon-to-output gap;
- the diagnosis of failure mode;
- the correction path and monitoring plan.
Protocol (AII-1 to AII-10)
AII-1: define the canon
Name the canonical corpus, freeze the version, and specify what counts as the source of truth. Without that step, later judgments will drift into opinion.
AII-2: establish the authority boundary
Clarify which sources are primary, secondary, contextual, or non-authoritative. The audit cannot distinguish fidelity from extrapolation if authority remains implicit.
AII-3: define response conditions
State under which conditions the system is allowed to answer, when it should clarify, and when legitimate non-response must prevail.
AII-4: build the test set
Construct prompts, retrieval conditions, agent flows, or evaluation cases that reflect actual use and known ambiguity points.
AII-5: capture outputs
Collect outputs under declared conditions. Preserve enough context to compare answers across runs or systems without pretending to exhaustive reconstruction.
AII-6: classify the canon-to-output relation
Sort outputs into explicit alignment, bounded inference, unresolved ambiguity, extrapolation, contradiction, or silence.
AII-7: gather proof
Record citations, traces, authority use, missing evidence, and any interpretive boundary that explains why the answer was or was not legitimate.
AII-8: diagnose the failure mode
Determine whether the issue comes from source hierarchy, retrieval quality, ambiguity, missing canon, category error, stale context, or unstable recommendation behaviour.
AII-9: organize correction
Correction may be endogenous (canon, structure, machine-first files) or exogenous (signal environment, discoverability, surrounding references). The audit should state which layer must move.
AII-10: monitor after correction
No audit is complete without continuity. A post-correction monitoring window is required to verify whether the measured gap narrows or simply changes form.
Why the protocol matters
Without an end-to-end audit, integrity remains rhetorical. The protocol gives the site a way to move from declared governance to tested interpretive behaviour.
Read also
- Canon vs inference
- Proof of fidelity
- Interpretive observability
- IIP-Scoring™
Artefacts normally expected
A mature audit usually leaves behind a canonical corpus definition, an authority map, a response-condition model, a captured output set, evidence traces, a gap classification, and a correction plan. Those artefacts matter because they allow later review instead of forcing each audit to start from memory.
Audit posture
The protocol should be run with a deliberately non-triumphal posture. Its job is not to prove that the site is always correct. Its job is to surface how close, bounded, or drifted the real outputs are under declared conditions.