Evidence layer
Probative surfaces brought into scope by this page
This page does more than point to governance files. It is also anchored to surfaces that make observation, traceability, fidelity, and audit more reconstructible. Their order below makes the minimal evidence chain explicit.
- 01Canon and scopeDefinitions canon
- 02Response authorizationQ-Layer: response legitimacy
- 03Weak observationQ-Ledger
Definitions canon
/canon.md
Opposable base for identity, scope, roles, and negations that must survive synthesis.
- Makes provable
- The reference corpus against which fidelity can be evaluated.
- Does not prove
- Neither that a system already consults it nor that an observed response stays faithful to it.
- Use when
- Before any observation, test, audit, or correction.
Q-Layer: response legitimacy
/response-legitimacy.md
Surface that explains when to answer, when to suspend, and when to switch to legitimate non-response.
- Makes provable
- The legitimacy regime to apply before treating an output as receivable.
- Does not prove
- Neither that a given response actually followed this regime nor that an agent applied it at runtime.
- Use when
- When a page deals with authority, non-response, execution, or restraint.
Q-Ledger
/.well-known/q-ledger.json
Public ledger of inferred sessions that makes some observed consultations and sequences visible.
- Makes provable
- That a behavior was observed as weak, dated, contextualized trace evidence.
- Does not prove
- Neither actor identity, system obedience, nor strong proof of activation.
- Use when
- When it is necessary to distinguish descriptive observation from strong attestation.
Accountability surface
This page owns the term “accountability surface” inside the interpretive governance corpus. It is the canonical definition for SERP ownership and internal routing.
An accountability surface is a page, artifact, log, policy, ledger or structured record that makes responsibility for an AI-mediated interpretation legible enough to assign, challenge, correct or delimit.
Short definition
An accountability surface is a page, artifact, log, policy, ledger or structured record that makes responsibility for an AI-mediated interpretation legible enough to assign, challenge, correct or delimit.
Why it matters
Accountability surfaces matter because responsibility cannot be inferred safely from visibility alone. A public page, a model answer, a support exchange, a third-party citation and an internal memory object may all influence interpretation. The accountability surface declares which part of the system can speak, which part is only contextual and which part is not authorized to create commitments.
In AI search, RAG and agentic environments, the problem usually appears after the output has left the generation interface. A response becomes part of a support exchange, a policy explanation, a decision path, a public summary, a workflow or a third-party representation. At that point, quality is no longer enough. The output must be assumable, challengeable and corrigible.
What it is not
An accountability surface is not a marketing page and not a guarantee of enforcement. It is a reading and responsibility layer. It helps distinguish authorship, authority, observation, policy, proof, correction and non-promises.
The distinction matters editorially. A blog post can illustrate the risk and a framework can operationalize the control, but this page is the canonical definition. Internal links should point to Accountability surface when the term itself is introduced.
Common failure modes
- a model answer appears attributable but no source accepts that role
- a policy artifact is read as operational enforcement
- a third-party summary reallocates authority silently
- a correction exists but no surface declares its scope
- a service page is mistaken for binding engagement terms
These failure modes are ordinary in systems that compress evidence, infer from incomplete material, hide arbitration, reuse stale state or treat retrieval as authorization.
Governance implication
The site should maintain accountability surfaces for identity, canon, source hierarchy, response legitimacy, observations, correction and non-promises. These surfaces support semantic accountability without pretending to control every downstream system.
For implementation, this term should be read with answer legitimacy, source hierarchy, proof of fidelity, interpretation trace, contestability and procedural validity.
Relation to phase 10 inference control
Phase 10 asks whether reasoning, completion and arbitration remain legitimate. Phase 11 asks whether the resulting output can survive reliance, challenge, correction and institutional review. A response can stay within an interpretive fidelity and still fail if it lacks a challenge path, a responsibility surface or a valid procedure.
Related canonical definitions
- Answer legitimacy
- Source hierarchy
- Proof of fidelity
- Response conditions
- Authority boundary
- Interpretation trace
- Opposability
- Enforceability
- Commitment boundary
- Liability reduction
- Contestability
- Procedural validity
- Challenge path
- Accountability surface
Supporting surfaces
Reading guidance
Use Accountability surface when an answer may be challenged, relied upon, escalated, contractualized, or used in a consequential environment. The issue is whether the response can be reconstructed, defended, limited, and contested under the rules that govern the context.
What to verify
- Whether the response has crossed from information into commitment.
- Whether the authority of the source, statement, and system is explicit enough to be challenged.
- Whether uncertainty, refusal, or qualification is preserved instead of smoothed away.
- Whether a reviewer can reconstruct the path from canon to output.
Practical boundary
This concept does not create legal enforceability by itself. It names the conditions that must be tested before a response is treated as assumable, opposable, or procedurally valid.