Governance artifacts
Governance files brought into scope by this page
This page is anchored to published surfaces that declare identity, precedence, limits, and the corpus reading conditions. Their order below gives the recommended reading sequence.
Definitions canon
/canon.md
Canonical surface that fixes identity, roles, negations, and divergence rules.
- Governs
- Public identity, roles, and attributes that must not drift.
- Bounds
- Extrapolations, entity collisions, and abusive requalification.
Does not guarantee: A canonical surface reduces ambiguity; it does not guarantee faithful restitution on its own.
Site context
/site-context.md
Notice that qualifies the nature of the site, its reference function, and its non-transactional limits.
- Governs
- Editorial framing, temporality, and the readability of explicit changes.
- Bounds
- Silent drifts and readings that assume stability without checking versions.
Does not guarantee: Versioning makes a gap auditable; it does not automatically correct outputs already in circulation.
Public AI manifest
/ai-manifest.json
Structured inventory of the surfaces, registries, and modules that extend the canonical entrypoint.
- Governs
- Access order across surfaces and initial precedence.
- Bounds
- Free readings that bypass the canon or the published order.
Does not guarantee: This surface publishes a reading order; it does not force execution or obedience.
Complementary artifacts (1)
These surfaces extend the main block. They add context, discovery, routing, or observation depending on the topic.
Canonical AI entrypoint
/.well-known/ai-governance.json
Neutral entrypoint that declares the governance map, precedence chain, and the surfaces to read first.
Evidence layer
Probative surfaces brought into scope by this page
This page does more than point to governance files. It is also anchored to surfaces that make observation, traceability, fidelity, and audit more reconstructible. Their order below makes the minimal evidence chain explicit.
- 01Canon and scopeDefinitions canon
- 02Weak observationQ-Ledger
- 03Derived measurementQ-Metrics
Definitions canon
/canon.md
Opposable base for identity, scope, roles, and negations that must survive synthesis.
- Makes provable
- The reference corpus against which fidelity can be evaluated.
- Does not prove
- Neither that a system already consults it nor that an observed response stays faithful to it.
- Use when
- Before any observation, test, audit, or correction.
Q-Ledger
/.well-known/q-ledger.json
Public ledger of inferred sessions that makes some observed consultations and sequences visible.
- Makes provable
- That a behavior was observed as weak, dated, contextualized trace evidence.
- Does not prove
- Neither actor identity, system obedience, nor strong proof of activation.
- Use when
- When it is necessary to distinguish descriptive observation from strong attestation.
Q-Metrics
/.well-known/q-metrics.json
Derived layer that makes some variations more comparable from one snapshot to another.
- Makes provable
- That an observed signal can be compared, versioned, and challenged as a descriptive indicator.
- Does not prove
- Neither the truth of a representation, the fidelity of an output, nor real steering on its own.
- Use when
- To compare windows, prioritize an audit, and document a before/after.
AI citation readiness and interpretive governance
AI citation readiness is the ability of a page, passage, entity or source to be accessible, retrievable, extractable, citable and governable inside AI-mediated answer systems. It does not guarantee citation, ranking, recommendation, traffic, model compliance or interpretive fidelity.
This hub separates a visible market question from a stricter governance question. The market question is usually: “How do we get cited by AI systems?” The governance question is harder: “When a system cites us, does the cited source govern the right claim, inside the right scope, with the right proof status?”
Why citation is not the final objective
A citation is an observable signal. It shows that a system selected or displayed a source in relation to an answer. It does not prove that the answer is faithful, complete, current, proportional or governed by the strongest available source.
A cited answer can still fail in several ways:
- the source is used ornamentally while another source governs the claim;
- the cited passage does not support the generated statement;
- an old page is cited because it remains known to the system;
- the entity is named correctly but placed in the wrong category;
- the source is retrieved but the final synthesis exceeds its authority;
- the answer is useful locally but not defensible under source hierarchy.
The purpose of this hub is to keep those states separate before an audit turns them into one vague visibility score.
The five states to distinguish
| State | What it means | Why it is not enough |
|---|---|---|
| Present | The entity appears in an answer | Presence can coexist with distortion |
| Retrieved | A source is likely used or surfaced | Retrieval may stay invisible and uncited |
| Cited | A URL or source is displayed | Citation may be decorative or weak |
| Understood | The answer preserves the local meaning | Local meaning can still miss scope |
| Governed | The right source constrains the right claim | This is the standard required for fidelity |
AI citation readiness improves the first four states. Interpretive governance is needed to qualify the fifth.
The five layers of citation readiness
1. Accessibility
The useful page, passage and source path must be reachable. A page cannot be cited if the relevant surface is blocked, hidden, unstable, inaccessible to search systems, or impossible to parse without excessive inference.
Accessibility includes crawl conditions, rendering conditions, preview controls, canonical URL behavior and the visibility of the claim itself. It does not mean that every crawler must be allowed everywhere. It means that access policy and citation expectations must not contradict each other.
2. Retrievability
The source must be findable not only for the visible query, but also for adjacent questions generated by the system. AI-mediated answer systems often work through decomposition: they expand a user request into related subquestions, then search for sources that cover the required angles.
A source that ranks only for one head query may be weaker than a source that appears across the query cluster. This is where fan-out query behavior, semantic coverage and topic cluster consistency become operational.
3. Extractability
The source must contain passages that can be lifted without losing their meaning. Strong extractability depends on clear headings, stable sections, explicit claims, concise definitions, visible tables, current dates and paragraphs that do not rely too heavily on hidden context.
For a stricter definition, read extractability and self-contained passage.
4. Citability
A source becomes citable when it can support a claim clearly enough to be selected as evidence. Citability depends on precision, source support, claim boundaries, entity consistency and the absence of contradictions that make the source risky to reuse.
A citable source is not automatically the governing source. It may be useful, illustrative or derivative. The role of the citation must therefore be qualified.
5. Governability
Governability is the missing layer in most citation-factor discussions. It asks whether the source can legitimately constrain the answer. This requires source hierarchy, answer legitimacy, proof of fidelity and a visible distinction between canonical, derivative, market-facing and contextual sources.
SEO, machine-first structure and governance
| Layer | Question | Typical mechanism | Main risk |
|---|---|---|---|
| SEO visibility | Can the page be found? | ranking, indexation, links, topical coverage | visibility without fidelity |
| Machine-first structure | Can the useful passage be recovered? | headings, sections, tables, definitions, internal routes | extraction without scope |
| Entity consistency | Can the system identify the subject correctly? | stable names, category, relations, schema, links | category drift |
| Source governance | Can the source legitimately govern the claim? | canon, hierarchy, policies, proof surfaces | ornamental citation |
| Audit discipline | Can the observation be reconstructed? | prompt, system, date, source role, evidence | screenshot-only diagnosis |
The point is not to oppose SEO to governance. SEO remains the floor. Machine-first structure increases retrieval and extraction. Governance sets the limits of legitimate interpretation.
Practical reading path
Start with AI citation readiness to define the concept. Then read AI citation tracking to separate citation frequency from citation role. Use citability to qualify whether a source is structurally usable. Use source hierarchy and proof of fidelity to decide whether the answer is legitimate.
For applied diagnosis, use the AI citation readiness audit and the AI citation readiness checklist.
Comparative routing layer
Use SEO visibility, AI citability and interpretive fidelity when the question is whether a visibility problem, a citation problem or a fidelity problem is being confused. This comparative route is useful before selecting an audit path.
For technical access questions, read preview control, AI-ready structure, machine-first routing and retrieval without citation. For citation quality, use citation fidelity and citation role.
Operational extensions
This cluster now separates the main operational questions behind AI citation readiness:
| Question | Reading route |
|---|---|
| How do SEO visibility, citability and fidelity differ? | SEO visibility, AI citability and interpretive fidelity |
| Can the useful passage be reached and previewed? | Robots, AI crawlers and citation accessibility |
| Is another source replacing the canonical source? | Source substitution in AI answers |
| Is the citation actually strong? | How to audit AI citation quality |
| Does time affect the claim? | Freshness and AI citation stability |
| Do schema signals support or contradict the page? | Structured data and AI citations |
| Does language or geography change the source? | Language, geography and AI citations |
| Is authority being confused with legitimacy? | Domain authority vs source legitimacy |
| Are the core claims extractable as blocks? | AI-ready content blocks |
| Are snippets and preview rules aligned with source hierarchy? | Preview control and snippet governance |
Key definitions added to this route include citation fidelity, citation quality, citation stability, citation accessibility, source legitimacy, preview control, retrieval without citation and AI-ready content block.
Technical and operational routes added to this hub
Citation readiness now has three complementary routes.
| Route | Use it when |
|---|---|
| Robots, AI crawlers and citation accessibility | the question is whether useful sources can be reached, rendered, previewed or extracted |
| How to structure a page for AI citations without weakening governance | the question is how to create answer-ready passages without losing scope |
| AI citation tracking audit: what must actually be measured | the question is how to observe citations after answers have been produced |
These routes should not be merged. Accessibility is upstream, structure is editorial and architectural, tracking is observational. Governance decides whether the cited source legitimately carries the claim.
Extended operational routes
The citation-readiness cluster now separates six applied routes:
| Route | When to use it |
|---|---|
| SEO visibility, AI citability and interpretive fidelity | When visibility, citation and fidelity are being merged into one diagnosis |
| Robots, AI crawlers and citation accessibility | When crawler access, preview control or hidden passages may block citation readiness |
| AI citation tracking audit: what must actually be measured | When a citation dashboard counts URLs without classifying source role |
| Freshness and AI citation stability | When older sources, obsolete states or unstable citation roles influence current answers |
| Structured data and AI citations | When schema is being treated as if it could replace source hierarchy |
| Language, geography and AI citations | When bilingual or regional source selection changes the answer |
Use these routes after the hub. They turn the general question “How do we get cited?” into a more precise diagnosis: access, retrieval, extraction, support, role, freshness, language, and governance.
What this hub does not promise
This hub does not promise citation by ChatGPT, Google AI Overviews, Gemini, Perplexity, Bing, Claude or any other answer system. It does not promise ranking, traffic, recommendation, model compliance or future stability.
Its purpose is narrower and more useful: to make citation readiness observable, improve the structure of the source, and keep citation optimization subordinate to interpretive fidelity.
Second-layer resources
The cluster now includes operational pages for the most common failure modes after basic citation readiness:
- SEO visibility, AI citability and interpretive fidelity separates search presence, citable evidence and legitimate synthesis.
- How to structure a page for AI citations without weakening governance explains how to add extractable passages without turning every page into an authority claim.
- Robots, AI crawlers and citation accessibility connects access policy, preview control and citation expectations.
- AI citation tracking audit: what must actually be measured defines the difference between citation frequency and citation quality.
- Source substitution in AI answers explains when a weaker source replaces the source that should govern the claim.
- Freshness and AI citation stability separates currentness, versioning, obsolete states and stable doctrine.
- Structured data and AI citations explains why schema can support interpretation without replacing source hierarchy.
- Domain authority vs source legitimacy separates broad domain strength from claim-scoped authority.
For scoring and production work, use the AI citation audit scoring matrix and the fan-out query map.
Extended reading path for citation quality
After the core readiness layer, use SEO visibility, AI citability and interpretive fidelity to separate the three regimes. Then read AI citation vs fidelity, source substitution in AI answers and how to audit AI citation quality.
For technical causes, use robots, AI crawlers and citation accessibility and structured data and AI citations. For market and authority causes, use language, geography and AI citations and domain authority vs source legitimacy.