Governance artifacts
Governance files brought into scope by this page
This page is anchored to published surfaces that declare identity, precedence, limits, and the corpus reading conditions. Their order below gives the recommended reading sequence.
Definitions canon
/canon.md
Canonical surface that fixes identity, roles, negations, and divergence rules.
- Governs
- Public identity, roles, and attributes that must not drift.
- Bounds
- Extrapolations, entity collisions, and abusive requalification.
Does not guarantee: A canonical surface reduces ambiguity; it does not guarantee faithful restitution on its own.
Site context
/site-context.md
Notice that qualifies the nature of the site, its reference function, and its non-transactional limits.
- Governs
- Editorial framing, temporality, and the readability of explicit changes.
- Bounds
- Silent drifts and readings that assume stability without checking versions.
Does not guarantee: Versioning makes a gap auditable; it does not automatically correct outputs already in circulation.
Public AI manifest
/ai-manifest.json
Structured inventory of the surfaces, registries, and modules that extend the canonical entrypoint.
- Governs
- Access order across surfaces and initial precedence.
- Bounds
- Free readings that bypass the canon or the published order.
Does not guarantee: This surface publishes a reading order; it does not force execution or obedience.
Evidence layer
Probative surfaces brought into scope by this page
This page does more than point to governance files. It is also anchored to surfaces that make observation, traceability, fidelity, and audit more reconstructible. Their order below makes the minimal evidence chain explicit.
- 01Canon and scopeDefinitions canon
- 02Weak observationQ-Ledger
Definitions canon
/canon.md
Opposable base for identity, scope, roles, and negations that must survive synthesis.
- Makes provable
- The reference corpus against which fidelity can be evaluated.
- Does not prove
- Neither that a system already consults it nor that an observed response stays faithful to it.
- Use when
- Before any observation, test, audit, or correction.
Q-Ledger
/.well-known/q-ledger.json
Public ledger of inferred sessions that makes some observed consultations and sequences visible.
- Makes provable
- That a behavior was observed as weak, dated, contextualized trace evidence.
- Does not prove
- Neither actor identity, system obedience, nor strong proof of activation.
- Use when
- When it is necessary to distinguish descriptive observation from strong attestation.
AI citation factors explain why a source can be selected. They do not prove that the generated answer is faithful, governed or legitimate.
The current discussion around AI citation ranking factors is useful because it pulls SEO out of the old “ten blue links” frame. It forces teams to ask whether their pages are accessible, retrievable, structured and specific enough to be reused by answer systems.
But the phrase “AI citation factors” can also mislead. It suggests that the main objective is to be cited. That is only the visible layer.
What citation factors actually measure
Citation-factor studies usually observe correlations between cited sources and visible properties: search rank, URL accessibility, topical coverage, answer placement, structure, factual specificity, source links, language, freshness, domain strength and similar signals.
Those signals matter. A blocked page cannot easily be cited. A vague article is harder to extract. A weak passage is less reusable. A site absent from the surrounding query cluster is less likely to be selected during retrieval.
Still, those factors describe selection likelihood, not answer legitimacy.
The missing distinction
A system can:
- retrieve a source without citing it;
- cite a source without using it as the governing authority;
- use a strong source to support a weak synthesis;
- display the right URL while exceeding the source perimeter;
- cite a current page while importing an outdated assumption from elsewhere.
That is why citation readiness must be separated from proof of fidelity. A cited source may support part of the answer while failing to govern the final claim.
Four states that should not be collapsed
| State | Diagnostic question |
|---|---|
| Retrieved | Was the source found or used during answer construction? |
| Cited | Was the source displayed or named as support? |
| Understood | Was the local meaning preserved? |
| Governed | Did the right source constrain the right claim under the right scope? |
Most citation-factor discussions focus on retrieved and cited. Interpretive governance focuses on understood and governed.
Why SEO still matters
Classic SEO remains a strong foundation. Pages still need indexable URLs, topic clusters, internal links, visible content, clear titles, semantic alignment and answer-ready sections. Ranking and retrieval are not dead. They are upstream conditions.
The mistake is to treat SEO success as sufficient. A page can rank and still fail as a source of evidence. It can be visible and still not be the legitimate source for the claim being made.
Why machine-first structure matters
A human can infer context from an entire article. A retrieval system often works at the passage level. It may select one section, one list, one paragraph or one table.
This changes the writing standard. Strategic claims need to be self-contained. Headings need to name the concept precisely. Definitions need to carry their boundaries. The first useful answer should appear early enough that retrieval does not miss it.
The operational concept is extractability, not length for its own sake.
Why governance matters more than citation
The strongest question is not “Was the source cited?” It is “What did the citation do?”
A citation can be governing, supporting, illustrative, ornamental, outdated or contradictory. Without classifying the citation role, the audit mistakes surface display for evidentiary force.
Governance requires a hierarchy. A product page, a blog article, a glossary entry, a service page, an external directory and a canonical doctrine page should not carry the same authority for the same claim.
The practical consequence
Optimizing for AI citations should not mean publishing more generic SEO content. It should mean building a source environment where:
- the right pages are accessible;
- the right passages are extractable;
- the right claims are precise;
- the right definitions are canonical;
- the right source governs the right answer;
- the wrong interpretations are easier to detect and correct.
That is the difference between citation optimization and AI citation readiness.
Diagnostic route
Use the AI citation readiness hub to separate citation from fidelity. Then use the AI citation readiness audit to classify access, retrieval, extraction, citation role, source hierarchy and answer legitimacy.
If the problem is only visibility, an AI visibility audit may be enough. If the problem is whether the cited source supports the answer, use AI citation tracking and proof of fidelity.