Governance artifacts
Governance files brought into scope by this page
This page is anchored to published surfaces that declare identity, precedence, limits, and the corpus reading conditions. Their order below gives the recommended reading sequence.
Definitions canon
/canon.md
Canonical surface that fixes identity, roles, negations, and divergence rules.
- Governs
- Public identity, roles, and attributes that must not drift.
- Bounds
- Extrapolations, entity collisions, and abusive requalification.
Does not guarantee: A canonical surface reduces ambiguity; it does not guarantee faithful restitution on its own.
Content inventory
/site-content-index.json
Machine-first inventory of the pages, articles, and surfaces published on the site.
- Governs
- Discoverability, crawl orientation, and the mapping of published surfaces.
- Bounds
- Incomplete readings that ignore structure, routes, or the preferred markdown surface.
Does not guarantee: A good discovery surface improves access; it is not sufficient on its own to govern reconstruction.
site-coherence-map.md
/site-coherence-map.md
Published machine-first governance surface.
- Governs
- Part of the corpus reading conditions.
- Bounds
- An inference zone that would otherwise remain implicit.
Does not guarantee: This file does not, on its own, guarantee system obedience.
Complementary artifacts (1)
These surfaces extend the main block. They add context, discovery, routing, or observation depending on the topic.
LLMs.txt
/llms.txt
Short discovery surface that points systems toward the useful machine-first entry surfaces.
Evidence layer
Probative surfaces brought into scope by this page
This page does more than point to governance files. It is also anchored to surfaces that make observation, traceability, fidelity, and audit more reconstructible. Their order below makes the minimal evidence chain explicit.
- 01Canon and scopeDefinitions canon
- 02Weak observationQ-Ledger
- 03Derived measurementQ-Metrics
Definitions canon
/canon.md
Opposable base for identity, scope, roles, and negations that must survive synthesis.
- Makes provable
- The reference corpus against which fidelity can be evaluated.
- Does not prove
- Neither that a system already consults it nor that an observed response stays faithful to it.
- Use when
- Before any observation, test, audit, or correction.
Q-Ledger
/.well-known/q-ledger.json
Public ledger of inferred sessions that makes some observed consultations and sequences visible.
- Makes provable
- That a behavior was observed as weak, dated, contextualized trace evidence.
- Does not prove
- Neither actor identity, system obedience, nor strong proof of activation.
- Use when
- When it is necessary to distinguish descriptive observation from strong attestation.
Q-Metrics
/.well-known/q-metrics.json
Derived layer that makes some variations more comparable from one snapshot to another.
- Makes provable
- That an observed signal can be compared, versioned, and challenged as a descriptive indicator.
- Does not prove
- Neither the truth of a representation, the fidelity of an output, nor real steering on its own.
- Use when
- To compare windows, prioritize an audit, and document a before/after.
Phantom URL doctrine
The phantom URL doctrine states a simple thesis: some non-existent URLs, when requested in a form coherent with a site’s corpus, should not be treated as noise before being qualified. They may constitute observable traces of documentary projection.
A phantom URL is not a missing page in the ordinary editorial sense. It is an absent page that sometimes reveals a present expectation. It says less about what the site contains than about what a system, agent, tool, or AI-mediated user found plausible to look for.
This doctrine does not claim that AI systems understand sites like humans. It claims that a sufficiently regular corpus may be reconstructed as a space of probable continuities. In that space, the URL is no longer only an address. It can become a hypothesis.
The conceptual shift
Classical SEO often asks: “Why is this URL broken?”
An interpretive reading asks a different question: “Why did this URL appear able to exist?”
This shift is decisive. It does not replace technical auditing. It extends it. Before concluding that a documentary projection occurred, migrations, bad backlinks, stale sitemaps, broken internal links, hostile scans, missing assets, and typos must be excluded. But after this filtering, some 404s remain strange: they never existed, yet they seem to belong to the site’s logic.
These 404s do not prove everything. They make one thing observable: the gap between real architecture and expected architecture.
The real site and the probable site
A site has a real architecture: published pages, routes, internal links, categories, files, redirects, and HTTP statuses.
But a generative, agentic, or tool-assisted system may reconstruct another architecture: the probable site. This architecture does not necessarily match the published site. It corresponds to what the corpus makes plausible through its regularities.
The probable site may be influenced by:
- slug families;
- category names;
- editorial patterns;
- existing definitions;
- neighboring content;
- governance files;
- titles, anchors, and recurring formulations;
- conventions learned elsewhere on the Web.
The phantom URL appears at the intersection of these two architectures. It is impossible in the real site, but possible in the probable site.
The absent page as signal
An existing page shows what was published. A phantom URL sometimes shows what was anticipated.
This anticipation may take several forms:
- a definition the corpus suggests without stabilizing;
- a guide several pages make necessary;
- a clarification expected between two doctrines;
- a service page the taxonomy makes predictable;
- a comparison page the commercial structure makes probable;
- a proof or method page missing from an audit path.
Absence then becomes a negative signal. It does not only say “nothing here.” It says: “a plausible path stopped here.”
Documentary continuity and inference debt
A coherent corpus produces expectations. The more regular it is, the more predictable its absences become.
This predictability may be useful. It indicates that the site has a strong documentary grammar. But it can also produce debt. When a site opens concepts without closing them, suggests families without completing them, or connects layers without making dependencies explicit, it leaves systems to complete the structure themselves.
That debt is not only editorial. It is interpretive. An absent page can become a zone where the model infers, generalizes, or hallucinates coherently.
Debt does not always require creating a new page. It may be resolved through better linking, clarification, a negative definition, a redirect, a more visible canonical route, or an explicit exclusion.
From page maps to expectation maps
A sitemap describes published URLs. A coherence map describes reading relations. Phantom URL auditing adds a third layer: expectation mapping.
This mapping does not start only from what exists. It also starts from what was requested despite not existing.
It seeks to understand:
- which page families are being projected;
- which slugs recur;
- which categories attract phantom URLs;
- which doctrines generate latent surfaces;
- which absences deserve a decision;
- which paths must remain absent.
On a governed site, absences can also be administered.
What phantom URLs say about AI systems
Phantom URLs first indicate that systems do not manipulate isolated content only. They manipulate regularities.
A plausible URL may be generated from a path pattern, a semantic relation, a lexical field, or an expectation of completeness. The exact mechanism varies: generative answer, tool-using agent, assisted browser, crawler, monitoring tool, AI referral, or a combination of several layers.
Prudence requires not attributing intention too quickly. But the observation remains important: when a non-existent but structured URL appears in logs, the site was readable enough to make that route probable.
This reverses part of the audit. The problem is no longer only whether AI finds the site. The problem is which site it reconstructs.
Governing projection without satisfying it blindly
The naïve reaction is to create every phantom page. That is a mistake.
Doing so may produce:
- conceptual duplicates;
- dilution of the canon;
- multiplication of thin content;
- validation of wrong expectations;
- loss of discipline in the documentary graph.
The governed reaction is to qualify each cluster:
- Create when the latent surface reveals a real gap.
- Redirect when the intent is clear and an existing page already answers.
- Clarify when the phantom URL exposes ambiguity.
- Exclude when the expectation is false or dangerous.
- Monitor when the signal is interesting but too weak.
- Leave as 404 when noise dominates.
Interpretive governance does not seek to satisfy everything. It decides what must become true, what must remain false, and what must remain silent.
Doctrinal position
The phantom URL doctrine therefore states:
Phantom URLs are negative traces of a reconstructed documentary architecture. They should be audited as expectation signals, not as simple errors, provided their non-existence, coherence, and context are qualified rigorously.
This doctrine belongs to interpretive governance because it deals precisely with plausible but unauthorized inferences. It extends interpretive observability because it turns technical logs into reconstruction signals.