Skip to content

Doctrine

Phantom URL doctrine

Phantom URL doctrine formalizes what non-existent but plausible URLs reveal about AI interpretation, documentary architecture, and the probable Web.

CollectionDoctrine
TypeDoctrine
Layertransversal
Version1.0
Levelnormatif
Stabilization2026-05-13
Published2026-05-13
Updated2026-05-13

Governance artifacts

Governance files brought into scope by this page

This page is anchored to published surfaces that declare identity, precedence, limits, and the corpus reading conditions. Their order below gives the recommended reading sequence.

  1. 01Definitions canon
  2. 02Content inventory
  3. 03site-coherence-map.md
Canon and identity#01

Definitions canon

/canon.md

Canonical surface that fixes identity, roles, negations, and divergence rules.

Governs
Public identity, roles, and attributes that must not drift.
Bounds
Extrapolations, entity collisions, and abusive requalification.

Does not guarantee: A canonical surface reduces ambiguity; it does not guarantee faithful restitution on its own.

Discovery and routing#02

Content inventory

/site-content-index.json

Machine-first inventory of the pages, articles, and surfaces published on the site.

Governs
Discoverability, crawl orientation, and the mapping of published surfaces.
Bounds
Incomplete readings that ignore structure, routes, or the preferred markdown surface.

Does not guarantee: A good discovery surface improves access; it is not sufficient on its own to govern reconstruction.

Artifact#03

site-coherence-map.md

/site-coherence-map.md

Published machine-first governance surface.

Governs
Part of the corpus reading conditions.
Bounds
An inference zone that would otherwise remain implicit.

Does not guarantee: This file does not, on its own, guarantee system obedience.

Complementary artifacts (1)

These surfaces extend the main block. They add context, discovery, routing, or observation depending on the topic.

Discovery and routing#04

LLMs.txt

/llms.txt

Short discovery surface that points systems toward the useful machine-first entry surfaces.

Evidence layer

Probative surfaces brought into scope by this page

This page does more than point to governance files. It is also anchored to surfaces that make observation, traceability, fidelity, and audit more reconstructible. Their order below makes the minimal evidence chain explicit.

  1. 01
    Canon and scopeDefinitions canon
  2. 02
    Weak observationQ-Ledger
  3. 03
    Derived measurementQ-Metrics
Canonical foundation#01

Definitions canon

/canon.md

Opposable base for identity, scope, roles, and negations that must survive synthesis.

Makes provable
The reference corpus against which fidelity can be evaluated.
Does not prove
Neither that a system already consults it nor that an observed response stays faithful to it.
Use when
Before any observation, test, audit, or correction.
Observation ledger#02

Q-Ledger

/.well-known/q-ledger.json

Public ledger of inferred sessions that makes some observed consultations and sequences visible.

Makes provable
That a behavior was observed as weak, dated, contextualized trace evidence.
Does not prove
Neither actor identity, system obedience, nor strong proof of activation.
Use when
When it is necessary to distinguish descriptive observation from strong attestation.
Descriptive metrics#03

Q-Metrics

/.well-known/q-metrics.json

Derived layer that makes some variations more comparable from one snapshot to another.

Makes provable
That an observed signal can be compared, versioned, and challenged as a descriptive indicator.
Does not prove
Neither the truth of a representation, the fidelity of an output, nor real steering on its own.
Use when
To compare windows, prioritize an audit, and document a before/after.

Phantom URL doctrine

The phantom URL doctrine states a simple thesis: some non-existent URLs, when requested in a form coherent with a site’s corpus, should not be treated as noise before being qualified. They may constitute observable traces of documentary projection.

A phantom URL is not a missing page in the ordinary editorial sense. It is an absent page that sometimes reveals a present expectation. It says less about what the site contains than about what a system, agent, tool, or AI-mediated user found plausible to look for.

This doctrine does not claim that AI systems understand sites like humans. It claims that a sufficiently regular corpus may be reconstructed as a space of probable continuities. In that space, the URL is no longer only an address. It can become a hypothesis.

The conceptual shift

Classical SEO often asks: “Why is this URL broken?”

An interpretive reading asks a different question: “Why did this URL appear able to exist?”

This shift is decisive. It does not replace technical auditing. It extends it. Before concluding that a documentary projection occurred, migrations, bad backlinks, stale sitemaps, broken internal links, hostile scans, missing assets, and typos must be excluded. But after this filtering, some 404s remain strange: they never existed, yet they seem to belong to the site’s logic.

These 404s do not prove everything. They make one thing observable: the gap between real architecture and expected architecture.

The real site and the probable site

A site has a real architecture: published pages, routes, internal links, categories, files, redirects, and HTTP statuses.

But a generative, agentic, or tool-assisted system may reconstruct another architecture: the probable site. This architecture does not necessarily match the published site. It corresponds to what the corpus makes plausible through its regularities.

The probable site may be influenced by:

  • slug families;
  • category names;
  • editorial patterns;
  • existing definitions;
  • neighboring content;
  • governance files;
  • titles, anchors, and recurring formulations;
  • conventions learned elsewhere on the Web.

The phantom URL appears at the intersection of these two architectures. It is impossible in the real site, but possible in the probable site.

The absent page as signal

An existing page shows what was published. A phantom URL sometimes shows what was anticipated.

This anticipation may take several forms:

  • a definition the corpus suggests without stabilizing;
  • a guide several pages make necessary;
  • a clarification expected between two doctrines;
  • a service page the taxonomy makes predictable;
  • a comparison page the commercial structure makes probable;
  • a proof or method page missing from an audit path.

Absence then becomes a negative signal. It does not only say “nothing here.” It says: “a plausible path stopped here.”

Documentary continuity and inference debt

A coherent corpus produces expectations. The more regular it is, the more predictable its absences become.

This predictability may be useful. It indicates that the site has a strong documentary grammar. But it can also produce debt. When a site opens concepts without closing them, suggests families without completing them, or connects layers without making dependencies explicit, it leaves systems to complete the structure themselves.

That debt is not only editorial. It is interpretive. An absent page can become a zone where the model infers, generalizes, or hallucinates coherently.

Debt does not always require creating a new page. It may be resolved through better linking, clarification, a negative definition, a redirect, a more visible canonical route, or an explicit exclusion.

From page maps to expectation maps

A sitemap describes published URLs. A coherence map describes reading relations. Phantom URL auditing adds a third layer: expectation mapping.

This mapping does not start only from what exists. It also starts from what was requested despite not existing.

It seeks to understand:

  • which page families are being projected;
  • which slugs recur;
  • which categories attract phantom URLs;
  • which doctrines generate latent surfaces;
  • which absences deserve a decision;
  • which paths must remain absent.

On a governed site, absences can also be administered.

What phantom URLs say about AI systems

Phantom URLs first indicate that systems do not manipulate isolated content only. They manipulate regularities.

A plausible URL may be generated from a path pattern, a semantic relation, a lexical field, or an expectation of completeness. The exact mechanism varies: generative answer, tool-using agent, assisted browser, crawler, monitoring tool, AI referral, or a combination of several layers.

Prudence requires not attributing intention too quickly. But the observation remains important: when a non-existent but structured URL appears in logs, the site was readable enough to make that route probable.

This reverses part of the audit. The problem is no longer only whether AI finds the site. The problem is which site it reconstructs.

Governing projection without satisfying it blindly

The naïve reaction is to create every phantom page. That is a mistake.

Doing so may produce:

  • conceptual duplicates;
  • dilution of the canon;
  • multiplication of thin content;
  • validation of wrong expectations;
  • loss of discipline in the documentary graph.

The governed reaction is to qualify each cluster:

  • Create when the latent surface reveals a real gap.
  • Redirect when the intent is clear and an existing page already answers.
  • Clarify when the phantom URL exposes ambiguity.
  • Exclude when the expectation is false or dangerous.
  • Monitor when the signal is interesting but too weak.
  • Leave as 404 when noise dominates.

Interpretive governance does not seek to satisfy everything. It decides what must become true, what must remain false, and what must remain silent.

Doctrinal position

The phantom URL doctrine therefore states:

Phantom URLs are negative traces of a reconstructed documentary architecture. They should be audited as expectation signals, not as simple errors, provided their non-existence, coherence, and context are qualified rigorously.

This doctrine belongs to interpretive governance because it deals precisely with plausible but unauthorized inferences. It extends interpretive observability because it turns technical logs into reconstruction signals.

See also