Engagement decision
How to recognize that this axis should be mobilized
Use this page as a decision page. The objective is not only to understand the concept, but to identify the symptoms, framing errors, use cases, and surfaces to open in order to correct the right problem.
Typical symptoms
- Organic or generative presence improves, but the scope remains badly understood.
- Systems attribute roles, services, or capabilities that were never declared.
- Representation changes with the prompt, language, or engine.
- GEO dashboards look healthy while fidelity remains weak.
Frequent framing errors
- Reducing interpretive SEO to a ranking strategy.
- Treating citation as proof of understanding.
- Working on visibility before canon, relations, and limits.
- Measuring downstream effects without publishing their upstream conditions.
Use cases
- Requalifying a site that is visible but poorly understood.
- Stabilizing the reconstruction of a brand, a method, or an offering.
- Connecting SEO, governance, entities, evidence, and observations.
- Arbitrating between gains in presence and gains in fidelity.
What gets corrected concretely
- Clarification of canon, relations, and exclusions.
- Realignment between visible surfaces and authority surfaces.
- Implementation of an observability and fidelity-proof protocol.
- Reduction of scope drift and default extrapolations.
Relevant machine-first artifacts
These surfaces bound the problem before detailed correction begins.
Governance files to open first
Useful evidence surfaces
These surfaces connect diagnosis, observation, fidelity, and audit.
References to open first
Governance artifacts
Governance files brought into scope by this page
This page is anchored to published surfaces that declare identity, precedence, limits, and the corpus reading conditions. Their order below gives the recommended reading sequence.
Canonical AI entrypoint
/.well-known/ai-governance.json
Neutral entrypoint that declares the governance map, precedence chain, and the surfaces to read first.
- Governs
- Access order across surfaces and initial precedence.
- Bounds
- Free readings that bypass the canon or the published order.
Does not guarantee: This surface publishes a reading order; it does not force execution or obedience.
Public AI manifest
/ai-manifest.json
Structured inventory of the surfaces, registries, and modules that extend the canonical entrypoint.
- Governs
- Access order across surfaces and initial precedence.
- Bounds
- Free readings that bypass the canon or the published order.
Does not guarantee: This surface publishes a reading order; it does not force execution or obedience.
LLMs.txt
/llms.txt
Short discovery surface that points systems toward the useful machine-first entry surfaces.
- Governs
- Discoverability, crawl orientation, and the mapping of published surfaces.
- Bounds
- Incomplete readings that ignore structure, routes, or the preferred markdown surface.
Does not guarantee: A good discovery surface improves access; it is not sufficient on its own to govern reconstruction.
Complementary artifacts (2)
These surfaces extend the main block. They add context, discovery, routing, or observation depending on the topic.
LLMs-full.txt
/llms-full.txt
Extended discovery surface for readers that consume richer context.
Robots.txt
/robots.txt
Crawl surface that improves discovery but does not, on its own, publish reading conditions.
Evidence layer
Probative surfaces brought into scope by this page
This page does more than point to governance files. It is also anchored to surfaces that make observation, traceability, fidelity, and audit more reconstructible. Their order below makes the minimal evidence chain explicit.
- 01Canon and scopeDefinitions canon
- 02Observation mapObservatory map
- 03Weak observationQ-Ledger
- 04Derived measurementQ-Metrics
Definitions canon
/canon.md
Opposable base for identity, scope, roles, and negations that must survive synthesis.
- Makes provable
- The reference corpus against which fidelity can be evaluated.
- Does not prove
- Neither that a system already consults it nor that an observed response stays faithful to it.
- Use when
- Before any observation, test, audit, or correction.
Observatory map
/observations/observatory-map.json
Machine-first index of published observation resources, snapshots, and comparison points.
- Makes provable
- Where the observation objects used in an evidence chain are located.
- Does not prove
- Neither the quality of a result nor the fidelity of a particular response.
- Use when
- To locate baselines, ledgers, snapshots, and derived artifacts.
Q-Ledger
/.well-known/q-ledger.json
Public ledger of inferred sessions that makes some observed consultations and sequences visible.
- Makes provable
- That a behavior was observed as weak, dated, contextualized trace evidence.
- Does not prove
- Neither actor identity, system obedience, nor strong proof of activation.
- Use when
- When it is necessary to distinguish descriptive observation from strong attestation.
Q-Metrics
/.well-known/q-metrics.json
Derived layer that makes some variations more comparable from one snapshot to another.
- Makes provable
- That an observed signal can be compared, versioned, and challenged as a descriptive indicator.
- Does not prove
- Neither the truth of a representation, the fidelity of an output, nor real steering on its own.
- Use when
- To compare windows, prioritize an audit, and document a before/after.
Complementary probative surfaces (1)
These artifacts extend the main chain. They help qualify an audit, an evidence level, a citation, or a version trajectory.
Citations
/citations.md
Minimal external reference surface used to contextualize some concepts without delegating canonical authority to them.
Interpretive SEO
This expertise axis aims to stabilize machine understanding beyond indexing: interpretation, attribution, semantic coherence, reconstruction fidelity, and perimeter drift prevention.
Interpretive SEO differs from a ranking logic: it focuses on what systems infer from a site, an entity, and a corpus, and on the stability of those inferences over time.
This axis is defined by Interpretive SEO and relies on Interpretive governance.
Problem
Content can be complete, credible, and well written while still producing unstable inferences: erroneous attribution, role confusion, association drift, abusive perimeter extension, or plausible but unfaithful synthesis.
The problem emerges when the meaning space remains too open: weak definitions, implicit relations, non-hierarchized authorities, missing exclusions, weak machine-first architecture, or absent governance files.
When this axis becomes critical
Interpretive SEO becomes a priority when:
- engines or assistants cite the site but distort its perimeter;
- a brand is visible yet poorly understood;
- generative responses attribute undeclared services, responsibilities, or capabilities;
- representation varies strongly from one formulation to another;
- presence metrics look acceptable while fidelity remains weak.
In those contexts, visibility, representation, stability, and governability must be read together, as explained in GEO metrics do not govern representation.
Typical consequences
- Divergent responses depending on engines, assistants, and formulations.
- Erroneous inferences about services, capabilities, or intervention perimeters.
- Identity dilution through repeated implicit associations.
- Unstable attributions of citations, projects, concepts, or responsibilities.
- Loss of control over what is treated as “central” or “true”.
What is corrected first
Interpretive SEO generally acts on five layers:
1. The canon
Clearly define concepts, roles, limits, and exclusions. See Definitions and the Machine-first canon.
2. Relations
Make explicit the links between person, organization, method, offer, document, project, and authority territory.
3. Machine-first surfaces
Strengthen the published architecture, entry points, and governance files. See the Machine-first visibility doctrine and Machine-first is not enough: why governance files change the reading regime.
4. Proof of fidelity
Compare what the canon states with what outputs reconstruct. See Proof of fidelity and Proof of fidelity: why citation is no longer enough.
5. Observability
Measure reconstruction stability, activation of governed surfaces, and the canon-output gap. See Q-Metrics and Interpretive observability: minimum metrics to log.
Conceptual levers
- Normative definitions: canonical registry of concepts used.
- Interpretive governance: bounding, hierarchies, negations, canonical referrals.
- Entities and relations: coherence between identifiers, pages, graphs, and mentions.
- Controlled redundancy: inter-surface stability without divergence.
- SSA-E + A2 + Dual Web architecture: machine-first implementation standard.
How to validate a correction
A correction is not validated because a dashboard goes up. It is validated when several signals converge:
- restitution remains faithful across models;
- entity representation becomes more stable under prompt variation;
- critical attributes stop being extended by default;
- citations become more coherent with the canon;
- drift remains auditable and versionable.
This logic is developed in Canon-output gap: measuring distortion instead of debating the “true” and GEO metrics see the effect, not the conditions.
Canonical references
Related reading
- Making governance measurable: Q-Metrics
- GEO metrics do not govern representation
- GEO metrics see the effect, not the conditions
Back to the map: Expertise.
Bridge vocabulary: LLM visibility, citability, and recommendability
Many teams reach this axis through the phrase LLM visibility. That phrase is useful, but too broad on its own.
Interpretive SEO works on the thresholds beneath it:
- presence versus absence;
- citability versus contradiction;
- recommendability versus weak comparability;
- stable representation versus prompt-dependent drift.
For the distinction, see LLM visibility vs citability vs recommendability.
Drift detection becomes meaningful here only under canon
Many teams reach this axis through Drift detection.
On this site, that label becomes meaningful only when drift is read against the canon-output gap, proof of fidelity, and interpretive observability. Otherwise the signal remains descriptive but weakly governable.