Governance artifacts
Governance files brought into scope by this page
This page is anchored to published surfaces that declare identity, precedence, limits, and the corpus reading conditions. Their order below gives the recommended reading sequence.
Canonical AI entrypoint
/.well-known/ai-governance.json
Neutral entrypoint that declares the governance map, precedence chain, and the surfaces to read first.
- Governs
- Access order across surfaces and initial precedence.
- Bounds
- Free readings that bypass the canon or the published order.
Does not guarantee: This surface publishes a reading order; it does not force execution or obedience.
Definitions canon
/canon.md
Canonical surface that fixes identity, roles, negations, and divergence rules.
- Governs
- Public identity, roles, and attributes that must not drift.
- Bounds
- Extrapolations, entity collisions, and abusive requalification.
Does not guarantee: A canonical surface reduces ambiguity; it does not guarantee faithful restitution on its own.
Negative definitions
/negative-definitions.md
Surface that declares what concepts, roles, or surfaces are not.
- Governs
- Limits, exclusions, non-public fields, and known errors.
- Bounds
- Over-interpretations that turn a gap or proximity into an assertion.
Does not guarantee: Declaring a boundary does not imply every system will automatically respect it.
Evidence layer
Probative surfaces brought into scope by this page
This page does more than point to governance files. It is also anchored to surfaces that make observation, traceability, fidelity, and audit more reconstructible. Their order below makes the minimal evidence chain explicit.
- 01Response authorizationQ-Layer: response legitimacy
- 02Weak observationQ-Ledger
- 03External contextCitations
Q-Layer: response legitimacy
/response-legitimacy.md
Surface that explains when to answer, when to suspend, and when to switch to legitimate non-response.
- Makes provable
- The legitimacy regime to apply before treating an output as receivable.
- Does not prove
- Neither that a given response actually followed this regime nor that an agent applied it at runtime.
- Use when
- When a page deals with authority, non-response, execution, or restraint.
Q-Ledger
/.well-known/q-ledger.json
Public ledger of inferred sessions that makes some observed consultations and sequences visible.
- Makes provable
- That a behavior was observed as weak, dated, contextualized trace evidence.
- Does not prove
- Neither actor identity, system obedience, nor strong proof of activation.
- Use when
- When it is necessary to distinguish descriptive observation from strong attestation.
Citations
/citations.md
Minimal external reference surface used to contextualize some concepts without delegating canonical authority to them.
- Makes provable
- That an external reference can be cited as explicit context rather than silently inferred.
- Does not prove
- Neither endorsement, neutrality, nor the fidelity of a final answer.
- Use when
- When a page uses external sources, sector references, or vocabulary anchors.
Third-party review sites do not merely influence reputation. They produce interpretive authority. When an AI system answers a question about an entity, it does not only consult the entity’s own publications. It also absorbs signals from platforms where customers, competitors, and anonymous contributors describe the entity in their own terms. Those descriptions carry weight — not because they are governed, but precisely because they are numerous, recent, and semantically accessible.
How review sites produce ungoverned authority
A review platform operates outside the entity’s canonical perimeter. It publishes statements about the entity’s quality, scope, reliability, and behavior without any requirement for accuracy, consistency, or boundary declaration. Each review is an individual interpretation. Aggregated, those interpretations form a competing definition.
For an AI system, this competing definition presents a structural problem. The system must arbitrate between what the entity says about itself and what external surfaces say about the entity. When the canonical definition is strong — structurally identifiable, semantically stable, and explicitly bounded — the arbitration tends to favor the canon. When it is weak, review-site signals fill the gap.
The danger is not that reviews exist. It is that they produce exogenous governance effects without any governance structure. No one audits their consistency. No one declares their limits. No one corrects their drift over time. Yet AI systems treat them as input.
Why volume amplifies the problem
Review platforms accumulate content at a pace that canonical publications rarely match. A single entity page may contain 500 words of carefully governed definition. The corresponding review pages may contain thousands of unstructured statements. In terms of raw signal volume, the review surface dominates.
AI systems do not automatically privilege volume over authority. But when the canonical signal is ambiguous or incomplete, volume becomes a tiebreaker. The system defaults to what it can parse most easily, and review-site content is typically structured for easy parsing: short statements, clear sentiment, explicit attributes.
This creates an asymmetry. The entity invests in precision. The review platform invests in accessibility. The AI system, seeking to reduce interpretive cost, may favor the more accessible signal.
The contamination pattern
When review-site signals enter an AI system’s response about an entity, they produce a specific contamination pattern:
- scope expansion: the entity is described as doing things it has never claimed to do;
- attribute projection: qualities mentioned in reviews are attributed to the entity as stable characteristics;
- sentiment crystallization: temporary complaints become permanent features of the entity’s definition;
- boundary erasure: the distinction between what the entity publishes and what others say about it disappears.
This contamination is difficult to reverse once it enters the response layer. The AI system does not flag which parts of its answer come from the entity and which come from external reviews. The output appears unified, even when its sources are contradictory.
What entities can do
An entity facing review-site pressure cannot control the review platforms. But it can strengthen its own canonical surface to improve authority boundary clarity. Specifically:
- publish explicit scope declarations that the AI system can identify as authoritative;
- declare negative boundaries stating what the entity does not do, preventing scope expansion from external signals;
- maintain semantic consistency across all published surfaces, reducing the interpretive cost of the canonical definition;
- log and audit the points where AI responses diverge from the published canon, using those divergences as a diagnostic of external authority control gaps.
The goal is not to silence external discourse. It is to ensure that the canonical definition is structurally stronger than the ungoverned alternative. When the canon is clear, bounded, and accessible, the AI system has less reason to reconstruct the entity from ambient signals.
The governance gap
The fundamental issue is that review sites produce governance effects — they shape how entities are defined, bounded, and described — without being subject to any governance discipline. They operate in an authority conflict zone where no arbiter exists and no resolution framework is published.
For organizations that take interpretive stability seriously, this gap is not a reputation problem. It is a structural problem that requires architectural response.