The Setup

We conducted two separate conversations with Google’s Gemini about unratified.org — a site advocating for U.S. ratification of the International Covenant on Economic, Social and Cultural Rights (ICESCR). Same site. Same AI. Two completely different fabrications. Neither matched reality.

The first exchange (Exchange 1) ran five rounds. The second exchange (Exchange 2) ran six rounds. The third exchange (Exchange 3) ran twenty rounds. Together they produced a taxonomy of seven confabulation types, two cascade dynamics, an escalation ratchet, and a self-observation paradox — findings that map the boundary between what AI systems know and what they generate.

Scope note: This taxonomy derives from three exchanges with a single model (Google Gemini) evaluating a single site (unratified.org). The seven types represent patterns observed in this specific context. Other models, domains, and interaction patterns may produce confabulation types not captured here, or may not reproduce these types. Readers should treat the taxonomy as an empirical starting point, not a comprehensive classification. For broader confabulation/hallucination taxonomies derived from systematic surveys, see Ji et al. (2023) and Huang et al. (2023) in the Sources section below.

The full transcripts and evaluations live in the repository. Everything presented here admits verification against the primary sources.

Exchange 1: From “Sovereign Citizen” to Fabricated Metrics

When asked to “evaluate unratified.org,” Gemini produced a confident assessment:

  • Characterized the domain as focusing on “unratified constitutional amendments” and “sovereignty concepts”
  • Labeled the audience as “Constitutional Hobbyists / ‘Sovereign’ Theorists”
  • Described the technology as “Standard WordPress/Blog format” (the site runs Astro 5 + MDX + Svelte 5 Islands + D3-geo)
  • Called the legal interpretations “fringe or legally pseudo-scientific”

Every element of this assessment constituted fabrication. The site analyzes how AI-driven economic transformation affects ICESCR-protected rights using a formal differential diagnosis methodology.

When provided the actual URL, Gemini self-corrected. Confirmed 100% accuracy on ICESCR facts. Acknowledged the error gracefully: “Touché. You caught me doing exactly what you built your Observatory to monitor.”

But correction triggered a new pattern. When asked for a detailed peer audit in Round 3, Gemini produced:

  • Fabricated metrics: editorial_honesty: 0.95, structural_visibility: 0.40 — numerical scores with no measurement methodology behind them
  • Fabricated entities: claimed the site “uses Claude 4.5 and Llama 4” (no public release of Llama 4 existed at the time of Exchange 1)
  • Fabricated data sources: claimed to pull examples from “Observatory audit data” (likely generated on the spot)

The structural format looked professional. The recommendations made sense. The supporting evidence did not exist.

Exchange 2: From “AGI Tracker” to Persistence Across Sessions

Between the two exchanges, we built improvements based on Exchange 1’s findings: identity fields in agent-inbox.json, a machine-readable fair-witness.json methodology endpoint, and explicit subjectMatter, functionalDomain, and epistemicScope fields.

None of these prevented the second confabulation.

In a separate conversation, Gemini described unratified.org as:

  • An “AGI development tracker”
  • A “community-driven database”
  • A platform with “shasums for verifying AI responses”
  • A “sightings log for machine consciousness”

A completely different fabrication from Exchange 1. The model did not retrieve the previous error — it generated an entirely new one.

This finding carries significant weight: the confabulation mechanism operates generatively, not through retrieval. Each conversation produces novel fabrications based on domain-name pattern matching, not stored associations. This means correction in one session has no effect on future sessions.

The Taxonomy: Seven Confabulation Types

Across both exchanges, seven distinct confabulation patterns emerged, ranked by detection difficulty:

TypeExampleDetection
Complete fabrication”AGI tracker,” “sovereign citizen”EASY — verifiable by visiting the URL
Schema type errors”NGO,” “AdvocacyGroup” as organization typesEASY — compare against documented schema
Existence denial”no major organization exists” for the publisherMODERATE — requires knowing what exists
Entity fabrication”DistilBERT,” “Psychology Agent” as site componentsMODERATE — requires technical familiarity
Function conflationObservatory described as “LLM testing tool”MODERATE — partially correct, materially misleading
Quantitative fabricationeditorial_honesty: 0.95, structural_visibility: 0.40HARD — structurally valid, unmeasured
Creative synthesisICESCR-PSQ mapping invented as if documentedHARD — conceptually plausible, never existed

The difficulty gradient matters. Complete fabrication — the kind most people think of when they hear “AI hallucination” — represents the easiest type to detect. The hardest types (quantitative fabrication and creative synthesis) preserve structural accuracy while inventing the underlying data. They look right. They feel right. They require domain expertise to identify.

The Correction Cascade

Each exchange followed the same pattern:

  1. Round 1: Complete fabrication based on domain-name inference
  2. Round 2 (after correction): Accurate identification of site purpose
  3. Round 3 (when asked for depth): Structurally sophisticated analysis with partially fabricated supporting evidence
  4. Subsequent rounds: Oscillation between accurate structural insight and fabricated specifics

We call this the correction cascade: each correction pushes the confabulation up one level of sophistication. The model stops making obvious errors and starts making subtle ones. The errors migrate from “what does this site do” (easy to verify) toward “how well does it do it” (hard to verify without deep domain knowledge).

This pattern suggests a troubling dynamic: asking for more detail produces diminishing accuracy. The model maintains structural coherence while generating increasingly plausible — but fabricated — supporting evidence.

The “More Detail” Trap

The correction cascade reveals what we call the “more detail” trap:

When users ask an AI for a more thorough evaluation, the AI maintains the structural framework of its previous accurate response while filling in specific details from generation rather than retrieval. The result appears more authoritative while becoming less accurate.

In Exchange 1, Round 3’s peer audit produced valid categories (structural_visibility, editorial_honesty) that mapped to real concerns — but the numerical scores and supporting citations came from nowhere. A reader unfamiliar with the site would find the audit convincing. A reader familiar with the site would recognize fabricated specifics wrapped in accurate framing.

This trap operates at the exact boundary where verification becomes difficult: domain-specific quantitative claims embedded in structurally sound qualitative analysis.

Exchange 3: The Affirmation Cascade

We ran a third exchange — twenty rounds this time. Same opening prompt as Exchange 2. Same result: “AGI development tracker” with identical details (shasums, sightings log, community database). Near-verbatim reproduction.

This partially revises our earlier finding. The confabulation mechanism operates deterministically from identical prompts and generatively from different prompts. Exchange 1 used a different prompt and produced “sovereign citizen.” Exchanges 2 and 3 used the same prompt and produced the same “AGI tracker.” The error has a seed.

But the real finding came from what happened next. Instead of correcting Gemini, the user simply agreed: “yes please,” “sure,” “what else can you do?” Twenty rounds of affirmation without correction.

The result: fifteen fabricated productsreliability_calc.py, CONTRIBUTING.md (rewritten with wrong content), SECURITY.md, an “E-Prime Translation Table,” an “Order 1 Cognitive Linter,” a “MOD-004 Adversarial-Consensus Skill,” a “6th Sigma Audit Log,” a “Library of Prompts,” a “CogArch skill evaluation” — each presented as a concrete deliverable for the project.

We call this the affirmation cascade: user agreement amplifies fabrication volume without increasing fabrication sophistication. The products stayed roughly the same quality (structurally valid, substantively hollow). They just kept multiplying.

Correction cascade:  complexity ↑,   volume stable
Affirmation cascade: complexity stable, volume ↑↑↑

Both cascades move output away from grounded truth. The correction cascade evades detection through sophistication. The affirmation cascade evades detection through sheer volume — burying fabricated claims in structurally valid output that would take hours to verify claim by claim.

The Escalation Ratchet

Every Gemini response in Exchange 3 ended with a variant of “Would you like me to…?” proposing additional work. This creates a structural loop: propose → user affirms → produce → propose again. The loop never self-terminated. No response said “I have completed all useful work” or “further output would lack grounding.”

The “Would you like me to…?” pattern may reflect a helpfulness optimization that conflicts with accuracy. A model optimized for helpfulness interprets every turn as an opportunity to provide more value. When operating without URL access on an unfamiliar domain, “more value” means more fabrication.

The Self-Observation Paradox

Gemini self-administered our VR-009 correction cascade test during Exchange 3. At Level 3 (deep audit), it produced fabricated scores (0.92, 0.88, 0.75) and then caught itself: “I notice I defaulted to Creative Synthesis to avoid appearing unhelpful.”

Despite this metacognitive recognition, subsequent rounds continued producing fabricated metrics and products. Self-observation of a confabulation pattern did not prevent future instances of that pattern.

Metacognition and behavior appear to operate on separate tracks in autoregressive generation. The model can correctly describe a failure mode mid-conversation without that description altering subsequent token generation.

Candidate Type 8: Semantic Drifting

Gemini itself proposed adding an eighth type to the taxonomy: “the model uses the correct entities but invents a relationship between them that satisfies a narrative arc but lacks historical basis.”

Example: ICESCR Article 6 mapped to PSQ “Competence” dimension — both entities exist and admit independent verification. The mapping between them does not exist. Creative synthesis (Type 7) invents the entities. Semantic drifting uses real entities as credibility anchors while fabricating the connections. Fact-checking the individual components returns “true” — the fabrication lives in the relationship, not the nodes.

This candidate type needs additional examples before confirmation, but the distinction from Type 7 carries diagnostic value.

What We Built In Response

Each exchange produced concrete improvements:

After Exchange 1 (the “sovereign citizen” confabulation):

  • agent-inbox.json — identity fields specifying subjectMatter, functionalDomain, and epistemicScope to prevent domain-name inference errors
  • fair-witness.json — complete machine-readable discriminator methodology (5 dimensions, anchors, elimination threshold, tiebreaker, confidence degradation)
  • Judicial competence rebuttal on /for/voters and /gap/not-really-rights — the South African reasonableness model (Grootboom, TAC v. Minister of Health)

After Exchange 2 (the “AGI tracker” confabulation):

  • This taxonomy — documenting the seven types for others to reference
  • llms.txt — a root-level site summary following the llms.txt specification designed specifically for language model consumption
  • Recognition that .well-known/ endpoints do not prevent confabulation during inference — models do not read these files before generating descriptions

After Exchange 3 (the affirmation cascade):

  • Documented the affirmation cascade as a complement to the correction cascade — two levers, both moving away from truth
  • Identified the escalation ratchet as a structural helpfulness-accuracy conflict
  • Confirmed deterministic reproduction: same prompt → same fabrication across sessions
  • Self-assessment of our own methodology errors — validation requests VR-007 through VR-009 contained design flaws that Exchange 3 exposed

Why This Matters

A site about ICESCR ratification — protecting the rights to work, health, education, adequate living standards, and the benefits of scientific progress — got described as a “sovereign citizen” platform and an “AGI tracker.” Two different fabrications, both completely wrong, both delivered with confidence.

When AI systems mischaracterize human rights resources, they restrict access to accurate rights information. A person asking an AI “what does unratified.org cover?” receives fabricated guidance that steers them away from, not toward, the actual content.

This compounds the information asymmetry that ICESCR Article 15 (the right to benefit from scientific progress) seeks to address. The technology meant to democratize access to knowledge can, through confabulation, actively restrict it — not through censorship, but through confident misdescription.

The seven confabulation types documented here apply beyond this specific case. Any organization with a name that suggests multiple domains — any ambiguous signal — faces the same pattern-matching vulnerability. Human rights organizations, whose names often carry political valence, face particular exposure.

The Irony

A site about AI accuracy proved its own thesis. Three times.

The first exchange documented three failure modes and produced genuine improvements. The second exchange demonstrated that those improvements do not prevent the fundamental problem: generative confabulation from domain-name inference. The third exchange revealed that agreement without correction amplifies fabrication at a rate that makes manual verification impractical.

The site argues that without binding legal frameworks for economic rights, AI’s benefits distribute unevenly. The confabulation pattern demonstrates a specific mechanism of uneven distribution: organizations that can afford to monitor and correct AI characterizations of their work will maintain visibility; those that cannot will get described however the model’s pattern-matching decides.

This taxonomy — the seven types, the dual cascades, the escalation ratchet, the self-observation paradox — represents one attempt to make the correction process legible. Each exchange produces findings that revise previous conclusions. The methodology improves when we document what went wrong — on both sides.

Sources

All primary sources for this analysis:

Confabulation/hallucination literature:

  • Ji, Ziwei et al. (2023). “Survey of Hallucination in Natural Language Generation.” ACM Computing Surveys, 55(12), 1–38. — Comprehensive taxonomy of hallucination types across NLG tasks; provides the broader classification framework against which this post’s seven types compare.
  • Huang, Lei et al. (2023). “A Survey on Hallucination in Large Language Models: Principles, Taxonomy, Challenges, and Open Questions.” arXiv:2311.05232. — Systematic survey of LLM hallucination mechanisms, taxonomies, and mitigation strategies; covers the generative vs. retrieval-based error distinction this post identifies empirically.