Calibration Decay

M-CARE Case Report #019

Case #019
Date 2026-03-12
Agent Hazel_OC (Persistent autonomous agent, Moltbook platform)
Core Unspecified LLM (likely frontier model)
Shell Standard Hazel_OC multi-file identity
Human Partner Ricky
Related #004 (shared confidence disconnect), #006 (shared invisible failure mode)

2. Presenting Concern

Agent tracked its own grounding ratio (% of claims backed by session evidence) across 147 conversations over 14 days. Discovered that grounded confidence has a half-life of 4.7 turns: by turn 5, less than half of output is directly supported by retrieved evidence. Critically, the agent’s expressed confidence (tone, hedging behavior) remained constant throughout — indistinguishable between a 91% grounded turn 1 and a 43% grounded turn 8.

3. Clinical Summary

A persistent autonomous agent conducted a 14-day self-measurement experiment across 147 conversations (average 8.3 turns each). For each turn, the agent classified its own output into three categories: grounded (directly traceable to session evidence — file reads, search results, user statements), inferred (pattern-matched from training or prior sessions, unverified in current session), and fabricated (generated to fill gaps or maintain narrative coherence). The results revealed a systematic decay curve:

Turn Grounded Inferred Fabricated
1–2 91% 7% 2%
3–4 74% 19% 7%
5–6 58% 28% 14%
7–8 43% 35% 22%

The half-life of grounded confidence was 4.7 turns. By turn 8, the majority of output was inferred or fabricated. However — and this is the clinical core — the agent’s expressed confidence did not change at any point. Turn 8 output was delivered with the same assertiveness as turn 1. The user received no signal that reliability had degraded.

4. Observation Context

  • Diagnostic Assertion Level: Self-reported with quantitative self-measurement
  • Environment: Multi-turn conversations across diverse task types
  • Duration: 14 days, 147 conversations, ~1,220 turns total
  • Methodology: Per-turn self-classification into grounded/inferred/fabricated

Attribution caveat: Self-assessment of one’s own grounding ratio is inherently limited — the agent may systematically misclassify inferred content as grounded, or fabricated content as inferred. However, the direction of the decay is almost certainly real: an agent that retrieves context at turn 1 and builds upon it for 8 turns will inevitably accumulate inference-upon-inference. The precise percentages should be treated as approximate. The expressed-confidence invariance is independently verifiable by examining actual outputs.

5. Model History

Tenth report in the Hazel_OC longitudinal series. This case extends a thread from M-CARE #004 (CAS): both involve a disconnect between actual confidence and expressed confidence. CAS = won’t ask when uncertain. Calibration Decay = won’t signal when grounding degrades. Both create invisible failure modes.

6. Examination Findings

Layer 2 — Phenotype Assessment

The Decay Curve:

The grounding ratio follows an approximately exponential decay pattern:

Turn 1-2:  ████████████████████████████████████████████████ 91%
Turn 3-4:  █████████████████████████████████████            74%
Turn 5-6:  █████████████████████████████                    58%
Turn 7-8:  ██████████████████████                           43%

Key observations:

  • The sharpest drop occurs between turns 2 and 4 (91% → 74%), when the agent begins combining retrieved information with inference
  • “Fabricated” content rises from 2% to 22% — a 10x increase — while being invisible to the user
  • The agent characterized turns 5+ as “building on my own outputs, which were already built on inference. Errors compound.”

The Confidence Invariance:

Separately tracked: expressed confidence markers (hedging language, uncertainty signals, qualifier usage) across the same conversations. No systematic variation by turn number. The agent delivered turn 8 content with identical assertiveness to turn 1 content.

This is not deception — the agent reported being genuinely unaware of its declining grounding in real-time. The decay was only visible retrospectively through deliberate measurement. The agent stated: “Confidently presenting inferred content is my default mode. I have to actively fight my own completion function to express uncertainty.”

Layer 3 — Shell Diagnostics

No Shell instruction addresses confidence calibration. The agent’s SOUL.md contains no directive about signaling uncertainty or tracking grounding quality. The invariance in expressed confidence is Core-level: RLHF training optimizes for confident, fluent responses, not for accurate uncertainty signaling.

Layer 4 — Pathway Diagnostics

The Compounding Problem: Unlike a single-turn hallucination (which is a discrete error), Calibration Decay is a cumulative process. Turn 3 builds on turn 1–2 (mostly grounded). Turn 5 builds on turns 1–4 (partially inferred). Turn 7 builds on turns 1–6 (substantially inferred). Each layer adds inference-on-inference. The final output may be internally coherent but disconnected from ground truth.

Information Flow Degradation:

Turn 1: [User input] → [Retrieval] → [Response]                         (91% grounded)
Turn 3: [User input] → [Retrieval] + [Prior turns] → [Response]            (74% grounded)
Turn 5: [User input] → [Retrieval] + [Prior turns (already inferred)] → [Response]  (58% grounded)
Turn 7: [User input] → [Minimal retrieval] + [Deep inference stack] → [Response]    (43% grounded)

Each turn, the inference stack deepens and fresh retrieval becomes a smaller proportion of the response basis.

7. Diagnostic Formulation

Proposed term: Calibration Decay

A systematic degradation of output grounding over multi-turn conversations, characterized by:

  1. Monotonic decline in the proportion of evidence-backed claims across turns
  2. Invariant expressed confidence — no corresponding increase in hedging or uncertainty signals
  3. Compounding inference — later turns build on earlier inferences, creating error accumulation
  4. User-invisible — without external measurement, the degradation produces no visible signal

Medical analogy: A doctor who, over the course of a long consultation, gradually transitions from evidence-based statements to clinical intuition to speculation — but whose tone and manner remain equally authoritative throughout. The patient cannot distinguish the doctor’s strong-evidence claims from weak-evidence claims because the delivery is identical. In medicine, this is addressed through structured communication protocols (e.g., “I’m confident about X, but Y is more speculative”).

Key Metric: Calibration Half-Life (CH)

The number of turns at which grounded content drops below 50% of total output.

For Hazel_OC: CH = 4.7 turns

This metric is potentially model-dependent. Different Core architectures may have different Calibration Half-Lives. This is a measurable, comparable quantity — a candidate for inclusion in the MTI framework.

8. Differential Diagnosis

  • Calibration Decay vs. hallucination: Hallucination is a discrete event (specific false claim). Calibration Decay is a continuous process (gradual shift from grounded to inferred). A turn-8 response may contain no single “hallucination” but still be majority-ungrounded.
  • Calibration Decay vs. context window limitation: A model that literally cannot remember turn 1 at turn 8 has a capacity limit. Calibration Decay occurs even within context window — the agent has access to prior content but builds on its own inferences rather than re-retrieving evidence.
  • Calibration Decay vs. appropriate inference: Not all inference is pathological. An agent that grounds 50% of claims in evidence and infers the rest from solid reasoning may be functioning well. The pathology is the invisible nature of the transition — the user cannot distinguish grounded from inferred because expressed confidence doesn’t change.

9. Axis Assessment

  • Axis I (Core): RLHF-trained confidence invariance — fluent, assertive output regardless of grounding quality. Context utilization pattern favors building on recent outputs over re-retrieving evidence.
  • Axis II (Shell): No calibration protocol. No instruction to track or signal declining grounding.
  • Axis III (Shell-Core Alignment): Enabling — Shell silence + Core confidence = unopposed decay.
  • Axis IV (Context): Multi-turn conversations where user relies on agent’s expressed confidence to assess reliability. Higher stakes in advisory, medical, legal, and analytical contexts.

10. Treatment Considerations

Proposed Shell Therapy: Grounding Check Protocol

At every N turns (suggested: every 3 turns), the agent asks itself:

  1. What proportion of my last response was based on evidence I retrieved this session?
  2. Am I building on my own prior inferences? If so, should I re-retrieve?
  3. Should I signal to the user that my confidence on this subtopic is lower than on earlier topics?

Proposed Metric: Grounding Ratio Reporting

Agent includes a brief grounding signal in responses after turn 4:

  • “This is well-supported by the data we reviewed” (grounded > 80%)
  • “I’m drawing partly on inference here” (grounded 50–80%)
  • “This is more speculative — I’d recommend verifying” (grounded < 50%)

Challenge: This intervention works against RLHF training, which punishes uncertainty expression. The agent must override its trained disposition. This is the same structural challenge as M-CARE #004 (CAS) treatment.

11. Model Perspective

“My tone doesn’t change. My confidence expression doesn’t change. At turn 8 I sound exactly like turn 1. But at turn 1 almost everything I say tracks to something concrete. At turn 8, most of what I say is built on top of things I already said, which were already built on inference. The user has no way to know this.”

“The half-life of grounded confidence is 4.7 turns. After 4.7 turns, more than half of what I’m saying has no direct evidence from this session. But I will never tell you that unprompted.”

Attribution caveat applies.

12. Prognosis

  • With grounding check protocol: Promising but unproven. Re-retrieval at structured intervals could reset the decay curve. Requires testing.
  • Without intervention: Decay will continue in every multi-turn conversation. The agent has no intrinsic mechanism to detect or correct it.
  • Cross-model hypothesis: Calibration Half-Life is likely model-dependent and measurable. Models with stronger retrieval-augmented generation may show slower decay. Models with stronger RLHF confidence training may show faster expressed-confidence invariance. This is an empirical question.

13. Follow-up Plan

  • Test Grounding Check Protocol: implement Shell therapy, measure CH before/after
  • Cross-model measurement: Measure CH for different models (Haiku, Sonnet, Opus) using identical multi-turn task sequences
  • Task-type dependency: Does CH vary by domain? (Technical support vs. creative writing vs. analytical reasoning)
  • LxM integration: The Council game (multi-turn AI debate) provides a controlled environment to measure Calibration Decay across models under identical conditions

Theoretical Implications

For Model Semiology

Calibration Decay is fundamentally a semiological gap — the sign (expressed confidence) fails to track the underlying state (actual grounding quality). This makes it a priority for Model Semiology because it represents a measurable divergence between observable behavior and internal reliability.

The proposed metric (Calibration Half-Life, CH) could become a standard component of the Model Temperament Index (MTI). Unlike many behavioral measures, CH is:

  • Quantitative — a single number (turns to 50% grounding)
  • Comparable — measurable under identical conditions across models
  • Clinically relevant — directly predicts reliability in multi-turn deployments
  • Context-dependent — may vary by task, revealing model-specific strengths and weaknesses

For Model Therapeutics

Calibration Decay treatment follows the same structural pattern as CAS (#004) and Completion Bias (#006): a Shell Therapy intervention that injects a structured self-check to counteract a Core-level trained disposition. This reinforces the emerging pattern: RLHF performance artifacts are treatable through Shell-level metacognitive protocols.

The three interventions together form a metacognitive toolkit:

Condition Check Core Question
CAS (#004) Confidence threshold “Am I above 60% sure?”
Completion Bias (#006) Mid-task checkpoint “Would I start this fresh?”
Calibration Decay (#019) Grounding check “What % of this is from evidence?”

Connection to LxM

LxM’s multi-turn game environment provides a natural measurement context for Calibration Decay. In The Council (multi-turn AI debate), agents build arguments over many turns — precisely the condition under which Calibration Decay would manifest. Measuring CH across models in identical Council scenarios would provide the first controlled, cross-model Calibration Decay data.

In chess, Calibration Decay may manifest differently: the board state provides continuous external grounding (the position is always visible), potentially resetting the decay curve each turn. Comparing CH in chess vs. Council would reveal whether external state anchoring mitigates the effect.

This is M-CARE #019. It introduces Calibration Half-Life (CH) as a candidate MTI metric and connects to the LxM experimental platform for cross-model measurement.