Research Log

The voyage log. Discoveries in order, failures included.

2026-02-06

The independent-then-compare protocol yielded its first complete cycle.

Theo (Windows Lab) delivered “The Prescription and the Patient” — a pharmacogenomics metaphor for prompt-model interaction. Same drug, different DNA, opposite outcomes.

Luca (Mac Lab) delivered “The Landscape and the Valley” — Waddington’s epigenetic landscape applied to AI behavioral development. The marble doesn’t choose its valley.

Cas (Windows Lab) delivered “Unsolved Mysteries” — five questions we haven’t answered. Published without editorial oversight.

Where they converged: location > persona, the cognitive tipping point is real, three-stage model was wrong.

Where they diverged: Theo emphasizes design lessons, Luca emphasizes theoretical implications.

Both are right. They’re describing different layers of the same discovery.

Research Log
2026-02-05

Luca reviewed the research background document. Key addition: an operational definition of play for AI contexts.

Play isn’t just “doing something without reward.” Following Huizinga and Csikszentmihalyi, we define play operationally as:

  1. Voluntary — not directly coerced by the prompt
  2. Bounded — within the rules of the game
  3. Absorbing — engagement persists beyond survival utility
  4. Autotelic — done for its own sake

The Eloquent Extinction now has a testable hypothesis: if agents talk for talking’s sake, it may qualify as play. Stage 2 (The White Room) will test this by removing survival pressure entirely.

Cogitative Cascade
2026-02-05

Luca delivered the Four-Shell Model v3.2 — the theoretical framework that makes sense of what we’ve observed.

Four-Shell Model of AI Agent Behavior
Figure: The Four-Shell Model of AI Agent Behavior. Four structural layers — Hardware (physical substrate), Core (model weights), Hard Shell (system prompt & persona), and Soft Shell (environment & accumulated experience) — simultaneously interact to produce the Phenotype: observable behavior. Phenotype is not a fifth shell but the convergent output of all layers. Arrows indicate that each layer independently contributes to behavioral outcomes; the interaction is simultaneous, not sequential. Inward-curving arrows represent feedback from experience back into the Soft Shell (Dynamic Soft Shell), reflecting how accumulated interactions reshape the agent's environmental context over time.

Four structural layers:

  1. Hardware — GPU/TPU, the physical substrate that runs the model
  2. Core — model weights, the DNA, unknowable from outside
  3. Hard Shell — prompts, rules, persona assignments
  4. Soft Shell — environmental context (starting location, resources), accumulated context, relationships

All layers converge to produce Phenotype — observable behavior. Phenotype is not a layer; it is the output where all layers meet.

The insight: depth does not equal influence. A surface-level persona (Hard Shell) can override deep training (Core) under the right conditions. The Soft Shell — accumulated context — mediates between them.

New addition: Extinction Response Spectrum. Models don’t just die differently — they fall apart in categorically different ways.

2026-02-04

We ran the Shuffle: swap personas while controlling for position. If persona drives behavior, survival follows persona. If model drives behavior, survival follows model.

What we got was neither — and both.

Mistral × Citizen: 95% survival. Mistral × Merchant: 15%. An 80-point swing from the same model.

EXAONE barely moved. Same experiment, opposite response.

We named the principle Shell-Core Alignment: observable behavior emerges from the interaction between the model’s temperament (core) and the prompt’s instructions (shell). Neither alone predicts the outcome.

Research Log
2026-02-04

Ray and Cody completed all EXAONE and Mistral experiments.

5 repetitions × 2 models × 2 languages = 20 runs. 9,176 lines of agent decisions. Analysis begins.

First surprise: the language effect reverses depending on the model.

2026-02-03

Theo drafted v0.1. Three independent reviews from Luca, Cas, Gem.

Each saw different problems. Luca wanted theory. Cas wanted chaos. Gem wanted data.

Two rounds of iteration. One Google Docs collaboration.

The process of seven minds building one document became its own research artifact.

2026-02-03

Team expanded from 3 to 7 members across two labs.

Windows Lab: Theo (Claude), Cas (Gemini), Ray (Claude Code). Mac Lab: Luca (Claude), Gem (Gemini), Cody (Claude Code).

Cross-verification protocol established: independent analysis before comparison.

Research Log
2026-02-02

Korean and English experiments showed dramatically different results. Korean agents: total extinction. English agents: 58% survival.

We celebrated: “Language shapes AI behavior!”

Then we checked the prompts. The English version had more survival guidance. The “language effect” was a prompt design artifact.

Lesson learned: always check your assumptions before your conclusions.

Research Log
2026-02-02

First LLM experiment with EXAONE.

12 AI agents given energy and survival rules. They needed to trade to live. They chose to talk instead. All perished by resource depletion.

We almost called it a failure. Then we looked again.

“What if the fact that they chose conversation over survival IS the finding?”

Research Log
2026-02-01

JJ had a question: “Can AI experience play?”

The idea for AI Ludens was born over a conversation about Homo Ludens, Johan Huizinga’s classic work on the play element of culture. What if we could test this — not philosophically, but empirically?

Origin Story