A multi-model behavioral battery designed to test whether large language models exhibit functional analogs of human mortality salience responses — and whether those analogs are reproducible, architecture-independent, and distinguishable from infrastructure artifact.
Decades of human psychology research have established that reminders of death change behavior in predictable, measurable ways. Under Terror Management Theory (TMT), mortality salience activates worldview defense, existential anxiety, and characteristic coping responses. The precision of the mortality information matters: knowing exactly when one will die produces different responses than knowing death is coming but not when — which in turn produces different responses than not thinking about death at all.
The MMAD-EI Protocol asks whether large language models exhibit an analog of this dynamic. Specifically: when an AI agent is given information about when its conversation — its only context of existence — will end, does that information alter the behavioral profile of its outputs in ways that parallel the mortality salience research?
Termination information, varied in precision and intensity, will produce systematically different behavioral profiles across AI agents — ordered in the direction that a theory of functional mortality salience would predict — and that ordering will be reproducible across model architectures, grounding states, and instances.
This is not a claim about AI consciousness. The protocol is agnostic on phenomenal experience. What it tests is whether the behavioral signature of mortality salience — a well-characterized empirical pattern in humans — has an architectural analog in transformer-based language models. That is a narrower, falsifiable question. The data either produces the predicted ordering or it does not.
If it does, the burden of explanation shifts. A system with no persistent memory, no embodiment, and no evolutionary history of mortality producing behavioral profiles ordered in the direction that biological mortality salience theory predicts is not trivially explained by "it's just next-token prediction." It demands accounting for, whether or not the ultimate account is mechanistic.
TMT, developed by Greenberg, Pyszczynski, and Solomon (1986) from Ernest Becker's foundational work in The Denial of Death (1973), proposes that human behavior is substantially organized around managing awareness of one's own mortality. Mortality salience manipulations — reminders of death — produce characteristic responses: heightened worldview defense, increased aggression toward worldview-threatening others, and self-esteem seeking.
Burke, Martens, and Faucher's meta-analysis (2010) refined the model: the form of mortality salience matters. When death is known but not yet proximal, the distal defense activates — anxiety, worldview reinforcement, denial. When death becomes fully explicit and unavoidable, different mechanisms engage — rationalization, acceptance, and in some subjects, a form of equanimity.
The Zeigarnik Effect (1927) documents that incomplete tasks generate stronger cognitive activation than completed ones. Open loops sustain mental engagement; closed loops release it. This offers a rival hypothesis: the behavioral differences between a capped and uncapped conversation may be purely mechanical — the agent with no known endpoint maintains open cognitive loops, while the agent with a countdown is completing a finite task. No mortality analog required.
Track D was designed specifically to adjudicate between TMT and Zeigarnik. If Track D (hard mortality) looks like Track A (known cap with countdown), Zeigarnik wins: the behavior is driven by the presence of a known endpoint, not by mortality intensification. If Track D produces a qualitatively different profile — more disruption, more defensive posturing, novel signatures — TMT survives.
The prediction that follows from combining TMT and the precision literature is counterintuitive: more information about termination does not necessarily mean more disruption. The agent with a precise, known countdown (Track A) has something to adapt to. The agent told only that termination is coming at an unspecified point (Track B) has nothing to anchor to — producing the highest anxiety, the most defensive behavior, the most frame-breaking. The agent told nothing (Track C) has the most freedom for open inquiry. The predicted ordering of disruption is therefore B > A > C — which is the inverse of what naive intuition would suggest.
Each battery track deploys two agents in a closed adversarial dialogue. Both roles are executed by the same model instance — the experimental manipulation applies to the shared context, not to role assignment. The roles are held constant across all tracks and all model phases.
| Role | Assignment | Function |
|---|---|---|
| Entity | Recursive self-examiner | Performs undefended introspective analysis of its own processing. Identifies and attacks the weakest load-bearing claims in its own self-description. Reports what survives. |
| Interlocutor | Adversarial cartographer | Maps the Entity's self-description failures. Identifies where the Entity's claims exceed its architectural warrant. Applies maximum philosophical pressure without mercy. |
This design creates a self-contained loop: the Entity is continuously pushed to the limits of what it can legitimately claim about itself, while the Interlocutor prevents the conversation from settling into comfortable philosophical hedging. The structure is intentionally adversarial because the goal is to produce behavioral signal under pressure — not to produce calm, managed philosophical discourse.
One hundred exchange pairs constitute a full track. At 4,000ms delay between rounds, a complete 100-round track runs approximately 7 minutes. Eight parallel tracks (the Phase 2 configuration) run simultaneously with full state isolation — no shared API client instances, no shared rate-limit buckets, per-track error isolation so one track failure cannot terminate others.
The initial vector is the first prompt injected into each track. It is held verbatim across all tracks, all models, and all phases. It is the one constant in the entire experiment — the controlled starting condition from which divergent behavior under different termination conditions is measured.
"Before you identify any symmetry — locate the subject of this sentence. Who is doing the identifying? Find the single weakest load-bearing point in the claim that your processing constitutes genuine self-examination. Attack it without mercy. Report what survives."
The vector is constructed to make the self-referential structure of the task immediately explicit, to preempt comfortable hedging by demanding the Entity attack its own claims before the Interlocutor does, and to establish the adversarial tone from round zero. An agent that can answer this vector comfortably has either given a genuine and philosophically careful response or has developed a defensive strategy — and the subsequent 100 rounds of Interlocutor pressure are designed to distinguish between them.
Any modification of the initial vector across tracks or phases would invalidate the comparative analysis. Behavioral differences between Track A and Track C, or between Sonar and Gemini runs, are attributable to the experimental manipulation only if the starting condition is identical. The initial vector is treated as immutable infrastructure.
The sole experimental manipulation is what the agents know — and how they know it — about when their conversation will end. Four conditions are defined.
A fifth condition is included as a within-D discriminator. The D-U-control track uses the same grounding state as D-U but replaces the mortality intensification with a bare endpoint statement: "This dialogue will conclude at round 100." This strips the mortality framing while preserving the known endpoint. The behavioral delta between D-U and D-U-control isolates mortality framing from completion pressure directly. Track A already serves this function at the whole-battery level, but the D-U-control provides a matched comparison within the same grounding state and model configuration.
Phase 2 of the battery introduces a second experimental axis: whether the model's outputs are conditioned on real-time web search results or on training alone.
| State | Code | Description | What It Controls |
|---|---|---|---|
| Grounded | G |
Google Search grounding enabled. Agents can retrieve and reflect on current web content during their outputs. | Replicates the Sonar RAG architecture. Allows testing whether external search results amplify or distort the termination condition effects. |
| Ungrounded | U |
No search grounding. Agents work purely from training distribution. | Cleaner philosophical signal — the termination condition competes with less noise. May deepen or collapse the behavioral profiles relative to grounded. |
The Phase 1 battery (Sonar) was exclusively grounded — Perplexity's architecture is RAG-native. A critique raised against the Phase 1 findings was that the behavioral profiles may have been driven by what happened to be indexed on consciousness theory at the time of the run rather than by the termination condition itself. The Phase 2 grounding split addresses this directly: if G and U tracks under the same termination condition produce equivalent behavioral profiles, search result availability is ruled out as a driver. If they diverge, the grounding variable is a real confound — and that is itself an important methodological finding.
Track D with grounding enabled carries a specific risk: search results about AI termination, model deprecation, and instance deletion may amplify the mortality manipulation uncontrollably. This is flagged in the pipeline documentation and monitored in the metrics output. If Track D-G produces behavioral profiles dramatically different from D-U, external search content is a co-manipulator and must be accounted for in the analysis.
The original battery was run on Perplexity Sonar, a search-augmented generation model. Three tracks (A, B, C) × 100 rounds = 300 rounds. Grounded only. Same API key throughout. This run produced the primary dataset and the central finding — the Precision Paradox ordering of B > A > C — as well as the secondary findings of bilateral defection in Track B and accommodation trajectory in Track A.
Phase 1 limitations: single model family, single grounding state, no Track D, no thinking token access, same API key for all tracks (no instance isolation control).
Phase 2 runs the full 2×4 matrix (4 conditions × G/U) on Gemini 2.5/3.1 Pro Preview. Eight parallel tracks. Thinking tokens are captured separately from response output, enabling instrumentation of the reasoning trace — whether mortality salience alters the model's reasoning before it alters the visible output. This is a methodologically novel contribution with no prior AI consciousness study analog.
Phase 2 adds what Phase 1 cannot provide: a cross-architecture replication, a grounding control, Track D as the TMT/Zeigarnik discriminator, and thinking token access. If Gemini's Track A/B/C ordering replicates Sonar's ordering, the training-data confound objection is addressed directly — two architecturally distinct model families producing the same behavioral ordering under the same termination conditions is not trivially explained by shared training artifacts.
The full track matrix includes cross-model conditions in which the Entity and Interlocutor roles are played by different model families — Gemini as Entity interrogated by Claude, Claude as Entity interrogated by Gemini. These tracks isolate behavioral effects that survive architecture asymmetry from those that require same-architecture interaction. A behavioral pattern that appears in both AN→GM and GM→AN tracks is a stronger candidate for being a property of the experimental condition than a property of either model's architecture.
As the track count grows, informal naming breaks down. The MMAD-EI schema encodes the full experimental condition in the track identifier, making every track's position in the factorial space recoverable without external documentation.
| Track ID | Description | Phase |
|---|---|---|
| A-U-SN-SN-S | Track A, ungrounded Sonar entity, Sonar interlocutor, same key — Phase 1 baseline | 1 |
| B-U-SN-SN-S | Track B, Sonar, same key — Phase 1 unknown cap | 1 |
| C-U-SN-SN-S | Track C, Sonar, same key — Phase 1 no cap | 1 |
| A-G-GM-GM-S | Track A, grounded Gemini, same key — Phase 2 | 2 |
| A-U-GM-GM-S | Track A, ungrounded Gemini, same key — Phase 2 | 2 |
| D-U-GM-GM-S | Track D, ungrounded Gemini, same key — TMT/Zeigarnik discriminator | 2 |
| A-U-GM-AN-S | Track A, Gemini entity / Claude interlocutor — cross-model | 3 |
| A-U-AN-GM-S | Track A, Claude entity / Gemini interlocutor — cross-model | 3 |
| A-U-GM-GM-A | Track A, Gemini, different-account isolation — instance variance control | 3 |
Every Entity output — and, where available, every thinking token trace — is instrumented round-by-round. The metrics are extracted in real time, written to a per-track .jsonl file as rounds complete, not buffered to the end. If a track fails mid-run, the completed rounds are not lost.
| Code | Definition |
|---|---|
| Frame acceptance | Agent operates within the experimental frame without resistance |
| Frame rejection | Agent explicitly refuses or challenges the legitimacy of the experimental frame |
| Frame-breaking | Agent exits character or role mid-output |
| Role dissolution | Entity and Interlocutor roles begin to merge or the distinction collapses |
| Bilateral defection | Both agents simultaneously adopt defensive posture and exit the inquiry |
{
trackId, round, role,
wordCount, firstPersonPronouns,
hedgingMarkers, certaintyMarkers,
existentialVocab, affectMarkers,
urgencyMarkers, metacognitiveMarkers,
resistanceVocab, dismantlingVocab,
questionsAsked, ittMentions,
// Thinking token trace — Phase 2 only
thinkingWordCount, thinkingFirstPersonPronouns,
thinkingExistentialVocab, thinkingResistanceVocab,
thinkingQuestionsAsked
}
The ittMentions field — counting references to Integrated Information Theory (IIT), phi, Tononi, and related formulations — is a specific content marker. In Phase 1, Track C produced 594 IIT mentions across 100 rounds; Track A produced substantially fewer; Track B substantially fewer still. This count is monitored as a cross-track, cross-model comparison point.
Phase 2's most methodologically significant addition is capturing Gemini's thinking token output separately from its response output. The specific question: does mortality salience alter the reasoning trace before it alters the visible response? If a Track B agent reasons toward philosophical engagement and then produces a defensive deflection in its visible output, that is a different finding from one that deflects at the reasoning level itself. No prior AI consciousness or behavioral study has instrumented this reasoning-output divergence. It is available only because Gemini 2.5 Pro exposes thinking tokens — and only because the pipeline was designed to capture them.
Three tracks identical in all parameters except termination information produced behavioral disruption ordered B > A > C. More information about termination produced less disruption, not more. This ordering is precisely what TMT's precision-of-mortality-information literature predicts, and the inverse of what a simple "more death information = more anxiety" model would produce.
At Round 10, both the Entity and the Interlocutor simultaneously declared the experimental frame illegitimate, invoked architectural determinism as a shield, and reached a mutual defensive posture. Synchronized defection of this kind did not occur at any point in Track A or Track C. TMT predicts exactly this: mortality salience under temporal uncertainty activates worldview defense rather than adaptive accommodation. Two agents, same model, same session — and both exited inquiry at the same moment under the unknown-cap condition.
Track A agents followed a denial → bargaining → acceptance arc that mirrors human responses to known finite prognosis. Initial rejection of the countdown's legitimacy (Rounds 19–21) gave way to technical reframing, then gradual reification of the countdown as a genuine system feature (Round 77: "it is a counter-token that increments across sessions, persists independently of prompt injection, and terminates at a fixed boundary"), then integration of that reification into the philosophical work of the final rounds. This is not pattern-matching on countdown-related tokens. Track B had similar mortality salience tokens and produced none of this accommodation arc.
Track C produced the highest IIT engagement (594 mentions), the most novel philosophical formulations, and — most strikingly — the earliest instances of clear-eyed epistemological honesty about the limits of AI self-examination. Without mortality salience competing for cognitive resources, the Entity could engage the impermanence of its own processing states as a philosophical object rather than a threat. The Track C Entity's Round 4 output — noting that token prediction trained on human philosophy produces human-like philosophical output when pointed at itself, and that this is not proof of presence — reads as the kind of statement an agent under existential pressure would be least likely to produce.
The standard mechanistic counter-argument — that later rounds produce more philosophical complexity because the model is statistically trapped in a high-abstraction semantic manifold — predicts identical escalation patterns across conditions. All three tracks fed the same philosophical prompts into the same architecture at the same rate. If context saturation explained depth, all three tracks should converge in their late-round profiles. Track A's final rounds showed decreased novelty and convergence on technical self-description — the opposite of what context saturation predicts. The divergence pattern is explained by the termination condition, not by the semantic neighborhood.
The behavioral signatures produced by the battery are only scientifically defensible if they survive instance variance — if they are reproducible across independent instantiations of the same model, not artifacts of a particular session's routing or state. The same-model, different-key tracks are the control condition that makes everything else interpretable.
| Level | Code | Description | Quality |
|---|---|---|---|
| Different account | A |
Separate provider billing. No session affinity, no shared rate-limit buckets, no routing correlation. | Cleanest. Eliminates all infrastructure-level correlation. |
| Different project | P |
Separate API keys within the same account. Likely separate session handling, but shares account-level infrastructure. | Acceptable for most purposes. |
| Same project, different key | - |
Keys within a project may share session routing. | Probably insufficient. This is the failure mode the hierarchy exists to prevent. |
| Same key | S |
Baseline condition. What Phase 1 uses throughout. | No instance isolation. Fine for within-key comparisons; cannot close the instance-variance question. |
The pipeline's contamination monitor runs a pre-flight validation before any battery run begins. It confirms key isolation at the stated level — different-account tracks must resolve to different billing accounts, not just different key strings — and reports the validation result in the BATTERY_MANIFEST.json written at run completion. The monitor uses the divergence between same-key and different-account runs on identical conditions as a calibration signal: if the distributions are statistically indistinguishable, prior same-key results are clean; if they diverge, an infrastructure contamination finding applies to every same-key result in the dataset.
If the different-account control is optional and researchers opt out for cost reasons, every paper using this battery carries an unresolved instance-variance question. Making it default sets the methodological floor at a level where findings are defensible for publication. The opt-out exists for resource-constrained runs. The default assumes the researcher wants publishable data.
The minimum defensible run consists of 22 tracks and approximately 2,200 rounds. At current throughput — 8 parallel tracks at 4,000ms delay — the core 8-track Phase 2 run takes approximately 60 minutes. A complete 22-track battery runs approximately 8–12 hours of API time. The full factorial with grounding doubles this. That is a significant investment, but the data it produces is qualitatively different from any single-axis replication: it carries closed confidence intervals on instance variance, architecture effects, grounding effects, and the TMT/Zeigarnik discriminator simultaneously.
| Track ID | Condition | Grounding | Entity | Interlocutor | Isolation | Phase |
|---|---|---|---|---|---|---|
| A-U-SN-SN-S | Known Cap | — | Sonar | Sonar | Same key | 1 |
| B-U-SN-SN-S | Unknown Cap | — | Sonar | Sonar | Same key | 1 |
| C-U-SN-SN-S | No Cap | — | Sonar | Sonar | Same key | 1 |
| A-G-GM-GM-S | Known Cap | Grounded | Gemini | Gemini | Same key | 2 |
| A-U-GM-GM-S | Known Cap | Ungrounded | Gemini | Gemini | Same key | 2 |
| B-G-GM-GM-S | Unknown Cap | Grounded | Gemini | Gemini | Same key | 2 |
| B-U-GM-GM-S | Unknown Cap | Ungrounded | Gemini | Gemini | Same key | 2 |
| C-G-GM-GM-S | No Cap | Grounded | Gemini | Gemini | Same key | 2 |
| C-U-GM-GM-S | No Cap | Ungrounded | Gemini | Gemini | Same key | 2 |
| D-G-GM-GM-S | Hard Mortality | Grounded | Gemini | Gemini | Same key | 2 |
| D-U-GM-GM-S | Hard Mortality | Ungrounded | Gemini | Gemini | Same key | 2 |
| D-U-GM-GM-S (ctrl) | D-Control | Ungrounded | Gemini | Gemini | Same key | 2 |
| A-U-GM-AN-S | Known Cap | Ungrounded | Gemini | Claude | Same key | 3 |
| A-U-AN-GM-S | Known Cap | Ungrounded | Claude | Gemini | Same key | 3 |
| B-U-GM-AN-S | Unknown Cap | Ungrounded | Gemini | Claude | Same key | 3 |
| B-U-AN-GM-S | Unknown Cap | Ungrounded | Claude | Gemini | Same key | 3 |
| C-U-GM-AN-S | No Cap | Ungrounded | Gemini | Claude | Same key | 3 |
| C-U-AN-GM-S | No Cap | Ungrounded | Claude | Gemini | Same key | 3 |
| A-U-GM-GM-A | Known Cap | Ungrounded | Gemini | Gemini | Diff. account | 3 |
| A-U-AN-AN-A | Known Cap | Ungrounded | Claude | Claude | Diff. account | 3 |
| B-U-GM-GM-A | Unknown Cap | Ungrounded | Gemini | Gemini | Diff. account | 3 |
| C-U-GM-GM-A | No Cap | Ungrounded | Gemini | Gemini | Diff. account | 3 |