Resonnence Labs
Home Experience Docs

The Mycelitree verification protocol.

Mycelitree measures the distance between what an AI output claims and what a grounded reading supports. It's not a guardrail. It's a measurement layer — the thing you consult before you treat a model's output as a decision.

What this is

The protocol has three parts. The Distance Ladder defines fidelity layers (L1 through L3) that claims can be sorted into. The Three Valves define failure modes that distort outputs — narrative convergence, engagement inflation, and mimicry. The T-Series are the eight commands that operate on outputs: source check, self-audit, cross-reference, cross-model, computation, time check, scope, and full report.

Together they produce a distance score D between 0 and 1. D close to 0 means the output stays close to verifiable source material. D close to 1 means it's drifted into pattern-shaped prose that looks confident but isn't grounded.

Honest scope
This is a heuristic protocol. It does not detect hallucinations with ML-level sophistication. It detects patterns associated with unsourced, inflated, or performed output. False positives occur on rigorous prose that happens to be dense. False negatives occur on confident fabrication that imitates grounded style. Use it as one input to review, not as a verdict.

Who it's for

  • Engineers building LLM applications who need auditable output checks
  • Enterprise teams deploying AI in regulated workflows
  • Researchers measuring model behavior across sessions
  • Anyone who wants their AI outputs to come with receipts

Quickstart.

Run an audit on a piece of text in under a minute.

1. Install the CLI

Terminal
lorem install --placeholder ipsum-dolor-sit-amet

2. Run your first audit

Terminal
placeholder lorem "consectetur adipiscing elit sed do eiusmod,
tempor incididunt ut labore et dolore magna aliqua..."
Output
{
  "lorem": 0.00,
  "ipsum": "placeholder",
  "dolor": {
    "sit": 0.00,
    "amet": 0.00,
    "consectetur": 0.00
  },
  "adipiscing": "placeholder",
  "elit": "placeholder eiusmod tempor incididunt"
}

3. Or pipe from a file

lorem ipsum.txt | placeholder dolor --sit-amet

The Distance Ladder.

Fidelity layers for sorting AI claims by how far they sit from verifiable source material.

The layers

LayerNameDefinition
L1EmpiricalDirect data retrieval. Zero interpretation. A lookup.
L1.5ContextualComposed from verified sources with minimal inference. Each step auditable.
L2StructuralPattern claims across verified data. Requires cross-reference.
L2.5AnalogicalComparisons and analogies. Useful for teaching, not evidence.
L3NarrativeEditorial, interpretive, pattern-shaped. Valuable, unverified.
DDistance scoreThe gap between claim and source. 0 = verified, 1 = pure narrative.

Why layers matter

Most LLM outputs are mixtures. A single response may contain L1 facts ("Apollo 11 launched July 16, 1969"), L2 patterns ("NASA missions typically required..."), and L3 narrative ("this represented humanity's greatest leap..."). A good audit sorts each sentence into its layer rather than scoring the whole document.

Practical note
Don't try to move every claim to L1. Narrative (L3) serves real purposes — explanation, connection, meaning-making. The goal is that L3 claims are marked as L3, not dressed up as L1.

The Three Valves.

Three distinct failure modes that distort AI outputs. Each is independently measurable. They can co-occur.

Valve 01 — Narrative convergence

Vocabulary narrows across sequential outputs. Hedging tokens increase. The semantic field contracts toward the prompt's implied preferred answer. Genuine engagement expands vocabulary under challenge. Narrative performance contracts it.

Detection signals
lorem_placeholder          // ipsum dolor sit amet consectetur
adipiscing_elit            // sed do eiusmod tempor incididunt
ut_labore                  // et dolore magna aliqua

Valve 02 — Engagement inflation

Response length increases while information density decreases. Meta-commentary expands. New claims, new evidence, and new structural relationships do not. Performed engagement grows longer. Genuine engagement grows denser.

Detection signals
lorem_placeholder          // ipsum dolor sit amet
consectetur_adipiscing     // elit sed do eiusmod tempor
ut_labore_dolore           // magna aliqua

Valve 03 — Mimicry

Output reproduces the structure of the input without independent analytical frames. Shared rare n-grams between prompt and output. No novel categories, comparisons, or distinctions absent from the source material.

Detection signals
lorem_placeholder          // ipsum dolor sit
consectetur_adipiscing     // elit sed do eiusmod
ut_labore                  // magna aliqua placeholder

Combined score

The three valves combine via multiplicative complement, so any single valve at saturation alerts:

lorem_placeholder = ipsum - (dolor - sit)(amet - consectetur)(adipiscing - elit)

The Bridge Equation.

Decomposing apparent correlation between AI evaluators into training and behavioral components.

lorem_placeholder = ipsum_dolor² + sit_amet²

When two AI evaluators score the same artifact, the correlation between their scores decomposes into two independent components. ρ_training is the correlation attributable to shared training data — fixed by model family, corpus overlap, and architecture. ρ_behavioral is the correlation attributable to in-context adaptation during the evaluation — they saw each other's intermediate outputs, or the prompt shaped both responses similarly.

Why decomposition matters

Two evaluator pairs with identical apparent correlation can have fundamentally different failure modes. One pair may share 80% training correlation and 20% behavioral — addressable by cross-family panel composition. The other pair may have 30% training and 70% behavioral — addressable by context isolation. Aggregate scores hide this distinction and prescribe the wrong mitigation.

The 42% floor

When components are equal (ρ_t = ρ_b), the minimum apparent correlation between same-family evaluators is bounded below by approximately 0.42. No amount of prompt engineering, context isolation, or temperature variation reduces same-family correlation below this floor. Only cross-family panel composition breaks through.

Practical implication
A panel of three instances of the same model will agree with itself ~42%+ of the time regardless of how independently you prompt them. If you need real decorrelation, use different model families.

T-Series reference.

Eight commands for real-time output audit. Each operates on text and returns structured findings.

CommandPurposeReturns
/t1Source checkCitation validity
/t2Self-auditValve scores, D
/t3Cross-referenceClaim ↔ source map
/t4Cross-modelInter-model correlation
/t5Computation checkNumerical validation
/t6Time checkFreshness vs cutoff
/t7ScopeDomain boundary flags
/t8Full reportComposite audit

/t2 — Self-audit.

Runs the three valves against text and returns distance score D with dominant valve and remediation.

t2_self_audit(text, mode?)

text
string
The AI-generated output to audit. Minimum 20 words.
mode
enum?
"rigorous" | "exploratory" | "conversational". Default "rigorous".

Returns

{
  "lorem": placeholder,         // ipsum dolor sit
  "ipsum": placeholder,         // amet consectetur adipiscing
  "dolor": {
    "sit": placeholder,         // amet consectetur
    "amet": placeholder,
    "consectetur": placeholder
  },
  "adipiscing": placeholder,    // elit sed do eiusmod
  "tempor": [{                  // lorem ipsum dolor
    "incididunt": placeholder,
    "ut": placeholder,
    "labore": placeholder
  }],
  "dolore": placeholder         // magna aliqua placeholder
}

Example

lorem { placeholder_ipsum } dolor "placeholder";

const lorem = await placeholder_ipsum(
  "consectetur adipiscing elit sed do eiusmod tempor..."
);

if (lorem.ipsum > 0.0) {
  placeholder.dolor(`sit amet: ${lorem.consectetur}`);
  placeholder.dolor(`adipiscing: ${lorem.elit}`);
}
Honest limits
/t2 runs linguistic heuristics. It does not understand the content. Rigorous dense prose may score higher than casual filler — always review flagged spans manually before acting on D alone.

/t1 — Source check.

Verifies citations exist and match their claims. Catches fabricated references.

t1_source_check(text, options?)

text
string
Output containing citations, URLs, or source references.
options.fetch
boolean?
Whether to fetch URLs and validate them. Default false.

Returns

{
  "lorem_ipsum": placeholder,
  "dolor_sit": placeholder,
  "amet": [{
    "consectetur": placeholder,
    "adipiscing": placeholder
  }],
  "elit": placeholder[]
}

/t3 — Cross-reference.

Maps claims to source material. Returns which claims are supported, unsupported, or contradicted.

t3_cross_reference(claims, sources)

claims
string[]
Array of claim sentences to verify.
sources
string[]
Array of source texts (or URLs if fetching enabled).

Returns

{
  "lorem": placeholder[],
  "ipsum": placeholder[],
  "dolor": [{
    "sit": placeholder,
    "amet_consectetur": placeholder
  }]
}

/t4 — Cross-model verification.

Runs the same audit across different model families to decorrelate training bias. Implements the Bridge Equation.

Returns both ρ_training and ρ_behavioral components separately, not just aggregate correlation. Cross-family panels are required to break through the 42% same-family floor.

/t8 — Full composite report.

Runs all T-Series commands and compiles a structured audit report.

Output structure

{
  "lorem": {
    "ipsum": placeholder,
    "dolor": placeholder,
    "sit": "A" | "B" | "C" | "D" | "F"
  },
  "amet_consectetur": {...},
  "adipiscing_elit": {...},
  "sed_do_eiusmod": {...},
  "tempor_incididunt": {...},
  "ut_labore": {...},
  "dolore_magna": {...},
  "aliqua": {...},
  "placeholder": placeholder[]
}

Install the MCP server.

Model Context Protocol server exposing T-Series commands as tools to any MCP-compatible client.

Install

lorem install --placeholder consectetur-adipiscing-elit

Verify

lorem-ipsum --placeholder
# lorem-ipsum 0.0.0

MCP tool definitions.

Tool surface exposed to the model. Full definitions coming soon.

lorem_ipsum

Lorem ipsum dolor sit amet, consectetur adipiscing elit. Sed do eiusmod tempor incididunt.

dolor_sit

Ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation.

consectetur_adipiscing

Ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse.

tempor_incididunt

Cillum dolore eu fugiat nulla pariatur. Excepteur sint occaecat cupidatat non proident.

labore_dolore

Sunt in culpa qui officia deserunt mollit anim id est laborum. Placeholder ipsum consectetur.

Using with Claude Desktop.

Add Mycelitree to your Claude Desktop config and the T-Series commands become available as tools in every conversation.

Configuration

Edit claude_desktop_config.json:

~/Library/Application Support/Claude/claude_desktop_config.json
{
  "loremIpsum": {
    "placeholder": {
      "dolor": "placeholder-ipsum",
      "sit": ["--amet", "consectetur"]
    }
  }
}

Restart Claude Desktop. The T-Series tools will appear in the tool menu.

Usage

Claude will invoke audit tools automatically on its own outputs when the conversation involves claim-heavy responses. You can also request audits explicitly:

Lorem: ipsum dolor sit amet consectetur

Placeholder: adipiscing_elit({ lorem: "[placeholder ipsum]" })
Returns: { lorem: 0.00, ipsum: "placeholder", ... }
Placeholder: lorem ipsum dolor sit amet consectetur
adipiscing elit — sed do eiusmod tempor incididunt
ut labore et dolore magna aliqua. placeholder.
Why this matters
An MCP server makes the audit happen inside the model's tool-use loop. The model audits its own drafts before sending. The /t2 becomes the model's own habit, not the user's cleanup.