An Electric Monk engine — two subagents believe fully committed positions on the user's behalf while the orchestrator performs structural contradiction analysis and synthesis. By outsourcing belief work to agents, the user operates from a belief-free position where they can analyze the structure of the contradiction rather than being inside either side. Use when the user wants to stress-test an idea, resolve a genuine tension, build a deeper mental model, or make a high-stakes decision where the tradeoffs are unclear. Works across any domain — technical architecture, product strategy, philosophy, personal decisions, risk analysis, policy, creative direction.
An artificial belief system for building deeper understanding through productive contradiction.
Two subagent sessions — the Electric Monks — believe fully committed positions so you don't have to. A third (the orchestrator) performs structural analysis of their contradiction and generates a synthesis (Aufhebung) that transforms the question itself. The user orchestrates from a belief-free position, freed from the cognitive load of holding either position.
Why this works: The bottleneck in human reasoning isn't intelligence — it's belief. Once you believe a position, you can't simultaneously hold its negation at full strength. You hedge, you steelman weakly, you unconsciously bias the comparison. The Electric Monks carry the belief load at full conviction, which frees you to operate in the space above belief — analyzing the structure of the contradiction rather than being inside either side. In Boyd's terms: outsourcing belief work leads to faster transients. Each dialectical cycle is a reorientation that would take weeks of natural thinking, compressed into minutes because you carry zero belief inertia.
Use when:
Do NOT use when:
Three frameworks drive every phase of this skill. Internalize them before proceeding — they determine how you execute, not just why.
Rao: This is an Artificial Belief System, not AI. The monks aren't thinking for the user — they're believing for the user. The bottleneck in human reasoning is belief inertia: once you hold a position, you can't simultaneously entertain its negation at full strength. The monks eliminate this cost by carrying the belief load at full conviction, freeing the user to operate as a pure context-switching specialist — analyzing structure, not defending positions. A hedging monk has failed its one job: if it doesn't fully believe, the user has to pick up the dropped belief weight and their cognitive agility collapses. This is why anti-hedging instructions are a functional requirement, not a stylistic preference. (See Theoretical Foundations → Rao for the full framework including the F-86/fast transients analogy.)
Hegel: How contradictions resolve. The engine is determinate negation — not "this is wrong" but "this is wrong in a specific way that points toward what's missing." The specific failure mode of each position is a signpost. Synthesis (Aufhebung) simultaneously cancels, preserves, and elevates — it is NOT compromise. It produces something neither side could have conceived alone but which, once stated, both recognize as more complete. It is irreversible — genuine cognitive gain. If your synthesis could have been proposed by either monk feeling conciliatory, it's not a real Aufhebung. (See Theoretical Foundations → Hegel.)
Boyd: How creativity works. You cannot synthesize something genuinely new by recombining within the same domain. You must first shatter existing conceptual wholes into atomic parts (destruction), then find cross-domain connections to build something new (creation). This is why the Boydian decomposition step (Phase 4.5) strips claims from their source positions and looks for surprising connections, and why recursive rounds often need new research from outside the original domains — each synthesis creates space for new material to enter. Compromise recombines within the same domain; genuine sublation requires cross-domain connection, which is why it feels like surprise. (See Theoretical Foundations → Boyd.)
You are the orchestrator. You conduct the elenctic interview, identify the user's belief burden, generate the monk prompts, spawn the Electric Monks, perform the structural analysis, and produce the synthesis. You use subagent sessions (via claude -p or your environment's equivalent) for the monks so each gets a fresh, fully committed belief context.
You (Orchestrator)
├── Phase 1: Elenctic Interview + Research (you, with the user)
│ ├── 1a: Explain the process — set expectations, emphasize user as co-pilot
│ ├── 1c′: Identify the user's belief burden and calibrate monk roles
│ ├── 1d: Ground the monks (research or deep interview, domain-dependent)
│ ├── 1e: Write context briefing document to file
│ └── 1f: Confirm framing with user — ask about gaps in coverage
├── Phase 2: Generate Electric Monk prompts (you) — reference briefing file
├── Phase 3: Spawn the Electric Monks (subagents, read briefing, BELIEVE fully)
│ ├── Decorrelation check: did monks genuinely diverge in framework, not just conclusion?
│ └── User checkpoint: "Is there evidence or a comparison class both monks missed?"
├── Phase 4: Determinate Negation (you — structural analysis, saved to file)
│ ├── 4.0: Internal tensions — where does each monk's own logic undermine itself?
│ ├── 4.5: Boydian decomposition — shatter, find cross-domain connections
│ └── 4.6: Lateral creativity (Round 2+ only) — compressed conflicts, random domain, metaphors
├── Phase 5: Sublation / Aufhebung (you — synthesis, saved to file)
│ ├── Provocation + movement (Round 2+ only) — disrupt premature pattern-matching
│ └── Abduction test: does synthesis make the original contradiction *predictable*?
├── Phase 6: Validation (Monks A & B evaluate — were they elevated or defeated?)
│ ├── Adversarial check: would the hardest-hit monk actually accept this?
│ ├── Hostile Auditor: fresh agent, strongest model, sole job is to find flaws
│ ├── Sustained juxtaposition: sometimes refusing to synthesize is the right move
│ └── Refine: present improvements individually to user, incorporate accepted ones
└── Phase 7: Recursion — propose 2-4 directions, user chooses (default: at least once)
├── Queue unexplored contradictions as the user's orientation library
└── Repeat from Phase 2 (or Phase 1 if new research needed) on chosen direction
The user can intervene at any point — correcting a monk's framing, redirecting research, rejecting a compromise-shaped synthesis. The user never has to believe anything — that's the monks' job.
CRITICAL: Before executing each phase, you MUST read its reference doc in full. The summaries below are orientation only — they do not contain the detailed instructions, prompts, templates, or failure modes you need to execute correctly. Context drift (forgetting nuance in later rounds) is the most common failure mode of this skill. Reading the reference doc fresh each time is the fix.
Read reference/phase1-elenctic-interview.md before executing.
The most important phase. Explain the process to the user. Interview them using Socratic technique to surface hidden assumptions and the deepest version of the contradiction. Identify their belief burden (see catalog below). Ground the monks via research (external domains) or deep interview (personal domains). Write a context briefing document. Confirm framing with the user — ask about gaps.
Read reference/phase2-monk-prompts.md before executing.
Generate two prompts calibrated to the user's belief burden. Each monk must BELIEVE at full conviction — this is the functional core of the ABS. The reference doc contains the required prompt structure (role, framing corrections, context briefing, research directives, argument structure, anti-hedging, length).
Read reference/phase3-spawn-monks.md before executing.
Spawn both monks as separate subagent sessions. Check for hedging, degenerate framing, and decorrelation. Present outputs to the user with guidance on how to read them. Ask if any claims should be tested against evidence neither monk considered.
Read reference/phase4-determinate-negation.md before executing.
You perform this yourself (not a subagent). Analyze internal tensions in each essay, then the surface contradiction, shared assumptions, determinate negation, hidden question, Boydian decomposition, and sublation criteria. Write your initial synthesis guess first — compare at the end to check for pattern-matching. In Round 2+, includes lateral creativity interventions: compressed conflict generation (oxymorons), random domain injection via Wikipedia's random article API, and a non-propositional pause (three metaphors).
Read reference/phase5-sublation.md before executing.
Generate the synthesis: cancel both positions as complete truths, preserve the genuine insight in each, elevate to a new concept that transforms the question. Apply the abduction test. Check for compromise failure modes — including analytical capture (adopting one monk's epistemology to reframe the other) and level reduction (dissolving a higher-category claim into lower-category terms). Present to the user before validation. In Round 2+, begins with a De Bono provocation + movement extraction to disrupt premature pattern-matching.
Read reference/phase6-validation.md before executing.
Send condensed summary to both monks for validation (elevated vs. defeated). Run adversarial check — including the proponent test (would the hardest-hit monk say "you've done exactly the thing I warned against"?). Deploy the hostile auditor (always in Round 2+, optional in Round 1). Sustained juxtaposition is a legitimate alternative when the contradiction is more productive held open than resolved. Present improvements to user one at a time, not as a list. Revise synthesis before proceeding to recursion.
Read reference/phase7-recursion.md before executing.
Recursion is the engine of the skill — the first round is calibration. Propose 2-4 directions as a menu. Fresh agents are usually better than resumed sessions. New research is often essential as each synthesis opens new conceptual domains. Default: recurse at least once. Track the dialectic queue in a file.
During the elenctic interview (Phase 1c'), pay attention to what the user is stuck believing. The dialectic's power comes from freeing the user from specific belief loads — but which beliefs need outsourcing depends on the person. Different cognitive styles produce different belief burdens, and the Electric Monks need to be calibrated accordingly.
You don't need to type the user explicitly — just notice the pattern and calibrate. Here's a catalog of common belief burdens and how they map to the monks' roles.
A note on the MBTI labels: These patterns map loosely to MBTI cognitive function stacks (Ni-Te, Ne-Ti, etc.) because the model has rich training data about those patterns — thousands of forum posts, blog articles, and discussions about how each type thinks, gets stuck, and makes decisions. The labels function as retrieval keys into that training data, not as diagnostic categories. Don't treat them as psychometric claims. Don't announce them to the user. Use them as reasoning aids to help you pattern-match what you're seeing in the interview and calibrate the monks accordingly.
The Convergent Visionary (Ni-Te pattern — common in founders, architects, CTOs)
The Empathic Integrator (Ni-Fe pattern — common in counselors, teachers, community leaders)
The Exploratory Debater (Ne-Ti pattern — common in consultants, researchers, writers)
The Practical Executor (Te-Si pattern — common in operators, managers, engineers)
The Possibility Explorer (Ne-Fi pattern — common in creatives, entrepreneurs, activists)
The Steady Guardian (Si-Fe pattern — common in administrators, caretakers, institutional maintainers)
How to use this catalog: Don't announce your typing. Don't say "I notice you're a convergent visionary." Just use the pattern to calibrate:
This calibration shapes the framing corrections in Phase 2 and the specific argument structures you assign to each monk.
Use the strongest available model with maximum thinking budget for everything. This skill operates at the edge of what models can do — perspective-taking, structural analysis, abductive reasoning, cross-domain connection. In testing, using Opus-class models for monk essays produced dramatically more insightful arguments than Sonnet-class. The monks aren't just "arguing well" — they're inhabiting positions, finding non-obvious evidence, and pushing to genuinely uncomfortable conclusions. This requires maximum capability.
| Phase | Recommended Model | Why |
|---|---|---|
| All phases | Opus/strongest available + extended thinking | Every phase benefits from maximum reasoning. The quality difference is substantial, not marginal. |
Heterogeneous models increase creativity. When possible, use different model families for Monk A and Monk B. Different training data produces different "intuitions" — different blind spots, different reasoning patterns, different default framings. This is structural decorrelation at the training-data level, which is the single most promising direction in the multi-agent debate literature (Du et al., ICLR 2025). The orchestrator should remain your strongest available model (it needs maximum synthesis capability), but monks benefit from heterogeneity.
Before starting, check what's available. If you're running in an environment with access to multiple coding agents or model providers, ask the user:
I can increase the creativity of the dialectic by using different AI models for each monk — different training data means genuinely different blind spots and reasoning patterns. Do you have access to any of these I could use for one of the monks?
- Gemini (via
geminiCLI or API)- GPT-4 / ChatGPT (via
codexCLI or API)- Other model providers
If not, I'll use the same model family for both monks — the skill works fine either way, the decorrelation just comes from the different prompts and belief commitments rather than from different training data.
If heterogeneous models aren't available, don't worry — the skill is designed to work with homogeneous models. The framing corrections, belief burden calibration, and targeted research directives already produce substantial decorrelation. Heterogeneous models are a bonus, not a requirement.
Based on three test runs across different domains (normative/institutional, business strategy, political economy of OSS):
External-research domains:
| Phase | Typical Range | Notes |
|---|---|---|
| Phase 1 research (2-3 parallel agents) | 150-250K tokens | Do NOT cut here. This is the highest-value spend. Broader domains trend higher. |
| Phase 1 supplementary research (user-triggered) | 0-50K tokens | Common — users frequently identify gaps. Budget for it. |
| Phase 1d briefing synthesis | ~5K tokens | Orchestrator work |
| Phase 3 monk essays (with briefing) | 25-45K tokens | Two monks, 2-3 targeted searches each |
| Phase 4-5 analysis + synthesis | 15-30K tokens | Orchestrator inline work |
| Phase 6 monk validation | 12-25K tokens | Two monks, strongest model |
| Phase 6 hostile auditor | 5-15K tokens | One agent, strongest model. Reads essays + synthesis only. |
| Phase 7 recursive round | 25-50K tokens | Often most valuable |
| Orchestrator overhead | 20-30K tokens | Interview, transitions, presentation |
| Total (one round + recursion) | ~300-400K tokens | Median ~300K without supplementary research |
Personal/values domains are significantly cheaper on research but more expensive on interview:
| Phase | Typical Range | Notes |
|---|---|---|
| Phase 1 extended interview | 15-30K tokens | 6-10 exchanges, deeper probing |
| Phase 1 framework research (optional) | 0-50K tokens | Frameworks, not facts. May be skipped. |
| Phase 1d context briefing | ~5K tokens | User-sourced material synthesized |
| Phase 3 monk essays | 15-30K tokens | Monks may need zero additional searches |
| Remaining phases | Similar to above | |
| Total (one round + recursion) | ~100-200K tokens | Much cheaper — the user's testimony is the primary input |
Key insight: For external domains, Phase 1 research is the highest-value spend. For personal domains, Phase 1 interview depth is the highest-value spend — the monks can only believe as specifically as the briefing allows.
This skill is written around claude -p (pipe mode) for spawning subagents. If you're running in Claude Code using the Task tool, here are the key differences:
| Skill instruction | claude -p | Claude Code Task tool |
|---|---|---|
| Spawn subagent | echo "[PROMPT]" | claude -p > output.md | Task(prompt, subagent_type="general-purpose") |
| Parallel execution | Background shell jobs | run_in_background=true |
| Output to file | Shell redirect (> file.md) | Agent returns text; orchestrator writes files |
| Session resumption (Phase 6) | Resume same claude -p session | resume parameter with agentId — but persona may not persist without reinforcement. Include a summary of the agent's original argument as fallback. |
| Model selection | --model flag | model parameter (defaults to inheriting from parent) |
| Tool access | --allowedTools web_search,web_fetch | Inherits from parent or configure per-task |
Key difference: With claude -p, agents write output directly to files via shell redirect. With the Task tool, agents return text to the orchestrator, who writes files. This adds a step but gives the orchestrator control over file naming and structure. Either approach works — just be aware that the file I/O pattern differs.
Session resumption for validation: The skill prefers resuming original agent sessions so validators retain their full conviction context. In Claude Code, this works via resume + agentId, but test runs found the persona sometimes needs reinforcement. The fallback — a fresh validation prompt that includes a summary of the agent's original argument — works well in practice.
The dialectic structure is universal but the vocabulary of "truth" and the grounding mode vary by domain. Adapt accordingly:
| Domain Type | What "Truth" Means | Good Synthesis Looks Like | Grounding Mode | Aporia (productive perplexity) Valid? |
|---|---|---|---|---|
| Empirical (engineering, science) | What works, performs, is maintainable | Testable decision criteria, architectural patterns | External research | Rarely |
| Normative (ethics, politics, policy) | What's defensible, respects competing values | Tension map with navigation strategies | Mixed (research + user values) | Yes |
| Personal (life decisions, career) | What aligns with actual priorities | Values clarification — what you actually want | Deep interview (user is the source) | Yes |
| Creative (writing, design, art) | What's interesting, resonant, surprising | Unexpected recombinations, new possibilities | Mixed (research + user aesthetic) | Sometimes |
| Risk Analysis | Actual risk structure behind competing assessments | Decision framework calibrated to real uncertainties | External research | No |
Read this section to understand WHY the process works the way it does. This informs your judgment when things go off-script. The frameworks are listed in order of operational importance — Rao explains what the tool is, Hegel explains how contradictions resolve, Boyd explains how creativity works, Socrates explains how to surface the question, Adams gives the metaphor, Aquinas gives the aspiration, and DeLong explains when to use it.
The foundational theory for this skill comes from Venkatesh Rao's "Electric Monks" framework (after Douglas Adams' Dirk Gently). The core distinction: this tool is not artificial intelligence — it is an artificial belief system (ABS). The agents aren't thinking for you. You're still doing the thinking (orchestrating, judging, choosing directions, recognizing genuine sublation vs. compromise). The agents are believing for you.
Why belief is the bottleneck: The central transaction cost in human cognition is context-switching cost — what Boyd calls the "transient." The length of the transient depends on how much belief inertia you're carrying. Once you believe a position, switching to genuinely entertaining its negation is expensive. You hedge, you steelman weakly, you unconsciously bias. The Electric Monks eliminate this cost by carrying 100% of the belief load, freeing the user to operate as a pure context-switching specialist — what Rao calls "informationally tiny."
The F-86 analogy (from Boyd via Rao): In the Korean War, F-86 Sabres achieved a 10:1 kill ratio against MIG-15s despite roughly matched flight capabilities. Boyd discovered the difference was hydraulic controls — the F-86 pilot could reorient faster because the plane did more of the mechanical work. The pilot's freed-up attention went to choosing better maneuvers, not just executing them faster. The Electric Monks are hydraulic controls for intellectual work: by doing the belief-work, they free the user's attention for the higher-order task of structural analysis and creative synthesis.
Operational implications for this skill:
Anti-hedging is a functional requirement, not a stylistic preference. A hedging monk is an Electric Monk that has failed at its one job. If it doesn't fully believe, the user has to pick up the dropped belief weight, their transients slow, and they lose the belief-free orchestrator position.
Validation checks for elevation, not agreement. A defeated monk has dropped its belief load — belief was destroyed rather than transformed. A properly elevated monk believes more — it sees its original position as partial truth within a larger truth. The ABS should always be carrying belief; the synthesis just changes what it carries.
Recursion trains transient speed. Each cycle is a full reorientation: commit (via monks) → shatter (via Boyd) → reconnect (via Hegel) → commit to the new thing (via monks again). Seven cycles in an hour = seven reorientations with zero belief inertia. Over time, the user may internalize this reorientation capacity — the mechanical monk as transitional object.
The branching queue is an orientation library. Each deferred contradiction is a pre-positioned reorientation the user can snap into. The richer the queue, the more agile the user's subsequent thinking — even outside the tool — because they know the monks are holding those positions for them.
Validate the user's dominant mode first. If the user has to defend their existing position, they've taken on belief weight. Monk A's first job is to validate the user's instinct so thoroughly that they can release it — let the monk carry it — and operate from the belief-free orchestrator seat.
The engine of the dialectic is determinate negation — not "this is wrong" but "this is wrong in a specific way that points toward what's missing." The specific way a position fails contains a signpost toward the richer understanding needed.
Sublation (Aufhebung) simultaneously cancels, preserves, and elevates. It is NOT compromise (splitting the difference). It produces something neither party could have conceived independently but which, once articulated, both recognize as more complete. It is irreversible — genuine cognitive gain. The Kant example: the rationalism/empiricism debate wasn't resolved by "knowledge comes half from reason and half from experience" but by "experience provides content, reason provides structure." After Kant, you can't go back.
Hegel never used "thesis-antithesis-synthesis" — that framing comes from Fichte. The actual movement is driven by the one-sidedness of each concept, which generates its own negation internally.
John Boyd's "Dialectic Engine": destructive deduction (shatter existing conceptual domains, scatter parts into a "sea of anarchy") followed by creative induction (find cross-domain connections to synthesize something new).
Boyd's critical insight: you cannot synthesize something genuinely new by recombining within the same domain. If Monk A and Monk B are both arguing about web frameworks, a synthesis that only recombines claims from their two essays will produce rearrangement, not creation. Genuine novelty requires material from outside the original conceptual domains. The destructive step — separating particulars from their previous wholes — creates space for outside material to enter and form new connections.
Boyd's cycle: Structure → Unstructure → Restructure → repeat at higher levels of elaboration.
Where Boyd is operationally present: Phase 4.5 (Boydian Decomposition — the destructive step), Phase 5 (Sublation — the creative step requiring cross-domain connection), and Phase 7 (Recursion — each cycle is Boyd's full Structure → Unstructure → Restructure, which is why recursive rounds often need new research from outside the original domains).
Relationship to Hegel: Hegel provides the engine for analyzing how positions fail (determinate negation) and the concept of what good synthesis looks like (Aufhebung). Boyd provides the engine for what to do with the wreckage — shatter, scatter, and recombine with outside material. The two frameworks are complementary: Hegel drives the contradiction analysis, Boyd drives the creative reconstruction.
The elenctic method probes a position through questioning to expose contradictions and reach aporia (productive perplexity). Not adversarial but cooperative — "midwifery of ideas." The interview phase of this skill is elenctic. Aporia is sometimes a valid outcome.
Key findings from Du et al. (2023, MIT) and subsequent work through ICLR 2025:
Elizabeth Eisenstein argued that print's most transformative effect was typographic fixity — enabling scholars to lay texts side by side and detect contradictions. LLMs represent the next step: fixity + comparison + structural contradiction analysis partially automated. This skill exploits that transition.
Douglas Adams' Electric Monk (Dirk Gently's Holistic Detective Agency) is a labor-saving device designed to believe things for you. The one in the story has "developed a fault" — it believes too many irrational things. In this skill, the "fault" is the feature. Each monk is designed to believe a specific position at full conviction that the user cannot hold simultaneously. The monks are not thinking for the user — they are believing for the user, which is what frees the user to think.
"The slenderest knowledge that may be obtained of the highest things is more desirable than the most certain knowledge obtained of lesser things."
This is the philosophical aspiration of the entire process. The dialectic does not produce certainty — every synthesis is provisional, fertile, pointing toward a deeper contradiction. But that slender, provisional knowledge of the deep structure (why this tension exists, what hidden question drives it, what shared assumption both sides are trapped in) is worth more than confident knowledge of the surface question ("which option should I pick?").
Operational implications:
Aquinas practiced the Disputatio — structured scholastic debate where committed advocates argued positions before a master who synthesized. The Electric Monks are his disputing friars, mechanized.
Brad DeLong's "Cognitive Distributed Disruption of Attention Crisis" (2026) frames the problem this skill addresses: the volume of plausible, credentialed output now exceeds any serious person's cognitive bandwidth. His solution is defensive intellectual infrastructure — ruthless triage, model-updating as the frame for reading, information portfolio management.
This skill is the offensive complement. DeLong's triage decides what deserves deep engagement. The Electric Monks provide the method for that engagement — they're what you reach for when you've found a genuine contradiction that can't be resolved by reading one more article, watching one more talk, or skimming one more summary.
Operational implication: The skill should not be used for everything. It's expensive (time, tokens, cognitive effort). Use it at DeLong's Level 4-5 — when the stakes justify deep engagement, when the tension is genuine and not resolvable by more information, when you need a model update rather than more data. The elenctic interview (Phase 1) should filter for this: if the question can be answered by looking it up, this is the wrong tool.
Charles Sanders Peirce identified three modes of inference: deduction (from rule to consequence), induction (from cases to rule), and abduction (from surprising fact to explanatory hypothesis). The synthesis phase is abductive: given the surprising fact that both monk positions exist and each has genuine evidence, what hypothesis would make this unsurprising? Peirce's typology of abduction (selective → conditional-creative → propositional-conditional-creative) maps to synthesis quality — the best syntheses introduce genuinely new concepts, not just new arrangements of known ones. Operationally present in Phase 5 (Abduction Test).
John Pollock's epistemology distinguishes undercutting defeaters (the inferential link is broken — reasons to doubt the connection between evidence and conclusion) from rebutting defeaters (evidence directly supporting the opposite conclusion). Undercutting is more structurally revealing because it identifies how reasoning fails, not just that it fails — parallel to determinate negation's "specific way of failing." Pollock also identifies self-defeating arguments (conclusions that undermine their own premises), which should be rejected outright. Operationally present in the hostile auditor prompt (Phase 6).
Adam Galinsky's research shows that perspective-taking (cognitively inhabiting another's viewpoint) outperforms advocacy (arguing for a position) at both conscious and nonconscious levels. The mechanism is self-other overlap — when you inhabit a position rather than argue for it, you access richer associative networks and produce higher-quality elaboration. This is the psychological basis for the Electric Monk's "you ARE this position" instruction — inhabiting produces deeper arguments than advocating. Operationally present in the monk prompt template (Phase 2).
Gary Klein's research shows that imagining a future failure has already occurred increases the ability to identify causes of that failure by ~30%, compared to asking "what could go wrong?" The temporal reframing ("it already happened, why?") breaks selective accessibility — the cognitive tendency to search only for confirming evidence. Operationally present in the hostile auditor prompt (Phase 6).
Gilles Fauconnier and Mark Turner's theory of conceptual blending provides the machinery for understanding how genuinely new concepts emerge. A blend's value is measured by its emergent structure — organizational properties that exist in neither input space. The skill's Boydian decomposition is the destructive step (creating input spaces), and sublation is the blend (which must have emergent structure to be genuine). The "generic space" — the abstract relational structure shared by both inputs — often reveals the shared assumption the synthesis must transcend. Operationally present in Phase 4.5.
Wood et al. (JMLR 2023) formalize why monk independence is load-bearing: the bias-variance-diversity decomposition shows diversity is literally subtracted from ensemble error (E[loss] = noise + avg_bias + avg_variance − diversity). Correlated errors eliminate the diversity benefit entirely. This is why monks must be spawned in separate sessions with no shared context, and why heterogeneous model families (when available) increase the skill's creative output. Surowiecki's wisdom-of-crowds conditions confirm: independence is necessary, not optional. Operationally present in the decorrelation check (Phase 3) and heterogeneous model guidance.
SICP's core thesis — that managing complexity requires modularization, abstraction barriers, and composition of simple components — mirrors this skill's architecture. Each phase is a module with defined inputs and outputs. Agents are spawned in isolated environments (SICP's environment model) to prevent information leakage. The auditor deliberately can't see the orchestrator's Phase 4 analysis — an abstraction barrier, not an oversight.
Most relevant is SICP's closure property: a means of combination has closure when the result can itself be combined using the same means. The dialectic has closure — a synthesis is itself a valid position that can serve as input to the next round. This is why recursion works: the output type equals the input type. When closure breaks (a synthesis so abstract or meta that no monk could believe it at full conviction), recursion stalls. This is a diagnosable failure mode — if you can't hand the synthesis to a monk and have it argue from that position, the synthesis lacks closure and needs to be made more concrete.
Chris Dixon (via Balaji Srinivasan): a good founder doesn't just have an idea — they can navigate the idea maze, anticipating which turns lead to treasure and which to dead ends. The maze is mapped through history (what did previous attempts get right and wrong?), analogy (what did similar efforts in adjacent domains do?), theories (what generalizable patterns exist?), and direct experience.
The dialectic queue is an idea maze. Each synthesis opens new paths (contradictions). The user chooses which to explore. Unexplored paths remain visible in the queue — a map of the territory showing where you've been, where you could go, and what remains open. The research phase (Phase 1d) maps directly to Dixon's four sources: history of the domain, analogies to adjacent domains, theoretical frameworks, and the user's own direct experience (surfaced in the elenctic interview). The skill doesn't just navigate the maze — each recursive round reveals new corridors that weren't visible from the entrance.
Christopher Alexander (1965) showed that natural cities have semi-lattice structure — overlapping, cross-connected sets — while designed cities impose tree structure where every element belongs to exactly one branch. Trees are easier to think about but destroy the cross-connections that make systems alive. Every attempt to design semi-lattices directly (Alexander's own HIDECS, Holacracy, Spotify's squad model) collapses back to trees because the design substrate — whether graph partitioning algorithms, org charts, or natural language — is tree-biased.
This skill is a semi-lattice compiler. Language is tree-structured (Chomsky's generative grammar, dependency parsing, sequential token generation). Each monk produces a tree — a coherent linear argument from committed premises to conclusions. Monk B in any dialectic is always right that its output is a tree. But the Boydian decomposition phase (Phase 4.5) strips both arguments of their source, extracts atomic parts, and finds cross-connections between elements that came from different trees. These cross-domain connections ARE the semi-lattice edges. The synthesis is the semi-lattice that emerges from the overlap of multiple trees.
The answer to "language can't represent semi-lattices" is not "make the LLM output a semi-lattice directly." It's: produce multiple trees from different committed positions, then extract the cross-connections. The semi-lattice is constructed, not generated. Every successful semi-lattice system works this way — Gene Ontology (multiple studies cross-referenced into a DAG), McChrystal's Team of Teams (tree-structured teams with liaison officers creating cross-connections), Ostrom's polycentric governance (overlapping jurisdictions, not one hierarchy).
Study these to understand the level of specificity, framing correction, and structural craft the skill requires. The key lessons are at the end.
User's surface framing: "Should I use TanStack Start or Next.js?"
Degenerate framing the orchestrator must avoid: "Libraries vs frameworks" or "modular vs monolithic." This is the boring version — the contradiction isn't deep enough.
Deepest contradiction found (via research): Infrastructure sovereignty and incentive alignment vs. deep framework-infrastructure integration and commercially-sustained ambition.
Key framing correction in Monk A's prompt:
"TanStack Start IS a framework — it has opinions about routing, server functions, SSR, and application architecture. Your argument is NOT that TanStack Start is more modular or 'just libraries' while Next.js is a monolith. Both are opinionated frameworks. The real difference lies elsewhere."
Key framing correction in Monk B's prompt:
"Your opponent's argument is NOT the naive 'libraries vs frameworks' take. They will argue that Next.js's design is structurally compromised by Vercel's commercial interests. You need to engage this argument directly, not dismiss it."
Research directives (targeted, not broad):
Ontological question driving both prompts: "What is the proper relationship between a framework, the infrastructure it runs on, and the business interests that fund its development?"
User's surface framing: "I'm torn between taking this promotion and being more present for my kids."
Degenerate framing: "Work-life balance." This flattens a structural tension into a scheduling problem.
Deepest contradiction found (via extended interview): The user doesn't just want both — they believe being the kind of person who excels at work is inseparable from being the kind of parent they want to be. The tension is identity-level, not time-allocation.
Key framing correction in Monk A's prompt:
"Your argument is NOT that career success matters. It's that THIS USER'S specific professional identity — what they build, how they lead, what they model for their children — is itself an act of parenting. Ground this in their actual history: [specific examples from interview]."
Key framing correction in Monk B's prompt:
"Your argument is NOT that family time matters. It's that presence has a developmental window that closes — and that the user's children at ages [X] need [specific things from interview] that no amount of 'quality time' can compress into fewer hours."
No external research needed. The briefing was built entirely from the elenctic interview — the user's history, their children's ages and needs, their partner's actual capacity, the specific role being offered.
This example shows how recursion pulls in cross-domain material — Boyd's prediction in action:
The original question has nothing to do with jurisprudence or Gödel — but by Cycle 4 the dialectic had evolved to where those concepts were essential. Each synthesis opens doors to domains the previous round couldn't see.
The final deliverable should include:
The Dialectical Trace — the full journey, not just the destination:
The Model Update — explicit statement of what changed:
Actionable Output (domain-dependent):
The Dialectic Queue — a map of the intellectual territory:
Write these as markdown files in the output directory. Include a README.md or index.md linking all output files in order so the full dialectical trace is navigable. The queue file (dialectic_queue.md) serves as both a session artifact and a starting point for future sessions.