Reviews a game design document for completeness, internal consistency, implementability, and adherence to project design standards. Run this before handing a design document to programmers.
Extract --depth [full|lean|solo] if present. Default is full when no flag is given.
Note: --depth controls the analysis depth of this skill (how many specialist agents are spawned). It is independent of the global review mode in production/review-mode.txt, which controls director gate spawning. These are two different concepts — --depth is about how thoroughly this skill analyses the document.
full: Complete review — all phases + specialist agent delegation (Phase 3b)lean: All phases, no specialist agents — faster, single-session analysissolo: Phases 1-4 only, no delegation, no Phase 5 next-step prompt — use when called from within another skillRead the target design document in full. Read CLAUDE.md to understand project context and standards. Read related design documents referenced or implied by the target doc (check for related systems).
design/gdd/Dependency graph validation: For every system listed in the Dependencies section, use Glob to check whether its GDD file exists in design/gdd/. Flag any that don't exist yet — these are broken references that downstream authors will hit.
Lore/narrative alignment: If design/gdd/game-concept.md or any file in design/narrative/ exists, read it. Note any mechanical choices in this GDD that contradict established world rules, tone, or design pillars. Pass this context to game-designer in Phase 3b.
Prior review check: Check whether design/gdd/reviews/[doc-name]-review-log.md exists. If it does, read the most recent entry — note what verdict was given and what blocking items were listed. This session is a re-review; track whether prior items were addressed.
Evaluate against the Design Document Standard checklist:
Internal consistency:
Implementability:
Cross-system consistency:
Skip this phase in lean or solo mode.
This phase is MANDATORY in full mode. Do not skip it.
Before spawning any agents, print this notice:
"Full review: spawning specialist agents in parallel. This typically takes 8–15 minutes. Use
--review leanfor faster single-session analysis."
Read the GDD and identify every domain present. A GDD can touch multiple domains simultaneously — be thorough. Common signals:
| If the GDD contains... | Spawn these agents |
|---|---|
| Costs, prices, drops, rewards, economy | economy-designer |
| Combat stats, damage, health, DPS | game-designer, systems-designer |
| AI behaviour, pathfinding, targeting | ai-programmer |
| Level layout, spawning, wave structure | level-designer |
| Player progression, XP, unlocks | economy-designer, game-designer |
| UI, HUD, menus, player-facing displays | ux-designer, ui-programmer |
| Dialogue, quests, story, lore | narrative-director |
| Animation, feel, timing, juice | gameplay-programmer |
| Multiplayer, sync, replication | network-programmer |
| Audio cues, music triggers | audio-director |
| Performance, draw calls, memory | performance-analyst |
| Engine-specific patterns or APIs | Primary engine specialist (from .claude/docs/technical-preferences.md) |
| Acceptance criteria, test coverage | qa-lead |
| Data schema, resource structure | systems-designer |
| Any gameplay system | game-designer (always) |
Always spawn game-designer and systems-designer as a baseline minimum. Every GDD touches their domain.
CRITICAL: Task in this skill spawns a SUBAGENT — a separate independent Claude session with its own context window. It is NOT task tracking. Do NOT simulate specialist perspectives internally. Do NOT reason through domain views yourself. You MUST issue actual Task calls. A simulated review is not a specialist review.
Issue all Task calls simultaneously. Do NOT spawn one at a time.
Prompt each specialist adversarially:
"Here is the GDD for [system] and the main review's structural findings so far. Your job is NOT to validate this design — your job is to find problems. Challenge the design choices from your domain expertise. What is wrong, underspecified, likely to cause problems, or missing entirely? Be specific and critical. Disagreement with the main review is welcome."
Additional instructions per agent type:
game-designer: Anchor your review to the Player Fantasy stated in Section B of this GDD. Does this design actually deliver that fantasy? Would a player feel the intended experience? Flag any rules that serve implementability but undermine the stated feeling.
systems-designer: For every formula in the GDD, plug in boundary values (minimum and maximum plausible inputs). Report whether any outputs go degenerate — negative values, division by zero, infinity, or nonsensical results at the extremes.
qa-lead: Review every acceptance criterion. Flag any that are not independently testable — phrases like "feels balanced", "works correctly", "performs well" are not ACs. Suggest concrete rewrites for any that fail this test.
After all specialists respond, spawn creative-director as the senior reviewer:
If specialists disagree with each other or with the creative-director, do NOT silently pick one view. Present the disagreement explicitly in Phase 4 so the user can adjudicate.
Mark every finding with its source: [game-designer], [economy-designer], [creative-director] etc.
## Design Review: [Document Title]
Specialists consulted: [list agents spawned]
Re-review: [Yes — prior verdict was X on YYYY-MM-DD / No — first review]
### Completeness: [X/8 sections present]
[List missing sections]
### Dependency Graph
[List each declared dependency and whether its GDD file exists on disk]
- ✓ enemy-definition-data.md — exists
- ✗ loot-system.md — NOT FOUND (file does not exist yet)
### Required Before Implementation
[Numbered list — blocking issues only. Each item tagged with source agent.]
### Recommended Revisions
[Numbered list — important but not blocking. Source-tagged.]
### Specialist Disagreements
[Any cases where agents disagreed with each other or with the main review.
Present both sides — do not silently resolve.]
### Nice-to-Have
[Minor improvements, low priority.]
### Senior Verdict [creative-director]
[Creative director's synthesis and overall assessment.]
### Scope Signal
Estimate implementation scope based on: dependency count, formula count,
systems touched, and whether new ADRs are required.
- **S** — single system, no formulas, no new ADRs, <3 dependencies
- **M** — moderate complexity, 1-2 formulas, 3-6 dependencies
- **L** — multi-system integration, 3+ formulas, may require new ADR
- **XL** — cross-cutting concern, 5+ dependencies, multiple new ADRs likely
Label clearly: "Rough scope signal: M (producer should verify before sprint planning)"
### Verdict: [APPROVED / NEEDS REVISION / MAJOR REVISION NEEDED]
This skill is read-only — no files are written during Phase 4.
Use AskUserQuestion for ALL closing interactions. Never plain text.
First widget — what to do next:
If APPROVED (first-pass, no revision needed), proceed directly to the systems-index widget, review-log widget, then the final closing widget. Do not show a separate "what to do" widget — the final closing widget covers next steps.
If NEEDS REVISION or MAJOR REVISION NEEDED, options:
[A] Revise the GDD now — address blocking items together[B] Stop here — revise in a separate session[C] Accept as-is and move on (only if all items are advisory)If user selects [A] — Revise now:
Work through all blocking items, asking for design decisions only where you cannot resolve the issue from the GDD and existing docs alone. Group all design-decision questions into a single multi-tab AskUserQuestion before making any edits — do not interrupt mid-revision for each blocker individually.
After all revisions are complete, show a summary table (blocker → fix applied) and use AskUserQuestion for a post-revision closing widget:
[A] Re-review in a new session — run /design-review [doc-path] after /clear[B] Accept revisions and mark Approved — update systems index, skip re-review[C] Move to next system — /design-system [next-system] (#N in design order)[D] Stop hereNever end the revision flow with plain text. Always close with this widget.
Second widget — systems index update (always show this separately):
Use a second AskUserQuestion:
design/gdd/systems-index.md to mark [system] as [In Review / Approved]?"[A] Yes — update it / [B] No — leave it as-isThird widget — review log (always offer):
Use a third AskUserQuestion:
design/gdd/reviews/[doc-name]-review-log.md? This creates a revision history so future re-reviews can track what changed."[A] Yes — append to review log / [B] No — skipIf yes, append an entry in this format:
## Review — [YYYY-MM-DD] — Verdict: [APPROVED / NEEDS REVISION / MAJOR REVISION NEEDED]
Scope signal: [S/M/L/XL]
Specialists: [list]
Blocking items: [count] | Recommended: [count]
Summary: [2-3 sentence summary of key findings from creative-director verdict]
Prior verdict resolved: [Yes / No / First review]
Final closing widget — always show after all file writes complete:
Once the systems-index and review-log widgets are answered, check project state and show one final AskUserQuestion:
Before building options, read:
design/gdd/systems-index.md — find any system with Status: In Review or NEEDS REVISION (other than the one just reviewed).md files in design/gdd/ (excluding game-concept.md, systems-index.md) to determine if /review-all-gdds is worth offering (≥2 GDDs)Build the option list dynamically — only include options that are genuinely next:
[_] Run /design-review [other-gdd-path] — [system name] is still [In Review / NEEDS REVISION] (include if another GDD needs review)[_] Run /consistency-check — verify this GDD's values don't conflict with existing GDDs (always include if ≥1 other GDD exists)[_] Run /review-all-gdds — holistic design-theory review across all designed systems (include if ≥2 GDDs exist)[_] Run /design-system [next-system] — next in design order (always include, name the actual system)[_] Stop hereAssign letters A, B, C… only to included options. Mark the most pipeline-advancing option as (recommended).
Never end the skill with plain text after file writes. Always close with this widget.