Full research workflow entry point. Sequences skills like solver's /solve.
/research <question>The master workflow that sequences the full research lifecycle. Like solver's
/solve command — a structured entry point that coordinates all other skills.
Tracks stages, handles branching, and always finalizes (updates plan + log).
1. SCOPE — Define the question precisely
2. HYPOTHESIZE — Generate approaches (hypothesis-generation skill)
3. DISCOVER — Find relevant papers (paper-discovery skill)
4. ANALYZE — Deep-read top papers (paper-analysis skill)
5. SYNTHESIZE — Combine findings into themes
6. VERIFY — Verify key claims (verification-cascade skill)
7. CRITIQUE — Self-evaluate the output (self-critique skill)
8. FINALIZE — Update plan, write log, save deliverable
Read research-plan.md if it exists. Ask the user:
State the scoped question in one sentence.
Load and follow the hypothesis-generation skill:
files_cat("/global/skills/hypothesis-generation.md")
Generate ≥5 approaches, quick-test top 3, recommend one with fallback. De-risk: test the riskiest assumption of the recommended approach first.
Load and follow paper-discovery:
files_cat("/global/skills/paper-discovery.md")
Use the recommended approach from Stage 2 to guide search strategy.
Branch: If discovery finds <3 relevant papers:
files_cat("/global/skills/pivot.md")Load and follow paper-analysis:
files_cat("/global/skills/paper-analysis.md")
Analyze top 3-5 papers from discovery. For more papers, use parallel-investigation:
files_cat("/global/skills/parallel-investigation.md")
Branch: If a paper contradicts the working hypothesis:
Read all analysis notes. Identify themes, consensus, contested points, and gaps. Organize by theme, not by paper.
This is the hardest stage. Watch for regression to the mean — your synthesis should be genuinely insightful, not a summary of summaries.
Load verification-cascade for key synthesis claims:
files_cat("/global/skills/verification-cascade.md")
Apply L2+ to all claims that will appear in the final output. Apply L3 (adversarial) to the central finding.
Load self-critique:
files_cat("/global/skills/self-critique.md")
Evaluate the entire output. Target ≥8/10 on all dimensions. If below threshold: revise and re-critique (max 2 iterations).
files_cat("/global/skills/research-log.md")
The orchestrator doesn't produce a single output file. Its outputs are:
research/discovery/<topic>research/papers/<id> per paperresearch/reviews/<topic>Input: /research "How does RLHF compare to DPO for LLM alignment?"
Execution:
/research <question>