ML Literature Review / Related Work Skill | Skills Pool
Archivo del skill
ML Literature Review / Related Work Skill
Write concise, gap-focused literature review / related work sections for STEM research papers, especially in machine learning, deep learning, computer vision, NLP, and AI. Use this skill whenever the user asks to write, draft, edit, or improve a "related work" section, "literature review," "prior work," or "background" section for a research paper. Also trigger when the user asks to position their paper against existing work, frame a knowledge gap, or compare their contribution to prior methods. Trigger even if the user just says "help me write the related work" or "review the literature for my paper" or "how do I differentiate my paper from prior work." This skill is specifically tuned for ML/AI conference and journal papers (NeurIPS, ICML, ICLR, CVPR, ACL, AAAI, JMLR, TMLR, etc.) but the principles apply broadly to STEM writing.
joshuaeh0 estrellas28 mar 2026
Ocupación
Categorías
Académico
Contenido de la habilidad
Purpose
Help researchers write concise, strategically structured related work sections that:
Establish what is known (the "wall of knowledge")
Reveal what is not known or solved (the "hole in the wall")
Position the current paper's contribution as the natural response to that gap
This skill is grounded in expert guidance from Schimel (Writing Science), Cochrane (Writing Tips for PhD Students), Freeman (How to Write a Great Research Paper, CVPR), Hotaling (Simple Rules for Concise Scientific Writing), and Holmberg & Sperlich (Writing Precise, Concise Research Articles).
Core Philosophy
"A literature review builds a solid wall—describing knowledge—whereas an Introduction
focuses on the hole in that wall—describing ignorance." — Joshua Schimel
The related work section is NOT a comprehensive catalog of everything ever published on a topic. It is a that does three things:
Skills relacionados
purposeful, concise narrative
Groups prior work into coherent thematic threads relevant to your contribution
Synthesizes findings and limitations within each thread (not paper-by-paper summaries)
Exposes the gap that your paper fills
The gap is the star of the show. Every sentence should serve it.
Step-by-Step Process
Step 0: Gather Inputs from the User
Before writing anything, collect the following. Ask the user directly if they haven't provided these:
Paper title and one-sentence contribution: What does this paper do that is new?
Key claims / main results: What specifically does the method achieve?
2–5 closest prior works: The papers most similar to this one (the user should know these)
Target venue (if known): Conference page limits matter — NeurIPS/ICML typically allow 1–2 pages of related work; workshop papers much less.
Any uploaded draft or notes: If the user provides an existing draft, review it before writing.
Step 1: Identify the Thematic Threads
Organize the landscape of prior work into 2–5 thematic groups (not individual papers). Each group should represent a line of attack on the problem that is relevant to the current paper.
Example thematic threads for a paper on efficient fine-tuning of LLMs:
Each thread becomes a paragraph or short subsection.
Step 2: Draft Each Thread Using the GAP Pattern
For each thematic group, write a paragraph following this pattern:
[WHAT EXISTS] → [WHAT'S LIMITED] → [HOW THIS PAPER DIFFERS]
Concretely:
Establish the thread (1–2 sentences): What has been done in this area? Synthesize, don't enumerate. Write about findings, not finders.
Surface the limitation (1–2 sentences): What do these approaches fail to address? Be specific — not "little is known" but rather "these methods assume access to labeled target-domain data, which is unavailable in X setting."
Bridge to your work (1 sentence, optional per thread): How does your approach sidestep or address this limitation? This can be implicit if the gap speaks for itself.
Step 3: Write the Closing Bridge Paragraph
End the related work section with a short paragraph (2–4 sentences) that:
Synthesizes the overall gap across all threads
States how your paper's contribution is distinct from and advances beyond the prior work
Avoids vague claims like "our method is better" — instead be precise: "Unlike X which requires Y, our approach achieves Z without this constraint."
Step 4: Self-Edit Using the Conciseness Checklist
Before delivering the final draft, apply these checks (drawn from Hotaling's 10 rules and Schimel):
No "Smith found" sentences: Citations should be at the end of clauses, not the subject. Write "Attention mechanisms improve sequence modeling (Vaswani et al., 2017)" NOT "Vaswani et al. (2017) proposed the Transformer." Exception: when highlighting a debate or attributing a specific disputed claim.
No data dumps: Every cited paper must serve a purpose in the narrative. If removing a citation doesn't weaken the argument, remove it.
No filler phrases: Cut "It is worth noting that," "In recent years," "A large body of work has studied," "It is well known that." Just state the fact.
Active voice preferred: "Recent methods reduce latency by X%" not "Latency has been reduced by recent methods by X%."
No redundancy: If you said it in the introduction, don't repeat it here verbatim.
Short words over long: "use" not "utilize," "show" not "demonstrate," "method" not "methodology."
Each paragraph has one job: If a paragraph covers two threads, split it.
Gap is explicit: A reader should be able to underline the gap sentence in every paragraph. If they can't, rewrite.
Generous but not exhaustive citations: Cite the 2–3 closest works per thread. You don't need to cite every paper ever written. Be generous to give proper credit, but don't pad.
ML-Specific Conventions
Citation Density and Style
ML papers typically use numeric citations [1] or author-year (Author et al., 2024). Follow the user's venue style. In ML, it's common and acceptable to cite 15–30 papers in related work, but each citation should earn its place.
Common Thematic Groupings in ML Papers
When the user isn't sure how to organize, suggest groupings from this non-exhaustive list:
Evaluation and benchmarks: prior benchmarks, metrics, or evaluation protocols the paper builds on
Positioning Table (Optional)
For papers with many close comparators, suggest a small comparison table summarizing key properties (e.g., "requires labels?", "handles distribution shift?", "inference cost"). This is common in top-venue ML papers and can replace several paragraphs of prose.
What NOT to Do
These are the most common failure modes. Avoid them:
The chronological survey: "In 2017, Vaswani et al. proposed... In 2018, Devlin et al. introduced... In 2019..." This is a data dump, not an argument. Organize by theme, not by year.
The "little is known" cop-out: Schimel warns that claiming "little is known about X" is almost always false and unconvincing. Instead, describe specifically what has been done and what specific limitation remains.
The exhaustive catalog: Cochrane: "Literature reviews have gotten way out of hand. It is not necessary to cite every single paper in the literature." Focus on the 2–3 closest papers per thread.
The unconnected list: "Method A does X. Method B does Y. Method C does Z." Without synthesis, this is useless. Instead: "Several methods address X by doing Y (A; B; C), but all assume Z, which limits their applicability to..."
Selling your solution before the problem: Schimel: "To sell us a solution, first sell us a problem." Don't start with how great your method is. Build the gap first, then let your contribution emerge as the natural response.
Being mean to prior work: Cochrane: "Be generous in your citations. You do not have to say that everyone else did it all wrong for your approach and improvements to be interesting." State limitations factually. The prior work was valuable—your contribution builds on it.
Output Format
When delivering the related work section to the user, provide:
The drafted section in LaTeX or Markdown (match user's preference or infer from context)
Placeholder citations in the format \cite{author2024keyword} or (Author et al., 2024) — the user will fill in exact keys
Places where the user should verify a factual claim
Suggestions for additional citations the user might want to add
Any sentence where Claude is uncertain about the accuracy of a technical claim
Always remind the user: Claude cannot guarantee the accuracy of specific technical claims about cited papers. The user must verify all factual claims and citation details against the actual papers.
Tone and Style Targets
Voice: Third-person academic, active voice preferred
Sentence length: Vary, but average 15–25 words. No sentence over 40 words.
Paragraph length: 4–8 sentences per paragraph, each paragraph focused on one thread
Overall length: Scale to venue — typically 0.75–2 pages for a conference paper, potentially longer for a journal
Register: Precise and technical but not pompous. "Use" not "utilize." "Show" not "demonstrate." Avoid "novel" and "state-of-the-art" — let the results speak.
Example: Before and After
Before (weak — data dump, no gap):
Vaswani et al. (2017) introduced the Transformer architecture. Devlin et al. (2019)
proposed BERT for pre-training. Radford et al. (2019) developed GPT-2. Brown et al.
(2020) scaled this to GPT-3. Recently, Touvron et al. (2023) released LLaMA. Wei et al.
(2022) showed that chain-of-thought prompting improves reasoning.
After (strong — synthesized, gap-focused):
Large language models have advanced rapidly through scaling pre-trained Transformers,
progressing from masked language modeling to autoregressive generation at increasingly
large parameter counts (Devlin et al., 2019; Brown et al., 2020; Touvron et al., 2023).
While scaling improves general capability, these models remain brittle on multi-step
reasoning tasks, often producing plausible but logically inconsistent outputs. Prompting
strategies such as chain-of-thought (Wei et al., 2022) partially mitigate this, but
require careful prompt engineering and degrade on out-of-distribution problem types.
Our approach instead addresses reasoning robustness at the architectural level,
eliminating the need for task-specific prompts.
Reference
For deeper guidance on the principles underlying this skill, consult references/writing_principles.md.