Conduct comprehensive, systematic literature reviews using multiple academic databases (PubMed, arXiv, bioRxiv, Semantic Scholar, etc.). This skill should be used when conducting systematic literature reviews, meta-analyses, research synthesis, or comprehensive literature searches across biomedical, scientific, and technical domains. Creates professionally formatted markdown documents and PDFs with verified citations in multiple citation styles (APA, Nature, Vancouver, etc.).
K-Dense-AI18,791 estrellas13 abr 2026
Ocupación
Categorías
Académico
Contenido de la habilidad
Overview
Conduct systematic, comprehensive literature reviews following rigorous academic methodology. Search multiple literature databases, synthesize findings thematically, verify all citations for accuracy, and generate professional output documents in markdown and PDF formats.
This skill uses the parallel-web skill (parallel-cli search) as the primary web search tool for broad academic literature discovery, supplemented by specialized database access skills (gget, bioservices, datacommons-client). It provides specialized tools for citation verification, result aggregation, and document generation.
When to Use This Skill
Use this skill when:
Conducting a systematic literature review for research or publication
Synthesizing current knowledge on a specific topic across multiple sources
Performing meta-analysis or scoping reviews
Writing the literature review section of a research paper or thesis
Investigating the state of the art in a research domain
Identifying research gaps and future directions
Requiring verified citations and professional formatting
Skills relacionados
Visual Enhancement with Scientific Schematics
⚠️ MANDATORY: Every literature review MUST include at least 1-2 AI-generated figures using the scientific-schematics skill.
This is not optional. Literature reviews without visual elements are incomplete. Before finalizing any document:
Generate at minimum ONE schematic or diagram (e.g., PRISMA flow diagram for systematic reviews)
Study designs (RCTs, observational, in vitro, etc.)
Document all criteria clearly
Phase 2: Systematic Literature Search
Multi-Database Search:
Select databases appropriate for the domain. Always start with parallel-web for broad academic coverage, then supplement with domain-specific databases.
Organize Results section by themes or research questions
Synthesize findings across multiple studies within each theme
Compare and contrast different approaches and results
Identify consensus areas and points of controversy
Highlight the strongest evidence
Example structure:
#### 3.3.1 Theme: CRISPR Delivery Methods
Multiple delivery approaches have been investigated for therapeutic
gene editing. Viral vectors (AAV) were used in 15 studies^1-15^ and
showed high transduction efficiency (65-85%) but raised immunogenicity
concerns^3,7,12^. In contrast, lipid nanoparticles demonstrated lower
efficiency (40-60%) but improved safety profiles^16-23^.
Critical Analysis:
Evaluate methodological strengths and limitations across studies
Assess quality and consistency of evidence
Identify knowledge gaps and methodological gaps
Note areas requiring future research
Write Discussion:
Interpret findings in broader context
Discuss clinical, practical, or research implications
Acknowledge limitations of the review itself
Compare with previous reviews if applicable
Propose specific future research directions
Phase 6: Citation Verification
CRITICAL: All citations must be verified for accuracy before final submission.
# Search PubMed
gget search pubmed "CRISPR gene editing" -l 100
# Search with filters
# Use PubMed Advanced Search Builder to construct complex queries
# Then execute via gget or direct Entrez API
Senior researchers with high h-index (>40 in established fields)
Leading research groups at recognized institutions (Harvard, Stanford, MIT, Oxford, etc.)
Authors with multiple Tier-1 publications in the relevant field
Researchers with recognized expertise (awards, editorial positions, society fellows)
Identifying Seminal Papers
For any topic, identify foundational work by:
High citation count (typically 500+ for papers 5+ years old)
Frequently cited by other included studies (appears in many reference lists)
Published in Tier-1 venues (Nature, Science, Cell family)
Written by field pioneers (often cited as establishing concepts)
Best Practices
Search Strategy
Start with parallel-web: Use parallel-cli search with academic domains for initial broad coverage before querying specialized databases
Use multiple databases (minimum 3): Ensures comprehensive coverage — parallel-web counts as one source
Include preprint servers: Captures latest unpublished findings
Document everything: Search strings, dates, result counts for reproducibility — save all parallel-cli output to sources/
Test and refine: Run pilot searches, review results, adjust search terms
Sort by citations: When available, sort search results by citation count to surface influential work first
Use parallel-cli extract: Fetch full content from promising URLs found during search to verify relevance before full-text screening
Screening and Selection
Use multiple databases (minimum 3): Ensures comprehensive coverage
Include preprint servers: Captures latest unpublished findings
Document everything: Search strings, dates, result counts for reproducibility
Test and refine: Run pilot searches, review results, adjust search terms
Screening and Selection
Use clear criteria: Document inclusion/exclusion criteria before screening
Screen systematically: Title → Abstract → Full text
Document exclusions: Record reasons for excluding studies
Consider dual screening: For systematic reviews, have two reviewers screen independently
Synthesis
Organize thematically: Group by themes, NOT by individual studies
Synthesize across studies: Compare, contrast, identify patterns
Be critical: Evaluate quality and consistency of evidence
Identify gaps: Note what's missing or understudied
Quality and Reproducibility
Assess study quality: Use appropriate quality assessment tools
Verify all citations: Run verify_citations.py script
Document methodology: Provide enough detail for others to reproduce
Follow guidelines: Use PRISMA for systematic reviews
Writing
Be objective: Present evidence fairly, acknowledge limitations
Be systematic: Follow structured template
Be specific: Include numbers, statistics, effect sizes where available
Be clear: Use clear headings, logical flow, thematic organization
Common Pitfalls to Avoid
Single database search: Misses relevant papers; always search multiple databases
No search documentation: Makes review irreproducible; document all searches
Study-by-study summary: Lacks synthesis; organize thematically instead
Unverified citations: Leads to errors; always run verify_citations.py
Too broad search: Yields thousands of irrelevant results; refine with specific terms
Too narrow search: Misses relevant papers; include synonyms and related terms
Ignoring preprints: Misses latest findings; include bioRxiv, medRxiv, arXiv
No quality assessment: Treats all evidence equally; assess and report quality
Publication bias: Only positive results published; note potential bias
Outdated search: Field evolves rapidly; clearly state search date
Example Workflow
Complete workflow for a biomedical literature review:
# 1. Create review document from template
cp assets/review_template.md crispr_sickle_cell_review.md
# 2. Start with parallel-web for broad academic search
parallel-cli search "CRISPR Cas9 sickle cell disease gene therapy efficacy" \
-q "CRISPR" -q "sickle cell" -q "gene therapy" \
--json --max-results 10 --excerpt-max-chars-total 27000 \
--include-domains "scholar.google.com,arxiv.org,pubmed.ncbi.nlm.nih.gov,semanticscholar.org,biorxiv.org,nature.com,science.org,cell.com,pnas.org,nih.gov" \
-o sources/litreview_crispr_scd-academic.json
parallel-cli search "CRISPR sickle cell disease clinical trials treatment" \
-q "CRISPR" -q "sickle cell" \
--json --max-results 10 --excerpt-max-chars-total 27000 \
-o sources/litreview_crispr_scd-general.json
# 3. Search specialized databases using appropriate skills
# - Use gget skill for PubMed, bioRxiv
# - Use direct API access for arXiv, Semantic Scholar
# - Export results in JSON format
# 4. Aggregate and process results (combine parallel-cli + database results)
python scripts/search_databases.py combined_results.json \
--deduplicate \
--rank citations \
--year-start 2015 \
--year-end 2024 \
--format markdown \
--output search_results.md \
--summary
# 5. Screen results and extract data
# - Use parallel-cli extract to fetch full content from promising URLs
# - Manually screen titles, abstracts, full texts
# - Extract key data into the review document
# - Organize by themes
# 6. Write the review following template structure
# - Introduction with clear objectives
# - Detailed methodology section
# - Results organized thematically
# - Critical discussion
# - Clear conclusions
# 7. Verify all citations
python scripts/verify_citations.py crispr_sickle_cell_review.md
# Review the citation report
cat crispr_sickle_cell_review_citation_report.json
# Fix any failed citations and re-verify
python scripts/verify_citations.py crispr_sickle_cell_review.md
# 8. Generate professional PDF
python scripts/generate_pdf.py crispr_sickle_cell_review.md \
--citation-style nature \
--output crispr_sickle_cell_review.pdf
# 9. Review final PDF and markdown outputs
Integration with Other Skills
This skill works seamlessly with other scientific skills:
Web Search & Extraction (parallel-web skill — PRIMARY)
parallel-cli search: Broad academic and general web search with domain filtering — use for initial scoping, finding papers, citation chaining, and supplementary searches
parallel-cli extract: Fetch full content from paper URLs, journal websites, and preprint servers — use for reading abstracts, extracting reference lists, and verifying paper details