Extracts and assesses research methodology, claims, evidence, and infrastructure from research papers in HASS disciplines. Evaluates transparency, reproducibility, and credibility through systematic extraction (eight-pass workflow, Pass 0-7) and credibility assessment (research approach classification, quality gating, and repliCATS Seven Signals evaluation adapted for HASS).
Systematic extraction and assessment framework for research methodology, argumentation, and reproducibility infrastructure in HASS disciplines (archaeology, palaeoecology, ethnography, ecology, literary studies, philology, etc.).
This skill enables comprehensive extraction of research content and infrastructure from academic papers, followed by credibility assessment, through a structured multi-pass workflow:
Extraction Phase (Passes 1-7):
Assessment Phase (Passes 8-9):
The extracted data enables systematic assessment of research transparency, reproducibility, and credibility.
Use when users request:
The complete workflow follows this sequence:
Blank JSON Template
↓
Pass 1: Evidence extraction (liberal)
↓
Pass 2: Claims + implicit arguments extraction + rationalization
↓
Pass 3: RDMAP extraction (liberal)
↓
Pass 4: RDMAP rationalization
↓
Pass 5: Research designs extraction
↓
Pass 6: Infrastructure extraction (PIDs, FAIR, funding, permits)
↓
Pass 7: Validation (integrity checks)
↓
extraction.json (complete)
↓
Pass 8: Research approach classification
↓
classification.json (approach + HARKing detection)
↓
Pass 9: Credibility assessment (quality-gated)
↓
assessment/ (cluster files + credibility report + assessment.json)
Key principles:
This skill provides:
The user provides:
Why this separation? Extraction prompts evolve frequently through testing and refinement. This architecture allows prompt tuning without modifying the skill package, minimizing versioning conflicts.
Users will typically request extraction at a specific pass. Listen for:
The user will provide the extraction prompt for the specific pass they want. These prompts are:
Claims/Evidence Extraction:
RDMAP Extraction:
Validation:
The prompts contain detailed instructions, examples, and decision frameworks for that specific extraction pass. Follow the prompt provided.
If you encounter uncertainty during extraction, consult:
Core Extraction Principles:
references/extraction-fundamentals.md - Universal sourcing requirements, explicit vs implicit extraction, systematic implicit RDMAP patterns, systematic implicit arguments patterns with 6 recognition patterns (ALWAYS read first for Passes 1-5)references/verbatim-quote-requirements.md - Strict verbatim quote requirements (prevents 40-50% validation failures)references/verification-procedures.md - Source verification for Pass 7 validationSchema & Structure:
references/schema/schema-guide.md - Complete object definitions with inline examplesDecision Frameworks:
references/checklists/tier-assignment-guide.md - Design vs Method vs Protocol decisionsreferences/research-design-operational-guide.md - Operational patterns for finding all Research Designs (4-6 expected)references/checklists/consolidation-patterns.md - When to lump vs split items, cross-reference repair procedure (CRITICAL for Passes 2 & 4)references/checklists/expected-information.md - Domain-specific completeness checklistsInfrastructure Assessment (Pass 6):
references/infrastructure/pid-systems-guide.md - Persistent identifiers (DOI, ORCID, RAiD, IGSN, software PIDs), PID graph connectivity scoring, HASS adoption contextreferences/infrastructure/fair-principles-guide.md - FAIR principles framework, metadata richness, controlled vocabularies, software-specific FAIR (FAIR4RS), computational reproducibility spectrum, machine-actionability, context-dependent assessmentreferences/infrastructure/fieldwork-permits-guide.md - Permit types, CARE principles integration, ethical restrictions assessmentreferences/infrastructure/credit-taxonomy.md - CReDIT contributor roles taxonomy (14 roles), format variationsExamples:
references/examples/sobotkova-example.md - Complete worked exampleFollow the workflow guidance to:
["M003", "M007"]Evidence = Raw observations requiring minimal interpretation (measurements, observations, data points)
Claims = Assertions that interpret or generalize (require reasoning or expertise to assess)
Test: "Does this require expertise to assess or just checking sources?"
For complete decision framework with examples and edge cases:
→ See references/checklists/evidence-vs-claims-guide.md
Research Designs (WHY), Methods (WHAT), Protocols (HOW).
For complete tier assignment guidance: See references/checklists/tier-assignment-guide.md
Evidence items with identical claim support patterns that are never cited independently should be consolidated.
For complete algorithm, examples, and cross-reference repair:
→ See references/checklists/consolidation-patterns.md
Classify research approach (inductive/deductive/abductive) with expressed vs revealed methodology comparison and HARKing detection.
Classification guidance:
references/credibility/approach-taxonomy.md - Definitions of deductive/inductive/abductive approaches, mixed-method characterisation, "none_stated" handlingreferences/credibility/harking-detection-guide.md - Expressed vs revealed comparison, mismatch types, assessment integrationreferences/schema/classification-schema.md - Complete output structure specificationAssess paper credibility using repliCATS Seven Signals adapted for HASS with approach-specific scoring anchors. Quality-gated workflow ensures assessment viability.
Step 1: Track A quality gating - Determines assessment pathway
Step 2: Signal cluster assessment (if quality ≥ moderate)
Assessment is organised into three pillars (see references/credibility/assessment-pillars.md):
Step 3: Report generation
If quality_state = "high" or "moderate":
track-a-quality.md - Quality assessmentcluster-1-foundational-clarity.md - Transparency pillar assessmentcluster-2-evidential-strength.md - Credibility pillar assessmentcluster-3-reproducibility.md - Reproducibility pillar assessmentcredibility-report-v1.md (or -CAVEATED.md if moderate)assessment.json - Canonical consolidationIf quality_state = "low":
track-a-only.md - Quality assessmentassessment-not-viable.md - Explanation of why assessment abortedCredibility assessment guidance:
references/credibility/assessment-pillars.md - Three pillars framework (Transparency, Credibility, Reproducibility)references/credibility/signal-definitions-hass.md - Seven Signals with approach-specific scoring anchors (0-100 scale for deductive/inductive/abductive)references/credibility/assessment-frameworks.md - Framework selection and signal emphasis by research approachreferences/credibility/track-a-quality-criteria.md - Quality gating decision logic (HIGH/MODERATE/LOW states)references/schema/assessment-schema.md - Cluster file and assessment.json structure specifications🚨 CRITICAL: Where to Find Code/Data Availability
For Transparency signal assessment, code/data availability is in reproducibility_infrastructure (NOT in evidence[]):
extraction.json → reproducibility_infrastructure
├── code_availability
│ ├── statement_present: true|false
│ ├── repositories: [{name, url, access_conditions}]
│ └── machine_actionability: {rating, rationale}
├── data_availability
│ ├── statement_present: true|false
│ ├── repositories: [{name, url, access_conditions}]
│ └── machine_actionability: {rating, rationale}
├── persistent_identifiers
│ └── software_pids: [{software_name, repository, doi, url}]
├── preregistration
│ └── preregistered: true|false
└── fair_assessment (if populated)
└── total_fair_score, fair_percentage
Always check these sections when assessing Transparency. Do NOT rely on evidence[] for code/data information.
Reproducibility = Analytic or Computational Reproducibility (NOT beginning-to-end reproducibility)
Approach-Specific Anchors:
CARE Principles Integration:
Use these exact field names. Do not improvise variants.
evidence_id (pattern: E###)evidence_textevidence_typeverbatim_quote ← REQUIREDlocation, supports_claims, source_verificationclaim_id (pattern: C###)claim_textclaim_type: empirical | interpretation | methodological_argument | theoreticalclaim_role: core | intermediate | supportingverbatim_quote ← REQUIREDlocation, supported_by, supports_claims, source_verificationimplicit_argument_id (pattern: IA###)argument_texttype: logical_implication | unstated_assumption | bridging_claim | design_assumption | methodological_assumptiontrigger_text ← REQUIRED (array of verbatim passages)trigger_locations ← REQUIRED (parallel array)inference_reasoning ← REQUIREDsupports_claims, source_verificationdesign_id (pattern: RD###)design_textdesign_typedesign_status: explicit | implicitverbatim_quote (if explicit) OR trigger_text + inference_reasoning (if implicit)method_id (pattern: M###)method_textmethod_typemethod_status: explicit | implicitimplements_designs, realized_through_protocolsprotocol_id (pattern: P###)protocol_textprotocol_typeprotocol_status: explicit | implicitimplements_methods, produces_evidencepaper_id - paper identifier (e.g., "penske-et-al-2023")run_id - run identifier (pattern: run-XX)classification_date - ISO date (YYYY-MM-DD)paper_type: empirical | methodological | theoretical | reviewpaper_type_confidence: high | medium | lowresearch_approach: deductive | inductive | abductive | interpretiveresearch_approach_confidence: high | medium | lowmixed_methods: booleancontext_flags: array (e.g., ["📦", "🔧"])classification_justification - brief rationale for classificationFor complete field definitions: See references/schema/schema-guide.md
For testing/debugging:
Expected outcomes:
Token efficiency:
Common user patterns:
Working with prompts:
Always:
The user will provide the detailed extraction prompt for each pass. Use this skill's reference materials to support decision-making during extraction.