Structures clinical trial data analysis with primary endpoint evaluation and safety reporting. Use when analyzing trial results, evaluating endpoints, or preparing statistical reports.
Clinical trial data analysis translates raw study data into evidence that regulators, clinicians, and payers act on. Errors in analysis — wrong population sets, incorrect handling of missing data, failure to control multiplicity — can result in complete response letters, advisory-committee failures, or post-marketing safety crises. This skill implements the statistical analysis workflow defined by ICH E9 (Statistical Principles for Clinical Trials), ICH E9(R1) (Estimands), FDA guidance on clinical trial endpoints, and CDISC standards to produce analyses that withstand regulatory scrutiny.
Checkpoint A — Intake and Scoping
Required Intake Questions
What is the study phase (I–IV) and therapeutic area?
What are the primary, secondary, and exploratory endpoints as defined in the protocol?
Is the Statistical Analysis Plan (SAP) finalized and locked before database lock?
What analysis populations are defined (ITT, mITT, per-protocol, safety)?
What is the randomization structure (stratification factors, adaptive elements)?
相關技能
Are there interim analyses with alpha-spending considerations?
What is the data format (CDISC SDTM, ADaM, legacy)?
Has database lock been confirmed? What is the lock date?
Are there any protocol deviations affecting the analysis populations?
What are the deliverable formats (tables, listings, figures — TLF shells)?
Required Source Documents
Finalized and signed Statistical Analysis Plan (SAP)
Randomization list (unblinded, post-database lock)
Protocol deviation log
Data Management Report (query resolution status, edit checks)
Medical coding dictionaries (MedDRA version, WHO Drug Dictionary version)
Step 1 — Define and Verify Analysis Populations
Construct each analysis population per the SAP:
Intent-to-Treat (ITT): All randomized participants analyzed as randomized, regardless of protocol adherence. This is the primary population for superiority trials per ICH E9.
Modified ITT (mITT): ITT minus participants who never received study drug or have no post-baseline efficacy assessment. Document specific exclusion criteria.
Per-Protocol (PP): Participants who completed the study without major protocol deviations. Define deviation types that trigger PP exclusion (wrong treatment, insufficient exposure, prohibited medications, missed primary-endpoint assessments).
Safety population: All participants who received at least one dose of study medication, analyzed as treated (not as randomized).
Reconcile population counts across datasets. Any discrepancy between randomization list and safety/ITT counts requires documentation.
Step 2 — Produce Demographic and Baseline Tables
Generate the CONSORT-required baseline comparison table:
Demographics: age (mean, SD, median, range), sex, race/ethnicity (per FDA and NIH reporting requirements), BMI
Report: median time-to-event per group (with 95% CI), HR, p-value, KM curves at key timepoints
For Count Data / Recurrent Events
Negative binomial regression or Andersen-Gill model
Report: rate per group, rate ratio with 95% CI, p-value
Step 4 — Handle Missing Data
Implement the pre-specified missing-data strategy per ICH E9(R1) estimand framework:
Primary approach: The method that aligns with the chosen estimand (e.g., MMRM under MAR assumption for treatment-policy estimand)
Sensitivity analyses (required — at least two):
Tipping-point analysis: Determine how extreme imputed values must be to reverse the conclusion
Pattern-mixture model (MNAR): Reference-based imputation (copy-reference, jump-to-reference)
Multiple imputation under various assumptions
Worst-case / best-case imputation for binary endpoints
Supplementary analyses: Complete-case analysis, last-observation-carried-forward (document as sensitivity only — LOCF is no longer acceptable as primary per FDA and EMA guidance)
Documentation: Report the number and reasons for missing data by treatment arm and visit
Step 5 — Analyze Secondary Endpoints with Multiplicity Control
Apply the pre-specified multiplicity adjustment strategy:
Hierarchical (fixed-sequence) testing: Test secondary endpoints in pre-specified order; stop testing at first non-significant result. This is the most common approach for pivotal trials.
Graphical approach (Bretz et al.): Allocate alpha across endpoints with pre-specified propagation rules
Hochberg or Holm step-up/step-down: When endpoints are independent or positively correlated
Gate-keeping strategies: For families of primary and secondary endpoints
Analyze each secondary endpoint using the same methodology specified for its data type. Report adjusted and nominal p-values.
Step 6 — Conduct Safety Analysis
Analyze the safety population:
Adverse Events
Summarize by System Organ Class (SOC) and Preferred Term (PT) using the specified MedDRA version
Present: any AE, drug-related AE, serious AE (SAE), AE leading to discontinuation, AE leading to death
Tabulate by severity grade (CTCAE v5 or mild/moderate/severe)
For each PT, report incidence (n, %) per treatment arm — not number of events (one participant with multiple episodes counts once)
Flag imbalances (≥2% difference between arms or ≥2× rate ratio) for medical review
Tables: formatted with proper headers, footnotes, population counts (N = per arm), and statistical references
Listings: participant-level data for SAEs, deaths, discontinuations, protocol deviations, and concomitant medications
Figures: KM plots, forest plots for subgroups, waterfall plots (oncology), spider plots (tumor response), bar/line charts for PRO scores
All output must be reproducible from ADaM datasets with documented programs
Checkpoint B — Analysis Review
Primary analysis matches the SAP exactly (no unplanned modifications)
All analysis populations are correctly derived and counts reconcile
Missing-data handling follows the estimand framework with sensitivity analyses
Multiplicity adjustment is correctly applied in the specified order
Safety tables use correct MedDRA version and incidence-based (not event-based) counting
Hy's Law assessment is completed for studies with hepatotoxicity potential
All TLFs match the pre-approved shells
Subgroup analyses (sex, age, race, region, baseline severity) are conducted for primary endpoint
Statistical programs are validated (double-programming or independent QC)
Unblinding log confirms no premature unblinding occurred
Quality Audit
SAP version matches the version referenced in the CSR
ADaM datasets are CDISC-compliant with submitted define.xml
All p-values are reported to the appropriate decimal precision (typically 4 decimal places)
Confidence intervals are consistently 2-sided 95% unless otherwise specified
KM curves include number-at-risk tables
Forest-plot subgroup analyses include interaction p-values
All post-hoc analyses are clearly labeled as exploratory
No results are presented that are not derivable from the submitted datasets
All [VERIFY] flags have been resolved or escalated
Guidelines
The SAP must be finalized before database lock and unblinding — any changes after unblinding must be documented and justified as pre-specified sensitivity or clearly labeled post-hoc
Never change the primary analysis method after seeing the data without regulatory disclosure
Use ITT as the primary population for superiority trials; per-protocol as co-primary for non-inferiority
LOCF is no longer acceptable as a primary missing-data method — use MMRM or multiple imputation
All statistical programs must have independent QC (double-programming or code review)
Report effect estimates with confidence intervals, not only p-values — p-values without effect sizes are insufficient
Safety analyses are descriptive — hypothesis testing of AE incidence rates is generally inappropriate
Apply the CONSORT flow diagram to document participant disposition through analysis populations
Mark any deviation from the SAP with [VERIFY] for biostatistics-lead review
This skill produces analysis results — interpretation for regulatory submission requires clinical and regulatory team review