Published on 24/12/2025
Using Real‑World Evidence to Strengthen Immunotherapy Research
Introduction: Why Real‑World Evidence Matters for Immunotherapy
Immuno‑oncology (IO) therapies such as PD‑1/PD‑L1 and CTLA‑4 inhibitors have reshaped cancer care, but traditional randomized trials can’t answer every question patients, payers, and regulators ask. Real‑world evidence (RWE)—clinical insights derived from routinely collected data like electronic health records (EHRs), cancer registries, and claims—helps fill gaps on effectiveness across diverse populations, long‑term safety, dosing schedules in practice, and treatment sequencing. For IO specifically, RWE is invaluable to characterize rare immune‑related adverse events (irAEs), assess outcomes beyond tightly controlled trial settings, and understand how biomarkers (e.g., PD‑L1 tiers, TMB) correlate with effectiveness in routine care.
Unlike trials with fixed visit windows and standardized assessments, real‑world data (RWD) are messy: irregular imaging, missing labs, and variable documentation. Turning this into decision‑grade RWE requires a protocolized plan for data curation, bias reduction, endpoint adjudication, and transparent reporting. When done well, RWE complements trials for label expansions, external control arms, post‑marketing commitments, and health‑technology assessments. Guidance from major agencies outlines how to ensure fitness‑for‑use, study replicability, and auditability for submissions in oncology. See foundational frameworks from the FDA for RWE program expectations.
RWD Sources
Common sources include EHR networks, disease‑specific registries, pathology and genomics labs, medical/Pharmacy claims, and mortality indexes. For IO use‑cases, linked datasets (EHR + imaging + genomics + claims) enable richer covariate balance and more accurate outcome ascertainment (e.g., time to next treatment, inpatient admissions for irAEs). Prior to analysis, perform a fitness‑for‑use assessment: completeness of key variables (stage, line of therapy, ECOG, PD‑L1 %), timeliness of data refresh, site coverage, and coding consistency (ICD‑10, HCPCS, LOINC).
Codify abstraction rules for unstructured notes (progress notes, radiology) and define quality checks: inter‑abstractor agreement ≥90%, reconciliation workflows, and lock procedures. Where biomarker labs underpin subgroup analyses (PD‑L1, MSI‑H, TMB), ensure analytical validation metadata are captured. The table below illustrates a small, illustrative quality/assay spec block often attached to RWE curation packets when integrating lab‑derived endpoints into IO datasets.
| Parameter | Spec / Example Value | Usage in RWE IO Study |
|---|---|---|
| LOD (PD‑L1 IHC assay) | 0.5 ng/mL (illustrative) | Supports sensitivity claims when mapping low‑expressors |
| LOQ (ctDNA TMB panel) | 1.5 ng/mL; TMB limit = 5 mut/Mb | Defines reliability threshold for subgroup assignment |
| PDE (safety threshold example) | 0.02 mg/day (illustrative) | Context for concomitant exposure risk notes |
| MACO (carryover example) | 12 mg (illustrative) | Manufacturing/cross‑contamination note for integrated datasets |
Note: PDE and MACO are manufacturing‑oriented constructs; they’re shown here as examples of documented thresholds when RWE packages incorporate lab/manufacturing context (e.g., companion diagnostic validation summaries) into inspection‑ready binders.
Study Designs for IO RWE: External Controls, Pragmatic Trials, and Hybrids
External control arms (ECAs): For single‑arm IO trials, matched real‑world cohorts can contextualize response rates or survival. Construct ECAs by mirroring trial inclusion/exclusion, index dates (e.g., start of first IO infusion), and follow‑up rules. Use rigorous pre‑specification for covariates (age, ECOG, stage, PD‑L1 strata, brain metastases, steroid pre‑use, comorbidities).
Pragmatic/point‑of‑care trials: Embed randomization into care pathways, with broad eligibility and minimal extra visits. For IO combinations (e.g., ICI + chemo in routine NSCLC care), pragmatic designs capture adherence to dosing intervals, dose holds for irAEs, and imaging cadence variability that reflects reality.
Hybrid designs: Augment ongoing trials with RWE extensions—post‑trial follow‑up via EHR linkages to quantify late irAEs or durability beyond the trial window. Always detail data provenance, curation SOPs, and change‑logs to maintain traceability from source to analysis dataset.
Endpoints in the Real World: Response, Progression, and Safety for IO
Endpoints must align with how care is delivered. Real‑world overall survival (rwOS) uses linked mortality sources. Real‑world PFS (rwPFS) is challenging because imaging timing is inconsistent; define progression as the earliest of radiology‑confirmed progression, switch of systemic therapy, or death, and document adjudication rules. Consider iRECIST‑aligned adjudication for suspected pseudoprogression: require a confirmatory scan window (e.g., ≤8 weeks) before classifying as rwPD when clinically stable.
For real‑world response (rwORR), create an abstraction guide for PR/CR calls from radiology text and tumor boards. For safety, quantify irAE curation pipelines: trigger terms (e.g., “immune‑mediated colitis”), steroid courses ≥20 mg prednisone‑equivalent, specialty consults, and relevant CPT/ICD patterns. Add patient‑reported outcomes where available (ePRO portals) to enrich fatigue/pruritus capture often under‑coded in EHRs.
Controlling Bias and Confounding: From Design Through Analysis
Key threats include confounding by indication (sicker patients preferentially selected for or against IO), immortal‑time bias (time between diagnosis and IO start), and informative censoring. Mitigate them with a layered strategy:
Design/Pre‑analysis
- Emulate trial criteria; align index dates; enforce baseline look‑back (≥6–12 months) to capture comorbidities and prior therapies.
- Specify covariates a priori (e.g., ECOG, PD‑L1 0/1–49/≥50%, TMB high/low, corticosteroid use >10 mg). Handle missingness with multiple imputation and report % missing by variable.
Analysis
- Propensity score matching (caliper 0.2 SD of logit) or inverse probability of treatment weighting (IPTW) with stabilized weights; present covariate balance (standardized mean differences <0.1).
- Competing‑risk models for time‑to‑event with death as competing event where applicable; sensitivity analyses with alternative index definitions.
Provide negative controls (outcomes unlikely related to IO) and tipping‑point analyses to show robustness to unmeasured confounding. Always publish a detailed SAP and protocol supplement for reproducibility.
Regulatory Expectations and Submission‑Ready RWE Packages
Agencies expect clarity on data provenance, traceability, and methodological rigor. A submission‑ready oncology RWE package typically includes: (1) protocol & SAP aligned to the research question (e.g., effectiveness of first‑line ICI in PD‑L1 ≥50% NSCLC), (2) data source characterization and site list, (3) curation SOPs with inter‑abstractor agreement metrics, (4) predefined endpoints and adjudication rules, (5) full code lists (ICD/LOINC/RxNorm), (6) diagnostics for balance and missingness, (7) sensitivity analyses, and (8) traceable programming records with version control. For cross‑referenceable regulatory reading, see EMA’s growing body of RWE guidance and Big Data network publications on methodological standards at the EMA.
When RWE supplements a single‑arm IO trial via an external control, document exchangeability arguments: comparability of assessment schedules, imaging technology, and steroid/immunosuppressant policies. Pre‑specify how you’ll address misalignment (e.g., anchor windows, re‑indexing rules) and show that results are consistent across analytic approaches.
Operationalizing IO RWE: Governance, Linkage, and Audit Readiness
Create a data governance charter that covers site onboarding, data sharing agreements, de‑identification, and patient privacy. For linkage (EHR↔claims↔mortality↔genomics), use tokenization with match confidence thresholds (e.g., ≥0.95) and persistent pseudo‑IDs. Build quality dashboards (e.g., ECOG completeness ≥85%, PD‑L1 captured in ≥70% where clinically indicated, imaging cadence metrics) and implement deviation CAPA workflows.
House all materials—source‑to‑target mapping, abstraction guides, QC logs—in an inspection‑ready TMF‑like repository. For practical SOP templates and inspection checklists, see resources at PharmaRegulatory, which many teams adapt to standardize oncology RWE operations across vendors and sites.
Illustrative Case Study and Practical Checklist
Case (hypothetical): Single‑arm Phase II PD‑1 inhibitor in metastatic urothelial carcinoma (n=145) reports ORR 28%. An external real‑world cohort (EHR + claims, n=420) is constructed from patients on platinum doublet with similar inclusion criteria. After IPTW (SMDs <0.1 for all key covariates), rwOS HR = 0.78 (95% CI 0.66–0.92), rwORR 24% vs 15% (adjudicated), and Grade ≥3 irAE‑related hospitalizations 4.2% vs 1.1% (chemo). Sensitivity analyses (on‑treatment vs intention‑to‑treat index; alternative death data sources) yield HR 0.76–0.81. Results inform a payer dossier and support a post‑marketing commitment to monitor endocrine irAEs at scale.
Checklist (ready‑to‑use):
- Define the estimand up front (population, variable, intercurrent events, summary measure).
- Lock covariates and endpoint rules pre‑analysis; publish SAP and code lists.
- Demonstrate data fitness (completeness, recency, site distribution) and inter‑abstractor agreement.
- Achieve covariate balance (SMD <0.1) and include diagnostics in the main report.
- Run sensitivity analyses (missing data, alternative index, competing risks, negative controls).
- Archive provenance artifacts and QC trails for audit.
