Published on 27/12/2025
Reliable Sources of Historical Site Performance Data for Informed Feasibility Planning
Introduction: Why Historical Data Matters in Site Selection
Feasibility assessments based solely on investigator reputation or generic questionnaire responses are no longer sufficient. Regulatory expectations under ICH E6(R2) and growing emphasis on quality-by-design demand data-driven decisions—particularly when selecting or requalifying clinical trial sites. One of the most powerful tools in this regard is historical site performance data.
However, such data is fragmented across multiple systems, stakeholders, and documents. To effectively use performance history, sponsors and CROs must first identify and validate reliable sources. This article outlines the key repositories—both internal and external—that house performance-related insights critical to clinical site evaluation.
1. Clinical Trial Management System (CTMS)
Primary Source: Site activity, enrollment metrics, deviation records, visit schedules
The CTMS is the most comprehensive internal repository of site-level performance data. When properly maintained, it provides structured, longitudinal records across multiple studies. Common metrics extracted include:
- Actual vs. planned enrollment timelines
- Screen failure and dropout rates
- Site activation duration (contracting to SIV)
- Protocol deviation frequencies
- Monitoring visit outcomes and action item resolution
Data from the CTMS can be exported into scoring algorithms or dashboards to rank sites against key performance thresholds.
2. Electronic Data Capture (EDC) Systems
Use
EDC systems provide real-time, timestamped evidence of a site’s data management performance. Sponsors should extract:
- Average time to resolve queries
- Number of queries per subject
- Frequency of inconsistent or missing entries
- Instances of backdated or corrected entries (audit trail review)
These indicators contribute to evaluating data integrity and operational discipline at the site level.
3. Monitoring Visit Reports (MVRs)
Source: CRAs’ documented observations and findings
MVRs provide qualitative and narrative context to complement quantitative CTMS data. They reveal:
- Site staff engagement and responsiveness
- Issues with IP storage or informed consent practices
- Monitoring delays and follow-up challenges
- Facility conditions and documentation practices
Feasibility teams should review MVRs from at least the last 2–3 studies conducted by the site.
4. Audit and Inspection Reports
Internal audits: Conducted by QA departments
Regulatory inspections: Conducted by FDA, EMA, MHRA, CDSCO, etc.
These reports are essential to understand the site’s compliance history. Key data points include:
- Number of audits conducted and frequency
- Findings classification: critical, major, minor
- CAPA effectiveness and recurrence of issues
- Regulatory warning letters or Form 483 issuance
For public access, regulators like the FDA provide searchable inspection records via [FDA Inspection Database](https://www.fda.gov/inspections-compliance-enforcement-and-criminal-investigations/inspection-database).
5. Trial Master File (TMF) and eTMF Systems
Documents Reviewed: Delegation logs, training records, IRB approvals, deviation logs
Sites with consistent TMF compliance typically demonstrate strong trial management systems. When reviewing TMFs:
- Check completeness and timeliness of submissions
- Evaluate site file organization and document version control
- Assess availability of GCP and protocol-specific training logs
eTMF metadata can also reveal submission patterns—frequent late uploads may suggest administrative inefficiencies.
6. Site Performance Dashboards (Sponsor-Created)
Many large sponsors build centralized dashboards that aggregate site metrics across studies. These may include:
- Site ranking based on custom KPIs
- Benchmarking across therapeutic areas
- Repeat participation history
- Real-time deviation and query alerts
These dashboards support feasibility reviews and can generate site profiles with graphical performance summaries.
7. CRO Reports and Vendor-Managed Portals
When feasibility and monitoring are outsourced, CROs often maintain site performance data in their proprietary systems. Sponsors should request:
- Study summary reports by site
- Aggregated site performance trends across portfolios
- Enrollment forecasting accuracy logs
- CRA-reported issues unresolved beyond timeline
Vendor qualification SOPs should include access to such performance data when selecting or renewing CRO partnerships.
8. External Clinical Trial Registries and Inspection Portals
These public databases can reveal past participation and regulatory scrutiny at global levels:
- ClinicalTrials.gov – Lists studies by site, sponsor, PI
- EU Clinical Trials Register – EU-based site and trial data
- CTRI (India) – Trial listings by Indian sites and ethics boards
- Japan PMDA RCT Portal – Site registrations and disclosures
- Canada Health Trial Database
While these don’t contain audit details, they reveal participation history, trial phases, and therapeutic experience.
9. Investigator CVs and Feasibility Questionnaires
Though often considered subjective, CVs and completed questionnaires provide context to objective data. Review:
- PI’s previous indications and study phases
- Training and GCP certifications
- Self-reported enrollment success and challenges
These should be cross-verified against actual performance data from CTMS and CRO portals.
Conclusion
Robust site selection and feasibility planning require a multi-source, cross-validated approach to historical performance data. By aggregating insights from internal systems (CTMS, EDC, TMF), monitoring reports, audits, and global registries, sponsors and CROs can develop objective, consistent, and inspection-ready criteria for site engagement. As clinical development becomes more digital, integrating these data streams will be critical not just for faster startup—but for trial success and regulatory compliance.
