Back to Insights
AI & Evidence

The Sound of Evidence: How AI Reads Clinical Trials

Pier Lasalvia, MD
Pier Lasalvia, MDCo-founder, CTO & Co-CEO
February 15, 2026 10 min read

Every year, over 400,000 clinical trial results are published across thousands of journals. For HTA analysts tasked with building evidence dossiers, this volume presents an impossible challenge: how do you find, read, extract, and synthesize every relevant study — without missing critical data or introducing human error?

The traditional answer has been brute force: teams of trained reviewers spending weeks reading abstracts, screening full texts, and manually extracting data into spreadsheets. It works, but it's slow, expensive, and surprisingly error-prone.

Podcast: AI in Evidence Synthesis — A 12-Minute Overview
12:34

The Evidence Mountain

Consider a typical systematic literature review (SLR) for an HTA submission:

  • Initial search: 3,000–8,000 records identified
  • Abstract screening: 2–3 reviewers spend 40–60 hours
  • Full-text review: 200–500 papers assessed in detail
  • Data extraction: 50–150 studies included, each requiring 30+ data points

The total effort? Anywhere from 6 to 12 weeks for a single therapeutic area, with costs ranging from $50,000 to $200,000 depending on scope.

6–12 weeks

Traditional SLR Timeline

Average time from search to completed evidence table for a single HTA submission.

How AI Changes the Game

Modern natural language processing (NLP) models can read and understand clinical trial publications at scale. But "understanding" means more than keyword matching — it requires parsing complex medical language, identifying specific outcomes, and extracting structured data from unstructured text.

The AI-assisted pipeline works in three stages:

Stage 1: Intelligent Screening

Instead of reading every abstract, the AI model scores each record on relevance to the research question. It considers population, intervention, comparator, and outcome (PICO) criteria simultaneously, flagging borderline cases for human review.

Stage 2: Structured Extraction

For included studies, the AI extracts predefined data points: study design, sample size, endpoints, effect sizes, confidence intervals, adverse events, and more. Each extraction is linked to the source sentence in the original paper, creating a transparent audit trail.

Stage 3: Quality Assessment

The system evaluates study quality using standard frameworks (Cochrane Risk of Bias, Newcastle-Ottawa Scale) and flags potential issues: small sample sizes, high dropout rates, or unusual statistical methods.

When to Trust AI Extractions

AI extraction works best for structured, quantitative data (sample sizes, hazard ratios, p-values). For nuanced qualitative judgments — like assessing whether a study's population matches your target market — human review remains essential. The best approach is AI-first, human-verified.

The Numbers: AI vs. Traditional

How does AI-assisted extraction compare to traditional methods? Toggle between accuracy and speed to see the difference.

Evidence Extraction Comparison
Manual Review72%
Basic NLP81%
AI-Assisted (Quantus)94%

The data tells a clear story: AI-assisted extraction achieves higher accuracy in less time, primarily because:

  1. Consistency: The AI applies the same criteria to every paper, eliminating reviewer fatigue and drift
  2. Completeness: It doesn't skip data points or miss studies buried in supplementary materials
  3. Speed: Processing 500 papers takes hours, not weeks

94%

Extraction Accuracy

AI-assisted accuracy for structured clinical data points, validated against expert reviewer gold standard.

A Real-World Case Study

A mid-size pharmaceutical company needed to update their evidence dossier for a rheumatoid arthritis therapy ahead of a NICE submission. The previous SLR, conducted manually 18 months earlier, had taken 10 weeks and $120,000.

Using AI-assisted methods:

  • Search updated in 1 day (vs. 1 week)
  • Screening completed in 4 hours (vs. 3 weeks)
  • Data extraction finished in 2 days (vs. 4 weeks)
  • Total cost reduced by 70%
  • 3 additional relevant studies identified that the manual review had missed

The team used the time savings to run a network meta-analysis that wasn't in the original budget — analysis that ultimately strengthened their submission.

PRISMA flow diagram showing AI-assisted screening pipeline
Figure 1: PRISMA flow diagram for the AI-assisted rheumatoid arthritis SLR update.

The Human-AI Partnership

AI doesn't replace expert reviewers — it amplifies them. The most effective model is a human-in-the-loop approach where:

  • AI handles high-volume, repetitive tasks (screening, extraction)
  • Humans focus on judgment calls (quality assessment, clinical interpretation)
  • Both contribute to a transparent, auditable evidence base

This partnership reduces total effort by 60–80% while maintaining or improving quality. More importantly, it frees expert time for the analytical work that actually drives better HTA outcomes.

What's Next

The frontier of AI in evidence synthesis is moving toward living systematic reviews — continuously updated evidence bases that incorporate new publications in real-time. Instead of periodic, expensive SLR updates, teams will have always-current evidence at their fingertips.

For HTA professionals, this means faster submissions, stronger evidence packages, and more time spent on strategy rather than data entry. The sound of evidence is getting clearer.