Back to blog

Become Compliant with AI

by
Sanjay Kunchakarra
September 5, 2025
Blog post image

Hear me out. The average phase 3 clinical trial today involves 3.5 million data points collected (up from 1 million in 2012). As protocols become more complex, the burden on study teams to collect, clean and analyze data continues to worsen, leading to mistakes, delays and rework.

In response, the global governing body on clinical trials (ICH) released their official guidance (a.k.a. ICH E6 R3) on how to deal with this issue earlier this year. They require teams to implement "risk-based quality management" within their processes. In simple terms, they want teams to work smart; to figure out the data that’s most important to that study and focus their energy on validating that critical information. Don’t waste time and resources on details that don’t tell you much about how safe or effective a drug is.

Clean, reliable data is the foundation of every trial.

An average trial might have 40 different case report forms x 20 fields per form + 5 external data feeds x 30 data points per feed. That get’s us to about 1,000 unique data fields per study. For each field, you might want to check for completeness, value plausibility, alignment with the protocol, consistency with other data points, etc.

The universe of possible data quality checks starts creeping into the thousands.

To be compliant with the ICHs recommendations, to adopt a true “risk-based” process, study teams now need to manually analyze 5,000+ possible data quality checks and decide which are the most important. These requirements meant to reduce unnecessary work have now created a new bottleneck.

As a result, most study teams just stick with the status quo – they use templated libraries and apply checks that are executable today vs. aligning with regulatory guidelines and building a system that ensures the validity of their most important data.

What we discovered:

One of the biggest unlocks we uncovered is the ability to use autonomous coding agents (e.g., Claude Code) to solve complex business problems. At surface level, one would think that autonomous coding agents might only be useful for software engineers – engineers write code and so do coding agents. But what we’ve understood is that coding agents represent an incredibly powerful combination of factors: the ability to think like a human expert, execute complex logic like a powerful software program and iterate like a tireless worker.

This has created an entirely new world of opportunities – tasks that were previously humanly impossible are now a reality. We can build agents that can understand a study and a protocol, generate a universe of possible rules, design a framework for evaluating importance, and execute evaluation of those rules 1000x faster than a human could.

The big unlock to the reality of risk-based practices that regulators have been looking for might have just come from the least likely of places: AI 😊

Explore more

September 5, 2025

It’s Time for Clinical Trial Data Review to Evolve

Before any real insights can be gleaned from a trials’ results, all collected data needs to be cleaned, verified and signed off on. But while the drug modalities and trial designs have seen incredible innovation over the last few decades, the way the clinical research industry reviews and cleans clinical trial data has not changed much.

Open article
Read more
Blog post image
September 5, 2025

Beyond Edit Checks & SAS: How AI Elevates Data Quality in Clinical Trials

Anyone who has been involved with clinical data review will recognize this cycle: build edit checks, wait for SAS listings, analyze monthly reports, and issue queries weeks after the data was first entered. And it works, sort of. These checks catch missing values, out-of-range numbers, and date inconsistencies. But they also leave behind a long tail of errors that only surface late in the trial, or worse, during database lock crunch time.

Open article
Read more
Blog post image
September 5, 2025

Enabling The 10X Data Manager

Data managers are the unsung heroes of the clinical trial process. They build a foundational level of hygiene into the study database, enabling all downstream analysis. They own the delivery of that final clean database, managing the challenges of protocol variability, site performance, and cross-functional responsibilities throughout the study. It's a thankless job candidly, and as studies get more complex and data volumes increase, the burden on data managers will only continue to grow.

Open article
Read more
Blog post image
Share:

See What a Prediction Engine Can Do for Your Program

We'll walk through your disease area, your data, and how Reveal Clinical fits into your development timeline.