Finding a middle-ground approach to balancing new solutions arising from data science with traditional requirements for data collection and submission.
There is a problem in clinical trial data structures. It isn’t a problem for any individual trial, or even a few. But, for those of us who deal with tens or hundreds or even thousands of trials worth of data, it is a pretty serious hurdle. It’s called scalability, and for the clinical trials industry, it presents a unique issue.
The reason for this is the expectation in clinical trials that data be maintained in a separate compartmentalized database. This is considered good clinical practice (GCP). Decades ago, when this practice was first introduced, it set a series of goals that were admirable, expected, and not difficult to follow. However, as the years went on, the rules were not updated. Much like trying to drive a modern car on streets designed for horse and carriage, what was once a good fit for the industry has become somewhat of a hindrance.
Simply put, a lot has changed technologically since the original principles were put in place decades ago. The tools we have at our disposal are on a level that the original guidelines could never have predicted. This means that the expectations for the use of data in general, not only in clinical trials, couldn’t have been anticipated.
Click to enlarge
So, how do we marry the expectations and solutions we have available to us from a data science perspective with requirements arising from these traditional viewpoints? Well, as an industry, we must recognize the limitations of the physical models we have in place currently. Then, we must talk about the possibilities that are available to us. Fortunately, other industries have solved these problems and are making advancements. So, while we are certainly lagging behind, there is a clear light at the end of the tunnel.
In order to begin advancing, we must first understand the exact problem. As mentioned, the problem that we must solve is that all clinical trial databases must be kept separate from each other. The debate is between physical separation and logical separation.
Logical separation provides the ability to be in one electronic database but to utilize filters to isolate the clinical trial database that is relevant for your analysis. Physical separation puts major hurdles around analysis and automation, while providing an additional layer of security and integrity protection. Both strategies are valid, but only one of them is scalable operationally, and allows for automation of data science, analysis, and monitoring.
Click to enlarge
A nice middle ground to this is the apartment model. Here, the clinical submission data flow path works in a physical separation “apartment” model while having a system that isn’t on the critical path of collection and submission house data in a warehouse or data lake for scalable reporting, operational monitoring, and aggregated data analysis, and aggregated data science statistical model development (see Figure 2).
This allows data collections to meet all previous regulatory concern requirements, get the most out of modern data architecture and data science solutions, and does so without ever compromising GCP principles.
If one wants to leverage image analysis, classification, machine learning, deep learning, or any of the other potentially groundbreaking technologies that are available on specific data, the data must be in a structure that normalizes and standardizes that data.
In other words, in order to get the most out of technology, while also staying within GCP, one will need to marry these two architecture solutions together and develop a comprehensive answer to these problems.
As we move further into the decade, we can expect to see some groundbreaking and new ways to use technology, both in our personal lives and in clinical trials. The question for us as an industry, however, is will we be ready when those advancements arrive? Or, will we still be stuck driving supercars on top of cobblestone?
Keith Aumiller, Senior Director, Data Services, Signant Health
Improving Relationships and Diversifying the Site Selection Process
April 17th 2025In this episode of the Applied Clinical Trials Podcast, Liz Beatty, co-founder and chief strategy officer, Inato, discusses a number of topics around site engagement including community-based sites, the role of technology in improving site/sponsor relationships, how increased operational costs are impacting the industry, and more.
Behind the Buzz: Why Clinical Research Leaders Flock to SCOPE Summit
February 7th 2025In this episode, we meet with Micah Lieberman, Executive Conference Director for SCOPE Summit (Summit for Clinical Ops Executives) at Cambridge Innovation Institute. We will dive deep into the critical role of collaboration within the clinical research ecosystem. How do we bring together diverse stakeholders—sponsors, CROs, clinical trial tech innovators, suppliers, patients, sites, advocacy organizations, investors, and non-profits—to share best practices in trial design, program planning, innovation, and clinical operations? We’ll explore why it’s vital for thought leaders to step beyond their own organizations and learn from others, exchanging ideas that drive advancements in clinical research. Additionally, we’ll discuss the pivotal role of scientific conferences like SCOPE Summit in fostering these essential connections and collaborations, helping shape the future of clinical trials. Join us as we uncover how collective wisdom and cross-industry partnerships are transforming the landscape of clinical research.
FDA-Approved Gene Therapy Beqvez Shows Sustained Efficacy, Safety in Long-Term Hemophilia B Trial
April 17th 2025Beqvez (fidanacogene elaparvovec), an FDA-approved one-time gene therapy for hemophilia B, demonstrated sustained factor IX expression, low bleeding rates, and a favorable safety profile over long-term follow-up.