Finding a middle-ground approach to balancing new solutions arising from data science with traditional requirements for data collection and submission.
There is a problem in clinical trial data structures. It isn’t a problem for any individual trial, or even a few. But, for those of us who deal with tens or hundreds or even thousands of trials worth of data, it is a pretty serious hurdle. It’s called scalability, and for the clinical trials industry, it presents a unique issue.
The reason for this is the expectation in clinical trials that data be maintained in a separate compartmentalized database. This is considered good clinical practice (GCP). Decades ago, when this practice was first introduced, it set a series of goals that were admirable, expected, and not difficult to follow. However, as the years went on, the rules were not updated. Much like trying to drive a modern car on streets designed for horse and carriage, what was once a good fit for the industry has become somewhat of a hindrance.
Simply put, a lot has changed technologically since the original principles were put in place decades ago. The tools we have at our disposal are on a level that the original guidelines could never have predicted. This means that the expectations for the use of data in general, not only in clinical trials, couldn’t have been anticipated.
Click to enlarge
So, how do we marry the expectations and solutions we have available to us from a data science perspective with requirements arising from these traditional viewpoints? Well, as an industry, we must recognize the limitations of the physical models we have in place currently. Then, we must talk about the possibilities that are available to us. Fortunately, other industries have solved these problems and are making advancements. So, while we are certainly lagging behind, there is a clear light at the end of the tunnel.
In order to begin advancing, we must first understand the exact problem. As mentioned, the problem that we must solve is that all clinical trial databases must be kept separate from each other. The debate is between physical separation and logical separation.
Logical separation provides the ability to be in one electronic database but to utilize filters to isolate the clinical trial database that is relevant for your analysis. Physical separation puts major hurdles around analysis and automation, while providing an additional layer of security and integrity protection. Both strategies are valid, but only one of them is scalable operationally, and allows for automation of data science, analysis, and monitoring.
Click to enlarge
A nice middle ground to this is the apartment model. Here, the clinical submission data flow path works in a physical separation “apartment” model while having a system that isn’t on the critical path of collection and submission house data in a warehouse or data lake for scalable reporting, operational monitoring, and aggregated data analysis, and aggregated data science statistical model development (see Figure 2).
This allows data collections to meet all previous regulatory concern requirements, get the most out of modern data architecture and data science solutions, and does so without ever compromising GCP principles.
If one wants to leverage image analysis, classification, machine learning, deep learning, or any of the other potentially groundbreaking technologies that are available on specific data, the data must be in a structure that normalizes and standardizes that data.
In other words, in order to get the most out of technology, while also staying within GCP, one will need to marry these two architecture solutions together and develop a comprehensive answer to these problems.
As we move further into the decade, we can expect to see some groundbreaking and new ways to use technology, both in our personal lives and in clinical trials. The question for us as an industry, however, is will we be ready when those advancements arrive? Or, will we still be stuck driving supercars on top of cobblestone?
Keith Aumiller, Senior Director, Data Services, Signant Health
Behind the Buzz: Why Clinical Research Leaders Flock to SCOPE Summit
February 7th 2025In this episode, we meet with Micah Lieberman, Executive Conference Director for SCOPE Summit (Summit for Clinical Ops Executives) at Cambridge Innovation Institute. We will dive deep into the critical role of collaboration within the clinical research ecosystem. How do we bring together diverse stakeholders—sponsors, CROs, clinical trial tech innovators, suppliers, patients, sites, advocacy organizations, investors, and non-profits—to share best practices in trial design, program planning, innovation, and clinical operations? We’ll explore why it’s vital for thought leaders to step beyond their own organizations and learn from others, exchanging ideas that drive advancements in clinical research. Additionally, we’ll discuss the pivotal role of scientific conferences like SCOPE Summit in fostering these essential connections and collaborations, helping shape the future of clinical trials. Join us as we uncover how collective wisdom and cross-industry partnerships are transforming the landscape of clinical research.
Reaching Diverse Patient Populations With Personalized Treatment Methods
January 20th 2025Daejin Abidoye, head of solid tumors, oncology development, AbbVie, discusses a number of topics around diversity in clinical research including industry’s greatest challenges in reaching diverse patient populations, personalized treatment methods, recruitment strategies, and more.
VERITAC-2 Trial Shows Vepdegestrant Significantly Improves Survival in ESR1-Mutant Breast Cancer
March 24th 2025Phase III VERITAC-2 trial results show vepdegestrant significantly improved progression-free survivalcompared to fulvestrant in patients with ESR1-mutant (ESR1m) advanced or metastatic breast cancer, but did not achieve statistical significance in the overall intent-to-treat population.