Clinical trials involve the generation of a large volume of clinical data, which are analyzed to assess a new therapy's safety and efficacy. Current complex cloud platform technologies facilitate this process by enabling users around the world to capture a multitude of data points with each patient visit. Such platform technologies systematically generate as a by-product a rich stream of operational data based on numerous ID fields (for sponsors, sites, patients, etc.) and event time stamps as users gather data.
Clinical trial professionals and their sponsors are investing massive amounts of resources into properly executing clinical trials, and are thus highly dependent on the clinical trial sites that conduct the trials. To identify and evaluate areas of operational inefficiency, comparing a clinical trial site to other similar clinical trial sites is useful in determining how well the site is performing regarding one or more operational metrics. Similarly, it is helpful to compare the operational performance of whole studies against similar prior studies or the operational performance of a pharmaceutical sponsor against a group of similar sponsors (e.g., a peer group) or the industry as a whole.
Where considered appropriate, reference numerals may be repeated among the drawings to indicate corresponding or analogous elements. Moreover, some of the blocks depicted in the drawings may be combined into a single function.
In the following detailed description, numerous specific details are set forth in order to provide a thorough understanding of embodiments of the invention. However, it will be understood by those of ordinary skill in the art that the embodiments of the present invention may be practiced without these specific details. In other instances, well-known methods, procedures, components, and circuits have not been described in detail so as not to obscure the present invention.
Operational data enable unique, quantitative perspectives on trial timelines, data quality, and costs across the life science industry. Quantifying important aspects of clinical trials creates a centralized, quantitative basis for managing operational aspects of such trials, which results in increased efficiencies in critical areas such as patient enrollment, data monitoring, trial cycle times, and costs. A system and method for determining relative operational performance in a clinical trial have been developed by using statistical analysis and appropriate transformations, including the z-transform.
Reference is now made to
The word “industry” may be used herein to encompass different combinations of entities that may be compared to the client and may include a peer group or any other combination of competitors or comparison group.
In one embodiment, data may be compared based on therapeutic area rather than on overall clinical trials. For example, client data may be compared to industry data just for clinical trials for oncology. Other therapeutic areas may include central nervous system, immunomodulation, endocrine systems, gastrointestinal, dermatologic, and pain and anesthesia. Data may also be compared based on medical indication (i.e., disease being treated).
In another embodiment, data may be compared based on trial phase (e.g., Phase I, Phase II, Phase III) rather than on overall clinical trials. This view may provide useful information to clients based on phase that may be masked by using an overall view. In another embodiment, data may be compared based on sponsor characteristics (e.g., large pharmaceutical sponsor, small pharmaceutical sponsor, biotech sponsor) rather than on overall clinical trials. In another embodiment, data may be compared based on CRO characteristics, such as determining how a CRO for a trial is performing compared to other CROs running similar trials. In another embodiment, data may be compared based on clinical trial site, to determine how a site in a trial is performing compared to other sites in the trial or in other trials.
Enrollment metrics may include enrollment rate, percentage of high enrolling sites, percentage of non-enrolling sites, number of sites per 100 subjects (or patients), and number of countries per 100 subjects.
Trial cycle times may include the first patient in (FPI) (or enrolled) to last patient in (LPI) (or enrolled) for a trial, DB (database) open to FPI, and last patient visit (LPV) to DB lock.
Monitoring or study conduct may include screen failure rate, on-site monitoring rate, and data correction rate.
Cost metrics may include principal investigator (PI) grant cost per patient and site cost per patient.
Regarding enrollment metrics, one measure of enrollment rate is the rate of enrollment for a trial site or a trial. Enrollment rate for a site may be calculated as the total number of enrolled subjects divided by the total enrollment time for a study site. Enrollment rate for a trial may be calculated as the total number of enrolled subjects divided by the total enrollment time across all sites for a trial.
One measure of percentage of high enrolling sites is the total number of high-enrolling sites divided by the total number of sites for a trial, multiplied by 100. A “high-enrolling site” may be a site that has enrolled more than 1.5 times the mean subject count across all sites for a given trial.
One measure of percentage of non-enrolling sites is the total number of non-enrolling sites divided by the total number of sites for a trial, multiplied by 100. A “non-enrolling site” may be one that does not contain any subjects or patients.
One measure of number of sites per 100 subjects is the total number of trial sites divided by the total number of enrolled subjects, multiplied by 100.
One measure of number of countries per 100 subjects is the total number of unique countries in the trial divided by the total number of enrolled subjects, multiplied by 100.
Regarding trial cycle times, one measure of the FPI to LPI for a trial is the time it takes for enrollment to take place across the whole trial. Some trials may be non-enrolling, in which case this metric would not be measured. In some instances, the system may use a minimum FPI to LPI, so if the actual FPI to LPI is less than such minimum (such as one month), FPI to LPI may be set to that minimum. (If a specified minimum value meets suitable criteria for top quality performance, then using a lower FPI to LPI value may bring no benefit and may even be counterproductive.)
One measure of DB (database) open to FPI is the total number of days from the date of the database launch to the date of enrollment for the first patient in a given trial
One measure of last patient visit (LPV) to DB lock is the time from the last patient last visit to the maximum lock date for all data points in a given trial.
Regarding monitoring or study conduct, screen failure rate may have site-level and study-level measures. One measure of site-level screen failure rate is the number of screen failures (subjects that attempted to enter a trial site but did not enroll) divided by the number of subjects that attempted to enter the trial site (number of screen failures plus number of enrolled subjects). One measure of trial-level screen failure rate is the sum of site-level screen failures divided by the sum of site-level subjects that attempted to enter the trial (sum of site-level screen failures plus the sum of site-level enrolled subjects).
On-site monitoring rate may have a site-level calculation and a trial-level calculation. One measure of a site-level on-site monitoring rate is the total number of days a monitor is on a site divided by the total number of active days for the site. One measure of a trial level on-site monitoring rate is the sum of the site-level on-site days divided by the sum of the site-level active days.
Data correction rate may have a site-level calculation and a trial-level calculation. One measure of a site-level data correction rate is the number of changed data points for a site divided by the total number of data points for that site. One measure of a trial level data correction rate is the sum of site-level changed data points divided by the sum of site-level data points for a given trial.
Regarding cost metrics, principal investigator (PI) grant cost per patient (also known as adjusted grant total per patient) may have a site-level calculation and a trial-level calculation. One measure of PI grant cost per patient is the grant total minus an IRB fee, a fixed fee, a failure fee, a grant adjustment, and lab cost (all in US Dollars), divided by the total number of patients for a given trial site. One measure of a trial level PI grant cost per patient is the sum of site-level adjusted grant total divided by the total number of patients for a given trial.
Site cost per patient may have a site-level calculation and a trial-level calculation. One measure of site cost per patient is the grant total in US Dollars divided by the total number of patients for a given study site. One measure of a trial level site cost per patient is the sum of site-level grant total divided by the total number of patients for a given trial.
This list of categories and metrics is not exclusive or exhaustive. Other categories of metrics and other metrics in these categories may be used to measure operational performance.
Referring back to
An appropriate transformation is any procedure that enables the comparison of client data to industry data on the same scale. The embodiments in the next few paragraphs are not exhaustive. Each of them is based on either the standardized normal distribution or the empirical cumulative distribution function.
One embodiment may use a z-transform, which may be calculated by taking each data point, subtracting the mean, and then dividing by the standard deviation. This converts the distribution of the data to a standardized distribution that has mean equal to 0 and standard deviation equal to 1, which allows client data to be compared to industry data on the same scale.
Embodiments of the present invention may use different statistical measures of “center” and “variability” for calculating the appropriate transformation. The z-transform embodiment described above uses the arithmetic mean as the statistical measure of center and the standard deviation as the statistical measure of variability.
Another embodiment may use the median as the statistical measure of center and the median absolute deviation as the statistical measure of variability. This embodiment is considered more robust to outliers than the z-transform.
Another embodiment may use a Winsorized mean as the statistical measure of center and a Winsorized standard deviation as the statistical measure of variability. A Winsorized distribution sets all values outside a specified percentile to that percentile value. Thus, an 80% Winsorized distribution sets all values above the 90th percentile to the value corresponding to the 90th percentile and all values below the 10th percentile to the value corresponding to the 10th percentile. Then the mean (i.e., Winsorized mean) and standard deviation (i.e., Winsorized standard deviation) of this modified distribution are calculated and used to compute the z-transform.
Similar to the Winsorized distribution is a trimmed distribution. A trimmed distribution truncates the tails by discarding all values outside a specified percentile. Thus, a 10% trimmed distribution deletes all values above the 90th percentile and below the 10th percentile. Then the mean and standard deviation of this modified distribution are calculated and used to compute the z-transform.
Other variations on the statistical measures of center and variability may be used, including using the mean absolute deviation instead of the standard deviation to calculate a modified z-transform.
Another embodiment may be based on the empirical cumulative distribution function (ECDF). This approach may use the industry data to calculate an ECDF and then evaluate the client data relative to this ECDF, assigning a suitable score to the client data. This approach transforms the client data into a score that corresponds to a position on the industry ECDF. An embodiment based on the ECDF may incorporate components that use methodology derived from kernel density estimation, polynomial spline regression, Bernstein polynomial estimation, and other methods applicable to ECDFs.
The ECDF determines the position of the client data in the distribution of the relevant industry data. For example, if the industry data consists of ten values—14, 18, 23, 28, 34, 42, 50, 59, 66, 72—and the client data value is 64, then the ECDF score of the client data value is 0.80, because the client data value is greater than or equal to eight of the ten industry values, and this fraction constitutes 8/10=0.80 of the industry data values.
The transformation value for each statistic for client data (however calculated as just described), Cn, may then be transmitted to data comparator and visualizer 40, a part of which is shown in
The parts and blocks shown in
Reference is now made to
Besides the operations shown in
One benefit of the present invention is that it provides a client with information regarding how it stands against others in its peer group or other grouping of competitors, both overall and for different therapeutic areas, indications, and phases. The present invention differs from other systems that provide information about clinical trial operational performance. For example, those systems may not use transformation values, thus making it difficult to compare different scopes of data.
Aspects of the present invention may be embodied in the form of a system, a computer program product, or a method. Similarly, aspects of the present invention may be embodied as hardware, software or a combination of both. Aspects of the present invention may be embodied as a computer program product saved on one or more computer-readable media in the form of computer-readable program code embodied thereon.
For example, the computer-readable medium may be a computer-readable signal medium or a computer-readable storage medium. A computer-readable storage medium may be, for example, an electronic, optical, magnetic, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination thereof.
A computer-readable signal medium may include a propagated data signal with computer-readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electromagnetic, optical, or any suitable combination thereof. A computer-readable signal medium may be any computer-readable medium that is not a computer-readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
Computer program code in embodiments of the present invention may be written in any suitable programming language. The program code may execute on a single computer, or on a plurality of computers. The computer may include a processing unit in communication with a computer-usable medium, wherein the computer-usable medium contains a set of instructions, and wherein the processing unit is designed to carry out the set of instructions.
The above discussion is meant to be illustrative of the principles and various embodiments of the present invention. Numerous variations and modifications will become apparent to those skilled in the art once the above disclosure is fully appreciated. It is intended that the following claims be interpreted to embrace all such variations and modifications.
Number | Date | Country | |
---|---|---|---|
Parent | 29535475 | Aug 2015 | US |
Child | 15498292 | US |