Next Article in Journal
Fever, Tachypnea, and Monocyte Distribution Width Predicts Length of Stay for Patients with COVID-19: A Pioneer Study
Previous Article in Journal
Circulating Tumor Cell Detection during Neoadjuvant Chemotherapy to Predict Early Response in Locally Advanced Oropharyngeal Cancers: A Prospective Pilot Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Personalized Risk-Based Screening Design for Comparative Two-Arm Group Sequential Clinical Trials

Department of Biostatistics and Medical Informatics, University of Wisconsin-Madison, Madison, WI 53705, USA
J. Pers. Med. 2022, 12(3), 448; https://doi.org/10.3390/jpm12030448
Submission received: 12 February 2022 / Revised: 7 March 2022 / Accepted: 10 March 2022 / Published: 12 March 2022

Abstract

:
Personalized medicine has been emerging to take into account individual variability in genes and environment. In the era of personalized medicine, it is critical to incorporate the patients’ characteristics and improve the clinical benefit for patients. The patients’ characteristics are incorporated in adaptive randomization to identify patients who are expected to get more benefit from the treatment and optimize the treatment allocation. However, it is challenging to control potential selection bias from using observed efficacy data and the effect of prognostic covariates in adaptive randomization. This paper proposes a personalized risk-based screening design using Bayesian covariate-adjusted response-adaptive randomization that compares the experimental screening method to a standard screening method based on indicators of having a disease. Personalized risk-based allocation probability is built for adaptive randomization, and Bayesian adaptive decision rules are calibrated to preserve error rates. A simulation study shows that the proposed design controls error rates and yields a much smaller number of failures and a larger number of patients allocated to a better intervention compared to existing randomized controlled trial designs. Therefore, the proposed design performs well for randomized controlled clinical trials under personalized medicine.

1. Introduction

Personalized medicine is a new paradigm motivated by the possibility that patients’ response to a particular treatment is heterogeneous, which may be due to biological covariates. Only a subset of patients is sensitive to, and benefit from, the treatment. Thus, a traditional one-size-fits-all remedy may not be the best option for some patients, even though the standard of care for a disease generally has a well-established track record. In the era of personalized medicine, molecularly targeted agents have been developed for disease treatment and prevention, e.g., trastuzumab [1,2], crizotinib [3,4], and erlotinib [5,6]. Novel statistical methods and clinical trial designs have been proposed for the novel targeted therapy. Park [7] reviewed statistical methods evaluating the effect of the targeted therapy with a certain genetic mutation on multiple disease types. Biomarker-based clinical trial designs have been proposed to address the one-size-fits-all issue [8,9,10,11]. Adaptive enrichment designs propose the enrichment rule to identify the patients who are expected to get more benefit from the experimental treatment and restrict the enrollment adaptively to the treatment sensitive patients [12,13]. In this paper, we are interested in how personalized medicine works on randomization of treatments for clinical trials.
Randomization is critical in clinical trials to remove any systematic bias for detecting the treatment effect and thus powerful to ensure validity in the comparative clinical trials. Most of randomized controlled trials use a fixed randomization to allocate participants to the treatments being compared, i.e., the allocation ratio 1:1 or 2:1 is commonly used in comparative two-arm clinical trials. The fixed randomization makes simple to execute the clinical trials. However, in the era of the personalized medicine, it would make investigators hesitant to assign equal number of patients to each treatment if the trial enrolls patients regardless of enrollment restriction to a targeted subgroup based on empirical evidence of the efficacy of the treatments. As an effective approach to address the ethical problem, adaptive randomization assigns more future patients to the better performing treatment based on the accumulating information on patients’ response to the treatments. Using the skewed allocation probability, response-adaptive randomization (RAR) designs for binary response trials have been proposed [14,15,16,17]. The optimal allocation probability to treatments was proposed in that sample size is minimized [18] or total number of failures is minimized [19,20]. To incorporate patients’ covariate information in RAR designs, the response probability conditioning on the covariates is estimated for RAR [17,21,22,23,24].
In this paper, we propose a personalized risk-based screening design for comparative two-arm group sequential clinical trials. The proposed design follows the group sequential manner with the first look used for a burn-in stage. It collects some preliminary data to facilitate the regression fitting and adaptive decision of the intervention assignment for the next stages. We propose personalized randomization using a Bayesian covariate-adjusted response-adaptive randomization based on adaptive regression of response on informative covariates to randomize a patient with the given vector of covariates to the intervention from which the patient is expected to get more benefit based on the accumulating information. Using risk factors to build the personalized risk-based allocation probability, the design provides individually tailored randomization of screening modality. Moreover, we propose a group sequential test in personalized allocation and Bayesian monitoring rule to compare screening effects and maintain the error rates.
The rest of this paper is organized as follows. In Section 2, we describe a motivating trial for cancer screening and propose a design structure, probability model, and methods for the personalized screening trial design. In Section 3, we evaluate the operating characteristics of the proposed design using simulation studies. We provide discussion in Section 4.

2. Personalized Risk-Based Screening Design

2.1. Motivating Trial

Tomosynthesis Mammographic Imaging Screening Trial (TMIST) is a Phase III trial study, which starts on July 2017 and will be completed by August 2030 (The study identifier is NCT03233191). TMIST randomizes women between the ages of 45 and 74 to either tomosynthesis mammography (3D mammography) or standard digital mammography (2D mammography) with equal probability and evaluates the mammographic accuracy for breast cancer screening. The primary endpoint of the study is the incidence of advanced breast cancer, and the trial was designed to compare the proportion of women diagnosed with an advanced breast cancer between two screening modalities. In an era of personalized medicine, it is essential to develop methods and trial designs for personalized risk-based screening using breast density, tumor subtyping, and genomics [25].

2.2. Design Structure

Motivated by TMIST considering two screening disparities, digital breast tomosynthesis mammography and standard digital mammography, we consider a comparative group sequential clinical trial with patients individually randomized to experimental treatment A or control B based on accumulating data.
Our design enrolls a maximum of N patients sequentially in cohorts of sizes n 1 , , n K with N = k = 1 K n k . The design uses a Bayesian group sequential monitoring, described in Section 2.5 below, for superiority or futility at interims to compare A to B in the adaptively randomized patients. The schema of the design is shown in Figure 1. The trial begins by enrolling patients according to the eligibility criteria for the first cohort of n 1 patients. It randomizes the patients to A or B with equal probability. When the n 1 patients have been enrolled and their outcomes are available, the superiority or futility of the experimental treatment A against the control B is monitored at the first interim. If the monitoring shows that A is superior or futile, the trial is terminated. However, if the trial is not stopped early, then we fit the regression model of response on a vector of patients’ characteristics and treatment to estimate the personalized allocation probability, given in (2) in Section 2.4 below. The allocation probability is updated to randomize the treatment adaptively and individually for the next enrollment of the second cohort. This procedure is repeated until the end of the trial. If the maximum sample size N is reached and the last patient’s outcome has been evaluated, a final analysis is performed.

2.3. Probability Model

Let G be an indicator of treatment group taking 1 for receiving experimental treatment A and 0 for receiving control B. Let Y be a binary indicator of events, e.g., deaths. For each patient, we assume that a vector of informative covariates x is available at enrollment.
We describe a probability distribution for Y assuming a probit regression model
Pr ( Y = 1 | G , x ) = Φ ( x ˜ β + G x ˜ γ ) ,
where Φ ( · ) denotes the cumulative distribution function of standard normal variable, x ˜ = ( 1 , x ) and θ ( β , γ ) denotes the regression coefficient parameter vector. Specifically, β is the vector of covariate main effects and γ is the vector of interaction effects between treatments and covariates including the main experimental versus control effect. Back to the motivating trial, Y is the indicator of having breast cancer. The probability in (1) indicates the chance of having an advanced breast cancer for the given screening method G and a vector of patients’ characteristics x . To interpret the breast cancer risk for screening, electronic health record, breast density, age, tumor subtyping, first-degree breast cancer family history, and genomics are candidates of the predictive covariates in the risk prediction model [26,27,28,29,30,31].
Assigning β and γ normal priors, the parameters are estimated by Bayesian inference. We used LearnBayes R package to fit Bayesian probit regression model.

2.4. Personalized Allocation for Adaptive Randomization

For each k = 1 , , K 1 , let D k be an accumulating data at the kth interim, i.e., a set of Y , G , x over the k cohorts. Let p A ( x ) = Pr ( Y = 1 | G = 1 , x ) , and p B ( x ) = Pr ( Y = 1 | G = 0 , x ) . Then, p A ( x ) p B ( x ) = Φ ( x ˜ β + x ˜ γ ) Φ ( x ˜ β ) , which is a function of unknown parameter θ = ( β , γ ) . To assign more patients to the better performing personalized treatment, we are interested in quantifying a likelihood of a patient with x benefiting more from the treatment A than B, i.e., p A ( x ) p B ( x ) < 0 . Let p k 1 ( x ) = Pr ( p A ( x ) < p B ( x ) | D k 1 ) denote the posterior probability that a patient with covariates x is less likely to have an event under treating A than treating B based on accumulating data D k 1 . Assuming normal prior on θ for Bayesian probit regression model in Section 2.3, samples of θ are generated from the posterior distribution
Pr ( θ | data ) = lik ( data | θ ) prior ( θ ) lik ( data | θ ) prior ( θ ) d θ
where lik ( data | θ ) denotes the likelihood function and prior ( θ ) denotes the prior distribution of parameter θ , and the posterior probability p k 1 ( x ) is calculated. We provide how to compute the posterior probability p k 1 ( x ) in Appendix A. The posterior probability is to reflect the personalized medicine, and patients’ characteristics x are incorporated into the posterior probability Pr ( p A < p B | D k 1 ) used in Bayesian adaptive randomization [32] Then, we define the probability of randomizing a patient with covariates x in the kth cohort to the treatment A as
π k , A ( x ) = p k 1 ( x ) p k 1 ( x ) + 1 p k 1 ( x ) .
This is an option considering the personalized allocation probability, which is a type of covariate-adjusted response adaptive randomization (CARA). To emphasize in the randomization ratio that patients can respond differently to the treatments, we prefer what we call personalized randomization over CARA. We use this allocation probability (2) for the proposed design to perform personalized randomization.
Alternative option is to consider another type of CARA given by
π k , A ( x ) = 1 p k 1 , A ( x ) 1 p k 1 , A ( x ) + 1 p k 1 , B ( x )
where p k 1 , A ( x ) = Pr ( Y = 1 | G = 1 , x , D k 1 ) and p k 1 , B ( x ) = Pr ( Y = 1 | G = 0 , x , D k 1 ) . The personalized allocation probability (3) uses the estimated response rates of treatment A and B denoted by p k 1 , A ( x ) and p k 1 , B ( x ) , which are obtained by posterior mean of parameter. In our motivating screening trial, the response is an event such as death. To build the personalized allocation probability which is skewed to patients who get more benefit, the allocation probability is proportional to 1 p k 1 , · ( x ) = Pr ( Y = 0 | G , x , D k 1 ) instead of Pr ( Y = 1 | G , x , D k 1 ) . This is the modified version using Bayesian inference from optimal allocation probability suggested by Rosenberger et al. [19].
The personalized allocation probabilities (2) and (3) are updated throughout the clinical trials based on the accumulating data. They change the treatment allocation probability and adaptively randomize more patients to the treatment arm that is superior according to the patients’ characteristics. Back to the motivating trial, using the risk predictive model in Section 2.3, we are able to perform data-driven personalized randomization. It builds the personalized risk-based allocation probability and randomizes more patients to the superior screening modality individually. The personalized randomization makes more reasonable in ethics and help clinicians and clinical trialists get more out of randomized clinical trials.

2.5. Group Sequential Test in Personalized Randomization

To effectively use the personalized randomizationin group sequential designs allowing early stopping, it is critical to preserve the overall type I error rate. As the response adaptive randomization (RAR) including CARA is considered based on the observed data, potential selection bias can occur. Moreover, the bias would be more serious if CARA is used when there exists an effect of informative covariates. Park [33] shows that group sequential designs using CARA are influenced by prognostic covariates and the overall type I error rate is not controlled. To address the issue of type I error rate inflation from using the personalized allocation and accommodate the possible change in eligibility of patients during the trial, it is required to propose an elaborate test statistic which preserves the error rates.
At the kth analysis, the trial enrolls patients of k cohorts sequentially. Based on the accumulated data D k from k successive cohorts, which might consist of k heterogeneous cohorts, the kth interim monitoring determines go or no-go of the trial. Let Δ k be an expected subgroup-averaged treatment effect based on the kth cohort. Assuming that x determines the subgroups, we suppose that there are I k subgroups in the kth cohort denoted by S i , i = 1 , , I k . In the case where x is continuous, the dichotomization can be considered to define the subgroups, e.g., young and old groups for the age variable. A comparative treatment effect of the kth cohort is obtained by
Δ k = i = 1 I k { p A ( x ) p B ( x ) } Pr ( x ) I ( x S i ) ,
where I ( · ) denotes the indicator function. It is a function of parameter θ and indicates the expected difference of the response probability with respect to x over the kth cohort. Then, a group sequential test statistic is proposed as the weighted sum of the comparative treatment effect based on k cohorts, i.e.,
T k = j = 1 k n j Δ j j = 1 k n j .
As the comparative treatment effect Δ k is calculated by marginalizing the difference of response probability with respect to x , the test statistic T k does not indicate the treatment effect of the individual patient. It indicates the overall treatment effect based on the accumulating data at the kth analysis.
When there are a few covariates, all possible combinations of subgroups are considered to obtain the comparative treatment effect Δ k . However, with more covariates, to avoid any computational burden or complexity, we suggest identifying the covariates whose main effect is significant so that they determine the subgroups in the kth cohort for the calculation of Δ k .
Let δ 1 denote the minimal improvement for the experimental treatment to be deemed superior to the control and δ 2 denote the minimal improvement so that the experimental treatment is considered worthy of further investigation. The values of δ 1 and δ 2 are pre-specified by clinicians or the study hypothesis. Let ϵ i , i = 1 , 2 , 3 be the pre-specified probability cutoffs for superiority and futility monitoring rule. They are design parameters obtained by preliminary simulation-based calibration, where ϵ 1 and ϵ 3 control type I error rate α and ϵ 2 controls type II error rate β . To save several rounds of calibrations, the initial cutoff values of ϵ 1 and ϵ 3 were selected as one minus target type I error rate, and the initial cutoff of ϵ 2 was selected as one minus target type II error rate. To make sense with experts’ experience and knowledge, the survey results can be used to determine the level of evidence and calibrate for the monitoring rules [34]. If the type I error rate is lower/higher than the desirable level, we decrease/increase the value of ϵ 1 and ϵ 3 , and if the calculated type II error rate is lower/higher than the desirable level, we decrease/increase the value of ϵ 2 . We repeat this calibration process until the desirable type I and II error rates are obtained. Then, the calibration procedure determines the cutoffs carefully to adjust the multiplicity of testing repeatedly over time and thus maintain the overall type I and II error rates at the nominal levels. It is widely used in Bayesian sequential designs [13,35,36,37]. Shi and Yin [38] provides the unified framework for the calibration procedure to search the cutoffs effectively.
Then, the Bayesian sequential monitoring rule is described as follows.
  • At each interim k = 1 , , K 1 , the trial is terminated for superiority if Pr ( T k < δ 1 | D k ) > ϵ 1 , or the trial is terminated for futility if Pr ( T k > δ 2 | D k ) > ϵ 2 .
  • When k = K (i.e., at final analysis), we argue that A is superior to B if Pr ( T K < δ 1 | D K ) > ϵ 3 , and otherwise, A is not superior to B.
The posterior probabilities Pr ( T k < δ 1 | D k ) and Pr ( T k > δ 2 | D k ) are comupted by Bayesian inference (see Appendix A). The values of δ 1 and δ 2 are not necessarily to be the same in the decision rules. The proposed rule allows unequal values of δ 1 and δ 2 to increase the flexibility of the study.

3. Simulation Study

We assumed maximum sample size 210, which yielded 80 % power to detect a response rate of 0.3 versus a null response rate of 0.5 based on a two-sample t-test with one-sided significance level α = 0.05 under the traditional randomized clinical trial using the fixed equal randomization. Each patient was randomized to either experimental treatment A or control B. Two interim analyses were performed when the first 70 and 140 enrolled patients completed the evaluation of the response. At interims, we monitored the superiority or futility of the treatment A against B. A final analysis was performed after the last patient completed follow-up to argue the experimental treatment A is superior to B.
In the following, we first identified the challenging issues in personalized allocation based on the conventional group sequential test. Next, we investigated the performance of the proposed design and verified if the issues are addressed.

3.1. Type I Error Rate Inflation

We considered four group sequential clinical trial designs: traditional randomization with 1:1 (Trad), response-adaptive randomization without incorporating covariates (RAR), and covariate-adjusted response-adaptive randomization using (2) and (3) (CARA1 and CARA2, respectively). For all designs, we used the fixed equal randomization for the first cohort of 70 patients but changed the randomization scheme at the first interim according to the design. Trad kept the fixed equal randomization throughout the trial, but other designs updated the allocation probability at each interim to randomize the patients for the next cohorts. RAR used the allocation probability which Rosenberger et al. [19] proposes. CARA1 and CARA2 used the personalized allocation probability described in (2) and (3), respectively. To make comparable, four designs performed the conventional group sequential test based on a chi-square test. We set the overall type I error rate to 0.05 for the group sequential test. The O’Brien–Fleming alpha spending function was used to specify the stopping boundaries for the sequential test in Trad, RAR, CARA1, and CARA2. To estimate the personalized allocation probability in CARA1 and CARA2, we fitted the Bayesian probit regression model assuming normal priors with the mean vector of the maximum likelihood estimate and diagonal covariance matrix with diagonal elements 4. The choice of prior was to avoid using the vague prior and help the error rates less inflated [33]. When we implemented the Bayesian inference, we ran 10,000 iterations and discarded the first 5000 iterations as burn-in.
We considered two binary covariates x = ( x 1 , x 2 ) which were generated from a Bernoulli distribution with response probability 0.5 . There were four possible subgroups of patients determined by the two covariates, i.e., patients with x = ( 1 , 1 ) , ( 1 , 0 ) , ( 0 , 1 ) , or ( 0 , 0 ) . Then, the response Y was generated from a Bernoulli distribution with the probability
Pr ( Y = 1 | G , x ) = Φ ( β 0 + β 1 x 1 + β 2 x 2 + γ 0 G + γ 1 G x 1 + γ 2 G x 2 ) .
We considered twenty scenarios, and the true parameters generating response in (6) were described in Table 1.
Table 1 provides the summary of the response rates p A = Pr ( Y = 1 | G = 1 , x ) and p B = Pr ( Y = 1 | G = 0 , x ) for the overall group and four subgroups. Scenarios 1–9 describe null scenarios where both an experimental treatment A and the control B have no difference in the response. Scenarios 10–20 describe alternative scenarios where the main experimental versus control effect exists. Scenario 1 shows the same response for A and B as 0.5 regardless of patients’ characteristics or treatment assignment, i.e., it has no main effect of covariates or the main experimental versus control effect. Scenarios 2 and 3 have the main effect of the first covariate (i.e., β 1 0 ), while Scenarios 4 and 5 have the main effect of the second covariates (i.e., β 2 0 ). Scenarios 6–9 have nonzero coefficients β 1 and β 2 implying that two covariates x 1 and x 2 have the main effect on the response. Thus, in Scenarios 2–9, the response rate depends on the covariates but does not depend on the treatment assignment. They indicate the cases where there is an effect of prognostic covariates. Scenario 10 does not have any effects of covariates but has the main experimental versus control effect, and it is the case the experimental treatment A has better efficacy in response (i.e., smaller response) than the control B. Compared to Scenarios 10–12 we consider the additional effect of prognostic covariate x 1 (i.e., β 1 0 ) to the main experimental versus control effect, while Scenarios 13 and 14 consider the additional effect of predictive covariate x 1 (i.e., γ 1 0 ) to the main experimental versus control effect. Furthermore, in Scenario 15, the first covariate x 1 has both prognostic and predictive effects. Scenarios 16 and 17 have an effect of prognostic covariate x 2 and the effect of treatment assignment. In Scenario 18, the second covariate x 2 has both prognostic and predictive effects. In Scenarios 19 and 20, both x 1 and x 2 have prognostic and predictive effects. Depending on the effects of prognostic or predictive covariates, in Scenarios 10–20, particular subgroups with the covariate profile are more likely to get benefit from one of the treatments than the other treatment. To better understand the subgroups of the covariate profile, we call A (or B)-sensitive patients if the patients with the covariate profile x are expected to respond better to A (or B) but not respond to B (or A). The better treatment for A-sensitive patients is A, and the better treatment for B-sensitive patients is B. For example, in Scenario 14, patients with x 1 = 0 are A-sensitive; in Scenario 18, patients with x 2 = 0 are A-sensitive; in scenario 19, patients except x = ( 1 , 0 ) are A-sensitive; and in scenario 20, patients with x = ( 1 , 1 ) are B-sensitive and patients with x = ( 0 , 0 ) are A-sensitive.
Table 2 shows the estimated rejection probability to detect the difference of the response rate between treatments A and B based on 1000 simulated trials. The rejection probability under the null scenarios (i.e., Scenarios 1–9) indicates the overall type I error rate, and the rejection probability under the alternative scenarios (i.e., Scenarios 10–20) indicates the power. Trad and RAR preserved the type I error rate at the target level of 0.05 for all null scenarios. In addition, CARA2 worked well to control the overall type I error rate except for Scenarios 5 and 7. Specifically, under CARA2 using the personalized allocation probability (3), the estimated type I error rates were inflated at 10–17% in Scenarios 5 and 7. However, CARA1 failed in most null scenarios when there exists an effect of the prognostic covariate(s). Specifically, CARA1 using the personalized allocation probability (2) led to serious error inflation at 25–40% by the prognostic covariates in Scenarios 5 and 7. To investigate the type I error rate inflation in Scenarios 5 and 7, we looked at the distribution of the subgroups in each treatment arm A or B for all designs. The mean and standard deviation of the allocation probability of the treatment for each subgroup are reported in Table 3. We observed that designs using personalized randomization, e.g., CARA1, CARA2, and BaCARA, led to the large variability of the distributions compared to Trad and RAR which controlled the overall type I error rate. Under CARA1 and CARA2, the conventional group sequential test did not work properly in the presence of the effect of prognostic covariate(s), and we observed large inflations of overall type I error rate. However, under BaCARA, the overall type I error rates were less likely to be inflated, which resulted from the proposed group sequential test statistics considering the differences in treatment effect within subgroups. Depending on the difference in the response rate for each covariate profile and the prevalence of the subgroups, the outcomes were influenced by the covariates.
Under the alternative scenarios, Trad yielded a power which ranged from 0.10 to 0.98 depending on the overall difference between p A and p B . As the power 80% was justified by the difference of 0.2 from the response probability of p B = 0.5 , the power for each scenario varied according to the smaller or larger treatment effect difference and the null response probability p B . RAR generally yielded similar or a little smaller power than Trad. CARA2 showed similar or larger power compared to Trad and RAR in most scenarios (except for Scenario 17). In most scenarios where the treatment effect difference or subgroup effect difference was less than 0.2 (i.e., Scenarios 10–16), CARA1 yielded similar or smaller power compared to Trad and RAR. However, when the treatment effect difference or subgroup effect difference became larger (i.e., in Scenarios 18–20), CARA1 led to much larger power than Trad and RAR. We also provided boxplots of the estimated difference between p A and p B at the final analysis for all designs in Figure 2. Therefore, CARA1 was more sensitive to the prognostic covariates than CARA2 and was more likely to inflate the error rates.
Table 4 shows other operating characteristics of the designs such as the average difference of the number of patients assigned to A and B and the average number of failures (i.e., events) across 1000 simulated trials. Compared to Trad, RAR and CARA change the allocation ratio and randomize more patients to the superior treatment. Under Trad, the averaged difference of the number of patients assigned to A and B ranged from −1.176 to 0.924 with an average of 0.065 across the alternative scenarios. Under RAR, the averaged difference of the number of patients assigned to A and B ranged from 0.482 to 3.082 with an average of 2.023 across the alternative scenarios. Under CARA1, the averaged difference of the number of patients assigned to A and B ranged from 16.808 to 48.808 with an average of 36.466 across the alternative scenarios. Under CARA2, the averaged difference of the number of patients assigned to A and B ranged from 0.316 to 17.598 with an average of 6.654 across the alternative scenarios. CARA1 and CARA2 showed a larger number of patients assigned to the superior treatment A than Trad and RAR. The gain was much larger when CARA1 is considered, which was resulted from the effective use of the personalized allocation probability based on the accumulating data. It also resulted in a smaller number of failures under CARA1 than other designs. Under Trad, the number of failures ranged from 33.47 to 135.25 with an average of 70.03; under RAR, the number of failures ranged from 33.21 to 134.73 with an average of 69.92; and under CARA2, the number of failures ranged from 32.73 to 133.56 with an average of 69.29. All three designs showed similar performance in the number of failures, i.e., CARA1 did not show apparent gain in the number of failures. However, under CARA1, the number of failures ranged from 30.12 to 130.58 with an average of 64.60.
The simulation study tells us that effective use of the personalized allocation probability can lead to the inflation of the overall type I error rate but is more ethical by assigning more patients to the superior treatment and yields a smaller number of failures. Therefore, it is critical to maintain the overall type I error rate in personalized allocation and improve clinical benefit while inheriting the advantages of CARA designs.

3.2. Evaluation of the Proposed Design: Preservation of Type I Error Rate

We observed the inflation of the overall type I error rate using CARA1 in Table 2, which came from the prognostic covariates’ effect and sequential personalized allocation. Using a conventional group sequential test to detect the overall treatment difference in the response probability did not work well when the randomization depended on patients’ characteristics. Patients receiving a certain treatment might not be homogeneous, and they responded differently to the treatment. To accommodate this heterogeneity and control the type I error rate, the group sequential test in personalized allocation was proposed in Section 2.5.
For convenience, we called the proposed design BaCARA, which used the personalized allocation probability (2) to randomize the patients and monitor the treatment effect based on the proposed group sequential test statistic (5) through the Bayesian sequential monitoring rule. We evaluated the operating characteristics of BaCARA through simulations. We followed the same simulation settings as in Table 2, Table 3 and Table 4. To compare the results with Trad, RAR, CARA1, and CARA2, we included the results of BaCARA in the last column of Table 2, Table 3 and Table 4. Assuming the minimal improvements δ 1 = δ 2 = 0 , we calibrated ϵ 1 = 0.995 , ϵ 2 = 0.75 , and ϵ 3 = 0.98 by preliminary simulations to control the error rates under the null scenarios (i.e., Scenarios 1–9 in Table 1) and the alternative scenario with p A = 0.3 and p B = 0.5 (i.e., scenario 10 in Table 1).
We observed in Table 2 that BaCARA preserved the overall type I error rate at the target level of 0.05, implying that the inflation issue of CARA1 was addressed. Compared with Trad and RAR designs, under BaCARA, the overall type II error rates seemed to be controlled well (i.e., in Scenarios 10–20). Similar to CARA1 and CARA2, BaCARA was more powerful than Trad and RAR in Scenarios 18–20 where patients’ response to the treatment was more heterogeneous, i.e., the treatment effect difference or subgroup effect difference was relatively larger. In addition, in Scenario 17 where CARA2 yielded a large inflation of type II error rate, BaCARA showed a large power compared to other designs. Thus, BaCARA improved the performance of CARA1 and CARA2 using the personalized allocation probability in that it preserved the overall type I and II error rates. BaCARA was appropriate to use for group sequential clinical trials incorporating patients’ characteristics into the adaptive randomization.
In Table 4, BaCARA showed better performance in the difference of the number of patients assigned A and B than Trad, RAR, and CARA2, but it had smaller differences of the number of patients assigned A and B than CARA1 in most scenarios. Under BaCARA, the difference of the number of patients assigned A and B ranged from 17.718 to 33.606 with an average of 26.634. However, BaCARA yielded a smaller number of failures across scenarios than Trad, RAR, CARA1, and CARA2. Under BaCARA, the number of failures ranged from 25.82 to 116.36 with an average of 57.84. Such an improvement came from the effective group sequential test as well as the personalized allocation. The proposed design led to the improvement of clinical benefit and provided a better suggestion to effectively use personalized randomization for personalized medicine.

4. Discussion

We proposed a personalized risk-based screening design using Bayesian covariate-adjusted response-adaptive randomization for comparative two-arm clinical trials. Following the group sequential procedure, we adaptively built the personalized allocation probability using the risk factors to randomize more patients to the most desirable individualized intervention and minimize the number of events. We also proposed a new group sequential test to address the challenging issues in the personalized allocation. The proposed Bayesian monitoring rule determined go or no-go of the trial at interims based on accumulating data, and the proposed design preserved the type I error rate through the calibrated cutoffs for the Bayesian monitoring rule.
We compared the performance of the proposed design to the randomized controlled trial designs such as traditional, RAR, and CARA designs. Even though RAR design assigned more patients to the better performing intervention and thus was ethical compared to the traditional randomized controlled trial design, the expected number of failures was not different, and the improvement of clinical benefit was not clear. In addition, in RAR, all eligible patients were enrolled and randomized without any restriction considering patients’ characteristics, which was not appropriate in personalized medicine. By incorporating patients’ characteristics into randomization, CARA design led to a larger allocation of patients to the better performing intervention than RAR design. However, CARA designs could be sensitive to the prognostic covariate effect and inflate the overall type I error rate. Furthermore, in our simulations, it was not clear to achieve a significant improvement in clinical benefit (i.e., the smaller number of events) compared to the traditional and RAR designs. Taking all of the above into account, the proposed design was the most appropriate to use for two-arm personalized screening clinical trials.
The proposed design is flexible and extended to the followings. First, assuming that informative covariates are not specified at the beginning of the trial, covariate selection methods can be carried out in the burn-in stage. The selected covariates with the significant effect are used in the remaining stages to randomize and test the screening effect. Second, our Bayesian sequential monitoring rule is flexible and can be modified according to the study objectives. For example, additional monitoring rules based on surrogate or safety endpoint can be included to make a data-driven decision throughout the trials. This also allows us to learn health systems along with the trials. Third, personalized randomization can be generalized for multi-arm trials, and each arm is compared to the control using the proposed test. For example, to calculate the allocation probability (2) of randomizing a patient with x to the treatment A, the posterior probability of p A ( x ) < p B ( x ) for comparing with the control B is replaced with the posterior probability that the treatment A offers the minimum response rates among all treatment arms [39]. Villar et al. [40], Ryan et al. [41], and Viele et al. [42] provide some directions under consideration for the multi-arm trials.

Funding

This work was supported in part by University of Wisconsin-Madison Office of the Vice Chancellor for Research and Graduate Education.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Appendix A. Calculation of Posterior Probability

We have used Bayesian inference for personalized randomization (Section 2.4) and monitoring rule (Section 2.5). In the Bayesian probit regression model, we compute the posterior distribution instead of computing the maximum likelihood estimator of the parameter. As mentioned in Section 2.3, we assume that the regression coefficient vector θ = ( β , γ ) follows the normal prior distributions. Given the observed data D , the posterior distribution of parameter θ is obtained to be proportional to prior distribution of parameter times the likelihood function. Then, the posterior sample is drawn from the posterior distribution using data augmentation and Gibbs sampling [43,44].
Given the accumulated data D k 1 and p-dimensional covariate vector x , we compute
p k 1 ( x ) = R 2 p + 2 F A , x ( g B , x ( θ ) | D k 1 ) f B , x ( θ | D k 1 ) d θ
where g B , x ( θ ) = Pr ( Y = 1 | G = 0 , x ) = Φ ( x ˜ β ) , F A , x ( z | D k 1 ) denotes the posterior cumulative distribution of Pr ( Y = 1 | G = 1 , x ) = Φ ( x ˜ β + G x ˜ γ ) evaluated at z , and f B , x ( θ | D K 1 ) denotes the posterior density of g B , x ( θ ) . Then, (A1) is approximated by a Monte Carlo simulation, i.e., the posterior probability p k 1 ( x ) is approximated by average of I { p A ( x ) < p B ( x ) } = I { Φ ( x ˜ β + G x ˜ γ ) Φ ( x ˜ β ) < 0 } over the posterior sample θ .
Similarly, for some value of δ , Pr ( T k < δ | D k ) is approximated by the average of I ( T k < δ ) over the posterior sample based on data D k , where T k is the weighted sum of Δ j , j = 1 , , k . In both cases, the posterior samples are obtained easily by using the R package LearnBayes.

References

  1. Slamon, D.J.; Leyland-Jones, B.; Shak, S.; Fuchs, H.; Paton, V.; Bajamonde, A.; Fleming, T.; Eiermann, W.; Wolter, J.; Pegram, M.; et al. Use of chemotherapy plus a monoclonal antibody against HER2 for metastatic breast cancer that overexpresses HER2. N. Engl. J. Med. 2001, 344, 783–792. [Google Scholar] [CrossRef]
  2. Gajria, D.; Chandarlapaty, S. HER2-amplified breast cancer: Mechanisms of trastuzumab resistance and novel targeted therapies. Expert Rev. Anticancer Ther. 2011, 11, 263–275. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Scagliotti, G.; Stahel, R.A.; Rosell, R.; Thatcher, N.; Soria, J.C. ALK translocation and crizotinib in non-small cell lung cancer: An evolving paradigm in oncology drug development. Eur. J. Cancer 2012, 48, 961–973. [Google Scholar] [CrossRef] [PubMed]
  4. Gandhi, L.; Jänne, P.A. Crizotinib for ALK-rearranged non–small cell lung cancer: A new targeted therapy for a new target. Clin. Cancer Res. 2012, 18, 3737–3742. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Piperdi, B.; Perez-Soler, R. Role of Erlotinib in the Treatment of Non-Small Cell Lung Cancer. Drugs 2012, 72, 11–19. [Google Scholar] [CrossRef] [Green Version]
  6. Landi, L.; Cappuzzo, F. Experience with erlotinib in the treatment of non-small cell lung cancer. Ther. Adv. Respir. Dis. 2015, 9, 146–163. [Google Scholar] [CrossRef] [PubMed]
  7. Park, Y. Review of Phase II Basket Trials for Precision Medicine. Ann. Biostat. Biom. Appl. 2019, 2. [Google Scholar] [CrossRef]
  8. Mandrekar, S.J.; Sargent, D.J. Clinical trial designs for predictive biomarker validation: One size does not fit all. J. Biopharm. Stat. 2009, 19, 530–542. [Google Scholar] [CrossRef] [Green Version]
  9. Mandrekar, S.J.; Sargent, D.J. Clinical trial designs for predictive biomarker validation: Theoretical considerations and practical challenges. J. Clin. Oncol. 2009, 27, 4027–4034. [Google Scholar] [CrossRef] [Green Version]
  10. Trippa, L.; Alexander, B.M. Bayesian Baskets: A Novel Design for Biomarker-Based Clinical Trials. J. Clin. Oncol. 2017, 35, 681–687. [Google Scholar] [CrossRef]
  11. Hu, C.; Dignam, J.J. Biomarker-driven oncology clinical trials: Key design elements, types, features, and practical considerations. JCO Precis. Oncol. 2019, 1, 1–12. [Google Scholar] [CrossRef] [PubMed]
  12. Simon, N.; Simon, R. Adaptive enrichment designs for clinical trials. Biostatistics 2013, 14, 613–625. [Google Scholar] [CrossRef] [PubMed]
  13. Park, Y.; Liu, S.; Thall, P.; Yuan, Y. Bayesian group sequential enrichment designs based on adaptive regression of response and survival time on baseline biomarkers. Biometrics 2021, 1–12. [Google Scholar] [CrossRef] [PubMed]
  14. Wei, L.; Durham, S. The randomized play-the-winner rule in medical trials. J. Am. Stat. Assoc. 1978, 73, 840–843. [Google Scholar] [CrossRef]
  15. Eisele, J.R. The doubly adaptive biased coin design for sequential clinical trials. J. Stat. Plan. Inference 1994, 38, 249–261. [Google Scholar] [CrossRef]
  16. Hu, F.; Zhang, L.X. Asymptotic properties of doubly adaptive biased coin designs for multitreatment clinical trials. Ann. Stat. 2004, 32, 268–301. [Google Scholar] [CrossRef]
  17. Villar, S.S.; Rosenberger, W.F. Covariate-adjusted response-adaptive randomization for multi-arm clinical trials using a modified forward looking Gittins index rule. Biometrics 2018, 74, 49–57. [Google Scholar] [CrossRef] [Green Version]
  18. Neyman, J. On the Two Different Aspects of the Representative Method: The Method of Stratified Sampling and the Method of Purposive Selection. J. R. Stat. Soc. 1934, 97, 558–606. [Google Scholar] [CrossRef]
  19. Rosenberger, W.F.; Stallard, N.; Ivanova, A.; Harper, C.N.; Ricks, M.L. Optimal adaptive designs for binary response trials. Biometrics 2001, 57, 909–913. [Google Scholar] [CrossRef]
  20. Tymofyeyev, Y.; Rosenberger, W.F.; Hu, F. Implementing optimal allocation in sequential binary response experiments. J. Am. Stat. Assoc. 2007, 102, 224–234. [Google Scholar] [CrossRef]
  21. Rosenberger, W.F.; Vidyashankar, A.; Agarwal, D.K. Covariate-adjusted response-adaptive designs for binary response. J. Biopharm. Stat. 2001, 11, 227–236. [Google Scholar] [CrossRef] [PubMed]
  22. Thall, P.F.; Wathen, J.K. Covariate-adjusted adaptive randomization in a sarcoma trial with multi-stage treatments. Stat. Med. 2005, 24, 1947–1964. [Google Scholar] [CrossRef] [PubMed]
  23. Eickhoff, J.C.; Kim, K.; Beach, J.; Kolesar, J.M.; Gee, J.R. A Bayesian adaptive design with biomarkers for targeted therapies. Clin. Trials 2010, 7, 546–556. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Hu, J.; Zhu, H.; Hu, F. A unified family of covariate-adjusted response-adaptive designs based on efficiency and ethics. J. Am. Stat. Assoc. 2015, 110, 357–367. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Lee, C.; McCaskill-Stevens, W. Tomosynthesis mammographic Imaging Screening Trial (TMIST): An invitation and opportunity for the National Medical Association Community to shape the future of precision screening for breast cancer. J. Natl. Med Assoc. 2020, 112, 613–618. [Google Scholar] [CrossRef] [PubMed]
  26. Wu, Y.; Fan, J.; Peissig, P.; Berg, R.; Tafti, A.P.; Yin, J.; Yuan, M.; Page, D.; Cox, J.; Burnside, E.S. Quantifying predictive capability of electronic health records for the most harmful breast cancer. In Medical Imaging 2018: Image Perception, Observer Performance, and Technology Assessment; International Society for Optics and Photonics: Houston, TX, USA, 2018; Volume 10577, p. 105770J. [Google Scholar]
  27. Tice, J.A.; Cummings, S.R.; Smith-Bindman, R.; Ichikawa, L.; Barlow, W.E.; Kerlikowske, K. Using clinical factors and mammographic breast density to estimate breast cancer risk: Development and validation of a new predictive model. Ann. Intern. Med. 2008, 148, 337–347. [Google Scholar] [CrossRef] [Green Version]
  28. Feld, S.I.; Fan, J.; Yuan, M.; Wu, Y.; Woo, K.M.; Alexandridis, R.; Burnside, E.S. Utility of genetic testing in addition to mammography for determining risk of breast cancer depends on patient age. AMIA Summits Transl. Sci. Proc. 2018, 2018, 81. [Google Scholar]
  29. Mavaddat, N.; Rebbeck, T.R.; Lakhani, S.R.; Easton, D.F.; Antoniou, A.C. Incorporating tumour pathology information into breast cancer risk prediction algorithms. Breast Cancer Res. 2010, 12, R28. [Google Scholar] [CrossRef] [Green Version]
  30. Black, M.H.; Li, S.; LaDuca, H.; Chen, J.; Hoiness, R.; Gutierrez, S.; Lu, H.M.; Dolinsky, J.S.; Xu, J.; Vachon, C.; et al. Polygenic risk score for breast cancer in high-risk women. J. Clin. Oncol. 2018, 36, 1508. [Google Scholar] [CrossRef]
  31. Van den Broek, J.J.; Schechter, C.B.; van Ravesteyn, N.T.; Janssens, A.C.J.; Wolfson, M.C.; Trentham-Dietz, A.; Simard, J.; Easton, D.F.; Mandelblatt, J.S.; Kraft, P.; et al. Personalizing breast cancer screening based on polygenic risk and family history. JNCI J. Natl. Cancer Inst. 2021, 113, 434–442. [Google Scholar] [CrossRef]
  32. Thall, P.; Fox, P.; Wathen, J. Statistical controversies in clinical research: Scientific and ethical problems with adaptive randomization in comparative clinical trials. Ann. Oncol. 2015, 26, 1621–1628. [Google Scholar] [CrossRef] [PubMed]
  33. Park, Y. Challenges and Opportunities in Biomarker-Driven Trial Design: Adaptive Randomization (accepted). Ann. Transl. Med. 2022. [Google Scholar] [CrossRef]
  34. Park, Y.; Fullerton, H.J.; Elm, J.J. A pragmatic, adaptive clinical trial design for a rare disease: The FOcal Cerebral Arteriopathy Steroid (FOCAS) trial. Contemp. Clin. Trials 2019, 86, 105852. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  35. Zhu, H.; Yu, Q. A Bayesian sequential design using alpha spending function to control type I error. Stat. Methods Med. Res. 2017, 26, 2184–2196. [Google Scholar] [CrossRef] [PubMed]
  36. Murray, T.A.; Thall, P.F.; Yuan, Y.; McAvoy, S.; Gomez, D.R. Robust treatment comparison based on utilities of semi-competing risks in non-small-cell lung cancer. J. Am. Stat. Assoc. 2017, 112, 11–23. [Google Scholar] [CrossRef] [Green Version]
  37. Murray, T.A.; Thall, P.F.; Yuan, Y. Utility-based designs for randomized comparative trials with categorical outcomes. Stat. Med. 2016, 35, 4285–4305. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  38. Shi, H.; Yin, G. Control of type I error rates in Bayesian sequential designs. Bayesian Anal. 2019, 14, 399–425. [Google Scholar] [CrossRef]
  39. Wathen, J.K.; Thall, P.F. A simulation study of outcome adaptive randomization in multi-arm clinical trials. Clin. Trials 2017, 14, 432–440. [Google Scholar] [CrossRef]
  40. Villar, S.S.; Wason, J.; Bowden, J. Response-adaptive randomization for multi-arm clinical trials using the forward looking Gittins index rule. Biometrics 2015, 71, 969–978. [Google Scholar] [CrossRef] [PubMed]
  41. Ryan, E.G.; Lamb, S.E.; Williamson, E.; Gates, S. Bayesian adaptive designs for multi-arm trials: An orthopaedic case study. Trials 2020, 21, 83. [Google Scholar] [CrossRef] [Green Version]
  42. Viele, K.; Broglio, K.; McGlothlin, A.; Saville, B.R. Comparison of methods for control allocation in multiple arm studies using response adaptive randomization. Clin. Trials 2020, 17, 52–60. [Google Scholar] [CrossRef] [PubMed]
  43. Albert, J.H.; Chib, S. Bayesian analysis of binary and polychotomous response data. J. Am. Stat. Assoc. 1993, 88, 669–679. [Google Scholar] [CrossRef]
  44. Held, L.; Holmes, C.C. Bayesian auxiliary variable models for binary and multinomial regression. Bayesian Anal. 2006, 1, 145–168. [Google Scholar] [CrossRef]
Figure 1. Schema of the proposed design.
Figure 1. Schema of the proposed design.
Jpm 12 00448 g001
Figure 2. Boxplots of the estimated difference in the response probability between A and B (i.e., effect size) at final analysis. The red dots indicate the true effect sizes of the scenarios.
Figure 2. Boxplots of the estimated difference in the response probability between A and B (i.e., effect size) at final analysis. The red dots indicate the true effect sizes of the scenarios.
Jpm 12 00448 g002
Table 1. Simulation scenarios: True model parameters when x 1 and x 2 are independently generated from a Bernoulli distribution with response probability 0.5. Note that “sc” denotes scenarios.
Table 1. Simulation scenarios: True model parameters when x 1 and x 2 are independently generated from a Bernoulli distribution with response probability 0.5. Note that “sc” denotes scenarios.
sc. β 0 β 1 β 2 γ 0 γ 1 γ 2 Overall x = ( 1 , 1 ) x = ( 1 , 0 ) x = ( 0 , 1 ) x = ( 0 , 0 )
p A p B p A p B p A p B p A p B p A p B
10000000.5000.4980.4960.4930.5030.5020.5000.5000.4990.498
2−0.50.500000.4070.3980.5020.4880.4970.4990.3180.3020.3100.296
3−0.5100000.4990.4990.6950.6890.6910.6940.3130.3000.2990.309
4−100.50000.2310.2330.3020.3110.1570.1550.3120.3120.1550.157
5−1020000.5010.4960.8450.8360.1590.1570.8440.8380.1560.152
6−0.50.50.50000.4980.5000.6890.6960.4990.5030.5000.4930.3030.308
7−0.5110000.6580.6560.9320.9310.6950.6980.6880.6920.3140.305
8−0.5−0.50.50000.3210.3190.3120.3040.1590.1590.5000.4970.3030.312
9−0.5−110000.3440.3450.3130.3090.0690.0700.6990.6950.3000.312
10000−0.5000.3120.4990.3100.5050.3110.5010.3150.4970.3090.491
11−0.50.50−0.5000.2350.4040.3040.5030.3140.5010.1580.3080.1580.305
12−0.5−0.20−0.5000.1380.2760.1140.2420.1150.2400.1580.3070.1630.317
13−0.500−0.5−0.500.1120.3110.0670.3150.0690.3180.1600.3040.1500.303
14−0.500−0.50.500.2340.3090.3060.3070.3150.3080.1560.3090.1640.312
15−0.50.50−0.5−0.500.1580.4010.1560.4930.1580.4980.1550.3040.1590.310
16−100.5−0.5000.1130.2330.1580.3060.0690.1580.1560.3090.0670.157
17−102−0.5000.3780.5040.6850.8420.0680.1660.6910.8410.0650.159
18−102−0.500.50.4530.5000.8370.8390.0710.1610.8410.8370.0680.161
190.50.5−0.5−0.50.5−0.50.5000.6800.4990.6920.8410.8390.1600.4940.4970.694
200.50.5−0.5−0.650.50.50.6250.6800.8020.6920.8000.8400.4440.4910.4490.690
Table 2. Simulation results: estimated rejection probability of the designs when x 1 and x 2 are independently generated from a Bernoulli distribution with response probability 0.5. Note that “sc” denotes scenarios. The bold indicates the inflation of error rates.
Table 2. Simulation results: estimated rejection probability of the designs when x 1 and x 2 are independently generated from a Bernoulli distribution with response probability 0.5. Note that “sc” denotes scenarios. The bold indicates the inflation of error rates.
sc.( p A , p B )TradRARCARA1CARA2BaCARA
1(0.500, 0.498)0.0560.0460.0550.0610.040
2(0.407, 0.398)0.0540.0520.0730.0590.038
3(0.499, 0.499)0.0510.0460.1270.0480.040
4(0.231, 0.233)0.0440.0350.0750.0520.031
5(0.501, 0.496)0.0380.0530.3800.1730.059
6(0.498, 0.500)0.0540.0560.1050.0460.037
7(0.658, 0.656)0.0440.0540.2450.0930.062
8(0.321, 0.319)0.0400.0630.0940.0450.039
9(0.344, 0.345)0.0530.0500.1900.0640.051
10(0.312, 0.499)0.7880.7930.7530.7960.806
11(0.235, 0.404)0.7580.7410.7230.7460.739
12(0.138, 0.276)0.7350.7000.6630.6900.684
13(0.112, 0.311)0.9420.9350.9410.9270.922
14(0.234, 0.309)0.2270.2300.2430.2200.230
15(0.158, 0.401)0.9810.9790.9470.9750.971
16(0.113, 0.233)0.6150.6500.6250.6360.648
17(0.378, 0.504)0.4160.3920.5160.2340.671
18(0.453, 0.500)0.0930.0950.6160.2760.203
19(0.500, 0.680)0.7520.7480.8890.7750.815
20(0.625, 0.680)0.1490.1360.2670.1750.189
Table 3. Distribution of biomarker subgroups in each treatment arm under scenarios 5 and 7 for each design: mean (standard deviation) of the allocation probability of the treatment for each subgroup is reported. Note that “sc” denotes scenarios.
Table 3. Distribution of biomarker subgroups in each treatment arm under scenarios 5 and 7 for each design: mean (standard deviation) of the allocation probability of the treatment for each subgroup is reported. Note that “sc” denotes scenarios.
sc.DesignArmSubgroups Determined by x = ( x 1 , x 2 )
(1,1)(1,0)(0,1)(0,0)
5TradA0.248 (0.044)0.250 (0.042)0.250 (0.041)0.252 (0.041)
B0.250 (0.042)0.248 (0.043)0.252 (0.043)0.250 (0.042)
RARA0.252 (0.043)0.250 (0.044)0.250 (0.043)0.248 (0.042)
B0.250 (0.041)0.248 (0.041)0.252 (0.042)0.251 (0.042)
CARA1A0.247 (0.071)0.252 (0.075)0.249 (0.074)0.252 (0.074)
B0.251 (0.071)0.252 (0.074)0.247 (0.074)0.250 (0.074)
CARA2A0.249 (0.065)0.254 (0.056)0.244 (0.064)0.253 (0.058)
B0.241 (0.066)0.255 (0.058)0.247 (0.064)0.257 (0.059)
BaCARAA0.250 (0.073)0.248 (0.071)0.248 (0.073)0.254 (0.073)
B0.248 (0.082)0.253 (0.087)0.247 (0.088)0.252 (0.085)
7TradA0.249 (0.041)0.250 (0.042)0.251 (0.042)0.250 (0.042)
B0.250 (0.044)0.252 (0.044)0.250 (0.043)0.248 (0.042)
RARA0.250 (0.043)0.250 (0.041)0.250 (0.042)0.250 (0.042)
B0.250 (0.044)0.251 (0.043)0.250 (0.041)0.250 (0.042)
CARA1A0.248 (0.071)0.246 (0.073)0.254 (0.076)0.253 (0.081)
B0.243 (0.074)0.251 (0.079)0.247 (0.079)0.259 (0.086)
CARA2A0.245 (0.065)0.250 (0.061)0.249 (0.062)0.255 (0.059)
B0.246 (0.065)0.249 (0.061)0.249 (0.062)0.256 (0.063)
BaCARAA0.244 (0.067)0.251 (0.067)0.250 (0.066)0.255 (0.074)
B0.252 (0.079)0.249 (0.082)0.246 (0.081)0.253 (0.091)
Table 4. Simulation results: other operating characteristics of the designs when x 1 and x 2 are independently generated from a Bernoulli distribution with response probability 0.5. Note that “sc” denotes scenarios.
Table 4. Simulation results: other operating characteristics of the designs when x 1 and x 2 are independently generated from a Bernoulli distribution with response probability 0.5. Note that “sc” denotes scenarios.
sc.( p A , p B )TradRARCARA1CARA2BaCARA
Difference of the number of patients between A and B
10(0.312, 0.499)0.1363.03841.1148.27828.990
11(0.235, 0.404)−0.3982.26641.8127.80028.904
12(0.138, 0.276)0.8521.60641.9944.58432.122
13(0.112, 0.311)0.0643.08248.8085.25627.088
14(0.234, 0.309)0.0021.50824.3443.22226.000
15(0.158, 0.401)0.9242.77245.1487.50819.454
16(0.113, 0.233)0.3241.92844.3224.15233.606
17(0.378, 0.504)−0.0862.04243.30617.59832.072
18(0.453, 0.500)−1.1760.73218.9720.31623.070
19(0.500, 0.680)0.6522.79634.5009.79223.950
20(0.625, 0.680)−0.5840.48216.8084.69217.718
Number of failures
10(0.312, 0.499)73.0273.3169.3972.8161.07
11(0.235, 0.404)58.4758.5855.6658.4749.02
12(0.138, 0.276)38.6639.2636.2238.6433.29
13(0.112, 0.311)34.8634.4430.1234.3225.82
14(0.234, 0.309)55.9455.9253.5355.4848.40
15(0.158, 0.401)44.1043.2939.6942.9431.20
16(0.113, 0.233)33.4733.2130.7832.7327.54
17(0.378, 0.504)87.9587.9280.1089.0172.76
18(0.453, 0.500)99.6699.7187.8297.7484.11
19(0.500, 0.680)108.95108.8096.76106.5086.62
20(0.625, 0.680)135.25134.73130.58133.56116.36
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Park, Y. Personalized Risk-Based Screening Design for Comparative Two-Arm Group Sequential Clinical Trials. J. Pers. Med. 2022, 12, 448. https://doi.org/10.3390/jpm12030448

AMA Style

Park Y. Personalized Risk-Based Screening Design for Comparative Two-Arm Group Sequential Clinical Trials. Journal of Personalized Medicine. 2022; 12(3):448. https://doi.org/10.3390/jpm12030448

Chicago/Turabian Style

Park, Yeonhee. 2022. "Personalized Risk-Based Screening Design for Comparative Two-Arm Group Sequential Clinical Trials" Journal of Personalized Medicine 12, no. 3: 448. https://doi.org/10.3390/jpm12030448

APA Style

Park, Y. (2022). Personalized Risk-Based Screening Design for Comparative Two-Arm Group Sequential Clinical Trials. Journal of Personalized Medicine, 12(3), 448. https://doi.org/10.3390/jpm12030448

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop