1 Introduction
The significance of streamlining clinical trials has been emphasized in the Critical Path Opportunities Report and List [60]. The FDA [61] revised their guidance on seamless clinical trials and re-iterated the importance of moving towards the broadening acceptance of seamless trials. The FDA [61] outlined the need to evaluate new therapies in a time-sensitive, cost-effective and ethical manner without compromising the integrity and validity of the development process.
The seamless phase II/III clinical trial can reduce the lead time between different phases, reduce the number of trials for comparing multiple treatments, efficiently combine the data from both phases, monitor patients from the phase II trial longer for safety issues, and decrease the sample size while maintaining power. Typically, multiple experimental treatments are compared against a control in the first stage. The empirically best candidates are then selected to enter the second stage together with the control arm. The final analysis based on the patients from both stages is performed such that the overall type I error rate is controlled [44, 55, 56]. Until 2016, there have been more than 40 active, first-in-human cancer trials that are using the seamless strategy [39]. A motivating example is the Indacaterol to Help Achieve New COPD Treatment Excellence (INHANCE) trial [4], an adaptive seamless phase II/III clinical trial of inhaled indacaterol to treat chronic obstructive pulmonary disease (COPD). Other real seamless phase II/III clinical trials include [68] and [16].
In practice, hypothesis testing with type I error control is the primary focus of a seamless phase II/III trial, with estimation being an essential but secondary target [9, 10, 19, 32, 33, 38, 40, 46, 51, 52, 58]. This paper will focus on the control of type I error rate, as well as the investigation of the advantages of implementing DBCD in seamless clinical trials. The closure principle [37] has been proposed to handle the multiple testing problem; certain combination methods such as the inverse ${\chi ^{2}}$ method [5] and the weighted inverse normal method [34] have been proposed to combine data from the two stages; and different approaches such as the Simes test [47] and the Dunnett test [20] have been proposed to test the intersection of more than two hypotheses constructed for applying the closure principle. [14] and [45] made use of these methods to control the familywise type I error rate (FWER) for ASD. This paper will employ this framework since FDA and the pharmaceutical industry will readily accept it. [31, 49, 50] allowed more than one experimental treatment to continue beyond the first interim analysis and sequential analyses in the second stage. [63] proposed a multi-stage drop-the-losers design and discussed the required sample size. [36] proposed methods for any number of treatment arms, any number of stages and any number of patients per treatment per stage in such trials. [35] provided the theoretical foundation for a general family of two-stage adaptive designs. ASD with different study endpoints in the two stages has been investigated by [18, 48, 57]. We leave all these extensions for future research on our proposed procedure.
Next, we introduce RAR. Clinical trials are complex and usually have multiple objectives such as increasing the power of detecting treatment differences, and assigning more patients to better treatments. Two families of RAR have been proposed to achieve these objectives: DBCD [25, 62, 70, 71] and urn models [64, 65, 69, 72]. RAR can achieve greater efficiency and ethical advantages by skewing the allocation proportion based on previous treatment assignments and responses. A popular formal RAR framework contains three steps. First, the design objectives are mathematically formulated, and it is usually expressed in an optimization problem. Second, the optimal allocation proportions to achieve this objective as the solution of the optimization problem are derived. Third, a specific RAR design is implemented to target the optimal allocation proportion. [25, 71, 72] studied the asymptotic properties and sequential monitoring of RAR. [24] showed that RAR could increase efficiency in certain clinical trials. [59] explored the derivation of optimal allocation proportion. Other discussions of the advantages of RAR can be found in [2, 8, 21, 26, 27, 30, 41]. Clinical trials using RAR designs include [1, 42, 53]. This paper focuses on using DBCD as randomization in seamless clinical trials.
Therefore, it is desirable to study how to benefit from the advantages of ASD and RAR in one clinical trial. However, both ASD and RAR pose a challenge in controlling the type I error rate, which is critical in confirmatory clinical trials. ASD tends to increase the type I error rate due to multiple testing and treatment selection at the interim look. RAR introduced extra difficulties with correlated responses and treatment assignments. In this paper, we overcame these difficulties and studied its asymptotic and finite-sample properties.
2 Seamless Clinical Trials with DBCD
2.1 Adaptive Seamless Design with DBCD
We first introduce the notation for DBCD with multiple treatments. Suppose $(K+1)$ treatments are under study in a clinical trial with sample size n. Let ${\boldsymbol{T}_{i}}=({T_{i0}},{T_{i1}},\dots ,{T_{iK}})$ denote the ith patient’s treatment assignment, where treatment 0 indicates the control arm, ${T_{ik}}=1$, $k=0,1,\dots ,K$ if the ith patient is in treatment k, and ${T_{ik}}=0$ otherwise. Let $\boldsymbol{N}(m)=\left({N_{0}}\left(m\right),{N_{1}}\left(m\right),\dots ,{N_{K}}\left(m\right)\right)$, where ${N_{k}}(m)={\textstyle\sum _{i=1}^{m}}{T_{ik}}$, $k=0,1,\dots ,K$ is the number of patients assigned to treatment k after m patients have entered the trial. Let ${\boldsymbol{X}_{i}}=({\boldsymbol{X}_{i0}},{\boldsymbol{X}_{i1}},\dots ,{\boldsymbol{X}_{iK}})$, $i=1,\dots ,n$ be a random matrix of response variables, where ${\boldsymbol{X}_{ik}},k=0,1,\dots ,K$, are d-dimensional random vectors. Here, if the ith patient is assigned to treatment k, only ${\boldsymbol{X}_{ik}}$ can be observed. In other words, ${\boldsymbol{X}_{ik}}$ is the ith patient’s response in the presence of treatment k and only observed if ${T_{ik}}=1$. Therefore, the variable ${T_{ik}}$ does not influence the expectation of ${\boldsymbol{X}_{ik}}$; it only determines if it is observed. Without loss of generality, we assume ${\boldsymbol{\theta }_{k}}=E({\boldsymbol{X}_{ik}})=({\theta _{k1}},\dots ,{\theta _{kd}})$, $k=0,1,\dots ,K$. Then the parameter estimator after responses of m patients have been observed is
Write
and
(2.1)
\[ {\hat{\boldsymbol{\theta }}_{k}}(m)=\frac{{\textstyle\textstyle\sum _{i=1}^{m}}{T_{ik}}{\boldsymbol{X}_{ik}}}{{N_{k}}(m)}.\]RAR can achieve various objectives by targeting different allocation proportions that will be functions of unknown parameters [59]. Let ${\boldsymbol{\rho }_{l}}(\boldsymbol{\theta })=({\rho _{l0}}(\boldsymbol{\theta }),{\rho _{l1}}(\boldsymbol{\theta }),\dots ,{\rho _{lK}}(\boldsymbol{\theta }))$, $l=1,2$, is the target allocation proportions for stage l, where ${\boldsymbol{\rho }_{l}}(\boldsymbol{\theta }):{\mathrm{\Re }^{d\times (K+1)}}\to {(0,1)^{(K+1)}}$ is the vector-valued functions satisfying ${\boldsymbol{\rho }_{l}}(\boldsymbol{\theta }){\mathbf{1}^{\prime }}=1$. Specific examples can be seen in Section 3.
Next, we introduce the procedure to conduct a seamless phase II/III clinical trials with a family of DBCD:
(i) In the first stage, we first assign ${m_{0}}$ patients to each of the $K+1$ treatments by fixed design to obtain initial parameter estimates. When the mth ($m\gt (K+1){m_{0}}$) patient enters the first stage of the trial, calculate $\hat{\boldsymbol{\theta }}(m-1)$ and ${\hat{\boldsymbol{\rho }}_{1}}={\boldsymbol{\rho }_{1}}(\hat{\boldsymbol{\theta }}(m-1))$ based on all the previous responses and treatment assignments.
(ii) Assign the mth patient to treatment k with probability
where ${s_{k}}={N_{k}}(m-1)/(m-1)$ and ${r_{k}}={\rho _{1k}}(\hat{\boldsymbol{\theta }}(m-1))$.
\[ {g_{1k}}\left(\boldsymbol{N}(m-1)/(m-1),{\boldsymbol{\rho }_{1}}(\hat{\boldsymbol{\theta }}(m-1))\right),\]
where ${g_{1k}}(\boldsymbol{s},\boldsymbol{r})={g_{1k}}(({s_{0}},{s_{1}},\dots ,{s_{K}}),({r_{0}},{r_{1}},\dots ,{r_{K}})):{(0,1)^{(K+1)}}\times {(0,1)^{(K+1)}}\to (0,1)$ satisfies ${\textstyle\sum _{k=0}^{K}}{g_{1k}}(\boldsymbol{s},\boldsymbol{r})=1$ [25]. We write ${\boldsymbol{g}_{1}}=({g_{10}},{g_{11}},\dots ,{g_{1K}})$. [25] proposed the following allocation probability function to the treatment k for the mth patient
(2.2)
\[ {g_{1k}}(\boldsymbol{s},\boldsymbol{r})=\frac{{r_{k}}{({r_{k}}/{s_{k}})^{2}}}{{\textstyle\textstyle\sum _{j=0}^{K}}\{{r_{j}}{({r_{j}}/{s_{j}})^{2}}\}},\](iii) At the end of the first stage, choose one (say treatment M) based on certain criteria to enter the second stage, along with the control arm. For example, we can choose the experimental treatment arm with the largest treatment effect to enter the second stage; we can also incorporate safety data into the criteria for choosing a treatment arm for the second stage.
(iv) Because we have only two treatment arms under study in the second stage, let ${\boldsymbol{\rho }_{2}}(\boldsymbol{\theta })=({\rho _{20}}(\boldsymbol{\theta }),{\rho _{2M}}(\boldsymbol{\theta }))$ be the target allocation proportions for the second stage, where ${\boldsymbol{\rho }_{2}}(\boldsymbol{\theta }):{\mathrm{\Re }^{d\times 2}}\to {(0,1)^{2}}$ is the vector-valued functions satisfying ${\boldsymbol{\rho }_{2}}(\boldsymbol{\theta }){\mathbf{1}^{\prime }}=1$. At the second stage, assign the mth patient to treatment $k,k=0,M$ with probability
\[ {g_{2k}}\left(\hspace{-0.1667em}\left({N_{0}}(m\hspace{-0.1667em}-\hspace{-0.1667em}1)/(m\hspace{-0.1667em}-\hspace{-0.1667em}1),{N_{M}}(m\hspace{-0.1667em}-\hspace{-0.1667em}1)/(m\hspace{-0.1667em}-\hspace{-0.1667em}1)\hspace{-0.1667em}\right),{\boldsymbol{\rho }_{2}}(\hat{\boldsymbol{\theta }}(m\hspace{-0.1667em}-\hspace{-0.1667em}1))\hspace{-0.1667em}\right)\hspace{-0.1667em},\]
where ${g_{2k}}(\boldsymbol{s},\boldsymbol{r}):{(0,1)^{2}}\times {(0,1)^{2}}\to (0,1)$ satisfies ${g_{20}}(\boldsymbol{s},\boldsymbol{r})+{g_{2M}}(\boldsymbol{s},\boldsymbol{r})=1$. We write ${\boldsymbol{g}_{2}}=({g_{20}},{g_{2M}})$.The above DBCD considers both the estimated targeted allocation proportions and the current allocation proportions in order to achieve different ethical and efficient objectives. A specific family of allocation probability functions will be given in Section 3. Other discussions and properties of DBCD can be seen in [71, 25].
2.2 Data Analysis Procedure
At the end of the clinical trial, one considers a general hypothesis test:
\[ {H_{0,M}}:h({\boldsymbol{\theta }_{M}})=h({\boldsymbol{\theta }_{0}})\hspace{2.5pt}\text{versus}\hspace{2.5pt}{H_{1,M}}:h({\boldsymbol{\theta }_{M}})\gt h({\boldsymbol{\theta }_{0}})\]
where $h({\boldsymbol{\theta }_{j}})$ is a ${\mathrm{\Re }^{d}}\to \mathrm{\Re }$ continuous and twice differentiable function in a small neighborhood of ${\boldsymbol{\theta }_{j}},\hspace{2.5pt}j=0,M$.We test the above hypothesis with the combined data from the two stages and follow the closure principle [37] to control the familywise type I error rate. The closure principle rejects ${H_{0,M}}$ at level α if each intersection hypothesis ${H_{0,I}}$ with $M\in I$, $I\subseteq \{1,\dots ,K\}$, is rejected at level α, where ${H_{0,I}}={\cap _{k\in I}}{H_{0,k}}$ with ${H_{0,k}}:h({\boldsymbol{\theta }_{k}})=h({\boldsymbol{\theta }_{0}})$. Each ${H_{0,I}}$ can be tested with the following inverse ${\chi ^{2}}$ method. Let ${P_{1,I}}$ and ${P_{2,I}}$ denote the p-values for ${H_{0,I}}$ based on the data from the first stage and the second stage, respectively. Then we reject ${H_{0,I}}$ if $-\log ({P_{1,I}}{P_{2,I}})\gt {\chi _{4}^{2}}(1-\alpha )/2$, where ${\chi _{4}^{2}}(1-\alpha )$ is the $(1-\alpha )$th quantile of the ${\chi ^{2}}$ distribution with 4 degrees of freedom. To calculate the adjusted p-values for each stage, ${P_{1,I}}$ and ${P_{2,I}}$, we use the Simes test [47] with the following test statistics for the elementary hypotheses ${H_{0,k}}$ in the intersection hypothesis ${H_{0,I}}$,
Here $\hat{Var}(h({\hat{\boldsymbol{\theta }}_{k}}(n)))$ and $\hat{Var}(h({\hat{\boldsymbol{\theta }}_{0}}(n)))$ are some consistent estimators of the variances of $h({\hat{\boldsymbol{\theta }}_{k}}(n))$ and $h({\hat{\boldsymbol{\theta }}_{0}}(n))$ respectively. We assume that for some functions ${\nu _{k}}$ and ${v_{0}}$
(2.3)
\[ {Z_{k}}\hspace{-0.1667em}\left(\frac{\boldsymbol{N}(n)}{n},\hat{\boldsymbol{\theta }}(n)\hspace{-0.1667em}\right)\hspace{-0.1667em}=\hspace{-0.1667em}\frac{h\hspace{-0.1667em}\left({\hat{\boldsymbol{\theta }}_{k}}(n)\right)-h\left({\hat{\boldsymbol{\theta }}_{0}}(n)\right)}{\sqrt{\hat{Var}\left(h\left({\hat{\boldsymbol{\theta }}_{k}}(n)\right)\right)\hspace{-0.1667em}+\hspace{-0.1667em}\hat{Var}\left(h\left({\hat{\boldsymbol{\theta }}_{0}}(n)\right)\hspace{-0.1667em}\right)}}.\]
\[\begin{aligned}{}& n\hat{Var}\left(h\left({\hat{\boldsymbol{\theta }}_{j}}\left(n\right)\right)\right)\\ {} & \hspace{1em}={\nu _{j}}\left(\frac{\boldsymbol{N}\left(n\right)}{n},\hat{\boldsymbol{\theta }}\left(n\right)\right)(1+o(1))\hspace{2.5pt}\hspace{2.5pt}\hspace{2.5pt}\hspace{2.5pt}\text{a.s.}\hspace{2.5pt}j=0,k.\end{aligned}\]
Both ${\nu _{j}}(\boldsymbol{y},\boldsymbol{z})$ and ${Z_{k}}(\boldsymbol{y},\boldsymbol{z})$ are ${\mathrm{\Re }^{(K+1)(1+d)}}\to \mathrm{\Re }$ function, where $\boldsymbol{y}$ is a $(K+1)$-dimensional vector and $\boldsymbol{z}$ is a $(K+1)d$-dimensional vector. Examples of using this formulation are given in Section 3.2.3 Asymptotic Results
Before we give the main theorem, we need the following conditions.
-
(A1) For some $\varepsilon \gt 0$, $E\| {\boldsymbol{X}_{1}}{\| ^{2+\varepsilon }}\lt \infty $;
-
(A2) ${g_{1k}}(\boldsymbol{s},\boldsymbol{r}),k=0,1,\dots ,K$, is jointly continuous and twice differentiable at $({\boldsymbol{\rho }_{1}},{\boldsymbol{\rho }_{1}})$, and ${g_{2k}}(\boldsymbol{s},\boldsymbol{r}),k=0,M$, is jointly continuous and twice differentiable at $({\boldsymbol{\rho }_{2}},{\boldsymbol{\rho }_{2}})$;
-
(A3) ${g_{1k}}(\boldsymbol{r},\boldsymbol{r})={r_{k}}$, $k=0,1,\dots ,K$, for all $\boldsymbol{r}\in {(0,1)^{(K+1)}}$ and ${\textstyle\sum _{k=0}^{K}}{r_{k}}=1$, ${g_{1k}}(\boldsymbol{s},\boldsymbol{r})$ is strictly decreasing in $\boldsymbol{s}\in {(0,1)^{(K+1)}}$ and strictly increasing in $\boldsymbol{r}\in {(0,1)^{(K+1)}}$, ${g_{2k}}(\boldsymbol{r},\boldsymbol{r})={r_{k}}$, $k=0,M$, for all $\boldsymbol{r}\in {(0,1)^{2}}$ and ${r_{0}}+{r_{M}}=1$, and ${g_{2k}}(\boldsymbol{s},\boldsymbol{r})$ is strictly decreasing in $\boldsymbol{s}\in {(0,1)^{2}}$ and strictly increasing in $\boldsymbol{r}\in {(0,1)^{2}}$;
-
(A4) ${\boldsymbol{\rho }_{1k}}(\boldsymbol{\theta }),k=0,1,\dots ,K$ and ${\boldsymbol{\rho }_{2k}}(\boldsymbol{\theta }),k=0,M$ are continuous functions and twice continuously differentiable in a small neighborhood of $\boldsymbol{\theta }$;
-
(A5) ${\nu _{j}}(\boldsymbol{y},\boldsymbol{z})$ is jointly continuous and twice differentiable in a small neighborhood of $(\boldsymbol{\rho },\boldsymbol{\theta })$;
-
(A6) ${Z_{k}}(\boldsymbol{y},\boldsymbol{z})$ is a continuous function and it is twice continuously differentiable in a small neighborhood of vector $(\boldsymbol{\rho },\boldsymbol{\theta })$.
Theorem 2.1.
Under Conditions (A1)–(A6), a valid type I error rate can be asymptotically obtained for the Simes test with the test statistics ${Z_{k}},k=1,\dots ,K$, for the proposed procedure. That is, for a given significance level α, when ${H_{0,M}}$ holds, the probability that we reject ${H_{0,M}}$ has a limit that is not larger than α.
Theorem 2.1 offers the theoretical justification for controlling the type I error rate for our procedure. All these conditions are easily satisfied. The well-known family of DBCD [25] meets all these requirements. Condition (A1) ensures consistency and asymptotic normality. All the examples in Chapter 5 [23] meet Conditions (A4)–(A6). In particular, Condition (A3) has practical meaning in clinical trials: if the current actual allocation proportion is equal to the target allocation proportion, the allocation probability for the next patient will equal to the target allocation proportion (${g_{jk}}(\boldsymbol{r},\boldsymbol{r})={r_{k}}$). On top of that, because the allocation probability function is strictly decreasing in the actual allocation proportion and strictly increasing in the estimated target allocation proportion, the proposed RAR design will asymptotically drive the actual allocation proportion to approach the theoretically targeted one for each stage (${\boldsymbol{\rho }_{1}}$ for stage 1 and ${\boldsymbol{\rho }_{2}}$ for stage 2), which is proved in [25]. The actual final allocation proportion for the two-stage seamless trial when the sample size is finite will be studied in the next section.
3 Numerical Studies
In this section, we study the finite-sample properties of our proposed procedure and offer three specific targeted allocation proportions.
Suppose 300 patients sequentially enter the trial with two experimental treatments and one control in the first stage. Let the responses ${X_{ik}},i=1,\dots ,300$, $k=0,1,2$, follow the Bernoulli distribution with success rate ${p_{k}}$, respectively. These patients will be sequentially randomly allocated to the treatment k with the following allocation probability function [25]
\[ {g_{1k}}(\boldsymbol{s},\boldsymbol{r})=\frac{{r_{k}}{({r_{k}}/{s_{k}})^{2}}}{{\textstyle\textstyle\sum _{j=0}^{K}}\{{r_{j}}{({r_{j}}/{s_{j}})^{2}}\}},\]
where ${s_{k}}={N_{k}}(m-1)/(m-1)$ and ${r_{k}}={\rho _{1k}}(\hat{\boldsymbol{\theta }}(m-1))$ when we are calculating the allocation probability for the mth patient. We will discuss three specific targeted allocation proportions later. The experimental treatment arm with a larger treatment effect, say treatment M, is chosen to continue to the second stage. In the second stage, 500 patients are sequentially randomly allocated to the control arm and treatment M with the following allocation probability function
\[ {g_{2k}}(\boldsymbol{s},\boldsymbol{r})=\frac{{r_{k}}{({r_{k}}/{s_{k}})^{2}}}{\{{r_{0}}{({r_{0}}/{s_{0}})^{2}}+{r_{M}}{({r_{M}}/{s_{M}})^{2}}\}},k=0,M.\]
At the end of the trial, we test
In this case, $d=1$, ${\theta _{k}}={p_{k}}$, and
The significance level is 0.025 for all the tests. All the results are based on 10, 000 replications.In the first scenario, let
\[ {\boldsymbol{\rho }_{1}}(\boldsymbol{\theta })=\left(\frac{{p_{0}}}{{p_{0}}+{p_{1}}+{p_{2}}},\frac{{p_{1}}}{{p_{0}}+{p_{1}}+{p_{2}}},\frac{{p_{2}}}{{p_{0}}+{p_{1}}+{p_{2}}}\right)\]
and
\[ {\boldsymbol{\rho }_{2}}(\boldsymbol{\theta })=\left(\frac{{q_{M}}}{{q_{0}}+{q_{M}}},\frac{{q_{0}}}{{q_{0}}+{q_{M}}}\right)\]
that is the urn allocation [64]. Urn allocation is used to assign more patients to the better treatment.In the second scenario, let
\[\begin{aligned}{}& {\boldsymbol{\rho }_{1}}(\boldsymbol{\theta })\\ {} & =\left(\frac{\sqrt{{p_{0}}}}{\sqrt{{p_{0}}}+\sqrt{{p_{1}}}+\sqrt{{p_{2}}}},\frac{\sqrt{{p_{1}}}}{\sqrt{{p_{0}}}+\sqrt{{p_{1}}}+\sqrt{{p_{2}}}},\frac{\sqrt{{p_{2}}}}{\sqrt{{p_{0}}}+\sqrt{{p_{1}}}+\sqrt{{p_{2}}}}\right)\end{aligned}\]
and
\[ {\boldsymbol{\rho }_{2}}(\boldsymbol{\theta })=\left(\frac{\sqrt{{p_{0}}}}{\sqrt{{p_{0}}}+\sqrt{{p_{M}}}},\frac{\sqrt{{p_{M}}}}{\sqrt{{p_{0}}}+\sqrt{{p_{M}}}}\right)\]
that is the optimal allocation [41]. The optimal allocation is used to minimize the total number of failures while fixing the power.In the third scenario, let
and
This target allocation proportion is intuitively assigning more patients to the better treatment.
Table 1
Performance of DBCD targeting the urn allocation under ${H_{0}}$ when three treatments are under study.
(${p_{1}}$, ${p_{2}}$, ${p_{0}}$) | Design | α | ${\hat{p}_{0}}$ | ${\rho _{0}}$ | Failure |
$(0.5,0.5,0.5)$ | DBCD | 0.024 | 0.500(0.027) | 0.438(0.020) | 400(14) |
$(0.5,0.5,0.5)$ | CR | 0.024 | 0.500(0.027) | 0.438(0.017) | 400(14) |
$(0.6,0.6,0.6)$ | DBCD | 0.022 | 0.599(0.026) | 0.437(0.022) | 320(14) |
$(0.6,0.6,0.6)$ | CR | 0.022 | 0.600(0.026) | 0.438(0.017) | 320(14) |
$(0.7,0.7,0.7)$ | DBCD | 0.024 | 0.700(0.025) | 0.437(0.025) | 240(13) |
$(0.7,0.7,0.7)$ | CR | 0.023 | 0.700(0.025) | 0.438(0.017) | 240(13) |
$(0.8,0.8,0.8)$ | DBCD | 0.023 | 0.799(0.021) | 0.437(0.032) | 160(11) |
$(0.8,0.8,0.8)$ | CR | 0.025 | 0.800(0.021) | 0.438(0.017) | 160(11) |
Table 2
Performance of DBCD targeting the optimal allocation under ${H_{0}}$ when three treatments are under study.
(${p_{1}}$, ${p_{2}}$, ${p_{0}}$) | Design | α | ${\hat{p}_{0}}$ | ${\rho _{0}}$ | Failure |
$(0.5,0.5,0.5)$ | DBCD | 0.023 | 0.499(0.027) | 0.438(0.012) | 400(14) |
$(0.5,0.5,0.5)$ | CR | 0.024 | 0.500(0.027) | 0.438(0.017) | 400(14) |
$(0.6,0.6,0.6)$ | DBCD | 0.024 | 0.600(0.026) | 0.438(0.011) | 320(14) |
$(0.6,0.6,0.6)$ | CR | 0.022 | 0.600(0.026) | 0.438(0.017) | 320(14) |
$(0.7,0.7,0.7)$ | DBCD | 0.026 | 0.700(0.025) | 0.438(0.010) | 240(13) |
$(0.7,0.7,0.7)$ | CR | 0.023 | 0.700(0.025) | 0.438(0.017) | 240(13) |
$(0.8,0.8,0.8)$ | DBCD | 0.024 | 0.800(0.022) | 0.438(0.010) | 160(11) |
$(0.8,0.8,0.8)$ | CR | 0.025 | 0.800(0.021) | 0.438(0.017) | 160(11) |
Table 3
Performance of DBCD targeting the intuitively ethical allocation proportion under ${H_{0}}$ when three treatments are under study.
(${p_{1}}$, ${p_{2}}$, ${p_{0}}$) | Design | α | ${\hat{p}_{0}}$ | ${\rho _{0}}$ | Failure |
$(0.5,0.5,0.5)$ | DBCD | 0.027 | 0.500(0.027) | 0.438(0.020) | 400(14) |
$(0.5,0.5,0.5)$ | CR | 0.024 | 0.500(0.027) | 0.438(0.017) | 400(14) |
$(0.6,0.6,0.6)$ | DBCD | 0.024 | 0.599(0.026) | 0.437(0.017) | 320(14) |
$(0.6,0.6,0.6)$ | CR | 0.022 | 0.600(0.026) | 0.438(0.017) | 320(14) |
$(0.7,0.7,0.7)$ | DBCD | 0.026 | 0.699(0.025) | 0.437(0.014) | 240(13) |
$(0.7,0.7,0.7)$ | CR | 0.023 | 0.700(0.025) | 0.438(0.017) | 240(13) |
$(0.8,0.8,0.8)$ | DBCD | 0.022 | 0.800(0.022) | 0.438(0.012) | 160(11) |
$(0.8,0.8,0.8)$ | CR | 0.025 | 0.800(0.021) | 0.438(0.017) | 160(11) |
In Tables 1–3, we studied and compared the performance of our methods under each of the above scenarios and complete randomization (CR) under ${H_{0,M}}$. In these tables, we found that, under ${H_{0,M}}$, our method can control the type I error rate (α) well. We reported ${\hat{p}_{0}}$ as a representative of the parameter estimators. We also reported the actual allocation proportion to the control group (${\rho _{0}}$) and the total number of failures (Failure). The standard deviations are in the parentheses. In all the tables, our methods and CR return almost the same values in terms of the allocation proportion and the total number of failures under ${H_{0,M}}$, since our designs are also targeting the equal allocation under ${H_{0,M}}$. Our methods can also estimate the parameter accurately.
Table 4
Performance of DBCD targeting the urn allocation under ${H_{1}}$ when three treatments are under study.
(${p_{1}}$, ${p_{2}}$, ${p_{0}}$) | Design | Power | ${\hat{p}_{0}}$ | ${\rho _{0}}$ | Failure |
$(0.4,0.45,0.3)$ | DBCD | 0.939 | 0.300(0.027) | 0.379(0.020) | 494(15) |
$(0.4,0.45,0.3)$ | CR | 0.946 | 0.300(0.025) | 0.438(0.017) | 501(15) |
$(0.5,0.55,0.4)$ | DBCD | 0.926 | 0.400(0.028) | 0.378(0.021) | 414(16) |
$(0.5,0.55,0.4)$ | CR | 0.930 | 0.400(0.026) | 0.438(0.017) | 420(15) |
$(0.6,0.65,0.5)$ | DBCD | 0.925 | 0.500(0.029) | 0.371(0.023) | 333(16) |
$(0.6,0.65,0.5)$ | CR | 0.929 | 0.500(0.027) | 0.438(0.017) | 340(15) |
$(0.7,0.75,0.6)$ | DBCD | 0.952 | 0.599(0.029) | 0.358(0.026) | 251(15) |
$(0.7,0.75,0.6)$ | CR | 0.955 | 0.600(0.026) | 0.438(0.017) | 260(14) |
Table 5
Performance of DBCD targeting the optimal allocation under ${H_{1}}$ when three treatments are under study.
(${p_{1}}$, ${p_{2}}$, ${p_{0}}$) | Design | Power | ${\hat{p}_{0}}$ | ${\rho _{0}}$ | Failure |
$(0.4,0.45,0.3)$ | DBCD | 0.938 | 0.300(0.026) | 0.395(0.015) | 495(15) |
$(0.4,0.45,0.3)$ | CR | 0.946 | 0.300(0.025) | 0.438(0.017) | 501(15) |
$(0.5,0.55,0.4)$ | DBCD | 0.928 | 0.400(0.027) | 0.404(0.013) | 417(15) |
$(0.5,0.55,0.4)$ | CR | 0.930 | 0.400(0.026) | 0.438(0.017) | 420(15) |
$(0.6,0.65,0.5)$ | DBCD | 0.928 | 0.500(0.028) | 0.410(0.017) | 337(15) |
$(0.6,0.65,0.5)$ | CR | 0.929 | 0.500(0.027) | 0.438(0.017) | 340(15) |
$(0.7,0.75,0.6)$ | DBCD | 0.951 | 0.600(0.027) | 0.414(0.010) | 258(14) |
$(0.7,0.75,0.6)$ | CR | 0.955 | 0.600(0.026) | 0.438(0.017) | 260(14) |
Table 6
Performance of DBCD targeting the intuitively ethical allocation proportion under ${H_{1}}$ when three treatments are under study.
(${p_{1}}$, ${p_{2}}$, ${p_{0}}$) | Design | Power | ${\hat{p}_{0}}$ | ${\rho _{0}}$ | Failure |
$(0.4,0.45,0.3)$ | DBCD | 0.939 | 0.299(0.028) | 0.354(0.027) | 491(15) |
$(0.4,0.45,0.3)$ | CR | 0.946 | 0.300(0.025) | 0.438(0.017) | 501(15) |
$(0.5,0.55,0.4)$ | DBCD | 0.926 | 0.399(0.029) | 0.372(0.023) | 413(16) |
$(0.5,0.55,0.4)$ | CR | 0.930 | 0.400(0.026) | 0.438(0.017) | 420(15) |
$(0.6,0.65,0.5)$ | DBCD | 0.927 | 0.500(0.029) | 0.383(0.019) | 334(15) |
$(0.6,0.65,0.5)$ | CR | 0.929 | 0.500(0.027) | 0.438(0.017) | 340(15) |
$(0.7,0.75,0.6)$ | DBCD | 0.954 | 0.599(0.028) | 0.391(0.016) | 255(14) |
$(0.7,0.75,0.6)$ | CR | 0.955 | 0.600(0.026) | 0.438(0.017) | 260(14) |
In Tables 4–6, we studied and compared the performance of our methods under each of the above scenarios and CR under ${H_{1,M}}$. In Table 4, we can see that our method can save up to around 10 patients while keeping the power at the same level as CR under ${H_{1,M}}$ for the first scenario. In Table 5, we can see that our method can assign more patients to the better treatment while keeping the power at the same level as CR under ${H_{1,M}}$. In Table 6, we found that the DBCD targeting this allocation proportion can also save up to 10 patients under the reported settings without sacrificing the power.
Table 7
Performance of DBCD targeting the urn allocation under ${H_{0}}$ when four treatments are under study.
(${p_{1}}$, ${p_{2}}$, ${p_{3}}$, ${p_{0}}$) | Design | α | ${\hat{p}_{0}}$ | ${\rho _{0}}$ | Failure |
$(0.5,0.5,0.5,0.5)$ | DBCD | 0.023 | 0.499(0.027) | 0.389(0.018) | 450(15) |
$(0.5,0.5,0.5,0.5)$ | CR | 0.025 | 0.500(0.027) | 0.389(0.016) | 450(15) |
$(0.6,0.6,0.6,0.6)$ | DBCD | 0.025 | 0.599(0.026) | 0.389(0.020) | 360(15) |
$(0.6,0.6,0.6,0.6)$ | CR | 0.023 | 0.600(0.026) | 0.389(0.016) | 360(15) |
$(0.7,0.7,0.7,0.7)$ | DBCD | 0.022 | 0.699(0.025) | 0.389(0.022) | 270(14) |
$(0.7,0.7,0.7,0.7)$ | CR | 0.025 | 0.700(0.025) | 0.389(0.016) | 270(14) |
$(0.8,0.8,0.8,0.8)$ | DBCD | 0.022 | 0.799(0.022) | 0.389(0.028) | 180(12) |
$(0.8,0.8,0.8,0.8)$ | CR | 0.025 | 0.800(0.022) | 0.389(0.016) | 180(12) |
Table 8
Performance of DBCD targeting the optimal allocation under ${H_{0}}$ when four treatments are under study.
(${p_{1}}$, ${p_{2}}$, ${p_{3}}$, ${p_{0}}$) | Design | α | ${\hat{p}_{0}}$ | ${\rho _{0}}$ | Failure |
$(0.5,0.5,0.5,0.5)$ | DBCD | 0.023 | 0.500(0.027) | 0.389(0.011) | 450(15) |
$(0.5,0.5,0.5,0.5)$ | CR | 0.025 | 0.500(0.027) | 0.389(0.016) | 450(15) |
$(0.6,0.6,0.6,0.6)$ | DBCD | 0.023 | 0.600(0.026) | 0.389(0.010) | 360(15) |
$(0.6,0.6,0.6,0.6)$ | CR | 0.023 | 0.600(0.026) | 0.389(0.016) | 360(15) |
$(0.7,0.7,0.7,0.7)$ | DBCD | 0.026 | 0.700(0.024) | 0.389(0.009) | 270(14) |
$(0.7,0.7,0.7,0.7)$ | CR | 0.025 | 0.700(0.025) | 0.389(0.016) | 270(14) |
$(0.8,0.8,0.8,0.8)$ | DBCD | 0.023 | 0.800(0.021) | 0.389(0.008) | 180(12) |
$(0.8,0.8,0.8,0.8)$ | CR | 0.025 | 0.800(0.022) | 0.389(0.016) | 180(12) |
Table 9
Performance of DBCD targeting the intuitively ethical allocation proportion under ${H_{0}}$ when four treatments are under study.
(${p_{1}}$, ${p_{2}}$, ${p_{3}}$, ${p_{0}}$) | Design | α | ${\hat{p}_{0}}$ | ${\rho _{0}}$ | Failure |
$(0.5,0.5,0.5,0.5)$ | DBCD | 0.022 | 0.499(0.027) | 0.389(0.018) | 450(15) |
$(0.5,0.5,0.5,0.5)$ | CR | 0.025 | 0.500(0.027) | 0.389(0.016) | 450(15) |
$(0.6,0.6,0.6,0.6)$ | DBCD | 0.025 | 0.599(0.026) | 0.389(0.016) | 360(15) |
$(0.6,0.6,0.6,0.6)$ | CR | 0.023 | 0.600(0.026) | 0.389(0.016) | 360(15) |
$(0.7,0.7,0.7,0.7)$ | DBCD | 0.024 | 0.700(0.025) | 0.389(0.013) | 270(14) |
$(0.7,0.7,0.7,0.7)$ | CR | 0.025 | 0.700(0.025) | 0.389(0.016) | 270(14) |
$(0.8,0.8,0.8,0.8)$ | DBCD | 0.020 | 0.800(0.021) | 0.389(0.011) | 180(12) |
$(0.8,0.8,0.8,0.8)$ | CR | 0.025 | 0.800(0.022) | 0.389(0.016) | 180(12) |
Table 10
Performance of DBCD targeting the urn allocation under ${H_{1}}$ when four treatments are under study.
(${p_{1}}$, ${p_{2}}$, ${p_{3}}$, ${p_{0}}$) | Design | Power | ${\hat{p}_{0}}$ | ${\rho _{0}}$ | Failure |
$(0.35,0.4,0.45,0.3)$ | DBCD | 0.910 | 0.299(0.026) | 0.339(0.019) | 559(17) |
$(0.35,0.4,0.45,0.3)$ | CR | 0.910 | 0.300(0.025) | 0.389(0.016) | 566(16) |
$(0.45,0.5,0.55,0.4)$ | DBCD | 0.892 | 0.400(0.028) | 0.338(0.019) | 469(17) |
$(0.45,0.5,0.55,0.4)$ | CR | 0.896 | 0.400(0.026) | 0.389(0.016) | 476(17) |
$(0.55,0.6,0.65,0.5)$ | DBCD | 0.894 | 0.499(0.029) | 0.332(0.021) | 379(17) |
$(0.55,0.6,0.65,0.5)$ | CR | 0.896 | 0.500(0.027) | 0.389(0.016) | 386(16) |
$(0.65,0.7,0.75,0.6)$ | DBCD | 0.923 | 0.598(0.029) | 0.320(0.024) | 287(17) |
$(0.65,0.7,0.75,0.6)$ | CR | 0.922 | 0.600(0.026) | 0.389(0.016) | 296(16) |
Table 11
Performance of DBCD targeting the optimal allocation under ${H_{1}}$ when four treatments are under study.
(${p_{1}}$, ${p_{2}}$, ${p_{3}}$, ${p_{0}}$) | Design | Power | ${\hat{p}_{0}}$ | ${\rho _{0}}$ | Failure |
$(0.35,0.4,0.45,0.3)$ | DBCD | 0.907 | 0.300(0.026) | 0.353(0.014) | 561(16) |
$(0.35,0.4,0.45,0.3)$ | CR | 0.910 | 0.300(0.025) | 0.389(0.016) | 566(16) |
$(0.45,0.5,0.55,0.4)$ | DBCD | 0.894 | 0.400(0.027) | 0.361(0.012) | 473(17) |
$(0.45,0.5,0.55,0.4)$ | CR | 0.896 | 0.400(0.026) | 0.389(0.016) | 476(17) |
$(0.55,0.6,0.65,0.5)$ | DBCD | 0.894 | 0.500(0.028) | 0.366(0.011) | 383(17) |
$(0.55,0.6,0.65,0.5)$ | CR | 0.896 | 0.500(0.027) | 0.389(0.016) | 386(16) |
$(0.65,0.7,0.75,0.6)$ | DBCD | 0.928 | 0.599(0.027) | 0.369(0.010) | 293(15) |
$(0.65,0.7,0.75,0.6)$ | CR | 0.922 | 0.600(0.026) | 0.389(0.016) | 296(16) |
Table 12
Performance of DBCD targeting the intuitively ethical allocation proportion under ${H_{1}}$ when four treatments are under study.
(${p_{1}}$, ${p_{2}}$, ${p_{3}}$, ${p_{0}}$) | Design | Power | ${\hat{p}_{0}}$ | ${\rho _{0}}$ | Failure |
$(0.35,0.4,0.45,0.3)$ | DBCD | 0.911 | 0.299(0.028) | 0.317(0.026) | 556(17) |
$(0.35,0.4,0.45,0.3)$ | CR | 0.910 | 0.300(0.025) | 0.389(0.016) | 566(16) |
$(0.45,0.5,0.55,0.4)$ | DBCD | 0.894 | 0.399(0.029) | 0.333(0.021) | 469(17) |
$(0.45,0.5,0.55,0.4)$ | CR | 0.896 | 0.400(0.026) | 0.389(0.016) | 476(17) |
$(0.55,0.6,0.65,0.5)$ | DBCD | 0.896 | 0.499(0.029) | 0.342(0.018) | 380(17) |
$(0.55,0.6,0.65,0.5)$ | CR | 0.896 | 0.500(0.027) | 0.389(0.016) | 386(16) |
$(0.65,0.7,0.75,0.6)$ | DBCD | 0.926 | 0.599(0.028) | 0.349(0.015) | 290(16) |
$(0.65,0.7,0.75,0.6)$ | CR | 0.922 | 0.600(0.026) | 0.389(0.016) | 296(16) |
We further performed numerical studies for clinical trials with one control arm and three experimental treatment arms representing the low, medium, and high doses of the experimental drugs. The success rates for the control arm and three experimental treatment arms are ${p_{0}}$, ${p_{1}}$, ${p_{2}}$ and ${p_{3}}$, respectively. We keep the same sample size for Stage 2 as Tables 1–3, but increase the sample size to 400 for Stage 1, considering we have four treatment arms in this stage.
In the first scenario (Tables 7 and 10 for ${H_{0,M}}$ and ${H_{1,M}}$), let
\[\begin{aligned}{}{\boldsymbol{\rho }_{1}}(\boldsymbol{\theta })=\Big(& \frac{{p_{0}}}{{p_{0}}+{p_{1}}+{p_{2}}+{p_{3}}},\frac{{p_{1}}}{{p_{0}}+{p_{1}}+{p_{2}}+{p_{3}}},\\ {} & \frac{{p_{2}}}{{p_{0}}+{p_{1}}+{p_{2}}+{p_{3}}},\frac{{p_{3}}}{{p_{0}}+{p_{1}}+{p_{2}}+{p_{3}}}\Big)\end{aligned}\]
and ${\boldsymbol{\rho }_{2}}(\boldsymbol{\theta })=\left(\frac{{q_{M}}}{{q_{0}}+{q_{M}}},\frac{{q_{0}}}{{q_{0}}+{q_{M}}}\right)$ that is the urn allocation [64]. In the second scenario (Table 8 and 11 for ${H_{0}}$ and ${H_{1}}$), let ${\rho _{1k}}(\boldsymbol{\theta })=\frac{\sqrt{{p_{k}}}}{\sqrt{{p_{0}}}+\sqrt{{p_{1}}}+\sqrt{{p_{2}}}+\sqrt{{p_{3}}}}$, $k=0,1,2,3$, and ${\boldsymbol{\rho }_{2}}(\boldsymbol{\theta })=\left(\frac{\sqrt{{p_{0}}}}{\sqrt{{p_{0}}}+\sqrt{{p_{M}}}},\frac{\sqrt{{p_{M}}}}{\sqrt{{p_{0}}}+\sqrt{{p_{M}}}}\right)$ that is the optimal allocation [41]. In the third scenario (Table 9 and 12 for ${H_{0}}$ and ${H_{1}}$), we use the intuitively better allocation proportion
and
We detected similar conclusions to the numerical studies with two experimental treatment arms above. While keeping the same power level, our methods can always assign more people to better treatment and reduce the number of failures by up to 10 patients.4 Conclusion
Clinical trials are complex and involve a variety of design features related to efficiency and ethics. ASD and RAR have been proposed to achieve different aims [15, 17, 22, 23]. The desire to reduce development costs and the time-to-market of new treatments has led to the development of ASD. DBCD is a well-known RAR design with a variety of favorable properties. However, there has been limited theoretical and numerical study of the combination of ASD and DBCD, which hinders the development and application of this procedure. In this paper, we proposed a versatile approach and studied this complex procedure’s theoretical and numerical properties. Our methods can lead to less failure without sacrificing power than traditional designs while controlling the type I error rate.
[73] also tried implementing DBCD in seamless clinical trials. However, their methods depend on the method in [43] to construct the test statistics and control the type I error rate. As a result, strictly speaking, their methods can only be used for normal responses, and other future investigations of the procedure will require new challenging theoretical proof. This is a severe limitation in practice. The current paper proposed more versatile approaches based on the closure principle combined with the combination test and methods to address multiplicity problems, which FDA and pharmaceutical industry will more readily accept. More importantly, many existing approaches based on this framework, such as its combination with sequential monitoring and other endpoints, can be directly used in future trials. We leave these for future research. Fundamentally, we proposed a totally different and more flexible approach to implement DBCD in seamless clinical trials, which will significantly promote the procedure in clinical trials.
It is also worth discussing the benefit-cost tradeoff of the adaptive designs. First, exploring the seamless phase II/III design is often desirable in pharmaceutical companies for various reasons. For example, the regulatory agencies often require the comparison of a new dose in addition to the proposed two-arm clinical trial design, so a seamless phase II/III design often becomes one of some natural choices. RAR might make the design more complex compared to a fixed design. However, with the development of technology such as central data monitoring, interactive voice response services, and interactive web response service, the complexity of implementing advanced designs such as RAR is much reduced. Second, the evaluation of the reduction of total failures depends on the disease characteristics. For lethal diseases like the Ebola virus, failure means quick death, and any savings could be worth it.
This paper opens the door to future research topics. First, there are two families of RAR designs, DBCD and urn models [64, 65, 72]. It is worth exploring the seamless clinical trials with urn models. Second, research on adaptive randomization design and ASD under the Bayesian framework includes but is not limited to [3, 6, 7, 12, 28, 29, 54, 66, 67]. We can comprehensively compare our methods with the bayesian approaches. Third, [18, 28, 57] investigated the ASD with different types of study endpoints in the two phases. All these factors can be explored for the proposed design. We leave all these for future research topics.