In the development of new cancer treatment, an essential step is to determine the maximum tolerated dose in a phase I clinical trial. In general, phase I trial designs can be classified as either model-based or algorithm-based approaches. Model-based phase I designs are typically more efficient by using all observed data, while there is a potential risk of model misspecification that may lead to unreliable dose assignment and incorrect maximum tolerated dose identification. In contrast, most of the algorithm-based designs are less efficient in using cumulative information, because they tend to focus on the observed data in the neighborhood of the current dose level for dose movement. To use the data more efficiently yet without any model assumption, we propose a novel approximate Bayesian computation approach to phase I trial design. Not only is the approximate Bayesian computation design free of any dose–toxicity curve assumption, but it can also aggregate all the available information accrued in the trial for dose assignment. Extensive simulation studies demonstrate its robustness and efficiency compared with other phase I trial designs. We apply the approximate Bayesian computation design to the MEK inhibitor selumetinib trial to demonstrate its satisfactory performance. The proposed design can be a useful addition to the family of phase I clinical trial designs due to its simplicity, efficiency and robustness.
In oncology research, a phase I clinical trial often aims to determine the maximum tolerated dose (MTD), which is typically defined as the dose with the dose-limiting toxicity (DLT) probability closest to the target toxicity rate.1 In the development of new drugs, phase I clinical trials play an essential role because the selected MTD will be further investigated in the subsequent phase II or III trials. Misidentification of the MTD would lead to an unreliable conclusion and may even cause termination of a trial immaturely. As a result, this leads to a waste of abundant resources as well as exposing patients at excessively toxic doses, which violates the ethical principle. Moreover, as the first-in-human study, subjects available for a phase I trial are rather limited, and the sample size is typically around ,2 and thus it is challenging to identify the MTD with such a small sample size.
Various phase I trial designs have been proposed with the goal to determine the MTD both efficiently and accurately. Depending on whether to adopt a model assumption on the dose–toxicity curve or not, these designs can generally be classified into two branches: the algorithm-based (model-free or curve-free) and the model-based (often imposing a parametric model assumption) approaches. In the algorithmic family, the design3 is the most commonly used one for phase I oncology trials due to its simplicity and conservativeness.4 However, the design has also been criticized for its poor performance due to inefficient use of the data. Alternatively, a variety of model-free designs have been developed to improve the trial efficiency in determining the MTD. Gasparini and Eisele5 presented a curve-free method in which the probabilities of toxicity are directly modeled as an unknown multidimensional parameter. Ivanova et al.6 proposed a cumulative cohort design (CCD) where the dose escalation or de-escalation criterion is based on the Markov chain theory. To enhance the trial safety, the modified toxicity probability interval (mTPI) design is proposed using a unit probability mass.7 Under the Bayesian framework, Liu and Yuan8 developed the Bayesian optimal interval (BOIN) design by minimizing the probability of incorrect dose allocation. Yan et al.9 proposed the keyboard design by partitioning the toxicity probability scale into more and shorter intervals. Enlightened by the uniformly most powerful Bayesian test,10 Lin and Yin11 developed the uniformly most powerful Bayesian interval (UMPBI) design for phase I dose-finding trials.
Along the line of directly modeling the dose–toxicity curve, many model-based phase I designs have been proposed. The continual reassessment method (CRM)12,13 is the most popular model-based design, which typically adopts a single unknown parameter to link the true and the prespecified toxicity probabilities at different dose levels. Cheung and Chappell14 extended the CRM by incorporating weights to account for the late-onset toxicity outcomes. Another extension of the CRM focuses on modeling bivariate competing outcomes.15 By incorporating the Bayesian model averaging (BMA) approach, Yin and Yuan16 developed the BMA-CRM to overcome the arbitrariness of skeleton specification (i.e., the prespecified toxicity probabilities of all doses) and thus enhance the model robustness. Further extensions of the CRM include O’Quigley and Paoletti17 Yuan et al.18 and Wages et al.19 among others. Due to safety concerns, the escalation with overdose control (EWOC)20 is designed to locate the MTD subject to the constraint that the predicted proportion of patients receiving overdoses does not exceed a specified threshold. Without assuming any parametric model, the nonparametric overdose control design is developed to enhance model robustness with little sacrifice on trial efficiency.21 For more detailed discussions on the characteristics of various phase I designs, refer to Zhou et al.22
Our work is motivated by a phase I trial on the MEK inhibitor selumetinib in children with progressive low-grade gliomas (LGG).23 There were three candidate doses in the trial: 25, 33, and 43 mg//dose bis in die. The target DLT rate was . Originally, the trial adopted the likelihood-based modified CRM using a two-parameter logistic model based on dosages adjusted for the body surface area. However, because the prior information on the dose–toxicity curve is very limited at the phase I trial stage, model-based designs might be at risk of violating the parametric assumption, which undermines the efficiency of the dose-finding procedure. Moreover, with only three doses under investigation, a parametric model may not fit the data well. On the other hand, if we choose an algorithm-based method to select a dose for an incoming cohort of patients, most of the existing methods only consider the data collected at the current dose level, which weakens the efficiency of the design by not fully using all the cumulative data in the trial thus far. For example, a typical interval design determines the next dose level solely based on the data from the current dose level.
To alleviate the risk of model misspecification and utilize all the available information accrued in the trial, we develop an approximate Bayesian computation (ABC) design to identify the MTD. By adopting the idea from the approximate Bayesian computation sampling methods,24 the ABC design generates the weighted posterior samples based on all the available data without any complex formulas and dose–toxicity model assumptions. Based on the weighted samples, decisions among dose escalation, de-escalation, or retaining the current dose are made for each cohort. In this way, our method avoids introducing any explicit dose–toxicity model, yet can utilize the cumulative information from all dose levels when selecting the next dose level. Extensive simulation studies demonstrate that the ABC design is efficient compared with the state-of-the-art methods, while it is robust due to its model-free or curve-free feature. The main contributions of our work are twofold. First, we propose a novel phase I design which is robust and efficient under different scenarios. Second, by incorporating the ABC sampling into the phase I design, our work can be easily extended to other more complicated phase I trials.
The rest of the paper is organized as follows. In Section 2., we introduce the ABC design for dose finding in phase I clinical trials. We present the simulation studies to evaluate the operating characteristics of the ABC design and compare it with several well-known phase I designs in Section 3. An application to the phase I trial of the MEK inhibitor selumetinib is provided in Section 4. The paper is concluded with a brief discussion in Section 5.
Methodology
Suppose that a phase I clinical trial aims to investigate dose levels with the corresponding DLT rates . The target toxicity rate is denoted as . After enrolling and treating the first cohorts of patients, the current dose level is denoted by . Thus far, we observe the cumulative data and , where represents the number of observed DLTs and represents the number of patients treated at dose level .
The main goal of the ABC design is to adopt all the available information when deciding the dose level for the next cohort without imposing any explicit dose–toxicity model assumption. In the Bayesian framework, we first assign a prior on , and then we can update the posterior distribution of as
where is the likelihood function. Based on the posterior distribution, the next dose level can be selected accordingly. The major difficulty lies on how to choose a suitable and simple prior while taking the monotonicity constraint into account. The dilemma is that a parametric model assumption typically undermines the robustness of the design, while due to the monotonicity constraint, the specific form of the prior without such a parametric model assumption can be complicated. The unusual complexity of the prior would also diminishes the flexibility of the design. Further, it causes difficulties in calculating the posterior distribution as well as the subsequent Bayesian inference.
Approximate Bayesian computation
To circumvent the aforementioned problem, we adopt the idea from the ABC sampling methods.24 The ABC methods are typically used to handle complex models, where the evaluation of the likelihood function is computationally costly or elusive. In such cases, the Bayesian inference with a closed-form posterior distribution is prohibitive. The ABC methods bypass the evaluation of the likelihood function with Monte Carlo simulations.
Suppose that we are interested in obtaining the posterior samples of parameter given the data and prior . A simple ABC method, known as the ABC rejection algorithm,25 is described as follows:
Draw a sample from the prior .
Given the sample , generate data under the likelihood model using the prior predictive distribution.
If the generated data is close to the observed data under some prespecified distance measure, we keep the sample as the posterior sample; otherwise, the sample is rejected and discarded.
The procedure is repeated for a large number of times to obtain adequate posterior samples of .
In the phase I trial setting, the evaluation of the likelihood is rather simple, while the prior model is complicated without a closed form due to the monotonicity constraint. On the other hand, the monotonicity constraint is can be naturally incorporated when we implement Monte Carlo simulations. For example, we can first generate the samples without any constraint and then sort them in an ascending order to preserve the monotonicity in the prior samples of . In Section B of the Supplemental Appendix, we present a toy example of using different approaches to sampling multivariate random variables under the monotonicity constraint, which shows that Monte Carlo simulation gives a simple and efficient way to solve such monotonically constrained sampling problem. Therefore, the ABC methods provide a very simple solution to obtain the constrained posterior samples of in a Monte Carlo manner.
Optimal dose selection
With the ABC rejection algorithm in hand, we are still not ready to obtain the posterior samples of for a phase I trial. The main challenge lies in the low efficiency of the ABC rejection method, because its acceptance rate can be very low which makes the procedure rather slow to obtain an adequate number of posterior samples for reliable and robust inference. Moreover, under the ABC rejection method, all the ’s leading to the generated ’s inconsistent with the observed data are equally discarded. However, even the discarded ’s can still provide some useful information about the posterior distribution. For example, some generated data ’s may be very different from the original data , while some may only moderately deviate from , and thus we should not blindly discard the corresponding ’s without discrimination.
Considering both issues as mentioned above (the low acceptance rate and equally discarding unqualified posterior samples), we propose to adopt the weighed ABC sampling for a phase I trial as follows:
Select a suitable prior distribution and a distance measure .
Generate the prior samples for a large number times from .
Given the prior samples , we generate the corresponding data, with .
Based on the generated data , the weight for each sample can be obtained as , where is the originally observed data.
Use the weighted samples as the posterior samples for statistical inference.
Motivated by the Gaussian kernel, we choose the distance measure as
where is the bandwidth and its selection will be discussed in Sensitivity Analysis Section.
The above procedure is essentially referred to as the ABC importance sampling when the proposal density is taken as the prior distribution.26 Therefore, it is guaranteed that the weighed samples are approximately generated from the posterior distribution. This weighted procedure overcomes the low acceptance rate problem of the ABC rejection method, because it adopts all the prior (weighted) samples for the posterior inference. We then estimate the toxicity rate for dose level as
where can be any function yielding a reasonable estimator . By default, we take as the weighted median due to its robustness. The weighted median function is the th weighted percentile of obtained as follows:
Sort in an ascending order to obtain attached with the corresponding weights .
The weighted median is selected as satisfying
We adopt the weighted median because it is a more robust estimator compared with the weighted mean and the posterior mode may be ambiguous due to the multi-modal posterior distribution. More detailed discussions on with a numerical study are given in Section C of the Supplemental Appendix. The optimal dose for cohort based on the cumulative data up to cohort , and , is defined as
Prior elicitation
While the ABC design is flexible with respect to the choice of the prior , the prior samples play an important role for efficiently identifying the MTD. To enhance the robustness of the ABC design, we should generate samples without any parametric dose–toxicity assumption, for which an intuitive way is to first generate samples from and then sort them in an ascending order. However, this method does not take the target rate into consideration and, as a result, the prior samples are less informative and lead to poor performance.
Before conducting a phase I trial, we are typically provided with the target toxicity rate as well as the toxicity probability monotone constraint. Thus, it is desirable to encode both parts of information in the prior to obtain more informative samples. The main goal of the phase I design is to find the dose level whose DLT rate is closed to . Because each dose could possibly be the MTD, we can generate prior samples of from possible models , where is the model that dose level is the MTD while indicates that all the dose levels are overly toxic (i.e. no MTD). This way of incorporating into the model provides more informative prior samples for the ABC design.
Consequently, we propose to generate the prior samples as follows:
Considering the trade-off between computation and performance, we generate prior samples of from each model , which leads to a total number of prior samples. The prior samples from are obtained from all models, . These prior samples can be generated before the trial conduct and saved for repeated use.
Given with ,
Set dose level as the target with a DLT rate where is a prespecified small number.
We independently generate samples from and sort them in an ascending order to obtain .
We independently generate samples from and sort them in an ascending order to obtain .
Under , we independently generate samples from and sort them in an ascending order to obtain .
The neighborhood parameter controls the distinguishability of the target dose level in the generated prior samples. A larger value of indicates that the MTD is easier to be determined in the prior samples, and vice versa. In Sensitivity Analysis Section, we conduct extensive simulation studies to show that the ABC design is not sensitive to the choice of , and we recommend as a default value in practice.
Dose-finding algorithm
To ensure the safety and benefit for the patients, we further impose an early stopping criterion in the ABC design. We terminate the trial when there is strong evidence indicating the lowest dose level is still overly toxic. We assign Jeffreys’ prior distribution to the DLT rate , and terminate the trial if .
The dose-finding procedure of the ABC design is detailed as follows:
Treat the first cohort of patients at the lowest or the physician-specified dose level.
After enrolling cohorts, select the optimal dose level based on (1).
According to the optimal dose level ,
If , then .
If , then .
If , then .
The trial can be either stopped after exhaustion of the maximum sample size, or be terminated early for safety if the lowest dose level is too toxic by the early stopping rule, .
At the end of the trial, the observed data are collected, where is the total number of cohorts. The MTD is then estimated with another round of ABC simulation, that is, is finally selected as
If the DLT rate of the recommended MTD must be lower than the target, we can easily make a modification by first identifying , while if is higher than the target then we simply choose the next lower dose as the MTD.
In the dose-finding algorithm, we adopt the same early termination rule as those in most of the interval designs8,11 for a fair comparison. However, as our prior samples involve model (i.e. all doses are overly toxic), it is possible to construct our own early termination rule as
where is the threshold value, for example, .
Simulation studies
Sensitivity analysis
We investigate the effect of the parameters and on the performance of the ABC design with the analysis of variance (ANOVA) method.27 We randomly generate dose–toxicity scenarios following the procedure in Paoletti et al.28 as described in Section A.2 of the Supplemental Appendix. To conduct a comprehensive analysis with randomly generated scenarios, we consider four influential factors in phase I trials, including the average probability difference around the target, the number of dose levels , the sample size as well as the target toxicity rate . The first three factors affect the difficulty of the MTD-identification task, where larger , smaller , and larger sample size would typically result in higher MTD selection percentages. The possible levels of the four factors are listed in Table 1, which yields different configurations. Under each configuration, we investigate the performance of the ABC design via randomly generated scenarios when takes a value of or is randomly selected from Uniform and is selected from .
The simulation factors that may affect the dose-finding performance of phase I trial design and the results of ANOVA in terms of the percentage of MTD selection. The ANOVA also includes all the pairwise interactions between the five simulation factors.
Factors
Levels of factor
DF
SS
MSE
Average probability difference around ()
{0.05, 0.07, 0.10, 0.15}
3
65.35
21.78
Bandwidth parameter
{0.1, 0.05, 0.01, 0.005}
3
4.32
1.44
Sample size
7
5.58
0.80
Number of dose levels
{3, 5, 7}
2
0.81
0.40
Target toxicity probability
{0.25, 0.30, 0.33}
2
0.34
0.17
Neighborhood parameter
{0.0, 0.05, 0.10, 0.15, 0.20, random}
5
0.54
0.11
Total variance
6911
83.39
MTD: maximum tolerated dose; DF: degree of freedom; SS: sum of squares; MSE: mean squared error (MSESS/DF).
After obtaining the percentage of MTD selection under each configuration with different values of , we perform ANOVA with regard to these percentages using the simulation factors including all the pairwise interactions in Table 1. In the ANOVA, we also regard the tuning parameters and as two additional factors in the evaluation of dose-finding performance. Thus, the degree of freedom of the total variance for ANOVA is . Among the six influential factors, the neighborhood parameter has the least effect on the performance of the ABC design in terms of the mean squared error (MSE). In particular, only accounts for of the MTD selection percentage variance, which indicates the ABC design is robust to the choice of . The performance of the ABC design is more sensitive to the bandwidth parameter , because it is the second most influential factor on the MTD selection percentage.
From Table 1, it is clear that the average probability difference around the target is the dominating factor for the percentage of MTD selection, as it contributes the largest MSE (significantly larger than the second) in the ANOVA. We further study the effect of the tuning parameters and on the MTD selection percentage under different values of . Figure 1 shows the MTD selection percentage versus the tuning parameters and for . Under various levels of trial difficulty (the difficulty of MTD identification decreases as increases), shows relatively minor effect on the performance of the ABC design, demonstrating the insensitivity of the design to . As for the bandwidth parameter , it is clear that a larger value of would undermine the performance of the ABC design. When is decreased close to , the performance of the ABC design is saturated and further reduction of would not improve the trial performance. In practice, it is recommended to choose any value of and for the ABC design, and by default we set and .
The maximum tolerated dose (MTD) selection percentage versus the neighborhood parameter (left) and bandwidth parameter (right) under the probability difference around the target . We set , respectively, and also consider the case of randomly choosing from Uniform and takes a value of .
Evaluation under random scenarios
To make an extensive comparison of the ABC design with existing methods, we select six state-of-the-art phase I designs, including the CRM design with the power model whose model skeleton is selected using the method of Lee and Cheung29 the CCD design6 which is based on the Markov chain theory, the modified toxicity probability interval (mTPI) design,7 the Bayesian optimal interval design (BOIN),8 the keyboard design,9 as well as the uniformly most powerful Bayesian interval (UMPBI) design.11 Among the six designs, the CRM is the only model-based one, while the other five are algorithm-based methods and, more specifically, they are all interval designs. Unless otherwise stated, all the six existing designs adopt the default parameters following the original papers and we utilize the same early stopping rule for the five interval designs and the ABC design, that is, if , we terminate the trial for safety. For the early termination rule under the CRM, the posterior probability of is calculated based on the CRM model and the threshold probability is still set at . The detailed settings for the six existing designs are given in Section A.1 of the Supplemental Appendix.
We investigate five dose levels for each trial and set the target toxicity rate and , respectively. The sample size is and patients are treated in a cohort size of . To avoid cherry-picking, we randomly generate dose–toxicity scenarios following the procedure in Paoletti et al.,28 which is detailed in Section A.2 of the Supplemental Appendix. The average probability difference around the target is controlled at , , and , respectively, and under each value of , we replicate simulations.
Four summary statistics are adopted to evaluate the performances of the seven designs under comparison. The two main measurements, reflecting the accuracy and efficiency of a design, are the percentage of MTD selection and the percentage of patients treated at the MTD (MTD allocation), for which larger values are preferred. The remaining two measurements quantify the safety aspects of a trial, including the percentage of trials that select overdoses as the MTD (overdose selection), and the percentage of patients allocated to overdoses (overdose allocation). A design with smaller values of the two safety statistics would be considered more ethical and desirable. To quantify the variability, we adopt the bootstrap method to construct the 95% confidence intervals for each measurement, that is, we resample trial results with replacement from a total of repetitions (random scenarios) and repeat the resampling procedure for times to obtain the confidence intervals.
The results with are shown in Figure 2. The performances of all the designs improve as increases. The efficiency of using all the available data under the ABC and CRM designs is reflected by the MTD allocation metric, for which both designs yield higher MTD allocation percentages compared with the other interval designs. The gap becomes more significant when increases. In terms of the MTD selection percentage, both the ABC and CRM designs are superior to the interval designs when is small; the gap diminishes when becomes large. When is large, the information from the MTD neighborhood data may be adequate to identify the MTD well, and thus using all the available data does not boost the performance greatly. Under the setting of , BOIN performs best, while CRM yields the lowest MTD selection percentage compared with the counterparts. The performances of the ABC design are robust across all different values of , which reveals the advantage of imposing no dose–toxicity assumption in ABC.
Simulation results with sample size based on 5000 randomly generated dose–toxicity scenarios under the average probability difference of , 0.07, 0.10 and 0.15 around the target toxicity probability , respectively. The error bar represents the 95% confidence interval for each measurement.
With regard to the two safety measurements, the CRM has the highest percentages of overdose selection and overdose allocation among the seven designs. The CCD design performs best in terms of the safety metrics, while its performances are worse for the two accuracy measurements. The other five designs are comparable in terms of the safety metrics.
The results with are presented in Figure 3. Overall, the performances of the seven designs are similar to those under the settings of . However, when , the mTPI design appears to be very aggressive. The mTPI yields similar performances to the ABC and CRM designs in terms of the MTD selection and allocation metrics, while it sacrifices safety significantly in contrast with other methods. In summary, numerical comparisons with the six well-known methods demonstrate the robustness and efficiency of the ABC design.
Simulation results with sample size based on 5000 randomly generated dose–toxicity scenarios under the average probability difference of , 0.07, 0.10 and 0.15 around the target toxicity probability , respectively. The error bar represents the 95% confidence interval for each measurement.
Evaluation under fixed scenarios
To gain more insight into the ABC design, we evaluate its performance under five representative fixed scenarios. For an objective comparison with no cherry-picking, we adopt the fixed scenarios in Table 1 of Cheung and Chappell.14 The number of dose levels is and the target toxicity rate is . The comparisons are again made with the six phase I designs under the similar settings as introduced in Section A.1 of the Supplemental Appendix. The total sample size is and patients are treated with a cohort size of . We use the same early stopping rules as those under the random scenarios. Under each scenario, we replicate 5000 simulations.
The detailed information and results under the fixed scenarios are summarized in Table 2. Under scenario 1, the CRM has the best performance, while the other designs yield comparable results. In scenario 2, all the dose levels are overly toxic, and the ABC design and all interval designs show satisfactory and comparable results by early stopping the trials, but the CRM yields a slightly lower non-selection percentage () compared with others. The ABC design leads to the best performance under both scenarios 3 and 4 and the gaps of MTD selection percentages are around under scenario 3. Under scenario 5 where the last dose level is the MTD, the mTPI design is substantially better than others, while it tends to select over-toxic dose level as the MTD as shown in scenarios 1, 3, and 4. Except for the mTPI design, the ABC design has slightly higher over-dose selection percentages under scenarios 1, 3, and 4, while the gaps are marginal. Overall, the ABC design yields satisfactory performances for all of the five fixed scenarios.
The percentages of MTD selection (the numbers of patients treated at each dose) under the ABC design in comparison with the BOIN, CCD, CRM, Keyboard, mTPI and UMPBI designs under six fixed scenarios with the target toxicity probability 0.20 in boldface when sample size is . None represents the percentage of trials of non-selection due to early stopping.
As an illustration, we apply the ABC design to the aforementioned phase I trial on the MEK inhibitor selumetinib in children with progressive LGG. The DLT outcomes were defined as any grade 4 toxicity (except lymphopenia), grade 3 neutropenia with fever, grade 3 thrombocytopenia with bleeding, any grade 3 or 4 toxicity possibly related to selumetinib, or any grade 2 toxicity persisting days that was medically significant or intolerable enough to interrupt or reduce the dose. Originally, the trial evaluated patients to estimate the MTD with the target toxicity rate . Patients were assigned in a cohort size of to one of the three dose levels of the MEK inhibitor selumetinib mg/m/dose bis in die via the CRM design based on the two-parameter logistic model. The observed data at dose levels 1–3 from the original clinical trial were
Based on the observed data in the trial, the estimated DLT rates were , and the MTD selected using the CRM method was dose level 1.23
We applied the ABC design with and to this trial based on the estimated DLT rates. There were 13 cohorts in total where the last cohort contained only one subject. For comparison, we also include the CRM design using the two-parameter logistic model, for which the detail is given in Section A.1 of the Supplemental Appendix. The entire procedure is repeated for times and the results are presented in Table 3. It is clear that the ABC design yields comparable performances to the CRM design under this trial example. Nevertheless, the ABC design is model-free, and thus it is more robust and easier to use in practice.
The percentages of MTD selection (the numbers of patients treated at each dose) under the ABC and CRM designs with 5000 replications. The target toxicity probability is 0.25 in the real trial application with 37 patients.
To better demonstrate the property of the ABC design, we present the detailed trial flow of one specific trial selected from the repetitions. The patient allocations and outcomes are shown in Figure 4. The trial started by treating the first cohort of patients at the lowest dose level. In the first cohort, there was no DLT observed, which yielded the estimated DLT rates . Thus, the trial decided to escalate to dose level 2 for the second cohort, where we observed two DLTs among three patients, resulting in the estimated DLT rates . Consequently, the next cohort was assigned back to dose level 1, where again no patient experienced the DLT, leading to . Consequently, the trial escalated to dose level 2 for cohort 4 and one out of the three patients experienced the DLT, which led to . As a result, the trial remained at dose level 2 for the next cohort, but two DLT outcomes were observed among the three patients and thus the trial de-escalated back to dose level 1. All the remaining cohorts were assigned to dose level 1. Finally, the observed data at dose levels 1–3 were
Upon the completion of the trial, we selected dose level 1 (i.e. the dose of 25 mg/m/dose bis in die) as the MTD, which is consistent with the original selection by the CRM design.23
Patient allocations and outcomes by the ABC design for the real trial illustration using a representative trial among the repetitions. The ABC recommended MTD is dose level 1.
MTD: maximum tolerated dose; ABC: approximate Bayesian computation.
Conclusion
We have proposed a new phase I design for identifying the MTD using the approximate Bayesian computation method. The ABC design possesses the merits of both model-free and model-based designs simultaneously. Because it is model-free, there is no need to impose any assumption on the dose–toxicity curve, which avoids the risk of model misspecification. Similar to the model-based methods, the ABC design is also efficient by aggregating all the available information when deciding dose assignment for each new cohort. As demonstrated by simulation studies, the ABC design performs well in terms of both the efficiency and robustness in phase I trials. Compared with other phase I designs, the ABC design shows advantages in terms of the MTD selection and patient allocation under the random scenarios. There are two tuning parameters and in the ABC design. The neighborhood parameter has rather minor effect on the design performance, as shown by the comprehensive simulation studies and ANOVA in Sensitivity Analysis Section, while the bandwidth parameter can be simply fixed at to achieve satisfactory performance. In practice, the ABC design is easy to use, because there is no need to carry out the extensive parameter calibration prior to the trial conduct. Therefore, the ABC design can be broadly used in phase I clinical trials due to its robust and efficient properties as well as the ease of implementation. Under some cases in the simulation studies, our design shows a slightly higher but acceptable overdose selection percentage compared with existing phase I designs as a sacrifice for higher MTD selection and MTD allocation percentages. This is a trade-off often encountered in dose finding: by exploring more doses, it would help to pin down the MTD more accurately, while at the same time more patients might be put at the risk of exploring untried (higher) dose levels. The ABC design does not allow inserting some intermediate dose level during the trial, because a dose–toxicity model is typically needed for such dose insertion. Nevertheless, one possibility along this direction is to incorporate a working model to pin down the MTD more precisely through interpolation and dose insertion.
By incorporating the ABC sampling, our design can be easily extended to other more complicated phase I trials. To account for late-onset outcomes, the ABC design can be combined with the fractional scheme30 in a straightforward way. The only modification is to tune the bandwidth parameter under the late-onset context. To accommodate the efficacy outcomes, we can introduce the admissible set and use a beta-binomial model to estimate the efficacy rates and thus deliver decisions through a trade-off between toxicity and efficacy. The ABC design can also be extended to the dose combination trials with an adaptation on generating prior samples, which warrants further investigation.
The implementation of the ABC design is simple and fast due to the fact that the prior samples can be generated beforehand. An application of our ABC design for dose finding (https://github.com/JINhuaqing/ABC) has been developed, where users can set various customized configurations for their trials and obtain visualization results of the ABC design. The R scripts for simulation studies as well as the real data application are available at https://github.com/JINhuaqing/ABC-simu.
Supplemental Material
sj-pdf-1-smm-10.1177_09622802221122402 - Supplemental material for Approximate Bayesian computation design for phase I clinical trials
Supplemental material, sj-pdf-1-smm-10.1177_09622802221122402 for Approximate Bayesian computation design for phase I clinical trials by Huaqing Jin, Wenbin Du and Guosheng Yin in Statistical Methods in Medical Research
Footnotes
Acknowledgements
We thank the two referees for their careful reviews of our manuscript and many insightful comments that have significantly improved the paper.
Declaration of conflicting interests
The authors declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Funding
The author(s) disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: This research was partially supported by funding from the Research Grants Council of Hong Kong (17308321).
ORCID iDs
Huaqing Jin
Wenbin Du
Guosheng Yin
Supplemental material
Supplemental material for this article is available online.
References
1.
YinG. Clinical trial design: Bayesian and frequentist adaptive methods. vol. 876. Hoboken, NJ: John Wiley & Sons, 2012.
2.
IasonosAO’QuigleyJ. Adaptive dose-finding studies: a review of model-guided phase I clinical trials. J Clin Oncol2014; 32: 2505.
3.
StorerBE. Design and analysis of phase I clinical trials. Biometrics1989; 45: 925–937.
4.
YuanYYinG. Bayesian hybrid dose-finding design in phase I oncology clinical trials. Stat Med2011; 30: 2098–2108.
5.
GaspariniMEiseleJ. A curve-free method for phase I clinical trials. Biometrics2000; 56: 609–615.
6.
IvanovaAFlournoyNChungY. Cumulative cohort design for dose-finding. J Stat Plan Inference2007; 137: 2316–2327.
7.
JiYLiuPLiY, et al. A modified toxicity probability interval method for dose-finding trials. Clinical Trials2010; 7: 653–663.
8.
LiuSYuanY. Bayesian optimal interval designs for phase I clinical trials. J R Stat Soc: Ser C (Applied Statistics)2015; 64: 507–523.
9.
YanFMandrekarSJYuanY. Keyboard: a novel Bayesian toxicity probability interval design for phase I clinical trials. Clin Cancer Res2017; 23: 3994–4003.
10.
JohnsonVE. Uniformly most powerful Bayesian tests. Ann Stat2013; 41: 1716.
11.
LinRYinG. Uniformly most powerful Bayesian interval design for phase I dose-finding trials. Pharm Stat2018; 17: 710–724.
12.
O’QuigleyJPepeMFisherL. Continual reassessment method: a practical design for phase 1 clinical trials in cancer. Biometrics1990; 46: 33–48.
13.
O’QuigleyJShenLZ. Continual reassessment method: a likelihood approach. Biometrics1996; 52: 673–684.
14.
CheungYKChappellR. Sequential designs for phase I clinical trials with late-onset toxicities. Biometrics2000; 56: 1177–1182.
15.
BraunTM. The bivariate continual reassessment method: extending the CRM to phase I trials of two competing outcomes. Control Clin Trials2002; 23: 240–256.
16.
YinGYuanY. Bayesian model averaging continual reassessment method in phase I clinical trials. J Am Stat Assoc2009; 104: 954–968.
17.
O’QuigleyJPaolettiX. Continual reassessment method for ordered groups. Biometrics2003; 59: 430–440.
18.
YuanZChappellRBaileyH. The continual reassessment method for multiple toxicity grades: a Bayesian quasi-likelihood approach. Biometrics2007; 63: 173–179.
19.
WagesNAConawayMRO’QuigleyJ. Continual reassessment method for partial ordering. Biometrics2011; 67: 1555–1563.
20.
BabbJRogatkoAZacksS. Cancer phase I clinical trials: efficient dose escalation with overdose control. Stat Med1998; 17: 1103–1120.
21.
LinRYinG. Nonparametric overdose control with late-onset toxicity in phase I clinical trials. Biostatistics2017; 18: 180–194.
22.
ZhouHYuanYNieL. Accuracy, safety, and reliability of novel phase I trial designs. Clin Cancer Res2018; 24: 4357–4364.
23.
BanerjeeAJakackiRIOnar-ThomasA, et al. A phase I trial of the mek inhibitor selumetinib (AZD6244) in pediatric patients with recurrent or refractory low-grade glioma: a pediatric brain tumor consortium (PBTC) study. Neuro-oncology2017; 19: 1135–1144.
24.
RubinDB. Bayesianly justifiable and relevant frequency calculations for the applied statistician. Ann Stat1984; 12: 1151–1172.
25.
BeaumontMA. Approximate Bayesian computation in evolution and ecology. Annu Rev Ecol Evol Syst2010; 41: 379–406.
26.
SissonSAFanYBeaumontM. Handbook of approximate Bayesian computation. CRC Press, 2018.
27.
CangulMChretienYRGutmanR, et al. Testing treatment effects in unconfounded studies under model misspecification: logistic regression, discretization, and their combination. Stat Med2009; 28: 2531–2551.
28.
PaolettiXO’QuigleyJMaccarioJ. Design efficiency in dose finding studies. Comput Stat Data Anal2004; 45: 197–214.
29.
LeeSMCheungYK. Model calibration in the continual reassessment method. Clinical Trials2009; 6: 227–238.
30.
YinGZhengSXuJ. Fractional dose-finding methods with late-onset toxicity in phase I clinical trials. J Biopharm Stat2013; 23: 856–870.
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.