Hostname: page-component-cd9895bd7-gxg78 Total loading time: 0 Render date: 2024-12-26T16:13:33.452Z Has data issue: false hasContentIssue false

Recent innovations in adaptive trial designs: A review of design opportunities in translational research

Published online by Cambridge University Press:  28 April 2023

Alexander M. Kaizer*
Affiliation:
Department of Biostatistics and Informatics, University of Colorado Anschutz Medical Campus, Aurora, CO, USA
Hayley M. Belli
Affiliation:
Department of Population Health, New York University Grossman School of Medicine, New York, NY, USA
Zhongyang Ma
Affiliation:
Department of Population Health, New York University Grossman School of Medicine, New York, NY, USA
Andrew G. Nicklawsky
Affiliation:
Department of Pediatrics, University of Colorado Anschutz Medical Campus, Aurora, CO, USA
Samantha C. Roberts
Affiliation:
Department of Biostatistics and Informatics, University of Colorado Anschutz Medical Campus, Aurora, CO, USA
Jessica Wild
Affiliation:
Department of Biostatistics and Informatics, University of Colorado Anschutz Medical Campus, Aurora, CO, USA
Adane F. Wogu
Affiliation:
Department of Biostatistics and Informatics, University of Colorado Anschutz Medical Campus, Aurora, CO, USA
Mengli Xiao
Affiliation:
Department of Biostatistics and Informatics, University of Colorado Anschutz Medical Campus, Aurora, CO, USA
Roy T. Sabo
Affiliation:
Department of Biostatistics, Virginia Commonwealth University, Richmond, VA, USA
*
Corresponding author: A. Kaizer, Email: alex.kaizer@cuanschutz.edu
Rights & Permissions [Opens in a new window]

Abstract

Clinical trials are constantly evolving in the context of increasingly complex research questions and potentially limited resources. In this review article, we discuss the emergence of “adaptive” clinical trials that allow for the preplanned modification of an ongoing clinical trial based on the accumulating evidence with application across translational research. These modifications may include terminating a trial before completion due to futility or efficacy, re-estimating the needed sample size to ensure adequate power, enriching the target population enrolled in the study, selecting across multiple treatment arms, revising allocation ratios used for randomization, or selecting the most appropriate endpoint. Emerging topics related to borrowing information from historic or supplemental data sources, sequential multiple assignment randomized trials (SMART), master protocol and seamless designs, and phase I dose-finding studies are also presented. Each design element includes a brief overview with an accompanying case study to illustrate the design method in practice. We close with brief discussions relating to the statistical considerations for these contemporary designs.

Type
Review Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2023. Published by Cambridge University Press on behalf of The Association for Clinical and Translational Science

Introduction

Clinical trials are considered the gold standard of evidence in clinical research [1,Reference Zabor, Kaizer and Hobbs2]. However, modern clinical research problems are becoming increasingly complex while available resources may be limited. Traditional clinical trial designs with fixed sample sizes and only one or two arms may be limited to efficiently address emerging research questions. One such pathway to address these limitations is through the use of “adaptive” clinical trial designs that allow for the prospective modification based on the accumulating data in a trial [1,Reference Li, Lin and Lin3Reference Chow, Chang and Pong7].

In 2019, the US Food and Drug Administration (FDA) published guidance on the use and implementation of adaptive designs for clinical trials [1]. This guidance included nonbinding recommendations for the design, conduct, and reporting of such clinical trials. The FDA noted that adaptive designs may have multiple advantages over traditional, nonadaptive designs for improving statistical efficiency of trial operating characteristics (e.g., type I error rates and power), addressing ethical considerations (e.g., stopping a trial early if an intervention shows safety concerns or limited benefit, increasing randomization to better-performing arms), adding to our understanding of treatment effects (e.g., enriching a trial with more participants expected to benefit), and being more acceptable to stakeholders (e.g., increased flexibility to successfully complete the trial) [1,8,9]. These adaptive designs have the potential to be applied across clinical trial phases, from phase I dose-finding studies to phase III confirmatory trials [1,Reference Zabor, Kaizer and Hobbs2].

This review article first presents the seven major adaptive design elements presented in the 2019 FDA guidance document, with a brief overview of the designs presented in Table 1. A case study is presented in each design subsection to illustrate the approach in practice. We follow this with an introduction of special topics relating to other novel trial designs and methodologies that have emerged regarding borrowing historic information, SMART designs, master protocol and seamless designs, and phase I dose-finding studies. The final section presents general statistical considerations for adaptive trial designs before concluding with a brief discussion.

Table 1. Brief summary of FDA guidance document adaptive design elements

FDA Adaptive Trial Elements

Group Sequential Designs

Group sequential designs allow for prospectively planned interim analyses and possible early stopping due to efficacy, futility, or harm [1]. These stopping rules may be binding (e.g., the trial must stop) or nonbinding (e.g., the trial is recommended to stop, but other considerations may warrant continuing the study). Since the early works of Pocock [Reference Pocock10] in 1977 and O’Brien and Fleming [Reference O'Brien and Fleming11] in 1979, the choice of efficacy-stopping boundaries that control the overall type I error rate have been broadly discussed in the literature (e.g., Lan and DeMets [Reference Demets and Lan12,Reference Gordon Lan and DeMets13]; Wang and Tsiatis [Reference Wang and Tsiatis14]; Jennison and Turnbull [Reference Jennison and Turnbull15]; Chow and Chang [Reference Chow and Chang5]). With regard to futility stopping rules, either binding or nonbinding, type I error rate control is guaranteed but type II error may be inflated [Reference Schüler, Kieser and Rauch16].

The related statistical approaches can be divided into frequentist methods, including conditional power [Reference Lachin17] and beta-spending functions [Reference Pampallona and Tsiatis18Reference Rudser and Emerson21]; and Bayesian methods based on the predictive power [Reference Spiegelhalter, Freedman and Blackburn22], posterior, or predictive probability [Reference Saville, Connor, Ayers and Alvarez23]. Variations of adaptive group sequential designs have been developed with modifications of sample size [Reference Cui, Hung and Wang24Reference Posch and Bauer26], treatment arm selection [Reference Stallard and Todd27Reference Kelly, Stallard and Todd29], and patient enrichment [Reference Müller and Schäfer30,Reference Magnusson and Turnbull31] and are discussed in later sections. It is worth noting that conventional methods tend to produce biased estimates and undesirable coverage probability of confidence intervals; therefore, various methods for the estimation of the treatment effect that appropriately adjust for group sequential stopping rules have been established [1,Reference Jennison and Turnbull15,Reference Perkins, Quinn and Deakin32].

Case Study: PARAMEDIC2 was a Phase III randomized, placebo-controlled trial to test the efficacy of epinephrine in out-of-hospital cardiac arrest patients on 30-day survival rates (EudraCT number 2014-000792-11) [Reference Perkins, Quinn and Deakin32,Reference Perkins, Ji and Deakin33]. A frequentist group sequential design was used with a total of 10 pre-specified interim analyses spaced every 3 months. Since epinephrine was the standard treatment, a higher level of evidence was required to stop early for futility. To account for this different expectation of evidence for futility and efficacy, asymmetric stopping boundaries were implemented including Pocock’s and O’Brien Fleming’s alpha-spending functions for efficacy and futility, respectively [Reference Perkins, Quinn and Deakin32]. Although epinephrine showed a significantly higher survival rate in the final analysis, the trial missed an opportunity to stop early for efficacy due to slower recruitment and lower survival rates than were expected when designing the study [Reference Perkins, Ji and Deakin33,Reference Ryan, Stallard, Lall, Ji, Perkins and Gates34]. Through simulation studies, the trial was re-constructed using three alternative Bayesian group sequential designs where the stopping decisions were based on the posterior probability of superiority and harm [Reference Ryan, Stallard, Lall, Ji, Perkins and Gates34]. The team then virtually re-executed the trial using PARAMEDIC2 data where only two designs recommended stopping early for benefit.

Sample Size Re-Estimation

Sample size re-estimation (SSR) is a form of clinical trial adaptation that accounts for potential uncertainty in the expected treatment effect. After the initial sample size calculation is performed to estimate the number of participants required to maintain prespecified power and type I error rate given an assumed treatment effect, the accumulated data at each interim analysis can then be used to re-estimate the sample size in order to maintain the same power if the observed treatment effect is weaker than expected but still clinically relevant. Many distinct methods exist to perform SSR, including blinded (i.e., individual group assignment remains hidden in the process of re-estimating) and unblinded (i.e., assignment is known for the process of re-estimating) methods [Reference Gould35]. Bayesian SSR methods also exist [Reference Brakenhoff, Roes and Nikolakopoulos36,Reference Sobel and Turkoz37], as do methods extending SSR to complex study designs [Reference Liu, Li, Lin, Lin and Hoffman38Reference Gao, Ware and Mehta40], less common statistical models [Reference Asendorf, Henderson, Schmidli and Friede41], and time to event data [Reference Schneider, Schmidli and Friede42]. SSR is most commonly used in mid to large scale phase II/III confirmatory trials [Reference Pritchett, Menon and Marchenko43]. Further, re-estimation methods can be tailored to equivalence, noninferiority, or superiority hypothesis testing contexts [Reference Friede and Kieser44Reference Golkowski, Friede and Kieser46]. Because it relies on interim analyses from incomplete data, SSR may inflate the type I error rate [Reference Wang and Chow47]. SSR which depends on re-estimating only nuisance parameters (e.g., variance) or uses a blinded approach has little impact on type I error rates, but unblinded methods may lead to an inflation of the type I error rate [Reference Proschan48Reference Friede and Schmidli50]. This type I error inflation can be addressed differently depending on how the re-estimation is performed and is always an important consideration when planning or implementing SSR. Used appropriately, SSR allows for uncertainty in the magnitude of the treatment effect, letting trials readjust their total sample size throughout the trial based on accumulated information without loss of statistical power.

Case Study: One adaptive trial with a planned SSR element is the ongoing Nasal Oxygen Therapy After Cardiac Surgery (NOTACS) trial (ClinicalTrials.gov number: NCT05308719) [Reference Dawson, Chiu, Klein, Earwaker and Villar51]. This trial is evaluating the use of high-flow nasal oxygen therapy (HFNT) compared to standard oxygen therapy (SOT) for cardiac surgery patients at increased risk for postoperative pulmonary complications. The sample size for this two-armed trial is restricted to be between 850 and 1252 patients, with a single interim sample size re-estimation analysis planned after at least 300 patients have accumulated 90 days of follow-up. This design allows the trial to increase the sample size above 850 patients if, based on interim data, the treatment effect of HFNT is lower than originally expected but still clinically meaningful enough to justify continuing the trial with the necessary higher sample size to maintain statistical power to detect an effect.

Adaptive Enrichment

In trials where it is believed that a subgroup of the population will experience more benefit from an intervention, an adaptive enrichment design provides the ability to drop lower performing subgroups at an interim analysis so that study resources are more efficiently allocated to those with a greater chance of benefit [1,9,Reference Thall52Reference Burnett and Jennison54]. In general, two broad classes exist: prognostic and predictive enrichment [9]. Prognostic designs attempt to choose patients more likely to have the study endpoint or have worsening conditions (e.g., event- or progression-based studies), whereas predictive designs attempt to identify patients with a higher chance of responding to the given treatment [9].

These designs are most common for confirmatory clinical trials, but may be used in earlier phases [1,Reference Burnett and Jennison54,Reference Lai, Lavori and Tsang55]. The a priori subgroups may be identified by either a single biomarker or multiple biomarkers, but designs may also include methods to identify adaptive thresholds for continuous biomarkers [Reference Thall52,Reference Simon and Simon53]. An advantage of enrichment designs compared to a study that only enrolls the targeted subgroup is that the overall population, including nontargeted subgroups, may be evaluated for potential benefit while providing the ability to drop the nontargeted subgroups with no benefits or worsening conditions [1]. A challenge of enrichment designs is accounting for the multiplicity of both repeated testing of study hypotheses and across multiple subgroups [1,Reference Thall52,Reference Burnett and Jennison54]. Burnett and Jennison discuss approaches to control the type I error rate within an adaptive enrichment trial, including a strong family-wise control [Reference Burnett and Jennison54]. Regardless of whether study arms are dropped at an interim analysis, the data from the overall study may be used for statistical inference [1].

Case Study: An example of a clinical trial that included an adaptive enrichment element is the Phase III Trial of TRC105 and Pazopanib Versus Pazopanib Alone in Patients With Advanced Angiosarcoma (TAPPAS; ClinicalTrials.gov number: NCT02979899) [Reference Jones, Ravi and Brohl56]. Due to uncertainty of the treatment effect for progression-free survival (PFS) among cutaneous and noncutaneous disease subgroups of angiosarcoma, an adaptive enrichment strategy based on conditional power was used after 40 events had occurred or within 30 days after enrollment of 120 participants to determine if the trial should continue without any modifications, should potentially increase the overall sample size (i.e., sample size re-estimation), or if the noncutaneous subgroup should be dropped with an increase of the sample size for the cutaneous group [Reference Jones, Ravi and Brohl56]. While there was no formal criteria for stopping for futility, the TAPPAS Trial design specified an “informal futility zone” that the Data Monitoring Committee for the study could use to stop the trial at an interim analysis [Reference Mehta, Liu and Theuer57]. Ultimately, the trial terminated for futility after 123 participants were enrolled because the study had entered the informal futility zone with a hazard ratio of 0.98 for PFS between the two groups (p = 0.95) [Reference Jones, Ravi and Brohl56].

Treatment Arm Selection

Adaptations to treatment arm selection are methods that modify the study design to allow for adding or dropping treatment arms. Some examples are adaptive dose-finding, drop-the-losers, adaptive seamless, and adaptive platform designs [Reference Chow4,Reference Chow and Chang6]. These adaptations can be used in both Bayesian and frequentist trial designs. Some of the applications are described below, as well as in subsequent sections detailing additional considerations of these types of design modifications.

Adaptive dose-finding designs are usually used in the early phases of studies to determine the minimum effective dose or the maximum tolerable dose for a drug to use for future trials. One method used for dose-finding designs is the continual reassessment method where the dose relationship is assessed throughout the trial using the data collected during the study [Reference Iasonos and O'Quigley58].

The adaptive seamless design combines the objectives of two trials into one. Seamless design trials can combine phases I and II or phases II and III. This type of design eliminates the time between the two phases of a trial and can lead to a smaller overall sample size required to conduct the final analysis where it uses data collected from subjects enrolled before and after the adaptation [Reference Maca, Bhattacharya, Dragalin, Gallo and Krams59].

Drop-the-losers clinical trial designs allow the dropping of the inferior treatment(s) or adding additional treatment arms. This is usually a two-stage trial and during the first stage of the trial, the inferior treatments are dropped based on pre-specified criteria or by conducting interim analysis. The winning treatment is then tested against a control [Reference Neal, Casella, Yang and Wu60].

Adaptive platform trial designs study multiple interventions for a common targeted condition. Based on a decision algorithm, treatments can be removed or added to the trial. The number and type of treatments can change throughout the study period. The benefit of this design is that the focus is on the condition rather than the treatment [61,Reference Lin and Bunn62].

Case Study: Cocaine use disorder (CUD) is a difficult condition to treat, and research has not found an effective treatment for the condition. Suchting et al. designed a drop-the-loser (DTL) Bayesian adaptive trial to determine the most effective dose of citalopram, a selective serotonin reuptake inhibitor that can treat CUD, to evaluate in a larger confirmatory trial (ClinicalTrials.gov number: NCT01535573) [Reference Suchting, Green and de Dios63]. Using Bayesian posterior probabilities, the researchers decided at the interim analysis which doses would be dropped from the trial (20 mg/day; 40 mg/day). In the implementation of the trial, the 20 mg/day dose was dropped after 50% of recruitment given that the 40 mg/day dose had a higher chance of success. Ultimately, the 40 mg/day dose provided “moderate-to-strong evidence” of positive effects at study conclusion [Reference Suchting, Green and de Dios63].

Adaptive Randomization

Randomization is the foundational basis for establishing efficacy of an investigational treatment in a randomized controlled trial (RCT). In theory, the randomness element seeks to establish study arms with similar patient characteristics between arms such that the only differing factor is the treatment assignment. Critical to the success of randomization is ensuring appropriate sample size allocation across arms as well as adjusting for any baseline covariates for which it is important that there is no imbalance.

Several randomization approaches in RCTs exist: The simplest approach is the coin flip; however, this method fails to control allocations across study arms and does not offer any direct covariate adjustment. To correct these issues, the idea of stratified block randomization emerged in the mid-twentieth century [Reference Fisher64,Reference Hill65]. This approach uses pre-specified randomization sequences within subgroups of patients and is relatively straightforward to implement. However, some limitations to stratified block randomization remain, namely requiring continuous stratification variables to be categorized, an inability to accommodate a large number of stratification covariates, and increased risk of selection bias near the end of a block when the allocation may be predicted [Reference Ciolino, Palac, Yang, Vaca and Belli66].

To address the limitations of stratified block randomization, covariate-adaptive designs were created to marginally balance multiple covariates at once rather than achieve balance within each stratum [Reference Taves67]. The minimization method is a covariate-adaptive method which was originally proposed by Taves [Reference Taves68] and Pocock and Simon [Reference Pocock and Simon69]. It implements an approach to minimize imbalance for each covariate of interest, thus allowing for more flexibility and options including weighting of covariates [Reference Endo, Nagatani, Hamada and Yoshimura70,Reference Nishi and Takaichi71]. However, minimization has been criticized for being more challenging to implement than stratified block randomization, requiring investment in software and personnel time from a study statistician.

Another strategy for randomization is the response-adaptive design, which modifies the allocation ratios based on the accumulating evidence of success in the trial arms [Reference Rosenberger, Sverdlov and Hu72,Reference Korn and Freidlin73]. This does not ensure balance of the baseline covariates, but attempts to allocate more participants to potentially effective study arms. Korn and Freidlin note adaptive randomization methods have the potential for increasing the total number of nonresponders relative to equal fixed allocation designs (i.e., 1:1 allocation). Further, response-adaptive randomization may introduce bias from temporal trends, lead to the unblinding of study arms and patients assigned to inferior arms due to variability in estimating effects based on smaller sample sizes [Reference Korn and Freidlin74,Reference Proschan and Evans75]. Recent work noted that response-adaptive randomization may be less susceptible to these issues in multiarm trials where the control maintains a fixed allocation, but caution should still be taken in implementing these methods [Reference Viele, Broglio, McGlothlin and Saville76].

Case Study: The Established Status Epilepticus Treatment Trial (ESETT; ClinicalTrials.gov number: NCT01960075) randomized participants to three intravenous anticonvulsive agents with an initial 1:1:1 allocation that changed based on a response-adaptive randomization design after 300 participants were observed [Reference Kapur, Elm and Chamberlain77]. ESETT also allowed for early termination for success or futility based on predefined stopping rules. After 300 participants, the randomization ratios were updated based on accumulating data to increase allocations to what appeared to be more promising study arms. However, the study was terminated for futility after 384 unique participants and it was determined that the probability of meeting full enrollment and detecting a significant effect was minimal [Reference Kapur, Elm and Chamberlain77].

Endpoint Selection

Endpoint selection is a critical part of designing a clinical trial. In adaptive clinical trials, in addition to the trial’s feasibility, its cost, and the intended goal of treatment, the modifications of trial procedures or statistical methodologies will impact the selection of endpoints and outcomes, especially for adaptive designs of confirmatory clinical trials [Reference Chow, Chang and Pong7,Reference Cerqueira, Jesus and Cotrim78].

Whether it is selected at the beginning of the trial or adapted during an on-going trial based on comparative interim results, the primary outcome selected for evaluation, (1) must address the trial objective and should be acknowledged as meaningful to clinicians, patients, and policymakers and (2) must be supported by enough scientific evidence to demonstrate its clinical relevance, i.e., it represents a current and reliable measure of clinical benefit in the target population [1,Reference McLeod, Norman, Litton, Saville, Webb and Snelling79].

Adaptation of the endpoint selection might be motivated by the uncertainty about the treatment effect sizes on multiple patient outcomes that are all considered acceptable as primary endpoints for a trial [1]. Regulatory and institutional guidelines elucidate directions in implementing the adaptation of endpoint selection, such as the adaptation rule should be pre-specified, and statistical hypothesis testing should account for the adaptive endpoint selection, as stated in the FDA’s 2019 Adaptive Trials Guidance [1]. Additionally, the FDA notes early discussion with the FDA review division is encouraged before considering adaptive endpoint selection methods [1]. On the other hand, the European Medicines Agency’s (EMA) Committee for Medicinal Products for Human Use (CHMP) in general warns against changing the primary endpoint in adaptive trials, stating the difficulty to justify changing the primary endpoints as they are chosen to describe a clinically relevant treatment effect and/or clinical benefit, which are defined during the study’s planning stages and cannot be changed based on interim results of the study [80].

Case Study: The use of adaptive endpoint selection is challenging with no straight-forward examples that the authors are aware of currently published in the literature. This may be due to the recommendation to consult with regulatory authorities, such as the FDA, prior to initiating such approaches or the strong recommendations against outcome adaptations from the EMA to avoid the appearance of bias or cherry picking by study investigators.

However, work is being done to move towards designs and contexts where adaptive endpoint selection may be advantageous. Filozof et al. discuss the scientific and logistical usage of an adaptive design-based approach to develop therapeutic strategies for patients with nonalcoholic steatohepatitis (NASH) [Reference Filozof, Chow and Dimick-Santos81]. The development of drugs for NASH has been substantially slow for a number of reasons, such as the heterogeneous nature of NASH with respect to the risk of progression to cirrhosis and the lack of a validated surrogate endpoint to clinical outcomes. Filozof et al. argue that, given the high unmet medical need and the lack of validated surrogate endpoints in NASH, the use of adaptive endpoint selection design methods appears reasonable as they provide the flexibility and efficiency for identifying potential signals of clinical benefit of the test treatment [Reference Filozof, Chow and Dimick-Santos81].

Recent work has been done to evaluate the statistical properties and potential benefits of these designs in certain contexts. Xu et al. proposed a design for rare diseases that maintains the family-wise type I error rate when selecting a primary endpoint based on an internal informational cohort when limited prior data exists and separate natural history cohort studies may be expensive or challenging to conduct [Reference Xu, Liu and Beckman82]. Roig et al. proposed an adaptive design that allows the modification of the primary endpoint based on blinded interim data while also recalculating the sample size accordingly [Reference Roig, Melis, Posch and Koenig83].

Adaptive Multiple Features

A clinical trial may also include multiple adaptive design features. As with any single trial design feature, it is important to ensure that the trial operating characteristics maintain the desired type I error rate and statistical power while enrolling a realistic sample size across a range of plausible future scenarios. Often, these are evaluated through extensive simulation studies [1].

Case Study: The vitamin C, thiamine, and steroids in sepsis (VICTAS; ClinicalTrials.gov number: NCT03509350) randomized clinical trial enrolled adults with sepsis-induced respiratory or cardiovascular dysfunction to examine if treatment with vitamin C, thiamine, and hydrocortisone result in more ventilator- and vasopressor-free days relative to placebo [Reference Sevransky, Rothman and Hager84,Reference Hager, Hooper and Bernard85]. The design incorporated both an adaptive sample size selection and Bayesian predictive probability monitoring for futility or efficacy [Reference Hager, Hooper and Bernard85]. If fewer than 400 participants were enrolled, interim monitoring would only consider stopping for efficacy, whereas sample sizes above 400 would consider stopping for either futility or efficacy [Reference Sevransky, Rothman and Hager84]. Ultimately, the trial terminated after 501 participants due to a change in funder’s priorities and no further enrollment occurred and the criteria for statistical significance not being met [Reference Sevransky, Rothman and Hager84]. This represents an unexpected conclusion to the study, but illustrates the potential for multiple adaptive elements to be included in one study.

Other Novel Trial Designs and Adaptations

Incorporating External/Supplemental Information

Information or data external to a trial may be useful to incorporate into analyses of a trial in order to increase the effective sample size. Some designs may attempt to incorporate this information to increase the sample size of treatment arms in the study, generally based on evaluating the exchangeability (i.e., equivalence) of the supplemental data with the current study data. Many approaches exist and include multisource exchangeability modeling [Reference Kaizer, Koopmeiners and Hobbs86,Reference Hobbs and Landin87], commensurate priors [Reference Hobbs, Sargent and Carlin88], power priors [Reference De Santis89,Reference Ibrahim and Chen90], and general Bayesian hierarchical models [Reference Neuenschwander, Capkun-Niggli, Branson and Spiegelhalter91]. It is also possible to include adaptive randomization methods when information is borrowed to maintain the overall proportion of trial data in a predetermined ratio (e.g., if historic control data are borrowed, more participants could be randomized to treatment arms) [Reference Kaizer, Hobbs and Koopmeiners92]. When considering methods that incorporate supplemental information, one must consider the sensitivity of methods to downweight nonexchangeable sources as to avoid potentially biasing the results towards historical data [Reference Kaizer, Koopmeiners and Hobbs86]. Additional challenges may be the specification of priors in study protocols given the uncertainty of future scenarios, but methods exist to account for this uncertainty in the design stage and to calibrate hyperparameters or tuning parameters [Reference Kaizer, Koopmeiners, Chen and Hobbs93].

Case Study: Putative Investigational Therapeutics in the Treatment of Patients With Known Ebola Infection (PREVAIL II; ClinicalTrials.gov number: NCT02363322) was a platform trial designed to sequentially evaluate potential therapies for the treatment of Ebola virus disease in the context of the West Africa Ebola outbreak in the mid-2010s with Bayesian posterior probabilities to facilitate frequent interim monitoring [94]. Given the sequential nature of the design, each “new” platform would always have at least one “old” platform of data that was just completed, but these data were not used in the PREVAIL II design due to concerns about changing disease outcomes over time. Kaizer, Hobbs, and Koopmeiners proposed a modified design that allowed for incorporating information from past platforms while also adapting the randomization ratio to enroll more participants to the new intervention if historic data were borrowed for the current control [Reference Kaizer, Hobbs and Koopmeiners92]. They demonstrated that information sharing could more quickly identify potentially effective therapies while randomizing more participants to the new arms with minimal bias based on the multisource exchangeability models used [Reference Kaizer, Hobbs and Koopmeiners92].

SMART Designs

The sequential multiple assignment randomized trial (SMART) is an adaptive research design for building optimal adaptive interventions. The general framework of a SMART design is as follows: every participant is randomized to an intervention arm initially, similar to the start of a classic, fixed RCT. Following this initial assignment, patients move through a series of stages with the option to either stay on or switch intervention arms, depending on their response to the intervention in the stage prior. If a participant switches intervention arms at the next stage, they will be re-randomized to a new treatment to maintain properties of causal inference associated with randomization within an RCT. The SMART framework mimics standard clinical practice in that, with time, patients will be assigned to more effective treatments. The SMART design therefore not only helps to build optimal dynamic treatment regimens, but also permits researchers to identify characteristics of patients for whom particular treatment regimens may be most effective [Reference Almirall, Nahum-Shani, Sherwood and Murphy95,Reference Murphy96]. SMART designs, while not technically adaptive since there are no prospectively defined modifications to the design, do represent an emerging class of designs that address pressing clinical questions relating to precision medicine and identifying the optimal treatment for any given patient.

Case Study: SMART designs are traditionally used in fields such as precision and behavioral medicine, mental health, or substance use where different combinations and sequences of treatments are administered [Reference Nahum-Shani, Ertefaie and Lu97]. The Establishing Moderators and Biosignatures of Antidepressant Response in Clinical Care (EMBARC; ClinicalTrials.gov number: NCT01407094) study is one example of a study that employed a SMART design [Reference Trivedi, McGrath and Fava98]. The primary aim of this two-stage, multisite, 8-week duration study was to discover candidate biomarkers as moderators of antidepressant treatment among depressed patients. A SMART design was selected due to the goal of investigating possible combinations of biomarkers and clinical characteristics as mediators and moderators to generate biosignatures for making personalized medication treatment prescriptions [Reference Petkova, Ogden and Tarpey99]. Meanwhile, data were collected 1 week after randomization to provide early indicators of patient response to treatment and thus to refine any subsequent treatment adaptations and predictions regarding the treatment response.

Master Protocol Designs

Master protocols are single, comprehensive trials developed to evaluate several concurrent sub-trials, involving assessment of multiple therapies, diseases, or subpopulations [Reference Park, Hsu, Siden, Thorlund and Mills100Reference Hirakawa, Asano, Sato and Teramukai104]. There are three general types: basket, umbrella, and platform trials. Basket trials evaluate single therapies across multiple diseases with common molecular alterations [Reference Park, Hsu, Siden, Thorlund and Mills100,Reference Kaizer, Koopmeiners, Kane, Roychoudhury, Hong and Hobbs105]. Sub-studies are often similarly designed single-arm trials, allowing pooled safety and efficacy data across subpopulations [Reference Lu, Li and Broglio101,Reference Hirakawa, Asano, Sato and Teramukai104]. Umbrella trials evaluate multiple therapies for a single disease stratified into subgroups [Reference Park, Hsu, Siden, Thorlund and Mills100,Reference Park, Siden and Zoratti102,Reference Janiaud, Serghiou and Ioannidis106]. Substudy designs can be single- or multiarm, though often include a control [Reference Hirakawa, Asano, Sato and Teramukai104,Reference Strzebonska and Waligora107]. Platform trials investigate multiple therapies in one or more diseases in an ongoing manner with arms added or dropped as new data and evidence appear [61,Reference Lu, Li and Broglio101,Reference Janiaud, Serghiou and Ioannidis106], often using Bayesian methods based on probabilities of treatment success or failure [61,Reference Saville and Berry108]. These designs may all include adaptive elements (e.g., arm dropping or group sequential methods) or be nonadaptive in their implementation.

Master protocols contain common guidelines for enrollment, measurements, data management, and statistical analysis; create a shared infrastructure across treatments; can reduce costs and increase efficiencies; and allow recruiting broader patient populations, helping bridge the translational gap toward clinical care [61,Reference Petkova, Ogden and Tarpey99,Reference Lu, Li and Broglio101,Reference Hirakawa, Asano, Sato and Teramukai104,Reference Redman and Allegra109]. Challenges include controlling false discovery rates [Reference Janiaud, Serghiou and Ioannidis106,Reference Saville and Berry108,Reference Meyer, Mesenbrink and Dunger-Baldauf110,Reference Zabor, Kane, Roychoudhury, Nie and Hobbs111], population drift [61,Reference Lu, Li and Broglio101,Reference Saville and Berry108], coordination among multiple partners [Reference Redman and Allegra109], intensive review and monitoring processes [Reference Sudhop, Brun, Riedel, Rosso, Broich and Senderovitz103], amendments [Reference Cecchini, Rubin and Blumenthal112], and informed consent [Reference Sudhop, Brun, Riedel, Rosso, Broich and Senderovitz103,Reference Cecchini, Rubin and Blumenthal112].

Case Studies: NCI-MATCH (Molecular Analysis for Therapy Choice; ClinicalTrials.gov number: NCT02465060) is a phase II basket trial for patients with advanced refractory solid tumors, lymphomas, or multiple myeloma [Reference Lih and Takebe113]. The Lung Cancer Master Protocol (Lung-MAP; ClinicalTrials.gov number: NCT02154490) is an umbrella trial for patients with advanced squamous nonsmall cell lung cancer, consisting of two-arm, seamless phase II–III randomized substudy designs [Reference Ferrarotto, Redman, Gandara, Herbst and Papadimitrakopoulou114]. I-SPY-2 (Investigation of serial Studies to Predict Your Therapeutic Response with Imaging and Molecular Analysis 2; ClinicalTrials.gov number: NCT01042379) is a multicenter phase II platform trial for patients with local metastatic breast cancer breast cancer, with promising drugs graduating to Phase III using Bayesian predictive probabilities [Reference Park, Liu and Yee115Reference Rugo, Olopade and DeMichele117].

Seamless Designs

Therapeutic development often occurs in distinct phases, with pauses between the “learning” and “confirming” phases [Reference Li, Lin and Lin3,Reference Stallard and Todd118Reference Richter, Friede and Rahnenführer120]. Seamless designs combine phases into a single protocol with no pause [Reference Li, Lin and Lin3,Reference Jennison and Turnbull121]. Operationally seamless designs combine phases into one trial, while inferentially seamless designs also combine data from distinct phases [Reference Li, Lin and Lin3,Reference Stallard and Todd118]. Seamless phase I/II designs simultaneously assess toxicity and efficacy [Reference Boonstra, Braun and Chase122], while seamless phase II/III designs combine the exploratory and confirmatory stages [Reference Bothwell, Avorn, Khan and Kesselheim123], where futile treatments are dropped and promising treatments are investigated further [Reference Li, Lin and Lin3,Reference Stallard and Todd118,Reference Richter, Friede and Rahnenführer120].

Seamless designs can accelerate the development timeline, reduce total sample size, and better estimate the risk of adverse events [Reference Li, Lin and Lin3,Reference Stallard and Todd118Reference Richter, Friede and Rahnenführer120,Reference Cuffe, Lawrence, Stone and Vandemeulebroecke124]. These designs require planning across stages, including prespecification of protocols, interim analysis schedule, statistical analysis plans, and data safety monitoring [Reference Li, Lin and Lin3,Reference Hobbs, Barata and Kanjanapan119], which can constrain flexibility and adaptability [Reference Cuffe, Lawrence, Stone and Vandemeulebroecke124]. Poor selection of short-term surrogate outcomes in early stages adversely affects later stages [Reference Cuffe, Lawrence, Stone and Vandemeulebroecke124,Reference Thall125]. Statistical methods that control type I error rates are required to reduce false positives [Reference Li, Lin and Lin3,Reference Richter, Friede and Rahnenführer120]. Seamless designs are generally appropriate for nonpivotal trials or for therapies with sufficient preliminary evidence [Reference Hobbs, Barata and Kanjanapan119,Reference Cuffe, Lawrence, Stone and Vandemeulebroecke124].

Case Studies: The Seamless Phase I/II Randomized Design for Immunotherapy Trials (SPIRIT) was proposed to investigate an immunotherapeutic agent that acts against the programed cell death ligand 1 checkpoint inhibitor in recurrent ovarian cancer [Reference Guo, Li and Yuan126]. The first stage determined the range of safe doses using a Bayesian optimal interval design, while admissible doses were explored in the second stage jointly modeling immune response and PFS [Reference Liu and Yuan127]. A seamless Phase IIb/III trial for chronic obstructive pulmonary disease identified treatments that exceeded a threshold for clinical relevance and were superior to control based on forced expiratory volume, then provided further evaluation based on percentage of days with poor control [Reference Barnes, Pocock and Magnussen128].

Phase I Studies

Clinical trials are defined by a series of phases depending on their overarching goal. A phase I clinical trial is a preliminary stage in this sequence characterized by its purpose to establish safety of an intervention and dosage in the case of a new drug. This includes identifying the maximum tolerated dose (MTD) which may be carried forward for future studies. Typically, these trials enroll a small number of subjects, thus utilizing designs that maximize the number of resources available is paramount.

Adaptive designs such as model-based or model-assisted approaches incorporate prior knowledge using Bayesian principles to allow for efficient use of data collected during the trial [Reference Yuan, Lee and Hilsenbeck129]. One such dose-finding design is the Bayesian Optimal Interval (BOIN) and its extensions [Reference Liu and Yuan127]. The BOIN design targets pre-specified toxicity level and with optimal boundaries. Based on the observed toxicities exhibited in a dosing cohort, the BOIN design recommends the optimal dose level for the next sequence of patients. Model-assisted designs generally provide an easier to interpret framework to clinicians as all operating characteristics can be pre-tabulated [Reference Yuan, Lee and Hilsenbeck129]. However, model-based approaches can require more input during the conduct of the trial as they need repeated model fitting and estimation. Care must be taken in the assumptions made to implement the design with clinical input and good statistical practice informing the final product, such as the type I error rate control. Design choices must be made a priori to adhere to the operating characteristics of the design.

Adaptive Phase I designs yield improved accuracy for identifying the MTD relative to standard approaches such as the 3 + 3 and allocate more patients to correct dosage level, improve safety outcomes, and increase the amount of information available for subsequent trials [Reference Yuan, Lee and Hilsenbeck129]. Recent advancements have also explored the use of “expansion cohorts” which take the identified dose from phase I seamlessly into an expanded phase II clinical trial [Reference Zabor, Kaizer, Garrett-Mayer and Hobbs130].

Case Study: A phase I trial examined the optimal dose of TG02 and temozolomide in treating high-grade gliomas that are traditionally highly resistant to treatment (ClinicalTrials.gov number: NCT02942264) [Reference Wu, Yuan and Cordova131]. A BOIN design was used with a target dose-limiting toxicity rate of 35% for the MTD. Ultimately, 38 participants were included and the combined dose for the two treatments was identified to use for designing a future phase II randomized trial for evaluating efficacy [Reference Wu, Yuan and Cordova131].

Statistical Considerations

Statistical Analysis for Adaptive Designs

Analysis for adaptive trials generally involves computing a suitable treatment effect estimator, inferences about the estimated effect, and the type I error rate control [Reference Wassmer and Brannath132]. The typical maximum likelihood estimation (MLE) for the treatment effect in the fixed sample size trial may be biased in adaptive designs because of the trial design adaptation (e.g., stopping rule) and selection of promising subgroups following interim analyses [Reference Robertson, Choodari-Oskooei, Dimairo, Flight, Pallmann and Jaki133]. Robertson et al. provided an overview of treatment effect estimators that improve the bias over MLE and their use in various types of adaptive designs [Reference Robertson, Choodari-Oskooei, Dimairo, Flight, Pallmann and Jaki133]. Statistical inference of the point estimate needs to have a correct coverage of confidence intervals (CIs) and account for design realizations such as multiple stages [Reference Jennison and Turnbull15,Reference Gao, Liu and Mehta134Reference Pallmann, Bedding and Choodari-Oskooei136]. Similar to CIs, hypothesis tests and p-values are combined across multiple stages of adaptive designs based on the conditional invariance assumption and it is advisable to pre-specify methods to generate p-values in trial protocols [Reference Jennison and Turnbull15,Reference Wassmer and Brannath132]. During hypothesis testing, the type I error rate needs to be adjusted when testing multiple hypotheses or when choosing a subpopulation or treatment, especially among confirmatory clinical trials [1,Reference Glimm, Maurer and Bretz137Reference Wang, James Hung and O'Neill139].

The above discussion so far focuses on the frequentist framework. However, Bayesian inference is commonly employed in adaptive designs since it can maintain the desired trial operating characteristics as demonstrated through simulation studies and has an adaptive nature [1,Reference Berry, Carlin, Lee and Muller140,Reference Jack Lee and Chu141]. For analysis implementation, software solutions in adaptive designs are summarized by Danielson et al. in Chapter 12 [Reference He, Pinheiro and Kuznetsova142]. In practice, the adaptive nature of a design should be accounted for in the analysis plan for the trial to avoid potential bias and unexpected coverage levels.

Trial Planning and Sample Size Calculations

A common theme throughout many of the designs is the importance of a priori planned adaptations and the identification of the target sample size. Adaptive trial designs need to account for the potential impact on the trial operating characteristics of making mid-trial changes [1]. This is most efficiently done via statistical simulation studies, but more traditional power calculations that do not account for the adaptations may serve as a convenient starting point to explore what changes (e.g., increased sample size) may be needed [1,Reference Li, Lin and Lin3]. As with any clinical trial, a range of possible scenarios should be explored with the resulting power and type I error rates summarized to determine if the properties are acceptable for a given context. These findings should be included in the study protocol with all assumptions and design elements clearly identified so that the findings are reproducible.

Secondary Endpoints

Secondary endpoints are often tested using either gatekeeping or hierarchical testing strategies in conventional trials as interim analysis or trial extension is not intended [Reference Dmitrienko, Tamhane, Wang and Chen143Reference Hung, Wang and O'Neill145]. These approaches are pre-specified in the study protocol, hence the overall type I error rate is strongly controlled. In adaptive trials, these techniques may not maintain a strong control of the overall type I error rate [Reference Pallmann, Bedding and Choodari-Oskooei136]. Hung et al. present a procedure for the statistical inference to test secondary endpoints after the primary endpoint reaches statistical significance, which is more complex in adaptive trials [Reference Hung, Wang and O'Neill145]. This complexity often arises from the fact that most secondary endpoints are correlated with the primary endpoint and hence more sophisticated approaches to adjust secondary endpoint analyses should be implemented [Reference Glimm, Maurer and Bretz137].

Appropriate caution should be taken along with the follow-up of proper regulatory guidelines in the analyses, adjustments, or interpretations of secondary endpoints. Further, if the adaptive trial requires the revision or modification of an endpoint, be it primary or secondary, the decision to revise or modify the endpoint should be independent of the data obtained from the trial and should not put the validity and integrity of the trial in question [Reference Evans146,Reference Wittes147].

Safety Considerations

Adaptive design elements may affect the availability of safety information for each study arm (e.g., terminating early for efficacy may not provide sufficient information to evaluate risk vs. benefit) or the adaptive element may place participants at excessive risk (e.g., early phase dose-escalation studies that permit rapid escalation) [1,Reference Pallmann, Bedding and Choodari-Oskooei136]. Where appropriate, a data safety monitoring board should be used to provide an external source of guidance and recommendation as to the safety of the overall study and to make recommendations to the trial investigators for potential modifications. As with any trial, the safety of participants is of utmost importance and adaptive designs should carefully consider trial modifications in the context of potential tradeoffs with participant safety.

Discussion

There is great potential for adaptive trial designs to improve the efficiency of clinical trials for future research problems with increasing complexity while potentially using fewer resources. Given the wide range of potential adaptations and emerging trial designs, it is clear that numerous designs for trials could be proposed based on the combinations of different adaptive elements. In general, while adaptive trials may provide increased flexibility, there is a need for sustained statistical support to ensure any prospectively planned modifications are made appropriately based on the accumulating data. Table 2 further highlights some of the general advantages and challenges in the implementation of adaptive trial designs. It is also worth noting that not all research questions warrant an adaptive design, and that nonadaptive designs may be more appropriate.

Table 2. General advantages and challenges of adaptive trials

Adaptive designs have the potential for furthering patient-centered research and recruiting more generalizable study populations. Modifications such as early termination or sample size re-estimation that aim to avoid wasted resources, either by stopping early due to substantial evidence or increasing the sample size to detect clinically meaningful differences, allow research results to more quickly disseminate to the communities which stand to benefit most. Designs with enrichment, adaptive randomization, or treatment arm selection attempt to address ethical concerns with randomizing individuals to arms or enrolling subgroups which may not benefit from the treatment, but special statistical considerations need to be made to maintain the trial operating characteristics. Currently, many adaptive designs examples occur in preclinical and clinical research (translational phases T1 and T2), but the adaptive methods may also be beneficial for clinical implementation and public health research (translational phases T3 and T4) to improve their efficiency, increase representation among diverse groups, and provide increased flexibility.

In summary, we provided a brief introduction to various adaptive design elements and emerging novel trial approaches with accompanying case studies to provide examples of the designs in practice. Additional novel designs and concepts are constantly emerging in the face of new challenges to address various research questions, and this article may serve as a starting point to introduce some of the design considerations to be used in practice. Finally, when designing any clinical trial, care should be taken to ensure the safety and integrity of the study for participants and the statistical trial operating characteristics.

Acknowledgments

AMK was supported in part by K01-HL151754 from the National Heart, Lung, and Blood Institute. HMB was supported in part by K01-AG075176 from the NIH National Institute on Aging. SCR, AFW, and MX are supported in part by UL1-TR002535 from the NIH National Center for Advancing Translational Sciences. RTS was supported in part by the Biostatistics, Epidemiology and Research Design (BERD) core of the C. Kenneth and Dianne Wright Center for Clinical and Translational Research UL1-TR002649 from the National Center for Advancing Translational Sciences. AGN is supported by P30-CA046934 from the National Cancer Institute through the Cancer Center Support Grant.

Disclosure

The authors have no conflicts of interest to declare.

References

US Food and Drug Administration. Adaptive designs for clinical trials of drugs and biologics guidance for industry. https://www.fda.gov/regulatory-information/search-fda-guidance-documents/adaptive-design-clinical-trials-drugs-and-biologics-guidance-industry.Google Scholar
Zabor, EC, Kaizer, AM, Hobbs, BP. Randomized controlled trials. Chest. 2020;158(1):S79S87.10.1016/j.chest.2020.03.013CrossRefGoogle ScholarPubMed
Li, Q, Lin, J, Lin, Y. Adaptive design implementation in confirmatory trials: Methods, practical considerations and case studies. Contemp Clin Trials. 2020;98:106096.CrossRefGoogle ScholarPubMed
Chow, S-C. Adaptive clinical trial design. Ann Rev Med. 2014;65(1):405415. doi: 10.1146/annurev-med-092012-112310.CrossRefGoogle ScholarPubMed
Chow, S-C, Chang, M. Adaptive Design Methods in Clinical Trials. Boca Raton, FL: Chapman and Hall/CRC; 2006.10.1201/9781584887775CrossRefGoogle Scholar
Chow, S-C, Chang, M. Adaptive design methods in clinical trials-a review. Orphanet J Rare Dis. 2008;3(1):113.CrossRefGoogle ScholarPubMed
Chow, S-C, Chang, M, Pong, A. Statistical consideration of adaptive methods in clinical development. J Biopharm Stat. 2005;15(4):575591.CrossRefGoogle ScholarPubMed
US Food and Drug Administration. Interacting with the FDA on complex innovative trial designs for drugs and biological products: Guidance for industry. https://www.fda.gov/regulatory-information/search-fda-guidance-documents/interacting-fda-complex-innovative-trial-designs-drugs-and-biological-products.Google Scholar
US Food and Drug Administration. Enrichment strategies for clinical trials to support determination of effectiveness of human drugs and biological products guidance for industry. https://www.fda.gov/regulatory-information/search-fda-guidance-documents/enrichment-strategies-clinical-trials-support-approval-human-drugs-and-biological-products.Google Scholar
Pocock, SJ. Group sequential methods in the design and analysis of clinical trials. Biometrika. 1977;64(2):191199.CrossRefGoogle Scholar
O'Brien, PC, Fleming, TR. A multiple testing procedure for clinical trials. Biometrics. 1979;35(3):549556.10.2307/2530245CrossRefGoogle ScholarPubMed
Demets, DL, Lan, KG. Interim analysis: The alpha spending function approach. Stat Med. 1994;13(13-14):13411352.10.1002/sim.4780131308CrossRefGoogle ScholarPubMed
Gordon Lan, K, DeMets, DL. Discrete sequential boundaries for clinical trials. Biometrika. 1983;70(3):659663.10.1093/biomet/70.3.659CrossRefGoogle Scholar
Wang, SK, Tsiatis, AA. Approximately optimal one-parameter boundaries for group sequential trials. Biometrics. 1987;43(1):193199.10.2307/2531959CrossRefGoogle ScholarPubMed
Jennison, C, Turnbull, BW. Group Sequential Methods With Applications to Clinical Trials. Boca Raton, FL: Chapman & Hall/CRC Press; 1999.CrossRefGoogle Scholar
Schüler, S, Kieser, M, Rauch, G. Choice of futility boundaries for group sequential designs with two endpoints. BMC Med Res Methodol. 2017;17(1):110.CrossRefGoogle Scholar
Lachin, JM. A review of methods for futility stopping based on conditional power. Stati Med. 2005;24(18):27472764.CrossRefGoogle ScholarPubMed
Pampallona, S, Tsiatis, AA. Group sequential designs for one-sided and two-sided hypothesis testing with provision for early stopping in favor of the null hypothesis. J Stat Plan Infer. 1994;42(1-2):1935.CrossRefGoogle Scholar
Chang, MN, Hwang, IK, Shin, WJ. Group sequential designs using both type I and type II error probability spending functions. Commun Stat Theory Methods. 1998;27(6):13231339.CrossRefGoogle Scholar
Pampallona, S, Tsiatis, AA, Kim, K. Interim monitoring of group sequential trials using spending functions for the type I and type II error probabilities. Drug Inform J. 2001;35(4):11131121.CrossRefGoogle Scholar
Rudser, KD, Emerson, SS. Implementing type I & type II error spending for two-sided group sequential designs. Contemp Clin Trials. 2008;29(3):351358.CrossRefGoogle ScholarPubMed
Spiegelhalter, DJ, Freedman, LS, Blackburn, PR. Monitoring clinical trials: Conditional or predictive power? Control Clin Trials. 1986;7(1):817.CrossRefGoogle ScholarPubMed
Saville, BR, Connor, JT, Ayers, GD, Alvarez, J. The utility of Bayesian predictive probabilities for interim monitoring of clinical trials. Clin Trials. 2014;11(4):485493.CrossRefGoogle ScholarPubMed
Cui, L, Hung, HJ, Wang, SJ. Modification of sample size in group sequential clinical trials. Biometrics. 1999;55(3):853857.CrossRefGoogle ScholarPubMed
Lehmacher, W, Wassmer, G. Adaptive sample size calculations in group sequential trials. Biometrics. 1999;55(4):12861290.CrossRefGoogle ScholarPubMed
Posch, M, Bauer, P. Interim analysis and sample size reassessment. Biometrics. 2000;56(4):11701176.CrossRefGoogle ScholarPubMed
Stallard, N, Todd, S. Sequential designs for phase III clinical trials incorporating treatment selection. Stat Med. 2003;22(5):689703.CrossRefGoogle ScholarPubMed
Stallard, N, Friede, T. A group-sequential design for clinical trials with treatment selection. Stat Med. 2008;27(29):62096227.CrossRefGoogle ScholarPubMed
Kelly, PJ, Stallard, N, Todd, S. An adaptive group sequential design for phase II/III clinical trials that select a single treatment from several. J Biopharm Stat. 2005;15(4):641658.CrossRefGoogle Scholar
Müller, HH, Schäfer, H. Adaptive group sequential designs for clinical trials: Combining the advantages of adaptive and of classical group sequential approaches. Biometrics. 2001;57(3):886891.CrossRefGoogle ScholarPubMed
Magnusson, BP, Turnbull, BW. Group sequential enrichment design incorporating subgroup selection. Stat Med. 2013;32(16):26952714.CrossRefGoogle ScholarPubMed
Perkins, GD, Quinn, T, Deakin, CD, et al. Pre-hospital assessment of the role of adrenaline: Measuring the effectiveness of drug administration in cardiac arrest (PARAMEDIC-2): Trial protocol. Resuscitation. 2016;108:7581.CrossRefGoogle ScholarPubMed
Perkins, GD, Ji, C, Deakin, CD, et al. A randomized trial of epinephrine in out-of-hospital cardiac arrest. N Engl J Med. 2018;379(8):711721.CrossRefGoogle ScholarPubMed
Ryan, EG, Stallard, N, Lall, R, Ji, C, Perkins, GD, Gates, S. Bayesian group sequential designs for phase III emergency medicine trials: A case study using the PARAMEDIC2 trial. Trials. 2020;21(1):113.CrossRefGoogle ScholarPubMed
Gould, AL. Sample size re-estimation: Recent developments and practical considerations. Stat Med. 2001;20(17-18):26252643.CrossRefGoogle ScholarPubMed
Brakenhoff, TB, Roes, KC, Nikolakopoulos, S. Bayesian sample size re-estimation using power priors. Stat Methods Med Res. 2019;28(6):16641675.CrossRefGoogle ScholarPubMed
Sobel, M, Turkoz, I. Bayesian blinded sample size re-estimation. Commun Stat Theory Methods. 2018;47(24):59165933.CrossRefGoogle Scholar
Liu, M, Li, Q, Lin, J, Lin, Y, Hoffman, E. Innovative trial designs and analyses for vaccine clinical development. Contemp Clin Trials. 2021;100:106225.CrossRefGoogle ScholarPubMed
Lake, S, Kammann, E, Klar, N, Betensky, R. Sample size re-estimation in cluster randomization trials. Stat Med. 2002;21(10):13371350.CrossRefGoogle ScholarPubMed
Gao, P, Ware, JH, Mehta, C. Sample size re-estimation for adaptive sequential design in clinical trials. J Biopharm Stat. 2008;18(6):11841196.CrossRefGoogle ScholarPubMed
Asendorf, T, Henderson, R, Schmidli, H, Friede, T. Sample size re-estimation for clinical trials with longitudinal negative binomial counts including time trends. Stat Med. 2019;38(9):15031528.CrossRefGoogle ScholarPubMed
Schneider, S, Schmidli, H, Friede, T. Blinded sample size re-estimation for recurrent event data with time trends. Stat Med. 2013;32(30):54485457.CrossRefGoogle ScholarPubMed
Pritchett, YL, Menon, S, Marchenko, O, et al. Sample size re-estimation designs in confirmatory clinical trials—current state, statistical considerations, and practical guidance. Stat Biopharm Res. 2015;7(4):309321.CrossRefGoogle Scholar
Friede, T, Kieser, M. Blinded sample size reassessment in non-inferiority and equivalence trials. Stat Med. 2003;22(6):9951007.CrossRefGoogle ScholarPubMed
Friede, T, Kieser, M. Blinded sample size re-estimation in superiority and noninferiority trials: Bias versus variance in variance estimation. Pharm Stat. 2013;12(3):141146.CrossRefGoogle ScholarPubMed
Golkowski, D, Friede, T, Kieser, M. Blinded sample size re-estimation in crossover bioequivalence trials. Pharm Stat. 2014;13(3):157162.CrossRefGoogle ScholarPubMed
Wang, P, Chow, S-C. Sample size re-estimation in clinical trials. Stat Med. 2021;40(27):61336149. doi: 10.1002/sim.9175.CrossRefGoogle ScholarPubMed
Proschan, MA. Two-stage sample size re-estimation based on a nuisance parameter: A review. J Biopharm Stat. 2005;15(4):559574.CrossRefGoogle ScholarPubMed
Asendorf, T, Henderson, R, Schmidli, H, Friede, T. Modelling and sample size reestimation for longitudinal count data with incomplete follow up. Stat Methods Med Res. 2019;28(1):117133.CrossRefGoogle ScholarPubMed
Friede, T, Schmidli, H. Blinded sample size reestimation with count data: Methods and applications in multiple sclerosis. Stat Med. 2010;29(10):11451156.CrossRefGoogle Scholar
Dawson, SN, Chiu, Y-D, Klein, AA, Earwaker, M, Villar, SS. Effect of high-flow nasal therapy on patient-centred outcomes in patients at high risk of postoperative pulmonary complications after cardiac surgery: A statistical analysis plan for NOTACS, a multicentre adaptive randomised controlled trial. Trials. 2022;23(1):18.CrossRefGoogle Scholar
Thall, PF. Adaptive enrichment designs in clinical trials. Annu Rev Stat Appl. 2021;8(1):393411.CrossRefGoogle ScholarPubMed
Simon, N, Simon, R. Adaptive enrichment designs for clinical trials. Biostatistics. 2013;14(4):613625.CrossRefGoogle ScholarPubMed
Burnett, T, Jennison, C. Adaptive enrichment trials: What are the benefits? Stat Med. 2021;40(3):690711.CrossRefGoogle ScholarPubMed
Lai, TL, Lavori, PW, Tsang, KW. Adaptive enrichment designs for confirmatory trials. Stat Med. 2019;38(4):613624.CrossRefGoogle ScholarPubMed
Jones, RL, Ravi, V, Brohl, AS, et al. Efficacy and safety of TRC105 plus pazopanib vs pazopanib alone for treatment of patients with advanced angiosarcoma: A randomized clinical trial. JAMA Oncol. 2022;8(5):740747.CrossRefGoogle ScholarPubMed
Mehta, C, Liu, L, Theuer, C. An adaptive population enrichment phase III trial of TRC105 and pazopanib versus pazopanib alone in patients with advanced angiosarcoma (TAPPAS trial). Ann Oncol. 2019;30(1):103108.CrossRefGoogle ScholarPubMed
Iasonos, A, O'Quigley, J. Adaptive dose-finding studies: A review of model-guided phase I clinical trials. J Clin Oncol. 2014;32(23):25052511.CrossRefGoogle ScholarPubMed
Maca, J, Bhattacharya, S, Dragalin, V, Gallo, P, Krams, M. Adaptive seamless phase II/III designs—background, operational aspects, and examples. Drug Inform J. 2006;40(4):463473.CrossRefGoogle Scholar
Neal, D, Casella, G, Yang, MC, Wu, SS. Interval estimation in two-stage, drop-the-losers clinical trials with flexible treatment selection. Stat Med. 2011;30(23):28042814.CrossRefGoogle ScholarPubMed
Nature Reviews Drug Discovery. Adaptive platform trials coalition. Adaptive platform trials: Definition, design, conduct and reporting considerations. Nat Rev Drug Discov. 2019;18(10):797807.CrossRefGoogle Scholar
Lin, J, Bunn, V. Comparison of multi-arm multi-stage design and adaptive randomization in platform clinical trials. Contemp Clin Trials. 2017;54:4859.CrossRefGoogle ScholarPubMed
Suchting, R, Green, CE, de Dios, C, et al. Citalopram for treatment of cocaine use disorder: A bayesian drop-the-loser randomized clinical trial. Drug Alcohol Depen. 2021;228:109054.CrossRefGoogle ScholarPubMed
Fisher, RA. The Design of Experiments. 2nd ed. London, UK: Macmillan; 1937.Google Scholar
Hill, AB. The clinical trial. N Engl J Med. 1952;247(4):113119.CrossRefGoogle ScholarPubMed
Ciolino, JD, Palac, HL, Yang, A, Vaca, M, Belli, HM. Ideal vs real: A systematic review on handling covariates in randomized controlled trials. BMC Med Res Methodol. 2019;19(1):111.CrossRefGoogle Scholar
Taves, DR. The use of minimization in clinical trials. Contemp Clin Trials. 2010;31(2):180184.CrossRefGoogle ScholarPubMed
Taves, DR. Minimization: A new method of assigning patients to treatment and control groups. Clin Pharmacol Ther. 1974;15(5):443453.CrossRefGoogle ScholarPubMed
Pocock, SJ, Simon, R. Sequential treatment assignment with balancing for prognostic factors in the controlled clinical trial. Biometrics. 1975;31(1):103115.CrossRefGoogle ScholarPubMed
Endo, A, Nagatani, F, Hamada, C, Yoshimura, I. Minimization method for balancing continuous prognostic variables between treatment and control groups using Kullback-Leibler divergence. Contemp Clin Trials. 2006;27(5):420431.CrossRefGoogle ScholarPubMed
Nishi, T, Takaichi, A. An extended minimization method to assure similar means of continuous prognostic variables between treatment groups. Jap J Biom. 2004;24(2):4355.Google Scholar
Rosenberger, WF, Sverdlov, O, Hu, F. Adaptive randomization for clinical trials. J Biopharm Stat. 2012;22(4):719736.CrossRefGoogle ScholarPubMed
Korn, EL, Freidlin, B. Outcome-adaptive randomization: Is it useful? J Clin Oncol. 2011;29(6):771776.CrossRefGoogle ScholarPubMed
Korn, EL, Freidlin, B. Time trends with response-adaptive randomization: The inevitability of inefficiency. Clin Trials. 2022;19(2):158161.CrossRefGoogle ScholarPubMed
Proschan, M, Evans, S. Resist the temptation of response-adaptive randomization. Clin Infect Dis. 2020;71(11):30023004.CrossRefGoogle ScholarPubMed
Viele, K, Broglio, K, McGlothlin, A, Saville, BR. Comparison of methods for control allocation in multiple arm studies using response adaptive randomization. Clin Trials. 2020;17(1):5260.CrossRefGoogle ScholarPubMed
Kapur, J, Elm, J, Chamberlain, JM, et al. Randomized trial of three anticonvulsant medications for status epilepticus. N Engl J Med. 2019;381(22):21032113.CrossRefGoogle ScholarPubMed
Cerqueira, FP, Jesus, AMC, Cotrim, MD. Adaptive design: A review of the technical, statistical, and regulatory aspects of implementation in a clinical trial. Ther Innov Regul Sci. 2020;54(1):246258.CrossRefGoogle ScholarPubMed
McLeod, C, Norman, R, Litton, E, Saville, BR, Webb, S, Snelling, TL. Choosing primary endpoints for clinical trials of health care interventions. Contemp Clin Trials Commun. 2019;16:100486.CrossRefGoogle ScholarPubMed
European Medicines Agency. Methodological issues in confirmatory clinical trials planned with an adaptive design - scientific guideline. https://www.ema.europa.eu/en/methodological-issues-confirmatory-clinical-trials-planned-adaptive-design-scientific-guideline.Google Scholar
Filozof, C, Chow, SC, Dimick-Santos, L, et al. Clinical endpoints and adaptive clinical trials in precirrhotic nonalcoholic steatohepatitis: Facilitating development approaches for an emerging epidemic. Hepatol Commun. 2017;1(7):577585.CrossRefGoogle ScholarPubMed
Xu, H, Liu, Y, Beckman, RA. Adaptive endpoints selection with application in rare disease. Stat Biopharm Res. 2023;just-accepted:118.CrossRefGoogle Scholar
Roig, MB, Melis, GG, Posch, M, Koenig, F. Adaptive clinical trial designs with blinded selection of binary composite endpoints and sample size reassessment. Biostatistics. 2022. doi: 10.1093/biostatistics/kxac040.CrossRefGoogle ScholarPubMed
Sevransky, JE, Rothman, RE, Hager, DN, et al. Effect of vitamin C, thiamine, and hydrocortisone on ventilator-and vasopressor-free days in patients with sepsis: The VICTAS randomized clinical trial. Jama. 2021;325(8):742750.CrossRefGoogle ScholarPubMed
Hager, DN, Hooper, MH, Bernard, GR, et al. The vitamin C, thiamine and steroids in sepsis (VICTAS) protocol: A prospective, multi-center, double-blind, adaptive sample size, randomized, placebo-controlled, clinical trial. Trials. 2019;20(1):116.CrossRefGoogle ScholarPubMed
Kaizer, AM, Koopmeiners, JS, Hobbs, BP. Bayesian hierarchical modeling based on multisource exchangeability. Biostatistics. 2018;19(2):169184.CrossRefGoogle ScholarPubMed
Hobbs, BP, Landin, R. Bayesian basket trial design with exchangeability monitoring. Stat Med. 2018;37(25):35573572.CrossRefGoogle ScholarPubMed
Hobbs, BP, Sargent, DJ, Carlin, BP. Commensurate priors for incorporating historical information in clinical trials using general and generalized linear models. Bayesian Anal. 2012;7(3):639.CrossRefGoogle ScholarPubMed
De Santis, F. Power priors and their use in clinical trials. Am Stat. 2006;60(2):122129.CrossRefGoogle Scholar
Ibrahim, JG, Chen, M-H. Power prior distributions for regression models. Stat Sci. 2000;15(1): 4660.Google Scholar
Neuenschwander, B, Capkun-Niggli, G, Branson, M, Spiegelhalter, DJ. Summarizing historical information on controls in clinical trials. Clin Trials. 2010;7(1):518.CrossRefGoogle ScholarPubMed
Kaizer, AM, Hobbs, BP, Koopmeiners, JS. A multi-source adaptive platform design for testing sequential combinatorial therapeutic strategies. Biometrics. 2018;74(3):10821094.CrossRefGoogle ScholarPubMed
Kaizer, AM, Koopmeiners, JS, Chen, N, Hobbs, BP. Statistical design considerations for trials that study multiple indications. Stat Methods Med Res. 2021;30(3):785798.CrossRefGoogle ScholarPubMed
Group TPIW, Team M-NPIS. A randomized, controlled trial of ZMapp for Ebola virus infection. N Engl J Med. 2016;375(15):14481456.CrossRefGoogle Scholar
Almirall, D, Nahum-Shani, I, Sherwood, NE, Murphy, SA. Introduction to SMART designs for the development of adaptive interventions: With application to weight loss research. Transl Behav Med. 2014;4(3):260274.CrossRefGoogle ScholarPubMed
Murphy, SA. An experimental design for the development of adaptive treatment strategies. Stat Med. 2005;24(10):14551481.CrossRefGoogle ScholarPubMed
Nahum-Shani, I, Ertefaie, A, Lu, X, et al. A SMART data analysis method for constructing adaptive treatment strategies for substance use disorders. Addiction. 2017;112(5):901909.CrossRefGoogle Scholar
Trivedi, MH, McGrath, PJ, Fava, M, et al. Establishing moderators and biosignatures of antidepressant response in clinical care (EMBARC): Rationale and design. J Psychiatr Res. 2016;78:1123.CrossRefGoogle ScholarPubMed
Petkova, E, Ogden, RT, Tarpey, T, et al. Statistical analysis plan for stage 1 EMBARC (Establishing moderators and biosignatures of antidepressant response for clinical care) study. Contemp Clin Trials Commun. 2017;6:2230.CrossRefGoogle Scholar
Park, JJ, Hsu, G, Siden, EG, Thorlund, K, Mills, EJ. An overview of precision oncology basket and umbrella trials for clinicians. CA: Cancer J Clin. 2020;70(2):125137.Google ScholarPubMed
Lu, CC, Li, XN, Broglio, K, et al. Practical considerations and recommendations for master protocol framework: Basket, umbrella and platform trials. Therap Innov Regul Sci. 2021;55(6):11451154.CrossRefGoogle ScholarPubMed
Park, JJ, Siden, E, Zoratti, MJ, et al. Systematic review of basket trials, umbrella trials, and platform trials: A landscape analysis of master protocols. Trials. 2019;20(1):110.CrossRefGoogle ScholarPubMed
Sudhop, T, Brun, NC, Riedel, C, Rosso, A, Broich, K, Senderovitz, T. Master protocols in clinical trials: A universal swiss army knife? Lancet Oncol. 2019;20(6):e336e342.CrossRefGoogle ScholarPubMed
Hirakawa, A, Asano, J, Sato, H, Teramukai, S. Master protocol trials in oncology: Review and new trial designs. Contemp Clin Trials Commun. 2018;12:18.CrossRefGoogle ScholarPubMed
Kaizer, AM, Koopmeiners, JS, Kane, MJ, Roychoudhury, S, Hong, DS, Hobbs, BP. Basket designs: Statistical considerations for oncology trials. JCO Precis Oncol. 2019;3(3):19.Google ScholarPubMed
Janiaud, P, Serghiou, S, Ioannidis, JP. New clinical trial designs in the era of precision medicine: An overview of definitions, strengths, weaknesses, and current use in oncology. Cancer Treat Rev. 2019;73:2030.CrossRefGoogle ScholarPubMed
Strzebonska, K, Waligora, M. Umbrella and basket trials in oncology: Ethical challenges. BMC Med Ethics. 2019;20(1):110.CrossRefGoogle ScholarPubMed
Saville, BR, Berry, SM. Efficiencies of platform clinical trials: A vision of the future. Clin Trials. 2016;13(3):358366.CrossRefGoogle ScholarPubMed
Redman, MW, Allegra, CJ. The master protocol concept. Elsevier. 2015;42(5):724730.Google ScholarPubMed
Meyer, EL, Mesenbrink, P, Dunger-Baldauf, C, et al. The evolution of master protocol clinical trial designs: A systematic literature review. Clin Therap. 2020;42(7):13301360.CrossRefGoogle ScholarPubMed
Zabor, EC, Kane, MJ, Roychoudhury, S, Nie, L, Hobbs, BP. Bayesian basket trial design with false-discovery rate control. Clin Trials. 2022;19(3):17407745211073624–306.CrossRefGoogle ScholarPubMed
Cecchini, M, Rubin, EH, Blumenthal, GM, et al. Challenges with novel clinical trial designs: Master protocolsChallenges with novel clinical trials. Clin Cancer Res. 2019;25(7):20492057.CrossRefGoogle Scholar
Lih, C-J, Takebe, N. Considerations of developing an NGS assay for clinical applications in precision oncology: The NCI-mATCH NGS assay experience. Curr Probl Cancer. 2017;41(3):201211.CrossRefGoogle ScholarPubMed
Ferrarotto, R, Redman, MW, Gandara, DR, Herbst, RS, Papadimitrakopoulou, VA. Lung-MAP—framework, overview, and design principles. Chin Clin Oncol. 2015;4(3):36.Google ScholarPubMed
Park, JW, Liu, MC, Yee, D, et al. Adaptive randomization of neratinib in early breast cancer. N Engl J Med. 2016;375(1):1122.CrossRefGoogle ScholarPubMed
Barker, A, Sigman, C, Kelloff, G, Hylton, N, Berry, D, Esserman, L. I-SPY 2: An adaptive breast cancer trial design in the setting of neoadjuvant chemotherapy. Clin Pharmacol Therap. 2009;86(1):97100.CrossRefGoogle ScholarPubMed
Rugo, HS, Olopade, OI, DeMichele, A, et al. Adaptive randomization of veliparib-carboplatin treatment in breast cancer. N Engl J Med. 2016;375(1):2334.CrossRefGoogle ScholarPubMed
Stallard, N, Todd, S. Seamless phase II/III designs. Stat Methods Med Res. 2011;20(6):623634.CrossRefGoogle ScholarPubMed
Hobbs, BP, Barata, PC, Kanjanapan, Y, et al. Seamless designs: Current practice and considerations for early-phase drug development in oncology. J Nat Cancer Inst. 2019;111(2):118128.CrossRefGoogle ScholarPubMed
Richter, J, Friede, T, Rahnenführer, J. Improving adaptive seamless designs through Bayesian optimization. Biom J. 2022;64(5):948963.CrossRefGoogle ScholarPubMed
Jennison, C, Turnbull, BW. Adaptive seamless designs: Selection and prospective testing of hypotheses. J Biopharm Stat. 2007;17(6):11351161.CrossRefGoogle ScholarPubMed
Boonstra, PS, Braun, TM, Chase, EC. A modular framework for early-phase seamless oncology trials. Clin Trials. 2021;18(3):303313.CrossRefGoogle ScholarPubMed
Bothwell, LE, Avorn, J, Khan, NF, Kesselheim, AS. Adaptive design clinical trials: A review of the literature and ClinicalTrials. gov. BMJ Open. 2018;8(2):e018320.CrossRefGoogle ScholarPubMed
Cuffe, RL, Lawrence, D, Stone, A, Vandemeulebroecke, M. When is a seamless study desirable? Case studies from different pharmaceutical sponsors. Pharm Stat. 2014;13(4):229237.CrossRefGoogle ScholarPubMed
Thall, PF. A review of phase 2-3 clinical trial designs. Lifetime Data Anal. 2008;14(1):3753.CrossRefGoogle ScholarPubMed
Guo, B, Li, D, Yuan, Y. SPIRIT: A seamless phase I/II randomized design for immunotherapy trials. Pharm Stat. 2018;17(5):527540.Google ScholarPubMed
Liu, S, Yuan, Y. Bayesian optimal interval designs for phase I clinical trials. J R Stat Soc Ser C Appl Stat. 2015;64(3):507523.CrossRefGoogle Scholar
Barnes, PJ, Pocock, SJ, Magnussen, H, et al. Integrating indacaterol dose selection in a clinical study in COPD using an adaptive seamless design. Pulm Pharmacol Ther. 2010;23(3):165171.CrossRefGoogle Scholar
Yuan, Y, Lee, JJ, Hilsenbeck, SG. Model-assisted designs for early-phase clinical trials: Simplicity meets superiority. JCO Prec Oncol. 2019;3(3):112.Google ScholarPubMed
Zabor, EC, Kaizer, AM, Garrett-Mayer, E, Hobbs, BP. Optimal sequential predictive probability designs for early-phase oncology expansion cohorts. JCO Prec Oncol. 2022;6(6):e2100390.CrossRefGoogle ScholarPubMed
Wu, J, Yuan, Y, Cordova, C, et al. Phase I trial of TG02 plus dose-dense or metronomic temozolomide for recurrent anaplastic astrocytoma and glioblastoma in adults. Am Soc Clin Oncol. 2019;37(15):2031.CrossRefGoogle Scholar
Wassmer, G, Brannath, W. Group Sequential and Confirmatory Adaptive Designs in Clinical Trials, vol. 301. Cham, Switzerland: Springer International Publishing Switzerland; 2016.CrossRefGoogle Scholar
Robertson, DS, Choodari-Oskooei, B, Dimairo, M, Flight, L, Pallmann, P, Jaki, T. Point estimation for adaptive trial designs I: A methodological review. Stat Med. 2022;42:122145.CrossRefGoogle Scholar
Gao, P, Liu, L, Mehta, C. Exact inference for adaptive group sequential designs. Stat Med. 2013;32(23):39914005.CrossRefGoogle ScholarPubMed
Kimani, PK, Todd, S, Stallard, N. A comparison of methods for constructing confidence intervals after phase II/III clinical trials. Biom J. 2014;56(1):107128.CrossRefGoogle ScholarPubMed
Pallmann, P, Bedding, AW, Choodari-Oskooei, B, et al. Adaptive designs in clinical trials: Why use them, and how to run and report them. BMC Med. 2018;16(1):115.CrossRefGoogle Scholar
Glimm, E, Maurer, W, Bretz, F. Hierarchical testing of multiple endpoints in group-sequential trials. Stat Med. 2010;29(2):219228.CrossRefGoogle ScholarPubMed
Bratton, DJ, Parmar, MK, Phillips, PP, Choodari-Oskooei, B. Type I error rates of multi-arm multi-stage clinical trials: Strong control and impact of intermediate outcomes. Trials. 2016;17(1):18.CrossRefGoogle ScholarPubMed
Wang, S-J, James Hung, H, O'Neill, R. Regulatory perspectives on multiplicity in adaptive design clinical trials throughout a drug development program. J Biopharm Stat. 2011;21(4):846859.CrossRefGoogle ScholarPubMed
Berry, SM, Carlin, BP, Lee, JJ, Muller, P. Bayesian Adaptive Methods for Clinical Trials. Boca Raton, FL: Chapman & Hall/CRC press; 2010.CrossRefGoogle Scholar
Jack Lee, J, Chu, CT. Bayesian clinical trials in action. Stat Med. 2012;31(25):29552972.CrossRefGoogle Scholar
He, W, Pinheiro, J, Kuznetsova, OM. Practical Considerations for Adaptive Trial Design and Implementation. New York, NY: Springer Science+Business Media New York; 2014.CrossRefGoogle Scholar
Dmitrienko, A, Tamhane, AC, Wang, X, Chen, X. Stepwise gatekeeping procedures in clinical trial applications. J Math Methods Biosci. 2006;48(6):984991.Google ScholarPubMed
Tamhane, AC, Mehta, CR, Liu, L. Testing a primary and a secondary endpoint in a group sequential design. Biometrics. 2010;66(4):11741184.CrossRefGoogle Scholar
Hung, HJ, Wang, S-J, O'Neill, R. Statistical considerations for testing multiple endpoints in group sequential or adaptive clinical trials. J Biopharm Stat. 2007;17(6):12011210.CrossRefGoogle ScholarPubMed
Evans, S. When and how can endpoints be changed after initiation of a randomized clinical trial. Plos Clin Trials. 2007;2(4):e18.CrossRefGoogle ScholarPubMed
Wittes, J. On changing a long-term clinical trial midstream. Stat Med. 2002;21(19):27892795.CrossRefGoogle ScholarPubMed
Figure 0

Table 1. Brief summary of FDA guidance document adaptive design elements

Figure 1

Table 2. General advantages and challenges of adaptive trials