1
|
Pulmonary Hypertension: Intensification and Personalization of Combination Rx (PHoenix): A phase IV randomized trial for the evaluation of dose-response and clinical efficacy of riociguat and selexipag using implanted technologies. Pulm Circ 2024; 14:e12337. [PMID: 38500737 PMCID: PMC10945040 DOI: 10.1002/pul2.12337] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/25/2023] [Revised: 12/19/2023] [Accepted: 01/02/2024] [Indexed: 03/20/2024] Open
Abstract
Approved therapies for the treatment of patients with pulmonary arterial hypertension (PAH) mediate pulmonary vascular vasodilatation by targeting distinct biological pathways. International guidelines recommend that patients with an inadequate response to dual therapy with a phosphodiesterase type-5 inhibitor (PDE5i) and endothelin receptor antagonist (ERA), are recommended to either intensify oral therapy by adding a selective prostacyclin receptor (IP) agonist (selexipag), or switching from PDE5i to a soluble guanylate-cyclase stimulator (sGCS; riociguat). The clinical equipoise between these therapeutic choices provides the opportunity for evaluation of individualized therapeutic effects. Traditionally, invasive/hospital-based investigations are required to comprehensively assess disease severity and demonstrate treatment benefits. Regulatory-approved, minimally invasive monitors enable equivalent measurements to be obtained while patients are at home. In this 2 × 2 randomized crossover trial, patients with PAH established on guideline-recommended dual therapy and implanted with CardioMEMS™ (a wireless pulmonary artery sensor) and ConfirmRx™ (an insertable cardiac rhythm monitor), will receive ERA + sGCS, or PDEi + ERA + IP agonist. The study will evaluate clinical efficacy via established clinical investigations and remote monitoring technologies, with remote data relayed through regulatory-approved online clinical portals. The primary aim will be the change in right ventricular systolic volume measured by magnetic resonance imaging (MRI) from baseline to maximal tolerated dose with each therapy. Using data from MRI and other outcomes, including hemodynamics, physical activity, physiological measurements, quality of life, and side effect reporting, we will determine whether remote technology facilitates early evaluation of clinical efficacy, and investigate intra-patient efficacy of the two treatment approaches.
Collapse
|
2
|
Bayesian sample size determination in basket trials borrowing information between subsets. Biostatistics 2023; 24:1000-1016. [PMID: 35993875 DOI: 10.1093/biostatistics/kxac033] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/27/2021] [Revised: 07/22/2022] [Accepted: 07/29/2022] [Indexed: 12/31/2022] Open
Abstract
Basket trials are increasingly used for the simultaneous evaluation of a new treatment in various patient subgroups under one overarching protocol. We propose a Bayesian approach to sample size determination in basket trials that permit borrowing of information between commensurate subsets. Specifically, we consider a randomized basket trial design where patients are randomly assigned to the new treatment or control within each trial subset ("subtrial" for short). Closed-form sample size formulae are derived to ensure that each subtrial has a specified chance of correctly deciding whether the new treatment is superior to or not better than the control by some clinically relevant difference. Given prespecified levels of pairwise (in)commensurability, the subtrial sample sizes are solved simultaneously. The proposed Bayesian approach resembles the frequentist formulation of the problem in yielding comparable sample sizes for circumstances of no borrowing. When borrowing is enabled between commensurate subtrials, a considerably smaller trial sample size is required compared to the widely implemented approach of no borrowing. We illustrate the use of our sample size formulae with two examples based on real basket trials. A comprehensive simulation study further shows that the proposed methodology can maintain the true positive and false positive rates at desired levels.
Collapse
|
3
|
The potential of innovative trial design for efficiently evaluating repurposed drugs. Nat Rev Drug Discov 2023; 22:681-682. [PMID: 37537382 DOI: 10.1038/d41573-023-00129-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/05/2023]
|
4
|
Online error rate control for platform trials. Stat Med 2023; 42:2475-2495. [PMID: 37005003 PMCID: PMC7614610 DOI: 10.1002/sim.9733] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/03/2021] [Revised: 01/20/2023] [Accepted: 03/18/2023] [Indexed: 04/04/2023]
Abstract
Platform trials evaluate multiple experimental treatments under a single master protocol, where new treatment arms are added to the trial over time. Given the multiple treatment comparisons, there is the potential for inflation of the overall type I error rate, which is complicated by the fact that the hypotheses are tested at different times and are not necessarily pre-specified. Online error rate control methodology provides a possible solution to the problem of multiplicity for platform trials where a relatively large number of hypotheses are expected to be tested over time. In the online multiple hypothesis testing framework, hypotheses are tested one-by-one over time, where at each time-step an analyst decides whether to reject the current null hypothesis without knowledge of future tests but based solely on past decisions. Methodology has recently been developed for online control of the false discovery rate as well as the familywise error rate (FWER). In this article, we describe how to apply online error rate control to the platform trial setting, present extensive simulation results, and give some recommendations for the use of this new methodology in practice. We show that the algorithms for online error rate control can have a substantially lower FWER than uncorrected testing, while still achieving noticeable gains in power when compared with the use of a Bonferroni correction. We also illustrate how online error rate control would have impacted a currently ongoing platform trial.
Collapse
|
5
|
Current practices in studies applying the target trial emulation framework: a protocol for a systematic review. BMJ Open 2023; 13:e070963. [PMID: 37369393 PMCID: PMC10410979 DOI: 10.1136/bmjopen-2022-070963] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 12/09/2022] [Accepted: 06/06/2023] [Indexed: 06/29/2023] Open
Abstract
INTRODUCTION Observational studies represent an alternative to estimate real-world causal effects in the absence of available randomised controlled trials (RCTs). Target trial emulation is a framework for the application of RCT design principles to emulate a hypothetical open-label RCT (the hypothetical target trial) using existing observational data as the primary data source as opposed to the prospective recruitment and measurement of randomised units. The aim of this systematic review is to investigate the practices of studies applying the target trial emulation framework to evaluate the effectiveness of interventions. METHODS AND ANALYSIS We will systematically search in Medline (via Ovid), Embase (via Ovid, entries from medRxiv are included), PsycINFO (via Ovid), SCOPUS, Web of Science, Cochrane Library, the ISRCTN registry and ClinicalTrials.gov for all study reports and protocols which used the trial emulation framework (without time restriction). We will extract information concerning study design, data source, analysis, results, interpretation and dissemination. Two reviewers will perform study selection, data extraction and quality assessment. Disagreements between reviewers will be resolved by a third reviewer. A narrative approach will be used to synthesise and report qualitative and quantitative data. Reporting of the review will be informed by Preferred Reporting Items for Systematic Review and Meta-Analysis guidance (PRISMA). ETHICS AND DISSEMINATION Ethical approval is not required as it is a protocol for a systematic review. Findings will be disseminated through peer-reviewed publications and conference presentations.
Collapse
|
6
|
Bayesian borrowing for basket trials with longitudinal outcomes. Stat Med 2023. [PMID: 37120858 DOI: 10.1002/sim.9751] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/07/2022] [Revised: 02/28/2023] [Accepted: 04/16/2023] [Indexed: 05/02/2023]
Abstract
Basket trials are a novel clinical trial design in which a single intervention is investigated in multiple patient subgroups, or "baskets." They offer the opportunity to share information between subgroups, potentially increasing power to detect treatment effects. Basket trials offer several advantages over running a series of separate trials, including reduced sample sizes, increased efficiency, and reduced costs. Primarily, basket trials have been undertaken in Phase II oncology settings, but could be a promising design in other areas where a shared underlying biological mechanism drives different diseases. One such area is chronic aging-related diseases. However, trials in this area frequently have longitudinal outcomes, and therefore suitable methods are needed to share information in this setting. In this paper, we extend three Bayesian borrowing methods for a basket design with continuous longitudinal endpoints. We demonstrate our methods on a real-world dataset and in a simulation study where the aim is to detect positive basketwise treatment effects. Methods are compared with standalone analysis of each basket without borrowing. Our results confirm that methods that share information can improve power to detect positive treatment effects and increase precision over independent analysis in many scenarios. In highly heterogeneous scenarios, there is a trade-off between increased power and increased risk of type I errors. Our proposed methods for basket trials with continuous longitudinal outcomes aim to facilitate their applicability in the area of aging related diseases. Choice of method should be made based on trial priorities and the expected basketwise distribution of treatment effects.
Collapse
|
7
|
Discussion on "Adaptive enrichment designs with a continuous biomarker" by Nigel Stallard. Biometrics 2023; 79:23-25. [PMID: 35266548 DOI: 10.1111/biom.13643] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/01/2021] [Revised: 12/17/2021] [Accepted: 12/23/2021] [Indexed: 11/28/2022]
|
8
|
Four 2×2 factorial trials of smartphone CBT to reduce subthreshold depression and to prevent new depressive episodes among adults in the community-RESiLIENT trial (Resilience Enhancement with Smartphone in LIving ENvironmenTs): a master protocol. BMJ Open 2023; 13:e067850. [PMID: 36828653 PMCID: PMC9972419 DOI: 10.1136/bmjopen-2022-067850] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 08/30/2022] [Accepted: 02/08/2023] [Indexed: 02/26/2023] Open
Abstract
INTRODUCTION The health burden due to depression is ever increasing in the world. Prevention is a key to reducing this burden. Guided internet cognitive-behavioural therapies (iCBT) appear promising but there is room for improvement because we do not yet know which of various iCBT skills are more efficacious than others, and for whom. In addition, there has been no platform for iCBT that can accommodate ongoing evolution of internet technologies. METHODS AND ANALYSIS Based on our decade-long experiences in developing smartphone CBT apps and examining them in randomised controlled trials, we have developed the Resilience Training App Version 2. This app now covers five CBT skills: cognitive restructuring, behavioural activation, problem-solving, assertion training and behaviour therapy for insomnia. The current study is designed as a master protocol including four 2×2 factorial trials using this app (1) to elucidate specific efficacies of each CBT skill, (2) to identify participants' characteristics that enable matching between skills and individuals, and (3) to allow future inclusion of new skills. We will recruit 3520 participants with subthreshold depression and ca 1700 participants without subthreshold depression, to examine the short-term efficacies of CBT skills to reduce depressive symptoms in the former and to explore the long-term efficacies in preventing depression in the total sample. The primary outcome for the short-term efficacies is the change in depressive symptoms as measured with the Patient Health Questionnaire-9 at week 6, and that for the long-term efficacies is the incidence of major depressive episodes as assessed by the computerised Composite International Diagnostic Interview by week 50. ETHICS AND DISSEMINATION The trial has been approved by the Ethics Committee of Kyoto University Graduate School of Medicine (C1556). TRIAL REGISTRATION NUMBER UMIN000047124.
Collapse
|
9
|
Bayesian modelling strategies for borrowing of information in randomised basket trials. J R Stat Soc Ser C Appl Stat 2022; 71:2014-2037. [PMID: 36636028 PMCID: PMC9827857 DOI: 10.1111/rssc.12602] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/27/2022] [Accepted: 09/01/2022] [Indexed: 02/01/2023]
Abstract
Basket trials are an innovative precision medicine clinical trial design evaluating a single targeted therapy across multiple diseases that share a common characteristic. To date, most basket trials have been conducted in early-phase oncology settings, for which several Bayesian methods permitting information sharing across subtrials have been proposed. With the increasing interest of implementing randomised basket trials, information borrowing could be exploited in two ways; considering the commensurability of either the treatment effects or the outcomes specific to each of the treatment groups between the subtrials. In this article, we extend a previous analysis model based on distributional discrepancy for borrowing over the subtrial treatment effects ('treatment effect borrowing', TEB) to borrowing over the subtrial groupwise responses ('treatment response borrowing', TRB). Simulation results demonstrate that both modelling strategies provide substantial gains over an approach with no borrowing. TRB outperforms TEB especially when subtrial sample sizes are small on all operational characteristics, while the latter has considerable gains in performance over TRB when subtrial sample sizes are large, or the treatment effects and groupwise mean responses are noticeably heterogeneous across subtrials. Further, we notice that TRB, and TEB can potentially lead to different conclusions in the analysis of real data.
Collapse
|
10
|
Design and analysis of umbrella trials: Where do we stand? Front Med (Lausanne) 2022; 9:1037439. [PMID: 36313987 PMCID: PMC9596938 DOI: 10.3389/fmed.2022.1037439] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/06/2022] [Accepted: 09/26/2022] [Indexed: 11/13/2022] Open
Abstract
Background The efficiencies that master protocol designs can bring to modern drug development have seen their increased utilization in oncology. Growing interest has also resulted in their consideration in non-oncology settings. Umbrella trials are one class of master protocol design that evaluates multiple targeted therapies in a single disease setting. Despite the existence of several reviews of master protocols, the statistical considerations of umbrella trials have received more limited attention. Methods We conduct a systematic review of the literature on umbrella trials, examining both the statistical methods that are available for their design and analysis, and also their use in practice. We pay particular attention to considerations for umbrella designs applied outside of oncology. Findings We identified 38 umbrella trials. To date, most umbrella trials have been conducted in early phase settings (73.7%, 28/38) and in oncology (92.1%, 35/38). The quality of statistical information available about conducted umbrella trials to date is poor; for example, it was impossible to ascertain how sample size was determined in the majority of trials (55.3%, 21/38). The literature on statistical methods for umbrella trials is currently sparse. Conclusions Umbrella trials have potentially great utility to expedite drug development, including outside of oncology. However, to enable lessons to be effectively learned from early use of such designs, there is a need for higher-quality reporting of umbrella trials. Furthermore, if the potential of umbrella trials is to be realized, further methodological research is required.
Collapse
|
11
|
Subgroup analyses in randomized controlled trials frequently categorized continuous subgroup information. J Clin Epidemiol 2022; 150:72-79. [PMID: 35788399 DOI: 10.1016/j.jclinepi.2022.06.017] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2022] [Revised: 06/15/2022] [Accepted: 06/28/2022] [Indexed: 10/17/2022]
Abstract
BACKGROUND AND OBJECTIVES To investigate how subgroup analyses of published Randomized Controlled Trials (RCTs) are performed when subgroups are created from continuous variables. METHODS We carried out a review of RCTs published in 2016-2021 that included subgroup analyses. Information was extracted on whether any of the subgroups were based on continuous variables and, if so, how they were analyzed. RESULTS Out of 428 reviewed papers, 258 (60.4%) reported RCTs with a subgroup analysis. Of these, 178/258 (69%) had at least one subgroup formed from a continuous variable and 14/258 (5.4%) were unclear. The vast majority (169/178, 94.9%) dichotomized the continuous variable and treated the subgroup as categorical. The most common way of dichotomizing was using a pre-specified cutpoint (129/169, 76.3%), followed by a data-driven cutpoint (26/169, 15.4%), such as the median. CONCLUSION It is common for subgroup analyses to use continuous variables to define subgroups. The vast majority dichotomize the continuous variable and, consequently, may lose substantial amounts of statistical information (equivalent to reducing the sample size by at least a third). More advanced methods that can improve efficiency, through optimally choosing cutpoints or directly using the continuous information, are rarely used.
Collapse
|
12
|
Practical guidance for planning resources required to support publicly-funded adaptive clinical trials. BMC Med 2022; 20:254. [PMID: 35945610 PMCID: PMC9364623 DOI: 10.1186/s12916-022-02445-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 02/01/2022] [Accepted: 06/20/2022] [Indexed: 11/15/2022] Open
Abstract
Adaptive designs are a class of methods for improving efficiency and patient benefit of clinical trials. Although their use has increased in recent years, research suggests they are not used in many situations where they have potential to bring benefit. One barrier to their more widespread use is a lack of understanding about how the choice to use an adaptive design, rather than a traditional design, affects resources (staff and non-staff) required to set-up, conduct and report a trial. The Costing Adaptive Trials project investigated this issue using quantitative and qualitative research amongst UK Clinical Trials Units. Here, we present guidance that is informed by our research, on considering the appropriate resourcing of adaptive trials. We outline a five-step process to estimate the resources required and provide an accompanying costing tool. The process involves understanding the tasks required to undertake a trial, and how the adaptive design affects them. We identify barriers in the publicly funded landscape and provide recommendations to trial funders that would address them. Although our guidance and recommendations are most relevant to UK non-commercial trials, many aspects are relevant more widely.
Collapse
|
13
|
Sample size estimation using a latent variable model for mixed outcome co-primary, multiple primary and composite endpoints. Stat Med 2022; 41:2303-2316. [PMID: 35199380 PMCID: PMC7612654 DOI: 10.1002/sim.9356] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/24/2020] [Revised: 01/28/2022] [Accepted: 02/02/2022] [Indexed: 12/30/2022]
Abstract
Mixed outcome endpoints that combine multiple continuous and discrete components are often employed as primary outcome measures in clinical trials. These may be in the form of co-primary endpoints, which conclude effectiveness overall if an effect occurs in all of the components, or multiple primary endpoints, which require an effect in at least one of the components. Alternatively, they may be combined to form composite endpoints, which reduce the outcomes to a one-dimensional endpoint. There are many advantages to joint modeling the individual outcomes, however in order to do this in practice we require techniques for sample size estimation. In this article we show how the latent variable model can be used to estimate the joint endpoints and propose hypotheses, power calculations and sample size estimation methods for each. We illustrate the techniques using a numerical example based on a four-dimensional endpoint and find that the sample size required for the co-primary endpoint is larger than that required for the individual endpoint with the smallest effect size. Conversely, the sample size required in the multiple primary case is similar to that needed for the outcome with the largest effect size. We show that the empirical power is achieved for each endpoint and that the FWER can be sufficiently controlled using a Bonferroni correction if the correlations between endpoints are less than 0.5. Otherwise, less conservative adjustments may be needed. We further illustrate empirically the efficiency gains that may be achieved in the composite endpoint setting.
Collapse
|
14
|
When is a two-stage single-arm trial efficient? An evaluation of the impact of outcome delay. Eur J Cancer 2022; 166:270-278. [PMID: 35344852 DOI: 10.1016/j.ejca.2022.02.010] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/29/2021] [Revised: 02/01/2022] [Accepted: 02/04/2022] [Indexed: 11/20/2022]
Abstract
BACKGROUND Simon's two-stage design is a widely used adaptive design, particularly in phase II oncology trials due to its simplicity and efficiency. However, its efficiency can be adversely affected when the primary end-point takes time to observe, as is common in practice. METHODS We propose an optimal design, taking the delay in observing treatment outcome into consideration and compare the efficiency gained from using Simon's design over a single-stage design for real-life oncology trials. Based on the results, we provide a general rule-of-thumb for determining whether a two-stage single-arm design can provide any added advantage over a single-stage design, given the recruitment rate and primary end-point length. RESULTS We observed an average 15-30% loss in the estimated efficiency gain in real oncology trials that used Simon's design due to the delay in observing the treatment outcome. The delay-optimal design provides some advantage over Simon's design in terms of reduced sample size when the delay is large compared to the recruitment length. DISCUSSION Simon's two-stage design provides large benefit over a single-stage design, in terms of reduced sample size, when the primary end-point length is no more than 10% of the total recruitment time. It provides no efficiency advantage when this ratio is above 50%.
Collapse
|
15
|
Adaptive Designs: Benefits and Cautions for Neurosurgery Trials. World Neurosurg 2022; 161:316-322. [PMID: 35505550 DOI: 10.1016/j.wneu.2021.07.061] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/27/2021] [Revised: 07/11/2021] [Accepted: 07/12/2021] [Indexed: 10/18/2022]
Abstract
BACKGROUND It is well accepted that randomized controlled trials provide the greatest quality of evidence about effectiveness and safety of new interventions. In neurosurgery, randomized controlled trials face challenges, with their use remaining relatively low compared with other clinical areas. Adaptive designs have emerged as a method for improving the efficiency and patient benefit of trials. They allow modifications to the trial design to be made as patient outcome data are collected. The benefit they provide is highly variable, predominantly governed by the time taken to observe the primary endpoint compared with the planned recruitment rate. They also face challenges in design, conduct, and reporting. METHODS We provide an overview of the benefits and challenges of adaptive designs, with a focus on neurosurgery applications. To investigate how often an adaptive design may be advantageous in neurosurgery, we extracted data on recruitment rates and endpoint lengths for ongoing neurosurgery trials registered in ClinicalTrials.gov. RESULTS We found that a majority of neurosurgery trials had a relatively short endpoint length compared with the planned recruitment period and therefore may benefit from an adaptive trial. However, we did not identify any ongoing ClinicalTrials.gov registered neurosurgery trials that mentioned using an adaptive design. CONCLUSIONS Adaptive designs may provide benefits to neurosurgery trials and should be considered for use more widely. Use of some types of adaptive design, such as multiarm multistage, may further increase the number of interventions that can be tested with limited patient and financial resources.
Collapse
|
16
|
Bayesian sample size determination using commensurate priors to leverage pre-experimental data. Biometrics 2022; 79:669-683. [PMID: 38523700 PMCID: PMC7614678 DOI: 10.1111/j.1541-0420.2005.00454.x] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/26/2024]
Abstract
This paper develops Bayesian sample size formulae for experiments comparing two groups, where relevant pre-experimental information from multiple sources can be incorporated in a robust prior to support both the design and analysis. We use commensurate predictive priors for borrowing of information, and further place Gamma mixture priors on the precisions to account for preliminary belief about the pairwise (in)commensurability between parameters that underpin the historical and new experiments. Averaged over the probability space of the new experimental data, appropriate sample sizes are found according to criteria that control certain aspects of the posterior distribution, such as the coverage probability or length of a defined density region. Our Bayesian methodology can be applied to circumstances that compare two normal means, proportions or event times. When nuisance parameters (such as variance) in the new experiment are unknown, a prior distribution can further be specified based on pre-experimental data. Exact solutions are available based on most of the criteria considered for Bayesian sample size determination, while a search procedure is described in cases for which there are no closed-form expressions. We illustrate the application of our sample size formulae in the design of clinical trials, where pre-trial information is available to be leveraged. Hypothetical data examples, motivated by a rare-disease trial with elicited expert prior opinion, and a comprehensive performance evaluation of the proposed methodology are presented.
Collapse
|
17
|
Advantages of multi-arm non-randomised sequentially allocated cohort designs for Phase II oncology trials. Br J Cancer 2022; 126:204-210. [PMID: 34750494 PMCID: PMC8770479 DOI: 10.1038/s41416-021-01613-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/28/2021] [Revised: 10/14/2021] [Accepted: 10/21/2021] [Indexed: 11/27/2022] Open
Abstract
BACKGROUND Efficient trial designs are required to prioritise promising drugs within Phase II trials. Adaptive designs are examples of such designs, but their efficiency is reduced if there is a delay in assessing patient responses to treatment. METHODS Motivated by the WIRE trial in renal cell carcinoma (NCT03741426), we compare three trial approaches to testing multiple treatment arms: (1) single-arm trials in sequence with interim analyses; (2) a parallel multi-arm multi-stage trial and (3) the design used in WIRE, which we call the Multi-Arm Sequential Trial with Efficient Recruitment (MASTER) design. The MASTER design recruits patients to one arm at a time, pausing recruitment to an arm when it has recruited the required number for an interim analysis. We conduct a simulation study to compare how long the three different trial designs take to evaluate a number of new treatment arms. RESULTS The parallel multi-arm multi-stage and the MASTER design are much more efficient than separate trials. The MASTER design provides extra efficiency when there is endpoint delay, or recruitment is very quick. CONCLUSIONS We recommend the MASTER design as an efficient way of testing multiple promising cancer treatments in non-comparative Phase II trials.
Collapse
|
18
|
Improving power in PSA response analyses of metastatic castration-resistant prostate cancer trials. BMC Cancer 2022; 22:111. [PMID: 35081926 PMCID: PMC8793251 DOI: 10.1186/s12885-022-09227-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/15/2021] [Accepted: 12/24/2021] [Indexed: 11/15/2022] Open
Abstract
BACKGROUND To determine how much an augmented analysis approach could improve the efficiency of prostate-specific antigen (PSA) response analyses in clinical practice. PSA response rates are commonly used outcome measures in metastatic castration-resistant prostate cancer (mCRPC) trial reports. PSA response is evaluated by comparing continuous PSA data (e.g., change from baseline) to a threshold (e.g., 50% reduction). Consequently, information in the continuous data is discarded. Recent papers have proposed an augmented approach that retains the conventional response rate, but employs the continuous data to improve precision of estimation. METHODS A literature review identified published prostate cancer trials that included a waterfall plot of continuous PSA data. This continuous data was extracted to enable the conventional and augmented approaches to be compared. RESULTS Sixty-four articles, reporting results for 78 mCRPC treatment arms, were re-analysed. The median efficiency gain from using the augmented analysis, in terms of the implied increase to the sample size of the original study, was 103.2% (IQR [89.8,190.9%]). CONCLUSIONS Augmented PSA response analysis requires no additional data to be collected and can be performed easily using available software. It improves precision of estimation to a degree that is equivalent to a substantial sample size increase. The implication of this work is that prostate cancer trials using PSA response as a primary endpoint could be delivered with fewer participants and, therefore, more rapidly with reduced cost.
Collapse
|
19
|
Response adaptive intervention allocation in stepped-wedge cluster randomized trials. Stat Med 2022; 41:1081-1099. [PMID: 35064595 PMCID: PMC7612601 DOI: 10.1002/sim.9317] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/12/2021] [Revised: 11/30/2021] [Accepted: 12/22/2021] [Indexed: 11/09/2022]
Abstract
BACKGROUND Stepped-wedge cluster randomized trial (SW-CRT) designs are often used when there is a desire to provide an intervention to all enrolled clusters, because of a belief that it will be effective. However, given there should be equipoise at trial commencement, there has been discussion around whether a pre-trial decision to provide the intervention to all clusters is appropriate. In pharmaceutical drug development, a solution to a similar desire to provide more patients with an effective treatment is to use a response adaptive (RA) design. METHODS We introduce a way in which RA design could be incorporated in an SW-CRT, permitting modification of the intervention allocation during the trial. The proposed framework explicitly permits a balance to be sought between power and patient benefit considerations. A simulation study evaluates the methodology. RESULTS In one scenario, for one particular RA design, the proportion of cluster-periods spent in the intervention condition was observed to increase from 32.2% to 67.9% as the intervention effect was increased. A cost of this was a 6.2% power drop compared to a design that maximized power by fixing the proportion of time in the intervention condition at 45.0%, regardless of the intervention effect. CONCLUSIONS An RA approach may be most applicable to settings for which the intervention has substantial individual or societal benefit considerations, potentially in combination with notable safety concerns. In such a setting, the proposed methodology may routinely provide the desired adaptability of the roll-out speed, with only a small cost to the study's power.
Collapse
|
20
|
Conditional power and friends: The why and how of (un)planned, unblinded sample size recalculations in confirmatory trials. Stat Med 2022; 41:877-890. [PMID: 35023184 PMCID: PMC9303654 DOI: 10.1002/sim.9288] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/06/2021] [Revised: 10/21/2021] [Accepted: 12/02/2021] [Indexed: 11/09/2022]
Abstract
Adapting the final sample size of a trial to the evidence accruing during the trial is a natural way to address planning uncertainty. Since the sample size is usually determined by an argument based on the power of the trial, an interim analysis raises the question of how the final sample size should be determined conditional on the accrued information. To this end, we first review and compare common approaches to estimating conditional power, which is often used in heuristic sample size recalculation rules. We then discuss the connection of heuristic sample size recalculation and optimal two-stage designs, demonstrating that the latter is the superior approach in a fully preplanned setting. Hence, unplanned design adaptations should only be conducted as reaction to trial-external new evidence, operational needs to violate the originally chosen design, or post hoc changes in the optimality criterion but not as a reaction to trial-internal data. We are able to show that commonly discussed sample size recalculation rules lead to paradoxical adaptations where an initially planned optimal design is not invariant under the adaptation rule even if the planning assumptions do not change. Finally, we propose two alternative ways of reacting to newly emerging trial-external evidence in ways that are consistent with the originally planned design to avoid such inconsistencies.
Collapse
|
21
|
Increasing power in the analysis of responder endpoints in rheumatology: a software tutorial. BMC Rheumatol 2021; 5:54. [PMID: 34872620 PMCID: PMC8650391 DOI: 10.1186/s41927-021-00224-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/03/2021] [Accepted: 08/16/2021] [Indexed: 11/10/2022] Open
Abstract
BACKGROUND Composite responder endpoints feature frequently in rheumatology due to the multifaceted nature of many of these conditions. Current analysis methods used to analyse these endpoints discard much of the data used to classify patients as responders and are therefore highly inefficient, resulting in low power. We highlight a novel augmented methodology that uses more of the information available to improve the precision of reported treatment effects. Since these methods are more challenging to implement, we developed free, user-friendly software available in a web-based interface and as R packages. The software consists of two programs: one that supports the analysis of responder endpoints; the second that facilitates sample size estimation. We demonstrate the use of the software to conduct the analysis with both the augmented and standard analysis method using the MUSE study, a phase IIb trial in patients with systemic lupus erythematosus. RESULTS The software outputs similar point estimates with smaller confidence intervals for the odds ratio, risk ratio and risk difference estimators using the augmented approach. The sample size required in each arm for a future trial using the novel approach based on the MUSE data is 50 versus 135 for the standard method, translating to a reduction in required sample size of approximately 63%. CONCLUSIONS We encourage trialists to use the software demonstrated to implement the augmented methodology in future studies to improve efficiency.
Collapse
|
22
|
Costs and staffing resource requirements for adaptive clinical trials: quantitative and qualitative results from the Costing Adaptive Trials project. BMC Med 2021; 19:251. [PMID: 34696781 PMCID: PMC8545558 DOI: 10.1186/s12916-021-02124-z] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 07/01/2021] [Accepted: 09/13/2021] [Indexed: 11/20/2022] Open
Abstract
BACKGROUND Adaptive designs offer great promise in improving the efficiency and patient-benefit of clinical trials. An important barrier to further increased use is a lack of understanding about which additional resources are required to conduct a high-quality adaptive clinical trial, compared to a traditional fixed design. The Costing Adaptive Trials (CAT) project investigated which additional resources may be required to support adaptive trials. METHODS We conducted a mock costing exercise amongst seven Clinical Trials Units (CTUs) in the UK. Five scenarios were developed, derived from funded clinical trials, where a non-adaptive version and an adaptive version were described. Each scenario represented a different type of adaptive design. CTU staff were asked to provide the costs and staff time they estimated would be needed to support the trial, categorised into specified areas (e.g. statistics, data management, trial management). This was calculated separately for the non-adaptive and adaptive version of the trial, allowing paired comparisons. Interviews with 10 CTU staff who had completed the costing exercise were conducted by qualitative researchers to explore reasons for similarities and differences. RESULTS Estimated resources associated with conducting an adaptive trial were always (moderately) higher than for the non-adaptive equivalent. The median increase was between 2 and 4% for all scenarios, except for sample size re-estimation which was 26.5% (as the adaptive design could lead to a lengthened study period). The highest increase was for statistical staff, with lower increases for data management and trial management staff. The percentage increase in resources varied across different CTUs. The interviews identified possible explanations for differences, including (1) experience in adaptive trials, (2) the complexity of the non-adaptive and adaptive design, and (3) the extent of non-trial specific core infrastructure funding the CTU had. CONCLUSIONS This work sheds light on additional resources required to adequately support a high-quality adaptive trial. The percentage increase in costs for supporting an adaptive trial was generally modest and should not be a barrier to adaptive designs being cost-effective to use in practice. Informed by the results of this research, guidance for investigators and funders will be developed on appropriately resourcing adaptive trials.
Collapse
|
23
|
Innovative trial approaches in immune-mediated inflammatory diseases: current use and future potential. BMC Rheumatol 2021; 5:21. [PMID: 34210348 PMCID: PMC8252241 DOI: 10.1186/s41927-021-00192-5] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/20/2020] [Accepted: 04/09/2021] [Indexed: 01/12/2023] Open
Abstract
BACKGROUND Despite progress that has been made in the treatment of many immune-mediated inflammatory diseases (IMIDs), there remains a need for improved treatments. Randomised controlled trials (RCTs) provide the highest form of evidence on the effectiveness of a potential new treatment regimen, but they are extremely expensive and time consuming to conduct. Consequently, much focus has been given in recent years to innovative design and analysis methods that could improve the efficiency of RCTs. In this article, we review the current use and future potential of these methods within the context of IMID trials. METHODS We provide a review of several innovative methods that would provide utility in IMID research. These include novel study designs (adaptive trials, Sequential Multi-Assignment Randomised Trials, basket, and umbrella trials) and data analysis methodologies (augmented analyses of composite responder endpoints, using high-dimensional biomarker information to stratify patients, and emulation of RCTs from routinely collected data). IMID trials are now well-placed to embrace innovative methods. For example, well-developed statistical frameworks for adaptive trial design are ready for implementation, whilst the growing availability of historical datasets makes the use of Bayesian methods particularly applicable. To assess whether and how these innovative methods have been used in practice, we conducted a review via PubMed of clinical trials pertaining to any of 51 IMIDs that were published between 2018 and 20 in five high impact factor clinical journals. RESULTS Amongst 97 articles included in the review, 19 (19.6%) used an innovative design method, but most of these were relatively straightforward examples of innovative approaches. Only two (2.1%) reported the use of evidence from routinely collected data, cohorts, or biobanks. Eight (9.2%) collected high-dimensional data. CONCLUSIONS Application of innovative statistical methodology to IMID trials has the potential to greatly improve efficiency, to generalise and extrapolate trial results, and to further personalise treatment strategies. Currently, such methods are infrequently utilised in practice. New research is required to ensure that IMID trials can benefit from the most suitable methods.
Collapse
|
24
|
Developing a predictive signature for two trial endpoints using the cross-validated risk scores method. Biostatistics 2021; 24:327-344. [PMID: 34165151 PMCID: PMC10102911 DOI: 10.1093/biostatistics/kxaa055] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/30/2020] [Revised: 11/11/2020] [Accepted: 11/30/2020] [Indexed: 11/13/2022] Open
Abstract
The existing cross-validated risk scores (CVRS) design has been proposed for developing and testing the efficacy of a treatment in a high-efficacy patient group (the sensitive group) using high-dimensional data (such as genetic data). The design is based on computing a risk score for each patient and dividing them into clusters using a nonparametric clustering procedure. In some settings, it is desirable to consider the tradeoff between two outcomes, such as efficacy and toxicity, or cost and effectiveness. With this motivation, we extend the CVRS design (CVRS2) to consider two outcomes. The design employs bivariate risk scores that are divided into clusters. We assess the properties of the CVRS2 using simulated data and illustrate its application on a randomized psychiatry trial. We show that CVRS2 is able to reliably identify the sensitive group (the group for which the new treatment provides benefit on both outcomes) in the simulated data. We apply the CVRS2 design to a psychology clinical trial that had offender status and substance use status as two outcomes and collected a large number of baseline covariates. The CVRS2 design yields a significant treatment effect for both outcomes, while the CVRS approach identified a significant effect for the offender status only after prefiltering the covariates.
Collapse
|
25
|
A Review of Bayesian Perspectives on Sample Size Derivation for Confirmatory Trials. AM STAT 2021; 75:424-432. [PMID: 34992303 PMCID: PMC7612172 DOI: 10.1080/00031305.2021.1901782] [Citation(s) in RCA: 21] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/29/2020] [Revised: 02/28/2021] [Accepted: 03/01/2021] [Indexed: 11/28/2022]
Abstract
Sample size derivation is a crucial element of planning any confirmatory trial. The required sample size is typically derived based on constraints on the maximal acceptable Type I error rate and minimal desired power. Power depends on the unknown true effect and tends to be calculated either for the smallest relevant effect or a likely point alternative. The former might be problematic if the minimal relevant effect is close to the null, thus requiring an excessively large sample size, while the latter is dubious since it does not account for the a priori uncertainty about the likely alternative effect. A Bayesian perspective on sample size derivation for a frequentist trial can reconcile arguments about the relative a priori plausibility of alternative effects with ideas based on the relevance of effect sizes. Many suggestions as to how such "hybrid" approaches could be implemented in practice have been put forward. However, key quantities are often defined in subtly different ways in the literature. Starting from the traditional entirely frequentist approach to sample size derivation, we derive consistent definitions for the most commonly used hybrid quantities and highlight connections, before discussing and demonstrating their use in sample size derivation for clinical trials.
Collapse
|
26
|
Bayesian design and analysis of external pilot trials for complex interventions. Stat Med 2021; 40:2877-2892. [PMID: 33733500 DOI: 10.1002/sim.8941] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/23/2019] [Revised: 02/15/2021] [Accepted: 02/17/2021] [Indexed: 11/08/2022]
Abstract
External pilot trials of complex interventions are used to help determine if and how a confirmatory trial should be undertaken, providing estimates of parameters such as recruitment, retention, and adherence rates. The decision to progress to the confirmatory trial is typically made by comparing these estimates to pre-specified thresholds known as progression criteria, although the statistical properties of such decision rules are rarely assessed. Such assessment is complicated by several methodological challenges, including the simultaneous evaluation of multiple endpoints, complex multi-level models, small sample sizes, and uncertainty in nuisance parameters. In response to these challenges, we describe a Bayesian approach to the design and analysis of external pilot trials. We show how progression decisions can be made by minimizing the expected value of a loss function, defined over the whole parameter space to allow for preferences and trade-offs between multiple parameters to be articulated and used in the decision-making process. The assessment of preferences is kept feasible by using a piecewise constant parametrization of the loss function, the parameters of which are chosen at the design stage to lead to desirable operating characteristics. We describe a flexible, yet computationally intensive, nested Monte Carlo algorithm for estimating operating characteristics. The method is used to revisit the design of an external pilot trial of a complex intervention designed to increase the physical activity of care home residents.
Collapse
|
27
|
Controlling type I error rates in multi-arm clinical trials: A case for the false discovery rate. Pharm Stat 2021; 20:109-116. [PMID: 32790026 PMCID: PMC7612170 DOI: 10.1002/pst.2059] [Citation(s) in RCA: 16] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/09/2019] [Revised: 06/04/2020] [Accepted: 07/16/2020] [Indexed: 11/30/2022]
Abstract
Multi-arm trials are an efficient way of simultaneously testing several experimental treatments against a shared control group. As well as reducing the sample size required compared to running each trial separately, they have important administrative and logistical advantages. There has been debate over whether multi-arm trials should correct for the fact that multiple null hypotheses are tested within the same experiment. Previous opinions have ranged from no correction is required, to a stringent correction (controlling the probability of making at least one type I error) being needed, with regulators arguing the latter for confirmatory settings. In this article, we propose that controlling the false-discovery rate (FDR) is a suitable compromise, with an appealing interpretation in multi-arm clinical trials. We investigate the properties of the different correction methods in terms of the positive and negative predictive value (respectively how confident we are that a recommended treatment is effective and that a non-recommended treatment is ineffective). The number of arms and proportion of treatments that are truly effective is varied. Controlling the FDR provides good properties. It retains the high positive predictive value of FWER correction in situations where a low proportion of treatments is effective. It also has a good negative predictive value in situations where a high proportion of treatments is effective. In a multi-arm trial testing distinct treatment arms, we recommend that sponsors and trialists consider use of the FDR.
Collapse
|
28
|
Graphical approaches for the control of generalized error rates. Stat Med 2020; 39:3135-3155. [PMID: 32557848 PMCID: PMC7612110 DOI: 10.1002/sim.8595] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/03/2020] [Accepted: 05/10/2020] [Indexed: 11/12/2022]
Abstract
When simultaneously testing multiple hypotheses, the usual approach in the context of confirmatory clinical trials is to control the familywise error rate (FWER), which bounds the probability of making at least one false rejection. In many trial settings, these hypotheses will additionally have a hierarchical structure that reflects the relative importance and links between different clinical objectives. The graphical approach of Bretz et al (2009) is a flexible and easily communicable way of controlling the FWER while respecting complex trial objectives and multiple structured hypotheses. However, the FWER can be a very stringent criterion that leads to procedures with low power, and may not be appropriate in exploratory trial settings. This motivates controlling generalized error rates, particularly when the number of hypotheses tested is no longer small. We consider the generalized familywise error rate (k-FWER), which is the probability of making k or more false rejections, as well as the tail probability of the false discovery proportion (FDP), which is the probability that the proportion of false rejections is greater than some threshold. We also consider asymptotic control of the false discovery rate, which is the expectation of the FDP. In this article, we show how to control these generalized error rates when using the graphical approach and its extensions. We demonstrate the utility of the resulting graphical procedures on three clinical trial case studies.
Collapse
|
29
|
Developing and testing high‐efficacy patient subgroups within a clinical trial using risk scores. Stat Med 2020; 39:3285-3298. [PMID: 32662542 PMCID: PMC7611900 DOI: 10.1002/sim.8665] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/12/2019] [Revised: 03/18/2020] [Accepted: 05/28/2020] [Indexed: 12/13/2022]
Abstract
There is the potential for high-dimensional information about patients collected in clinical trials (such as genomic, imaging, and data from wearable technologies) to be informative for the efficacy of a new treatment in situations where only a subset of patients benefits from the treatment. The adaptive signature design (ASD) method has been proposed for developing and testing the efficacy of a treatment in a high-efficacy patient group (the sensitive group) using genetic data. The method requires selection of three tuning parameters which may be highly computationally expensive. We propose a variation to the ASD method, the cross-validated risk scores (CVRS) design method, that does not require selection of any tuning parameters. The method is based on computing a risk score for each patient and dividing them into clusters using a nonparametric clustering procedure.We assess the properties of CVRS against the originally proposed cross-validated ASD using simulation data and a real psychiatry trial. CVRS, as assessed for various sample sizes and response rates, has a substantial reduction in the computational time required. In many simulation scenarios, there is a substantial improvement in the ability to correctly identify the sensitive group and the power of the design to detect a treatment effect in the sensitive group.We illustrate the application of the CVRS method on the psychiatry trial.
Collapse
|
30
|
Borrowing of information across patient subgroups in a basket trial based on distributional discrepancy. Biostatistics 2020; 23:120-135. [PMID: 32380518 PMCID: PMC8759447 DOI: 10.1093/biostatistics/kxaa019] [Citation(s) in RCA: 23] [Impact Index Per Article: 5.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/12/2019] [Revised: 03/20/2020] [Accepted: 03/25/2020] [Indexed: 12/02/2022] Open
Abstract
Basket trials have emerged as a new class of efficient approaches in oncology to evaluate a new treatment in several patient subgroups simultaneously. In this article, we extend the key ideas to disease areas outside of oncology, developing a robust Bayesian methodology for randomized, placebo-controlled basket trials with a continuous endpoint to enable borrowing of information across subtrials with similar treatment effects. After adjusting for covariates, information from a complementary subtrial can be represented into a commensurate prior for the parameter that underpins the subtrial under consideration. We propose using distributional discrepancy to characterize the commensurability between subtrials for appropriate borrowing of information through a spike-and-slab prior, which is placed on the prior precision factor. When the basket trial has at least three subtrials, commensurate priors for point-to-point borrowing are combined into a marginal predictive prior, according to the weights transformed from the pairwise discrepancy measures. In this way, only information from subtrial(s) with the most commensurate treatment effect is leveraged. The marginal predictive prior is updated to a robust posterior by the contemporary subtrial data to inform decision making. Operating characteristics of the proposed methodology are evaluated through simulations motivated by a real basket trial in chronic diseases. The proposed methodology has advantages compared to other selected Bayesian analysis models, for (i) identifying the most commensurate source of information and (ii) gauging the degree of borrowing from specific subtrials. Numerical results also suggest that our methodology can improve the precision of estimates and, potentially, the statistical power for hypothesis testing.
Collapse
|
31
|
Simple MRI score aids prediction of dementia in cerebral small vessel disease. Neurology 2020; 94:e1294-e1302. [PMID: 32123050 PMCID: PMC7274929 DOI: 10.1212/wnl.0000000000009141] [Citation(s) in RCA: 60] [Impact Index Per Article: 15.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/07/2019] [Accepted: 10/01/2019] [Indexed: 12/22/2022] Open
Abstract
Objective To determine whether a simple small vessel disease (SVD) score, which uses information available on rapid visual assessment of clinical MRI scans, predicts risk of cognitive decline and dementia, above that provided by simple clinical measures. Methods Three prospective longitudinal cohort studies (SCANS [St George's Cognition and Neuroimaging in Stroke], RUN DMC [Radboud University Nijmegen Diffusion Imaging and Magnetic Resonance Imaging Cohort], and the ASPS [Austrian Stroke Prevention Study]), which covered a range of SVD severity from mild and asymptomatic to severe and symptomatic, were included. In all studies, MRI was performed at baseline, cognitive tests repeated during follow-up, and progression to dementia recorded prospectively. Outcome measures were cognitive decline and onset of dementia during follow-up. We determined whether the SVD score predicted risk of cognitive decline and future dementia. We also determined whether using the score to select a group of patients with more severe disease would reduce sample sizes for clinical intervention trials. Results In a pooled analysis of all 3 cohorts, the score improved prediction of dementia (area under the curve [AUC], 0.85; 95% confidence interval [CI], 0.81–0.89) compared with that from clinical risk factors alone (AUC, 0.76; 95% CI, 0.71–0.81). Predictive performance was higher in patients with more severe SVD. Power calculations showed selecting patients with a higher score reduced sample sizes required for hypothetical clinical trials by 40%–66% depending on the outcome measure used. Conclusions A simple SVD score, easily obtainable from clinical MRI scans and therefore applicable in routine clinical practice, aided prediction of future dementia risk.
Collapse
|
32
|
Abstract
BACKGROUND Adaptive designs are a wide class of methods focused on improving the power, efficiency and participant benefit of clinical trials. They do this through allowing information gathered during the trial to be used to make changes in a statistically robust manner - the changes could include which treatment arms patients are enrolled to (e.g. dropping non-promising treatment arms), the allocation ratios, the target sample size or the enrolment criteria of the trial. Generally, we are enthusiastic about adaptive designs and advocate their use in many clinical situations. However, they are not always advantageous. In some situations, they provide little efficiency advantage or are even detrimental to the quality of information provided by the trial. In our experience, factors that reduce the efficiency of adaptive designs are routinely downplayed or ignored in methodological papers, which may lead researchers into believing they are more beneficial than they actually are. MAIN TEXT In this paper, we discuss situations where adaptive designs may not be as useful, including situations when the outcomes take a long time to observe, when dropping arms early may cause issues and when increased practical complexity eliminates theoretical efficiency gains. CONCLUSION Adaptive designs often provide notable efficiency benefits. However, it is important for investigators to be aware that they do not always provide an advantage. There should always be careful consideration of the potential benefits and disadvantages of an adaptive design.
Collapse
|
33
|
Abstract
We describe and compare two methods for the group sequential design of two-arm experiments with Poisson distributed data, which are based on a normal approximation and exact calculations respectively. A framework to determine near-optimal stopping boundaries is also presented. Using this framework, for a considered example, we demonstrate that a group sequential design could reduce the expected sample size under the null hypothesis by as much as 44% compared to a fixed sample approach. We conclude with a discussion of the advantages and disadvantages of the two presented procedures.
Collapse
|
34
|
Familywise error control in multi-armed response-adaptive trials. Biometrics 2019; 75:885-894. [PMID: 30714095 PMCID: PMC6739232 DOI: 10.1111/biom.13042] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/13/2018] [Accepted: 01/25/2019] [Indexed: 01/21/2023]
Abstract
Response-adaptive designs allow the randomization probabilities to change during the course of a trial based on cumulated response data so that a greater proportion of patients can be allocated to the better performing treatments. A major concern over the use of response-adaptive designs in practice, particularly from a regulatory viewpoint, is controlling the type I error rate. In particular, we show that the naïve z-test can have an inflated type I error rate even after applying a Bonferroni correction. Simulation studies have often been used to demonstrate error control but do not provide a guarantee. In this article, we present adaptive testing procedures for normally distributed outcomes that ensure strong familywise error control by iteratively applying the conditional invariance principle. Our approach can be used for fully sequential and block randomized trials and for a large class of adaptive randomization rules found in the literature. We show there is a high price to pay in terms of power to guarantee familywise error control for randomization schemes with extreme allocation probabilities. However, for proposed Bayesian adaptive randomization schemes in the literature, our adaptive tests maintain or increase the power of the trial compared to the z-test. We illustrate our method using a three-armed trial in primary hypercholesterolemia.
Collapse
|
35
|
A latent variable model for improving inference in trials assessing the effect of dose on toxicity and composite efficacy endpoints. Stat Methods Med Res 2019. [PMID: 30799777 PMCID: PMC6986906 DOI: 10.1177/tobeassigned] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
Abstract
It is often of interest to explore how dose affects the toxicity and efficacy properties of a novel treatment. In oncology, efficacy is often assessed through response, which is defined by a patient having no new tumour lesions and their tumour size shrinking by 30%. Usually response and toxicity are analysed as binary outcomes in early phase trials. Methods have been proposed to improve the efficiency of analysing response by utilising the continuous tumour size information instead of dichotomising it. However, these methods do not allow for toxicity or for different doses. Motivated by a phase II trial testing multiple doses of a treatment against placebo, we propose a latent variable model that can estimate the probability of response and no toxicity (or other related outcomes) for different doses. We assess the confidence interval coverage and efficiency properties of the method, compared to methods that do not use the continuous tumour size, in a simulation study and the real study. The coverage is close to nominal when model assumptions are met, although can be below nominal when the model is misspecified. Compared to methods that treat response as binary, the method has confidence intervals with 30-50% narrower widths. The method adds considerable efficiency but care must be taken that the model assumptions are reasonable.
Collapse
|
36
|
Abstract
Bioequivalence (BE) studies are most often conducted as crossover trials, and therefore establishing their required sample size necessitates specification of the within-person variance. Given that this specification is often difficult in practice, there has been great interest in recent years in the use of adaptive designs for BE trials. However, while numerous methods for this have now been presented, their focus has been solely on two-treatment BE studies. In some instances, it will be desired to incorporate more than a single test and reference formulation into a BE trial. It would therefore be useful to establish methodology for the design of adaptive multi-treatment BE trials, to acquire the benefits in the two-treatment setting in this more complex situation. Here, we achieve this for three-treatment studies by extending previously proposed designs for two-treatment trials. First, we discuss the additional design considerations that arise when multiple comparisons are made. Next, an extensive simulation study is employed to compare the performance of the proposed procedures. With this, we demonstrate that two-stage designs with desirable statistical operating characteristics can be readily identified for three-treatment BE trials. Supplementary materials for this article are available online.
Collapse
|
37
|
Multi-arm multi-stage trials can improve the efficiency of finding effective treatments for stroke: a case study. BMC Cardiovasc Disord 2018; 18:215. [PMID: 30482176 PMCID: PMC6260683 DOI: 10.1186/s12872-018-0956-4] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/11/2018] [Accepted: 11/15/2018] [Indexed: 01/07/2023] Open
Abstract
Background Many recent Stroke trials fail to show a beneficial effect of the intervention late in the development. Currently a large number of new treatment options are being developed. Multi-arm multi-stage (MAMS) designs offer one potential strategy to avoid lengthy studies of treatments without beneficial effects while at the same time allowing evaluation of several novel treatments. In this paper we provide a review of what MAMS designs are and argue that they are of particular value for Stroke trials. We illustrate this benefit through a case study based on previous published trials of endovascular treatment for acute ischemic stroke. We show in this case study that MAMS trials provide additional power for the same sample size compared to alternative trial designs. This level of additional power depends on the recruitment length of the trial, with most efficiency gained when recruitment is relatively slow. We conclude with a discussion of additional considerations required when starting a MAMS trial. Conclusion MAMS trial designs are potentially very useful for stroke trials due to their improved statistical power compared to the traditional approach.
Collapse
|
38
|
Admissible multiarm stepped-wedge cluster randomized trial designs. Stat Med 2018; 38:1103-1119. [PMID: 30402914 PMCID: PMC6491976 DOI: 10.1002/sim.8022] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/27/2017] [Revised: 08/23/2018] [Accepted: 10/10/2018] [Indexed: 11/24/2022]
Abstract
Numerous publications have now addressed the principles of designing, analyzing, and reporting the results of stepped‐wedge cluster randomized trials. In contrast, there is little research available pertaining to the design and analysis of multiarm stepped‐wedge cluster randomized trials, utilized to evaluate the effectiveness of multiple experimental interventions. In this paper, we address this by explaining how the required sample size in these multiarm trials can be ascertained when data are to be analyzed using a linear mixed model. We then go on to describe how the design of such trials can be optimized to balance between minimizing the cost of the trial and minimizing some function of the covariance matrix of the treatment effect estimates. Using a recently commenced trial that will evaluate the effectiveness of sensor monitoring in an occupational therapy rehabilitation program for older persons after hip fracture as an example, we demonstrate that our designs could reduce the number of observations required for a fixed power level by up to 58%. Consequently, when logistical constraints permit the utilization of any one of a range of possible multiarm stepped‐wedge cluster randomized trial designs, researchers should consider employing our approach to optimize their trials efficiency.
Collapse
|
39
|
Group sequential crossover trial designs with strong control of the familywise error rate. Seq Anal 2018; 37:174-203. [PMID: 30393467 PMCID: PMC6199128 DOI: 10.1080/07474946.2018.1466528] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/03/2017] [Revised: 01/08/2018] [Accepted: 03/31/2018] [Indexed: 12/05/2022]
Abstract
Crossover designs are an extremely useful tool to investigators, and group
sequential methods have proven highly proficient at improving the efficiency of
parallel group trials. Yet, group sequential methods and crossover designs have
rarely been paired together. One possible explanation for this could be the
absence of a formal proof of how to strongly control the familywise error rate
in the case when multiple comparisons will be made. Here, we provide this proof,
valid for any number of initial experimental treatments and any number of
stages, when results are analyzed using a linear mixed model. We then establish
formulae for the expected sample size and expected number of observations of
such a trial, given any choice of stopping boundaries. Finally, utilizing the
four-treatment, four-period TOMADO trial as an example, we demonstrate that
group sequential methods in this setting could have reduced the trials expected
number of observations under the global null hypothesis by over 33%.
Collapse
|
40
|
Blinded and unblinded sample size reestimation procedures for stepped-wedge cluster randomized trials. Biom J 2018; 60:903-916. [PMID: 30073685 PMCID: PMC6175439 DOI: 10.1002/bimj.201700125] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/26/2017] [Revised: 06/26/2018] [Accepted: 06/27/2018] [Indexed: 11/05/2022]
Abstract
The ability to accurately estimate the sample size required by a stepped-wedge (SW) cluster randomized trial (CRT) routinely depends upon the specification of several nuisance parameters. If these parameters are misspecified, the trial could be overpowered, leading to increased cost, or underpowered, enhancing the likelihood of a false negative. We address this issue here for cross-sectional SW-CRTs, analyzed with a particular linear-mixed model, by proposing methods for blinded and unblinded sample size reestimation (SSRE). First, blinded estimators for the variance parameters of a SW-CRT analyzed using the Hussey and Hughes model are derived. Following this, procedures for blinded and unblinded SSRE after any time period in a SW-CRT are detailed. The performance of these procedures is then examined and contrasted using two example trial design scenarios. We find that if the two key variance parameters were underspecified by 50%, the SSRE procedures were able to increase power over the conventional SW-CRT design by up to 41%, resulting in an empirical power above the desired level. Thus, though there are practical issues to consider, the performance of the procedures means researchers should consider incorporating SSRE in to future SW-CRTs.
Collapse
|
41
|
Blinded and unblinded sample size reestimation in crossover trials balanced for period. Biom J 2018; 60:917-933. [PMID: 30073679 PMCID: PMC6175184 DOI: 10.1002/bimj.201700092] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2017] [Revised: 06/06/2018] [Accepted: 06/07/2018] [Indexed: 11/06/2022]
Abstract
The determination of the sample size required by a crossover trial typically depends on the specification of one or more variance components. Uncertainty about the value of these parameters at the design stage means that there is often a risk a trial may be under- or overpowered. For many study designs, this problem has been addressed by considering adaptive design methodology that allows for the re-estimation of the required sample size during a trial. Here, we propose and compare several approaches for this in multitreatment crossover trials. Specifically, regulators favor reestimation procedures to maintain the blinding of the treatment allocations. We therefore develop blinded estimators for the within and between person variances, following simple or block randomization. We demonstrate that, provided an equal number of patients are allocated to sequences that are balanced for period, the proposed estimators following block randomization are unbiased. We further provide a formula for the bias of the estimators following simple randomization. The performance of these procedures, along with that of an unblinded approach, is then examined utilizing three motivating examples, including one based on a recently completed four-treatment four-period crossover trial. Simulation results show that the performance of the proposed blinded procedures is in many cases similar to that of the unblinded approach, and thus they are an attractive alternative.
Collapse
|
42
|
Group sequential clinical trial designs for normally distributed outcome variables. THE STATA JOURNAL 2018; 18:416-431. [PMID: 35125974 PMCID: PMC7612318 DOI: 10.1177/1536867x1801800208] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
In a group sequential clinical trial, accumulated data are analyzed at numerous time points to allow early decisions about a hypothesis of interest. These designs have historically been recommended for their ethical, administrative, and economic benefits. In this article, we first discuss a collection of new commands for computing the stopping boundaries and required group size of various classical group sequential designs, assuming a normally distributed outcome variable. Then, we demonstrate how the performance of several designs can be compared graphically.
Collapse
|
43
|
Improving the analysis of composite endpoints in rare disease trials. Orphanet J Rare Dis 2018; 13:81. [PMID: 29788976 PMCID: PMC5964664 DOI: 10.1186/s13023-018-0819-1] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/13/2017] [Accepted: 05/01/2018] [Indexed: 12/31/2022] Open
Abstract
BACKGROUND Composite endpoints are recommended in rare diseases to increase power and/or to sufficiently capture complexity. Often, they are in the form of responder indices which contain a mixture of continuous and binary components. Analyses of these outcomes typically treat them as binary, thus only using the dichotomisations of continuous components. The augmented binary method offers a more efficient alternative and is therefore especially useful for rare diseases. Previous work has indicated the method may have poorer statistical properties when the sample size is small. Here we investigate small sample properties and implement small sample corrections. METHODS We re-sample from a previous trial with sample sizes varying from 30 to 80. We apply the standard binary and augmented binary methods and determine the power, type I error rate, coverage and average confidence interval width for each of the estimators. We implement Firth's adjustment for the binary component models and a small sample variance correction for the generalized estimating equations, applying the small sample adjusted methods to each sub-sample as before for comparison. RESULTS For the log-odds treatment effect the power of the augmented binary method is 20-55% compared to 12-20% for the standard binary method. Both methods have approximately nominal type I error rates. The difference in response probabilities exhibit similar power but both unadjusted methods demonstrate type I error rates of 6-8%. The small sample corrected methods have approximately nominal type I error rates. On both scales, the reduction in average confidence interval width when using the adjusted augmented binary method is 17-18%. This is equivalent to requiring a 32% smaller sample size to achieve the same statistical power. CONCLUSIONS The augmented binary method with small sample corrections provides a substantial improvement for rare disease trials using composite endpoints. We recommend the use of the method for the primary analysis in relevant rare disease trials. We emphasise that the method should be used alongside other efforts in improving the quality of evidence generated from rare disease trials rather than replace them.
Collapse
|
44
|
Abstract
Adaptive designs can make clinical trials more flexible by utilising results accumulating in the trial to modify the trial's course in accordance with pre-specified rules. Trials with an adaptive design are often more efficient, informative and ethical than trials with a traditional fixed design since they often make better use of resources such as time and money, and might require fewer participants. Adaptive designs can be applied across all phases of clinical research, from early-phase dose escalation to confirmatory trials. The pace of the uptake of adaptive designs in clinical research, however, has remained well behind that of the statistical literature introducing new methods and highlighting their potential advantages. We speculate that one factor contributing to this is that the full range of adaptations available to trial designs, as well as their goals, advantages and limitations, remains unfamiliar to many parts of the clinical community. Additionally, the term adaptive design has been misleadingly used as an all-encompassing label to refer to certain methods that could be deemed controversial or that have been inadequately implemented.We believe that even if the planning and analysis of a trial is undertaken by an expert statistician, it is essential that the investigators understand the implications of using an adaptive design, for example, what the practical challenges are, what can (and cannot) be inferred from the results of such a trial, and how to report and communicate the results. This tutorial paper provides guidance on key aspects of adaptive designs that are relevant to clinical triallists. We explain the basic rationale behind adaptive designs, clarify ambiguous terminology and summarise the utility and pitfalls of adaptive designs. We discuss practical aspects around funding, ethical approval, treatment supply and communication with stakeholders and trial participants. Our focus, however, is on the interpretation and reporting of results from adaptive design trials, which we consider vital for anyone involved in medical research. We emphasise the general principles of transparency and reproducibility and suggest how best to put them into practice.
Collapse
|
45
|
Stepped wedge cluster randomized controlled trial designs: a review of reporting quality and design features. Trials 2017; 18:33. [PMID: 28109321 PMCID: PMC5251280 DOI: 10.1186/s13063-017-1783-0] [Citation(s) in RCA: 44] [Impact Index Per Article: 6.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/29/2016] [Accepted: 01/03/2017] [Indexed: 11/13/2022] Open
Abstract
Background The stepped wedge (SW) cluster randomized controlled trial (CRCT) design is being used with increasing frequency. However, there is limited published research on the quality of reporting of SW-CRCTs. We address this issue by conducting a literature review. Methods Medline, Ovid, Web of Knowledge, the Cochrane Library, PsycINFO, the ISRCTN registry, and ClinicalTrials.gov were searched to identify investigations employing the SW-CRCT design up to February 2015. For each included completed study, information was extracted on a selection of criteria, based on the CONSORT extension to CRCTs, to assess the quality of reporting. Results A total of 123 studies were included in our review, of which 39 were completed trial reports. The standard of reporting of SW-CRCTs varied in quality. The percentage of trials reporting each criterion varied to as low as 15.4%, with a median of 66.7%. Conclusions There is much room for improvement in the quality of reporting of SW-CRCTs. This is consistent with recent findings for CRCTs. A CONSORT extension for SW-CRCTs is warranted to standardize the reporting of SW-CRCTs. Electronic supplementary material The online version of this article (doi:10.1186/s13063-017-1783-0) contains supplementary material, which is available to authorized users.
Collapse
|
46
|
Improving the power of clinical trials of rheumatoid arthritis by using data on continuous scales when analysing response rates: an application of the augmented binary method. Rheumatology (Oxford) 2016; 55:1796-802. [PMID: 27338084 PMCID: PMC5034221 DOI: 10.1093/rheumatology/kew263] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/02/2015] [Indexed: 12/22/2022] Open
Abstract
Objective. In clinical trials of RA, it is common to assess effectiveness using end points based upon dichotomized continuous measures of disease activity, which classify patients as responders or non-responders. Although dichotomization generally loses statistical power, there are good clinical reasons to use these end points; for example, to allow for patients receiving rescue therapy to be assigned as non-responders. We adopt a statistical technique called the augmented binary method to make better use of the information provided by these continuous measures and account for how close patients were to being responders. Methods. We adapted the augmented binary method for use in RA clinical trials. We used a previously published randomized controlled trial (Oral SyK Inhibition in Rheumatoid Arthritis-1) to assess its performance in comparison to a standard method treating patients purely as responders or non-responders. The power and error rate were investigated by sampling from this study. Results. The augmented binary method reached similar conclusions to standard analysis methods but was able to estimate the difference in response rates to a higher degree of precision. Results suggested that CI widths for ACR responder end points could be reduced by at least 15%, which could equate to reducing the sample size of a study by 29% to achieve the same statistical power. For other end points, the gain was even higher. Type I error rates were not inflated. Conclusion. The augmented binary method shows considerable promise for RA trials, making more efficient use of patient data whilst still reporting outcomes in terms of recognized response end points.
Collapse
|
47
|
A Bayesian adaptive design for biomarker trials with linked treatments. Br J Cancer 2015; 113:699-705. [PMID: 26263479 PMCID: PMC4559835 DOI: 10.1038/bjc.2015.278] [Citation(s) in RCA: 24] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/03/2015] [Revised: 06/15/2015] [Accepted: 07/02/2015] [Indexed: 11/09/2022] Open
Abstract
BACKGROUND Response to treatments is highly heterogeneous in cancer. Increased availability of biomarkers and targeted treatments has led to the need for trial designs that efficiently test new treatments in biomarker-stratified patient subgroups. METHODS We propose a novel Bayesian adaptive randomisation (BAR) design for use in multi-arm phase II trials where biomarkers exist that are potentially predictive of a linked treatment's effect. The design is motivated in part by two phase II trials that are currently in development. The design starts by randomising patients to the control treatment or to experimental treatments that the biomarker profile suggests should be active. At interim analyses, data from treated patients are used to update the allocation probabilities. If the linked treatments are effective, the allocation remains high; if ineffective, the allocation changes over the course of the trial to unlinked treatments that are more effective. RESULTS Our proposed design has high power to detect treatment effects if the pairings of treatment with biomarker are correct, but also performs well when alternative pairings are true. The design is consistently more powerful than parallel-groups stratified trials. CONCLUSIONS This BAR design is a powerful approach to use when there are pairings of biomarkers with treatments available for testing simultaneously.
Collapse
|
48
|
The power of phase II end-points for different possible mechanisms of action of an experimental treatment. Eur J Cancer 2015; 51:984-92. [PMID: 25840669 PMCID: PMC4435668 DOI: 10.1016/j.ejca.2015.03.002] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/30/2015] [Revised: 03/02/2015] [Accepted: 03/04/2015] [Indexed: 11/16/2022]
Abstract
BACKGROUND The high failure rate in phase III oncology trials is partly because the signal obtained from phase II trials is often weak. Several papers have considered the appropriateness of various phase II end-points for individual trials, but there has not been a systematic comparison using simulated data to determine which end-point should be used in which situation. METHODS In this paper we carry out simulation studies to compare the power of several Response Evaluation Criteria in Solid Tumours (RECIST) response-based end-points for one-arm and two-arm trials, together with progression-free survival (PFS) and testing the tumour-shrinkage directly for two-arm trials. We consider six scenarios: (1) short-term cytotoxic therapy; (2) continuous cytotoxic therapy; (3+4) cytostatic therapy; (5+6) delayed tumour-shrinkage effect (seen in some immunotherapies). We also consider measurement error in the assessment of tumour size. RESULTS Measurement error affects the type-I error rate and power of single-arm trials, and the power of two-arm trials. Generally no single end-point performed well in all scenarios. Best observed response rate, PFS and directly testing the tumour-shrinkages performed best for a number of scenarios. PFS performed very poorly when the effect of the treatment was short-lived. In scenario 6, where the delay in effect was long, no end-point performed well. CONCLUSIONS A clinician setting up a phase II trial should consider the likely mechanism of action the drug will have and choose an end-point that provides high power for that scenario. Testing the difference in tumour-shrinkage is often powerful. Alternative end-points are required for therapies with a long delayed effect.
Collapse
|
49
|
|
50
|
Correcting for multiple-testing in multi-arm trials: is it necessary and is it done? Trials 2014; 15:364. [PMID: 25230772 PMCID: PMC4177585 DOI: 10.1186/1745-6215-15-364] [Citation(s) in RCA: 97] [Impact Index Per Article: 9.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/25/2014] [Accepted: 09/08/2014] [Indexed: 11/21/2022] Open
Abstract
Background Multi-arm trials enable the evaluation of multiple treatments within a single trial. They provide a way of substantially increasing the efficiency of the clinical development process. However, since multi-arm trials test multiple hypotheses, some regulators require that a statistical correction be made to control the chance of making a type-1 error (false-positive). Several conflicting viewpoints are expressed in the literature regarding the circumstances in which a multiple-testing correction should be used. In this article we discuss these conflicting viewpoints and review the frequency with which correction methods are currently used in practice. Methods We identified all multi-arm clinical trials published in 2012 by four major medical journals. Summary data on several aspects of the trial design were extracted, including whether the trial was exploratory or confirmatory, whether a multiple-testing correction was applied and, if one was used, what type it was. Results We found that almost half (49%) of published multi-arm trials report using a multiple-testing correction. The percentage that corrected was higher for trials in which the experimental arms included multiple doses or regimens of the same treatments (67%). The percentage that corrected was higher in exploratory than confirmatory trials, although this is explained by a greater proportion of exploratory trials testing multiple doses and regimens of the same treatment. Conclusions A sizeable proportion of published multi-arm trials do not correct for multiple-testing. Clearer guidance about whether multiple-testing correction is needed for multi-arm trials that test separate treatments against a common control group is required. Electronic supplementary material The online version of this article (doi:10.1186/1745-6215-15-364) contains supplementary material, which is available to authorized users.
Collapse
|