Binary endpoint

  • 文章类型: Journal Article
    背景:组顺序设计结合了在中期分析的时间点停止徒劳的选项,可以节省时间和资源。因此,无用边界的选择对设计产生的性能特性有重要影响,包括正确或错误地停止徒劳的权力和可能性。几位作者为选择有益的徒劳界限这一主题做出了贡献。对于二进制端点,西蒙的设计(对照试验10:1-10,1989年)是常用的两阶段设计,用于单臂第二阶段研究,包括徒劳停止。然而,西蒙的最优设计经常在第一阶段后错误地宣布无用的可能性很高,在西蒙的minimax设计中,通常在中期分析中已经评估了计划样本大小的高比例,在早期停止的情况下仅留下有限的好处。
    方法:这项工作的重点是Schüler等人引入的最优性标准。(BMCMedResMethodol17:119,2017),并将其方法扩展到单臂II期研究中的二元终点。介绍了一种推导优化无用边界的算法,和研究设计实现这个概念的最佳无用边界的性能进行比较,常见的西蒙的minimax和最优设计,以及Kim等人对这些设计的修改版本。(Oncotarget10:4255-61,2019年)。
    结果:引入的优化的无用边界旨在最大限度地提高在小的或相反的影响的情况下正确停止无用的概率,同时也设置对中期分析的时间点的约束,功率损耗,以及错误停止研究的可能性,即停止研究,即使治疗效果显示出希望。总的来说,操作特性,如最大样本量和预期样本量,与经典和修改后的西蒙设计相当,有时更好。不像西蒙的设计,有约束力的停止规则,此处提出的优化的无用性边界未进行调整以耗尽全部目标标称显著性水平,因此对于非约束性应用仍然有效。
    结论:无效边界的选择和中期分析的时间点对研究设计的性质有重大影响。因此,他们应该在规划阶段进行彻底调查。引入的选择最佳无效边界的方法为Simon的设计提供了更灵活的替代方案,该方案具有无约束力的停止规则。错误停止无效的可能性被最小化,并且优化的无效边界没有表现出错误宣布无效的不期望的高概率或在临时时间点评估的计划样本的高比例的不利特性。
    BACKGROUND: Group sequential designs incorporating the option to stop for futility at the time point of an interim analysis can save time and resources. Thereby, the choice of the futility boundary importantly impacts the design\'s resulting performance characteristics, including the power and probability to correctly or wrongly stop for futility. Several authors contributed to the topic of selecting good futility boundaries. For binary endpoints, Simon\'s designs (Control Clin Trials 10:1-10, 1989) are commonly used two-stage designs for single-arm phase II studies incorporating futility stopping. However, Simon\'s optimal design frequently yields an undesirably high probability of falsely declaring futility after the first stage, and in Simon\'s minimax design often a high proportion of the planned sample size is already evaluated at the interim analysis leaving only limited benefit in case of an early stop.
    METHODS: This work focuses on the optimality criteria introduced by Schüler et al. (BMC Med Res Methodol 17:119, 2017) and extends their approach to binary endpoints in single-arm phase II studies. An algorithm for deriving optimized futility boundaries is introduced, and the performance of study designs implementing this concept of optimal futility boundaries is compared to the common Simon\'s minimax and optimal designs, as well as modified versions of these designs by Kim et al. (Oncotarget 10:4255-61, 2019).
    RESULTS: The introduced optimized futility boundaries aim to maximize the probability of correctly stopping for futility in case of small or opposite effects while also setting constraints on the time point of the interim analysis, the power loss, and the probability of stopping the study wrongly, i.e. stopping the study even though the treatment effect shows promise. Overall, the operating characteristics, such as maximum sample size and expected sample size, are comparable to those of the classical and modified Simon\'s designs and sometimes better. Unlike Simon\'s designs, which have binding stopping rules, the optimized futility boundaries proposed here are not adjusted to exhaust the full targeted nominal significance level and are thus still valid for non-binding applications.
    CONCLUSIONS: The choice of the futility boundary and the time point of the interim analysis have a major impact on the properties of the study design. Therefore, they should be thoroughly investigated at the planning stage. The introduced method of selecting optimal futility boundaries provides a more flexible alternative to Simon\'s designs with non-binding stopping rules. The probability of wrongly stopping for futility is minimized and the optimized futility boundaries don\'t exhibit the unfavorable properties of an undesirably high probability of falsely declaring futility or a high proportion of the planned sample evaluated at the interim time point.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    背景:样本量计算是临床试验计划的核心方面。样本量是根据参数假设计算的,比如治疗效果和终点的方差。这种方法的基本问题是在试验之前不知道真实的分布参数。因此,样本量计算总是包含一定程度的不确定性,导致试验能力不足或规模过大的风险。应对这种不确定性的一种方法是自适应设计。自适应设计允许在中期分析期间调整样本大小。有大量这样的重新计算规则可供选择。为了指导选择合适的自适应设计,并重新计算样本大小,以前的文献表明,对于具有正态分布终点的研究,有条件的表现得分。然而,在临床试验中也经常使用二元终点,并且尚未研究条件性能评分对二元终点的应用。
    方法:我们通过提出相关的一维分数参数化,将条件性能分数的理论扩展到二元端点。此外,我们还进行了仿真研究,以评估运行特性并说明应用。
    结果:我们发现分数定义可以扩展而无需修改二进制端点的情况。我们用单个分布参数表示评分结果,因此得出一个单一的效应测度,其中包含干预组和对照组之间的比例差异[公式:见文本],以及对照组中的终点比例[公式:见正文]。
    结论:本研究将条件性能得分的理论扩展到二元端点,并证明了其在实践中的应用。
    BACKGROUND: Sample size calculation is a central aspect in planning of clinical trials. The sample size is calculated based on parameter assumptions, like the treatment effect and the endpoint\'s variance. A fundamental problem of this approach is that the true distribution parameters are not known before the trial. Hence, sample size calculation always contains a certain degree of uncertainty, leading to the risk of underpowering or oversizing a trial. One way to cope with this uncertainty are adaptive designs. Adaptive designs allow to adjust the sample size during an interim analysis. There is a large number of such recalculation rules to choose from. To guide the choice of a suitable adaptive design with sample size recalculation, previous literature suggests a conditional performance score for studies with a normally distributed endpoint. However, binary endpoints are also frequently applied in clinical trials and the application of the conditional performance score to binary endpoints is not yet investigated.
    METHODS: We extend the theory of the conditional performance score to binary endpoints by suggesting a related one-dimensional score parametrization. We moreover perform a simulation study to evaluate the operational characteristics and to illustrate application.
    RESULTS: We find that the score definition can be extended without modification to the case of binary endpoints. We represent the score results by a single distribution parameter, and therefore derive a single effect measure, which contains the difference in proportions [Formula: see text] between the intervention and the control group, as well as the endpoint proportion [Formula: see text] in the control group.
    CONCLUSIONS: This research extends the theory of the conditional performance score to binary endpoints and demonstrates its application in practice.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    近年来,自适应随机化方法在临床研究和试验设计中获得了显著的普及,因为它们能够在调整正在进行的临床试验的统计程序方面提供效率和灵活性.对于一项比较多种治疗的研究,多臂两阶段设计可用于从第一阶段中选择最佳治疗,并进一步将该治疗与第二阶段的对照进行比较。传统设计在两个阶段都使用相等的随机化。为了更好地利用第一阶段的中期结果,我们建议为具有二元结局的多组临床试验开发响应自适应随机化两阶段设计.考虑了两种分配方法:(1)基于顺序设计的最佳分配;(2)赢家规则。在三个标准下获得了最佳的多臂两阶段设计:最小化预期的故障数量,最小化平均预期样本量,并最小化零假设下的预期样本量。仿真研究表明,基于获胜者规则的自适应设计具有良好的性能。针对胰腺腺癌和种系BRCA/PALB2突变患者的II期试验用于说明所提出的反应自适应随机化设计的应用。
    In recent years, adaptive randomization methods have gained significant popularity in clinical research and trial design due to their ability to provide both efficiency and flexibility in adjusting the statistical procedures of ongoing clinical trials. For a study to compare multiple treatments, a multi-arm two-stage design could be utilized to select the best treatment from the first stage and further compare that treatment with control in the second stage. The traditional design used equal randomization in both stages. To better utilize the interim results from the first stage, we propose to develop response adaptive randomization two-stage designs for a multi-arm clinical trial with binary outcome. Two allocation methods are considered: (1) an optimal allocation based on a sequential design; (2) the play-the-winner rule. Optimal multi-arm two-stage designs are obtained under three criteria: minimizing the expected number of failures, minimizing the average expected sample size, and minimizing the expected sample size under the null hypothesis. Simulation studies show that the proposed adaptive design based on the play-the-winner rule has good performance. A phase II trial for patients with pancreas adenocarcinoma and a germline BRCA/PALB2 mutation was used to illustrate the application of the proposed response adaptive randomization designs.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    自适应无缝设计将II期和III期结合到单个试验中,已显示出对提高药物开发效率的兴趣与日俱增。成为最常见的自适应设计类型。它通常由两个阶段组成,每个阶段的审判目标往往不同。主要目的是在第一阶段中选择最佳的实验治疗组,并在第二阶段中比较所选择的治疗组和对照组之间的功效。在这篇文章中,我们专注于两阶段自适应无缝设计,其中治疗选择基于短期二元终点,治疗比较基于长期二元终点。因此,我们提出了一个精确的条件检验作为最终分析,基于双变量二项分布,并从中期分析中选择具有最有希望的短期终点反应率的治疗方法。此外,引入了中期p$p$$方法,以提高精确测试的保守性。进行了仿真研究,以将所提出的方法与基于组合测试的方法进行比较。所提出的精确方法在标称水平上控制I型错误率,无论初始治疗的数量或短期和长期终点之间的相关性。在治疗比较能力方面,所提出的方法比基于场景组合测试的方法更强大,只有一种治疗是有效的。
    The adaptive seamless design combining phases II and III into a single trial has been shown growing interest for improving the efficiency of drug development, becoming the most frequent adaptive design type. It typically consists of two stages, the trial objectives being often different in each stage. The primary objectives are to select optimal experimental treatment group(s) in the first stage and compare the efficacy between the selected treatment and control groups in the second stage. In this article, we focus on a two-stage adaptive seamless design, for which treatment selection is based on the short-term binary endpoint and treatment comparison is based on the long-term binary endpoint. We thus propose an exact conditional test as a final analysis, based on the bivariate binomial distribution and given the selected treatment with the most promising short-term endpoint response rate from an interim analysis. Additionally, the mid- p $$ p $$ approach is incorporated to improve conservativeness for an exact test. Simulation studies were conducted to compare the proposed methods with a method based on the combination test. The proposed exact method controlled for type I error rate at the nominal level, regardless of the number of initial treatments or the correlation between short- and long-term endpoints. In terms of the treatment comparison power, the proposed methods are more powerful than that based on the combination test in the scenarios, with only one treatment being effective.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    在药物开发过程中,生物标志物有时被确定为将患者人群分为从评估的治疗中获益较多和较少的人群。因此,后来的研究可能是有针对性的,而较早的是在混合患者人群中进行的。这对证据综合提出了挑战,特别是如果只有聚合数据可用。从这个场景开始,我们研究了三种常用的网络元分析估计方法,天真的估计方法,独立分析,和网络元回归。此外,我们调整和修改两种方法,用于证据综合,将随机对照试验与观察性研究相结合,通过加权富集法,和信息先验估计。我们使用偏差评估了32种情景的模拟研究中的所有五种方法,均方根误差,覆盖范围,精度,和权力。此外,我们重新访问临床数据集以举例说明和讨论应用。在模拟研究中,在所有调查的情况下,没有一种方法被观察到明显有利。然而,在所有评估的性能度量和模拟场景中,独立分析和天真的估计的性能与其他方法相当或更差,因此不建议使用。虽然实质性的试验间异质性对所有估计方法都是具有挑战性的,网络元回归的性能,通过加权方法和信息先验方法取决于模拟场景和感兴趣的性能度量。此外,由于这些估计方法得出的假设略有不同,其中一些需要额外的信息进行估计,我们建议尽可能进行敏感性分析。
    During drug development, a biomarker is sometimes identified as separating a patient population into those with more and those with less benefit from evaluated treatments. Consequently, later studies might be targeted, while earlier ones are performed in mixed patient populations. This poses a challenge in evidence synthesis, especially if only aggregated data are available. Starting from this scenario, we investigate three commonly used network meta-analytic estimation methods, the naive estimation approach, the stand-alone analysis, and the network meta-regression. Additionally, we adapt and modify two methods, which are used in evidence synthesis to combine randomized controlled trials with observational studies, the enrichment-through-weighting approach, and the informative prior estimation. We evaluate all five methods in a simulation study with 32 scenarios using bias, root-mean-squared-error, coverage, precision, and power. Additionally, we revisit a clinical data set to exemplify and discuss the application. In the simulation study, none of the methods was observed to be clearly favorable over all investigated scenarios. However, the stand-alone analysis and the naive estimation performed comparably or worse than the other methods in all evaluated performance measures and simulation scenarios and are therefore not recommended. While substantial between-trial heterogeneity is challenging for all estimation approaches, the performance of the network meta-regression, the enriching-through weighting approach and the informative prior approach was dependent on the simulation scenario and the performance measure of interest. Furthermore, as these estimation methods are drawing slightly different assumptions, some of which require the presence of additional information for estimation, we recommend sensitivity-analyses wherever possible.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    在II期肿瘤学试验中,经常使用两阶段设计,允许因徒劳和/或功效而提前停止。然而,这种基于频率统计方法的设计不能保证从贝叶斯的角度来看参加预先指定的临床感兴趣率的高后验概率.这里,我们提出了一种新的贝叶斯设计,使早期终止的功效以及徒劳。除了临床上无趣和有趣的反应率,响应率的先验分布,设计中规定了最小后验阈值概率和最高后验密度区间的长度.最后,我们定义了具有最高总有效预测概率的可行设计。我们研究了拟议设计的特性,并将其应用于肿瘤学试验作为示例。所提出的设计确保了观察到的反应率在预定的后验概率水平内。所提出的设计为单臂两阶段试验提供了替代设计。
    In phase II oncology trials, two-stage design allowing early stopping for futility and/or efficacy is frequently used. However, this design based on frequentist statistical approaches could not guarantee a high posterior probability of attending the pre-specified clinically interesting rate from a Bayesian perspective. Here, we proposed a new Bayesian design enabling early terminating for efficacy as well as futility. In addition to the clinically uninteresting and interesting response rate, a prior distribution of response rate, the minimum posterior threshold probabilities and the lengths of the highest posterior density intervals were specified in the design. Finally, we defined the feasible design with the highest total effective predictive probability. We studied the properties of the proposed design and applied it to an oncology trial as an example. The proposed design ensured that the observed response rate fell within prespecified levels of posterior probability. The proposed design provides an alternative design to single-arm two-stage trials.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    drop-the-losers设计结合了k治疗的2期试验和单一适应性协议下的验证性3期试验,从而获得比传统的临床开发方法的效率。这种设计在罕见疾病环境中可能特别有用,其中保存样本量至关重要,控制武器可能不可行。我们提出了一个无条件的精确似然(UEL)测试和推理程序,这些设计的二进制端点使用小样本大小,将其操作特性与现有方法进行比较。评估了其他实际考虑因素,包括逐级样本量的选择和领带的效果。
    The drop-the-losers design combines a phase 2 trial of k treatments and a confirmatory phase 3 trial under a single adaptive protocol, thereby gaining efficiency over a traditional clinical development approach. Such designs may be particularly useful in the rare disease setting, where conserving sample size is paramount, and control arms may not be feasible. We propose an unconditional exact likelihood (UEL) testing and inference procedure for these designs for a binary endpoint using small sample sizes, comparing its operating characteristics to existing methods. Additional practical considerations are evaluated, including the choice of stagewise sample sizes and effect of ties.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

  • 文章类型: Journal Article
    The Multiple Comparison Procedure - Modelling (MCP-Mod) method was qaulified by regulatory agencies (e.g., EMA in 2014 and FDA in 2016) as an efficient statistical method for Phase 2 dose-finding studies when there is uncertainty about dose-response relationship. As this is a relatively new approach, there is limited literature providing practical guidance on its application. In this paper, we evaluated the performance of the MCP-Mod method for clinical trials with a binary primary endpoint, focusing on (1) the impact of sample size, data variability and treatment effect size on the performance of the MCP-Mod, (2) the impact of candidate model mis-specification, and (3) optimal sample allocation under a fixed sample size. The evaluation was performed via simulations under different scenarios.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

       PDF(Pubmed)

  • 文章类型: Journal Article
    Phase II clinical trials make a critical decision of go or no-go to a subsequent phase III studies. A considerable proportion of promising drugs identified in phase II trials fail the confirmative efficacy test in phase III. Recognizing the low posterior probabilities of H1 when accepting the drug under Simon\'s two-stage design, the Bayesian enhancement two-stage (BET) design is proposed to strengthen the passing criterion. Under the BET design, the lengths of highest posterior density (HPD) intervals, posterior probabilities of H0 and H1 are computed to calibrate the design parameters, aiming to improve the stability of the trial characteristics and strengthen the evidence for proceeding the drug development forward. However, from a practical perspective, the HPD interval length lacks transparency and interpretability. To circumvent this problem, we propose the BET design with error control (BETEC) by replacing the HPD interval length with the posterior error rate. The BETEC design can achieve a balance between the posterior false positive rate and false negative rate and, more importantly, it has an intuitive and clear interpretation. We compare our method with the BET design and Simon\'s design through extensive simulation studies. As an illustration, we further apply BETEC to two recent clinical trials, and investigate its performance in comparison with other competitive designs. Being both efficient and intuitive, the BETEC design can serve as an alternative toolbox for implementing phase II single-arm trials.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

  • 文章类型: Journal Article
    Individualized therapies for patients with biomarkers are moving more and more into the focus of research interest when developing new treatments. Hereby, the term individualized (or targeted) therapy denotes a treatment specifically developed for biomarker-positive patients. A network meta-analysis model for a binary endpoint combining the evidence for a targeted therapy from individual patient data with the evidence for a non-targeted therapy from aggregate data is presented and investigated. The biomarker status of the patients is either available at patient-level in individual patient data or at study-level in aggregate data. Both types of biomarker information have to be included. The evidence synthesis model follows a Bayesian approach and applies a meta-regression to the studies with aggregate data. In a simulation study, we address three treatment arms, one of them investigating a targeted therapy. The bias and the root-mean-square error of the treatment effect estimate for the subgroup of biomarker-positive patients based on studies with aggregate data are investigated. Thereby, the meta-regression approach is compared to approaches applying alternative solutions. The regression approach has a surprisingly small bias even in the presence of few studies. By contrast, the root-mean-square error is relatively greater. An illustrative example is provided demonstrating implementation of the presented network meta-analysis model in a clinical setting.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

公众号