Pub Date : 2024-09-16DOI: 10.1080/10543406.2024.2398036
Jianrong Wu, Yimei Li, Liang Zhu, Tushar Patni
Traditional two-arm randomized trial designs have played a pivotal role in establishing the efficacy of medical interventions. However, their efficiency is often compromised when confronted with multiple experimental treatments or limited resources. In response to these challenges, the multi-arm multi-stage designs have emerged, enabling the simultaneous evaluation of multiple treatments within a single trial. In such an approach, if an arm meets efficacy success criteria at an interim stage, the whole trial stops and the arm is selected for further study. However when multiple treatment arms are active, stopping the trial at the moment one arm achieves success diminishes the probability of selecting the best arm. To address this issue, we have developed a group sequential multi-arm multi-stage survival trial design with an arm-specific stopping rule. The proposed method controls the familywise type I error in a strong sense and selects the best promising treatment arm with a high probability.
传统的双臂随机试验设计在确定医疗干预措施的疗效方面发挥了关键作用。然而,当面临多种实验治疗或资源有限时,其效率往往会大打折扣。为应对这些挑战,多臂多阶段设计应运而生,可在一次试验中同时评估多种治疗方法。在这种方法中,如果某一治疗臂在中期阶段达到疗效成功标准,整个试验就会停止,并选择该治疗臂进行进一步研究。然而,当多个治疗方案同时进行时,如果在一个治疗方案取得成功时停止试验,就会降低选择最佳治疗方案的概率。为了解决这个问题,我们开发了一种分组顺序多臂多阶段生存试验设计,其中包含针对特定臂的停止规则。所提出的方法能有效控制族式 I 型误差,并能高概率地选择最佳治疗臂。
{"title":"Multi-arm multi-stage survival trial design with arm-specific stopping rule.","authors":"Jianrong Wu, Yimei Li, Liang Zhu, Tushar Patni","doi":"10.1080/10543406.2024.2398036","DOIUrl":"https://doi.org/10.1080/10543406.2024.2398036","url":null,"abstract":"<p><p>Traditional two-arm randomized trial designs have played a pivotal role in establishing the efficacy of medical interventions. However, their efficiency is often compromised when confronted with multiple experimental treatments or limited resources. In response to these challenges, the multi-arm multi-stage designs have emerged, enabling the simultaneous evaluation of multiple treatments within a single trial. In such an approach, if an arm meets efficacy success criteria at an interim stage, the whole trial stops and the arm is selected for further study. However when multiple treatment arms are active, stopping the trial at the moment one arm achieves success diminishes the probability of selecting the best arm. To address this issue, we have developed a group sequential multi-arm multi-stage survival trial design with an arm-specific stopping rule. The proposed method controls the familywise type I error in a strong sense and selects the best promising treatment arm with a high probability.</p>","PeriodicalId":54870,"journal":{"name":"Journal of Biopharmaceutical Statistics","volume":" ","pages":"1-12"},"PeriodicalIF":1.2,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142301312","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-11DOI: 10.1080/10543406.2024.2395532
Benjamin Duncan
Proper and timely characterization of the safety profile of a pharmaceutical product under development is imperative for assessing the overall benefit-risk relationship of the product and for making key development decisions. For ongoing clinical development, a comprehensive and robust safety monitoring and safety signal detection program which is based upon quantitative statistical reasoning is critical. Methods presented here can be applied to safety signal detection and periodic safety monitoring. Various statistical properties, distributions, and models, all utilizing a Bayesian framework are considered and further examined in order to identify robust methods applicable to a broad set of scenarios and situations. Methods developed for incidence counts (including those with under-dispersed distributions) with variable time-at-risk and with underlying constant or non-constant hazard rates, are proposed and compared to traditional methods designed to assess adverse event incidence rates or binomial incidence proportions (which assume an underlying constant hazard rate and subsequent Poisson distribution for modeling event counts).
{"title":"Robust safety monitoring and signal detection using alternatives to the standard poisson distribution.","authors":"Benjamin Duncan","doi":"10.1080/10543406.2024.2395532","DOIUrl":"https://doi.org/10.1080/10543406.2024.2395532","url":null,"abstract":"Proper and timely characterization of the safety profile of a pharmaceutical product under development is imperative for assessing the overall benefit-risk relationship of the product and for making key development decisions. For ongoing clinical development, a comprehensive and robust safety monitoring and safety signal detection program which is based upon quantitative statistical reasoning is critical. Methods presented here can be applied to safety signal detection and periodic safety monitoring. Various statistical properties, distributions, and models, all utilizing a Bayesian framework are considered and further examined in order to identify robust methods applicable to a broad set of scenarios and situations. Methods developed for incidence counts (including those with under-dispersed distributions) with variable time-at-risk and with underlying constant or non-constant hazard rates, are proposed and compared to traditional methods designed to assess adverse event incidence rates or binomial incidence proportions (which assume an underlying constant hazard rate and subsequent Poisson distribution for modeling event counts).","PeriodicalId":54870,"journal":{"name":"Journal of Biopharmaceutical Statistics","volume":"1 1","pages":"1-18"},"PeriodicalIF":1.1,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142178780","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-31DOI: 10.1080/10543406.2024.2395548
Belmiro P M Duarte, Anthony C Atkinson
We study optimal designs for clinical trials when the value of the response and its variance depend on treatment and covariates are included in the response model. Such designs are generalizations of Neyman allocation, commonly used in personalized medicine when external factors may have differing effects on the response depending on subgroups of patients. We develop theoretical results for D-, A-, E- and D-optimal designs and construct semidefinite programming (SDP) formulations that support their numerical computation. D-, A-, and E-optimal designs are appropriate for efficient estimation of distinct properties of the parameters of the response models. Our formulation allows finding optimal allocation schemes for a general number of treatments and of covariates. Finally, we study frequentist sequential clinical trial allocation within contexts where response parameters and their respective variances remain unknown. We illustrate, with a simulated example and with a redesigned clinical trial on the treatment of neuro-degenerative disease, that both theoretical and SDP results, derived under the assumption of known variances, converge asymptotically to allocations obtained through the sequential scheme. Procedures to use static and sequential allocation are proposed.
我们研究了当反应值及其方差取决于治疗方法,且反应模型中包含协变量时临床试验的最佳设计。这种设计是奈曼分配的一般化,常用于外部因素可能因患者亚群的不同而对反应产生不同影响的个性化医疗中。我们提出了 D-、A-、E- 和 D A-最优设计的理论结果,并构建了支持其数值计算的半有限编程(SDP)公式。D-、A- 和 E-最优设计适用于有效估计响应模型参数的不同属性。我们的计算公式允许为一般数量的处理和协变量找到最优分配方案。最后,我们研究了在反应参数及其各自方差未知的情况下的频数主义序贯临床试验分配。我们通过一个模拟例子和一个重新设计的治疗神经退行性疾病的临床试验来说明,在已知方差的假设条件下得出的理论结果和 SDP 结果都会渐进地趋近于通过顺序方案得到的分配结果。提出了使用静态和顺序分配的程序。
{"title":"Optimum designs for clinical trials in personalized medicine when response variance depends on treatment.","authors":"Belmiro P M Duarte, Anthony C Atkinson","doi":"10.1080/10543406.2024.2395548","DOIUrl":"https://doi.org/10.1080/10543406.2024.2395548","url":null,"abstract":"<p><p>We study optimal designs for clinical trials when the value of the response and its variance depend on treatment and covariates are included in the response model. Such designs are generalizations of Neyman allocation, commonly used in personalized medicine when external factors may have differing effects on the response depending on subgroups of patients. We develop theoretical results for D-, A-, E- and D<math><msub><mi> </mi><mrow><mrow><mi>A</mi></mrow></mrow></msub></math>-optimal designs and construct semidefinite programming (SDP) formulations that support their numerical computation. D-, A-, and E-optimal designs are appropriate for efficient estimation of distinct properties of the parameters of the response models. Our formulation allows finding optimal allocation schemes for a general number of treatments and of covariates. Finally, we study frequentist sequential clinical trial allocation within contexts where response parameters and their respective variances remain unknown. We illustrate, with a simulated example and with a redesigned clinical trial on the treatment of neuro-degenerative disease, that both theoretical and SDP results, derived under the assumption of known variances, converge asymptotically to allocations obtained through the sequential scheme. Procedures to use static and sequential allocation are proposed.</p>","PeriodicalId":54870,"journal":{"name":"Journal of Biopharmaceutical Statistics","volume":" ","pages":"1-18"},"PeriodicalIF":1.2,"publicationDate":"2024-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142114910","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A non-inferiority trial is usually conducted to investigate whether a new drug/treatment is no worse than a reference drug/treatment by a small, pre-specified, non-inferiority margin. This study aimed to assess the non-inferiority of the difference between two binary-outcome treatments in a matched-pairs design based on the method of variance of estimates recovery (MOVER). The processes for estimating the confidence interval of a single proportion included in the MOVER are the Wilson score interval, Agresti - Coull interval, Jeffreys interval, modified Jeffreys interval, score method with continuity correction, and arcsin interval. The performance of the six MOVER tests, the fiducial test, and the restricted maximum likelihood estimation test were evaluated by comparing their type I error rates and power at different pre-assigned levels and with varying combinations of parameters. The evaluation results showed that the modified Jeffreys MOVER test can be a competitive alternative to the other recommended tests. It can control type I errors well, and its power is not inferior to other methods. The proposed tests were illustrated with three real-world examples.
非劣效性试验通常是为了研究一种新药/治疗方法与参考药物/治疗方法相比,是否有很小的、预先指定的非劣效性差值。本研究旨在根据估计值恢复方差法(MOVER),在配对设计中评估两种二元结果治疗之间差异的非劣效性。MOVER 中包含的估算单一比例置信区间的方法有 Wilson 评分区间、Agresti - Coull 区间、Jeffreys 区间、修正 Jeffreys 区间、带连续性校正的评分法和 arcsin 区间。通过比较不同预设水平和不同参数组合下的 I 类错误率和功率,评估了六种 MOVER 检验、fiducial 检验和受限最大似然估计检验的性能。评估结果表明,修改后的 Jeffreys MOVER 检验可以替代其他推荐检验。它能很好地控制 I 型误差,其功率也不比其他方法差。我们用三个实际案例对所提出的检验方法进行了说明。
{"title":"MOVER tests for non-inferiority of the difference between two binary-outcome treatments in the matched-pairs design.","authors":"Liangchang Xiu, Linlin Xie, Haiyi Yan, Chunxin Wu, Huansheng Liu, Chao Chen","doi":"10.1080/10543406.2024.2390888","DOIUrl":"https://doi.org/10.1080/10543406.2024.2390888","url":null,"abstract":"<p><p>A non-inferiority trial is usually conducted to investigate whether a new drug/treatment is no worse than a reference drug/treatment by a small, pre-specified, non-inferiority margin. This study aimed to assess the non-inferiority of the difference between two binary-outcome treatments in a matched-pairs design based on the method of variance of estimates recovery (MOVER). The processes for estimating the confidence interval of a single proportion included in the MOVER are the Wilson score interval, Agresti - Coull interval, Jeffreys interval, modified Jeffreys interval, score method with continuity correction, and arcsin interval. The performance of the six MOVER tests, the fiducial test, and the restricted maximum likelihood estimation test were evaluated by comparing their type I error rates and power at different pre-assigned levels and with varying combinations of parameters. The evaluation results showed that the modified Jeffreys MOVER test can be a competitive alternative to the other recommended tests. It can control type I errors well, and its power is not inferior to other methods. The proposed tests were illustrated with three real-world examples.</p>","PeriodicalId":54870,"journal":{"name":"Journal of Biopharmaceutical Statistics","volume":" ","pages":"1-14"},"PeriodicalIF":1.2,"publicationDate":"2024-08-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142114909","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-11DOI: 10.1080/10543406.2024.2387364
Qiqi Deng, Lili Zhu, Brendan Weiss, Praveen Aanur, Lei Gao
Dose optimization is a critical challenge in drug development. Historically, dose determination in oncology has followed a divergent path from other non-oncology therapeutic areas due to the unique characteristics and requirements in Oncology. However, with the emergence of new drug modalities and mechanisms of drugs in oncology, such as immune therapies, radiopharmaceuticals, targeted therapies, cytostatic agents, and others, the dose-response relationship for efficacy and toxicity could be vastly varied compared to the cytotoxic chemotherapies. The doses below the MTD may demonstrate similar efficacy to the MTD with an improved tolerability profile, resembling what is commonly observed in non-oncology treatments. Hence, alternate strategies for dose optimization are required for new modalities in oncology drug development. This paper delves into the historical evolution of dose finding methods from non-oncology to oncology, highlighting examples and summarizing the underlying drivers of change. Subsequently, a practical framework and guidance are provided to illustrate how dose optimization can be incorporated into various stages of the development program. We provide the following general recommendations: 1) The objective for phase I is to identify a dose range rather than a single MTD dose for subsequent development to better characterize the safety and tolerability profile within the dose range. 2) At least two doses separable by PK are recommended for dose optimization in phase II. 3) Ideally, dose optimization should be performed before launching the confirmatory study. Nevertheless, innovative designs such as seamless II/III design can be implemented for dose selection and may accelerate the drug development program.
剂量优化是药物开发中的一项重要挑战。从历史上看,由于肿瘤学的独特性和要求,肿瘤学的剂量确定一直与其他非肿瘤学治疗领域不同。然而,随着免疫疗法、放射性药物、靶向疗法、细胞抑制剂等新的药物模式和药物机制在肿瘤学中的出现,与细胞毒性化疗相比,疗效和毒性的剂量反应关系可能会有很大的不同。低于MTD的剂量可能具有与MTD相似的疗效,但耐受性有所改善,这与非肿瘤治疗中常见的情况类似。因此,在肿瘤药物开发的新模式中,需要有剂量优化的替代策略。本文深入探讨了从非肿瘤学到肿瘤学的剂量寻找方法的历史演变,重点举例说明并总结了变化的根本原因。随后,本文提供了一个实用框架和指南,说明如何将剂量优化纳入开发计划的各个阶段。我们提出以下一般性建议:1) I 期的目标是为后续开发确定一个剂量范围,而不是单一的 MTD 剂量,以便更好地描述剂量范围内的安全性和耐受性特征。2)建议在 II 期进行剂量优化时至少使用两个可通过 PK 分离的剂量。3) 理想情况下,剂量优化应在启动确证研究之前进行。然而,创新设计(如无缝 II/III 设计)可用于剂量选择,并可加快药物开发计划。
{"title":"Strategies for successful dose optimization in oncology drug development: a practical guide.","authors":"Qiqi Deng, Lili Zhu, Brendan Weiss, Praveen Aanur, Lei Gao","doi":"10.1080/10543406.2024.2387364","DOIUrl":"https://doi.org/10.1080/10543406.2024.2387364","url":null,"abstract":"<p><p>Dose optimization is a critical challenge in drug development. Historically, dose determination in oncology has followed a divergent path from other non-oncology therapeutic areas due to the unique characteristics and requirements in Oncology. However, with the emergence of new drug modalities and mechanisms of drugs in oncology, such as immune therapies, radiopharmaceuticals, targeted therapies, cytostatic agents, and others, the dose-response relationship for efficacy and toxicity could be vastly varied compared to the cytotoxic chemotherapies. The doses below the MTD may demonstrate similar efficacy to the MTD with an improved tolerability profile, resembling what is commonly observed in non-oncology treatments. Hence, alternate strategies for dose optimization are required for new modalities in oncology drug development. This paper delves into the historical evolution of dose finding methods from non-oncology to oncology, highlighting examples and summarizing the underlying drivers of change. Subsequently, a practical framework and guidance are provided to illustrate how dose optimization can be incorporated into various stages of the development program. We provide the following general recommendations: 1) The objective for phase I is to identify a dose range rather than a single MTD dose for subsequent development to better characterize the safety and tolerability profile within the dose range. 2) At least two doses separable by PK are recommended for dose optimization in phase II. 3) Ideally, dose optimization should be performed before launching the confirmatory study. Nevertheless, innovative designs such as seamless II/III design can be implemented for dose selection and may accelerate the drug development program.</p>","PeriodicalId":54870,"journal":{"name":"Journal of Biopharmaceutical Statistics","volume":" ","pages":"1-15"},"PeriodicalIF":1.2,"publicationDate":"2024-08-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141914624","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01Epub Date: 2023-07-26DOI: 10.1080/10543406.2023.2236218
Yiqi Zhao, Ping Yan, Xinfeng Yang
In this paper, we aim to show the process of simulating survival data when the distribution of the overall population and one subgroup (called "positive subgroup") as well as the proportion of the subgroup is known, while the distribution of the other subgroup (called "negative subgroup") is unknown. We propose a combination method which generates survival data of the positive subgroup and negative subgroup, respectively, and survival data of the overall population are the combination of the two subgroups. The parameters of the overall population and the positive subgroup need to satisfy certain constraints, otherwise the parameters may lead to contradictions. From simulation, we show that our proposed combination method can reflect the correlation between the test statistics of overall population and positive subgroup, which makes the simulated data more realistic and the results of simulation more reliable. Moreover, for a multiplicity control in trial design, the combination method can help to determine the splitting strategy between primary endpoints, and is helpful in designs of clinical trials as shown in three applications.
{"title":"Simulating survival data when one subgroup lacks information.","authors":"Yiqi Zhao, Ping Yan, Xinfeng Yang","doi":"10.1080/10543406.2023.2236218","DOIUrl":"10.1080/10543406.2023.2236218","url":null,"abstract":"<p><p>In this paper, we aim to show the process of simulating survival data when the distribution of the overall population and one subgroup (called \"positive subgroup\") as well as the proportion of the subgroup is known, while the distribution of the other subgroup (called \"negative subgroup\") is unknown. We propose a combination method which generates survival data of the positive subgroup and negative subgroup, respectively, and survival data of the overall population are the combination of the two subgroups. The parameters of the overall population and the positive subgroup need to satisfy certain constraints, otherwise the parameters may lead to contradictions. From simulation, we show that our proposed combination method can reflect the correlation between the test statistics of overall population and positive subgroup, which makes the simulated data more realistic and the results of simulation more reliable. Moreover, for a multiplicity control in trial design, the combination method can help to determine the <math><mrow><mrow><mi>α</mi></mrow></mrow></math> splitting strategy between primary endpoints, and is helpful in designs of clinical trials as shown in three applications.</p>","PeriodicalId":54870,"journal":{"name":"Journal of Biopharmaceutical Statistics","volume":" ","pages":"613-625"},"PeriodicalIF":1.2,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9873752","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01Epub Date: 2023-09-07DOI: 10.1080/10543406.2023.2251578
Linxi Han, Qiqi Deng, Zhangyi He, Frank Fleischer, Feng Yu
The Multiple Comparison Procedure and Modelling (MCPMod) approach has been shown to be a powerful statistical technique that can significantly improve the design and analysis of dose-finding studies under model uncertainty. Due to its frequentist nature, however, it is difficult to incorporate information into MCPMod from historical trials on the same drug. BMCPMod, a recently introduced Bayesian version of MCPMod, is designed to take into account historical information on the placebo dose group. We introduce a Bayesian hierarchical framework capable of incorporating historical information on an arbitrary number of dose groups, including both placebo and active ones, taking into account the relationship between responses of these dose groups. Our approach can also model both prognostic and predictive between-trial heterogeneity and is particularly useful in situations where the effect sizes of two trials are different. Our goal is to reduce the necessary sample size in the dose-finding trial while maintaining its target power.
{"title":"Bayesian hierarchical model for dose-finding trial incorporating historical data.","authors":"Linxi Han, Qiqi Deng, Zhangyi He, Frank Fleischer, Feng Yu","doi":"10.1080/10543406.2023.2251578","DOIUrl":"10.1080/10543406.2023.2251578","url":null,"abstract":"<p><p>The Multiple Comparison Procedure and Modelling (MCPMod) approach has been shown to be a powerful statistical technique that can significantly improve the design and analysis of dose-finding studies under model uncertainty. Due to its frequentist nature, however, it is difficult to incorporate information into MCPMod from historical trials on the same drug. BMCPMod, a recently introduced Bayesian version of MCPMod, is designed to take into account historical information on the placebo dose group. We introduce a Bayesian hierarchical framework capable of incorporating historical information on an arbitrary number of dose groups, including both placebo and active ones, taking into account the relationship between responses of these dose groups. Our approach can also model both prognostic and predictive between-trial heterogeneity and is particularly useful in situations where the effect sizes of two trials are different. Our goal is to reduce the necessary sample size in the dose-finding trial while maintaining its target power.</p>","PeriodicalId":54870,"journal":{"name":"Journal of Biopharmaceutical Statistics","volume":" ","pages":"646-660"},"PeriodicalIF":1.2,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10161812","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01Epub Date: 2023-10-11DOI: 10.1080/10543406.2023.2269251
Dong Wang, Sue-Jane Wang, Samir Lababidi
The development of next-generation sequencing (NGS) opens opportunities for new applications such as liquid biopsy, in which tumor mutation genotypes can be determined by sequencing circulating tumor DNA after blood draws. However, with highly diluted samples like those obtained with liquid biopsy, NGS invariably introduces a certain level of misclassification, even with improved technology. Recently, there has been a high demand to use mutation genotypes as biomarkers for predicting prognosis and treatment selection. Many methods have also been proposed to build classifiers based on multiple loci with machine learning algorithms as biomarkers. How the higher misclassification rate introduced by liquid biopsy will affect the performance of these biomarkers has not been thoroughly investigated. In this paper, we report the results from a simulation study focused on the clinical utility of biomarkers when misclassification is present due to the current technological limit of NGS in the liquid biopsy setting. The simulation covers a range of performance profiles for current NGS platforms with different machine learning algorithms and uses actual patient genotypes. Our results show that, at the high end of the performance spectrum, the misclassification introduced by NGS had very little effect on the clinical utility of the biomarker. However, in more challenging applications with lower accuracy, misclassification could have a notable effect on clinical utility. The pattern of this effect can be complex, especially for machine learning-based classifiers. Our results show that simulation can be an effective tool for assessing different scenarios of misclassification.
{"title":"The impact of misclassification errors on the performance of biomarkers based on next-generation sequencing, a simulation study.","authors":"Dong Wang, Sue-Jane Wang, Samir Lababidi","doi":"10.1080/10543406.2023.2269251","DOIUrl":"10.1080/10543406.2023.2269251","url":null,"abstract":"<p><p>The development of next-generation sequencing (NGS) opens opportunities for new applications such as liquid biopsy, in which tumor mutation genotypes can be determined by sequencing circulating tumor DNA after blood draws. However, with highly diluted samples like those obtained with liquid biopsy, NGS invariably introduces a certain level of misclassification, even with improved technology. Recently, there has been a high demand to use mutation genotypes as biomarkers for predicting prognosis and treatment selection. Many methods have also been proposed to build classifiers based on multiple loci with machine learning algorithms as biomarkers. How the higher misclassification rate introduced by liquid biopsy will affect the performance of these biomarkers has not been thoroughly investigated. In this paper, we report the results from a simulation study focused on the clinical utility of biomarkers when misclassification is present due to the current technological limit of NGS in the liquid biopsy setting. The simulation covers a range of performance profiles for current NGS platforms with different machine learning algorithms and uses actual patient genotypes. Our results show that, at the high end of the performance spectrum, the misclassification introduced by NGS had very little effect on the clinical utility of the biomarker. However, in more challenging applications with lower accuracy, misclassification could have a notable effect on clinical utility. The pattern of this effect can be complex, especially for machine learning-based classifiers. Our results show that simulation can be an effective tool for assessing different scenarios of misclassification.</p>","PeriodicalId":54870,"journal":{"name":"Journal of Biopharmaceutical Statistics","volume":" ","pages":"700-718"},"PeriodicalIF":1.2,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41220408","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01Epub Date: 2023-11-21DOI: 10.1080/10543406.2023.2275765
Quynh Nguyen, Katharina Hees, Benjamin Hofner
Platform trials offer a framework to study multiple interventions in one trial with the opportunity of opening and closing arms. The use of common controls can increase efficiency as compared to individual controls. The need for multiplicity adjustment because of common controls is currently a debate among researchers, pharmaceutical companies, and regulators. The impact of common controls on the type one error in a fixed platform trial, i.e. when all treatments start and end recruitment at the same time, has been discussed in the literature before. We complement these findings by investigating the impact of a common control on the type one error and power in a flexible platform trial, i.e. when one arm joins the platform later. We derived the correlation of test statistics to assess the impact of the overlap and compared the results to a trial with individual controls. Furthermore, we evaluate the power, and the impact of multiplicity adjustment on the power in fixed and flexible platform trials. These methodological considerations are complemented by a regulatory guideline review. With multiple arms, the FWER is inflated when no multiplicity adjustment is applied. However, the FWER inflation is smaller with common controls than with individual controls. Even after multiplicity adjustment, a trial with common controls is often beneficial in terms of sample size and power. However, in some cases, the trial with common controls loses the efficiency gain and it might be advisable to run a separate trial rather than joining a platform trial.
{"title":"Adaptive platform trials: the impact of common controls on type one error and power.","authors":"Quynh Nguyen, Katharina Hees, Benjamin Hofner","doi":"10.1080/10543406.2023.2275765","DOIUrl":"10.1080/10543406.2023.2275765","url":null,"abstract":"<p><p>Platform trials offer a framework to study multiple interventions in one trial with the opportunity of opening and closing arms. The use of common controls can increase efficiency as compared to individual controls. The need for multiplicity adjustment because of common controls is currently a debate among researchers, pharmaceutical companies, and regulators. The impact of common controls on the type one error in a fixed platform trial, i.e. when all treatments start and end recruitment at the same time, has been discussed in the literature before. We complement these findings by investigating the impact of a common control on the type one error and power in a flexible platform trial, i.e. when one arm joins the platform later. We derived the correlation of test statistics to assess the impact of the overlap and compared the results to a trial with individual controls. Furthermore, we evaluate the power, and the impact of multiplicity adjustment on the power in fixed and flexible platform trials. These methodological considerations are complemented by a regulatory guideline review. With multiple arms, the FWER is inflated when no multiplicity adjustment is applied. However, the FWER inflation is smaller with common controls than with individual controls. Even after multiplicity adjustment, a trial with common controls is often beneficial in terms of sample size and power. However, in some cases, the trial with common controls loses the efficiency gain and it might be advisable to run a separate trial rather than joining a platform trial.</p>","PeriodicalId":54870,"journal":{"name":"Journal of Biopharmaceutical Statistics","volume":" ","pages":"719-736"},"PeriodicalIF":1.2,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138292447","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01Epub Date: 2023-08-10DOI: 10.1080/10543406.2023.2244055
Yuichiro Kaneko, Satoshi Morita
The delayed treatment effect, which manifests as a separation of survival curves after a change point, has often been observed in immunotherapy clinical trials. A late effect of this kind may violate the proportional hazards assumption, resulting in the non-negligible loss of statistical power of an ordinary log-rank test when comparing survival curves. The Fleming-Harrington (FH) test, a weighted log-rank test, is configured to mitigate the loss of power by incorporating a weight function with two parameters, one each for early and late treatment effects. The two parameters need to be appropriately determined, but no helpful guides have been fully established. Since the late effect is expected in immunotherapy trials, we focus on the late effect parameter in this study. We consider parameterizing the late effect in a readily interpretable fashion and determining the optimal late effect parameter in the FH test to maintain statistical power in reference to the asymptotic relative efficiency (ARE). The optimization is carried out under three lag models (i.e. linear, threshold, and generalized linear lag), where the optimal weights are proportional to the lag functions characterized by the change points. Extensive simulation studies showed that the FH test with the selected late parameter reliably provided sufficient power even when the change points in the lag models were misspecified. This finding suggests that the FH test with the ARE-guided late parameter may be a reasonable and practical choice for the primary analysis in immunotherapy clinical trials.
在免疫疗法临床试验中经常可以观察到延迟治疗效应,这种效应表现为变化点之后生存曲线的分离。这种延迟效应可能会违反比例危险假设,从而导致在比较生存曲线时,普通对数秩检验的统计能力出现不可忽略的损失。Fleming-Harrington(FH)检验是一种加权对数rank检验,通过加入一个带有两个参数(早期和晚期治疗效果各一个)的加权函数来减轻统计能力的损失。这两个参数需要适当确定,但目前还没有完全确定的有用指南。由于在免疫疗法试验中预期会出现晚期效应,因此我们在本研究中将重点放在晚期效应参数上。我们考虑以一种易于解释的方式确定晚期效应参数,并参照渐近相对效率(ARE)确定 FH 试验中的最佳晚期效应参数,以保持统计功率。优化在三种滞后模型(即线性滞后、阈值滞后和广义线性滞后)下进行,其中最优权重与变化点所表征的滞后函数成正比。广泛的模拟研究表明,即使滞后模型中的变化点被错误地指定,使用选定的后期参数进行的 FH 检验也能可靠地提供足够的功率。这一结果表明,在免疫疗法临床试验的主要分析中,使用 ARE 指导的后期参数进行 FH 检验可能是一个合理而实用的选择。
{"title":"Determining the late effect parameter in the Fleming-Harrington test using asymptotic relative efficiency in cancer immunotherapy clinical trials.","authors":"Yuichiro Kaneko, Satoshi Morita","doi":"10.1080/10543406.2023.2244055","DOIUrl":"10.1080/10543406.2023.2244055","url":null,"abstract":"<p><p>The delayed treatment effect, which manifests as a separation of survival curves after a change point, has often been observed in immunotherapy clinical trials. A late effect of this kind may violate the proportional hazards assumption, resulting in the non-negligible loss of statistical power of an ordinary log-rank test when comparing survival curves. The Fleming-Harrington (FH) test, a weighted log-rank test, is configured to mitigate the loss of power by incorporating a weight function with two parameters, one each for early and late treatment effects. The two parameters need to be appropriately determined, but no helpful guides have been fully established. Since the late effect is expected in immunotherapy trials, we focus on the late effect parameter in this study. We consider parameterizing the late effect in a readily interpretable fashion and determining the optimal late effect parameter in the FH test to maintain statistical power in reference to the asymptotic relative efficiency (ARE). The optimization is carried out under three lag models (i.e. linear, threshold, and generalized linear lag), where the optimal weights are proportional to the lag functions characterized by the change points. Extensive simulation studies showed that the FH test with the selected late parameter reliably provided sufficient power even when the change points in the lag models were misspecified. This finding suggests that the FH test with the ARE-guided late parameter may be a reasonable and practical choice for the primary analysis in immunotherapy clinical trials.</p>","PeriodicalId":54870,"journal":{"name":"Journal of Biopharmaceutical Statistics","volume":" ","pages":"626-645"},"PeriodicalIF":1.2,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10014033","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}