Thomas Drury, Jonathan W Bartlett, David Wright, Oliver N Keene
The creation of the ICH E9 (R1) estimands framework has led to more precise specification of the treatment effects of interest in the design and statistical analysis of clinical trials. However, it is unclear how the new framework relates to causal inference, as both approaches appear to define what is being estimated and have a quantity labeled an estimand. Using illustrative examples, we show that both approaches can be used to define a population-based summary of an effect on an outcome for a specified population and highlight the similarities and differences between these approaches. We demonstrate that the ICH E9 (R1) estimand framework offers a descriptive, structured approach that is more accessible to non-mathematicians, facilitating clearer communication of trial objectives and results. We then contrast this with the causal inference framework, which provides a mathematically precise definition of an estimand and allows the explicit articulation of assumptions through tools such as causal graphs. Despite these differences, the two paradigms should be viewed as complementary rather than competing. The combined use of both approaches enhances the ability to communicate what is being estimated. We encourage those familiar with one framework to appreciate the concepts of the other to strengthen the robustness and clarity of clinical trial design, analysis, and interpretation.
ICH E9 (R1)估计框架的创建导致了对临床试验设计和统计分析中感兴趣的治疗效果的更精确规范。然而,尚不清楚新框架与因果推理的关系,因为两种方法似乎都定义了被估计的内容,并有一个标记为估计的数量。使用说明性的例子,我们表明这两种方法都可以用来定义对特定人群的结果的影响的基于人群的总结,并强调这些方法之间的异同。我们证明,ICH E9 (R1)估算框架提供了一种描述性的、结构化的方法,非数学家更容易理解,有助于更清晰地沟通试验目标和结果。然后,我们将其与因果推理框架进行对比,因果推理框架提供了估算的数学精确定义,并允许通过因果图等工具明确表达假设。尽管存在这些差异,但这两种模式应被视为互补而非竞争。两种方法的结合使用增强了沟通评估内容的能力。我们鼓励熟悉其中一个框架的人了解另一个框架的概念,以加强临床试验设计、分析和解释的稳健性和清晰度。
{"title":"The Estimand Framework and Causal Inference: Complementary Not Competing Paradigms.","authors":"Thomas Drury, Jonathan W Bartlett, David Wright, Oliver N Keene","doi":"10.1002/pst.70035","DOIUrl":"https://doi.org/10.1002/pst.70035","url":null,"abstract":"<p><p>The creation of the ICH E9 (R1) estimands framework has led to more precise specification of the treatment effects of interest in the design and statistical analysis of clinical trials. However, it is unclear how the new framework relates to causal inference, as both approaches appear to define what is being estimated and have a quantity labeled an estimand. Using illustrative examples, we show that both approaches can be used to define a population-based summary of an effect on an outcome for a specified population and highlight the similarities and differences between these approaches. We demonstrate that the ICH E9 (R1) estimand framework offers a descriptive, structured approach that is more accessible to non-mathematicians, facilitating clearer communication of trial objectives and results. We then contrast this with the causal inference framework, which provides a mathematically precise definition of an estimand and allows the explicit articulation of assumptions through tools such as causal graphs. Despite these differences, the two paradigms should be viewed as complementary rather than competing. The combined use of both approaches enhances the ability to communicate what is being estimated. We encourage those familiar with one framework to appreciate the concepts of the other to strengthen the robustness and clarity of clinical trial design, analysis, and interpretation.</p>","PeriodicalId":19934,"journal":{"name":"Pharmaceutical Statistics","volume":"24 5","pages":"e70035"},"PeriodicalIF":1.4,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144964825","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Correction to \"Sample Size Estimation for Correlated Count Data With Changes in Dispersion\".","authors":"","doi":"10.1002/pst.70034","DOIUrl":"https://doi.org/10.1002/pst.70034","url":null,"abstract":"","PeriodicalId":19934,"journal":{"name":"Pharmaceutical Statistics","volume":"24 5","pages":"e70034"},"PeriodicalIF":1.4,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145023976","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Michaela Maria Freitag, Dario Zocholl, Elias Laurin Meyer, Stefan M Gold, Marta Bofill Roig, Heidi De Smedt, Martin Posch, Franz König
Major depressive disorder (MDD) is one of the leading causes of disability globally. Despite its prevalence, approximately one-third of patients do not benefit sufficiently from available treatments, and few new drugs have been developed recently. Consequently, more efficient methods are needed to evaluate a broader range of treatment options quickly. Platform trials offer a promising solution, as they allow for the assessment of multiple investigational treatments simultaneously by sharing control groups and by reducing both trial activation and patient recruitment times. The objective of this simulation study was to support the design and optimisation of a phase II superiority platform trial for MDD, considering the disease-specific characteristics. In particular, we assessed the efficiency of platform trials compared to traditional two-arm trials by investigating key design elements, including allocation and randomisation strategies, as well as per-treatment arm sample sizes and interim futility analyses. Through extensive simulations, we refined these design components and evaluated their impact on trial performance. The results demonstrated that platform trials not only enhance efficiency but also achieve higher statistical power in evaluating individual treatments compared to conventional trials. The efficiency of platform trials is particularly prominent when interim futility analyses are performed to eliminate treatments that have either no or a negligible treatment effect early. Overall, this work provides valuable insights into the design of platform trials in the superiority setting and underscores their potential to accelerate therapy development in MDD and other therapeutic areas, providing a flexible and powerful alternative to traditional trial designs.
{"title":"Design Considerations for a Phase II Platform Trial in Major Depressive Disorder.","authors":"Michaela Maria Freitag, Dario Zocholl, Elias Laurin Meyer, Stefan M Gold, Marta Bofill Roig, Heidi De Smedt, Martin Posch, Franz König","doi":"10.1002/pst.70025","DOIUrl":"https://doi.org/10.1002/pst.70025","url":null,"abstract":"<p><p>Major depressive disorder (MDD) is one of the leading causes of disability globally. Despite its prevalence, approximately one-third of patients do not benefit sufficiently from available treatments, and few new drugs have been developed recently. Consequently, more efficient methods are needed to evaluate a broader range of treatment options quickly. Platform trials offer a promising solution, as they allow for the assessment of multiple investigational treatments simultaneously by sharing control groups and by reducing both trial activation and patient recruitment times. The objective of this simulation study was to support the design and optimisation of a phase II superiority platform trial for MDD, considering the disease-specific characteristics. In particular, we assessed the efficiency of platform trials compared to traditional two-arm trials by investigating key design elements, including allocation and randomisation strategies, as well as per-treatment arm sample sizes and interim futility analyses. Through extensive simulations, we refined these design components and evaluated their impact on trial performance. The results demonstrated that platform trials not only enhance efficiency but also achieve higher statistical power in evaluating individual treatments compared to conventional trials. The efficiency of platform trials is particularly prominent when interim futility analyses are performed to eliminate treatments that have either no or a negligible treatment effect early. Overall, this work provides valuable insights into the design of platform trials in the superiority setting and underscores their potential to accelerate therapy development in MDD and other therapeutic areas, providing a flexible and powerful alternative to traditional trial designs.</p>","PeriodicalId":19934,"journal":{"name":"Pharmaceutical Statistics","volume":"24 5","pages":"e70025"},"PeriodicalIF":1.4,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12384050/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144964785","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
For toxicology studies, the validation of the concurrent control group by historical control data (HCD) has become requirements. This validation is usually done by historical control limits (HCL), which should cover the observations of the concurrent control with a predefined level of confidence. In many applications, HCL are applied to dichotomous data, for example, the number of rats with a tumor versus the number of rats without a tumor (carcinogenicity studies) or the number of cells with a micronucleus out of a total number of cells. Dichotomous HCD may be overdispersed and can be heavily right- (or left-) skewed, which is usually not taken into account in the practical applications of HCL. To overcome this problem, four different prediction intervals (two frequentist, two Bayesian), that can be applied to such data, are proposed. Based on comprehensive Monte-Carlo simulations, the coverage probabilities of the proposed prediction intervals were compared to heuristical HCL typically used in daily toxicological routine (historical range, limits of the np-chart, mean 2 SD). Our simulations reveal, that frequentist bootstrap calibrated prediction intervals control the type-1-error best, but, also prediction intervals calculated based on Bayesian generalized linear mixed models appear to be practically applicable. Contrary, all heuristics fail to control the type-1-error. The application of HCL is demonstrated based on a real life data set containing historical controls from long-term carcinogenicity studies run on behalf of the U.S. National Toxicology Program. The proposed frequentist prediction intervals are publicly available from the R package predint, whereas R code for the computation of the two Bayesian prediction intervals is provided via GitHub.
{"title":"Prediction Intervals for Overdispersed Binomial Endpoints and Their Application to Toxicological Historical Control Data.","authors":"Max Menssen, Jonathan Rathjens","doi":"10.1002/pst.70033","DOIUrl":"10.1002/pst.70033","url":null,"abstract":"<p><p>For toxicology studies, the validation of the concurrent control group by historical control data (HCD) has become requirements. This validation is usually done by historical control limits (HCL), which should cover the observations of the concurrent control with a predefined level of confidence. In many applications, HCL are applied to dichotomous data, for example, the number of rats with a tumor versus the number of rats without a tumor (carcinogenicity studies) or the number of cells with a micronucleus out of a total number of cells. Dichotomous HCD may be overdispersed and can be heavily right- (or left-) skewed, which is usually not taken into account in the practical applications of HCL. To overcome this problem, four different prediction intervals (two frequentist, two Bayesian), that can be applied to such data, are proposed. Based on comprehensive Monte-Carlo simulations, the coverage probabilities of the proposed prediction intervals were compared to heuristical HCL typically used in daily toxicological routine (historical range, limits of the np-chart, mean <math> <semantics><mrow><mo>±</mo></mrow> <annotation>$$ pm $$</annotation></semantics> </math> 2 SD). Our simulations reveal, that frequentist bootstrap calibrated prediction intervals control the type-1-error best, but, also prediction intervals calculated based on Bayesian generalized linear mixed models appear to be practically applicable. Contrary, all heuristics fail to control the type-1-error. The application of HCL is demonstrated based on a real life data set containing historical controls from long-term carcinogenicity studies run on behalf of the U.S. National Toxicology Program. The proposed frequentist prediction intervals are publicly available from the R package predint, whereas R code for the computation of the two Bayesian prediction intervals is provided via GitHub.</p>","PeriodicalId":19934,"journal":{"name":"Pharmaceutical Statistics","volume":"24 5","pages":"e70033"},"PeriodicalIF":1.4,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12433933/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145065356","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Hajime Uno, Lu Tian, Miki Horiguchi, Satoshi Hattori
{"title":"Comment on \"Average Hazard as Harmonic Mean\" by Chiba (2025).","authors":"Hajime Uno, Lu Tian, Miki Horiguchi, Satoshi Hattori","doi":"10.1002/pst.70032","DOIUrl":"10.1002/pst.70032","url":null,"abstract":"","PeriodicalId":19934,"journal":{"name":"Pharmaceutical Statistics","volume":"24 5","pages":"e70032"},"PeriodicalIF":1.4,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12806506/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144799848","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A major emphasis in personalized medicine is to optimally treat subgroups of patients who may benefit from certain therapeutic agents. One relevant study design is the targeted design, in which patients have consented for their specimens to be obtained at baseline and the specimens are sent to a laboratory for assessing the biomarker status prior to randomization. Here, only biomarker-positive patients will be randomized to either an experimental or the standard of care arms. Many biomarkers, however, are derived from patient tissue specimens, which are heterogeneous leading to variability in the biomarker levels and status. This heterogeneity would have an adverse impact on the power of an enriched biomarker clinical trial. In this article, we show the adverse effect of using the uncorrected sample size and overcome this challenge by presenting an approach to adjust for misclassification for the targeted design. Specifically, we propose a sample size formula that adjusts for misclassification and apply it in the design of two phase III clinical trials in renal and prostate cancer.
{"title":"Sample Size for Enriched Biomarker Designs With Measurement Error for Time-to-Event Outcomes.","authors":"Siyuan Guo, Susan Halabi, Aiyi Liu","doi":"10.1002/pst.70027","DOIUrl":"10.1002/pst.70027","url":null,"abstract":"<p><p>A major emphasis in personalized medicine is to optimally treat subgroups of patients who may benefit from certain therapeutic agents. One relevant study design is the targeted design, in which patients have consented for their specimens to be obtained at baseline and the specimens are sent to a laboratory for assessing the biomarker status prior to randomization. Here, only biomarker-positive patients will be randomized to either an experimental or the standard of care arms. Many biomarkers, however, are derived from patient tissue specimens, which are heterogeneous leading to variability in the biomarker levels and status. This heterogeneity would have an adverse impact on the power of an enriched biomarker clinical trial. In this article, we show the adverse effect of using the uncorrected sample size and overcome this challenge by presenting an approach to adjust for misclassification for the targeted design. Specifically, we propose a sample size formula that adjusts for misclassification and apply it in the design of two phase III clinical trials in renal and prostate cancer.</p>","PeriodicalId":19934,"journal":{"name":"Pharmaceutical Statistics","volume":"24 5","pages":"e70027"},"PeriodicalIF":1.4,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12683776/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144708386","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Sample size determination in Bayesian randomized phase II trial design often relies on computationally intensive search methods, presenting challenges in terms of feasibility and efficiency. We propose a novel approach that greatly reduces the computing time of sample size calculations for Bayesian trial designs. Our approach innovatively connects group sequential design with Bayesian trial design and leverages the proportional relationship between sample size and the squared drift parameter. This results in a faster algorithm. By employing regression analysis, our method can accurately pinpoint the required sample size with significantly reduced computational burden. Through theoretical justification and extensive numerical evaluations, we validate our approach and illustrate its efficiency across a wide range of common trial scenarios, including binary endpoint with Beta-Binomial model, normal endpoint, binary/ordinal endpoint under Bayesian generalized linear model, and survival endpoints under Bayesian piecewise exponential models. To facilitate the use of our methods, we create an R package named "BayesSize" on GitHub.
{"title":"Drift Parameter Based Sample Size Determination in Multi-Stage Bayesian Randomized Clinical Trials.","authors":"Yueyang Han, Haolun Shi, Jiguo Cao, Ruitao Lin","doi":"10.1002/pst.70037","DOIUrl":"10.1002/pst.70037","url":null,"abstract":"<p><p>Sample size determination in Bayesian randomized phase II trial design often relies on computationally intensive search methods, presenting challenges in terms of feasibility and efficiency. We propose a novel approach that greatly reduces the computing time of sample size calculations for Bayesian trial designs. Our approach innovatively connects group sequential design with Bayesian trial design and leverages the proportional relationship between sample size and the squared drift parameter. This results in a faster algorithm. By employing regression analysis, our method can accurately pinpoint the required sample size with significantly reduced computational burden. Through theoretical justification and extensive numerical evaluations, we validate our approach and illustrate its efficiency across a wide range of common trial scenarios, including binary endpoint with Beta-Binomial model, normal endpoint, binary/ordinal endpoint under Bayesian generalized linear model, and survival endpoints under Bayesian piecewise exponential models. To facilitate the use of our methods, we create an R package named \"BayesSize\" on GitHub.</p>","PeriodicalId":19934,"journal":{"name":"Pharmaceutical Statistics","volume":"24 5","pages":"e70037"},"PeriodicalIF":1.4,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144855998","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In this paper, we carried out extensive simulation studies to compare the performances of partial and maximum likelihood based methods for estimating the area under the bi-Weibull ROC curve. Further, real data sets from HIV/AIDS research were analyzed for illustrative purposes. Simulation results suggest that both methods perform well and yield similar results for Weibull data. However, for non-Weibull data, both methods perform poorly. The bi-Weibull model yields smooth estimates of ROC curves and a closed-form expression for the area under the ROC curve. Moreover, by adjusting its shape parameter, the bi-Weibull model can represent a variety of distributions, such as exponential, Rayleigh, normal, and extreme value distributions. Its compatibility with Cox's proportional hazards model facilitates the derivation of covariate-adjusted ROC curves and supports analyses involving correlated and longitudinal biomarkers. These properties make the model very useful in the ROC curve analyses. Thus, the bi-Weibull model should be considered as an alternative when the restrictive distributional assumptions of the commonly used parametric models (e.g., binormal model) are not met.
{"title":"Comparing Estimation Methods for the Area Under the Bi-Weibull ROC Curve.","authors":"Ruhul Ali Khan, Musie Ghebremichael","doi":"10.1002/pst.70038","DOIUrl":"https://doi.org/10.1002/pst.70038","url":null,"abstract":"<p><p>In this paper, we carried out extensive simulation studies to compare the performances of partial and maximum likelihood based methods for estimating the area under the bi-Weibull ROC curve. Further, real data sets from HIV/AIDS research were analyzed for illustrative purposes. Simulation results suggest that both methods perform well and yield similar results for Weibull data. However, for non-Weibull data, both methods perform poorly. The bi-Weibull model yields smooth estimates of ROC curves and a closed-form expression for the area under the ROC curve. Moreover, by adjusting its shape parameter, the bi-Weibull model can represent a variety of distributions, such as exponential, Rayleigh, normal, and extreme value distributions. Its compatibility with Cox's proportional hazards model facilitates the derivation of covariate-adjusted ROC curves and supports analyses involving correlated and longitudinal biomarkers. These properties make the model very useful in the ROC curve analyses. Thus, the bi-Weibull model should be considered as an alternative when the restrictive distributional assumptions of the commonly used parametric models (e.g., binormal model) are not met.</p>","PeriodicalId":19934,"journal":{"name":"Pharmaceutical Statistics","volume":"24 5","pages":"e70038"},"PeriodicalIF":1.4,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145001214","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
With the ongoing advancements in cancer drug development, a subset of patients can live quite long, or are even considered cured in certain cancer types. Additionally, nonproportional hazards, such as delayed treatment effects and crossing hazards, are commonly observed in cancer clinical trials with immunotherapy. To address these challenges, various cure models have been proposed to integrate the cure rate into trial designs and accommodate delayed treatment effects. In this article, we introduce a unified approach for calculating sample sizes, taking into account different cure rate models and nonproportional hazards. Our approach supports both the traditional weighted logrank test and the Maxcombo test, which demonstrates robust performance under nonproportional hazards. Furthermore, we assess the accuracy of our sample size estimation through Monte Carlo simulations across various scenarios and compare our method with existing approaches. Several illustrative examples are provided to demonstrate the proposed method.
{"title":"A General Approach for Sample Size Calculation With Nonproportional Hazards and Cure Rates.","authors":"Huan Cheng, Xiaoyun Li, Jianghua He","doi":"10.1002/pst.70024","DOIUrl":"https://doi.org/10.1002/pst.70024","url":null,"abstract":"<p><p>With the ongoing advancements in cancer drug development, a subset of patients can live quite long, or are even considered cured in certain cancer types. Additionally, nonproportional hazards, such as delayed treatment effects and crossing hazards, are commonly observed in cancer clinical trials with immunotherapy. To address these challenges, various cure models have been proposed to integrate the cure rate into trial designs and accommodate delayed treatment effects. In this article, we introduce a unified approach for calculating sample sizes, taking into account different cure rate models and nonproportional hazards. Our approach supports both the traditional weighted logrank test and the Maxcombo test, which demonstrates robust performance under nonproportional hazards. Furthermore, we assess the accuracy of our sample size estimation through Monte Carlo simulations across various scenarios and compare our method with existing approaches. Several illustrative examples are provided to demonstrate the proposed method.</p>","PeriodicalId":19934,"journal":{"name":"Pharmaceutical Statistics","volume":"24 4","pages":"e70024"},"PeriodicalIF":1.3,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144576028","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Claire Watkins, Eva Kleine, Miguel Miranda, Emmanuel Bourmaud, Orlando Doehring
The objective of this article is to bring together the key current information on practical considerations when conducting statistical analyses adjusting long-term outcomes for treatment switching, combining it with learnings from our own experience, thus providing a useful reference tool for analysts. When patients switch from their randomised treatment to another therapy that affects a subsequently observed outcome such as overall survival, there may be interest in estimating the treatment effect under a hypothetical scenario without the intercurrent event of switching. We describe the theory and provide guidance on how and when to conduct analyses using three commonly used complex approaches: rank preserving structural failure time models (RPSFTM), two-stage estimation (TSE), and inverse probability of censoring weighting (IPCW). Extensions and alternatives to the standard approaches are summarised. Important and sometimes misunderstood concepts such as recensoring and sources of variability are explained. An overview of available software and programming guidance is provided, along with an R code repository for a worked example, reporting recommendations, and a review of the current acceptability of these methods to regulatory and health technology assessment agencies. Since the current guidance on this topic is scattered across multiple sources, it is difficult for an analyst to obtain a good overview of all options and potential pitfalls. This paper is intended to save statisticians time and effort by summarizing important information in a single source. By also including recommendations for best practice, it aims to improve the quality of the analyses and reporting when adjusting time-to-event outcomes for treatment switching.
{"title":"Further Practical Guidance on Adjusting Time-To-Event Outcomes for Treatment Switching.","authors":"Claire Watkins, Eva Kleine, Miguel Miranda, Emmanuel Bourmaud, Orlando Doehring","doi":"10.1002/pst.70019","DOIUrl":"10.1002/pst.70019","url":null,"abstract":"<p><p>The objective of this article is to bring together the key current information on practical considerations when conducting statistical analyses adjusting long-term outcomes for treatment switching, combining it with learnings from our own experience, thus providing a useful reference tool for analysts. When patients switch from their randomised treatment to another therapy that affects a subsequently observed outcome such as overall survival, there may be interest in estimating the treatment effect under a hypothetical scenario without the intercurrent event of switching. We describe the theory and provide guidance on how and when to conduct analyses using three commonly used complex approaches: rank preserving structural failure time models (RPSFTM), two-stage estimation (TSE), and inverse probability of censoring weighting (IPCW). Extensions and alternatives to the standard approaches are summarised. Important and sometimes misunderstood concepts such as recensoring and sources of variability are explained. An overview of available software and programming guidance is provided, along with an R code repository for a worked example, reporting recommendations, and a review of the current acceptability of these methods to regulatory and health technology assessment agencies. Since the current guidance on this topic is scattered across multiple sources, it is difficult for an analyst to obtain a good overview of all options and potential pitfalls. This paper is intended to save statisticians time and effort by summarizing important information in a single source. By also including recommendations for best practice, it aims to improve the quality of the analyses and reporting when adjusting time-to-event outcomes for treatment switching.</p>","PeriodicalId":19934,"journal":{"name":"Pharmaceutical Statistics","volume":"24 4","pages":"e70019"},"PeriodicalIF":1.3,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12181803/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144340348","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}