Pub Date : 2024-10-30Epub Date: 2024-08-21DOI: 10.1002/sim.10203
Amin Yarahmadi, Lori E Dodd, Thomas Jaki, Peter Horby, Nigel Stallard
Motivated by the experience of COVID-19 trials, we consider clinical trials in the setting of an emerging disease in which the uncertainty of natural disease course and potential treatment effects makes advance specification of a sample size challenging. One approach to such a challenge is to use a group sequential design to allow the trial to stop on the basis of interim analysis results as soon as a conclusion regarding the effectiveness of the treatment under investigation can be reached. As such a trial may be halted before a formal stopping boundary is reached, we consider the final analysis under such a scenario, proposing alternative methods for when the decision to halt the trial is made with or without knowledge of interim analysis results. We address the problems of ensuring that the type I error rate neither exceeds nor falls unnecessarily far below the nominal level. We also propose methods in which there is no maximum sample size, the trial continuing either until the stopping boundary is reached or it is decided to halt the trial.
受 COVID-19 试验经验的启发,我们考虑了新发疾病背景下的临床试验,在这种情况下,自然病程和潜在治疗效果的不确定性使得提前确定样本量具有挑战性。应对这种挑战的一种方法是采用分组顺序设计,以便一旦对所研究的治疗效果得出结论,就可以根据中期分析结果停止试验。由于这种试验可能会在达到正式停止界限之前就停止,因此我们考虑了这种情况下的最终分析,提出了在了解或不了解中期分析结果的情况下决定停止试验的替代方法。我们解决了确保 I 类错误率既不超过名义水平,也不会不必要地远远低于名义水平的问题。我们还提出了不设最大样本量的方法,即试验一直持续到达到停止边界或决定停止试验为止。
{"title":"Group sequential designs for clinical trials when the maximum sample size is uncertain.","authors":"Amin Yarahmadi, Lori E Dodd, Thomas Jaki, Peter Horby, Nigel Stallard","doi":"10.1002/sim.10203","DOIUrl":"10.1002/sim.10203","url":null,"abstract":"<p><p>Motivated by the experience of COVID-19 trials, we consider clinical trials in the setting of an emerging disease in which the uncertainty of natural disease course and potential treatment effects makes advance specification of a sample size challenging. One approach to such a challenge is to use a group sequential design to allow the trial to stop on the basis of interim analysis results as soon as a conclusion regarding the effectiveness of the treatment under investigation can be reached. As such a trial may be halted before a formal stopping boundary is reached, we consider the final analysis under such a scenario, proposing alternative methods for when the decision to halt the trial is made with or without knowledge of interim analysis results. We address the problems of ensuring that the type I error rate neither exceeds nor falls unnecessarily far below the nominal level. We also propose methods in which there is no maximum sample size, the trial continuing either until the stopping boundary is reached or it is decided to halt the trial.</p>","PeriodicalId":21879,"journal":{"name":"Statistics in Medicine","volume":" ","pages":"4667-4683"},"PeriodicalIF":1.8,"publicationDate":"2024-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142009561","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-30Epub Date: 2024-08-27DOI: 10.1002/sim.10199
Michael Y C Chong, Marija Pejchinovska, Monica Alexander
Understanding the underlying causes of maternal death across all regions of the world is essential to inform policies and resource allocation to reduce the mortality burden. However, in many countries there exists very little data on the causes of maternal death, and data that do exist do not capture the entire population at risk. In this article, we present a Bayesian hierarchical multinomial model to estimate maternal cause of death distributions globally, regionally, and for all countries worldwide. The framework combines data from various sources to inform estimates, including data from civil registration and vital systems, smaller-scale surveys and studies, and high-quality data from confidential enquiries and surveillance systems. The framework accounts for varying data quality and coverage, and allows for situations where one or more causes of death are missing. We illustrate the results of the model on three case-study countries that have different data availability situations.
{"title":"Estimating causes of maternal death in data-sparse contexts.","authors":"Michael Y C Chong, Marija Pejchinovska, Monica Alexander","doi":"10.1002/sim.10199","DOIUrl":"10.1002/sim.10199","url":null,"abstract":"<p><p>Understanding the underlying causes of maternal death across all regions of the world is essential to inform policies and resource allocation to reduce the mortality burden. However, in many countries there exists very little data on the causes of maternal death, and data that do exist do not capture the entire population at risk. In this article, we present a Bayesian hierarchical multinomial model to estimate maternal cause of death distributions globally, regionally, and for all countries worldwide. The framework combines data from various sources to inform estimates, including data from civil registration and vital systems, smaller-scale surveys and studies, and high-quality data from confidential enquiries and surveillance systems. The framework accounts for varying data quality and coverage, and allows for situations where one or more causes of death are missing. We illustrate the results of the model on three case-study countries that have different data availability situations.</p>","PeriodicalId":21879,"journal":{"name":"Statistics in Medicine","volume":" ","pages":"4702-4735"},"PeriodicalIF":1.8,"publicationDate":"2024-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142073902","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-15Epub Date: 2024-08-01DOI: 10.1002/sim.10186
J Hoogland, O Efthimiou, T L Nguyen, T P A Debray
In recent years, there has been a growing interest in the prediction of individualized treatment effects. While there is a rapidly growing literature on the development of such models, there is little literature on the evaluation of their performance. In this paper, we aim to facilitate the validation of prediction models for individualized treatment effects. The estimands of interest are defined based on the potential outcomes framework, which facilitates a comparison of existing and novel measures. In particular, we examine existing measures of discrimination for benefit (variations of the c-for-benefit), and propose model-based extensions to the treatment effect setting for discrimination and calibration metrics that have a strong basis in outcome risk prediction. The main focus is on randomized trial data with binary endpoints and on models that provide individualized treatment effect predictions and potential outcome predictions. We use simulated data to provide insight into the characteristics of the examined discrimination and calibration statistics under consideration, and further illustrate all methods in a trial of acute ischemic stroke treatment. The results show that the proposed model-based statistics had the best characteristics in terms of bias and accuracy. While resampling methods adjusted for the optimism of performance estimates in the development data, they had a high variance across replications that limited their accuracy. Therefore, individualized treatment effect models are best validated in independent data. To aid implementation, a software implementation of the proposed methods was made available in R.
近年来,人们对个性化治疗效果的预测越来越感兴趣。尽管有关此类模型开发的文献迅速增加,但有关其性能评估的文献却很少。本文旨在促进个体化治疗效果预测模型的验证。我们根据潜在结果框架来定义感兴趣的估算对象,这有助于对现有的和新的测量方法进行比较。特别是,我们研究了现有的收益区分度(c-收益的变体),并提出了基于模型的治疗效果设定区分度和校准指标的扩展,这些指标在结果风险预测方面具有坚实的基础。主要重点是具有二元终点的随机试验数据,以及提供个体化治疗效果预测和潜在结果预测的模型。我们使用模拟数据来深入分析所研究的判别和校准统计量的特点,并在急性缺血性中风治疗试验中进一步说明所有方法。结果表明,所提出的基于模型的统计方法在偏差和准确性方面具有最佳特性。虽然重采样方法可以调整开发数据中性能估计的乐观程度,但它们在重复中的方差较大,限制了其准确性。因此,个体化治疗效果模型最好在独立数据中进行验证。为了帮助实施,我们用 R 语言提供了建议方法的软件实施。
{"title":"Evaluating individualized treatment effect predictions: A model-based perspective on discrimination and calibration assessment.","authors":"J Hoogland, O Efthimiou, T L Nguyen, T P A Debray","doi":"10.1002/sim.10186","DOIUrl":"10.1002/sim.10186","url":null,"abstract":"<p><p>In recent years, there has been a growing interest in the prediction of individualized treatment effects. While there is a rapidly growing literature on the development of such models, there is little literature on the evaluation of their performance. In this paper, we aim to facilitate the validation of prediction models for individualized treatment effects. The estimands of interest are defined based on the potential outcomes framework, which facilitates a comparison of existing and novel measures. In particular, we examine existing measures of discrimination for benefit (variations of the c-for-benefit), and propose model-based extensions to the treatment effect setting for discrimination and calibration metrics that have a strong basis in outcome risk prediction. The main focus is on randomized trial data with binary endpoints and on models that provide individualized treatment effect predictions and potential outcome predictions. We use simulated data to provide insight into the characteristics of the examined discrimination and calibration statistics under consideration, and further illustrate all methods in a trial of acute ischemic stroke treatment. The results show that the proposed model-based statistics had the best characteristics in terms of bias and accuracy. While resampling methods adjusted for the optimism of performance estimates in the development data, they had a high variance across replications that limited their accuracy. Therefore, individualized treatment effect models are best validated in independent data. To aid implementation, a software implementation of the proposed methods was made available in R.</p>","PeriodicalId":21879,"journal":{"name":"Statistics in Medicine","volume":" ","pages":"4481-4498"},"PeriodicalIF":16.4,"publicationDate":"2024-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141876030","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-15Epub Date: 2024-08-07DOI: 10.1002/sim.10182
Melissa J Smith, Leslie A McClure, D Leann Long
A causal decomposition analysis allows researchers to determine whether the difference in a health outcome between two groups can be attributed to a difference in each group's distribution of one or more modifiable mediator variables. With this knowledge, researchers and policymakers can focus on designing interventions that target these mediator variables. Existing methods for causal decomposition analysis either focus on one mediator variable or assume that each mediator variable is conditionally independent given the group label and the mediator-outcome confounders. In this article, we propose a flexible causal decomposition analysis method that can accommodate multiple correlated and interacting mediator variables, which are frequently seen in studies of health behaviors and studies of environmental pollutants. We extend a Monte Carlo-based causal decomposition analysis method to this setting by using a multivariate mediator model that can accommodate any combination of binary and continuous mediator variables. Furthermore, we state the causal assumptions needed to identify both joint and path-specific decomposition effects through each mediator variable. To illustrate the reduction in bias and confidence interval width of the decomposition effects under our proposed method, we perform a simulation study. We also apply our approach to examine whether differences in smoking status and dietary inflammation score explain any of the Black-White differences in incident diabetes using data from a national cohort study.
{"title":"Path-specific causal decomposition analysis with multiple correlated mediator variables.","authors":"Melissa J Smith, Leslie A McClure, D Leann Long","doi":"10.1002/sim.10182","DOIUrl":"10.1002/sim.10182","url":null,"abstract":"<p><p>A causal decomposition analysis allows researchers to determine whether the difference in a health outcome between two groups can be attributed to a difference in each group's distribution of one or more modifiable mediator variables. With this knowledge, researchers and policymakers can focus on designing interventions that target these mediator variables. Existing methods for causal decomposition analysis either focus on one mediator variable or assume that each mediator variable is conditionally independent given the group label and the mediator-outcome confounders. In this article, we propose a flexible causal decomposition analysis method that can accommodate multiple correlated and interacting mediator variables, which are frequently seen in studies of health behaviors and studies of environmental pollutants. We extend a Monte Carlo-based causal decomposition analysis method to this setting by using a multivariate mediator model that can accommodate any combination of binary and continuous mediator variables. Furthermore, we state the causal assumptions needed to identify both joint and path-specific decomposition effects through each mediator variable. To illustrate the reduction in bias and confidence interval width of the decomposition effects under our proposed method, we perform a simulation study. We also apply our approach to examine whether differences in smoking status and dietary inflammation score explain any of the Black-White differences in incident diabetes using data from a national cohort study.</p>","PeriodicalId":21879,"journal":{"name":"Statistics in Medicine","volume":" ","pages":"4519-4541"},"PeriodicalIF":1.8,"publicationDate":"2024-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11570347/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141898265","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-15Epub Date: 2024-08-15DOI: 10.1002/sim.10197
Kenneth J Nieser, Alex H S Harris
Quality measurement plays an increasing role in U.S. health care. Measures inform quality improvement efforts, public reporting of variations in quality of care across providers and hospitals, and high-stakes financial decisions. To be meaningful in these contexts, measures should be reliable and not heavily impacted by chance variations in sampling or measurement. Several different methods are used in practice by measure developers and endorsers to evaluate reliability; however, there is uncertainty and debate over differences between these methods and their interpretations. We review methods currently used in practice, pointing out differences that can lead to disparate reliability estimates. We compare estimates from 14 different methods in the case of two sets of mental health quality measures within a large health system. We find that estimates can differ substantially and that these discrepancies widen when sample size is reduced.
{"title":"Comparing methods for assessing the reliability of health care quality measures.","authors":"Kenneth J Nieser, Alex H S Harris","doi":"10.1002/sim.10197","DOIUrl":"10.1002/sim.10197","url":null,"abstract":"<p><p>Quality measurement plays an increasing role in U.S. health care. Measures inform quality improvement efforts, public reporting of variations in quality of care across providers and hospitals, and high-stakes financial decisions. To be meaningful in these contexts, measures should be reliable and not heavily impacted by chance variations in sampling or measurement. Several different methods are used in practice by measure developers and endorsers to evaluate reliability; however, there is uncertainty and debate over differences between these methods and their interpretations. We review methods currently used in practice, pointing out differences that can lead to disparate reliability estimates. We compare estimates from 14 different methods in the case of two sets of mental health quality measures within a large health system. We find that estimates can differ substantially and that these discrepancies widen when sample size is reduced.</p>","PeriodicalId":21879,"journal":{"name":"Statistics in Medicine","volume":" ","pages":"4575-4594"},"PeriodicalIF":16.4,"publicationDate":"2024-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141983251","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-15Epub Date: 2024-07-30DOI: 10.1002/sim.10164
Judith J Lok
<p><p>We often estimate a parameter of interest <math> <semantics><mrow><mi>ψ</mi></mrow> <annotation>$$ psi $$</annotation></semantics> </math> when the identifying conditions involve a finite-dimensional nuisance parameter <math> <semantics><mrow><mi>θ</mi> <mo>∈</mo> <msup><mrow><mi>ℝ</mi></mrow> <mrow><mi>d</mi></mrow> </msup> </mrow> <annotation>$$ theta in {mathbb{R}}^d $$</annotation></semantics> </math> . Examples from causal inference are inverse probability weighting, marginal structural models and structural nested models, which all lead to unbiased estimating equations. This article presents a consistent sandwich estimator for the variance of estimators <math> <semantics> <mrow> <mover><mrow><mi>ψ</mi></mrow> <mo>^</mo></mover> </mrow> <annotation>$$ hat{psi} $$</annotation></semantics> </math> that solve unbiased estimating equations including <math> <semantics><mrow><mi>θ</mi></mrow> <annotation>$$ theta $$</annotation></semantics> </math> which is also estimated by solving unbiased estimating equations. This article presents four additional results for settings where <math> <semantics> <mrow> <mover><mrow><mi>θ</mi></mrow> <mo>^</mo></mover> </mrow> <annotation>$$ hat{theta} $$</annotation></semantics> </math> solves (partial) score equations and <math> <semantics><mrow><mi>ψ</mi></mrow> <annotation>$$ psi $$</annotation></semantics> </math> does not depend on <math> <semantics><mrow><mi>θ</mi></mrow> <annotation>$$ theta $$</annotation></semantics> </math> . This includes many causal inference settings where <math> <semantics><mrow><mi>θ</mi></mrow> <annotation>$$ theta $$</annotation></semantics> </math> describes the treatment probabilities, missing data settings where <math> <semantics><mrow><mi>θ</mi></mrow> <annotation>$$ theta $$</annotation></semantics> </math> describes the missingness probabilities, and measurement error settings where <math> <semantics><mrow><mi>θ</mi></mrow> <annotation>$$ theta $$</annotation></semantics> </math> describes the error distribution. These four additional results are: (1) Counter-intuitively, the asymptotic variance of <math> <semantics> <mrow> <mover><mrow><mi>ψ</mi></mrow> <mo>^</mo></mover> </mrow> <annotation>$$ hat{psi} $$</annotation></semantics> </math> is typically smaller when <math> <semantics><mrow><mi>θ</mi></mrow> <annotation>$$ theta $$</annotation></semantics> </math> is estimated. (2) If estimating <math> <semantics><mrow><mi>θ</mi></mrow> <annotation>$$ theta $$</annotation></semantics> </math> is ignored, the sandwich estimator for the variance of <math> <semantics> <mrow> <mover><mrow><mi>ψ</mi></mrow> <mo>^</mo></mover> </mrow> <annotation>$$ hat{psi} $$</annotation></semantics> </math> is conservative. (3) A consistent sandwich estimator for the variance of <math> <semantics> <mrow> <mover><mrow><mi>ψ</mi></mrow> <mo>^</mo></mover> </mrow> <annotation>$$ hat{psi} $$</annotation></semantics> </math> . (4) If <math> <semantics> <mrow> <mover><mrow><mi>ψ</mi></mrow> <mo>^
{"title":"How estimating nuisance parameters can reduce the variance (with consistent variance estimation).","authors":"Judith J Lok","doi":"10.1002/sim.10164","DOIUrl":"10.1002/sim.10164","url":null,"abstract":"<p><p>We often estimate a parameter of interest <math> <semantics><mrow><mi>ψ</mi></mrow> <annotation>$$ psi $$</annotation></semantics> </math> when the identifying conditions involve a finite-dimensional nuisance parameter <math> <semantics><mrow><mi>θ</mi> <mo>∈</mo> <msup><mrow><mi>ℝ</mi></mrow> <mrow><mi>d</mi></mrow> </msup> </mrow> <annotation>$$ theta in {mathbb{R}}^d $$</annotation></semantics> </math> . Examples from causal inference are inverse probability weighting, marginal structural models and structural nested models, which all lead to unbiased estimating equations. This article presents a consistent sandwich estimator for the variance of estimators <math> <semantics> <mrow> <mover><mrow><mi>ψ</mi></mrow> <mo>^</mo></mover> </mrow> <annotation>$$ hat{psi} $$</annotation></semantics> </math> that solve unbiased estimating equations including <math> <semantics><mrow><mi>θ</mi></mrow> <annotation>$$ theta $$</annotation></semantics> </math> which is also estimated by solving unbiased estimating equations. This article presents four additional results for settings where <math> <semantics> <mrow> <mover><mrow><mi>θ</mi></mrow> <mo>^</mo></mover> </mrow> <annotation>$$ hat{theta} $$</annotation></semantics> </math> solves (partial) score equations and <math> <semantics><mrow><mi>ψ</mi></mrow> <annotation>$$ psi $$</annotation></semantics> </math> does not depend on <math> <semantics><mrow><mi>θ</mi></mrow> <annotation>$$ theta $$</annotation></semantics> </math> . This includes many causal inference settings where <math> <semantics><mrow><mi>θ</mi></mrow> <annotation>$$ theta $$</annotation></semantics> </math> describes the treatment probabilities, missing data settings where <math> <semantics><mrow><mi>θ</mi></mrow> <annotation>$$ theta $$</annotation></semantics> </math> describes the missingness probabilities, and measurement error settings where <math> <semantics><mrow><mi>θ</mi></mrow> <annotation>$$ theta $$</annotation></semantics> </math> describes the error distribution. These four additional results are: (1) Counter-intuitively, the asymptotic variance of <math> <semantics> <mrow> <mover><mrow><mi>ψ</mi></mrow> <mo>^</mo></mover> </mrow> <annotation>$$ hat{psi} $$</annotation></semantics> </math> is typically smaller when <math> <semantics><mrow><mi>θ</mi></mrow> <annotation>$$ theta $$</annotation></semantics> </math> is estimated. (2) If estimating <math> <semantics><mrow><mi>θ</mi></mrow> <annotation>$$ theta $$</annotation></semantics> </math> is ignored, the sandwich estimator for the variance of <math> <semantics> <mrow> <mover><mrow><mi>ψ</mi></mrow> <mo>^</mo></mover> </mrow> <annotation>$$ hat{psi} $$</annotation></semantics> </math> is conservative. (3) A consistent sandwich estimator for the variance of <math> <semantics> <mrow> <mover><mrow><mi>ψ</mi></mrow> <mo>^</mo></mover> </mrow> <annotation>$$ hat{psi} $$</annotation></semantics> </math> . (4) If <math> <semantics> <mrow> <mover><mrow><mi>ψ</mi></mrow> <mo>^","PeriodicalId":21879,"journal":{"name":"Statistics in Medicine","volume":" ","pages":"4456-4480"},"PeriodicalIF":1.8,"publicationDate":"2024-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11570876/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141856610","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Why does a beneficial treatment effect on a longitudinal biomarker not translate into overall treatment benefit on survival, when the biomarker is in fact a prognostic factor of survival? In a recent exploratory data analysis in oncology, we were faced with this seemingly paradoxical result. To address this problem, we applied a theoretically principled methodology called dynamic path analysis, which allows us to perform mediation analysis with a longitudinal mediator and survival outcome. The aim of the analysis is to decompose the total treatment effect into a direct treatment effect and an indirect treatment effect mediated through a carefully constructed mediation path. The dynamic nature of the underlying methodology enables us to describe how these effects evolve over time, which can add to the mechanistic understanding of the underlying processes. In this paper, we present a detailed description of the dynamic path analysis framework and illustrate its application to survival mediation analysis using simulated and real data. The use case analysis provides clarity on the specific exploratory question of interest while the methodology generalizes to a wide range of applications in drug development where time-to-event is the primary clinical outcome of interest.
{"title":"Dynamic path analysis for exploring treatment effect mediation processes in clinical trials with time-to-event endpoints.","authors":"Matthias Kormaksson, Markus Reiner Lange, David Demanse, Susanne Strohmaier, Jiawei Duan, Qing Xie, Mariana Carbini, Claudia Bossen, Achim Guettner, Antonella Maniero","doi":"10.1002/sim.10191","DOIUrl":"10.1002/sim.10191","url":null,"abstract":"<p><p>Why does a beneficial treatment effect on a longitudinal biomarker not translate into overall treatment benefit on survival, when the biomarker is in fact a prognostic factor of survival? In a recent exploratory data analysis in oncology, we were faced with this seemingly paradoxical result. To address this problem, we applied a theoretically principled methodology called dynamic path analysis, which allows us to perform mediation analysis with a longitudinal mediator and survival outcome. The aim of the analysis is to decompose the total treatment effect into a direct treatment effect and an indirect treatment effect mediated through a carefully constructed mediation path. The dynamic nature of the underlying methodology enables us to describe how these effects evolve over time, which can add to the mechanistic understanding of the underlying processes. In this paper, we present a detailed description of the dynamic path analysis framework and illustrate its application to survival mediation analysis using simulated and real data. The use case analysis provides clarity on the specific exploratory question of interest while the methodology generalizes to a wide range of applications in drug development where time-to-event is the primary clinical outcome of interest.</p>","PeriodicalId":21879,"journal":{"name":"Statistics in Medicine","volume":" ","pages":"4614-4634"},"PeriodicalIF":16.4,"publicationDate":"2024-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141898263","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-15Epub Date: 2024-07-30DOI: 10.1002/sim.10188
David Adenyo, Jason R Guertin, Bernard Candas, Caroline Sirois, Denis Talbot
Marginal structural models have been increasingly used by analysts in recent years to account for confounding bias in studies with time-varying treatments. The parameters of these models are often estimated using inverse probability of treatment weighting. To ensure that the estimated weights adequately control confounding, it is possible to check for residual imbalance between treatment groups in the weighted data. Several balance metrics have been developed and compared in the cross-sectional case but have not yet been evaluated and compared in longitudinal studies with time-varying treatment. We have first extended the definition of several balance metrics to the case of a time-varying treatment, with or without censoring. We then compared the performance of these balance metrics in a simulation study by assessing the strength of the association between their estimated level of imbalance and bias. We found that the Mahalanobis balance performed best. Finally, the method was illustrated for estimating the cumulative effect of statins exposure over one year on the risk of cardiovascular disease or death in people aged 65 and over in population-wide administrative data. This illustration confirms the feasibility of employing our proposed metrics in large databases with multiple time-points.
{"title":"Evaluation and comparison of covariate balance metrics in studies with time-dependent confounding.","authors":"David Adenyo, Jason R Guertin, Bernard Candas, Caroline Sirois, Denis Talbot","doi":"10.1002/sim.10188","DOIUrl":"10.1002/sim.10188","url":null,"abstract":"<p><p>Marginal structural models have been increasingly used by analysts in recent years to account for confounding bias in studies with time-varying treatments. The parameters of these models are often estimated using inverse probability of treatment weighting. To ensure that the estimated weights adequately control confounding, it is possible to check for residual imbalance between treatment groups in the weighted data. Several balance metrics have been developed and compared in the cross-sectional case but have not yet been evaluated and compared in longitudinal studies with time-varying treatment. We have first extended the definition of several balance metrics to the case of a time-varying treatment, with or without censoring. We then compared the performance of these balance metrics in a simulation study by assessing the strength of the association between their estimated level of imbalance and bias. We found that the Mahalanobis balance performed best. Finally, the method was illustrated for estimating the cumulative effect of statins exposure over one year on the risk of cardiovascular disease or death in people aged 65 and over in population-wide administrative data. This illustration confirms the feasibility of employing our proposed metrics in large databases with multiple time-points.</p>","PeriodicalId":21879,"journal":{"name":"Statistics in Medicine","volume":" ","pages":"4437-4455"},"PeriodicalIF":16.4,"publicationDate":"2024-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141856609","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-15Epub Date: 2024-08-07DOI: 10.1002/sim.10189
Yongwu Shao, Zhishen Ye, Zhiwei Zhang
The Cox proportional hazards model is commonly used to analyze time-to-event data in clinical trials. Standard inference procedures for the Cox model are based on asymptotic approximations and may perform poorly when there are few events in one or both treatment groups, as may be the case when the event of interest is rare or when the experimental treatment is highly efficacious. In this article, we propose an exact test of equivalence and efficacy under a proportional hazard model with treatment effect as the only fixed effect, together with an exact confidence interval that is obtained by inverting the exact test. The proposed test is based on a conditional error method originally proposed for sample size reestimation problems. In the present context, the conditional error method is used to combine information from a sequence of hypergeometric distributions, one at each observed event time. The proposed procedures are evaluated in simulation studies and illustrated using real data from an HIV prevention trial. A companion R package "ExactCox" is available for download on CRAN.
Cox 比例危险模型常用于分析临床试验中的时间到事件数据。Cox 模型的标准推断程序基于渐近线近似值,当一个或两个治疗组中的事件较少时,可能会表现不佳,而当感兴趣的事件罕见或试验性治疗非常有效时,情况就会如此。在本文中,我们提出了在以治疗效果为唯一固定效应的比例危险模型下,对等效性和疗效进行精确检验的方法,以及通过倒置精确检验得到的精确置信区间。本文提出的检验方法基于一种条件误差法,该方法最初是针对样本量再估计问题提出的。在目前的情况下,条件误差法被用于结合来自超几何分布序列的信息,每个观测事件时间都有一个超几何分布。我们在模拟研究中对所提出的程序进行了评估,并使用一项艾滋病预防试验的真实数据进行了说明。可在 CRAN 上下载配套的 R 软件包 "ExactCox"。
{"title":"Exact test and exact confidence interval for the Cox model.","authors":"Yongwu Shao, Zhishen Ye, Zhiwei Zhang","doi":"10.1002/sim.10189","DOIUrl":"10.1002/sim.10189","url":null,"abstract":"<p><p>The Cox proportional hazards model is commonly used to analyze time-to-event data in clinical trials. Standard inference procedures for the Cox model are based on asymptotic approximations and may perform poorly when there are few events in one or both treatment groups, as may be the case when the event of interest is rare or when the experimental treatment is highly efficacious. In this article, we propose an exact test of equivalence and efficacy under a proportional hazard model with treatment effect as the only fixed effect, together with an exact confidence interval that is obtained by inverting the exact test. The proposed test is based on a conditional error method originally proposed for sample size reestimation problems. In the present context, the conditional error method is used to combine information from a sequence of hypergeometric distributions, one at each observed event time. The proposed procedures are evaluated in simulation studies and illustrated using real data from an HIV prevention trial. A companion R package \"ExactCox\" is available for download on CRAN.</p>","PeriodicalId":21879,"journal":{"name":"Statistics in Medicine","volume":" ","pages":"4499-4518"},"PeriodicalIF":16.4,"publicationDate":"2024-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141898264","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-15Epub Date: 2024-08-15DOI: 10.1002/sim.10193
Shuyan Chen, Zhiqing Fang, Zhong Li, Xin Liu
Joint models for longitudinal and time-to-event data are receiving increasing attention owing to its capability of capturing the possible association between these two types of data. Typically, a joint model consists of a longitudinal submodel for longitudinal processes and a survival submodel for the time-to-event response, and links two submodels by common covariates that may carry both fixed and random effects. However, research gaps still remain on how to simultaneously select fixed and random effects from the two submodels under the joint modeling framework efficiently and effectively. In this article, we propose a novel block-coordinate gradient descent (BCGD) algorithm to simultaneously select multiple longitudinal covariates that may carry fixed and random effects in the joint model. Specifically, for the multiple longitudinal processes, a linear mixed effect model is adopted where random intercepts and slopes serve as essential covariates of the trajectories, and for the survival submodel, the popular proportional hazard model is employed. A penalized likelihood estimation is used to control the dimensionality of covariates in the joint model and estimate the unknown parameters, especially when estimating the covariance matrix of random effects. The proposed BCGD method can successfully capture the useful covariates of both fixed and random effects with excellent selection power, and efficiently provide a relatively accurate estimate of fixed and random effects empirically. The simulation results show excellent performance of the proposed method and support its effectiveness. The proposed BCGD method is further applied on two real data sets, and we examine the risk factors for the effects of different heart valves, differing on type of tissue, implanted in the aortic position and the risk factors for the diagnosis of primary biliary cholangitis.
{"title":"A novel block-coordinate gradient descent algorithm for simultaneous grouped selection of fixed and random effects in joint modeling.","authors":"Shuyan Chen, Zhiqing Fang, Zhong Li, Xin Liu","doi":"10.1002/sim.10193","DOIUrl":"10.1002/sim.10193","url":null,"abstract":"<p><p>Joint models for longitudinal and time-to-event data are receiving increasing attention owing to its capability of capturing the possible association between these two types of data. Typically, a joint model consists of a longitudinal submodel for longitudinal processes and a survival submodel for the time-to-event response, and links two submodels by common covariates that may carry both fixed and random effects. However, research gaps still remain on how to simultaneously select fixed and random effects from the two submodels under the joint modeling framework efficiently and effectively. In this article, we propose a novel block-coordinate gradient descent (BCGD) algorithm to simultaneously select multiple longitudinal covariates that may carry fixed and random effects in the joint model. Specifically, for the multiple longitudinal processes, a linear mixed effect model is adopted where random intercepts and slopes serve as essential covariates of the trajectories, and for the survival submodel, the popular proportional hazard model is employed. A penalized likelihood estimation is used to control the dimensionality of covariates in the joint model and estimate the unknown parameters, especially when estimating the covariance matrix of random effects. The proposed BCGD method can successfully capture the useful covariates of both fixed and random effects with excellent selection power, and efficiently provide a relatively accurate estimate of fixed and random effects empirically. The simulation results show excellent performance of the proposed method and support its effectiveness. The proposed BCGD method is further applied on two real data sets, and we examine the risk factors for the effects of different heart valves, differing on type of tissue, implanted in the aortic position and the risk factors for the diagnosis of primary biliary cholangitis.</p>","PeriodicalId":21879,"journal":{"name":"Statistics in Medicine","volume":" ","pages":"4595-4613"},"PeriodicalIF":16.4,"publicationDate":"2024-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141983250","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}