首页 > 最新文献

Statistical Methods in Medical Research最新文献

英文 中文
Permutation-based global rank test with adaptive weights for multiple primary endpoints. 基于置换的多主端点自适应权重全局秩检验。
IF 1.6 3区 医学 Q3 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-06-01 Epub Date: 2025-05-14 DOI: 10.1177/09622802251334886
Satoshi Yoshida, Yusuke Yamaguchi, Kazushi Maruo, Masahiko Gosho

Multiple efficacy endpoints are investigated in clinical trials, and selecting the appropriate primary endpoints is key to the study's success. The global test is an analysis approach that can handle multiple endpoints without multiplicity adjustment. This test, which aggregates the statistics from multiple primary endpoints into a single statistic using weights for the statistical comparison, has been gaining increasing attention. A key consideration in the global test is determination of the weights. In this study, we propose a novel global rank test in which the weights for each endpoint are estimated based on the current study data to maximize the test statistic, and the permutation test is applied to control the type I error rate. Simulation studies conducted to compare the proposed test with other global tests show that the proposed test can control the type I error rate at the nominal level, regardless of the number of primary endpoints and correlations between endpoints. Additionally, the proposed test offers higher statistical powers when the efficacy is considerably different between endpoints or when endpoints are moderately correlated, such as when the correlation coefficient is greater than or equal to 0.5.

临床试验研究了多个疗效终点,选择合适的主要终点是研究成功的关键。全局测试是一种可以处理多个端点而不需要进行多重性调整的分析方法。该测试使用权重将来自多个主要端点的统计信息聚合为一个统计信息,以进行统计比较,该测试已获得越来越多的关注。全局测试中的一个关键考虑因素是权重的确定。在本研究中,我们提出了一种新的全局秩检验方法,该方法基于当前研究数据估计每个端点的权重以最大化检验统计量,并应用置换检验来控制I型错误率。将提议的测试与其他全局测试进行比较的模拟研究表明,无论主要端点的数量和端点之间的相关性如何,提议的测试都可以将第一类错误率控制在名义水平上。此外,当终点之间的疗效差异很大或当终点适度相关时,例如相关系数大于或等于0.5时,建议的检验提供更高的统计能力。
{"title":"Permutation-based global rank test with adaptive weights for multiple primary endpoints.","authors":"Satoshi Yoshida, Yusuke Yamaguchi, Kazushi Maruo, Masahiko Gosho","doi":"10.1177/09622802251334886","DOIUrl":"10.1177/09622802251334886","url":null,"abstract":"<p><p>Multiple efficacy endpoints are investigated in clinical trials, and selecting the appropriate primary endpoints is key to the study's success. The global test is an analysis approach that can handle multiple endpoints without multiplicity adjustment. This test, which aggregates the statistics from multiple primary endpoints into a single statistic using weights for the statistical comparison, has been gaining increasing attention. A key consideration in the global test is determination of the weights. In this study, we propose a novel global rank test in which the weights for each endpoint are estimated based on the current study data to maximize the test statistic, and the permutation test is applied to control the type I error rate. Simulation studies conducted to compare the proposed test with other global tests show that the proposed test can control the type I error rate at the nominal level, regardless of the number of primary endpoints and correlations between endpoints. Additionally, the proposed test offers higher statistical powers when the efficacy is considerably different between endpoints or when endpoints are moderately correlated, such as when the correlation coefficient is greater than or equal to 0.5.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1254-1266"},"PeriodicalIF":1.6,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144080760","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Rank-based estimators of global treatment effects for cluster randomized trials with multiple endpoints on different scales. 在不同尺度上具有多个终点的聚类随机试验的总体治疗效果的基于秩的估计。
IF 1.9 3区 医学 Q3 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-06-01 Epub Date: 2025-05-14 DOI: 10.1177/09622802251338387
Emma Davies Smith, Vipul Jairath, Guangyong Zou

Cluster randomized trials commonly employ multiple endpoints. When a single summary of treatment effects across endpoints is of primary interest, global methods represent a common analysis strategy. However, specification of the required joint distribution is non-trivial, particularly when the endpoints have different scales. We develop rank-based interval estimators for a global treatment effect referred to here as the "global win probability, or the mean of multiple Wilcoxon Mann-Whitney probabilities, and interpreted as the probability that a treatment individual responds better than a control individual on average. Using endpoint-specific ranks among the combined sample and within each arm, each individual-level observation is converted to a "win fraction" which quantifies the proportion of wins experienced over every observation in the comparison arm. An individual's multiple observations are then replaced with a single "global win fraction" by averaging win fractions across endpoints. A linear mixed model is applied directly to the global win fractions to obtain point, variance, and interval estimates adjusted for clustering. Simulation demonstrates our approach performs well concerning confidence interval coverage and type I error, and methods are easily implemented using standard software. A case study using public data is provided with corresponding R and SAS code.

聚类随机试验通常采用多个终点。当主要关注的是跨端点的治疗效果的单一总结时,全局方法代表了一种常见的分析策略。然而,指定所需的联合分布是非平凡的,特别是当端点具有不同的尺度时。我们开发了基于等级的区间估计器,用于全局治疗效果,这里称为“全局获胜概率”,或多个Wilcoxon Mann-Whitney概率的平均值,并解释为治疗个体平均比对照个体反应更好的概率。在联合样本和每个组中使用特定终点的排名,每个个人水平的观察被转换为“获胜分数”,该分数量化了在比较组中每个观察中经历的获胜比例。然后,通过在端点上平均获胜分数,将个人的多个观察结果替换为单个“全局获胜分数”。线性混合模型直接应用于全局赢分数,以获得点,方差和区间估计调整聚类。仿真结果表明,该方法在置信区间覆盖和I型误差方面具有良好的性能,并且易于使用标准软件实现。提供了一个使用公共数据的案例研究,并提供了相应的R和SAS代码。
{"title":"Rank-based estimators of global treatment effects for cluster randomized trials with multiple endpoints on different scales.","authors":"Emma Davies Smith, Vipul Jairath, Guangyong Zou","doi":"10.1177/09622802251338387","DOIUrl":"10.1177/09622802251338387","url":null,"abstract":"<p><p>Cluster randomized trials commonly employ multiple endpoints. When a single summary of treatment effects across endpoints is of primary interest, global methods represent a common analysis strategy. However, specification of the required joint distribution is non-trivial, particularly when the endpoints have different scales. We develop rank-based interval estimators for a global treatment effect referred to here as the \"global win probability, or the mean of multiple Wilcoxon Mann-Whitney probabilities, and interpreted as the probability that a treatment individual responds better than a control individual on average. Using endpoint-specific ranks among the combined sample and within each arm, each individual-level observation is converted to a \"win fraction\" which quantifies the proportion of wins experienced over every observation in the comparison arm. An individual's multiple observations are then replaced with a single \"global win fraction\" by averaging win fractions across endpoints. A linear mixed model is applied directly to the global win fractions to obtain point, variance, and interval estimates adjusted for clustering. Simulation demonstrates our approach performs well concerning confidence interval coverage and type I error, and methods are easily implemented using standard software. A case study using public data is provided with corresponding R and SAS code.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1267-1289"},"PeriodicalIF":1.9,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12478089/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144080761","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
G-formula with multiple imputation for causal inference with incomplete data. 不完全数据下的多重归因g公式。
IF 1.6 3区 医学 Q3 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-06-01 Epub Date: 2025-03-31 DOI: 10.1177/09622802251316971
Jonathan W Bartlett, Camila Olarte Parra, Emily Granger, Ruth H Keogh, Erik W van Zwet, Rhian M Daniel

G-formula is a popular approach for estimating the effects of time-varying treatments or exposures from longitudinal data. G-formula is typically implemented using Monte-Carlo simulation, with non-parametric bootstrapping used for inference. In longitudinal data settings missing data are a common issue, which are often handled using multiple imputation, but it is unclear how G-formula and multiple imputation should be combined. We show how G-formula can be implemented using Bayesian multiple imputation methods for synthetic data, and that by doing so, we can impute missing data and simulate the counterfactuals of interest within a single coherent approach. We describe how this can be achieved using standard multiple imputation software and explore its performance using a simulation study and an application from cystic fibrosis.

g公式是一种流行的方法,用于估计时变处理或纵向数据暴露的影响。g公式通常使用蒙特卡罗模拟实现,非参数自举用于推理。在纵向数据设置中,数据缺失是一个常见的问题,通常使用多次输入来处理,但g公式和多次输入如何结合还不清楚。我们展示了如何使用合成数据的贝叶斯多重输入方法来实现g公式,并且通过这样做,我们可以输入缺失的数据并在单一连贯的方法中模拟感兴趣的反事实。我们描述了如何使用标准的多重植入软件来实现这一点,并通过模拟研究和囊性纤维化的应用来探索其性能。
{"title":"G-formula with multiple imputation for causal inference with incomplete data.","authors":"Jonathan W Bartlett, Camila Olarte Parra, Emily Granger, Ruth H Keogh, Erik W van Zwet, Rhian M Daniel","doi":"10.1177/09622802251316971","DOIUrl":"10.1177/09622802251316971","url":null,"abstract":"<p><p>G-formula is a popular approach for estimating the effects of time-varying treatments or exposures from longitudinal data. G-formula is typically implemented using Monte-Carlo simulation, with non-parametric bootstrapping used for inference. In longitudinal data settings missing data are a common issue, which are often handled using multiple imputation, but it is unclear how G-formula and multiple imputation should be combined. We show how G-formula can be implemented using Bayesian multiple imputation methods for synthetic data, and that by doing so, we can impute missing data and simulate the counterfactuals of interest within a single coherent approach. We describe how this can be achieved using standard multiple imputation software and explore its performance using a simulation study and an application from cystic fibrosis.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1130-1143"},"PeriodicalIF":1.6,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12209542/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143753917","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimising error rates in programmes of pilot and definitive trials using Bayesian statistical decision theory. 利用贝叶斯统计决策理论优化试点和最终试验方案的错误率。
IF 1.6 3区 医学 Q3 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-06-01 Epub Date: 2025-03-31 DOI: 10.1177/09622802251322987
Duncan T Wilson, Andrew Hall, Julia M Brown, Rebecca Ea Walwyn

Pilot trials are often conducted in advance of definitive trials to assess their feasibility and to inform their design. Although pilot trials typically collect primary endpoint data, preliminary tests of effectiveness have been discouraged given their typically low power. Power could be increased at the cost of a higher type I error rate, but there is little methodological guidance on how to determine the optimal balance between these operating characteristics. We consider a Bayesian decision-theoretic approach to this problem, introducing a utility function and defining an optimal pilot and definitive trial programme as that which maximises expected utility. We base utility on changes in average primary outcome, the cost of sampling, treatment costs, and the decision-maker's attitude to risk. We apply this approach to re-design OK-Diabetes, a pilot trial of a complex intervention with a continuous primary outcome with known standard deviation. We then examine how optimal programme characteristics vary with the parameters of the utility function. We find that the conventional approach of not testing for effectiveness in pilot trials can be considerably sub-optimal.

试点试验通常在确定性试验之前进行,以评估其可行性并为其设计提供依据。尽管先导试验通常会收集主要终点数据,但由于其功率通常较低,因此不鼓励对有效性进行初步测试。虽然可以通过提高 I 类错误率来提高试验的有效性,但如何在这些操作特征之间取得最佳平衡,目前还没有什么方法论指导。我们考虑采用贝叶斯决策理论方法来解决这一问题,引入效用函数,并将最佳试点和最终试验方案定义为预期效用最大化的方案。我们将平均主要结果的变化、取样成本、治疗成本以及决策者对风险的态度作为效用的基础。我们运用这种方法重新设计了 OK-糖尿病试验,这是一项复杂干预的试点试验,其主要结果是连续的,标准偏差已知。然后,我们研究了最佳方案特征如何随效用函数参数的变化而变化。我们发现,在试点试验中不测试有效性的传统方法可能在很大程度上不是最佳方法。
{"title":"Optimising error rates in programmes of pilot and definitive trials using Bayesian statistical decision theory.","authors":"Duncan T Wilson, Andrew Hall, Julia M Brown, Rebecca Ea Walwyn","doi":"10.1177/09622802251322987","DOIUrl":"10.1177/09622802251322987","url":null,"abstract":"<p><p>Pilot trials are often conducted in advance of definitive trials to assess their feasibility and to inform their design. Although pilot trials typically collect primary endpoint data, preliminary tests of effectiveness have been discouraged given their typically low power. Power could be increased at the cost of a higher type I error rate, but there is little methodological guidance on how to determine the optimal balance between these operating characteristics. We consider a Bayesian decision-theoretic approach to this problem, introducing a utility function and defining an optimal pilot and definitive trial programme as that which maximises expected utility. We base utility on changes in average primary outcome, the cost of sampling, treatment costs, and the decision-maker's attitude to risk. We apply this approach to re-design OK-Diabetes, a pilot trial of a complex intervention with a continuous primary outcome with known standard deviation. We then examine how optimal programme characteristics vary with the parameters of the utility function. We find that the conventional approach of not testing for effectiveness in pilot trials can be considerably sub-optimal.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1162-1177"},"PeriodicalIF":1.6,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12209550/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143754028","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An iterative matrix uncertainty selector for high-dimensional generalized linear models with measurement errors. 具有测量误差的高维广义线性模型的迭代矩阵不确定性选择器。
IF 1.6 3区 医学 Q3 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-06-01 Epub Date: 2025-03-19 DOI: 10.1177/09622802251316963
Betrand Fesuh Nono, Georges Nguefack-Tsague, Martin Kegnenlezom, Eugène-Patrice N Nguéma

Measurement error is a prevalent issue in high-dimensional generalized linear regression that existing regularization techniques may inadequately address. Most require estimating error distributions, which can be computationally prohibitive or unrealistic. We introduce an error distribution-free approach for variable selection called the Iterative Matrix Uncertainty Selector (IMUS). IMUS employs the matrix uncertainty selector framework for linear models, which is known for its selection consistency properties. It features an efficient iterative algorithm easily implemented for any generalized linear model within the exponential family. Empirically, we demonstrate that IMUS performs well in simulations and on three microarray gene expression datasets, achieving effective covariate selection with smoother convergence and clearer elbow criteria compared to other error distribution free methods. Notably, simulation studies in logistic and Poisson regression showed that IMUS exhibited smoother convergence and clearer elbow criteria, performing comparably to the Generalized Matrix Uncertainty Selector (GMUS) and Generalized Matrix Uncertainty Lasso (GMUL) in covariate selection. In many scenarios, IMUS had smaller estimation errors than GMUL and GMUS, measured by both the 1- and 2-norms. In applications to three microarray datasets with noisy measurements, GMUS faced convergence issues, while GMUL converged but lacked well-defined elbows for two datasets. In contrast, IMUS converged with well-defined elbows for all datasets, providing a potentially effective solution for high dimensional regression problems involving measurement errors.

测量误差是高维广义线性回归中普遍存在的问题,现有的正则化技术可能无法充分解决。大多数都需要估计误差分布,这在计算上是不允许的或不现实的。我们引入了一种误差无分布的变量选择方法,称为迭代矩阵不确定性选择器(IMUS)。IMUS对线性模型采用矩阵不确定性选择器框架,以其选择一致性而闻名。它的特点是一个有效的迭代算法,易于实现任何广义线性模型在指数族。经验表明,IMUS在模拟和三个微阵列基因表达数据集上表现良好,与其他无误差分布的方法相比,实现了有效的协变量选择,收敛更平滑,肘部标准更清晰。值得注意的是,逻辑回归和泊松回归的模拟研究表明,与广义矩阵不确定性选择器(GMUS)和广义矩阵不确定性套索(GMUL)相比,IMUS在协变量选择方面具有更平滑的收敛性和更清晰的肘部准则。在许多情况下,IMUS比GMUL和GMUS具有更小的估计误差,通过1和2规范测量。在三个带有噪声测量的微阵列数据集的应用中,GMUS面临收敛问题,而GMUL在两个数据集上收敛但缺乏明确的肘部。相比之下,IMUS收敛于所有数据集的定义良好的弯头,为涉及测量误差的高维回归问题提供了潜在的有效解决方案。
{"title":"An iterative matrix uncertainty selector for high-dimensional generalized linear models with measurement errors.","authors":"Betrand Fesuh Nono, Georges Nguefack-Tsague, Martin Kegnenlezom, Eugène-Patrice N Nguéma","doi":"10.1177/09622802251316963","DOIUrl":"10.1177/09622802251316963","url":null,"abstract":"<p><p>Measurement error is a prevalent issue in high-dimensional generalized linear regression that existing regularization techniques may inadequately address. Most require estimating error distributions, which can be computationally prohibitive or unrealistic. We introduce an error distribution-free approach for variable selection called the Iterative Matrix Uncertainty Selector (IMUS). IMUS employs the matrix uncertainty selector framework for linear models, which is known for its selection consistency properties. It features an efficient iterative algorithm easily implemented for any generalized linear model within the exponential family. Empirically, we demonstrate that IMUS performs well in simulations and on three microarray gene expression datasets, achieving effective covariate selection with smoother convergence and clearer elbow criteria compared to other error distribution free methods. Notably, simulation studies in logistic and Poisson regression showed that IMUS exhibited smoother convergence and clearer elbow criteria, performing comparably to the Generalized Matrix Uncertainty Selector (GMUS) and Generalized Matrix Uncertainty Lasso (GMUL) in covariate selection. In many scenarios, IMUS had smaller estimation errors than GMUL and GMUS, measured by both the 1- and 2-norms. In applications to three microarray datasets with noisy measurements, GMUS faced convergence issues, while GMUL converged but lacked well-defined elbows for two datasets. In contrast, IMUS converged with well-defined elbows for all datasets, providing a potentially effective solution for high dimensional regression problems involving measurement errors.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1114-1129"},"PeriodicalIF":1.6,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143664493","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sequential design for paired ordinal categorical outcome. 配对有序分类结果的序贯设计。
IF 1.6 3区 医学 Q3 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-06-01 Epub Date: 2025-03-31 DOI: 10.1177/09622802251322990
Baoshan Zhang, Yuan Wu

This study addresses a critical gap in the design of clinical trials that use grouped sequential designs for one-sample or paired ordinal categorical outcomes. Single-arm experiments, such as those using the modified Rankin Scale in stroke trials, underscore the necessity of our work. We present a novel method for applying the Wilcoxon signed-rank test to grouped sequences in these contexts. Our approach provides a practical and theoretical framework for assessing treatment effects, detailing variance formulas and demonstrating the asymptotic normality of the U-statistic. Through simulation studies and real data analysis, we validate the empirical Type I error rates and power. Additionally, we include a comprehensive flowchart to guide researchers in determining the required sample size to achieve specified power levels while controlling Type I error rates, thereby enhancing the design process of sequential trials.

本研究弥补了临床试验设计中的一个重要空白,即对单样本或配对序数分类结果采用分组序列设计。单臂试验(如在中风试验中使用修正的 Rankin 量表)强调了我们工作的必要性。我们提出了一种将 Wilcoxon 符号秩检验应用于这些情况下分组序列的新方法。我们的方法为评估治疗效果提供了一个实用的理论框架,详细说明了方差公式,并证明了 U 统计量的渐近正态性。通过模拟研究和真实数据分析,我们验证了经验 I 类错误率和功率。此外,我们还提供了一个全面的流程图,指导研究人员确定所需的样本量,以便在控制 I 类错误率的同时达到指定的功率水平,从而改进顺序试验的设计过程。
{"title":"Sequential design for paired ordinal categorical outcome.","authors":"Baoshan Zhang, Yuan Wu","doi":"10.1177/09622802251322990","DOIUrl":"10.1177/09622802251322990","url":null,"abstract":"<p><p>This study addresses a critical gap in the design of clinical trials that use grouped sequential designs for one-sample or paired ordinal categorical outcomes. Single-arm experiments, such as those using the modified Rankin Scale in stroke trials, underscore the necessity of our work. We present a novel method for applying the Wilcoxon signed-rank test to grouped sequences in these contexts. Our approach provides a practical and theoretical framework for assessing treatment effects, detailing variance formulas and demonstrating the asymptotic normality of the U-statistic. Through simulation studies and real data analysis, we validate the empirical Type I error rates and power. Additionally, we include a comprehensive flowchart to guide researchers in determining the required sample size to achieve specified power levels while controlling Type I error rates, thereby enhancing the design process of sequential trials.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1144-1161"},"PeriodicalIF":1.6,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143754224","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Estimand-based inference in the presence of long-term survivors. 在长期幸存者面前的基于估计的推断。
IF 1.6 3区 医学 Q3 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-06-01 Epub Date: 2025-03-31 DOI: 10.1177/09622802251327686
Yi-Cheng Tai, Weijing Wang, Martin T Wells

In this article, we develop nonparametric inference methods for comparing survival data across two samples, beneficial for clinical trials of novel cancer therapies where long-term survival is critical. These therapies, including immunotherapies and other advanced treatments, aim to establish durable effects. They often exhibit distinct survival patterns such as crossing or delayed separation and potentially leveling-off at the tails of survival curves, violating the proportional hazards assumption and rendering the hazard ratio inappropriate for measuring treatment effects. Our methodology uses the mixture cure framework to separately analyze cure rates of long-term survivors and the survival functions of susceptible individuals. We evaluated a nonparametric estimator for the susceptible survival function in a one-sample setting. Under sufficient follow-up, it is expressed as a location-scale-shift variant of the Kaplan-Meier estimator. It retains desirable features of the Kaplan-Meier estimator, including inverse-probability-censoring weighting, product-limit estimation, self-consistency, and nonparametric efficiency. Under insufficient follow-up, it can be adapted by incorporating a suitable cure rate estimator. In the two-sample setting, in addition to using the difference in cure rates to measure long-term effects, we propose a graphical estimand to compare relative treatment effects on susceptible subgroups. This process, inspired by Kendall's tau, compares the order of survival times among susceptible individuals. Large-sample properties of the proposed methods are derived for inference and their finite-sample properties are evaluated through simulations. The methodology is applied to analyze digitized data from the CheckMate 067 trial.

在本文中,我们开发了非参数推断方法来比较两个样本的生存数据,这有利于新型癌症治疗的临床试验,因为长期生存是至关重要的。这些疗法,包括免疫疗法和其他先进疗法,旨在建立持久的效果。它们通常表现出明显的生存模式,如交叉或延迟分离,并可能在生存曲线的尾部趋于平稳,违反了比例风险假设,使风险比不适合衡量治疗效果。我们的方法采用混合治疗框架,分别分析长期幸存者的治愈率和易感个体的生存功能。我们评估了单样本环境下易感生存函数的非参数估计量。在充分跟踪条件下,它被表示为Kaplan-Meier估计量的位置尺度位移变体。它保留了Kaplan-Meier估计量的理想特征,包括逆概率滤波加权、乘积极限估计、自一致性和非参数效率。在随访不足的情况下,可以通过纳入合适的治愈率估计器来调整。在两样本设置中,除了使用治愈率的差异来衡量长期效果外,我们还提出了一个图形估计来比较易感亚组的相对治疗效果。这个过程受到肯德尔tau的启发,比较了易感个体的生存时间顺序。推导了所提方法的大样本性质,并通过仿真对其有限样本性质进行了评价。该方法被应用于分析CheckMate 067试验的数字化数据。
{"title":"Estimand-based inference in the presence of long-term survivors.","authors":"Yi-Cheng Tai, Weijing Wang, Martin T Wells","doi":"10.1177/09622802251327686","DOIUrl":"10.1177/09622802251327686","url":null,"abstract":"<p><p>In this article, we develop nonparametric inference methods for comparing survival data across two samples, beneficial for clinical trials of novel cancer therapies where long-term survival is critical. These therapies, including immunotherapies and other advanced treatments, aim to establish durable effects. They often exhibit distinct survival patterns such as crossing or delayed separation and potentially leveling-off at the tails of survival curves, violating the proportional hazards assumption and rendering the hazard ratio inappropriate for measuring treatment effects. Our methodology uses the mixture cure framework to separately analyze cure rates of long-term survivors and the survival functions of susceptible individuals. We evaluated a nonparametric estimator for the susceptible survival function in a one-sample setting. Under sufficient follow-up, it is expressed as a location-scale-shift variant of the Kaplan-Meier estimator. It retains desirable features of the Kaplan-Meier estimator, including inverse-probability-censoring weighting, product-limit estimation, self-consistency, and nonparametric efficiency. Under insufficient follow-up, it can be adapted by incorporating a suitable cure rate estimator. In the two-sample setting, in addition to using the difference in cure rates to measure long-term effects, we propose a graphical estimand to compare relative treatment effects on susceptible subgroups. This process, inspired by Kendall's tau, compares the order of survival times among susceptible individuals. Large-sample properties of the proposed methods are derived for inference and their finite-sample properties are evaluated through simulations. The methodology is applied to analyze digitized data from the CheckMate 067 trial.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1178-1191"},"PeriodicalIF":1.6,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143753827","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A computationally efficient approach to false discovery rate control and power maximisation via randomisation and mirror statistic. 一种通过随机化和镜像统计实现错误发现率控制和功率最大化的高效计算方法。
IF 1.6 3区 医学 Q3 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-06-01 Epub Date: 2025-03-31 DOI: 10.1177/09622802251329768
Marco Molinari, Magne Thoresen

Simultaneously performing variable selection and inference in high-dimensional regression models is an open challenge in statistics and machine learning. The increasing availability of vast amounts of variables requires the adoption of specific statistical procedures to accurately select the most important predictors in a high-dimensional space, while controlling the false discovery rate (FDR) associated with the variable selection procedure. In this paper, we propose the joint adoption of the Mirror Statistic approach to FDR control, coupled with outcome randomisation to maximise the statistical power of the variable selection procedure, measured through the true positive rate. Through extensive simulations, we show how our proposed strategy allows us to combine the benefits of the two techniques. The Mirror Statistic is a flexible method to control FDR, which only requires mild model assumptions, but requires two sets of independent regression coefficient estimates, usually obtained after splitting the original dataset. Outcome randomisation is an alternative to data splitting that allows to generate two independent outcomes, which can then be used to estimate the coefficients that go into the construction of the Mirror Statistic. The combination of these two approaches provides increased testing power in a number of scenarios, such as highly correlated covariates and high percentages of active variables. Moreover, it is scalable to very high-dimensional problems, since the algorithm has a low memory footprint and only requires a single run on the full dataset, as opposed to iterative alternatives such as multiple data splitting.

在高维回归模型中同时进行变量选择和推理是统计学和机器学习中的一个开放挑战。大量变量的可用性不断增加,需要采用特定的统计程序来准确地选择高维空间中最重要的预测因子,同时控制与变量选择过程相关的错误发现率(FDR)。在本文中,我们建议联合采用镜像统计方法来控制FDR,再加上结果随机化,以最大限度地提高变量选择过程的统计能力,通过真阳性率来衡量。通过广泛的模拟,我们展示了我们提出的策略如何使我们能够结合这两种技术的优点。镜像统计是一种灵活的控制FDR的方法,它只需要温和的模型假设,但需要两组独立的回归系数估计,通常是在原始数据集分裂后得到的。结果随机化是数据分割的另一种选择,它允许生成两个独立的结果,然后可以用来估计进入镜像统计构建的系数。这两种方法的结合在许多场景中提供了更高的测试能力,例如高度相关的协变量和高百分比的活动变量。此外,它可以扩展到非常高维的问题,因为该算法具有低内存占用,并且只需要在整个数据集上运行一次,而不是迭代替代方案,如多次数据分割。
{"title":"A computationally efficient approach to false discovery rate control and power maximisation via randomisation and mirror statistic.","authors":"Marco Molinari, Magne Thoresen","doi":"10.1177/09622802251329768","DOIUrl":"10.1177/09622802251329768","url":null,"abstract":"<p><p>Simultaneously performing variable selection and inference in high-dimensional regression models is an open challenge in statistics and machine learning. The increasing availability of vast amounts of variables requires the adoption of specific statistical procedures to accurately select the most important predictors in a high-dimensional space, while controlling the false discovery rate (FDR) associated with the variable selection procedure. In this paper, we propose the joint adoption of the Mirror Statistic approach to FDR control, coupled with outcome randomisation to maximise the statistical power of the variable selection procedure, measured through the true positive rate. Through extensive simulations, we show how our proposed strategy allows us to combine the benefits of the two techniques. The Mirror Statistic is a flexible method to control FDR, which only requires mild model assumptions, but requires two sets of independent regression coefficient estimates, usually obtained after splitting the original dataset. Outcome randomisation is an alternative to data splitting that allows to generate two independent outcomes, which can then be used to estimate the coefficients that go into the construction of the Mirror Statistic. The combination of these two approaches provides increased testing power in a number of scenarios, such as highly correlated covariates and high percentages of active variables. Moreover, it is scalable to very high-dimensional problems, since the algorithm has a low memory footprint and only requires a single run on the full dataset, as opposed to iterative alternatives such as multiple data splitting.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1233-1253"},"PeriodicalIF":1.6,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12209545/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143754551","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dose selection criteria to identify the optimal dose based on ranked efficacy-toxicity outcomes without reliance on clinical utilities. 剂量选择标准,根据分级的疗效-毒性结果确定最佳剂量,而不依赖于临床效用。
IF 1.6 3区 医学 Q3 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-06-01 Epub Date: 2025-03-31 DOI: 10.1177/09622802251327691
Sydney Porter, Anne Eaton, Thomas A Murray

Recently, targeted and immunotherapy cancer treatments have motivated dose-finding based on efficacy-toxicity trade-offs rather than toxicity alone. The EffTox and utility-based Bayesian optimal interval (U-BOIN) dose-finding designs were developed in response to this need, but may be sensitive to elicited subjective design parameters that reflect the trade-off between efficacy and toxicity. To ease elicitation and reduce subjectivity, we propose dose desirability criteria that only depend on a preferential ordering of the joint efficacy-toxicity outcomes. We propose two novel order-based criteria and compare them with utility-based and contour-based criteria when paired with the design framework and probability models of EffTox and U-BOIN. The proposed dose desirability criteria simplify implementation and improve robustness to the elicited subjective design parameters and perform similarly in simulation studies to the established EffTox and U-BOIN designs when the ordering of the joint outcomes is equivalent. We also propose an alternative dose admissibility criteria based on the joint efficacy and toxicity profile of a dose rather than its marginal toxicity and efficacy profile. We argue that this alternative joint criterion is more consistent with defining dose desirability in terms of efficacy-toxicity trade-offs than the standard marginal admissibility criteria. The proposed methods enhance the usability and robustness of dose-finding designs that account for efficacy-toxicity trade-offs to identify the optimal biological dose.

最近,靶向和免疫疗法的癌症治疗已经开始基于疗效和毒性的权衡来确定剂量,而不仅仅是毒性。EffTox和基于效用的贝叶斯最优区间(U-BOIN)剂量发现设计是为了满足这一需求而开发的,但可能对反映疗效和毒性之间权衡的主观设计参数很敏感。为了简化引出和减少主观性,我们提出了剂量可取性标准,该标准仅依赖于联合疗效-毒性结果的优先顺序。我们提出了两个新的基于顺序的标准,并将它们与基于效用和基于轮廓的标准结合EffTox和U-BOIN的设计框架和概率模型进行了比较。所提出的剂量理想性准则简化了实施,提高了对所得主观设计参数的鲁棒性,并且在模拟研究中,当联合结果的顺序相等时,与已建立的EffTox和U-BOIN设计相似。我们还提出了一种替代剂量允许标准,基于剂量的联合疗效和毒性概况,而不是其边际毒性和疗效概况。我们认为,与标准的边际可接受性标准相比,这一替代性联合标准更符合在药效-毒性权衡方面定义剂量的可取性。提出的方法提高了剂量发现设计的可用性和稳健性,这些设计考虑了药效-毒性权衡,以确定最佳生物剂量。
{"title":"Dose selection criteria to identify the optimal dose based on ranked efficacy-toxicity outcomes without reliance on clinical utilities.","authors":"Sydney Porter, Anne Eaton, Thomas A Murray","doi":"10.1177/09622802251327691","DOIUrl":"10.1177/09622802251327691","url":null,"abstract":"<p><p>Recently, targeted and immunotherapy cancer treatments have motivated dose-finding based on efficacy-toxicity trade-offs rather than toxicity alone. The EffTox and utility-based Bayesian optimal interval (U-BOIN) dose-finding designs were developed in response to this need, but may be sensitive to elicited subjective design parameters that reflect the trade-off between efficacy and toxicity. To ease elicitation and reduce subjectivity, we propose dose desirability criteria that only depend on a preferential ordering of the joint efficacy-toxicity outcomes. We propose two novel order-based criteria and compare them with utility-based and contour-based criteria when paired with the design framework and probability models of EffTox and U-BOIN. The proposed dose desirability criteria simplify implementation and improve robustness to the elicited subjective design parameters and perform similarly in simulation studies to the established EffTox and U-BOIN designs when the ordering of the joint outcomes is equivalent. We also propose an alternative dose admissibility criteria based on the joint efficacy and toxicity profile of a dose rather than its marginal toxicity and efficacy profile. We argue that this alternative joint criterion is more consistent with defining dose desirability in terms of efficacy-toxicity trade-offs than the standard marginal admissibility criteria. The proposed methods enhance the usability and robustness of dose-finding designs that account for efficacy-toxicity trade-offs to identify the optimal biological dose.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1219-1232"},"PeriodicalIF":1.6,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143754568","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Weibull mixture cure frailty model for high-dimensional covariates. 高维协变量的Weibull混合治疗脆弱性模型。
IF 1.6 3区 医学 Q3 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-06-01 Epub Date: 2025-03-31 DOI: 10.1177/09622802251327687
Fatih Kızılaslan, David Michael Swanson, Valeria Vitelli

A novel mixture cure frailty model is introduced for handling censored survival data. Mixture cure models are preferable when the existence of a cured fraction among patients can be assumed. However, such models are heavily underexplored: frailty structures within cure models remain largely undeveloped, and furthermore, most existing methods do not work for high-dimensional datasets, when the number of predictors is significantly larger than the number of observations. In this study, we introduce a novel extension of the Weibull mixture cure model that incorporates a frailty component, employed to model an underlying latent population heterogeneity with respect to the outcome risk. Additionally, high-dimensional covariates are integrated into both the cure rate and survival part of the model, providing a comprehensive approach to employ the model in the context of high-dimensional omics data. We also perform variable selection via an adaptive elastic-net penalization, and propose a novel approach to inference using the expectation-maximization (EM) algorithm. Extensive simulation studies are conducted across various scenarios to demonstrate the performance of the model, and results indicate that our proposed method outperforms competitor models. We apply the novel approach to analyze RNAseq gene expression data from bulk breast cancer patients included in The Cancer Genome Atlas (TCGA) database. A set of prognostic biomarkers is then derived from selected genes, and subsequently validated via both functional enrichment analysis and comparison to the existing biological literature. Finally, a prognostic risk score index based on the identified biomarkers is proposed and validated by exploring the patients' survival.

提出了一种新的混合治疗脆弱性模型,用于处理截尾存活数据。当可以假设患者中存在治愈部分时,混合治疗模型是优选的。然而,这样的模型还未被充分开发:治疗模型中的脆弱结构在很大程度上仍未开发,而且,大多数现有方法不适用于高维数据集,当预测因子的数量明显大于观测值的数量时。在这项研究中,我们引入了Weibull混合治疗模型的新扩展,该模型包含了一个脆弱性成分,用于模拟潜在的群体异质性,涉及结果风险。此外,高维协变量被整合到模型的治愈率和生存率部分,为在高维组学数据的背景下使用该模型提供了一种全面的方法。我们还通过自适应弹性网络惩罚进行变量选择,并提出了一种使用期望最大化(EM)算法进行推理的新方法。在各种情况下进行了广泛的仿真研究,以证明模型的性能,结果表明我们提出的方法优于竞争对手的模型。我们应用这种新方法分析了癌症基因组图谱(TCGA)数据库中大量乳腺癌患者的RNAseq基因表达数据。然后从选定的基因中衍生出一组预后生物标志物,随后通过功能富集分析和与现有生物学文献的比较进行验证。最后,提出了基于识别的生物标志物的预后风险评分指数,并通过探索患者的生存来验证。
{"title":"A Weibull mixture cure frailty model for high-dimensional covariates.","authors":"Fatih Kızılaslan, David Michael Swanson, Valeria Vitelli","doi":"10.1177/09622802251327687","DOIUrl":"10.1177/09622802251327687","url":null,"abstract":"<p><p>A novel mixture cure frailty model is introduced for handling censored survival data. Mixture cure models are preferable when the existence of a cured fraction among patients can be assumed. However, such models are heavily underexplored: frailty structures within cure models remain largely undeveloped, and furthermore, most existing methods do not work for high-dimensional datasets, when the number of predictors is significantly larger than the number of observations. In this study, we introduce a novel extension of the Weibull mixture cure model that incorporates a frailty component, employed to model an underlying latent population heterogeneity with respect to the outcome risk. Additionally, high-dimensional covariates are integrated into both the cure rate and survival part of the model, providing a comprehensive approach to employ the model in the context of high-dimensional omics data. We also perform variable selection via an adaptive elastic-net penalization, and propose a novel approach to inference using the expectation-maximization (EM) algorithm. Extensive simulation studies are conducted across various scenarios to demonstrate the performance of the model, and results indicate that our proposed method outperforms competitor models. We apply the novel approach to analyze RNAseq gene expression data from bulk breast cancer patients included in The Cancer Genome Atlas (TCGA) database. A set of prognostic biomarkers is then derived from selected genes, and subsequently validated via both functional enrichment analysis and comparison to the existing biological literature. Finally, a prognostic risk score index based on the identified biomarkers is proposed and validated by exploring the patients' survival.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1192-1218"},"PeriodicalIF":1.6,"publicationDate":"2025-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12209551/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143754552","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Statistical Methods in Medical Research
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1