Pub Date : 2024-08-01Epub Date: 2024-06-12DOI: 10.1177/09622802241259174
Yilei Ma, Youpeng Su, Peng Wang, Ping Yin
Estimation of the 100p percent lethal dose () is of great interest to pharmacologists for assessing the toxicity of certain compounds. However, most existing literature focuses on the interval estimation of and little attention has been paid to its point estimation. Currently, the most commonly used method for estimating the is the maximum likelihood estimator (MLE), which can be represented as a ratio estimator, with the denominator being the slope estimated from the logistic regression model. However, the MLE can be seriously biased when the sample size is small, a common nature in such studies, or when the dose-response curve is relatively flat (i.e. the slope approaches zero). In this study, we address these issues by developing a novel penalised maximum likelihood estimator (PMLE) that can prevent the denominator of the ratio from being close to zero. Similar to the MLE, the PMLE is computationally simple and thus can be conveniently used in practice. Moreover, with a suitable penalty parameter, we show that the PMLE can (a) reduce the bias to the second order with respect to the sample size and (b) avoid extreme estimates. Through simulation studies and real data applications, we show that the PMLE generally outperforms the existing methods in terms of bias and root mean square error.
{"title":"Point estimation of the 100<i>p</i> percent lethal dose using a novel penalised likelihood approach.","authors":"Yilei Ma, Youpeng Su, Peng Wang, Ping Yin","doi":"10.1177/09622802241259174","DOIUrl":"10.1177/09622802241259174","url":null,"abstract":"<p><p>Estimation of the 100<i>p</i> percent lethal dose (<math><msub><mtext>LD</mtext><mrow><mn>100</mn><mi>p</mi></mrow></msub></math>) is of great interest to pharmacologists for assessing the toxicity of certain compounds. However, most existing literature focuses on the interval estimation of <math><msub><mtext>LD</mtext><mrow><mn>100</mn><mi>p</mi></mrow></msub></math> and little attention has been paid to its point estimation. Currently, the most commonly used method for estimating the <math><msub><mtext>LD</mtext><mrow><mn>100</mn><mi>p</mi></mrow></msub></math> is the maximum likelihood estimator (MLE), which can be represented as a ratio estimator, with the denominator being the slope estimated from the logistic regression model. However, the MLE can be seriously biased when the sample size is small, a common nature in such studies, or when the dose-response curve is relatively flat (i.e. the slope approaches zero). In this study, we address these issues by developing a novel penalised maximum likelihood estimator (PMLE) that can prevent the denominator of the ratio from being close to zero. Similar to the MLE, the PMLE is computationally simple and thus can be conveniently used in practice. Moreover, with a suitable penalty parameter, we show that the PMLE can (a) reduce the bias to the second order with respect to the sample size and (b) avoid extreme estimates. Through simulation studies and real data applications, we show that the PMLE generally outperforms the existing methods in terms of bias and root mean square error.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1331-1341"},"PeriodicalIF":1.6,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141306885","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01Epub Date: 2024-07-25DOI: 10.1177/09622802241262521
Xinyang Jiang, Wen Li, Kang Wang, Ruosha Li, Jing Ning
This study investigates the heterogeneity of a biomarker's discriminative performance for predicting subsequent time-to-event outcomes across different patient subgroups. While the area under the curve (AUC) for the time-dependent receiver operating characteristic curve is commonly used to assess biomarker performance, the partial time-dependent AUC (PAUC) provides insights that are often more pertinent for population screening and diagnostic testing. To achieve this objective, we propose a regression model tailored for PAUC and develop two distinct estimation procedures for discrete and continuous covariates, employing a pseudo-partial likelihood method. Simulation studies are conducted to assess the performance of these procedures across various scenarios. We apply our model and inference procedure to the Alzheimer's Disease Neuroimaging Initiative data set to evaluate potential heterogeneities in the discriminative performance of biomarkers for early Alzheimer's disease diagnosis based on patients' characteristics.
本研究调查了生物标记物在不同患者亚群中预测后续时间到事件结果的鉴别性能的异质性。虽然随时间变化的接收者操作特征曲线的曲线下面积(AUC)通常用于评估生物标记物的性能,但部分随时间变化的AUC(PAUC)提供的见解往往与人群筛查和诊断检测更相关。为实现这一目标,我们提出了一个为 PAUC 量身定制的回归模型,并采用伪偏似方法为离散和连续协变量开发了两种不同的估计程序。我们进行了模拟研究,以评估这些程序在各种情况下的性能。我们将模型和推理过程应用于阿尔茨海默病神经影像倡议数据集,以评估基于患者特征的早期阿尔茨海默病诊断生物标记物的鉴别性能的潜在异质性。
{"title":"Analyzing heterogeneity in biomarker discriminative performance through partial time-dependent receiver operating characteristic curve modeling.","authors":"Xinyang Jiang, Wen Li, Kang Wang, Ruosha Li, Jing Ning","doi":"10.1177/09622802241262521","DOIUrl":"10.1177/09622802241262521","url":null,"abstract":"<p><p>This study investigates the heterogeneity of a biomarker's discriminative performance for predicting subsequent time-to-event outcomes across different patient subgroups. While the area under the curve (AUC) for the time-dependent receiver operating characteristic curve is commonly used to assess biomarker performance, the partial time-dependent AUC (PAUC) provides insights that are often more pertinent for population screening and diagnostic testing. To achieve this objective, we propose a regression model tailored for PAUC and develop two distinct estimation procedures for discrete and continuous covariates, employing a pseudo-partial likelihood method. Simulation studies are conducted to assess the performance of these procedures across various scenarios. We apply our model and inference procedure to the Alzheimer's Disease Neuroimaging Initiative data set to evaluate potential heterogeneities in the discriminative performance of biomarkers for early Alzheimer's disease diagnosis based on patients' characteristics.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1424-1436"},"PeriodicalIF":1.6,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11449645/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141760988","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01Epub Date: 2024-07-25DOI: 10.1177/09622802241262527
Guy Cafri, Stephen Fortin, Peter C Austin
Observational studies are frequently used in clinical research to estimate the effects of treatments or exposures on outcomes. To reduce the effects of confounding when estimating treatment effects, covariate balancing methods are frequently implemented. This study evaluated, using extensive Monte Carlo simulation, several methods of covariate balancing, and two methods for propensity score estimation, for estimating the average treatment effect on the treated using a hazard ratio from a Cox proportional hazards model. With respect to minimizing bias and maximizing accuracy (as measured by the mean square error) of the treatment effect, the average treatment effect on the treated weighting, fine stratification, and optimal full matching with a conventional logistic regression model for the propensity score performed best across all simulated conditions. Other methods performed well in specific circumstances, such as pair matching when sample sizes were large (n = 5000) and the proportion treated was < 0.25. Statistical power was generally higher for weighting methods than matching methods, and Type I error rates were at or below the nominal level for balancing methods with unbiased treatment effect estimates. There was also a decreasing effective sample size with an increasing number of strata, therefore for stratification-based weighting methods, it may be important to consider fewer strata. Generally, we recommend methods that performed well in our simulations, although the identification of methods that performed well is necessarily limited by the specific features of our simulation. The methods are illustrated using a real-world example comparing beta blockers and angiotensin-converting enzyme inhibitors among hypertensive patients at risk for incident stroke.
{"title":"Minimizing confounding in comparative observational studies with time-to-event outcomes: An extensive comparison of covariate balancing methods using Monte Carlo simulation.","authors":"Guy Cafri, Stephen Fortin, Peter C Austin","doi":"10.1177/09622802241262527","DOIUrl":"10.1177/09622802241262527","url":null,"abstract":"<p><p>Observational studies are frequently used in clinical research to estimate the effects of treatments or exposures on outcomes. To reduce the effects of confounding when estimating treatment effects, covariate balancing methods are frequently implemented. This study evaluated, using extensive Monte Carlo simulation, several methods of covariate balancing, and two methods for propensity score estimation, for estimating the average treatment effect on the treated using a hazard ratio from a Cox proportional hazards model. With respect to minimizing bias and maximizing accuracy (as measured by the mean square error) of the treatment effect, the average treatment effect on the treated weighting, fine stratification, and optimal full matching with a conventional logistic regression model for the propensity score performed best across all simulated conditions. Other methods performed well in specific circumstances, such as pair matching when sample sizes were large (n = 5000) and the proportion treated was <u><</u> 0.25. Statistical power was generally higher for weighting methods than matching methods, and Type I error rates were at or below the nominal level for balancing methods with unbiased treatment effect estimates. There was also a decreasing effective sample size with an increasing number of strata, therefore for stratification-based weighting methods, it may be important to consider fewer strata. Generally, we recommend methods that performed well in our simulations, although the identification of methods that performed well is necessarily limited by the specific features of our simulation. The methods are illustrated using a real-world example comparing beta blockers and angiotensin-converting enzyme inhibitors among hypertensive patients at risk for incident stroke.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1437-1460"},"PeriodicalIF":1.6,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141760992","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01Epub Date: 2024-07-25DOI: 10.1177/09622802241262523
Harrison T Reeder, Sebastien Haneuse, Kyu Ha Lee
An important task in health research is to characterize time-to-event outcomes such as disease onset or mortality in terms of a potentially high-dimensional set of risk factors. For example, prospective cohort studies of Alzheimer's disease (AD) typically enroll older adults for observation over several decades to assess the long-term impact of genetic and other factors on cognitive decline and mortality. The accelerated failure time model is particularly well-suited to such studies, structuring covariate effects as "horizontal" changes to the survival quantiles that conceptually reflect shifts in the outcome distribution due to lifelong exposures. However, this modeling task is complicated by the enrollment of adults at differing ages, and intermittent follow-up visits leading to interval-censored outcome information. Moreover, genetic and clinical risk factors are not only high-dimensional, but characterized by underlying grouping structures, such as by function or gene location. Such grouped high-dimensional covariates require shrinkage methods that directly acknowledge this structure to facilitate variable selection and estimation. In this paper, we address these considerations directly by proposing a Bayesian accelerated failure time model with a group-structured lasso penalty, designed for left-truncated and interval-censored time-to-event data. We develop an R package with a Markov chain Monte Carlo sampler for estimation. We present a simulation study examining the performance of this method relative to an ordinary lasso penalty and apply the proposed method to identify groups of predictive genetic and clinical risk factors for AD in the Religious Orders Study and Memory and Aging Project prospective cohort studies of AD and dementia.
健康研究的一项重要任务是根据一组潜在的高维风险因素来描述疾病发病或死亡率等时间到事件的结果。例如,阿尔茨海默病(AD)的前瞻性队列研究通常会招募老年人进行数十年的观察,以评估遗传和其他因素对认知能力下降和死亡率的长期影响。加速失效时间模型特别适合此类研究,它将协变量效应结构化为生存量化值的 "水平 "变化,从概念上反映了终生暴露导致的结果分布变化。然而,这项建模任务因不同年龄段的成人入组和间歇性随访而变得复杂,这导致了间隔删失的结果信息。此外,遗传和临床风险因素不仅是高维的,而且具有潜在的分组结构,如按功能或基因位置分组。这种分组的高维协变量需要直接承认这种结构的收缩方法,以促进变量的选择和估计。在本文中,我们提出了一种具有分组结构套索惩罚的贝叶斯加速失效时间模型,该模型专为左截断和区间校验的时间到事件数据而设计,从而直接解决了这些问题。我们开发了一个 R 软件包,其中包含一个用于估计的马尔科夫链蒙特卡罗采样器。我们介绍了一项模拟研究,检验了该方法相对于普通拉索惩罚的性能,并将所提出的方法应用于在 "宗教习俗研究"(Religious Orders Study)和 "记忆与老龄化项目"(Memory and Aging Project)的前瞻性队列研究中识别AD的预测性遗传和临床风险因素组。
{"title":"Group lasso priors for Bayesian accelerated failure time models with left-truncated and interval-censored data.","authors":"Harrison T Reeder, Sebastien Haneuse, Kyu Ha Lee","doi":"10.1177/09622802241262523","DOIUrl":"10.1177/09622802241262523","url":null,"abstract":"<p><p>An important task in health research is to characterize time-to-event outcomes such as disease onset or mortality in terms of a potentially high-dimensional set of risk factors. For example, prospective cohort studies of Alzheimer's disease (AD) typically enroll older adults for observation over several decades to assess the long-term impact of genetic and other factors on cognitive decline and mortality. The accelerated failure time model is particularly well-suited to such studies, structuring covariate effects as \"horizontal\" changes to the survival quantiles that conceptually reflect shifts in the outcome distribution due to lifelong exposures. However, this modeling task is complicated by the enrollment of adults at differing ages, and intermittent follow-up visits leading to interval-censored outcome information. Moreover, genetic and clinical risk factors are not only high-dimensional, but characterized by underlying grouping structures, such as by function or gene location. Such grouped high-dimensional covariates require shrinkage methods that directly acknowledge this structure to facilitate variable selection and estimation. In this paper, we address these considerations directly by proposing a Bayesian accelerated failure time model with a group-structured lasso penalty, designed for left-truncated and interval-censored time-to-event data. We develop an R package with a Markov chain Monte Carlo sampler for estimation. We present a simulation study examining the performance of this method relative to an ordinary lasso penalty and apply the proposed method to identify groups of predictive genetic and clinical risk factors for AD in the Religious Orders Study and Memory and Aging Project prospective cohort studies of AD and dementia.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1412-1423"},"PeriodicalIF":1.6,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11833807/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141760991","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01Epub Date: 2022-02-24DOI: 10.1177/09622802211070256
Tasnim Hamza, Toshi A Furukawa, Nicola Orsini, Andrea Cipriani, Cynthia P Iglesias, Georgia Salanti
Network meta-analysis has been used to answer a range of clinical questions about the preferred intervention for a given condition. Although the effectiveness and safety of pharmacological agents depend on the dose administered, network meta-analysis applications typically ignore the role that drugs dosage plays in the results. This leads to more heterogeneity in the network. In this paper, we present a suite of network meta-analysis models that incorporate the dose-effect relationship using restricted cubic splines. We extend existing models into a dose-effect network meta-regression to account for study-level covariates and for groups of agents in a class-effect dose-effect network meta-analysis model. We apply our models to a network of aggregate data about the efficacy of 21 antidepressants and placebo for depression. We find that all antidepressants are more efficacious than placebo after a certain dose. Also, we identify the dose level at which each antidepressant's effect exceeds that of placebo and estimate the dose beyond which the effect of antidepressants no longer increases. When covariates were introduced to the model, we find that studies with small sample size tend to exaggerate antidepressants efficacy for several of the drugs. Our dose-effect network meta-analysis model with restricted cubic splines provides a flexible approach to modelling the dose-effect relationship in multiple interventions. Decision-makers can use our model to inform treatment choice.
{"title":"A dose-effect network meta-analysis model with application in antidepressants using restricted cubic splines.","authors":"Tasnim Hamza, Toshi A Furukawa, Nicola Orsini, Andrea Cipriani, Cynthia P Iglesias, Georgia Salanti","doi":"10.1177/09622802211070256","DOIUrl":"10.1177/09622802211070256","url":null,"abstract":"<p><p>Network meta-analysis has been used to answer a range of clinical questions about the preferred intervention for a given condition. Although the effectiveness and safety of pharmacological agents depend on the dose administered, network meta-analysis applications typically ignore the role that drugs dosage plays in the results. This leads to more heterogeneity in the network. In this paper, we present a suite of network meta-analysis models that incorporate the dose-effect relationship using restricted cubic splines. We extend existing models into a dose-effect network meta-regression to account for study-level covariates and for groups of agents in a class-effect dose-effect network meta-analysis model. We apply our models to a network of aggregate data about the efficacy of 21 antidepressants and placebo for depression. We find that all antidepressants are more efficacious than placebo after a certain dose. Also, we identify the dose level at which each antidepressant's effect exceeds that of placebo and estimate the dose beyond which the effect of antidepressants no longer increases. When covariates were introduced to the model, we find that studies with small sample size tend to exaggerate antidepressants efficacy for several of the drugs. Our dose-effect network meta-analysis model with restricted cubic splines provides a flexible approach to modelling the dose-effect relationship in multiple interventions. Decision-makers can use our model to inform treatment choice.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1461-1472"},"PeriodicalIF":1.6,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11462779/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"39824807","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-01Epub Date: 2024-07-25DOI: 10.1177/09622802241259177
Florian Stijven, Ariel Alonso, Geert Molenberghs
The selection of the primary endpoint in a clinical trial plays a critical role in determining the trial's success. Ideally, the primary endpoint is the clinically most relevant outcome, also termed the true endpoint. However, practical considerations, like extended follow-up, may complicate this choice, prompting the proposal to replace the true endpoint with so-called surrogate endpoints. Evaluating the validity of these surrogate endpoints is crucial, and a popular evaluation framework is based on the proportion of treatment effect explained (PTE). While methodological advancements in this area have focused primarily on estimation methods, interpretation remains a challenge hindering the practical use of the PTE. We review various ways to interpret the PTE. These interpretations-two causal and one non-causal-reveal connections between the PTE principal surrogacy, causal mediation analysis, and the prediction of trial-level treatment effects. A common limitation across these interpretations is the reliance on unverifiable assumptions. As such, we argue that the PTE is only meaningful when researchers are willing to make very strong assumptions. These challenges are also illustrated in an analysis of three hypothetical vaccine trials.
{"title":"Proportion of treatment effect explained: An overview of interpretations.","authors":"Florian Stijven, Ariel Alonso, Geert Molenberghs","doi":"10.1177/09622802241259177","DOIUrl":"10.1177/09622802241259177","url":null,"abstract":"<p><p>The selection of the primary endpoint in a clinical trial plays a critical role in determining the trial's success. Ideally, the primary endpoint is the clinically most relevant outcome, also termed the true endpoint. However, practical considerations, like extended follow-up, may complicate this choice, prompting the proposal to replace the true endpoint with so-called surrogate endpoints. Evaluating the validity of these surrogate endpoints is crucial, and a popular evaluation framework is based on the proportion of treatment effect explained (PTE). While methodological advancements in this area have focused primarily on estimation methods, interpretation remains a challenge hindering the practical use of the PTE. We review various ways to interpret the PTE. These interpretations-two causal and one non-causal-reveal connections between the PTE principal surrogacy, causal mediation analysis, and the prediction of trial-level treatment effects. A common limitation across these interpretations is the reliance on unverifiable assumptions. As such, we argue that the PTE is only meaningful when researchers are willing to make very strong assumptions. These challenges are also illustrated in an analysis of three hypothetical vaccine trials.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1278-1296"},"PeriodicalIF":1.6,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141760993","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-01Epub Date: 2024-05-08DOI: 10.1177/09622802241247719
Xinzhou Guo, Florence T Bourgeois, Tianxi Cai
When the primary endpoints in randomized clinical trials require long term follow-up or are costly to measure, it is often desirable to assess treatment effects on surrogate instead of clinical endpoints. Prior to adopting a surrogate endpoint for such purposes, the extent of its surrogacy on the primary endpoint must be assessed. There is a rich statistical literature on assessing surrogacy in the overall population, much of which is based on quantifying the proportion of treatment effect on the primary endpoint that is explained by the treatment effect on the surrogate endpoint. However, the surrogacy of an endpoint may vary across different patient subgroups according to baseline demographic characteristics, and limited methods are currently available to assess overall surrogacy in the presence of potential surrogacy heterogeneity. In this paper, we propose methods that incorporate covariates for baseline information, such as age, to improve overall surrogacy assessment. We use flexible semi-non-parametric modeling strategies to adjust for covariate effects and derive a robust estimate for the proportion of treatment effect of the covariate-adjusted surrogate endpoint. Simulation results suggest that the adjusted surrogate endpoint has greater proportion of treatment effect compared to the unadjusted surrogate endpoint. We apply the proposed method to data from a clinical trial of infliximab and assess the adequacy of the surrogate endpoint in the presence of age heterogeneity.
{"title":"Quantifying proportion of treatment effect by surrogate endpoint under heterogeneity.","authors":"Xinzhou Guo, Florence T Bourgeois, Tianxi Cai","doi":"10.1177/09622802241247719","DOIUrl":"10.1177/09622802241247719","url":null,"abstract":"<p><p>When the primary endpoints in randomized clinical trials require long term follow-up or are costly to measure, it is often desirable to assess treatment effects on surrogate instead of clinical endpoints. Prior to adopting a surrogate endpoint for such purposes, the extent of its surrogacy on the primary endpoint must be assessed. There is a rich statistical literature on assessing surrogacy in the overall population, much of which is based on quantifying the proportion of treatment effect on the primary endpoint that is explained by the treatment effect on the surrogate endpoint. However, the surrogacy of an endpoint may vary across different patient subgroups according to baseline demographic characteristics, and limited methods are currently available to assess overall surrogacy in the presence of potential surrogacy heterogeneity. In this paper, we propose methods that incorporate covariates for baseline information, such as age, to improve overall surrogacy assessment. We use flexible semi-non-parametric modeling strategies to adjust for covariate effects and derive a robust estimate for the proportion of treatment effect of the covariate-adjusted surrogate endpoint. Simulation results suggest that the adjusted surrogate endpoint has greater proportion of treatment effect compared to the unadjusted surrogate endpoint. We apply the proposed method to data from a clinical trial of infliximab and assess the adequacy of the surrogate endpoint in the presence of age heterogeneity.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1152-1162"},"PeriodicalIF":1.6,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140877360","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-01Epub Date: 2024-05-23DOI: 10.1177/09622802241254251
Shun Yu, Yuehan Yang
In this paper, we focus on the modeling problem of estimating data with non-sparse structures, specifically focusing on biological data that exhibit a high degree of relevant features. Various fields, such as biology and finance, face the challenge of non-sparse estimation. We address the problems using the proposed method, called structured iterative division. Structured iterative division effectively divides data into non-sparse and sparse structures and eliminates numerous irrelevant variables, significantly reducing the error while maintaining computational efficiency. Numerical and theoretical results demonstrate the competitive advantage of the proposed method on a wide range of problems, and the proposed method exhibits excellent statistical performance in numerical comparisons with several existing methods. We apply the proposed algorithm to two biology problems, gene microarray datasets, and chimeric protein datasets, to the prognostic risk of distant metastasis in breast cancer and Alzheimer's disease, respectively. Structured iterative division provides insights into gene identification and selection, and we also provide meaningful results in anticipating cancer risk and identifying key factors.
{"title":"A structured iterative division approach for non-sparse regression models and applications in biological data analysis.","authors":"Shun Yu, Yuehan Yang","doi":"10.1177/09622802241254251","DOIUrl":"10.1177/09622802241254251","url":null,"abstract":"<p><p>In this paper, we focus on the modeling problem of estimating data with non-sparse structures, specifically focusing on biological data that exhibit a high degree of relevant features. Various fields, such as biology and finance, face the challenge of non-sparse estimation. We address the problems using the proposed method, called structured iterative division. Structured iterative division effectively divides data into non-sparse and sparse structures and eliminates numerous irrelevant variables, significantly reducing the error while maintaining computational efficiency. Numerical and theoretical results demonstrate the competitive advantage of the proposed method on a wide range of problems, and the proposed method exhibits excellent statistical performance in numerical comparisons with several existing methods. We apply the proposed algorithm to two biology problems, gene microarray datasets, and chimeric protein datasets, to the prognostic risk of distant metastasis in breast cancer and Alzheimer's disease, respectively. Structured iterative division provides insights into gene identification and selection, and we also provide meaningful results in anticipating cancer risk and identifying key factors.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1233-1248"},"PeriodicalIF":1.6,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141082254","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-01Epub Date: 2024-06-05DOI: 10.1177/09622802241254211
Jean Marie Boher, Thomas Filleron, Patrick Sfumato, Pierre Bunouf, Richard J Cook
Despite the widespread use of Cox regression for modeling treatment effects in clinical trials, in immunotherapy oncology trials and other settings therapeutic benefits are not immediately realized thereby violating the proportional hazards assumption. Weighted logrank tests and the so-called Maxcombo test involving the combination of multiple logrank test statistics have been advocated to increase power for detecting effects in these and other settings where hazards are nonproportional. We describe a testing framework based on supremum logrank statistics created by successively analyzing and excluding early events, or obtained using a moving time window. We then describe how such tests can be conducted in a group sequential trial with interim analyses conducted for potential early stopping of benefit. The crossing boundaries for the interim test statistics are determined using an easy-to-implement Monte Carlo algorithm. Numerical studies illustrate the good frequency properties of the proposed group sequential methods.
{"title":"Group sequential methods based on supremum logrank statistics under proportional and nonproportional hazards.","authors":"Jean Marie Boher, Thomas Filleron, Patrick Sfumato, Pierre Bunouf, Richard J Cook","doi":"10.1177/09622802241254211","DOIUrl":"10.1177/09622802241254211","url":null,"abstract":"<p><p>Despite the widespread use of Cox regression for modeling treatment effects in clinical trials, in immunotherapy oncology trials and other settings therapeutic benefits are not immediately realized thereby violating the proportional hazards assumption. Weighted logrank tests and the so-called Maxcombo test involving the combination of multiple logrank test statistics have been advocated to increase power for detecting effects in these and other settings where hazards are nonproportional. We describe a testing framework based on supremum logrank statistics created by successively analyzing and excluding early events, or obtained using a moving time window. We then describe how such tests can be conducted in a group sequential trial with interim analyses conducted for potential early stopping of benefit. The crossing boundaries for the interim test statistics are determined using an easy-to-implement Monte Carlo algorithm. Numerical studies illustrate the good frequency properties of the proposed group sequential methods.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1137-1151"},"PeriodicalIF":1.6,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141262803","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-01Epub Date: 2024-05-20DOI: 10.1177/09622802241254195
Daxuan Deng, Vernon M Chinchilli, Hao Feng, Chixiang Chen, Ming Wang
In clinical and observational studies, secondary outcomes are frequently collected alongside the primary outcome for each subject, yet their potential to improve the analysis efficiency remains underutilized. Moreover, missing data, commonly encountered in practice, can introduce bias to estimates if not appropriately addressed. This article presents an innovative approach that enhances the empirical likelihood-based information borrowing method by integrating missing-data techniques, ensuring robust data integration. We introduce a plug-in inverse probability weighting estimator to handle missingness in the primary analysis, demonstrating its equivalence to the standard joint estimator under mild conditions. To address potential bias from missing secondary outcomes, we propose a uniform mapping strategy, imputing incomplete secondary outcomes into a unified space. Extensive simulations highlight the effectiveness of our method, showing consistent, efficient, and robust estimators under various scenarios involving missing data and/or misspecified secondary models. Finally, we apply our proposal to the Uniform Data Set from the National Alzheimer's Coordinating Center, exemplifying its practical application.
{"title":"Robust integration of secondary outcomes information into primary outcome analysis in the presence of missing data.","authors":"Daxuan Deng, Vernon M Chinchilli, Hao Feng, Chixiang Chen, Ming Wang","doi":"10.1177/09622802241254195","DOIUrl":"10.1177/09622802241254195","url":null,"abstract":"<p><p>In clinical and observational studies, secondary outcomes are frequently collected alongside the primary outcome for each subject, yet their potential to improve the analysis efficiency remains underutilized. Moreover, missing data, commonly encountered in practice, can introduce bias to estimates if not appropriately addressed. This article presents an innovative approach that enhances the empirical likelihood-based information borrowing method by integrating missing-data techniques, ensuring robust data integration. We introduce a plug-in inverse probability weighting estimator to handle missingness in the primary analysis, demonstrating its equivalence to the standard joint estimator under mild conditions. To address potential bias from missing secondary outcomes, we propose a uniform mapping strategy, imputing incomplete secondary outcomes into a unified space. Extensive simulations highlight the effectiveness of our method, showing consistent, efficient, and robust estimators under various scenarios involving missing data and/or misspecified secondary models. Finally, we apply our proposal to the Uniform Data Set from the National Alzheimer's Coordinating Center, exemplifying its practical application.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1249-1263"},"PeriodicalIF":1.6,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141065604","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}