Pub Date : 2024-10-23DOI: 10.1177/09622802241280792
Yanqin Feng, Sijie Wu, Jieli Ding
Clustered current status data frequently occur in many fields of survival studies. Some potential factors related to the hazards of interest cannot be directly observed but are characterized through multiple correlated observable surrogates. In this article, we propose a joint modeling method for regression analysis of clustered current status data with latent variables and potentially informative cluster sizes. The proposed models consist of a factor analysis model to characterize latent variables through their multiple surrogates and an additive hazards frailty model to investigate covariate effects on the failure time and incorporate intra-cluster correlations. We develop an estimation procedure that combines the expectation-maximization algorithm and the weighted estimating equations. The consistency and asymptotic normality of the proposed estimators are established. The finite-sample performance of the proposed method is assessed via a series of simulation studies. This procedure is applied to analyze clustered current status data from the National Toxicology Program on a tumorigenicity study given by the United States Department of Health and Human Services.
{"title":"Joint regression analysis of clustered current status data with latent variables.","authors":"Yanqin Feng, Sijie Wu, Jieli Ding","doi":"10.1177/09622802241280792","DOIUrl":"https://doi.org/10.1177/09622802241280792","url":null,"abstract":"<p><p>Clustered current status data frequently occur in many fields of survival studies. Some potential factors related to the hazards of interest cannot be directly observed but are characterized through multiple correlated observable surrogates. In this article, we propose a joint modeling method for regression analysis of clustered current status data with latent variables and potentially informative cluster sizes. The proposed models consist of a factor analysis model to characterize latent variables through their multiple surrogates and an additive hazards frailty model to investigate covariate effects on the failure time and incorporate intra-cluster correlations. We develop an estimation procedure that combines the expectation-maximization algorithm and the weighted estimating equations. The consistency and asymptotic normality of the proposed estimators are established. The finite-sample performance of the proposed method is assessed via a series of simulation studies. This procedure is applied to analyze clustered current status data from the National Toxicology Program on a tumorigenicity study given by the United States Department of Health and Human Services.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"9622802241280792"},"PeriodicalIF":1.6,"publicationDate":"2024-10-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142508325","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-21DOI: 10.1177/09622802241287707
Peter C Austin
Restricted cubic splines (RCS) allow analysts to model nonlinear relations between continuous covariates and the outcome in a regression model. When using RCS with the Cox proportional hazards model, there is no longer a single hazard ratio for the continuous variable. Instead, the hazard ratio depends on the values of the covariate for the two individuals being compared. Thus, using age as an example, when one assumes a linear relation between age and the log-hazard of the outcome there is a single hazard ratio comparing any two individuals whose age differs by 1 year. However, when allowing for a nonlinear relation between age and the log-hazard of the outcome, the hazard ratio comparing the hazard of the outcome between a 31- and a 30-year-old may differ from the hazard ratio comparing the hazard of the outcome between an 81- and an 80-year-old. We describe four methods to describe graphically the relation between a continuous variable and the outcome when using RCS with a Cox model. These graphical methods are based on plots of relative hazard ratios, cumulative incidence, hazards, and cumulative hazards against the continuous variable. Using a case study of patients presenting to hospital with heart failure and a series of mathematical derivations, we illustrate that the four methods will produce qualitatively similar conclusions about the nature of the relation between a continuous variable and the outcome. Use of these methods will allow for an intuitive communication of the nature of the relation between the variable and the outcome.
{"title":"Graphical methods to illustrate the nature of the relation between a continuous variable and the outcome when using restricted cubic splines with a Cox proportional hazards model.","authors":"Peter C Austin","doi":"10.1177/09622802241287707","DOIUrl":"https://doi.org/10.1177/09622802241287707","url":null,"abstract":"<p><p>Restricted cubic splines (RCS) allow analysts to model nonlinear relations between continuous covariates and the outcome in a regression model. When using RCS with the Cox proportional hazards model, there is no longer a single hazard ratio for the continuous variable. Instead, the hazard ratio depends on the values of the covariate for the two individuals being compared. Thus, using age as an example, when one assumes a linear relation between age and the log-hazard of the outcome there is a single hazard ratio comparing any two individuals whose age differs by 1 year. However, when allowing for a nonlinear relation between age and the log-hazard of the outcome, the hazard ratio comparing the hazard of the outcome between a 31- and a 30-year-old may differ from the hazard ratio comparing the hazard of the outcome between an 81- and an 80-year-old. We describe four methods to describe graphically the relation between a continuous variable and the outcome when using RCS with a Cox model. These graphical methods are based on plots of relative hazard ratios, cumulative incidence, hazards, and cumulative hazards against the continuous variable. Using a case study of patients presenting to hospital with heart failure and a series of mathematical derivations, we illustrate that the four methods will produce qualitatively similar conclusions about the nature of the relation between a continuous variable and the outcome. Use of these methods will allow for an intuitive communication of the nature of the relation between the variable and the outcome.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"9622802241287707"},"PeriodicalIF":1.6,"publicationDate":"2024-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142475114","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-01Epub Date: 2024-08-30DOI: 10.1177/09622802241275382
Annabel Louisa Davies, Anthony Cc Coolen, Tobias Galla
Predicting patient survival probabilities based on observed covariates is an important assessment in clinical practice. These patient-specific covariates are often measured over multiple follow-up appointments. It is then of interest to predict survival based on the history of these longitudinal measurements, and to update predictions as more observations become available. The standard approaches to these so-called 'dynamic prediction' assessments are joint models and landmark analysis. Joint models involve high-dimensional parameterizations, and their computational complexity often prohibits including multiple longitudinal covariates. Landmark analysis is simpler, but discards a proportion of the available data at each 'landmark time'. In this work, we propose a 'delayed kernel' approach to dynamic prediction that sits somewhere in between the two standard methods in terms of complexity. By conditioning hazard rates directly on the covariate measurements over the observation time frame, we define a model that takes into account the full history of covariate measurements but is more practical and parsimonious than joint modelling. Time-dependent association kernels describe the impact of covariate changes at earlier times on the patient's hazard rate at later times. Under the constraints that our model (a) reduces to the standard Cox model for time-independent covariates, and (b) contains the instantaneous Cox model as a special case, we derive two natural kernel parameterizations. Upon application to three clinical data sets, we find that the predictive accuracy of the delayed kernel approach is comparable to that of the two existing standard methods.
{"title":"Delayed kernels for longitudinal survival analysis and dynamic prediction.","authors":"Annabel Louisa Davies, Anthony Cc Coolen, Tobias Galla","doi":"10.1177/09622802241275382","DOIUrl":"10.1177/09622802241275382","url":null,"abstract":"<p><p>Predicting patient survival probabilities based on observed covariates is an important assessment in clinical practice. These patient-specific covariates are often measured over multiple follow-up appointments. It is then of interest to predict survival based on the history of these longitudinal measurements, and to update predictions as more observations become available. The standard approaches to these so-called 'dynamic prediction' assessments are joint models and landmark analysis. Joint models involve high-dimensional parameterizations, and their computational complexity often prohibits including multiple longitudinal covariates. Landmark analysis is simpler, but discards a proportion of the available data at each 'landmark time'. In this work, we propose a 'delayed kernel' approach to dynamic prediction that sits somewhere in between the two standard methods in terms of complexity. By conditioning hazard rates directly on the covariate measurements over the observation time frame, we define a model that takes into account the full history of covariate measurements but is more practical and parsimonious than joint modelling. Time-dependent association kernels describe the impact of covariate changes at earlier times on the patient's hazard rate at later times. Under the constraints that our model (a) reduces to the standard Cox model for time-independent covariates, and (b) contains the instantaneous Cox model as a special case, we derive two natural kernel parameterizations. Upon application to three clinical data sets, we find that the predictive accuracy of the delayed kernel approach is comparable to that of the two existing standard methods.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1836-1858"},"PeriodicalIF":1.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11577694/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142112180","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-01DOI: 10.1177/09622802241277956
Jeffrey E Harris
We describe a novel approach for recovering the underlying parameters of the SIR dynamic epidemic model from observed data on case incidence. We formulate a discrete-time approximation of the original continuous-time model and search for the parameter vector that minimizes the standard least squares criterion function. We show that the gradient vector and matrix of second-order derivatives of the criterion function with respect to the parameters adhere to their own systems of difference equations and thus can be exactly calculated iteratively. Applying our new approach, we estimated a four-parameter SIR model from daily reported cases of COVID-19 during the SARS-CoV-2 Omicron/BA.1 surge of December 2021-March 2022 in New York City. The estimated SIR model showed a tight fit to the observed data, but less so when we excluded residual cases attributable to the Delta variant during the initial upswing of the wave in December. Our analyses of both the real-world COVID-19 data and simulated case incidence data revealed an important problem of weak parameter identification. While our methods permitted for the separate estimation of the infection transmission parameter and the infection persistence parameter, only a linear combination of these two key parameters could be estimated with precision. The SIR model appears to be an adequate reduced-form description of the Omicron surge, but it is not necessarily the correct structural model. Prior information above and beyond case incidence data may be required to sharply identify the parameters and thus distinguish between alternative epidemic models.
我们介绍了一种从病例发生率观测数据中恢复 SIR 动态流行病模型基本参数的新方法。我们提出了原始连续时间模型的离散时间近似值,并寻找能使标准最小二乘法准则函数最小化的参数向量。我们证明,判据函数相对于参数的梯度向量和二阶导数矩阵遵循各自的差分方程组,因此可以通过迭代精确计算。应用我们的新方法,我们从 2021 年 12 月至 2022 年 3 月纽约市 SARS-CoV-2 Omicron/BA.1 高峰期间每日报告的 COVID-19 病例中估算出了一个四参数 SIR 模型。估算出的 SIR 模型与观察到的数据非常吻合,但当我们剔除了 12 月病例潮最初上升阶段的三角洲变异体残留病例后,与观察到的数据的吻合程度就降低了。我们对实际 COVID-19 数据和模拟病例发病率数据的分析表明了一个重要的问题,即参数识别能力较弱。虽然我们的方法允许对感染传播参数和感染持续参数进行单独估计,但只能对这两个关键参数的线性组合进行精确估计。SIR 模型似乎是对奥米克龙激增的充分简化描述,但并不一定是正确的结构模型。除了病例发生率数据之外,可能还需要其他先验信息来精确确定参数,从而区分其他流行病模型。
{"title":"A tight fit of the SIR dynamic epidemic model to daily cases of COVID-19 reported during the 2021-2022 Omicron surge in New York City: A novel approach.","authors":"Jeffrey E Harris","doi":"10.1177/09622802241277956","DOIUrl":"10.1177/09622802241277956","url":null,"abstract":"<p><p>We describe a novel approach for recovering the underlying parameters of the SIR dynamic epidemic model from observed data on case incidence. We formulate a discrete-time approximation of the original continuous-time model and search for the parameter vector that minimizes the standard least squares criterion function. We show that the gradient vector and matrix of second-order derivatives of the criterion function with respect to the parameters adhere to their own systems of difference equations and thus can be exactly calculated iteratively. Applying our new approach, we estimated a four-parameter SIR model from daily reported cases of COVID-19 during the SARS-CoV-2 Omicron/BA.1 surge of December 2021-March 2022 in New York City. The estimated SIR model showed a tight fit to the observed data, but less so when we excluded residual cases attributable to the Delta variant during the initial upswing of the wave in December. Our analyses of both the real-world COVID-19 data and simulated case incidence data revealed an important problem of weak parameter identification. While our methods permitted for the separate estimation of the infection transmission parameter and the infection persistence parameter, only a linear combination of these two key parameters could be estimated with precision. The SIR model appears to be an adequate reduced-form description of the Omicron surge, but it is not necessarily the correct structural model. Prior information above and beyond case incidence data may be required to sharply identify the parameters and thus distinguish between alternative epidemic models.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":"33 10","pages":"1877-1898"},"PeriodicalIF":1.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11577707/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142669223","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In this article, we present a joint modeling approach for zero-inflated longitudinal count measurements and time-to-event outcomes. For the longitudinal sub-model, a mixed effects Hurdle model is utilized, incorporating various distributional assumptions such as zero-inflated Poisson, zero-inflated negative binomial, or zero-inflated generalized Poisson. For the time-to-event sub-model, a Cox proportional hazard model is applied. For the functional form linking the longitudinal outcome history to the hazard of the event, a linear combination is used. This combination is derived from the current values of the linear predictors of Hurdle mixed effects. Some other forms are also considered, including a linear combination of the current slopes of the linear predictors of Hurdle mixed effects as well as the shared random effects. A Markov chain Monte Carlo method is implemented for Bayesian parameter estimation. Dynamic prediction using joint modeling is highly valuable in personalized medicine, as discussed here for joint modeling of zero-inflated longitudinal count measurements and time-to-event outcomes. We assess and demonstrate the effectiveness of the proposed joint models through extensive simulation studies, with a specific emphasis on parameter estimation and dynamic predictions for both over-dispersed and under-dispersed data. We finally apply the joint model to longitudinal microbiome pregnancy and HIV data sets.
{"title":"Joint modeling of zero-inflated longitudinal measurements and time-to-event outcomes with applications to dynamic prediction.","authors":"Mojtaba Ganjali, Taban Baghfalaki, Narayanaswamy Balakrishnan","doi":"10.1177/09622802241268466","DOIUrl":"10.1177/09622802241268466","url":null,"abstract":"<p><p>In this article, we present a joint modeling approach for zero-inflated longitudinal count measurements and time-to-event outcomes. For the longitudinal sub-model, a mixed effects Hurdle model is utilized, incorporating various distributional assumptions such as zero-inflated Poisson, zero-inflated negative binomial, or zero-inflated generalized Poisson. For the time-to-event sub-model, a Cox proportional hazard model is applied. For the functional form linking the longitudinal outcome history to the hazard of the event, a linear combination is used. This combination is derived from the current values of the linear predictors of Hurdle mixed effects. Some other forms are also considered, including a linear combination of the current slopes of the linear predictors of Hurdle mixed effects as well as the shared random effects. A Markov chain Monte Carlo method is implemented for Bayesian parameter estimation. Dynamic prediction using joint modeling is highly valuable in personalized medicine, as discussed here for joint modeling of zero-inflated longitudinal count measurements and time-to-event outcomes. We assess and demonstrate the effectiveness of the proposed joint models through extensive simulation studies, with a specific emphasis on parameter estimation and dynamic predictions for both over-dispersed and under-dispersed data. We finally apply the joint model to longitudinal microbiome pregnancy and HIV data sets.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1731-1767"},"PeriodicalIF":1.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142381665","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-01Epub Date: 2024-09-09DOI: 10.1177/09622802241269646
Yi Liu, Huiyue Li, Yunji Zhou, Roland A Matsouaka
The use of propensity score methods has become ubiquitous in causal inference. At the heart of these methods is the positivity assumption. Violation of the positivity assumption leads to the presence of extreme propensity score weights when estimating average causal effects, which affects statistical inference. To circumvent this issue, trimming or truncating methods have been widely used. Unfortunately, these methods require that we pre-specify a threshold. There are a number of alternative methods to deal with the lack of positivity when we estimate the average treatment effect (ATE). However, no other methods exist beyond trimming and truncation to deal with the same issue when the goal is to estimate the average treatment effect on the treated (ATT). In this article, we propose a propensity score weight-based alternative for the ATT, called overlap weighted average treatment effect on the treated. The appeal of our proposed method lies in its ability to obtain similar or even better results than trimming and truncation while relaxing the constraint to choose an a priori threshold (or related measures). The performance of the proposed method is illustrated via a series of Monte Carlo simulations and a data analysis on racial disparities in health care expenditures.
在因果推断中,倾向得分法的使用已变得无处不在。这些方法的核心是积极性假设。违反积极性假设会导致在估计平均因果效应时出现极端倾向得分权重,从而影响统计推断。为了规避这一问题,修剪或截断方法得到了广泛应用。遗憾的是,这些方法要求我们预先指定一个阈值。当我们估计平均治疗效果(ATE)时,有许多替代方法可以解决缺乏正向性的问题。然而,当我们的目标是估计平均治疗效果(ATT)时,除了修剪和截断外,还没有其他方法来解决同样的问题。在本文中,我们提出了一种基于倾向得分权重的 ATT 替代方法,称为 "重叠加权平均治疗效果"。我们提出的方法的吸引力在于,它能够获得与修剪和截断相似甚至更好的结果,同时放宽了选择先验阈值(或相关测量方法)的限制。我们通过一系列蒙特卡罗模拟和医疗支出种族差异的数据分析来说明所提方法的性能。
{"title":"Average treatment effect on the treated, under lack of positivity.","authors":"Yi Liu, Huiyue Li, Yunji Zhou, Roland A Matsouaka","doi":"10.1177/09622802241269646","DOIUrl":"10.1177/09622802241269646","url":null,"abstract":"<p><p>The use of propensity score methods has become ubiquitous in causal inference. At the heart of these methods is the positivity assumption. Violation of the positivity assumption leads to the presence of extreme propensity score weights when estimating average causal effects, which affects statistical inference. To circumvent this issue, trimming or truncating methods have been widely used. Unfortunately, these methods require that we pre-specify a threshold. There are a number of alternative methods to deal with the lack of positivity when we estimate the average treatment effect (ATE). However, no other methods exist beyond trimming and truncation to deal with the same issue when the goal is to estimate the average treatment effect on the treated (ATT). In this article, we propose a propensity score weight-based alternative for the ATT, called overlap weighted average treatment effect on the treated. The appeal of our proposed method lies in its ability to obtain similar or even better results than trimming and truncation while relaxing the constraint to choose an a priori threshold (or related measures). The performance of the proposed method is illustrated via a series of Monte Carlo simulations and a data analysis on racial disparities in health care expenditures.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1689-1717"},"PeriodicalIF":1.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142154972","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-01Epub Date: 2024-09-09DOI: 10.1177/09622802241269010
Hortense Doms, Philippe Lambert, Catherine Legrand
In medical studies, repeated measurements of biomarkers and time-to-event data are often collected during the follow-up period. To assess the association between these two outcomes, joint models are frequently considered. The most common approach uses a linear mixed model for the longitudinal part and a proportional hazard model for the survival part. The latter assumes a linear relationship between the survival covariates and the log hazard. In this work, we propose an extension allowing the inclusion of nonlinear covariate effects in the survival model using Bayesian penalized B-splines. Our model is valid for non-Gaussian longitudinal responses since we use a generalized linear mixed model for the longitudinal process. A simulation study shows that our method gives good statistical performance and highlights the importance of taking into account the possible nonlinear effects of certain survival covariates. Data from patients with a first progression of glioblastoma are analysed to illustrate the method.
在医学研究中,通常会在随访期间重复测量生物标志物和时间到事件数据。为了评估这两种结果之间的关联,通常会考虑联合模型。最常见的方法是纵向部分使用线性混合模型,生存部分使用比例危险模型。后者假定生存协变量与对数危险之间存在线性关系。在这项工作中,我们提出了一种扩展方法,允许在使用贝叶斯惩罚性 B 样条的生存模型中加入非线性协变量效应。我们的模型适用于非高斯纵向响应,因为我们对纵向过程使用了广义线性混合模型。模拟研究表明,我们的方法具有良好的统计性能,并强调了考虑某些生存协变量可能产生的非线性效应的重要性。我们对胶质母细胞瘤首次进展期患者的数据进行了分析,以说明该方法。
{"title":"Flexible joint model for time-to-event and non-Gaussian longitudinal outcomes.","authors":"Hortense Doms, Philippe Lambert, Catherine Legrand","doi":"10.1177/09622802241269010","DOIUrl":"10.1177/09622802241269010","url":null,"abstract":"<p><p>In medical studies, repeated measurements of biomarkers and time-to-event data are often collected during the follow-up period. To assess the association between these two outcomes, joint models are frequently considered. The most common approach uses a linear mixed model for the longitudinal part and a proportional hazard model for the survival part. The latter assumes a linear relationship between the survival covariates and the log hazard. In this work, we propose an extension allowing the inclusion of nonlinear covariate effects in the survival model using Bayesian penalized B-splines. Our model is valid for non-Gaussian longitudinal responses since we use a generalized linear mixed model for the longitudinal process. A simulation study shows that our method gives good statistical performance and highlights the importance of taking into account the possible nonlinear effects of certain survival covariates. Data from patients with a first progression of glioblastoma are analysed to illustrate the method.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1783-1799"},"PeriodicalIF":1.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142154973","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-01Epub Date: 2024-08-20DOI: 10.1177/09622802241268504
J Andres Christen, F Javier Rubio
The hazard function represents one of the main quantities of interest in the analysis of survival data. We propose a general approach for parametrically modelling the dynamics of the hazard function using systems of autonomous ordinary differential equations (ODEs). This modelling approach can be used to provide qualitative and quantitative analyses of the evolution of the hazard function over time. Our proposal capitalises on the extensive literature on ODEs which, in particular, allows for establishing basic rules or laws on the dynamics of the hazard function via the use of autonomous ODEs. We show how to implement the proposed modelling framework in cases where there is an analytic solution to the system of ODEs or where an ODE solver is required to obtain a numerical solution. We focus on the use of a Bayesian modelling approach, but the proposed methodology can also be coupled with maximum likelihood estimation. A simulation study is presented to illustrate the performance of these models and the interplay of sample size and censoring. Two case studies using real data are presented to illustrate the use of the proposed approach and to highlight the interpretability of the corresponding models. We conclude with a discussion on potential extensions of our work and strategies to include covariates into our framework. Although we focus on examples of Medical Statistics, the proposed framework is applicable in any context where the interest lies in estimating and interpreting the dynamics of the hazard function.
{"title":"Dynamic survival analysis: Modelling the hazard function via ordinary differential equations.","authors":"J Andres Christen, F Javier Rubio","doi":"10.1177/09622802241268504","DOIUrl":"10.1177/09622802241268504","url":null,"abstract":"<p><p>The hazard function represents one of the main quantities of interest in the analysis of survival data. We propose a general approach for parametrically modelling the <i>dynamics</i> of the hazard function using systems of autonomous ordinary differential equations (ODEs). This modelling approach can be used to provide qualitative and quantitative analyses of the evolution of the hazard function over time. Our proposal capitalises on the extensive literature on ODEs which, in particular, allows for establishing basic rules or laws on the dynamics of the hazard function via the use of autonomous ODEs. We show how to implement the proposed modelling framework in cases where there is an analytic solution to the system of ODEs or where an ODE solver is required to obtain a numerical solution. We focus on the use of a Bayesian modelling approach, but the proposed methodology can also be coupled with maximum likelihood estimation. A simulation study is presented to illustrate the performance of these models and the interplay of sample size and censoring. Two case studies using real data are presented to illustrate the use of the proposed approach and to highlight the interpretability of the corresponding models. We conclude with a discussion on potential extensions of our work and strategies to include covariates into our framework. Although we focus on examples of Medical Statistics, the proposed framework is applicable in any context where the interest lies in estimating and interpreting the dynamics of the hazard function.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1768-1782"},"PeriodicalIF":1.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11577698/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142005270","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-01Epub Date: 2024-10-07DOI: 10.1177/09622802241279109
Chun-Che Wen, Nathaniel Baker, Rajib Paul, Elizabeth Hill, Kelly Hunt, Hong Li, Kevin Gray, Brian Neelon
In a recent 12-week smoking cessation trial, varenicline tartrate failed to show significant improvements in enhancing end-of-treatment abstinence when compared with placebo among adolescents and young adults. The original analysis aimed to assess the average effect across the entire population using timeline followback methods, which typically involve overdispersed binomial counts. We instead propose to investigate treatment effect heterogeneity among latent classes of participants using a Bayesian beta-binomial piecewise linear growth mixture model specifically designed to address longitudinal overdispersed binomial responses. Within each class, we fit a piecewise linear beta-binomial mixed model with random changepoints for each study group to detect critical windows of treatment efficacy. Using this model, we can cluster subjects who share similar characteristics, estimate the class-specific mean abstinence trends for each study group, and quantify the treatment effect over time within each class. Our analysis identified two classes of subjects: one comprising high-abstinent individuals, typically young adults and light smokers, in which varenicline led to improved abstinence; and another comprising low-abstinent individuals for whom varenicline showed no discernible effect. These findings highlight the importance of tailoring varenicline to specific participant subgroups, thereby advancing precision medicine in smoking cessation studies.
{"title":"A Bayesian beta-binomial piecewise growth mixture model for longitudinal overdispersed binomial data.","authors":"Chun-Che Wen, Nathaniel Baker, Rajib Paul, Elizabeth Hill, Kelly Hunt, Hong Li, Kevin Gray, Brian Neelon","doi":"10.1177/09622802241279109","DOIUrl":"10.1177/09622802241279109","url":null,"abstract":"<p><p>In a recent 12-week smoking cessation trial, varenicline tartrate failed to show significant improvements in enhancing end-of-treatment abstinence when compared with placebo among adolescents and young adults. The original analysis aimed to assess the average effect across the entire population using timeline followback methods, which typically involve overdispersed binomial counts. We instead propose to investigate treatment effect heterogeneity among latent classes of participants using a Bayesian beta-binomial piecewise linear growth mixture model specifically designed to address longitudinal overdispersed binomial responses. Within each class, we fit a piecewise linear beta-binomial mixed model with random changepoints for each study group to detect critical windows of treatment efficacy. Using this model, we can cluster subjects who share similar characteristics, estimate the class-specific mean abstinence trends for each study group, and quantify the treatment effect over time within each class. Our analysis identified two classes of subjects: one comprising high-abstinent individuals, typically young adults and light smokers, in which varenicline led to improved abstinence; and another comprising low-abstinent individuals for whom varenicline showed no discernible effect. These findings highlight the importance of tailoring varenicline to specific participant subgroups, thereby advancing precision medicine in smoking cessation studies.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1859-1876"},"PeriodicalIF":1.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142381663","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-01Epub Date: 2024-09-30DOI: 10.1177/09622802241275413
Jianing Wang, David M Kline, Laura Forsberg White
Approaches to population size estimation are of importance across a wide spectrum of disciplines, especially when census and simple random sampling are impractical. The capture-recapture method and the multiplier-benchmark method are two commonly used approaches that use data that partially capture the target population and overlap in a known way. Due to similarities in required data structures, the approaches are often used interchangeably without a critical appraisal of the underlying assumptions, especially in the two-sample case. Here, we describe the similarities and differences of the sampling mechanisms and assumptions underlying both approaches. We emphasize that the capture-recapture method assumes data sources as random samples and describes two-way inclusion histories, while in multiplier-benchmark method, one source captures a fixed sub-population, and the one-way inclusion histories are modeled. We also discuss the implications of these differences through simulation and real data to guide the choice of method in practice. A careful study of the data structures, relationships, and data generation processes is crucial for assessing the appropriateness of using these methods.
{"title":"On the estimation of population size-A comparison of capture-recapture and multiplier-benchmark methods.","authors":"Jianing Wang, David M Kline, Laura Forsberg White","doi":"10.1177/09622802241275413","DOIUrl":"10.1177/09622802241275413","url":null,"abstract":"<p><p>Approaches to population size estimation are of importance across a wide spectrum of disciplines, especially when census and simple random sampling are impractical. The capture-recapture method and the multiplier-benchmark method are two commonly used approaches that use data that partially capture the target population and overlap in a known way. Due to similarities in required data structures, the approaches are often used interchangeably without a critical appraisal of the underlying assumptions, especially in the two-sample case. Here, we describe the similarities and differences of the sampling mechanisms and assumptions underlying both approaches. We emphasize that the capture-recapture method assumes data sources as random samples and describes two-way inclusion histories, while in multiplier-benchmark method, one source captures a fixed sub-population, and the one-way inclusion histories are modeled. We also discuss the implications of these differences through simulation and real data to guide the choice of method in practice. A careful study of the data structures, relationships, and data generation processes is crucial for assessing the appropriateness of using these methods.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1818-1835"},"PeriodicalIF":1.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142354183","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}