Pub Date : 2025-10-01Epub Date: 2025-10-15DOI: 10.1007/s10985-025-09672-z
Yiyuan Huang, Douglas Schaubel, Min Zhang
In many clinical trials, one is interested in evaluating the treatment effect based on different types of outcomes, including recurrent and terminal events. The most popular approach is the time-to-first-event analysis (TTFE), based on the composite outcome of the time to the first event among all events of interest. The motivation for the composite outcome approach is to increase the number of events and potentially increase power. Other composite outcome or composite analysis methods are also studied in the literature, but are less adopted in practice. In this article, we first review the mainstream composite analysis methods and classify them into three categories: (A) Composite-outcome Methods, which combine multiple events into a composite outcome before analysis, e.g., combining events into a time-to-event outcome in TTFE and into a single recurrent event process in the combined-recurrent-event analysis (CRE); (B) Joint-analysis Methods, which test for the recurrent event process and the terminal event jointly, e.g., Joint Frailty Model (JFM), Ghosh-Lin Method (GL), and Nelsen-Aalen Method (NA); (C) Win-ratio type Methods that account for the ordering of two types of events, e.g., Win-fraction Regression (WR). We conduct comprehensive simulation studies to evaluate the performance of various types of methods in terms of type I error control and power under a wide range of scenarios. We found that the non-parametric joint testing approach (GL/NA) and CRE have overall the best performance. However, TTFE and WR exhibit relatively low power. Also, adding events that have no or weak association with treatment usually decreases power.
{"title":"Statistical methods for composite analysis of recurrent and terminal events in clinical trials.","authors":"Yiyuan Huang, Douglas Schaubel, Min Zhang","doi":"10.1007/s10985-025-09672-z","DOIUrl":"10.1007/s10985-025-09672-z","url":null,"abstract":"<p><p>In many clinical trials, one is interested in evaluating the treatment effect based on different types of outcomes, including recurrent and terminal events. The most popular approach is the time-to-first-event analysis (TTFE), based on the composite outcome of the time to the first event among all events of interest. The motivation for the composite outcome approach is to increase the number of events and potentially increase power. Other composite outcome or composite analysis methods are also studied in the literature, but are less adopted in practice. In this article, we first review the mainstream composite analysis methods and classify them into three categories: (A) Composite-outcome Methods, which combine multiple events into a composite outcome before analysis, e.g., combining events into a time-to-event outcome in TTFE and into a single recurrent event process in the combined-recurrent-event analysis (CRE); (B) Joint-analysis Methods, which test for the recurrent event process and the terminal event jointly, e.g., Joint Frailty Model (JFM), Ghosh-Lin Method (GL), and Nelsen-Aalen Method (NA); (C) Win-ratio type Methods that account for the ordering of two types of events, e.g., Win-fraction Regression (WR). We conduct comprehensive simulation studies to evaluate the performance of various types of methods in terms of type I error control and power under a wide range of scenarios. We found that the non-parametric joint testing approach (GL/NA) and CRE have overall the best performance. However, TTFE and WR exhibit relatively low power. Also, adding events that have no or weak association with treatment usually decreases power.</p>","PeriodicalId":49908,"journal":{"name":"Lifetime Data Analysis","volume":" ","pages":"810-829"},"PeriodicalIF":1.0,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145294258","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-01Epub Date: 2025-10-14DOI: 10.1007/s10985-025-09671-0
Miki Horiguchi, Lu Tian, Kenneth L Kehl, Hajime Uno
Delayed treatment effects on time-to-event outcomes are commonly observed in randomized controlled trials of cancer immunotherapies. When the treatment effect has a delayed onset, the conventional test/estimation approach-using the log-rank test for between-group comparison and Cox's hazard ratio to quantify the treatment effect-can be suboptimal. The log-rank test may lack power in such scenarios, and the interpretation of the hazard ratio is often ambiguous. Recently, alternative test/estimation approaches have been proposed to address these limitations. One such approach is based on long-term restricted mean survival time (LT-RMST), while another is based on average hazard with survival weight (AH-SW). This paper integrates these two concepts and introduces a novel long-term average hazard (LT-AH) approach with survival weight for both hypothesis testing and estimation. Numerical studies highlight specific scenarios where the proposed LT-AH method achieves higher power than the existing alternatives. The LT-AH for each group can be estimated nonparametrically, and the proposed between-group comparison maintains test/estimation coherency. Because the difference and ratio of LT-AH do not rely on model assumptions about the relationship between two groups, the LT-AH approach provides a robust framework for estimating the magnitude of between-group differences. Furthermore, LT-AH allows for treatment effect quantification in both absolute (difference in LT-AH) and relative (ratio of LT-AH) terms, aligning with guideline recommendations and addressing practical needs. Given its interpretability and improved power in certain settings, the proposed LT-AH approach offers a useful alternative to conventional hazard-based methods, particularly when delayed treatment effects are expected.
{"title":"Assessing delayed treatment benefits of immunotherapy using long-term average hazard: a novel test/estimation approach.","authors":"Miki Horiguchi, Lu Tian, Kenneth L Kehl, Hajime Uno","doi":"10.1007/s10985-025-09671-0","DOIUrl":"10.1007/s10985-025-09671-0","url":null,"abstract":"<p><p>Delayed treatment effects on time-to-event outcomes are commonly observed in randomized controlled trials of cancer immunotherapies. When the treatment effect has a delayed onset, the conventional test/estimation approach-using the log-rank test for between-group comparison and Cox's hazard ratio to quantify the treatment effect-can be suboptimal. The log-rank test may lack power in such scenarios, and the interpretation of the hazard ratio is often ambiguous. Recently, alternative test/estimation approaches have been proposed to address these limitations. One such approach is based on long-term restricted mean survival time (LT-RMST), while another is based on average hazard with survival weight (AH-SW). This paper integrates these two concepts and introduces a novel long-term average hazard (LT-AH) approach with survival weight for both hypothesis testing and estimation. Numerical studies highlight specific scenarios where the proposed LT-AH method achieves higher power than the existing alternatives. The LT-AH for each group can be estimated nonparametrically, and the proposed between-group comparison maintains test/estimation coherency. Because the difference and ratio of LT-AH do not rely on model assumptions about the relationship between two groups, the LT-AH approach provides a robust framework for estimating the magnitude of between-group differences. Furthermore, LT-AH allows for treatment effect quantification in both absolute (difference in LT-AH) and relative (ratio of LT-AH) terms, aligning with guideline recommendations and addressing practical needs. Given its interpretability and improved power in certain settings, the proposed LT-AH approach offers a useful alternative to conventional hazard-based methods, particularly when delayed treatment effects are expected.</p>","PeriodicalId":49908,"journal":{"name":"Lifetime Data Analysis","volume":" ","pages":"784-809"},"PeriodicalIF":1.0,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12586407/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145287554","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-01Epub Date: 2025-07-11DOI: 10.1007/s10985-025-09665-y
Zhiguo Li
Data analysis methods have been well developed for analyzing data to make inferences about adaptive treatment strategies in sequential multiple assignment randomized trials (SMART), when data are continuous or right-censored. However, in some clinical studies, time-to-event outcomes are interval censored, meaning that, for example, the time of interest is only observed between two random visit times to the clinic, which is common in some areas such as psychology studies. In this case, the appropriate analysis methods in SMART studies have not been considered in the literature. This article tries to fill this gap by developing methods for this purpose. Based on a proportional hazards model, we propose to use a weighted spline-based sieve maximum likelihood method to make inference about the group differences using a Wald test. Asymptotic properties of the estimator for the hazard ratio are derived, and variance estimation is considered. We conduct a simulation to assess its finite sample performance, and then analyze data from the Sequenced Treatment Alternatives to Relieve Depression (STAR*D) trial.
数据分析方法已经发展得很好,当数据是连续的或右删节的,用于分析数据以推断顺序多任务随机试验(SMART)中的自适应治疗策略。然而,在一些临床研究中,时间到事件的结果是间隔审查的,这意味着,例如,感兴趣的时间只在两次随机就诊时间之间观察到,这在心理学研究等某些领域很常见。在这种情况下,文献中没有考虑SMART研究中适当的分析方法。本文试图通过开发用于此目的的方法来填补这一空白。基于比例风险模型,我们建议使用加权样条筛选最大似然方法来推断使用Wald检验的组差异。导出了风险比估计量的渐近性质,并考虑了方差估计。我们进行模拟以评估其有限样本性能,然后分析来自Sequenced Treatment Alternatives to ease Depression (STAR*D)试验的数据。
{"title":"Analysis of interval censored survival data in sequential multiple assignment randomized trials.","authors":"Zhiguo Li","doi":"10.1007/s10985-025-09665-y","DOIUrl":"10.1007/s10985-025-09665-y","url":null,"abstract":"<p><p>Data analysis methods have been well developed for analyzing data to make inferences about adaptive treatment strategies in sequential multiple assignment randomized trials (SMART), when data are continuous or right-censored. However, in some clinical studies, time-to-event outcomes are interval censored, meaning that, for example, the time of interest is only observed between two random visit times to the clinic, which is common in some areas such as psychology studies. In this case, the appropriate analysis methods in SMART studies have not been considered in the literature. This article tries to fill this gap by developing methods for this purpose. Based on a proportional hazards model, we propose to use a weighted spline-based sieve maximum likelihood method to make inference about the group differences using a Wald test. Asymptotic properties of the estimator for the hazard ratio are derived, and variance estimation is considered. We conduct a simulation to assess its finite sample performance, and then analyze data from the Sequenced Treatment Alternatives to Relieve Depression (STAR*D) trial.</p>","PeriodicalId":49908,"journal":{"name":"Lifetime Data Analysis","volume":" ","pages":"852-868"},"PeriodicalIF":1.0,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144621007","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-01Epub Date: 2025-10-09DOI: 10.1007/s10985-025-09673-y
Emily M Damone, Matthew A Psioda, Joseph G Ibrahim
Many methods exist to jointly model either recurrent and related terminal survival events or longitudinal outcome measures and related terminal survival event. However, few methods exist which can account for the dependency between all three outcomes of interest, and none allow for the modeling of all three outcomes without strong correlation assumptions. We propose a joint model which uses subject-specific random effects to connect the survival model (terminal and recurrent events) with a longitudinal outcome model. In the proposed method, proportional hazards models with shared frailties are used to model dependence between the recurrent and terminal events, while a separate (but correlated) set of random effects are utilized in a generalized linear mixed model to model dependence with longitudinal outcome measures. All random effects are related based on an assumed multivariate normal distribution. The proposed joint modeling approach allows for flexible models, particularly for unique longitudinal trajectories, that can be utilized in a wide range of health applications. We evaluate the model through simulation studies as well as through an application to data from the Atherosclerosis Risk in Communities (ARIC) study.
{"title":"Bayesian joint models for longitudinal, recurrent, and terminal event data.","authors":"Emily M Damone, Matthew A Psioda, Joseph G Ibrahim","doi":"10.1007/s10985-025-09673-y","DOIUrl":"10.1007/s10985-025-09673-y","url":null,"abstract":"<p><p>Many methods exist to jointly model either recurrent and related terminal survival events or longitudinal outcome measures and related terminal survival event. However, few methods exist which can account for the dependency between all three outcomes of interest, and none allow for the modeling of all three outcomes without strong correlation assumptions. We propose a joint model which uses subject-specific random effects to connect the survival model (terminal and recurrent events) with a longitudinal outcome model. In the proposed method, proportional hazards models with shared frailties are used to model dependence between the recurrent and terminal events, while a separate (but correlated) set of random effects are utilized in a generalized linear mixed model to model dependence with longitudinal outcome measures. All random effects are related based on an assumed multivariate normal distribution. The proposed joint modeling approach allows for flexible models, particularly for unique longitudinal trajectories, that can be utilized in a wide range of health applications. We evaluate the model through simulation studies as well as through an application to data from the Atherosclerosis Risk in Communities (ARIC) study.</p>","PeriodicalId":49908,"journal":{"name":"Lifetime Data Analysis","volume":" ","pages":"932-949"},"PeriodicalIF":1.0,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145253405","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-01Epub Date: 2025-08-18DOI: 10.1007/s10985-025-09667-w
Wei-En Lu, Ai Ni
In large observational studies with survival outcome and low event rates, the case-cohort design is commonly used to reduce the cost associated with covariate measurement. The restricted mean survival time (RMST) difference has been increasingly used as an alternative to hazard ratio when estimating the causal effect on survival outcomes. We investigate the estimation of marginal causal effect on RMST under the stratified case-cohort design while adjusting for measured confounders through propensity score stratification. The asymptotic normality of the estimator is established, and its variance formula is derived. Simulation studies are performed to evaluate the finite sample performance of the proposed method compared to several alternative methods. Finally, we apply the proposed method to the Atherosclerosis Risk in Communities study to estimate the marginal causal effect of high-sensitivity C-reactive protein level on coronary heart disease-free survival.
{"title":"Causal effect estimation on restricted mean survival time under case-cohort design via propensity score stratification.","authors":"Wei-En Lu, Ai Ni","doi":"10.1007/s10985-025-09667-w","DOIUrl":"10.1007/s10985-025-09667-w","url":null,"abstract":"<p><p>In large observational studies with survival outcome and low event rates, the case-cohort design is commonly used to reduce the cost associated with covariate measurement. The restricted mean survival time (RMST) difference has been increasingly used as an alternative to hazard ratio when estimating the causal effect on survival outcomes. We investigate the estimation of marginal causal effect on RMST under the stratified case-cohort design while adjusting for measured confounders through propensity score stratification. The asymptotic normality of the estimator is established, and its variance formula is derived. Simulation studies are performed to evaluate the finite sample performance of the proposed method compared to several alternative methods. Finally, we apply the proposed method to the Atherosclerosis Risk in Communities study to estimate the marginal causal effect of high-sensitivity C-reactive protein level on coronary heart disease-free survival.</p>","PeriodicalId":49908,"journal":{"name":"Lifetime Data Analysis","volume":" ","pages":"898-931"},"PeriodicalIF":1.0,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12586416/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144876417","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-01Epub Date: 2025-08-27DOI: 10.1007/s10985-025-09666-x
Yuchen Mao, Lianming Wang, Xuemei Sui
Joint modeling of longitudinal responses and survival time has gained great attention in statistics literature over the last few decades. Most existing works focus on joint analysis of longitudinal data and right-censored data. In this article, we propose a new frailty model for joint analysis of a longitudinal response and interval-censored survival time. Such data commonly arise in real-life studies where participants are examined at periodical or irregular follow-up times. The proposed joint model contains a nonlinear mixed effects submodel for the longitudinal response and a semiparametric probit submodel for the survival time given a shared normal frailty. The proposed joint model allows the regression coefficients to be interpreted as the marginal effects up to a multiplicative constant on both the longitudinal and survival responses. Adopting splines allows us to approximate the unknown baseline functions in both submodels with only a finite number of unknown coefficients while providing great modeling flexibility. An efficient Gibbs sampler is developed for posterior computation, in which all parameters and latent variables can be sampled easily from their full conditional distributions. The proposed method shows a good estimation performance in simulation studies and is further illustrated by a real-life application to the patient data from the Aerobics Center Longitudinal Study. The R code for the proposed methodology is made available for public use.
{"title":"Bayesian joint analysis of longitudinal data and interval-censored failure time data.","authors":"Yuchen Mao, Lianming Wang, Xuemei Sui","doi":"10.1007/s10985-025-09666-x","DOIUrl":"10.1007/s10985-025-09666-x","url":null,"abstract":"<p><p>Joint modeling of longitudinal responses and survival time has gained great attention in statistics literature over the last few decades. Most existing works focus on joint analysis of longitudinal data and right-censored data. In this article, we propose a new frailty model for joint analysis of a longitudinal response and interval-censored survival time. Such data commonly arise in real-life studies where participants are examined at periodical or irregular follow-up times. The proposed joint model contains a nonlinear mixed effects submodel for the longitudinal response and a semiparametric probit submodel for the survival time given a shared normal frailty. The proposed joint model allows the regression coefficients to be interpreted as the marginal effects up to a multiplicative constant on both the longitudinal and survival responses. Adopting splines allows us to approximate the unknown baseline functions in both submodels with only a finite number of unknown coefficients while providing great modeling flexibility. An efficient Gibbs sampler is developed for posterior computation, in which all parameters and latent variables can be sampled easily from their full conditional distributions. The proposed method shows a good estimation performance in simulation studies and is further illustrated by a real-life application to the patient data from the Aerobics Center Longitudinal Study. The R code for the proposed methodology is made available for public use.</p>","PeriodicalId":49908,"journal":{"name":"Lifetime Data Analysis","volume":" ","pages":"950-969"},"PeriodicalIF":1.0,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144976304","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-07-01Epub Date: 2025-06-14DOI: 10.1007/s10985-025-09658-x
Chi Wing Chu, Hok Kan Ling
We study shape-constrained nonparametric estimation of the underlying survival function in a cross-sectional study without follow-up. Assuming the rate of initiation event is stationary over time, the observed current duration becomes a length-biased and multiplicatively censored counterpart of the underlying failure time of interest. We focus on two shape constraints for the underlying survival function, namely, log-concavity and convexity. The log-concavity constraint is versatile as it allows for log-concave densities, bi-log-concave distributions, increasing densities, and multi-modal densities. We establish the consistency and pointwise asymptotic distribution of the shape-constrained estimators. Specifically, the proposed estimator under log-concavity is consistent and tuning-parameter-free, thus circumventing the well-known inconsistency issue of the Grenander estimator at 0, where correction methods typically involve tuning parameters.
{"title":"Shape-constrained estimation for current duration data in cross-sectional studies.","authors":"Chi Wing Chu, Hok Kan Ling","doi":"10.1007/s10985-025-09658-x","DOIUrl":"10.1007/s10985-025-09658-x","url":null,"abstract":"<p><p>We study shape-constrained nonparametric estimation of the underlying survival function in a cross-sectional study without follow-up. Assuming the rate of initiation event is stationary over time, the observed current duration becomes a length-biased and multiplicatively censored counterpart of the underlying failure time of interest. We focus on two shape constraints for the underlying survival function, namely, log-concavity and convexity. The log-concavity constraint is versatile as it allows for log-concave densities, bi-log-concave distributions, increasing densities, and multi-modal densities. We establish the consistency and pointwise asymptotic distribution of the shape-constrained estimators. Specifically, the proposed estimator under log-concavity is consistent and tuning-parameter-free, thus circumventing the well-known inconsistency issue of the Grenander estimator at 0, where correction methods typically involve tuning parameters.</p>","PeriodicalId":49908,"journal":{"name":"Lifetime Data Analysis","volume":" ","pages":"595-630"},"PeriodicalIF":1.0,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144295231","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-07-01Epub Date: 2025-07-02DOI: 10.1007/s10985-025-09662-1
Dongfeng Wu
A probability method to estimate cancer risk for asymptomatic individuals for the rest of life was developed based on one's current age and screening history using the disease progressive model. The risk is a function of the transition probability density from the disease-free to the preclinical state, the sojourn time in the preclinical state and the screening sensitivity if one had a screening history with negative results. The method can be applied to any chronic disease. As an example, the method was applied to estimate women's breast cancer risk using parameters estimated from the Health Insurance Plan of Greater New York under two scenarios: with and without a screening history, and obtain some meaningful results.
基于一个人的当前年龄和使用疾病进展模型的筛查史,开发了一种估计无症状个体余生癌症风险的概率方法。风险是从无病到临床前状态的转移概率密度、临床前状态的停留时间和筛查敏感性(如果有筛查史结果为阴性)的函数。这种方法适用于任何慢性疾病。以该方法为例,在有筛查史和无筛查史两种情况下,利用大纽约健康保险计划(Health Insurance Plan of Greater New York)估算的参数对女性乳腺癌风险进行了估计,并获得了一些有意义的结果。
{"title":"Estimating the risk of cancer with and without a screening history.","authors":"Dongfeng Wu","doi":"10.1007/s10985-025-09662-1","DOIUrl":"10.1007/s10985-025-09662-1","url":null,"abstract":"<p><p>A probability method to estimate cancer risk for asymptomatic individuals for the rest of life was developed based on one's current age and screening history using the disease progressive model. The risk is a function of the transition probability density from the disease-free to the preclinical state, the sojourn time in the preclinical state and the screening sensitivity if one had a screening history with negative results. The method can be applied to any chronic disease. As an example, the method was applied to estimate women's breast cancer risk using parameters estimated from the Health Insurance Plan of Greater New York under two scenarios: with and without a screening history, and obtain some meaningful results.</p>","PeriodicalId":49908,"journal":{"name":"Lifetime Data Analysis","volume":" ","pages":"702-712"},"PeriodicalIF":1.0,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144555552","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-07-01Epub Date: 2025-04-29DOI: 10.1007/s10985-025-09654-1
Xin Ye, Shu Yang, Xiaofei Wang, Yanyan Liu
In this study, we focus on estimating the heterogeneous treatment effect (HTE) for survival outcome. The outcome is subject to censoring and the number of covariates is high-dimensional. We utilize data from both the randomized controlled trial (RCT), considered as the gold standard, and real-world data (RWD), possibly affected by hidden confounding factors. To achieve a more efficient HTE estimate, such integrative analysis requires great insight into the data generation mechanism, particularly the accurate characterization of unmeasured confounding effects/bias. With this aim, we propose a penalized-regression-based integrative approach that allows for the simultaneous estimation of parameters, selection of variables, and identification of the existence of unmeasured confounding effects. The consistency, asymptotic normality, and efficiency gains are rigorously established for the proposed estimate. Finally, we apply the proposed method to estimate the HTE of lobar/sublobar resection on the survival of lung cancer patients. The RCT is a multicenter non-inferiority randomized phase 3 trial, and the RWD comes from a clinical oncology cancer registry in the United States. The analysis reveals that the unmeasured confounding exists and the integrative approach does enhance the efficiency for the HTE estimation.
{"title":"Integrative analysis of high-dimensional RCT and RWD subject to censoring and hidden confounding.","authors":"Xin Ye, Shu Yang, Xiaofei Wang, Yanyan Liu","doi":"10.1007/s10985-025-09654-1","DOIUrl":"10.1007/s10985-025-09654-1","url":null,"abstract":"<p><p>In this study, we focus on estimating the heterogeneous treatment effect (HTE) for survival outcome. The outcome is subject to censoring and the number of covariates is high-dimensional. We utilize data from both the randomized controlled trial (RCT), considered as the gold standard, and real-world data (RWD), possibly affected by hidden confounding factors. To achieve a more efficient HTE estimate, such integrative analysis requires great insight into the data generation mechanism, particularly the accurate characterization of unmeasured confounding effects/bias. With this aim, we propose a penalized-regression-based integrative approach that allows for the simultaneous estimation of parameters, selection of variables, and identification of the existence of unmeasured confounding effects. The consistency, asymptotic normality, and efficiency gains are rigorously established for the proposed estimate. Finally, we apply the proposed method to estimate the HTE of lobar/sublobar resection on the survival of lung cancer patients. The RCT is a multicenter non-inferiority randomized phase 3 trial, and the RWD comes from a clinical oncology cancer registry in the United States. The analysis reveals that the unmeasured confounding exists and the integrative approach does enhance the efficiency for the HTE estimation.</p>","PeriodicalId":49908,"journal":{"name":"Lifetime Data Analysis","volume":" ","pages":"473-497"},"PeriodicalIF":1.0,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12317910/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144003384","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In survival analysis, researchers commonly focus on variable selection issues in real-world data, particularly when complex network structures exist among covariates. Additionally, due to factors such as data collection costs and delayed entry, real-world data often exhibit censoring and truncation phenomena.This paper addresses left-truncated current status data by employing a copula-based approach to model the relationship between censoring time and failure time. Based on this, we investigate the problem of variable selection in the context of complex network structures among covariates. To this end, we integrate Markov Random Field (MRF) with the Proportional Hazards (PH) model, and extend the latter to more flexibly characterize the correlation structure among covariates. For solving the constructed model, we propose a penalized optimization method and utilize spline functions to estimate the baseline hazard function. Through numerical simulation experiments and case studies of clinical trial data, we comprehensively evaluate the effectiveness and performance of the proposed model and its parameter inference strategy. This evaluation not only demonstrates the robustness of the proposed model in handling complex disease data but also further verifies the high precision and reliability of the parameter estimation method.
{"title":"Regression analysis of a graphical proportional hazards model for informatively left-truncated current status data.","authors":"Mengyue Zhang, Shishu Zhao, Shuying Wang, Xiaolin Xu","doi":"10.1007/s10985-025-09655-0","DOIUrl":"10.1007/s10985-025-09655-0","url":null,"abstract":"<p><p>In survival analysis, researchers commonly focus on variable selection issues in real-world data, particularly when complex network structures exist among covariates. Additionally, due to factors such as data collection costs and delayed entry, real-world data often exhibit censoring and truncation phenomena.This paper addresses left-truncated current status data by employing a copula-based approach to model the relationship between censoring time and failure time. Based on this, we investigate the problem of variable selection in the context of complex network structures among covariates. To this end, we integrate Markov Random Field (MRF) with the Proportional Hazards (PH) model, and extend the latter to more flexibly characterize the correlation structure among covariates. For solving the constructed model, we propose a penalized optimization method and utilize spline functions to estimate the baseline hazard function. Through numerical simulation experiments and case studies of clinical trial data, we comprehensively evaluate the effectiveness and performance of the proposed model and its parameter inference strategy. This evaluation not only demonstrates the robustness of the proposed model in handling complex disease data but also further verifies the high precision and reliability of the parameter estimation method.</p>","PeriodicalId":49908,"journal":{"name":"Lifetime Data Analysis","volume":" ","pages":"498-542"},"PeriodicalIF":1.0,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144047308","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}