Pub Date : 2023-10-01Epub Date: 2023-07-12DOI: 10.1007/s10985-023-09604-9
Ryan Sun, Dayu Sun, Liang Zhu, Jianguo Sun
In modern biomedical datasets, it is common for recurrent outcomes data to be collected in an incomplete manner. More specifically, information on recurrent events is routinely recorded as a mixture of recurrent event data, panel count data, and panel binary data; we refer to this structure as general mixed recurrent event data. Although the aforementioned data types are individually well-studied, there does not appear to exist an established approach for regression analysis of the three component combination. Often, ad-hoc measures such as imputation or discarding of data are used to homogenize records prior to the analysis, but such measures lead to obvious concerns regarding robustness, loss of efficiency, and other issues. This work proposes a maximum likelihood regression estimation procedure for the combination of general mixed recurrent event data and establishes the asymptotic properties of the proposed estimators. In addition, we generalize the approach to allow for the existence of terminal events, a common complicating feature in recurrent event analysis. Numerical studies and application to the Childhood Cancer Survivor Study suggest that the proposed procedures work well in practical situations.
{"title":"Regression analysis of general mixed recurrent event data.","authors":"Ryan Sun, Dayu Sun, Liang Zhu, Jianguo Sun","doi":"10.1007/s10985-023-09604-9","DOIUrl":"10.1007/s10985-023-09604-9","url":null,"abstract":"<p><p>In modern biomedical datasets, it is common for recurrent outcomes data to be collected in an incomplete manner. More specifically, information on recurrent events is routinely recorded as a mixture of recurrent event data, panel count data, and panel binary data; we refer to this structure as general mixed recurrent event data. Although the aforementioned data types are individually well-studied, there does not appear to exist an established approach for regression analysis of the three component combination. Often, ad-hoc measures such as imputation or discarding of data are used to homogenize records prior to the analysis, but such measures lead to obvious concerns regarding robustness, loss of efficiency, and other issues. This work proposes a maximum likelihood regression estimation procedure for the combination of general mixed recurrent event data and establishes the asymptotic properties of the proposed estimators. In addition, we generalize the approach to allow for the existence of terminal events, a common complicating feature in recurrent event analysis. Numerical studies and application to the Childhood Cancer Survivor Study suggest that the proposed procedures work well in practical situations.</p>","PeriodicalId":49908,"journal":{"name":"Lifetime Data Analysis","volume":" ","pages":"807-822"},"PeriodicalIF":1.2,"publicationDate":"2023-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11334736/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9829612","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-10-01Epub Date: 2023-05-09DOI: 10.1007/s10985-023-09599-3
Jad Beyhum, Jean-Pierre Florens, Ingrid Van Keilegom
This paper discusses nonparametric identification and estimation of the causal effect of a treatment in the presence of confounding, competing risks and random right-censoring. Our identification strategy is based on an instrumental variable. We show that the competing risks model generates a nonparametric quantile instrumental regression problem. Quantile treatment effects on the subdistribution function can be recovered from the regression function. A distinguishing feature of the model is that censoring and competing risks prevent identification at some quantiles. We characterize the set of quantiles for which exact identification is possible and give partial identification results for other quantiles. We outline an estimation procedure and discuss its properties. The finite sample performance of the estimator is evaluated through simulations. We apply the proposed method to the Health Insurance Plan of Greater New York experiment.
{"title":"A nonparametric instrumental approach to confounding in competing risks models.","authors":"Jad Beyhum, Jean-Pierre Florens, Ingrid Van Keilegom","doi":"10.1007/s10985-023-09599-3","DOIUrl":"10.1007/s10985-023-09599-3","url":null,"abstract":"<p><p>This paper discusses nonparametric identification and estimation of the causal effect of a treatment in the presence of confounding, competing risks and random right-censoring. Our identification strategy is based on an instrumental variable. We show that the competing risks model generates a nonparametric quantile instrumental regression problem. Quantile treatment effects on the subdistribution function can be recovered from the regression function. A distinguishing feature of the model is that censoring and competing risks prevent identification at some quantiles. We characterize the set of quantiles for which exact identification is possible and give partial identification results for other quantiles. We outline an estimation procedure and discuss its properties. The finite sample performance of the estimator is evaluated through simulations. We apply the proposed method to the Health Insurance Plan of Greater New York experiment.</p>","PeriodicalId":49908,"journal":{"name":"Lifetime Data Analysis","volume":" ","pages":"709-734"},"PeriodicalIF":1.3,"publicationDate":"2023-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9493533","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-10-01Epub Date: 2023-07-02DOI: 10.1007/s10985-023-09603-w
Eun Ryung Lee, Seyoung Park, Sang Kyu Lee, Hyokyoung G Hong
Despite the urgent need for an effective prediction model tailored to individual interests, existing models have mainly been developed for the mean outcome, targeting average people. Additionally, the direction and magnitude of covariates' effects on the mean outcome may not hold across different quantiles of the outcome distribution. To accommodate the heterogeneous characteristics of covariates and provide a flexible risk model, we propose a quantile forward regression model for high-dimensional survival data. Our method selects variables by maximizing the likelihood of the asymmetric Laplace distribution (ALD) and derives the final model based on the extended Bayesian Information Criterion (EBIC). We demonstrate that the proposed method enjoys a sure screening property and selection consistency. We apply it to the national health survey dataset to show the advantages of a quantile-specific prediction model. Finally, we discuss potential extensions of our approach, including the nonlinear model and the globally concerned quantile regression coefficients model.
{"title":"Quantile forward regression for high-dimensional survival data.","authors":"Eun Ryung Lee, Seyoung Park, Sang Kyu Lee, Hyokyoung G Hong","doi":"10.1007/s10985-023-09603-w","DOIUrl":"10.1007/s10985-023-09603-w","url":null,"abstract":"<p><p>Despite the urgent need for an effective prediction model tailored to individual interests, existing models have mainly been developed for the mean outcome, targeting average people. Additionally, the direction and magnitude of covariates' effects on the mean outcome may not hold across different quantiles of the outcome distribution. To accommodate the heterogeneous characteristics of covariates and provide a flexible risk model, we propose a quantile forward regression model for high-dimensional survival data. Our method selects variables by maximizing the likelihood of the asymmetric Laplace distribution (ALD) and derives the final model based on the extended Bayesian Information Criterion (EBIC). We demonstrate that the proposed method enjoys a sure screening property and selection consistency. We apply it to the national health survey dataset to show the advantages of a quantile-specific prediction model. Finally, we discuss potential extensions of our approach, including the nonlinear model and the globally concerned quantile regression coefficients model.</p>","PeriodicalId":49908,"journal":{"name":"Lifetime Data Analysis","volume":" ","pages":"769-806"},"PeriodicalIF":1.3,"publicationDate":"2023-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9730710","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-07-01DOI: 10.1007/s10985-023-09589-5
Kathrin Möllenhoff, Achim Tresch
The classical approach to analyze time-to-event data, e.g. in clinical trials, is to fit Kaplan-Meier curves yielding the treatment effect as the hazard ratio between treatment groups. Afterwards, a log-rank test is commonly performed to investigate whether there is a difference in survival or, depending on additional covariates, a Cox proportional hazard model is used. However, in numerous trials these approaches fail due to the presence of non-proportional hazards, resulting in difficulties of interpreting the hazard ratio and a loss of power. When considering equivalence or non-inferiority trials, the commonly performed log-rank based tests are similarly affected by a violation of this assumption. Here we propose a parametric framework to assess equivalence or non-inferiority for survival data. We derive pointwise confidence bands for both, the hazard ratio and the difference of the survival curves. Further we propose a test procedure addressing non-inferiority and equivalence by directly comparing the survival functions at certain time points or over an entire range of time. Once the model's suitability is proven the method provides a noticeable power benefit, irrespectively of the shape of the hazard ratio. On the other hand, model selection should be carried out carefully as misspecification may cause type I error inflation in some situations. We investigate the robustness and demonstrate the advantages and disadvantages of the proposed methods by means of a simulation study. Finally, we demonstrate the validity of the methods by a clinical trial example.
{"title":"Investigating non-inferiority or equivalence in time-to-event data under non-proportional hazards.","authors":"Kathrin Möllenhoff, Achim Tresch","doi":"10.1007/s10985-023-09589-5","DOIUrl":"https://doi.org/10.1007/s10985-023-09589-5","url":null,"abstract":"<p><p>The classical approach to analyze time-to-event data, e.g. in clinical trials, is to fit Kaplan-Meier curves yielding the treatment effect as the hazard ratio between treatment groups. Afterwards, a log-rank test is commonly performed to investigate whether there is a difference in survival or, depending on additional covariates, a Cox proportional hazard model is used. However, in numerous trials these approaches fail due to the presence of non-proportional hazards, resulting in difficulties of interpreting the hazard ratio and a loss of power. When considering equivalence or non-inferiority trials, the commonly performed log-rank based tests are similarly affected by a violation of this assumption. Here we propose a parametric framework to assess equivalence or non-inferiority for survival data. We derive pointwise confidence bands for both, the hazard ratio and the difference of the survival curves. Further we propose a test procedure addressing non-inferiority and equivalence by directly comparing the survival functions at certain time points or over an entire range of time. Once the model's suitability is proven the method provides a noticeable power benefit, irrespectively of the shape of the hazard ratio. On the other hand, model selection should be carried out carefully as misspecification may cause type I error inflation in some situations. We investigate the robustness and demonstrate the advantages and disadvantages of the proposed methods by means of a simulation study. Finally, we demonstrate the validity of the methods by a clinical trial example.</p>","PeriodicalId":49908,"journal":{"name":"Lifetime Data Analysis","volume":"29 3","pages":"483-507"},"PeriodicalIF":1.3,"publicationDate":"2023-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10258187/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9624247","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-07-01DOI: 10.1007/s10985-022-09587-z
Marie Böhnstedt, Jutta Gampe, Monique A A Caljouw, Hein Putter
In studies of recurrent events, joint modeling approaches are often needed to allow for potential dependent censoring by a terminal event such as death. Joint frailty models for recurrent events and death with an additional dependence parameter have been studied for cases in which individuals are observed from the start of the event processes. However, samples are often selected at a later time, which results in delayed entry so that only individuals who have not yet experienced the terminal event will be included. In joint frailty models such left truncation has effects on the frailty distribution that need to be accounted for in both the recurrence process and the terminal event process, if the two are associated. We demonstrate, in a comprehensive simulation study, the effects that not adjusting for late entry can have and derive the correctly adjusted marginal likelihood, which can be expressed as a ratio of two integrals over the frailty distribution. We extend the estimation method of Liu and Huang (Stat Med 27:2665-2683, 2008. https://doi.org/10.1002/sim.3077 ) to include potential left truncation. Numerical integration is performed by Gaussian quadrature, the baseline intensities are specified as piecewise constant functions, potential covariates are assumed to have multiplicative effects on the intensities. We apply the method to estimate age-specific intensities of recurrent urinary tract infections and mortality in an older population.
在反复事件的研究中,通常需要联合建模方法,以允许潜在的依赖于死亡等终端事件的审查。对于从事件过程开始就观察到个体的情况,研究了带有附加依赖参数的复发事件和死亡的联合脆弱性模型。然而,样本通常是在稍后的时间选择的,这导致延迟进入,因此只有尚未经历过终端事件的个体将被包括在内。在联合脆弱性模型中,这种左截断对脆弱性分布有影响,如果在复发过程和终止事件过程中两者都有关联,则需要考虑这种影响。在全面的模拟研究中,我们证明了不调整晚进入的影响,并推导出正确调整的边际似然,它可以表示为脆弱性分布上两个积分的比率。我们推广了Liu和Huang (Stat Med 27:2665-2683, 2008)的估计方法。https://doi.org/10.1002/sim.3077)包括潜在的左截断。采用高斯正交法进行数值积分,将基线强度指定为分段常数函数,假设潜在协变量对强度具有乘法效应。我们应用该方法来估计老年人群中复发性尿路感染的年龄特异性强度和死亡率。
{"title":"Incorporating delayed entry into the joint frailty model for recurrent events and a terminal event.","authors":"Marie Böhnstedt, Jutta Gampe, Monique A A Caljouw, Hein Putter","doi":"10.1007/s10985-022-09587-z","DOIUrl":"https://doi.org/10.1007/s10985-022-09587-z","url":null,"abstract":"<p><p>In studies of recurrent events, joint modeling approaches are often needed to allow for potential dependent censoring by a terminal event such as death. Joint frailty models for recurrent events and death with an additional dependence parameter have been studied for cases in which individuals are observed from the start of the event processes. However, samples are often selected at a later time, which results in delayed entry so that only individuals who have not yet experienced the terminal event will be included. In joint frailty models such left truncation has effects on the frailty distribution that need to be accounted for in both the recurrence process and the terminal event process, if the two are associated. We demonstrate, in a comprehensive simulation study, the effects that not adjusting for late entry can have and derive the correctly adjusted marginal likelihood, which can be expressed as a ratio of two integrals over the frailty distribution. We extend the estimation method of Liu and Huang (Stat Med 27:2665-2683, 2008. https://doi.org/10.1002/sim.3077 ) to include potential left truncation. Numerical integration is performed by Gaussian quadrature, the baseline intensities are specified as piecewise constant functions, potential covariates are assumed to have multiplicative effects on the intensities. We apply the method to estimate age-specific intensities of recurrent urinary tract infections and mortality in an older population.</p>","PeriodicalId":49908,"journal":{"name":"Lifetime Data Analysis","volume":"29 3","pages":"585-607"},"PeriodicalIF":1.3,"publicationDate":"2023-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9615411","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-07-01DOI: 10.1007/s10985-022-09584-2
Peng Liu, Kwun Chuen Gary Chan, Ying Qing Chen
Retrospective sampling can be useful in epidemiological research for its convenience to explore an etiological association. One particular retrospective sampling is that disease outcomes of the time-to-event type are collected subject to right truncation, along with other covariates of interest. For regression analysis of the right-truncated time-to-event data, the so-called proportional reverse-time hazards model has been proposed, but the interpretation of its regression parameters tends to be cumbersome, which has greatly hampered its application in practice. In this paper, we instead consider the proportional odds model, an appealing alternative to the popular proportional hazards model. Under the proportional odds model, there is an embedded relationship between the reverse-time hazard function and the usual hazard function. Building on this relationship, we provide a simple procedure to estimate the regression parameters in the proportional odds model for the right truncated data. Weighted estimations are also studied.
{"title":"On a simple estimation of the proportional odds model under right truncation.","authors":"Peng Liu, Kwun Chuen Gary Chan, Ying Qing Chen","doi":"10.1007/s10985-022-09584-2","DOIUrl":"https://doi.org/10.1007/s10985-022-09584-2","url":null,"abstract":"<p><p>Retrospective sampling can be useful in epidemiological research for its convenience to explore an etiological association. One particular retrospective sampling is that disease outcomes of the time-to-event type are collected subject to right truncation, along with other covariates of interest. For regression analysis of the right-truncated time-to-event data, the so-called proportional reverse-time hazards model has been proposed, but the interpretation of its regression parameters tends to be cumbersome, which has greatly hampered its application in practice. In this paper, we instead consider the proportional odds model, an appealing alternative to the popular proportional hazards model. Under the proportional odds model, there is an embedded relationship between the reverse-time hazard function and the usual hazard function. Building on this relationship, we provide a simple procedure to estimate the regression parameters in the proportional odds model for the right truncated data. Weighted estimations are also studied.</p>","PeriodicalId":49908,"journal":{"name":"Lifetime Data Analysis","volume":"29 3","pages":"537-554"},"PeriodicalIF":1.3,"publicationDate":"2023-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10258175/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9614963","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-07-01DOI: 10.1007/s10985-023-09593-9
Yichen Lou, Peijie Wang, Jianguo Sun
The case-cohort design was developed to reduce costs when disease incidence is low and covariates are difficult to obtain. However, most of the existing methods are for right-censored data and there exists only limited research on interval-censored data, especially on regression analysis of bivariate interval-censored data. Interval-censored failure time data frequently occur in many areas and a large literature on their analyses has been established. In this paper, we discuss the situation of bivariate interval-censored data arising from case-cohort studies. For the problem, a class of semiparametric transformation frailty models is presented and for inference, a sieve weighted likelihood approach is developed. The large sample properties, including the consistency of the proposed estimators and the asymptotic normality of the regression parameter estimators, are established. Moreover, a simulation is conducted to assess the finite sample performance of the proposed method and suggests that it performs well in practice.
{"title":"A semi-parametric weighted likelihood approach for regression analysis of bivariate interval-censored outcomes from case-cohort studies.","authors":"Yichen Lou, Peijie Wang, Jianguo Sun","doi":"10.1007/s10985-023-09593-9","DOIUrl":"https://doi.org/10.1007/s10985-023-09593-9","url":null,"abstract":"<p><p>The case-cohort design was developed to reduce costs when disease incidence is low and covariates are difficult to obtain. However, most of the existing methods are for right-censored data and there exists only limited research on interval-censored data, especially on regression analysis of bivariate interval-censored data. Interval-censored failure time data frequently occur in many areas and a large literature on their analyses has been established. In this paper, we discuss the situation of bivariate interval-censored data arising from case-cohort studies. For the problem, a class of semiparametric transformation frailty models is presented and for inference, a sieve weighted likelihood approach is developed. The large sample properties, including the consistency of the proposed estimators and the asymptotic normality of the regression parameter estimators, are established. Moreover, a simulation is conducted to assess the finite sample performance of the proposed method and suggests that it performs well in practice.</p>","PeriodicalId":49908,"journal":{"name":"Lifetime Data Analysis","volume":"29 3","pages":"628-653"},"PeriodicalIF":1.3,"publicationDate":"2023-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9615456","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-07-01DOI: 10.1007/s10985-023-09597-5
Erik T Parner, Per K Andersen, Morten Overgaard
Jack-knife pseudo-observations have in recent decades gained popularity in regression analysis for various aspects of time-to-event data. A limitation of the jack-knife pseudo-observations is that their computation is time consuming, as the base estimate needs to be recalculated when leaving out each observation. We show that jack-knife pseudo-observations can be closely approximated using the idea of the infinitesimal jack-knife residuals. The infinitesimal jack-knife pseudo-observations are much faster to compute than jack-knife pseudo-observations. A key assumption of the unbiasedness of the jack-knife pseudo-observation approach is on the influence function of the base estimate. We reiterate why the condition on the influence function is needed for unbiased inference and show that the condition is not satisfied for the Kaplan-Meier base estimate in a left-truncated cohort. We present a modification of the infinitesimal jack-knife pseudo-observations that provide unbiased estimates in a left-truncated cohort. The computational speed and medium and large sample properties of the jack-knife pseudo-observations and infinitesimal jack-knife pseudo-observation are compared and we present an application of the modified infinitesimal jack-knife pseudo-observations in a left-truncated cohort of Danish patients with diabetes.
{"title":"Regression models for censored time-to-event data using infinitesimal jack-knife pseudo-observations, with applications to left-truncation.","authors":"Erik T Parner, Per K Andersen, Morten Overgaard","doi":"10.1007/s10985-023-09597-5","DOIUrl":"https://doi.org/10.1007/s10985-023-09597-5","url":null,"abstract":"<p><p>Jack-knife pseudo-observations have in recent decades gained popularity in regression analysis for various aspects of time-to-event data. A limitation of the jack-knife pseudo-observations is that their computation is time consuming, as the base estimate needs to be recalculated when leaving out each observation. We show that jack-knife pseudo-observations can be closely approximated using the idea of the infinitesimal jack-knife residuals. The infinitesimal jack-knife pseudo-observations are much faster to compute than jack-knife pseudo-observations. A key assumption of the unbiasedness of the jack-knife pseudo-observation approach is on the influence function of the base estimate. We reiterate why the condition on the influence function is needed for unbiased inference and show that the condition is not satisfied for the Kaplan-Meier base estimate in a left-truncated cohort. We present a modification of the infinitesimal jack-knife pseudo-observations that provide unbiased estimates in a left-truncated cohort. The computational speed and medium and large sample properties of the jack-knife pseudo-observations and infinitesimal jack-knife pseudo-observation are compared and we present an application of the modified infinitesimal jack-knife pseudo-observations in a left-truncated cohort of Danish patients with diabetes.</p>","PeriodicalId":49908,"journal":{"name":"Lifetime Data Analysis","volume":"29 3","pages":"654-671"},"PeriodicalIF":1.3,"publicationDate":"2023-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10258172/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9622679","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-07-01DOI: 10.1007/s10985-022-09585-1
Zhiqiang Tan
For discrete-time survival data, conditional likelihood inference in Cox's hazard odds model is theoretically desirable but exact calculation is numerical intractable with a moderate to large number of tied events. Unconditional maximum likelihood estimation over both regression coefficients and baseline hazard probabilities can be problematic with a large number of time intervals. We develop new methods and theory using numerically simple estimating functions, along with model-based and model-robust variance estimation, in hazard probability and odds models. For the probability hazard model, we derive as a consistent estimator the Breslow-Peto estimator, previously known as an approximation to the conditional likelihood estimator in the hazard odds model. For the hazard odds model, we propose a weighted Mantel-Haenszel estimator, which satisfies conditional unbiasedness given the numbers of events in addition to the risk sets and covariates, similarly to the conditional likelihood estimator. Our methods are expected to perform satisfactorily in a broad range of settings, with small or large numbers of tied events corresponding to a large or small number of time intervals. The methods are implemented in the R package dSurvival.
{"title":"Consistent and robust inference in hazard probability and odds models with discrete-time survival data.","authors":"Zhiqiang Tan","doi":"10.1007/s10985-022-09585-1","DOIUrl":"https://doi.org/10.1007/s10985-022-09585-1","url":null,"abstract":"<p><p>For discrete-time survival data, conditional likelihood inference in Cox's hazard odds model is theoretically desirable but exact calculation is numerical intractable with a moderate to large number of tied events. Unconditional maximum likelihood estimation over both regression coefficients and baseline hazard probabilities can be problematic with a large number of time intervals. We develop new methods and theory using numerically simple estimating functions, along with model-based and model-robust variance estimation, in hazard probability and odds models. For the probability hazard model, we derive as a consistent estimator the Breslow-Peto estimator, previously known as an approximation to the conditional likelihood estimator in the hazard odds model. For the hazard odds model, we propose a weighted Mantel-Haenszel estimator, which satisfies conditional unbiasedness given the numbers of events in addition to the risk sets and covariates, similarly to the conditional likelihood estimator. Our methods are expected to perform satisfactorily in a broad range of settings, with small or large numbers of tied events corresponding to a large or small number of time intervals. The methods are implemented in the R package dSurvival.</p>","PeriodicalId":49908,"journal":{"name":"Lifetime Data Analysis","volume":"29 3","pages":"555-584"},"PeriodicalIF":1.3,"publicationDate":"2023-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9613466","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-07-01Epub Date: 2023-03-08DOI: 10.1007/s10985-023-09591-x
Wende Clarence Safari, Ignacio López-de-Ullibarri, María Amalia Jácome
This paper addresses the problem of estimating the conditional survival function of the lifetime of the subjects experiencing the event (latency) in the mixture cure model when the cure status information is partially available. The approach of past work relies on the assumption that long-term survivors are unidentifiable because of right censoring. However, in some cases this assumption is invalid since some subjects are known to be cured, e.g., when a medical test ascertains that a disease has entirely disappeared after treatment. We propose a latency estimator that extends the nonparametric estimator studied in López-Cheda et al. (TEST 26(2):353-376, 2017b) to the case when the cure status is partially available. We establish the asymptotic normality distribution of the estimator, and illustrate its performance in a simulation study. Finally, the estimator is applied to a medical dataset to study the length of hospital stay of COVID-19 patients requiring intensive care.
{"title":"Latency function estimation under the mixture cure model when the cure status is available.","authors":"Wende Clarence Safari, Ignacio López-de-Ullibarri, María Amalia Jácome","doi":"10.1007/s10985-023-09591-x","DOIUrl":"10.1007/s10985-023-09591-x","url":null,"abstract":"<p><p>This paper addresses the problem of estimating the conditional survival function of the lifetime of the subjects experiencing the event (latency) in the mixture cure model when the cure status information is partially available. The approach of past work relies on the assumption that long-term survivors are unidentifiable because of right censoring. However, in some cases this assumption is invalid since some subjects are known to be cured, e.g., when a medical test ascertains that a disease has entirely disappeared after treatment. We propose a latency estimator that extends the nonparametric estimator studied in López-Cheda et al. (TEST 26(2):353-376, 2017b) to the case when the cure status is partially available. We establish the asymptotic normality distribution of the estimator, and illustrate its performance in a simulation study. Finally, the estimator is applied to a medical dataset to study the length of hospital stay of COVID-19 patients requiring intensive care.</p>","PeriodicalId":49908,"journal":{"name":"Lifetime Data Analysis","volume":"29 3","pages":"608-627"},"PeriodicalIF":1.3,"publicationDate":"2023-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9994787/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9619729","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}