Pub Date : 2025-01-16eCollection Date: 2025-01-01DOI: 10.1080/02664763.2025.2452964
Fuzhi Xu, Shuangge Ma, Qingzhao Zhang
Practical scenarios often present instances where the types of responses are different between multi-source different datasets, reflecting distinct attributes or characteristics. In this paper, an integrative rank-based regression is proposed to facilitate information sharing among varied datasets with multi-type responses. Taking advantage of the rank-based regression, our proposed approach adeptly tackles differences in the magnitude of loss functions. In addition, it can robustly handle outliers and data contamination, and effectively mitigate model misspecification. Extensive numerical simulations demonstrate the superior and competitive performance of the proposed approach in model estimation and variable selection. Analysis of genetic data on HNSC and LUAD yields results with biological explanations and confirms its practical usefulness.
{"title":"Integrative rank-based regression for multi-source high-dimensional data with multi-type responses.","authors":"Fuzhi Xu, Shuangge Ma, Qingzhao Zhang","doi":"10.1080/02664763.2025.2452964","DOIUrl":"10.1080/02664763.2025.2452964","url":null,"abstract":"<p><p>Practical scenarios often present instances where the types of responses are different between multi-source different datasets, reflecting distinct attributes or characteristics. In this paper, an integrative rank-based regression is proposed to facilitate information sharing among varied datasets with multi-type responses. Taking advantage of the rank-based regression, our proposed approach adeptly tackles differences in the magnitude of loss functions. In addition, it can robustly handle outliers and data contamination, and effectively mitigate model misspecification. Extensive numerical simulations demonstrate the superior and competitive performance of the proposed approach in model estimation and variable selection. Analysis of genetic data on HNSC and LUAD yields results with biological explanations and confirms its practical usefulness.</p>","PeriodicalId":15239,"journal":{"name":"Journal of Applied Statistics","volume":"52 11","pages":"2011-2030"},"PeriodicalIF":1.1,"publicationDate":"2025-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12404076/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144992765","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-15eCollection Date: 2025-01-01DOI: 10.1080/02664763.2025.2451977
Jaine de Moura Carvalho, Frank Gomes-Silva, Josimar M Vasconcelos, Gauss M Cordeiro
Several continuous distributions have been proposed recently to provide more flexibility in modeling lifetime data. Among these, the Slashed class of models, particularly the Slashed Lomax ( ) distribution, has gained special attention. This asymmetric model has positive support and it is notable for its stochastic representation and ability to fit heavy-tailed datasets. Despite the increasing number of new continuous models catering to specific samples, there have been few statistical tools introduced to evaluate their goodness-of-fits. To address this deficit, we employ the methodology outlined in J.M. Nicolas [Introduction aux statistiques de deuxième espèce: Applications des logs-moments et des logs-cumulants à l'analyse des lois d'images radar, TS, Trait. Signal 19 (2002), pp. 139-167] derived from the Mellin Transform (MT) to provide new goodness-of-fit measures for the distribution. These measures consider both qualitative and quantitative aspects. We derive the MT for the distribution, calculate the log-cumulants, and construct the log-cumulant diagram. Further, we introduce a test statistic using a combination of Hotelling's statistic and the multivariate Delta method to test hypotheses about the log-cumulants. We apply the new methodology to two real databases in the context of survival analysis to show its effectiveness in evaluating the fit criteria. We conduct bootstrap experiments to assess the power of the proposed test and to evaluate the performance of the estimators. The results revealed that the adjustment tools performed well and that the log-cumulant method proved to be an effective estimation criterion.
最近提出了几个连续分布,以便在建模生命周期数据时提供更大的灵活性。在这些型号中,slash类型号,特别是slash Lomax (SL)分布,获得了特别的关注。这种非对称模型有积极的支持,它的随机表示和适应重尾数据集的能力是值得注意的。尽管有越来越多的新的连续模型迎合特定的样本,但很少有统计工具被引入来评估它们的拟合优度。为了解决这个问题,我们采用了J.M. Nicolas [Introduction aux statisques de deuxime esp: Applications des logs-moments和des logs- cumulative, l'analyse des lois d'images radar, TS, Trait]中概述的方法。信号19(2002),第139-167页)从Mellin变换(MT)中导出,为SL分布提供了新的拟合优度度量。这些措施考虑到定性和定量两个方面。推导了SL分布的MT,计算了对数累积量,构造了对数累积图。此外,我们引入了一个检验统计量,使用霍特林的T 2统计量和多元Delta方法的组合来检验关于对数累积量的假设。我们将新方法应用于生存分析背景下的两个真实数据库,以显示其在评估拟合标准方面的有效性。我们进行自举实验来评估所提出的测试的能力,并评估估计器的性能。结果表明,平差工具性能良好,对数累积量法是一种有效的估计准则。
{"title":"The slashed Lomax distribution: new properties and Mellin-type statistical measures for inference.","authors":"Jaine de Moura Carvalho, Frank Gomes-Silva, Josimar M Vasconcelos, Gauss M Cordeiro","doi":"10.1080/02664763.2025.2451977","DOIUrl":"https://doi.org/10.1080/02664763.2025.2451977","url":null,"abstract":"<p><p>Several continuous distributions have been proposed recently to provide more flexibility in modeling lifetime data. Among these, the Slashed class of models, particularly the Slashed Lomax ( <math><mrow><mi>SL</mi></mrow> </math> ) distribution, has gained special attention. This asymmetric model has positive support and it is notable for its stochastic representation and ability to fit heavy-tailed datasets. Despite the increasing number of new continuous models catering to specific samples, there have been few statistical tools introduced to evaluate their goodness-of-fits. To address this deficit, we employ the methodology outlined in J.M. Nicolas [<i>Introduction aux statistiques de deuxième espèce: Applications des logs-moments et des logs-cumulants à l'analyse des lois d'images radar, TS</i>, Trait. Signal 19 (2002), pp. 139-167] derived from the Mellin Transform (MT) to provide new goodness-of-fit measures for the <math><mrow><mi>SL</mi></mrow> </math> distribution. These measures consider both qualitative and quantitative aspects. We derive the MT for the <math><mrow><mi>SL</mi></mrow> </math> distribution, calculate the log-cumulants, and construct the log-cumulant diagram. Further, we introduce a test statistic using a combination of Hotelling's <math><msup><mi>T</mi> <mn>2</mn></msup> </math> statistic and the multivariate Delta method to test hypotheses about the log-cumulants. We apply the new methodology to two real databases in the context of survival analysis to show its effectiveness in evaluating the fit criteria. We conduct bootstrap experiments to assess the power of the proposed test and to evaluate the performance of the estimators. The results revealed that the adjustment tools performed well and that the log-cumulant method proved to be an effective estimation criterion.</p>","PeriodicalId":15239,"journal":{"name":"Journal of Applied Statistics","volume":"52 10","pages":"1984-2006"},"PeriodicalIF":1.1,"publicationDate":"2025-01-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12320265/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144789266","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-15eCollection Date: 2025-01-01DOI: 10.1080/02664763.2025.2451972
A Batsidis, G Tzavelas, P Economou
The present research deals with statistical inference for the expectation of a function of a random vector based on biased samples. After highlighting with the help of a motivating example the need for conducting this study, using the concept of multivariate weighted distributions, a consistent and asymptotically normally distributed estimator is proposed and utilized for developing statistical inference. A Monte Carlo study is carried out to examine the performance of the estimator proposed. Finally, the analysis of a real-world data set illustrates the benefits of using the proposed methods for statistical inference.
{"title":"Inference under multivariate size-biased sampling.","authors":"A Batsidis, G Tzavelas, P Economou","doi":"10.1080/02664763.2025.2451972","DOIUrl":"https://doi.org/10.1080/02664763.2025.2451972","url":null,"abstract":"<p><p>The present research deals with statistical inference for the expectation of a function of a random vector based on biased samples. After highlighting with the help of a motivating example the need for conducting this study, using the concept of multivariate weighted distributions, a consistent and asymptotically normally distributed estimator is proposed and utilized for developing statistical inference. A Monte Carlo study is carried out to examine the performance of the estimator proposed. Finally, the analysis of a real-world data set illustrates the benefits of using the proposed methods for statistical inference.</p>","PeriodicalId":15239,"journal":{"name":"Journal of Applied Statistics","volume":"52 10","pages":"1968-1983"},"PeriodicalIF":1.1,"publicationDate":"2025-01-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12320264/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144789263","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-09eCollection Date: 2025-01-01DOI: 10.1080/02664763.2024.2449396
Qi Zhou, Haonan He, Jie Zhao, Joon Jin Song
Observational data often exhibit clustered structure, which leads to inaccurate estimates of exposure effect if such structure is ignored. To overcome the challenges of modelling the complex confounder effects in clustered data, we propose a Bayesian doubly robust estimator of causal effects with random intercept BART to enhance the robustness against model misspecification. The proposed approach incorporates the uncertainty in the estimation of the propensity score, potential outcomes and the distribution of individual-level and cluster-level confounders into the exposure effect estimation, thereby improving the coverage probability of interval estimation. We evaluate the proposed method in the simulation study compared with frequentist doubly robust estimators with parametric and nonparametric multilevel modelling strategies. The proposed method is applied to estimate the effect of limited food access on the mortality of cardiovascular disease in the senior population.
{"title":"Bayesian doubly robust estimation of causal effects for clustered observational data.","authors":"Qi Zhou, Haonan He, Jie Zhao, Joon Jin Song","doi":"10.1080/02664763.2024.2449396","DOIUrl":"10.1080/02664763.2024.2449396","url":null,"abstract":"<p><p>Observational data often exhibit clustered structure, which leads to inaccurate estimates of exposure effect if such structure is ignored. To overcome the challenges of modelling the complex confounder effects in clustered data, we propose a Bayesian doubly robust estimator of causal effects with random intercept BART to enhance the robustness against model misspecification. The proposed approach incorporates the uncertainty in the estimation of the propensity score, potential outcomes and the distribution of individual-level and cluster-level confounders into the exposure effect estimation, thereby improving the coverage probability of interval estimation. We evaluate the proposed method in the simulation study compared with frequentist doubly robust estimators with parametric and nonparametric multilevel modelling strategies. The proposed method is applied to estimate the effect of limited food access on the mortality of cardiovascular disease in the senior population.</p>","PeriodicalId":15239,"journal":{"name":"Journal of Applied Statistics","volume":"52 10","pages":"1931-1949"},"PeriodicalIF":1.1,"publicationDate":"2025-01-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12320258/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144789262","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-08eCollection Date: 2025-01-01DOI: 10.1080/02664763.2024.2449413
N Muriel
Zero autocorrelation test statistics of the portmanteau type are studied under dependence. The asymptotic distribution of statistics formed with weighted averages of the autocorrelation and partial autocorrelation functions is theoretically obtained and its accuracy is then analyzed via simulation and in an empirical application. In the simulation study, we find that the proposed statistics provide test with sizes quite close to their nominal, intended sizes and with power functions which show high sensitivity to deviations from the null. It also reveals, for all the lags studied, that the tests are increasingly precise as the sample size increases. An application to financial time series modeling is given where the importance of using robust portmanteau statistics is illustrated. Specifically, we show that traditional tests incur in large deviations from their nominal size, whereas robust tests do not.
{"title":"Weighted portmanteau statistics for testing for zero autocorrelation in dependent data.","authors":"N Muriel","doi":"10.1080/02664763.2024.2449413","DOIUrl":"10.1080/02664763.2024.2449413","url":null,"abstract":"<p><p>Zero autocorrelation test statistics of the portmanteau type are studied under dependence. The asymptotic distribution of statistics formed with weighted averages of the autocorrelation and partial autocorrelation functions is theoretically obtained and its accuracy is then analyzed via simulation and in an empirical application. In the simulation study, we find that the proposed statistics provide test with sizes quite close to their nominal, intended sizes and with power functions which show high sensitivity to deviations from the null. It also reveals, for all the lags studied, that the tests are increasingly precise as the sample size increases. An application to financial time series modeling is given where the importance of using robust portmanteau statistics is illustrated. Specifically, we show that traditional tests incur in large deviations from their nominal size, whereas robust tests do not.</p>","PeriodicalId":15239,"journal":{"name":"Journal of Applied Statistics","volume":"52 10","pages":"1950-1967"},"PeriodicalIF":1.1,"publicationDate":"2025-01-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12320263/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144789267","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-04eCollection Date: 2025-01-01DOI: 10.1080/02664763.2024.2445237
Subhankar Dutta, Suchandan Kayal
In this article, we consider statistical inference based on dependent competing risks data from Marshall-Olkin bivariate Weibull distribution. The maximum likelihood estimates of the unknown model parameters have been computed by using Newton-Raphson method under adaptive Type II progressive hybrid censoring with partially observed failure causes. Existence and uniqueness of maximum likelihood estimates are derived. Approximate confidence intervals have been constructed via the observed Fisher information matrix using asymptotic normality property of the maximum likelihood estimates. Bayes estimates and highest posterior density credible intervals have been calculated under gamma-Dirichlet prior distribution by using Markov chain Monte Carlo technique. Convergence of Markov chain Monte Carlo samples is tested. In addition, a Monte Carlo simulation is carried out to compare the effectiveness of the proposed methods. Further, three different optimality criteria have been taken into account to obtain the most effective censoring plans. From these simulation study results it has been concluded that Bayesian technique produces superior outcomes. Finally, a real-life data set has been analyzed to illustrate the operability and applicability of the proposed methods.
{"title":"Statistical inference for dependent competing risks data under adaptive Type-II progressive hybrid censoring.","authors":"Subhankar Dutta, Suchandan Kayal","doi":"10.1080/02664763.2024.2445237","DOIUrl":"10.1080/02664763.2024.2445237","url":null,"abstract":"<p><p>In this article, we consider statistical inference based on dependent competing risks data from Marshall-Olkin bivariate Weibull distribution. The maximum likelihood estimates of the unknown model parameters have been computed by using Newton-Raphson method under adaptive Type II progressive hybrid censoring with partially observed failure causes. Existence and uniqueness of maximum likelihood estimates are derived. Approximate confidence intervals have been constructed via the observed Fisher information matrix using asymptotic normality property of the maximum likelihood estimates. Bayes estimates and highest posterior density credible intervals have been calculated under gamma-Dirichlet prior distribution by using Markov chain Monte Carlo technique. Convergence of Markov chain Monte Carlo samples is tested. In addition, a Monte Carlo simulation is carried out to compare the effectiveness of the proposed methods. Further, three different optimality criteria have been taken into account to obtain the most effective censoring plans. From these simulation study results it has been concluded that Bayesian technique produces superior outcomes. Finally, a real-life data set has been analyzed to illustrate the operability and applicability of the proposed methods.</p>","PeriodicalId":15239,"journal":{"name":"Journal of Applied Statistics","volume":"52 10","pages":"1871-1903"},"PeriodicalIF":1.1,"publicationDate":"2025-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12320270/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144789265","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-12-31eCollection Date: 2025-01-01DOI: 10.1080/02664763.2024.2447324
Mengyu Li, Xiaoguang Wang
Survival predictions for patients are becoming increasingly important in clinical practice as they play a crucial role in aiding healthcare professionals to make more informed diagnoses and treatment decisions. The nested case-control designs have been extensively utilized as a cost-effective solution in many large cohort studies across epidemiology and other research fields. To achieve accurate survival predictions of individuals from nested case-control studies, we propose a semiparametric model averaging approach based on the partly linear additive proportional hazards structure to avoid the curse of dimensionality. The inverse probability weighting method is considered to estimate the parameters of submodels used in model averaging. We choose the weights by maximizing the pseudo-likelihood function constructed for the aggregated model and discuss the asymptotic optimality of selected weights. Simulation studies are conducted to assess the performance of our proposed model averaging method in the nested case-control study. Furthermore, we apply the proposed approach to real data to demonstrate its superiority.
{"title":"Semiparametric model averaging prediction in nested case-control studies.","authors":"Mengyu Li, Xiaoguang Wang","doi":"10.1080/02664763.2024.2447324","DOIUrl":"10.1080/02664763.2024.2447324","url":null,"abstract":"<p><p>Survival predictions for patients are becoming increasingly important in clinical practice as they play a crucial role in aiding healthcare professionals to make more informed diagnoses and treatment decisions. The nested case-control designs have been extensively utilized as a cost-effective solution in many large cohort studies across epidemiology and other research fields. To achieve accurate survival predictions of individuals from nested case-control studies, we propose a semiparametric model averaging approach based on the partly linear additive proportional hazards structure to avoid the curse of dimensionality. The inverse probability weighting method is considered to estimate the parameters of submodels used in model averaging. We choose the weights by maximizing the pseudo-likelihood function constructed for the aggregated model and discuss the asymptotic optimality of selected weights. Simulation studies are conducted to assess the performance of our proposed model averaging method in the nested case-control study. Furthermore, we apply the proposed approach to real data to demonstrate its superiority.</p>","PeriodicalId":15239,"journal":{"name":"Journal of Applied Statistics","volume":"52 10","pages":"1904-1930"},"PeriodicalIF":1.1,"publicationDate":"2024-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12320267/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144789264","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-12-24eCollection Date: 2025-01-01DOI: 10.1080/02664763.2024.2444649
Albert Whata, Justine B Nasejje, Najmeh Nakhaei Rad, Tshilidzi Mulaudzi, Ding-Geng Chen
The Extended Cox model provides an alternative to the proportional hazard Cox model for modelling data including time-varying covariates. Incorporating time-varying covariates is particularly beneficial when dealing with survival data, as it can improve the precision of survival function estimation. Deep learning methods, in particular, the Deep-pseudo survival neural network (DSNN) model have demonstrated a high potential for accurately predicting right-censored survival data when dealing with time-invariant variables. The DSNN's ability to discretise survival times makes it a natural choice for extending its application to scenarios involving time-varying covariates. This study adapts the DSNN to predict survival probabilities for data with time-varying covariates. To demonstrate this, we considered two scenarios: significant and non-significant time-varying covariates. For significant covariates, the Brier scores were below 0.25 at all considered specific time points, while, in the non-significant case, the Brier scores were above 0.25. The results illustrate that the DSNN performed comparably to the extended Cox, the Dynamic-DeepHit and mulitivariate joint models and on the simulated data. A real-world data application further confirms the predictive potential of the DSNN model in modelling survival data with time-varying covariates.
{"title":"Adapting and evaluating deep-pseudo neural network for survival data with time-varying covariates.","authors":"Albert Whata, Justine B Nasejje, Najmeh Nakhaei Rad, Tshilidzi Mulaudzi, Ding-Geng Chen","doi":"10.1080/02664763.2024.2444649","DOIUrl":"10.1080/02664763.2024.2444649","url":null,"abstract":"<p><p>The Extended Cox model provides an alternative to the proportional hazard Cox model for modelling data including time-varying covariates. Incorporating time-varying covariates is particularly beneficial when dealing with survival data, as it can improve the precision of survival function estimation. Deep learning methods, in particular, the Deep-pseudo survival neural network (DSNN) model have demonstrated a high potential for accurately predicting right-censored survival data when dealing with time-invariant variables. The DSNN's ability to discretise survival times makes it a natural choice for extending its application to scenarios involving time-varying covariates. This study adapts the DSNN to predict survival probabilities for data with time-varying covariates. To demonstrate this, we considered two scenarios: significant and non-significant time-varying covariates. For significant covariates, the Brier scores were below 0.25 at all considered specific time points, while, in the non-significant case, the Brier scores were above 0.25. The results illustrate that the DSNN performed comparably to the extended Cox, the Dynamic-DeepHit and mulitivariate joint models and on the simulated data. A real-world data application further confirms the predictive potential of the DSNN model in modelling survival data with time-varying covariates.</p>","PeriodicalId":15239,"journal":{"name":"Journal of Applied Statistics","volume":"52 10","pages":"1847-1870"},"PeriodicalIF":1.1,"publicationDate":"2024-12-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12320266/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144789261","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-12-16eCollection Date: 2025-01-01DOI: 10.1080/02664763.2024.2435997
Jung Wun Lee, Hayley Dunnack Yackel
The latent class profile model (LCPM) is a widely used technique for identifying distinct subgroups within a sample based on observations' longitudinal responses to categorical items. This paper proposes an expanded version of LCPM by embedding time-specific structures. Such development allows analysts to investigate associations between latent class memberships and time-dependent predictors at specific time points. We suggest a simultaneous estimation of latent class measurement parameters via the expectation-maximization (EM) algorithm, which yields valid point and interval estimators of associations between latent class memberships and covariates. We illustrate the validity of our estimation strategy via numerical studies. In addition, we demonstrate the novelty of the proposed model by analyzing the head and neck cancer data set.
{"title":"Latent class profile model with time-dependent covariates: a study on symptom patterning of patients for head and neck cancer.","authors":"Jung Wun Lee, Hayley Dunnack Yackel","doi":"10.1080/02664763.2024.2435997","DOIUrl":"10.1080/02664763.2024.2435997","url":null,"abstract":"<p><p>The latent class profile model (LCPM) is a widely used technique for identifying distinct subgroups within a sample based on observations' longitudinal responses to categorical items. This paper proposes an expanded version of LCPM by embedding time-specific structures. Such development allows analysts to investigate associations between latent class memberships and time-dependent predictors at specific time points. We suggest a simultaneous estimation of latent class measurement parameters via the expectation-maximization (EM) algorithm, which yields valid point and interval estimators of associations between latent class memberships and covariates. We illustrate the validity of our estimation strategy via numerical studies. In addition, we demonstrate the novelty of the proposed model by analyzing the head and neck cancer data set.</p>","PeriodicalId":15239,"journal":{"name":"Journal of Applied Statistics","volume":"52 8","pages":"1628-1648"},"PeriodicalIF":1.2,"publicationDate":"2024-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12147489/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144266340","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-12-15eCollection Date: 2025-01-01DOI: 10.1080/02664763.2024.2438795
Ayesha Talib, Sajid Ali, Ismail Shah
Control charts are sophisticated graphical tools used to detect and control aberrant variations. Different control schemes are designed to continuously monitor and improve the process stability and performance. This study proposes a bivariate exponentially weighted moving average chart for joint monitoring of the mean vector of Gumbel's bivariate geometric (GBG) data. The performance of the proposed chart is compared with Hotelling's chart. The results of the study indicated that the proposed control chart performs uniformly and substantially better than Hotelling's chart. In addition to two real-life examples, an example based on simulated data is also considered and compared to existing charts to verify the superiority of the proposed chart. Based on the comparisons, it turns out that the MEWMA (GBG) chart outperforms Hotelling's chart and individual EWMA control chart.
{"title":"A control chart for bivariate discrete data monitoring.","authors":"Ayesha Talib, Sajid Ali, Ismail Shah","doi":"10.1080/02664763.2024.2438795","DOIUrl":"10.1080/02664763.2024.2438795","url":null,"abstract":"<p><p>Control charts are sophisticated graphical tools used to detect and control aberrant variations. Different control schemes are designed to continuously monitor and improve the process stability and performance. This study proposes a bivariate exponentially weighted moving average chart for joint monitoring of the mean vector of Gumbel's bivariate geometric (GBG) data. The performance of the proposed chart is compared with Hotelling's <math><msup><mi>T</mi> <mn>2</mn></msup> </math> chart. The results of the study indicated that the proposed control chart performs uniformly and substantially better than Hotelling's <math><msup><mi>T</mi> <mn>2</mn></msup> </math> chart. In addition to two real-life examples, an example based on simulated data is also considered and compared to existing charts to verify the superiority of the proposed chart. Based on the comparisons, it turns out that the MEWMA (GBG) chart outperforms Hotelling's <math><msup><mi>T</mi> <mn>2</mn></msup> </math> chart and individual EWMA control chart.</p>","PeriodicalId":15239,"journal":{"name":"Journal of Applied Statistics","volume":"52 9","pages":"1713-1741"},"PeriodicalIF":1.1,"publicationDate":"2024-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12217121/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144560246","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}