首页 > 最新文献

Statistical Methods in Medical Research最新文献

英文 中文
Generalized framework for identifying meaningful heterogenous treatment effects in observational studies: A parametric data-adaptive G-computation approach.
IF 1.6 3区 医学 Q3 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-02-24 DOI: 10.1177/09622802251316969
Roch A Nianogo, Stephen O'Neill, Kosuke Inoue

There has been a renewed interest in identifying heterogenous treatment effects (HTEs) to guide personalized medicine. The objective was to illustrate the use of a step-by-step transparent parametric data-adaptive approach (the generalized HTE approach) based on the G-computation algorithm to detect heterogenous subgroups and estimate meaningful conditional average treatment effects (CATE). The following seven steps implement the generalized HTE approach: Step 1: Select variables that satisfy the backdoor criterion and potential effect modifiers; Step 2: Specify a flexible saturated model including potential confounders and effect modifiers; Step 3: Apply a selection method to reduce overfitting; Step 4: Predict potential outcomes under treatment and no treatment; Step 5: Contrast the potential outcomes for each individual; Step 6: Fit cluster modeling to identify potential effect modifiers; Step 7: Estimate subgroup CATEs. We illustrated the use of this approach using simulated and real data. Our generalized HTE approach successfully identified HTEs and subgroups defined by all effect modifiers using simulated and real data. Our study illustrates that it is feasible to use a step-by-step parametric and transparent data-adaptive approach to detect effect modifiers and identify meaningful HTEs in an observational setting. This approach should be more appealing to epidemiologists interested in explanation.

{"title":"Generalized framework for identifying meaningful heterogenous treatment effects in observational studies: A parametric data-adaptive G-computation approach.","authors":"Roch A Nianogo, Stephen O'Neill, Kosuke Inoue","doi":"10.1177/09622802251316969","DOIUrl":"https://doi.org/10.1177/09622802251316969","url":null,"abstract":"<p><p>There has been a renewed interest in identifying heterogenous treatment effects (HTEs) to guide personalized medicine. The objective was to illustrate the use of a step-by-step transparent parametric data-adaptive approach (the generalized HTE approach) based on the G-computation algorithm to detect heterogenous subgroups and estimate meaningful conditional average treatment effects (CATE). The following seven steps implement the generalized HTE approach: Step 1: Select variables that satisfy the backdoor criterion and potential effect modifiers; Step 2: Specify a flexible saturated model including potential confounders and effect modifiers; Step 3: Apply a selection method to reduce overfitting; Step 4: Predict potential outcomes under treatment and no treatment; Step 5: Contrast the potential outcomes for each individual; Step 6: Fit cluster modeling to identify potential effect modifiers; Step 7: Estimate subgroup CATEs. We illustrated the use of this approach using simulated and real data. Our generalized HTE approach successfully identified HTEs and subgroups defined by all effect modifiers using simulated and real data. Our study illustrates that it is feasible to use a step-by-step parametric and transparent data-adaptive approach to detect effect modifiers and identify meaningful HTEs in an observational setting. This approach should be more appealing to epidemiologists interested in explanation.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"9622802251316969"},"PeriodicalIF":1.6,"publicationDate":"2025-02-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143493492","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Extension of Fisher's least significant difference method to multi-armed group-sequential response-adaptive designs.
IF 1.6 3区 医学 Q3 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-02-24 DOI: 10.1177/09622802251319896
Wenyu Liu, D Stephen Coad

Multi-armed multi-stage designs evaluate experimental treatments using a control arm at interim analyses. Incorporating response-adaptive randomisation in these designs allows early stopping, faster treatment selection and more patients to be assigned to the more promising treatments. Existing frequentist multi-armed multi-stage designs demonstrate that the family-wise error rate is strongly controlled, but they may be too conservative and lack power when the experimental treatments are very different therapies rather than doses of the same drug. Moreover, the designs use a fixed allocation ratio. In this article, Fisher's least significant difference method extended to group-sequential response-adaptive designs is investigated. It is shown mathematically that the information time continues after dropping inferior arms, and hence the error-spending approach can be used to control the family-wise error rate. Two optimal allocations were considered. One ensures efficient estimation of the treatment effects and the other maximises the power subject to a fixed total sample size. Operating characteristics of the group-sequential response-adaptive design for normal and censored survival outcomes based on simulation and redesigning the NeoSphere trial were compared with those of a fixed-sample design. Results show that the adaptive design attains efficient and ethical advantages, and that the family-wise error rate is well controlled.

{"title":"Extension of Fisher's least significant difference method to multi-armed group-sequential response-adaptive designs.","authors":"Wenyu Liu, D Stephen Coad","doi":"10.1177/09622802251319896","DOIUrl":"https://doi.org/10.1177/09622802251319896","url":null,"abstract":"<p><p>Multi-armed multi-stage designs evaluate experimental treatments using a control arm at interim analyses. Incorporating response-adaptive randomisation in these designs allows early stopping, faster treatment selection and more patients to be assigned to the more promising treatments. Existing frequentist multi-armed multi-stage designs demonstrate that the family-wise error rate is strongly controlled, but they may be too conservative and lack power when the experimental treatments are very different therapies rather than doses of the same drug. Moreover, the designs use a fixed allocation ratio. In this article, Fisher's least significant difference method extended to group-sequential response-adaptive designs is investigated. It is shown mathematically that the information time continues after dropping inferior arms, and hence the error-spending approach can be used to control the family-wise error rate. Two optimal allocations were considered. One ensures efficient estimation of the treatment effects and the other maximises the power subject to a fixed total sample size. Operating characteristics of the group-sequential response-adaptive design for normal and censored survival outcomes based on simulation and redesigning the NeoSphere trial were compared with those of a fixed-sample design. Results show that the adaptive design attains efficient and ethical advantages, and that the family-wise error rate is well controlled.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"9622802251319896"},"PeriodicalIF":1.6,"publicationDate":"2025-02-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143493488","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The relative efficiency of staircase and stepped wedge cluster randomised trial designs.
IF 1.6 3区 医学 Q3 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-02-16 DOI: 10.1177/09622802251317613
Kelsey L Grantham, Andrew B Forbes, Richard Hooper, Jessica Kasza

The stepped wedge design is an appealing longitudinal cluster randomised trial design. However, it places a large burden on participating clusters by requiring all clusters to collect data in all periods of the trial. The staircase design may be a desirable alternative: treatment sequences consist of a limited number of measurement periods before and after the implementation of the intervention. In this article, we explore the relative efficiency of the stepped wedge design to several variants of the 'basic staircase' design, which has one control followed by one intervention period in each sequence. We model outcomes using linear mixed models and consider a sampling scheme where each participant is measured once. We first consider a basic staircase design embedded within the stepped wedge design, then basic staircase designs with either more clusters or larger cluster-period sizes, with the same total number of participants and with fewer total participants than the stepped wedge design. The relative efficiency of the designs depends on the intracluster correlation structure, correlation parameters and the trial configuration, including the number of sequences and cluster-period size. For a wide range of realistic trial settings, a basic staircase design will deliver greater statistical power than a stepped wedge design with the same number of participants, and in some cases, with even fewer total participants.

{"title":"The relative efficiency of staircase and stepped wedge cluster randomised trial designs.","authors":"Kelsey L Grantham, Andrew B Forbes, Richard Hooper, Jessica Kasza","doi":"10.1177/09622802251317613","DOIUrl":"https://doi.org/10.1177/09622802251317613","url":null,"abstract":"<p><p>The stepped wedge design is an appealing longitudinal cluster randomised trial design. However, it places a large burden on participating clusters by requiring all clusters to collect data in all periods of the trial. The staircase design may be a desirable alternative: treatment sequences consist of a limited number of measurement periods before and after the implementation of the intervention. In this article, we explore the relative efficiency of the stepped wedge design to several variants of the 'basic staircase' design, which has one control followed by one intervention period in each sequence. We model outcomes using linear mixed models and consider a sampling scheme where each participant is measured once. We first consider a basic staircase design embedded within the stepped wedge design, then basic staircase designs with either more clusters or larger cluster-period sizes, with the same total number of participants and with fewer total participants than the stepped wedge design. The relative efficiency of the designs depends on the intracluster correlation structure, correlation parameters and the trial configuration, including the number of sequences and cluster-period size. For a wide range of realistic trial settings, a basic staircase design will deliver greater statistical power than a stepped wedge design with the same number of participants, and in some cases, with even fewer total participants.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"9622802251317613"},"PeriodicalIF":1.6,"publicationDate":"2025-02-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143433771","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Long-term Dagum-power variance function frailty regression model: Application in health studies.
IF 1.6 3区 医学 Q3 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-02-12 DOI: 10.1177/09622802241304113
Agatha Sacramento Rodrigues, Patrick Borges

Survival models with cure fractions, known as long-term survival models, are widely used in epidemiology to account for both immune and susceptible patients regarding a failure event. In such studies, it is also necessary to estimate unobservable heterogeneity caused by unmeasured prognostic factors. Moreover, the hazard function may exhibit a non-monotonic shape, specifically, an unimodal hazard function. In this article, we propose a long-term survival model based on a defective version of the Dagum distribution, incorporating a power variance function frailty term to account for unobservable heterogeneity. This model accommodates survival data with cure fractions and non-monotonic hazard functions. The distribution is reparameterized in terms of the cure fraction, with covariates linked via a logit link, allowing for direct interpretation of covariate effects on the cure fraction-an uncommon feature in defective approaches. We present maximum likelihood estimation for model parameters, assess performance through Monte Carlo simulations, and illustrate the model's applicability using two health-related datasets: severe COVID-19 in pregnant and postpartum women and patients with malignant skin neoplasms.

{"title":"Long-term Dagum-power variance function frailty regression model: Application in health studies.","authors":"Agatha Sacramento Rodrigues, Patrick Borges","doi":"10.1177/09622802241304113","DOIUrl":"https://doi.org/10.1177/09622802241304113","url":null,"abstract":"<p><p>Survival models with cure fractions, known as long-term survival models, are widely used in epidemiology to account for both immune and susceptible patients regarding a failure event. In such studies, it is also necessary to estimate unobservable heterogeneity caused by unmeasured prognostic factors. Moreover, the hazard function may exhibit a non-monotonic shape, specifically, an unimodal hazard function. In this article, we propose a long-term survival model based on a defective version of the Dagum distribution, incorporating a power variance function frailty term to account for unobservable heterogeneity. This model accommodates survival data with cure fractions and non-monotonic hazard functions. The distribution is reparameterized in terms of the cure fraction, with covariates linked via a logit link, allowing for direct interpretation of covariate effects on the cure fraction-an uncommon feature in defective approaches. We present maximum likelihood estimation for model parameters, assess performance through Monte Carlo simulations, and illustrate the model's applicability using two health-related datasets: severe COVID-19 in pregnant and postpartum women and patients with malignant skin neoplasms.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"9622802241304113"},"PeriodicalIF":1.6,"publicationDate":"2025-02-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143400106","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Jointly assessing multiple endpoints in pilot and feasibility studies. 在试点和可行性研究中联合评估多个终点。
IF 1.6 3区 医学 Q3 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-02-12 DOI: 10.1177/09622802241311219
Robert N Montgomery, Amy E Bodde, Eric D Vidoni

Pilot and feasibility studies are routinely used to determine whether a definitive trial should be pursued; however, the methodologies used to assess feasibility endpoints are often basic and are rarely informed by the requirements of the planned future trial. We propose a new method for analyzing feasibility outcomes which can incorporate relationships between endpoints, utilize a preliminary study design for a future trial and allow for multiple types of feasibility endpoints. The approach specifies a Joint Feasibility Space (JFS) which is the combination of feasibility outcomes that would render a future trial feasible. We estimate the probability of being in the JFS using Bayesian methods and use simulation to create a decision rule based on frequentist operating characteristics. We compare our approach to other general-purpose methods in the literature with simulation and show that our approach has approximately the same performance when analyzing a single feasibility endpoint but is more efficient with more than one endpoint. Feasibility endpoints should be the focus of pilot and feasibility studies. The analyses of these endpoints deserve more attention than they are given, and we have provided a new, effective method their assessment.

{"title":"Jointly assessing multiple endpoints in pilot and feasibility studies.","authors":"Robert N Montgomery, Amy E Bodde, Eric D Vidoni","doi":"10.1177/09622802241311219","DOIUrl":"https://doi.org/10.1177/09622802241311219","url":null,"abstract":"<p><p>Pilot and feasibility studies are routinely used to determine whether a definitive trial should be pursued; however, the methodologies used to assess feasibility endpoints are often basic and are rarely informed by the requirements of the planned future trial. We propose a new method for analyzing feasibility outcomes which can incorporate relationships between endpoints, utilize a preliminary study design for a future trial and allow for multiple types of feasibility endpoints. The approach specifies a Joint Feasibility Space (JFS) which is the combination of feasibility outcomes that would render a future trial feasible. We estimate the probability of being in the JFS using Bayesian methods and use simulation to create a decision rule based on frequentist operating characteristics. We compare our approach to other general-purpose methods in the literature with simulation and show that our approach has approximately the same performance when analyzing a single feasibility endpoint but is more efficient with more than one endpoint. Feasibility endpoints should be the focus of pilot and feasibility studies. The analyses of these endpoints deserve more attention than they are given, and we have provided a new, effective method their assessment.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"9622802241311219"},"PeriodicalIF":1.6,"publicationDate":"2025-02-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143400103","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Robust propensity score estimation via loss function calibration.
IF 1.6 3区 医学 Q3 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-02-12 DOI: 10.1177/09622802241308709
Yimeng Shang, Yu-Han Chiu, Lan Kong

Propensity score estimation is often used as a preliminary step to estimate the average treatment effect with observational data. Nevertheless, misspecification of propensity score models undermines the validity of effect estimates in subsequent analyses. Prediction-based machine learning algorithms are increasingly used to estimate propensity scores to allow for more complex relationships between covariates. However, these approaches may not necessarily achieve covariates balancing. We propose a calibration-based method to better incorporate covariate balance properties in a general modeling framework. Specifically, we calibrate the loss function by adding a covariate imbalance penalty to standard parametric (e.g. logistic regressions) or machine learning models (e.g. neural networks). Our approach may mitigate the impact of model misspecification by explicitly taking into account the covariate balance in the propensity score estimation process. The empirical results show that the proposed method is robust to propensity score model misspecification. The integration of loss function calibration improves the balance of covariates and reduces the root-mean-square error of causal effect estimates. When the propensity score model is misspecified, the neural-network-based model yields the best estimator with less bias and smaller variance as compared to other methods considered.

{"title":"Robust propensity score estimation via loss function calibration.","authors":"Yimeng Shang, Yu-Han Chiu, Lan Kong","doi":"10.1177/09622802241308709","DOIUrl":"https://doi.org/10.1177/09622802241308709","url":null,"abstract":"<p><p>Propensity score estimation is often used as a preliminary step to estimate the average treatment effect with observational data. Nevertheless, misspecification of propensity score models undermines the validity of effect estimates in subsequent analyses. Prediction-based machine learning algorithms are increasingly used to estimate propensity scores to allow for more complex relationships between covariates. However, these approaches may not necessarily achieve covariates balancing. We propose a calibration-based method to better incorporate covariate balance properties in a general modeling framework. Specifically, we calibrate the loss function by adding a covariate imbalance penalty to standard parametric (e.g. logistic regressions) or machine learning models (e.g. neural networks). Our approach may mitigate the impact of model misspecification by explicitly taking into account the covariate balance in the propensity score estimation process. The empirical results show that the proposed method is robust to propensity score model misspecification. The integration of loss function calibration improves the balance of covariates and reduces the root-mean-square error of causal effect estimates. When the propensity score model is misspecified, the neural-network-based model yields the best estimator with less bias and smaller variance as compared to other methods considered.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"9622802241308709"},"PeriodicalIF":1.6,"publicationDate":"2025-02-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143411114","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Using Bayesian evidence synthesis to quantify uncertainty in population trends in smoking behaviour. 利用贝叶斯证据综合法量化人群吸烟行为趋势的不确定性。
IF 1.6 3区 医学 Q3 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-02-12 DOI: 10.1177/09622802241310326
Stephen Wade, Peter Sarich, Pavla Vaneckova, Silvia Behar-Harpaz, Preston J Ngo, Paul B Grogan, Sonya Cressman, Coral E Gartner, John M Murray, Tony Blakely, Emily Banks, Martin C Tammemagi, Karen Canfell, Marianne F Weber, Michael Caruana

Simulation models of smoking behaviour provide vital forecasts of exposure to inform policy targets, estimates of the burden of disease, and impacts of tobacco control interventions. A key element of useful model-based forecasts is a clear picture of uncertainty due to the data used to inform the model, however, assessment of this parameter uncertainty is incomplete in almost all tobacco control models. As a remedy, we demonstrate a Bayesian approach to model calibration that quantifies parameter uncertainty. With a model calibrated to Australian data, we observed that the smoking cessation rate in Australia has increased with calendar year since the late 20th century, and in 2016 people who smoked would quit at a rate of 4.7 quit-events per 100 person-years (90% equal-tailed interval (ETI): 4.5-4.9). We found that those who quit smoking before age 30 years switched to reporting that they never smoked at a rate of approximately 2% annually (90% ETI: 1.9-2.2%). The Bayesian approach demonstrated here can be used as a blueprint to model other population behaviours that are challenging to measure directly, and to provide a clearer picture of uncertainty to decision-makers.

{"title":"Using Bayesian evidence synthesis to quantify uncertainty in population trends in smoking behaviour.","authors":"Stephen Wade, Peter Sarich, Pavla Vaneckova, Silvia Behar-Harpaz, Preston J Ngo, Paul B Grogan, Sonya Cressman, Coral E Gartner, John M Murray, Tony Blakely, Emily Banks, Martin C Tammemagi, Karen Canfell, Marianne F Weber, Michael Caruana","doi":"10.1177/09622802241310326","DOIUrl":"https://doi.org/10.1177/09622802241310326","url":null,"abstract":"<p><p>Simulation models of smoking behaviour provide vital forecasts of exposure to inform policy targets, estimates of the burden of disease, and impacts of tobacco control interventions. A key element of useful model-based forecasts is a clear picture of uncertainty due to the data used to inform the model, however, assessment of this parameter uncertainty is incomplete in almost all tobacco control models. As a remedy, we demonstrate a Bayesian approach to model calibration that quantifies parameter uncertainty. With a model calibrated to Australian data, we observed that the smoking cessation rate in Australia has increased with calendar year since the late 20th century, and in 2016 people who smoked would quit at a rate of 4.7 quit-events per 100 person-years (90% equal-tailed interval (ETI): 4.5-4.9). We found that those who quit smoking before age 30 years switched to reporting that they never smoked at a rate of approximately 2% annually (90% ETI: 1.9-2.2%). The Bayesian approach demonstrated here can be used as a blueprint to model other population behaviours that are challenging to measure directly, and to provide a clearer picture of uncertainty to decision-makers.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"9622802241310326"},"PeriodicalIF":1.6,"publicationDate":"2025-02-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143400108","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Bayesian mixture models for phylogenetic source attribution from consensus sequences and time since infection estimates.
IF 1.6 3区 医学 Q3 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-02-12 DOI: 10.1177/09622802241309750
Alexandra Blenkinsop, Lysandros Sofocleous, Francesco Di Lauro, Evangelia Georgia Kostaki, Ard van Sighem, Daniela Bezemer, Thijs van de Laar, Peter Reiss, Godelieve de Bree, Nikos Pantazis, Oliver Ratmann

In stopping the spread of infectious diseases, pathogen genomic data can be used to reconstruct transmission events and characterize population-level sources of infection. Most approaches for identifying transmission pairs do not account for the time passing since the divergence of pathogen variants in individuals, which is problematic in viruses with high within-host evolutionary rates. This prompted us to consider possible transmission pairs in terms of phylogenetic data and additional estimates of time since infection derived from clinical biomarkers. We develop Bayesian mixture models with an evolutionary clock as a signal component and additional mixed effects or covariate random functions describing the mixing weights to classify potential pairs into likely and unlikely transmission pairs. We demonstrate that although sources cannot be identified at the individual level with certainty, even with the additional data on time elapsed, inferences into the population-level sources of transmission are possible, and more accurate than using only phylogenetic data without time since infection estimates. We apply the proposed approach to estimate age-specific sources of HIV infection in Amsterdam tranamission networks among men who have sex with men between 2010 and 2021. This study demonstrates that infection time estimates provide informative data to characterize transmission sources, and shows how phylogenetic source attribution can then be done with multi-dimensional mixture models.

{"title":"Bayesian mixture models for phylogenetic source attribution from consensus sequences and time since infection estimates.","authors":"Alexandra Blenkinsop, Lysandros Sofocleous, Francesco Di Lauro, Evangelia Georgia Kostaki, Ard van Sighem, Daniela Bezemer, Thijs van de Laar, Peter Reiss, Godelieve de Bree, Nikos Pantazis, Oliver Ratmann","doi":"10.1177/09622802241309750","DOIUrl":"https://doi.org/10.1177/09622802241309750","url":null,"abstract":"<p><p>In stopping the spread of infectious diseases, pathogen genomic data can be used to reconstruct transmission events and characterize population-level sources of infection. Most approaches for identifying transmission pairs do not account for the time passing since the divergence of pathogen variants in individuals, which is problematic in viruses with high within-host evolutionary rates. This prompted us to consider possible transmission pairs in terms of phylogenetic data and additional estimates of time since infection derived from clinical biomarkers. We develop Bayesian mixture models with an evolutionary clock as a signal component and additional mixed effects or covariate random functions describing the mixing weights to classify potential pairs into likely and unlikely transmission pairs. We demonstrate that although sources cannot be identified at the individual level with certainty, even with the additional data on time elapsed, inferences into the population-level sources of transmission are possible, and more accurate than using only phylogenetic data without time since infection estimates. We apply the proposed approach to estimate age-specific sources of HIV infection in Amsterdam tranamission networks among men who have sex with men between 2010 and 2021. This study demonstrates that infection time estimates provide informative data to characterize transmission sources, and shows how phylogenetic source attribution can then be done with multi-dimensional mixture models.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"9622802241309750"},"PeriodicalIF":1.6,"publicationDate":"2025-02-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143400099","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hierarchical Bayesian bivariate spatial modeling of small area proportions with application to health survey data. 应用于健康调查数据的小面积比例的层次贝叶斯双变量空间建模。
IF 1.6 3区 医学 Q3 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-02-11 DOI: 10.1177/09622802251316968
Hanjun Yu, Xinyi Xu, Lichao Yu

In this article, we propose bivariate small area estimation methods for proportions based on the logit-normal mixed models with latent spatial dependence. We incorporate multivariate conditional autoregressive structures for the random effects under the hierarchical Bayesian modeling framework, and extend the methods to accommodate non-sampled regions. Posterior inference is obtained via adaptive Markov chain Monte Carlo algorithms. Extensive simulation studies are carried out to demonstrate the effectiveness of the proposed bivariate spatial models. The results suggest that the proposed methods are more efficient than the univariate and non-spatial methods in estimation and prediction, particularly when bivariate spatial dependence exists. Practical guidelines for model selection based on the simulation results are provided. We further illustrate the application of our methods by estimating the province-level heart disease rates and dyslipidemia rates among the middle-aged and elderly population in China's 31 mainland provinces in 2020.

{"title":"Hierarchical Bayesian bivariate spatial modeling of small area proportions with application to health survey data.","authors":"Hanjun Yu, Xinyi Xu, Lichao Yu","doi":"10.1177/09622802251316968","DOIUrl":"https://doi.org/10.1177/09622802251316968","url":null,"abstract":"<p><p>In this article, we propose bivariate small area estimation methods for proportions based on the logit-normal mixed models with latent spatial dependence. We incorporate multivariate conditional autoregressive structures for the random effects under the hierarchical Bayesian modeling framework, and extend the methods to accommodate non-sampled regions. Posterior inference is obtained via adaptive Markov chain Monte Carlo algorithms. Extensive simulation studies are carried out to demonstrate the effectiveness of the proposed bivariate spatial models. The results suggest that the proposed methods are more efficient than the univariate and non-spatial methods in estimation and prediction, particularly when bivariate spatial dependence exists. Practical guidelines for model selection based on the simulation results are provided. We further illustrate the application of our methods by estimating the province-level heart disease rates and dyslipidemia rates among the middle-aged and elderly population in China's 31 mainland provinces in 2020.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"9622802251316968"},"PeriodicalIF":1.6,"publicationDate":"2025-02-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143392034","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Statistical considerations for evaluating treatment effect under various non-proportional hazard scenarios.
IF 1.6 3区 医学 Q3 HEALTH CARE SCIENCES & SERVICES Pub Date : 2025-02-11 DOI: 10.1177/09622802241313297
Xinyu Zhang, Erich J Greene, Ondrej Blaha, Wei Wei

We conducted a systematic comparison of statistical methods used for the analysis of time-to-event outcomes under various proportional and non-proportional hazard (NPH) scenarios. Our study used data from recently published oncology trials to compare the Log-rank test, still by far the most widely used option, against some available alternatives, including the MaxCombo test, the Restricted Mean Survival Time difference test, the Generalized Gamma model and the Generalized F model. Power, type I error rate, and time-dependent bias with respect to the survival probability and median survival time were used to evaluate and compare the performance of these methods. In addition to the real data, we simulated three hypothetical scenarios with crossing hazards chosen so that the early and late effects "cancel out" and used them to evaluate the ability of the aforementioned methods to detect time-specific and overall treatment effects. We implemented novel metrics for assessing the time-dependent bias in treatment effect estimates to provide a more comprehensive evaluation in NPH scenarios. Recommendations under each NPH scenario are provided by examining the type I error rate, power, and time-dependent bias associated with each statistical approach.

{"title":"Statistical considerations for evaluating treatment effect under various non-proportional hazard scenarios.","authors":"Xinyu Zhang, Erich J Greene, Ondrej Blaha, Wei Wei","doi":"10.1177/09622802241313297","DOIUrl":"https://doi.org/10.1177/09622802241313297","url":null,"abstract":"<p><p>We conducted a systematic comparison of statistical methods used for the analysis of time-to-event outcomes under various proportional and non-proportional hazard (NPH) scenarios. Our study used data from recently published oncology trials to compare the Log-rank test, still by far the most widely used option, against some available alternatives, including the MaxCombo test, the Restricted Mean Survival Time difference test, the Generalized Gamma model and the Generalized F model. Power, type I error rate, and time-dependent bias with respect to the survival probability and median survival time were used to evaluate and compare the performance of these methods. In addition to the real data, we simulated three hypothetical scenarios with crossing hazards chosen so that the early and late effects \"cancel out\" and used them to evaluate the ability of the aforementioned methods to detect time-specific and overall treatment effects. We implemented novel metrics for assessing the time-dependent bias in treatment effect estimates to provide a more comprehensive evaluation in NPH scenarios. Recommendations under each NPH scenario are provided by examining the type I error rate, power, and time-dependent bias associated with each statistical approach.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"9622802241313297"},"PeriodicalIF":1.6,"publicationDate":"2025-02-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143392085","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Statistical Methods in Medical Research
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1