Pub Date : 2025-10-06DOI: 10.1177/09622802251380625
Rajenki Das, Nina Deliu, Mark R Toshner, Sofía S Villar
Although response-adaptive randomisation (RAR) has gained substantial attention in the literature, it still has limited use in clinical trials. Amongst other reasons, the implementation of RAR in real world trials raises important practical questions, often neglected in the technical literature. Motivated by an innovative phase-II stratified RAR rare-disease trial, this paper addresses two challenges: (1) How to ensure that RAR allocations are desirable, that is, both acceptable and faithful to the intended probabilities, particularly in small samples? and (2) What adaptations to trigger after interim analyses in the presence of missing data? To answer (1), we propose a Mapping strategy that discretises the randomisation probabilities into a vector of allocation ratios, resulting in improved frequentist errors. Under the implementation of Mapping, we answer (2) by analysing the impact of missing data on operating characteristics in selected scenarios. Finally, we discuss additional concerns including: pooling data across trial strata, analysing the level of blinding in the trial, and reporting safety results.
{"title":"Implementing response-adaptive randomisation in stratified rare-disease trials: Design challenges and practical solutions.","authors":"Rajenki Das, Nina Deliu, Mark R Toshner, Sofía S Villar","doi":"10.1177/09622802251380625","DOIUrl":"https://doi.org/10.1177/09622802251380625","url":null,"abstract":"<p><p>Although response-adaptive randomisation (RAR) has gained substantial attention in the literature, it still has limited use in clinical trials. Amongst other reasons, the implementation of RAR in real world trials raises important practical questions, often neglected in the technical literature. Motivated by an innovative phase-II stratified RAR rare-disease trial, this paper addresses two challenges: (1) How to ensure that RAR allocations are desirable, that is, both acceptable and faithful to the intended probabilities, particularly in small samples? and (2) What adaptations to trigger after interim analyses in the presence of missing data? To answer (1), we propose a <i>Mapping</i> strategy that discretises the randomisation probabilities into a vector of allocation ratios, resulting in improved frequentist errors. Under the implementation of <i>Mapping</i>, we answer (2) by analysing the impact of missing data on operating characteristics in selected scenarios. Finally, we discuss additional concerns including: pooling data across trial strata, analysing the level of blinding in the trial, and reporting safety results.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"9622802251380625"},"PeriodicalIF":1.9,"publicationDate":"2025-10-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145239772","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-01Epub Date: 2025-07-09DOI: 10.1177/09622802251354921
Xiaoyu Cai, Wei Zhang, Huiyun Li, Zhaohai Li, Aiyi Liu
The receiver operating characteristic curve is a popular tool for evaluating the discriminative ability of a diagnostic biomarker. Parametric and nonparametric methods exist in the literature for estimation of a receiver operating characteristic curve and its associated summary measures using data usually collected from a case-control study. Since the receiver operating characteristic curve remains unchanged under a monotone transformation, the biomarker data from both cases (diseased subjects) and controls (non-diseased subjects) are often transformed based on a common Box-Cox transformation (or other appropriate transformation) prior to the application of a parametric estimation method. However, careful examination of the data often reveals that the biomarker values in the diseased and non-diseased population can only be normally approximated via different transformations. In this situation, existing estimation methods cannot be directly applied to the heterogeneously-transformed data. In this article, we deal with the situation that biomarker data from both diseased and non-diseased population are normally distributed after being transformed with different Box-Cox transformations. Under this assumption, we show that existing methods based on a common Box-Cox transformation are invalid in that they possess substantial biases. We move on to propose a method to estimate the underlying receiver operating characteristic curve and its area under the curve, and investigate its performance as compared to the nonparametric estimator that ignores any distributional assumptions as well as the estimators based on a common Box-Cox transformation assumptions. The method is exemplified with HIV infection data from the National Health and Nutrition Examination Survey (NHANES).
{"title":"Estimation of receiver operating characteristic curve when case and control require different transformations for normality.","authors":"Xiaoyu Cai, Wei Zhang, Huiyun Li, Zhaohai Li, Aiyi Liu","doi":"10.1177/09622802251354921","DOIUrl":"10.1177/09622802251354921","url":null,"abstract":"<p><p>The receiver operating characteristic curve is a popular tool for evaluating the discriminative ability of a diagnostic biomarker. Parametric and nonparametric methods exist in the literature for estimation of a receiver operating characteristic curve and its associated summary measures using data usually collected from a case-control study. Since the receiver operating characteristic curve remains unchanged under a monotone transformation, the biomarker data from both cases (diseased subjects) and controls (non-diseased subjects) are often transformed based on a common Box-Cox transformation (or other appropriate transformation) prior to the application of a parametric estimation method. However, careful examination of the data often reveals that the biomarker values in the diseased and non-diseased population can only be normally approximated via different transformations. In this situation, existing estimation methods cannot be directly applied to the heterogeneously-transformed data. In this article, we deal with the situation that biomarker data from both diseased and non-diseased population are normally distributed after being transformed with different Box-Cox transformations. Under this assumption, we show that existing methods based on a common Box-Cox transformation are invalid in that they possess substantial biases. We move on to propose a method to estimate the underlying receiver operating characteristic curve and its area under the curve, and investigate its performance as compared to the nonparametric estimator that ignores any distributional assumptions as well as the estimators based on a common Box-Cox transformation assumptions. The method is exemplified with HIV infection data from the National Health and Nutrition Examination Survey (NHANES).</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1982-1993"},"PeriodicalIF":1.9,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12710658/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144592388","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-01Epub Date: 2025-07-17DOI: 10.1177/09622802251340250
Yaxian Chen, Yeonhee Park
Driven by evolving Food and Drug Administration recommendations, modern clinical trials demand innovative designs that strike a balance between statistical rigor and ethical considerations. Covariate-adjusted response-adaptive randomization (CARA) designs bridge this gap by utilizing patient attributes and responses to skew treatment allocation in favor of the treatment to be best for an individual patient's profiles. However, existing CARA designs for survival outcomes often rely on specific parametric models, constraining their applicability in clinical practice. To overcome this limitation, we propose a novel CARA method for survival outcomes (called CARAS) based on the Cox model, which improves model flexibility and mitigate risks of model misspecification. Additionally, we introduce a group sequential overlap-weighted log-rank test to preserve the type I error rate in group sequential trials using CARAS. Comprehensive simulation studies and a real-world trial example demonstrate the proposed method's clinical benefit, statistical efficiency, and robustness to model misspecification compared to traditional randomized controlled trial designs and response-adaptive randomization designs.
{"title":"Group sequential designs for survival outcomes with adaptive randomization.","authors":"Yaxian Chen, Yeonhee Park","doi":"10.1177/09622802251340250","DOIUrl":"10.1177/09622802251340250","url":null,"abstract":"<p><p>Driven by evolving Food and Drug Administration recommendations, modern clinical trials demand innovative designs that strike a balance between statistical rigor and ethical considerations. Covariate-adjusted response-adaptive randomization (CARA) designs bridge this gap by utilizing patient attributes and responses to skew treatment allocation in favor of the treatment to be best for an individual patient's profiles. However, existing CARA designs for survival outcomes often rely on specific parametric models, constraining their applicability in clinical practice. To overcome this limitation, we propose a novel CARA method for survival outcomes (called CARAS) based on the Cox model, which improves model flexibility and mitigate risks of model misspecification. Additionally, we introduce a group sequential overlap-weighted log-rank test to preserve the type I error rate in group sequential trials using CARAS. Comprehensive simulation studies and a real-world trial example demonstrate the proposed method's clinical benefit, statistical efficiency, and robustness to model misspecification compared to traditional randomized controlled trial designs and response-adaptive randomization designs.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1952-1967"},"PeriodicalIF":1.9,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144650606","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Despite the widespread use of time-to-event data in precision medicine, existing research has often neglected the presence of the cure fraction, assuming that all individuals will inevitably experience the event of interest. When a cure fraction is present, the cure rate and survival time of uncured patients should be considered in estimating the optimal individualized treatment regimes. In this study, we propose direct methods for estimating the optimal individualized treatment regimes that either maximize the cure rate or mean survival time of uncured patients. Additionally, we propose two optimal individualized treatment regimes that balance the tradeoff between the cure rate and mean survival time of uncured patients based on a constrained estimation framework for a more comprehensive assessment of individualized treatment regimes. This framework allows us to estimate the optimal individualized treatment regime that maximizes the population's cure rate without significantly compromising the mean survival time of those who remain uncured or maximizes the mean survival time of uncured patients while having the cure rate controlled at a desired level. The exterior-point algorithm is adopted to expedite the resolution of the constrained optimization problem and statistical validity is rigorously established. Furthermore, the advantages of the proposed methods are demonstrated via simulations and analysis of esophageal cancer data.
{"title":"Optimal treatment regimes in the presence of a cure fraction.","authors":"Chenrui Qi, Zicheng Lin, Baqun Zhang, Cunjie Lin, Zishu Zhan","doi":"10.1177/09622802251338399","DOIUrl":"10.1177/09622802251338399","url":null,"abstract":"<p><p>Despite the widespread use of time-to-event data in precision medicine, existing research has often neglected the presence of the cure fraction, assuming that all individuals will inevitably experience the event of interest. When a cure fraction is present, the cure rate and survival time of uncured patients should be considered in estimating the optimal individualized treatment regimes. In this study, we propose direct methods for estimating the optimal individualized treatment regimes that either maximize the cure rate or mean survival time of uncured patients. Additionally, we propose two optimal individualized treatment regimes that balance the tradeoff between the cure rate and mean survival time of uncured patients based on a constrained estimation framework for a more comprehensive assessment of individualized treatment regimes. This framework allows us to estimate the optimal individualized treatment regime that maximizes the population's cure rate without significantly compromising the mean survival time of those who remain uncured or maximizes the mean survival time of uncured patients while having the cure rate controlled at a desired level. The exterior-point algorithm is adopted to expedite the resolution of the constrained optimization problem and statistical validity is rigorously established. Furthermore, the advantages of the proposed methods are demonstrated via simulations and analysis of esophageal cancer data.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1935-1951"},"PeriodicalIF":1.9,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144650607","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-01Epub Date: 2025-08-06DOI: 10.1177/09622802251360413
Alessandra Giovagnoli, Monia Lupparelli
Starting from a Bayesian perspective, this paper proposes a novel response adaptive randomization rule based on the use of the predictive distribution. The intent is to design a randomized mechanism that favors the allocation of the next patient to the "best" treatment, considering the expected future outcomes obtained by combining accrued data with prior information. This predictive rule also stems from a decision-theoretic approach. The method is driven by patients' beneficial motivations, fully debated in this work, but also accounts for essential inferential purposes in clinical trials discussed within the framework of frequentist inference. Some asymptotic properties of the proposed rule are proved and also shown through numerical studies, which compare this method with other competing ones, as the notable Thompson rule for the special case of binary outcomes.
{"title":"A note on response-adaptive randomization from a Bayesian prediction viewpoint.","authors":"Alessandra Giovagnoli, Monia Lupparelli","doi":"10.1177/09622802251360413","DOIUrl":"10.1177/09622802251360413","url":null,"abstract":"<p><p>Starting from a Bayesian perspective, this paper proposes a novel response adaptive randomization rule based on the use of the predictive distribution. The intent is to design a randomized mechanism that favors the allocation of the next patient to the \"best\" treatment, considering the expected future outcomes obtained by combining accrued data with prior information. This predictive rule also stems from a decision-theoretic approach. The method is driven by patients' beneficial motivations, fully debated in this work, but also accounts for essential inferential purposes in clinical trials discussed within the framework of frequentist inference. Some asymptotic properties of the proposed rule are proved and also shown through numerical studies, which compare this method with other competing ones, as the notable Thompson rule for the special case of binary outcomes.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"2053-2068"},"PeriodicalIF":1.9,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144795539","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-01Epub Date: 2025-07-14DOI: 10.1177/09622802251356592
Yichen Lou, Mingyue Du, Peijie Wang, Xinyuan Song
This article discusses regression analysis of interval-censored failure time data in the presence of a cure fraction and nonignorable missing covariates. To address the challenges caused by interval censoring, missing covariates and the existence of a cure subgroup, we propose a joint semiparametric modeling framework that simultaneously models the failure time of interest and the missing covariates. In particular, we present a class of semiparametric nonmixture cure models for the failure time and a semiparametric density ratio model for the missing covariates. A two-step likelihood-based estimation procedure is developed and the large sample properties of the resulting estimators are established. An extensive numerical study demonstrates the good performance of the proposed method in practical settings and the proposed approach is applied to an Alzheimer's disease study that motivated this study.
{"title":"Semiparametric regression analysis of interval-censored failure time data with a cure subgroup and nonignorable missing covariates.","authors":"Yichen Lou, Mingyue Du, Peijie Wang, Xinyuan Song","doi":"10.1177/09622802251356592","DOIUrl":"10.1177/09622802251356592","url":null,"abstract":"<p><p>This article discusses regression analysis of interval-censored failure time data in the presence of a cure fraction and nonignorable missing covariates. To address the challenges caused by interval censoring, missing covariates and the existence of a cure subgroup, we propose a joint semiparametric modeling framework that simultaneously models the failure time of interest and the missing covariates. In particular, we present a class of semiparametric nonmixture cure models for the failure time and a semiparametric density ratio model for the missing covariates. A two-step likelihood-based estimation procedure is developed and the large sample properties of the resulting estimators are established. An extensive numerical study demonstrates the good performance of the proposed method in practical settings and the proposed approach is applied to an Alzheimer's disease study that motivated this study.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1994-2010"},"PeriodicalIF":1.9,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144627017","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-01Epub Date: 2025-08-11DOI: 10.1177/09622802251360409
Jingxia Liu, Fan Li, Siobhan Sutcliffe, Graham A Colditz
The optimal designs (ODs) for parallel-arm longitudinal cluster randomized trials, multiple-period cluster randomized crossover (CRXO) trials, and stepped wedge cluster randomized trials (SW-CRTs), including closed-cohort and repeat cross-sectional designs, have been studied separately under a cost-efficiency framework based on generalized estimating equations (GEEs). However, whether a global OD exists across longitudinal designs and randomization schedules remains unknown. Therefore, this research addresses a critical gap by comparing OD feature across complete longitudinal cluster randomized trial designs with two treatment conditions and continuous outcomes. We define the OD as the design with either the lowest cost to obtain a desired level of power or the largest power given a fixed budget. For each of these ODs, we obtain the optimal number of clusters and the optimal cluster-period size (number of participants per cluster per period). To ensure equitable comparisons, we consider the GEE treatment effect estimator with the same block exchangeable correlation structure and develop OD algorithms with the lowest cost for each of six study designs. To obtain OD with the largest power, we summarize the previous and propose new OD algorithms and formulae. We suggest using the number of treatment sequences , where T is the number of time-periods, in both the optimal closed-cohort and repeated cross-sectional SW-CRTs to have the lowest cost. This is consistent with our previous findings for ODs with the largest power in SW-CRTs. Comparing all six ODs, we conclude that optimal closed-cohort CRXO trials are global ODs, yielding both the lowest cost and largest power.
{"title":"Selecting the optimal longitudinal cluster randomized design with a continuous outcome: Parallel-arm, crossover, or stepped-wedge.","authors":"Jingxia Liu, Fan Li, Siobhan Sutcliffe, Graham A Colditz","doi":"10.1177/09622802251360409","DOIUrl":"10.1177/09622802251360409","url":null,"abstract":"<p><p>The optimal designs (ODs) for parallel-arm longitudinal cluster randomized trials, multiple-period cluster randomized crossover (CRXO) trials, and stepped wedge cluster randomized trials (SW-CRTs), including closed-cohort and repeat cross-sectional designs, have been studied separately under a cost-efficiency framework based on generalized estimating equations (GEEs). However, whether a global OD exists across longitudinal designs and randomization schedules remains unknown. Therefore, this research addresses a critical gap by comparing OD feature across complete longitudinal cluster randomized trial designs with two treatment conditions and continuous outcomes. We define the OD as the design with either the lowest cost to obtain a desired level of power or the largest power given a fixed budget. For each of these ODs, we obtain the optimal number of clusters and the optimal cluster-period size (number of participants per cluster per period). To ensure equitable comparisons, we consider the GEE treatment effect estimator with the same block exchangeable correlation structure and develop OD algorithms with the lowest cost for each of six study designs. To obtain OD with the largest power, we summarize the previous and propose new OD algorithms and formulae. We suggest using the number of treatment sequences <math><mi>L</mi><mo>=</mo><mi>T</mi><mo>-</mo><mn>1</mn></math>, where <i>T</i> is the number of time-periods, in both the optimal closed-cohort and repeated cross-sectional SW-CRTs to have the lowest cost. This is consistent with our previous findings for ODs with the largest power in SW-CRTs. Comparing all six ODs, we conclude that optimal closed-cohort CRXO trials are global ODs, yielding both the lowest cost and largest power.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"2069-2090"},"PeriodicalIF":1.9,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144817525","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-01Epub Date: 2025-05-21DOI: 10.1177/09622802251327680
Ornella Moro, Inger Torhild Gram, Maja-Lisa Løchen, Marit B Veierød, Ana Maria Wägner, Giovanni Sebastiani
Future occurrence of a disease can be highly influenced by some specific risk factors. This work presents a comprehensive approach to quantify the event probability as a function of each separate risk factor by means of a parametric model. The proposed methodology is mainly described and applied here in the case of a linear model, but the non-linear case is also addressed. To improve estimation accuracy, three distinct methods are developed and their results are integrated. One of them is Bayesian, based on a non-informative prior. Each of the other two, uses aggregation of sample elements based on their factor values, which is optimized by means of a different specific criterion. For one of these two, optimization is performed by Simulated Annealing. The methodology presented is applicable across various diseases but here we quantify the risk for cardiovascular diseases in subjects with type 1 diabetes. The results obtained combining the three different methods show accurate estimates of cardiovascular risk variation rates for the factors considered. Furthermore, the detection of a biological activation phenomenon for one of the factors is also illustrated. To quantify the performances of the proposed methodology and to compare them with those from a known method used for this type of models, a large simulation study is done, whose results are illustrated here.
{"title":"Quantification of the influence of risk factors with application to cardiovascular diseases in subjects with type 1 diabetes.","authors":"Ornella Moro, Inger Torhild Gram, Maja-Lisa Løchen, Marit B Veierød, Ana Maria Wägner, Giovanni Sebastiani","doi":"10.1177/09622802251327680","DOIUrl":"10.1177/09622802251327680","url":null,"abstract":"<p><p>Future occurrence of a disease can be highly influenced by some specific risk factors. This work presents a comprehensive approach to quantify the event probability as a function of each separate risk factor by means of a parametric model. The proposed methodology is mainly described and applied here in the case of a linear model, but the non-linear case is also addressed. To improve estimation accuracy, three distinct methods are developed and their results are integrated. One of them is Bayesian, based on a non-informative prior. Each of the other two, uses aggregation of sample elements based on their factor values, which is optimized by means of a different specific criterion. For one of these two, optimization is performed by Simulated Annealing. The methodology presented is applicable across various diseases but here we quantify the risk for cardiovascular diseases in subjects with type 1 diabetes. The results obtained combining the three different methods show accurate estimates of cardiovascular risk variation rates for the factors considered. Furthermore, the detection of a biological activation phenomenon for one of the factors is also illustrated. To quantify the performances of the proposed methodology and to compare them with those from a known method used for this type of models, a large simulation study is done, whose results are illustrated here.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"1901-1919"},"PeriodicalIF":1.9,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144111965","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-01Epub Date: 2025-08-08DOI: 10.1177/09622802251362659
Garazi Retegui, Alan E Gelfand, Jaione Etxeberria, María Dolores Ugarte
Disease mapping attempts to explain observed health event counts across areal units, typically using Markov random field models. These models rely on spatial priors to account for variation in raw relative risk or rate estimates. Spatial priors introduce some degree of smoothing, wherein, for any particular unit, empirical risk or incidence estimates are either adjusted towards a suitable mean or incorporate neighbor-based smoothing. While model explanation may be the primary focus, the literature lacks a comparison of the amount of smoothing introduced by different spatial priors. Additionally, there has been no investigation into how varying the parameters of these priors influences the resulting smoothing. This study examines seven commonly used spatial priors through both simulations and real data analyses. Using areal maps of peninsular Spain and England, we analyze smoothing effects using two datasets with associated populations at risk. We propose empirical metrics to quantify the smoothing achieved by each model and theoretical metrics to calibrate the expected extent of smoothing as a function of model parameters. We employ areal maps in order to quantitatively characterize the extent of smoothing within and across the models as well as to link the theoretical metrics to the empirical metrics.
{"title":"On prior smoothing with discrete spatial data in the context of disease mapping.","authors":"Garazi Retegui, Alan E Gelfand, Jaione Etxeberria, María Dolores Ugarte","doi":"10.1177/09622802251362659","DOIUrl":"10.1177/09622802251362659","url":null,"abstract":"<p><p>Disease mapping attempts to explain observed health event counts across areal units, typically using Markov random field models. These models rely on spatial priors to account for variation in raw relative risk or rate estimates. Spatial priors introduce some degree of smoothing, wherein, for any particular unit, empirical risk or incidence estimates are either adjusted towards a suitable mean or incorporate neighbor-based smoothing. While model explanation may be the primary focus, the literature lacks a comparison of the amount of smoothing introduced by different spatial priors. Additionally, there has been no investigation into how varying the parameters of these priors influences the resulting smoothing. This study examines seven commonly used spatial priors through both simulations and real data analyses. Using areal maps of peninsular Spain and England, we analyze smoothing effects using two datasets with associated populations at risk. We propose empirical metrics to quantify the smoothing achieved by each model and theoretical metrics to calibrate the expected extent of smoothing as a function of model parameters. We employ areal maps in order to quantitatively characterize the extent of smoothing within and across the models as well as to link the theoretical metrics to the empirical metrics.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"2091-2107"},"PeriodicalIF":1.9,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144800318","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-01Epub Date: 2025-07-17DOI: 10.1177/09622802251357021
Junhan Fang, Donna Spiegelman, Ashley L Buchanan, Laura Forastiere
Many public health interventions are conducted in settings where individuals are connected and the intervention assigned to some individuals may spill over to other individuals. In these settings, we can assess: (a) the individual effect on the treated, (b) the spillover effect on untreated individuals through an indirect exposure to the intervention, and (c) the overall effect on the whole population. Here, we consider an egocentric network-based randomized design in which a set of index participants is recruited and randomly assigned to treatment, while data are also collected on their untreated network members. Such a design is common in peer education interventions conceived to leverage behavioral influence among peers. Using the potential outcomes framework, we first clarify the assumptions required to rely on an identification strategy that is commonly used in the well-studied two-stage randomized design. Under these assumptions, causal effects can be jointly estimated using a regression model with a block-diagonal structure. We then develop sample size formulas for detecting individual, spillover, and overall effects for single and joint hypothesis tests, and investigate the role of different parameters. Finally, we illustrate the use of our sample size formulas for an egocentric network-based randomized experiment to evaluate a peer education intervention for HIV prevention.
{"title":"Design of egocentric network-based studies to estimate causal effects under interference.","authors":"Junhan Fang, Donna Spiegelman, Ashley L Buchanan, Laura Forastiere","doi":"10.1177/09622802251357021","DOIUrl":"10.1177/09622802251357021","url":null,"abstract":"<p><p>Many public health interventions are conducted in settings where individuals are connected and the intervention assigned to some individuals may spill over to other individuals. In these settings, we can assess: (a) the individual effect on the treated, (b) the spillover effect on untreated individuals through an indirect exposure to the intervention, and (c) the overall effect on the whole population. Here, we consider an egocentric network-based randomized design in which a set of index participants is recruited and randomly assigned to treatment, while data are also collected on their untreated network members. Such a design is common in peer education interventions conceived to leverage behavioral influence among peers. Using the potential outcomes framework, we first clarify the assumptions required to rely on an identification strategy that is commonly used in the well-studied two-stage randomized design. Under these assumptions, causal effects can be jointly estimated using a regression model with a block-diagonal structure. We then develop sample size formulas for detecting individual, spillover, and overall effects for single and joint hypothesis tests, and investigate the role of different parameters. Finally, we illustrate the use of our sample size formulas for an egocentric network-based randomized experiment to evaluate a peer education intervention for HIV prevention.</p>","PeriodicalId":22038,"journal":{"name":"Statistical Methods in Medical Research","volume":" ","pages":"2034-2052"},"PeriodicalIF":1.9,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144650605","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}