Relative risks have become a popular measure of treatment effect for binary outcomes in randomized controlled trials (RCTs). Relative risks can be estimated directly using log binomial regression but the model may fail to converge. Alternative methods are available for estimating relative risks but these have generally only been evaluated for independent data. As some of these methods are now being applied in cluster RCTs, investigation of their performance in this context is needed. We compare log binomial regression and three alternative methods (expanded logistic regression, log Poisson regression and log normal regression) for estimating relative risks in cluster RCTs. Clustering is taken into account using generalized estimating equations (GEEs) with an independence or exchangeable working correlation structure. The results of our large simulation study show that the log binomial GEE generally performs well for clustered data but suffers from convergence problems, as expected. Both the log Poisson GEE and log normal GEE have advantages in certain settings in terms of type I error, bias and coverage. The expanded logistic GEE can perform poorly and is sensitive to the chosen working correlation structure. Conclusions about the effectiveness of treatment often differ depending on the method used, highlighting the need to pre-specify an analysis approach. We recommend pre-specifying that either the log Poisson GEE or log normal GEE will be used in the event that the log binomial GEE fails to converge.
{"title":"Relative Risk Estimation in Cluster Randomized Trials: A Comparison of Generalized Estimating Equation Methods","authors":"L. Yelland, A. Salter, Philip Ryan","doi":"10.2202/1557-4679.1323","DOIUrl":"https://doi.org/10.2202/1557-4679.1323","url":null,"abstract":"Relative risks have become a popular measure of treatment effect for binary outcomes in randomized controlled trials (RCTs). Relative risks can be estimated directly using log binomial regression but the model may fail to converge. Alternative methods are available for estimating relative risks but these have generally only been evaluated for independent data. As some of these methods are now being applied in cluster RCTs, investigation of their performance in this context is needed. We compare log binomial regression and three alternative methods (expanded logistic regression, log Poisson regression and log normal regression) for estimating relative risks in cluster RCTs. Clustering is taken into account using generalized estimating equations (GEEs) with an independence or exchangeable working correlation structure. The results of our large simulation study show that the log binomial GEE generally performs well for clustered data but suffers from convergence problems, as expected. Both the log Poisson GEE and log normal GEE have advantages in certain settings in terms of type I error, bias and coverage. The expanded logistic GEE can perform poorly and is sensitive to the chosen working correlation structure. Conclusions about the effectiveness of treatment often differ depending on the method used, highlighting the need to pre-specify an analysis approach. We recommend pre-specifying that either the log Poisson GEE or log normal GEE will be used in the event that the log binomial GEE fails to converge.","PeriodicalId":50333,"journal":{"name":"International Journal of Biostatistics","volume":"7 1","pages":""},"PeriodicalIF":1.2,"publicationDate":"2011-05-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.2202/1557-4679.1323","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"68718384","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ruimin Xu, P. McNicholas, A. Desmond, G. Darlington
We investigate a competing risks model, using the specification of the Gompertz distribution for failure times from competing causes and the inverse Gaussian distribution for failure times from the cause of interest. The expectation-maximization algorithm is used for parameter estimation and the model is applied to real data on breast cancer and melanoma. In these applications, our models compare favourably with existing techniques. The proposed method provides a useful technique that may be more broadly applicable than existing alternatives.
{"title":"A First Passage Time Model for Long-Term Survivors with Competing Risks","authors":"Ruimin Xu, P. McNicholas, A. Desmond, G. Darlington","doi":"10.2202/1557-4679.1224","DOIUrl":"https://doi.org/10.2202/1557-4679.1224","url":null,"abstract":"We investigate a competing risks model, using the specification of the Gompertz distribution for failure times from competing causes and the inverse Gaussian distribution for failure times from the cause of interest. The expectation-maximization algorithm is used for parameter estimation and the model is applied to real data on breast cancer and melanoma. In these applications, our models compare favourably with existing techniques. The proposed method provides a useful technique that may be more broadly applicable than existing alternatives.","PeriodicalId":50333,"journal":{"name":"International Journal of Biostatistics","volume":"7 1","pages":""},"PeriodicalIF":1.2,"publicationDate":"2011-05-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.2202/1557-4679.1224","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"68717157","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This work considers the estimation of the size N of a closed population using incomplete lists of its members. Capture histories are constructed by establishing the presence or the absence of each individual in all the lists available. Models for data featuring a heterogeneous catchability and list dependencies are considered. A log-linear model leading to a lower bound for the population size is derived for a known set of list dependencies and a latent catchability variable with an arbitrary distribution. This generalizes Chao’s lower bound to models with interactions. The proposed model can be used to carry out a search for important list interactions. It also provides diagnostic information about the nature of the underlying heterogeneity. Indeed, it is shown that the Poisson maximum likelihood estimator of N under a dichotomous latent class model does not exist for a particular set of LB models. Several distributions for the heterogeneous catchability are considered; they allow to investigate the sensitivity of the population size estimate to the model for the heterogeneous catchability.
{"title":"A Lower Bound Model for Multiple Record Systems Estimation with Heterogeneous Catchability","authors":"L. Rivest","doi":"10.2202/1557-4679.1283","DOIUrl":"https://doi.org/10.2202/1557-4679.1283","url":null,"abstract":"This work considers the estimation of the size N of a closed population using incomplete lists of its members. Capture histories are constructed by establishing the presence or the absence of each individual in all the lists available. Models for data featuring a heterogeneous catchability and list dependencies are considered. A log-linear model leading to a lower bound for the population size is derived for a known set of list dependencies and a latent catchability variable with an arbitrary distribution. This generalizes Chao’s lower bound to models with interactions. The proposed model can be used to carry out a search for important list interactions. It also provides diagnostic information about the nature of the underlying heterogeneity. Indeed, it is shown that the Poisson maximum likelihood estimator of N under a dichotomous latent class model does not exist for a particular set of LB models. Several distributions for the heterogeneous catchability are considered; they allow to investigate the sensitivity of the population size estimate to the model for the heterogeneous catchability.","PeriodicalId":50333,"journal":{"name":"International Journal of Biostatistics","volume":"41 1","pages":""},"PeriodicalIF":1.2,"publicationDate":"2011-05-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.2202/1557-4679.1283","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"68717594","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This Reader’s Reaction contains some brief remarks regarding Pearl’s concerns regarding the value of principal stratification.
这个读者的反应包含了一些关于珀尔对主要分层的价值的关注的简短评论。
{"title":"Reaction to Pearl's Critique of Principal Stratification","authors":"Arvid Sjolander","doi":"10.2202/1557-4679.1324","DOIUrl":"https://doi.org/10.2202/1557-4679.1324","url":null,"abstract":"This Reader’s Reaction contains some brief remarks regarding Pearl’s concerns regarding the value of principal stratification.","PeriodicalId":50333,"journal":{"name":"International Journal of Biostatistics","volume":"7 1","pages":""},"PeriodicalIF":1.2,"publicationDate":"2011-04-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.2202/1557-4679.1324","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"68718456","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Variable selection for high dimensional data has recently received a great deal of attention. However, due to the complex structure of the likelihood, only limited developments have been made for time-to-event data where censoring is present. In this paper, we propose a Bayesian variable selection scheme for a Bayesian semiparametric survival model for right censored survival data sets. A special shrinkage prior on the coefficients corresponding to the predictor variables is used to handle cases when the explanatory variables are of very high-dimension. The shrinkage prior is obtained through a scale mixture representation of Normal and Gamma distributions. Our proposed variable selection prior corresponds to the well known lasso penalty. The likelihood function is based on the Cox proportional hazards model framework, where the cumulative baseline hazard function is modeled a priori by a gamma process. We assign a prior on the tuning parameter of the shrinkage prior and adaptively control the sparsity of our model. The primary use of the proposed model is to identify the important covariates relating to the survival curves. To implement our methodology, we have developed a fast Markov chain Monte Carlo algorithm with an adaptive jumping rule. We have successfully applied our method on simulated data sets under two different settings and real microarray data sets which contain right censored survival time. The performance of our Bayesian variable selection model compared with other competing methods is also provided to demonstrate the superiority of our method. A short description of the biological relevance of the selected genes in the real data sets is provided, further strengthening our claims.
{"title":"Bayesian Variable Selection in Semiparametric Proportional Hazards Model for High Dimensional Survival Data","authors":"Kyu Ha Lee, S. Chakraborty, Jianguo Sun","doi":"10.2202/1557-4679.1301","DOIUrl":"https://doi.org/10.2202/1557-4679.1301","url":null,"abstract":"Variable selection for high dimensional data has recently received a great deal of attention. However, due to the complex structure of the likelihood, only limited developments have been made for time-to-event data where censoring is present. In this paper, we propose a Bayesian variable selection scheme for a Bayesian semiparametric survival model for right censored survival data sets. A special shrinkage prior on the coefficients corresponding to the predictor variables is used to handle cases when the explanatory variables are of very high-dimension. The shrinkage prior is obtained through a scale mixture representation of Normal and Gamma distributions. Our proposed variable selection prior corresponds to the well known lasso penalty. The likelihood function is based on the Cox proportional hazards model framework, where the cumulative baseline hazard function is modeled a priori by a gamma process. We assign a prior on the tuning parameter of the shrinkage prior and adaptively control the sparsity of our model. The primary use of the proposed model is to identify the important covariates relating to the survival curves. To implement our methodology, we have developed a fast Markov chain Monte Carlo algorithm with an adaptive jumping rule. We have successfully applied our method on simulated data sets under two different settings and real microarray data sets which contain right censored survival time. The performance of our Bayesian variable selection model compared with other competing methods is also provided to demonstrate the superiority of our method. A short description of the biological relevance of the selected genes in the real data sets is provided, further strengthening our claims.","PeriodicalId":50333,"journal":{"name":"International Journal of Biostatistics","volume":"7 1","pages":""},"PeriodicalIF":1.2,"publicationDate":"2011-04-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.2202/1557-4679.1301","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"68717639","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A meta-analysis that uses individual-level data instead of study-level data is widely considered to be a gold standard approach, in part because it allows a time-to-event analysis. Unfortunately, with the common practice of presenting Kaplan-Meier survival curves after pooling subjects across randomized trials, using individual-level data can actually be a step backwards; a Simpson's paradox can occur in which pooling incorrectly reverses the direction of an association. We introduce a nonparametric procedure for synthesizing survival curves across studies that is designed to avoid this difficulty and preserve the integrity of randomization. The technique is based on a counterfactual formulation in which we ask what pooled survival curves would look like if all subjects in all studies had been assigned treatment, or if all subjects had been assigned to control arms. The method is related to a Kaplan-Meier adjustment proposed in 2005 by Xie and Liu to correct for confounding in nonrandomized studies, but is formulated for the meta-analysis setting. The procedure is discussed in the context of examining rosiglitazone and cardiovascular adverse events.
{"title":"An Alternative to Pooling Kaplan-Meier Curves in Time-to-Event Meta-Analysis","authors":"D. Rubin","doi":"10.2202/1557-4679.1289","DOIUrl":"https://doi.org/10.2202/1557-4679.1289","url":null,"abstract":"A meta-analysis that uses individual-level data instead of study-level data is widely considered to be a gold standard approach, in part because it allows a time-to-event analysis. Unfortunately, with the common practice of presenting Kaplan-Meier survival curves after pooling subjects across randomized trials, using individual-level data can actually be a step backwards; a Simpson's paradox can occur in which pooling incorrectly reverses the direction of an association. We introduce a nonparametric procedure for synthesizing survival curves across studies that is designed to avoid this difficulty and preserve the integrity of randomization. The technique is based on a counterfactual formulation in which we ask what pooled survival curves would look like if all subjects in all studies had been assigned treatment, or if all subjects had been assigned to control arms. The method is related to a Kaplan-Meier adjustment proposed in 2005 by Xie and Liu to correct for confounding in nonrandomized studies, but is formulated for the meta-analysis setting. The procedure is discussed in the context of examining rosiglitazone and cardiovascular adverse events.","PeriodicalId":50333,"journal":{"name":"International Journal of Biostatistics","volume":"7 1","pages":""},"PeriodicalIF":1.2,"publicationDate":"2011-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.2202/1557-4679.1289","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"68717777","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
When cost data are collected in a clinical study, interest centers on the between-treatment difference in mean cost. When censoring is present, the resulting loss of information can be limited by collecting cost data for several pre-specified time intervals, leading to censored longitudinal cost data. Most models for marginal costs stratify by time interval. However, in few other areas of biostatistics would we stratify by default. We argue that there are benefits to considering more general models: for example, in some settings, pooling regression coefficients across intervals can improve the precision of the estimated between-treatment difference in mean cost. Previous work has used inverse-probability-weighted GEEs coupled with an independent working variance to estimate parameters from these more general models. We show that the greatest precision benefits of non-stratified models are achieved by using more sophisticated working variance matrices.
{"title":"Marginal Models for Censored Longitudinal Cost Data: Appropriate Working Variance Matrices in Inverse-Probability-Weighted GEEs Can Improve Precision","authors":"E. Pullenayegum, A. Willan","doi":"10.2202/1557-4679.1170","DOIUrl":"https://doi.org/10.2202/1557-4679.1170","url":null,"abstract":"When cost data are collected in a clinical study, interest centers on the between-treatment difference in mean cost. When censoring is present, the resulting loss of information can be limited by collecting cost data for several pre-specified time intervals, leading to censored longitudinal cost data. Most models for marginal costs stratify by time interval. However, in few other areas of biostatistics would we stratify by default. We argue that there are benefits to considering more general models: for example, in some settings, pooling regression coefficients across intervals can improve the precision of the estimated between-treatment difference in mean cost. Previous work has used inverse-probability-weighted GEEs coupled with an independent working variance to estimate parameters from these more general models. We show that the greatest precision benefits of non-stratified models are achieved by using more sophisticated working variance matrices.","PeriodicalId":50333,"journal":{"name":"International Journal of Biostatistics","volume":"7 1","pages":""},"PeriodicalIF":1.2,"publicationDate":"2011-02-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.2202/1557-4679.1170","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"68716427","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In disease classification, a traditional technique is the receiver operative characteristic (ROC) curve and the area under the curve (AUC). With high-dimensional data, the ROC techniques are needed to conduct classification and variable selection. The current ROC methods do not explicitly incorporate unequal misclassification costs or do not have a theoretical grounding for optimizing the AUC. Empirical studies in the literature have demonstrated that optimizing the hinge loss can maximize the AUC approximately. In theory, minimizing the hinge rank loss is equivalent to minimizing the AUC in the asymptotic limit. In this article, we propose a novel nonparametric method HingeBoost to optimize a weighted hinge loss incorporating misclassification costs. HingeBoost can be used to construct linear and nonlinear classifiers. The estimation and variable selection for the hinge loss are addressed by a new boosting algorithm. Furthermore, the proposed twin HingeBoost can select more sparse predictors. Some properties of HingeBoost are studied as well. To compare HingeBoost with existing classification methods, we present empirical study results using data from simulations and a prostate cancer study with mass spectrometry-based proteomics.
{"title":"HingeBoost: ROC-Based Boost for Classification and Variable Selection","authors":"Zhuo Wang","doi":"10.2202/1557-4679.1304","DOIUrl":"https://doi.org/10.2202/1557-4679.1304","url":null,"abstract":"In disease classification, a traditional technique is the receiver operative characteristic (ROC) curve and the area under the curve (AUC). With high-dimensional data, the ROC techniques are needed to conduct classification and variable selection. The current ROC methods do not explicitly incorporate unequal misclassification costs or do not have a theoretical grounding for optimizing the AUC. Empirical studies in the literature have demonstrated that optimizing the hinge loss can maximize the AUC approximately. In theory, minimizing the hinge rank loss is equivalent to minimizing the AUC in the asymptotic limit. In this article, we propose a novel nonparametric method HingeBoost to optimize a weighted hinge loss incorporating misclassification costs. HingeBoost can be used to construct linear and nonlinear classifiers. The estimation and variable selection for the hinge loss are addressed by a new boosting algorithm. Furthermore, the proposed twin HingeBoost can select more sparse predictors. Some properties of HingeBoost are studied as well. To compare HingeBoost with existing classification methods, we present empirical study results using data from simulations and a prostate cancer study with mass spectrometry-based proteomics.","PeriodicalId":50333,"journal":{"name":"International Journal of Biostatistics","volume":"7 1","pages":""},"PeriodicalIF":1.2,"publicationDate":"2011-02-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.2202/1557-4679.1304","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"68717746","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
We undertake here a comprehensive simulation study of the theoretical properties that we derive in a companion article devoted to the asymptotic study of adaptive group sequential designs in the case of randomized clinical trials (RCTs) with binary treatment, binary outcome and no covariate. By adaptive design, we mean in this setting a RCT design that allows the investigator to dynamically modify its course through data-driven adjustment of the randomization probability based on data accrued so far without negatively impacting on the statistical integrity of the trial. By adaptive group sequential design, we refer to the fact that group sequential testing methods can be equally well applied on top of adaptive designs. The simulation study validates the theory. It notably shows in the estimation framework that the confidence intervals we obtain achieve the desired coverage even for moderate sample sizes. In addition, it shows in the testing framework that type I error control at the prescribed level is guaranteed and that all sampling procedures only suffer from a very slight increase of the type II error. A three-sentence take-home message is “Adaptive designs do learn the targeted optimal design and inference and testing can be carried out under adaptive sampling as they would under the targeted optimal randomization probability iid sampling. In particular, adaptive designs achieve the same efficiency as the fixed oracle design. This is confirmed by a simulation study, at least for moderate or large sample sizes, across a large collection of targeted randomization probabilities.”
{"title":"Targeting the Optimal Design in Randomized Clinical Trials with Binary Outcomes and No Covariate: Simulation Study","authors":"A. Chambaz, M. J. van der Laan","doi":"10.2202/1557-4679.1310","DOIUrl":"https://doi.org/10.2202/1557-4679.1310","url":null,"abstract":"We undertake here a comprehensive simulation study of the theoretical properties that we derive in a companion article devoted to the asymptotic study of adaptive group sequential designs in the case of randomized clinical trials (RCTs) with binary treatment, binary outcome and no covariate. By adaptive design, we mean in this setting a RCT design that allows the investigator to dynamically modify its course through data-driven adjustment of the randomization probability based on data accrued so far without negatively impacting on the statistical integrity of the trial. By adaptive group sequential design, we refer to the fact that group sequential testing methods can be equally well applied on top of adaptive designs. The simulation study validates the theory. It notably shows in the estimation framework that the confidence intervals we obtain achieve the desired coverage even for moderate sample sizes. In addition, it shows in the testing framework that type I error control at the prescribed level is guaranteed and that all sampling procedures only suffer from a very slight increase of the type II error. A three-sentence take-home message is “Adaptive designs do learn the targeted optimal design and inference and testing can be carried out under adaptive sampling as they would under the targeted optimal randomization probability iid sampling. In particular, adaptive designs achieve the same efficiency as the fixed oracle design. This is confirmed by a simulation study, at least for moderate or large sample sizes, across a large collection of targeted randomization probabilities.”","PeriodicalId":50333,"journal":{"name":"International Journal of Biostatistics","volume":"7 1","pages":""},"PeriodicalIF":1.2,"publicationDate":"2011-01-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.2202/1557-4679.1310","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"68718288","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The relative risk is a clinically important measure of the effect of treatment on binary outcomes in randomized controlled trials (RCTs). An adjusted relative risk can be estimated using log binomial regression; however, convergence problems are common with this model. While alternative methods have been proposed for estimating relative risks, comparisons between methods have been limited, particularly in the context of RCTs. We compare ten different methods for estimating relative risks under a variety of scenarios relevant to RCTs with independent observations. Results of a large simulation study show that some methods may fail to overcome the convergence problems of log binomial regression, while others may substantially overestimate the treatment effect or produce inaccurate confidence intervals. Further, conclusions about the effectiveness of treatment may differ depending on the method used. We give recommendations for choosing a method for estimating relative risks in the context of RCTs with independent observations.
{"title":"Relative Risk Estimation in Randomized Controlled Trials: A Comparison of Methods for Independent Observations","authors":"L. Yelland, A. Salter, Philip Ryan","doi":"10.2202/1557-4679.1278","DOIUrl":"https://doi.org/10.2202/1557-4679.1278","url":null,"abstract":"The relative risk is a clinically important measure of the effect of treatment on binary outcomes in randomized controlled trials (RCTs). An adjusted relative risk can be estimated using log binomial regression; however, convergence problems are common with this model. While alternative methods have been proposed for estimating relative risks, comparisons between methods have been limited, particularly in the context of RCTs. We compare ten different methods for estimating relative risks under a variety of scenarios relevant to RCTs with independent observations. Results of a large simulation study show that some methods may fail to overcome the convergence problems of log binomial regression, while others may substantially overestimate the treatment effect or produce inaccurate confidence intervals. Further, conclusions about the effectiveness of treatment may differ depending on the method used. We give recommendations for choosing a method for estimating relative risks in the context of RCTs with independent observations.","PeriodicalId":50333,"journal":{"name":"International Journal of Biostatistics","volume":"7 1","pages":""},"PeriodicalIF":1.2,"publicationDate":"2011-01-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.2202/1557-4679.1278","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"68717935","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}