{"title":"A randomization-based theory for preliminary testing of covariate balance in controlled trials","authors":"Anqi Zhao, Peng Ding","doi":"10.1080/19466315.2023.2267774","DOIUrl":null,"url":null,"abstract":"AbstractRandomized trials balance all covariates on average and are the gold standard for estimating treatment effects. Chance imbalances nevertheless exist more or less in realized treatment allocations and intrigue an important question: what should we do if the treatment groups differ with respect to some important baseline characteristics? A common strategy is to conduct a preliminary test of the balance of baseline covariates after randomization, and invoke covariate adjustment for subsequent inference if and only if the realized allocation fails some prespecified criterion. Although such practice is intuitive and popular among practitioners, the existing literature has so far only evaluated its properties under strong parametric model assumptions in theory and simulation, yielding results of limited generality. To fill this gap, we examine two strategies for conducting preliminary test-based covariate adjustment by regression, and evaluate the validity and efficiency of the resulting inferences from the randomization-based perspective. The main result is twofold. First, the preliminary-test estimator based on the analysis of covariance can be even less efficient than the unadjusted difference in means, and risks anticonservative confidence intervals based on normal approximation even with the robust standard error. Second, the preliminary-test estimator based on the fully interacted specification is less efficient than its counterpart under the always-adjust strategy, and yields overconservative confidence intervals based on normal approximation. In addition, although the Fisher randomization test is still finite-sample exact for testing the sharp null hypothesis of no treatment effect on any individual, it is no longer valid for testing the weak null hypothesis of zero average treatment effect in large samples even with properly studentized test statistics. These undesirable properties are due to the asymptotic non-normality of the preliminary-test estimators. Based on theory and simulation, we echo the existing literature and do not recommend the preliminary-test procedure for covariate adjustment in randomized trials.Keywords: Causal inferencedesign-based inferenceefficiencyFisher randomization testregression adjustmentrerandomizationDisclaimerAs a service to authors and researchers we are providing this version of an accepted manuscript (AM). Copyediting, typesetting, and review of the resulting proofs will be undertaken on this manuscript before final publication of the Version of Record (VoR). During production and pre-press, errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal relate to these versions also. FundingThe author(s) reported there is no funding associated with the work featured in this article.","PeriodicalId":51280,"journal":{"name":"Statistics in Biopharmaceutical Research","volume":"46 1","pages":"0"},"PeriodicalIF":1.5000,"publicationDate":"2023-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Statistics in Biopharmaceutical Research","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1080/19466315.2023.2267774","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"MATHEMATICAL & COMPUTATIONAL BIOLOGY","Score":null,"Total":0}
引用次数: 0
Abstract
AbstractRandomized trials balance all covariates on average and are the gold standard for estimating treatment effects. Chance imbalances nevertheless exist more or less in realized treatment allocations and intrigue an important question: what should we do if the treatment groups differ with respect to some important baseline characteristics? A common strategy is to conduct a preliminary test of the balance of baseline covariates after randomization, and invoke covariate adjustment for subsequent inference if and only if the realized allocation fails some prespecified criterion. Although such practice is intuitive and popular among practitioners, the existing literature has so far only evaluated its properties under strong parametric model assumptions in theory and simulation, yielding results of limited generality. To fill this gap, we examine two strategies for conducting preliminary test-based covariate adjustment by regression, and evaluate the validity and efficiency of the resulting inferences from the randomization-based perspective. The main result is twofold. First, the preliminary-test estimator based on the analysis of covariance can be even less efficient than the unadjusted difference in means, and risks anticonservative confidence intervals based on normal approximation even with the robust standard error. Second, the preliminary-test estimator based on the fully interacted specification is less efficient than its counterpart under the always-adjust strategy, and yields overconservative confidence intervals based on normal approximation. In addition, although the Fisher randomization test is still finite-sample exact for testing the sharp null hypothesis of no treatment effect on any individual, it is no longer valid for testing the weak null hypothesis of zero average treatment effect in large samples even with properly studentized test statistics. These undesirable properties are due to the asymptotic non-normality of the preliminary-test estimators. Based on theory and simulation, we echo the existing literature and do not recommend the preliminary-test procedure for covariate adjustment in randomized trials.Keywords: Causal inferencedesign-based inferenceefficiencyFisher randomization testregression adjustmentrerandomizationDisclaimerAs a service to authors and researchers we are providing this version of an accepted manuscript (AM). Copyediting, typesetting, and review of the resulting proofs will be undertaken on this manuscript before final publication of the Version of Record (VoR). During production and pre-press, errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal relate to these versions also. FundingThe author(s) reported there is no funding associated with the work featured in this article.
期刊介绍:
Statistics in Biopharmaceutical Research ( SBR), publishes articles that focus on the needs of researchers and applied statisticians in biopharmaceutical industries; academic biostatisticians from schools of medicine, veterinary medicine, public health, and pharmacy; statisticians and quantitative analysts working in regulatory agencies (e.g., U.S. Food and Drug Administration and its counterpart in other countries); statisticians with an interest in adopting methodology presented in this journal to their own fields; and nonstatisticians with an interest in applying statistical methods to biopharmaceutical problems.
Statistics in Biopharmaceutical Research accepts papers that discuss appropriate statistical methodology and information regarding the use of statistics in all phases of research, development, and practice in the pharmaceutical, biopharmaceutical, device, and diagnostics industries. Articles should focus on the development of novel statistical methods, novel applications of current methods, or the innovative application of statistical principles that can be used by statistical practitioners in these disciplines. Areas of application may include statistical methods for drug discovery, including papers that address issues of multiplicity, sequential trials, adaptive designs, etc.; preclinical and clinical studies; genomics and proteomics; bioassay; biomarkers and surrogate markers; models and analyses of drug history, including pharmacoeconomics, product life cycle, detection of adverse events in clinical studies, and postmarketing risk assessment; regulatory guidelines, including issues of standardization of terminology (e.g., CDISC), tolerance and specification limits related to pharmaceutical practice, and novel methods of drug approval; and detection of adverse events in clinical and toxicological studies. Tutorial articles also are welcome. Articles should include demonstrable evidence of the usefulness of this methodology (presumably by means of an application).
The Editorial Board of SBR intends to ensure that the journal continually provides important, useful, and timely information. To accomplish this, the board strives to attract outstanding articles by seeing that each submission receives a careful, thorough, and prompt review.
Authors can choose to publish gold open access in this journal.