Bayesian meta-analysis is a promising approach for rare events meta-analysis. However, the inference of the overall effect in rare events meta-analysis is sensitive to the choice of prior distribution for the heterogeneity parameter. Therefore, it is crucial to assign a convincing prior specification and ensure that it is both plausible and transparent. Various priors for the heterogeneity parameter have been proposed; however, the comparative performance of alternative prior specifications in rare events meta-analysis is poorly understood. Based on a binomial-normal hierarchical model, we conducted a comprehensive simulation study to compare seven heterogeneity prior specifications for binary outcomes, using the odds ratio as the metric. We compared their performance in terms of coverage, median percentage bias, width of the 95% credible interval, and root mean square error (RMSE). We illustrate the results with two recently published rare events meta-analyses of a few studies. The results show that the half-normal prior (with a scale of 0.5), the prior proposed by Turner et al. for the general healthcare setting (without restriction to a specific type of outcome) and for the adverse event setting perform well when the degree of heterogeneity is not relatively high, yielding smaller bias and shorter interval widths with similar coverage and RMSE in most cases compared to other prior specifications. None of the priors performed better when the heterogeneity between-studies were significantly extreme.
{"title":"Comparison of Prior Distributions for the Heterogeneity Parameter in a Rare Events Meta-Analysis of a Few Studies.","authors":"Minghong Yao, Fan Mei, Kang Zou, Ling Li, Xin Sun","doi":"10.1002/pst.2448","DOIUrl":"https://doi.org/10.1002/pst.2448","url":null,"abstract":"<p><p>Bayesian meta-analysis is a promising approach for rare events meta-analysis. However, the inference of the overall effect in rare events meta-analysis is sensitive to the choice of prior distribution for the heterogeneity parameter. Therefore, it is crucial to assign a convincing prior specification and ensure that it is both plausible and transparent. Various priors for the heterogeneity parameter have been proposed; however, the comparative performance of alternative prior specifications in rare events meta-analysis is poorly understood. Based on a binomial-normal hierarchical model, we conducted a comprehensive simulation study to compare seven heterogeneity prior specifications for binary outcomes, using the odds ratio as the metric. We compared their performance in terms of coverage, median percentage bias, width of the 95% credible interval, and root mean square error (RMSE). We illustrate the results with two recently published rare events meta-analyses of a few studies. The results show that the half-normal prior (with a scale of 0.5), the prior proposed by Turner et al. for the general healthcare setting (without restriction to a specific type of outcome) and for the adverse event setting perform well when the degree of heterogeneity is not relatively high, yielding smaller bias and shorter interval widths with similar coverage and RMSE in most cases compared to other prior specifications. None of the priors performed better when the heterogeneity between-studies were significantly extreme.</p>","PeriodicalId":19934,"journal":{"name":"Pharmaceutical Statistics","volume":" ","pages":""},"PeriodicalIF":1.3,"publicationDate":"2024-10-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142505736","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In covariate-adaptive or response-adaptive randomization, the treatment assignment and outcome can be correlated. Under this situation, the re-randomization test is a straightforward and attractive method to provide valid statistical inferences. In this paper, we investigate the number of repetitions in tests. This is motivated by a group sequential design in clinical trials, where the nominal significance bound can be very small at an interim analysis. Accordingly, re-randomization tests lead to a very large number of required repetitions, which may be computationally intractable. To reduce the number of repetitions, we propose an adaptive procedure and compare it with multiple approaches under predefined criteria. Monte Carlo simulations are conducted to show the performance of different approaches in a limited sample size. We also suggest strategies to reduce total computation time and provide practical guidance in preparing, executing, and reporting before and after data are unblinded at an interim analysis, so one can complete the computation within a reasonable time frame.
{"title":"Number of Repetitions in Re-Randomization Tests.","authors":"Yilong Zhang, Yujie Zhao, Bingjun Wang, Yiwen Luo","doi":"10.1002/pst.2438","DOIUrl":"https://doi.org/10.1002/pst.2438","url":null,"abstract":"<p><p>In covariate-adaptive or response-adaptive randomization, the treatment assignment and outcome can be correlated. Under this situation, the re-randomization test is a straightforward and attractive method to provide valid statistical inferences. In this paper, we investigate the number of repetitions in tests. This is motivated by a group sequential design in clinical trials, where the nominal significance bound can be very small at an interim analysis. Accordingly, re-randomization tests lead to a very large number of required repetitions, which may be computationally intractable. To reduce the number of repetitions, we propose an adaptive procedure and compare it with multiple approaches under predefined criteria. Monte Carlo simulations are conducted to show the performance of different approaches in a limited sample size. We also suggest strategies to reduce total computation time and provide practical guidance in preparing, executing, and reporting before and after data are unblinded at an interim analysis, so one can complete the computation within a reasonable time frame.</p>","PeriodicalId":19934,"journal":{"name":"Pharmaceutical Statistics","volume":" ","pages":""},"PeriodicalIF":1.3,"publicationDate":"2024-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142472207","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Several indices were suggested to determine the follow up duration in oncology trials from either maturity or stability perspective, by maximizing time such that the index was either greater or less than a pre-defined cutoff value. However, the selection of cutoff value was subjective and usually no commonly agreed cutoff value existed; sometimes one had to resort to simulations. To solve this problem, a new balance index was proposed, which integrated both data stability and data maturity. Its theoretical properties and relationships with other indices were investigated; then its performance was demonstrated through a case study. The highlights of the index are: (1) easy to calculate; (2) free of cutoff value selection; (3) generally consistent with the other indices while sometimes able to shorten the follow-up duration thus more flexible. For the cases where the new balance index cannot be calculated, a modified balance index was also proposed and discussed. For either single arm trial or randomized clinical trial, the two new balance indices can be implemented to widespread situations such as designing a new trial from scratch, or using aggregated trial information to inform the decision-making in the middle of trial conduct.
有人提出了几种指数,从成熟或稳定的角度来确定肿瘤试验的随访时间,方法是最大限度地延长时间 t $$ t $$,使指数大于或小于预先确定的临界值。然而,临界值的选择是主观的,通常不存在共同认可的临界值,有时不得不求助于模拟。为了解决这个问题,我们提出了一个新的平衡指数,它综合了数据稳定性和数据成熟度。我们研究了该指数的理论属性以及与其他指数的关系,然后通过案例研究证明了该指数的性能。该指数的亮点在于(1) 计算简便;(2) 无需选择临界值;(3) 与其他指数基本一致,有时还能缩短跟踪时间,因此更加灵活。对于无法计算新平衡指数的情况,还提出并讨论了修正平衡指数。对于单臂试验或随机临床试验,这两种新的平衡指数可广泛应用于各种情况,如从零开始设计新的试验,或在试验进行过程中利用汇总的试验信息为决策提供参考。
{"title":"Balance Index to Determine the Follow-Up Duration of Oncology Trials.","authors":"Lei Yang, Feinan Lu","doi":"10.1002/pst.2442","DOIUrl":"https://doi.org/10.1002/pst.2442","url":null,"abstract":"<p><p>Several indices were suggested to determine the follow up duration in oncology trials from either maturity or stability perspective, by maximizing time <math> <semantics><mrow><mi>t</mi></mrow> <annotation>$$ t $$</annotation></semantics> </math> such that the index was either greater or less than a pre-defined cutoff value. However, the selection of cutoff value was subjective and usually no commonly agreed cutoff value existed; sometimes one had to resort to simulations. To solve this problem, a new balance index was proposed, which integrated both data stability and data maturity. Its theoretical properties and relationships with other indices were investigated; then its performance was demonstrated through a case study. The highlights of the index are: (1) easy to calculate; (2) free of cutoff value selection; (3) generally consistent with the other indices while sometimes able to shorten the follow-up duration thus more flexible. For the cases where the new balance index cannot be calculated, a modified balance index was also proposed and discussed. For either single arm trial or randomized clinical trial, the two new balance indices can be implemented to widespread situations such as designing a new trial from scratch, or using aggregated trial information to inform the decision-making in the middle of trial conduct.</p>","PeriodicalId":19934,"journal":{"name":"Pharmaceutical Statistics","volume":" ","pages":""},"PeriodicalIF":1.3,"publicationDate":"2024-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142400956","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A three-arm comparative clinical endpoint bioequivalence (BE) study is often used to establish bioequivalence (BE) between a locally acting generic drug (T) and reference drug (R), where superiority needs to be established for T and R over Placebo (P) and equivalence needs to be established for T vs. R. Sometimes, when study design parameters are uncertain, a fixed design study may be under- or over-powered and result in study failure or unnecessary cost. In this paper, we propose a two-stage adaptive clinical endpoint BE study with unblinded sample size re-estimation, standard or maximum combination method, optimized allocation ratio, optional re-estimation of the effect size based on likelihood estimation, and optional re-estimation of the R and P treatment means at interim analysis, which have not been done previously. Our proposed method guarantees control of Type 1 error rate analytically. It helps to reduce the average sample size when the original fixed design is overpowered and increases the sample size and power when the original study and group sequential design are under-powered. Our proposed adaptive design can help generic drug sponsors cut cost and improve success rate, making clinical study endpoint BE studies more affordable and more generic drugs accessible to the public.
三臂比较临床终点生物等效性(BE)研究通常用于确定局部作用仿制药(T)和参比药(R)之间的生物等效性(BE),其中需要确定T和R相对于安慰剂(P)的优越性,以及T相对于R的等效性。有时,当研究设计参数不确定时,固定设计的研究可能功率不足或过高,导致研究失败或不必要的成本。在本文中,我们提出了一种两阶段自适应临床终点 BE 研究,其中包括非盲法样本量重新估计、标准或最大组合法、优化分配比例、基于似然估计的可选效应大小重新估计、中期分析时可选的 R 和 P 治疗均值重新估计,这些都是以前没有做过的。我们提出的方法保证了对第一类错误率的分析控制。当原来的固定设计功率过大时,它有助于减少平均样本量;当原来的研究和分组序列设计功率不足时,它有助于增加样本量和功率。我们提出的自适应设计可以帮助仿制药申办者降低成本,提高成功率,使临床研究终点 BE 研究更加经济实惠,让更多的公众可以获得仿制药。
{"title":"An Adaptive Three-Arm Comparative Clinical Endpoint Bioequivalence Study Design With Unblinded Sample Size Re-Estimation and Optimized Allocation Ratio.","authors":"David Hinds, Wanjie Sun","doi":"10.1002/pst.2439","DOIUrl":"https://doi.org/10.1002/pst.2439","url":null,"abstract":"<p><p>A three-arm comparative clinical endpoint bioequivalence (BE) study is often used to establish bioequivalence (BE) between a locally acting generic drug (T) and reference drug (R), where superiority needs to be established for T and R over Placebo (P) and equivalence needs to be established for T vs. R. Sometimes, when study design parameters are uncertain, a fixed design study may be under- or over-powered and result in study failure or unnecessary cost. In this paper, we propose a two-stage adaptive clinical endpoint BE study with unblinded sample size re-estimation, standard or maximum combination method, optimized allocation ratio, optional re-estimation of the effect size based on likelihood estimation, and optional re-estimation of the R and P treatment means at interim analysis, which have not been done previously. Our proposed method guarantees control of Type 1 error rate analytically. It helps to reduce the average sample size when the original fixed design is overpowered and increases the sample size and power when the original study and group sequential design are under-powered. Our proposed adaptive design can help generic drug sponsors cut cost and improve success rate, making clinical study endpoint BE studies more affordable and more generic drugs accessible to the public.</p>","PeriodicalId":19934,"journal":{"name":"Pharmaceutical Statistics","volume":" ","pages":""},"PeriodicalIF":1.3,"publicationDate":"2024-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142392402","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Fenny Ong, Geert Molenberghs, Andrea Callegaro, Wim Van der Elst, Florian Stijven, Geert Verbeke, Ingrid Van Keilegom, Ariel Alonso
In a causal inference framework, a new metric has been proposed to quantify surrogacy for a continuous putative surrogate and a binary true endpoint, based on information theory. The proposed metric, termed the individual causal association (ICA), was quantified using a joint causal inference model for the corresponding potential outcomes. Due to the non-identifiability inherent in this type of models, a sensitivity analysis was introduced to study the behavior of the ICA as a function of the non-identifiable parameters characterizing the aforementioned model. In this scenario, to reduce uncertainty, several plausible yet untestable assumptions like monotonicity, independence, conditional independence or homogeneous variance-covariance, are often incorporated into the analysis. We assess the robustness of the methodology regarding these simplifying assumptions via simulation. The practical implications of the findings are demonstrated in the analysis of a randomized clinical trial evaluating an inactivated quadrivalent influenza vaccine.
{"title":"Assessing the Operational Characteristics of the Individual Causal Association as a Metric of Surrogacy in the Binary Continuous Setting.","authors":"Fenny Ong, Geert Molenberghs, Andrea Callegaro, Wim Van der Elst, Florian Stijven, Geert Verbeke, Ingrid Van Keilegom, Ariel Alonso","doi":"10.1002/pst.2437","DOIUrl":"https://doi.org/10.1002/pst.2437","url":null,"abstract":"<p><p>In a causal inference framework, a new metric has been proposed to quantify surrogacy for a continuous putative surrogate and a binary true endpoint, based on information theory. The proposed metric, termed the individual causal association (ICA), was quantified using a joint causal inference model for the corresponding potential outcomes. Due to the non-identifiability inherent in this type of models, a sensitivity analysis was introduced to study the behavior of the ICA as a function of the non-identifiable parameters characterizing the aforementioned model. In this scenario, to reduce uncertainty, several plausible yet untestable assumptions like monotonicity, independence, conditional independence or homogeneous variance-covariance, are often incorporated into the analysis. We assess the robustness of the methodology regarding these simplifying assumptions via simulation. The practical implications of the findings are demonstrated in the analysis of a randomized clinical trial evaluating an inactivated quadrivalent influenza vaccine.</p>","PeriodicalId":19934,"journal":{"name":"Pharmaceutical Statistics","volume":" ","pages":""},"PeriodicalIF":1.3,"publicationDate":"2024-09-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142351785","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Jonathan Jones, Bairu Zhang, Xiang Zhang, Peter Konings, Pia Hansson, Anna Backmark, Alessia Serrano, Ulrike Künzel, Steven Novick
Quality by Design (QbD) is an approach to assay development to determine the design space, which is the range of assay variable settings that should result in satisfactory assay quality. Typically, QbD is applied in manufacturing, but it works just as well in the preclinical space. Through three examples, we illustrate the QbD approach with experimental design and associated data analysis to determine the design space for preclinical assays.
{"title":"Quality by Design for Preclinical In Vitro Assay Development.","authors":"Jonathan Jones, Bairu Zhang, Xiang Zhang, Peter Konings, Pia Hansson, Anna Backmark, Alessia Serrano, Ulrike Künzel, Steven Novick","doi":"10.1002/pst.2430","DOIUrl":"https://doi.org/10.1002/pst.2430","url":null,"abstract":"<p><p>Quality by Design (QbD) is an approach to assay development to determine the design space, which is the range of assay variable settings that should result in satisfactory assay quality. Typically, QbD is applied in manufacturing, but it works just as well in the preclinical space. Through three examples, we illustrate the QbD approach with experimental design and associated data analysis to determine the design space for preclinical assays.</p>","PeriodicalId":19934,"journal":{"name":"Pharmaceutical Statistics","volume":" ","pages":""},"PeriodicalIF":1.3,"publicationDate":"2024-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142351786","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
I would like to reconsider a recent analysis by Prof. Senn on the statistics of the Pfizer‐BioNTech vaccine trial, to express some different opinions and to clarify some theoretical points, especially regarding the clinical applications of Bayesian statistics.
{"title":"On Some Modeling Issues in Estimating Vaccine Efficacy","authors":"Mauro Gasparini","doi":"10.1002/pst.2440","DOIUrl":"https://doi.org/10.1002/pst.2440","url":null,"abstract":"I would like to reconsider a recent analysis by Prof. Senn on the statistics of the Pfizer‐BioNTech vaccine trial, to express some different opinions and to clarify some theoretical points, especially regarding the clinical applications of Bayesian statistics.","PeriodicalId":19934,"journal":{"name":"Pharmaceutical Statistics","volume":"6 1","pages":""},"PeriodicalIF":1.5,"publicationDate":"2024-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142214992","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A common feature in cohort studies is when there is a baseline measurement of the continuous follow-up or outcome variable. Common examples include baseline measurements of physiological characteristics such as blood pressure or heart rate in studies where the outcome is post-baseline measurement of the same variable. Methods incorporating the propensity score are increasingly being used to estimate the effects of treatments using observational studies. We examined six methods for incorporating the baseline value of the follow-up variable when using propensity score matching or weighting. These methods differed according to whether the baseline value of the follow-up variable was included or excluded from the propensity score model, whether subsequent regression adjustment was conducted in the matched or weighted sample to adjust for the baseline value of the follow-up variable, and whether the analysis estimated the effect of treatment on the follow-up variable or on the change from baseline. We used Monte Carlo simulations with 750 scenarios. While no analytic method had uniformly superior performance, we provide the following recommendations: first, when using weighting and the ATE is the target estimand, use an augmented inverse probability weighted estimator or include the baseline value of the follow-up variable in the propensity score model and subsequently adjust for the baseline value of the follow-up variable in a regression model. Second, when the ATT is the target estimand, regardless of whether using weighting or matching, analyze change from baseline using a propensity score that excludes the baseline value of the follow-up variable.
队列研究的一个共同特点是对连续随访变量或结果变量进行基线测量。常见的例子包括在研究中对血压或心率等生理特征进行基线测量,而结果则是对同一变量进行基线后测量。纳入倾向得分的方法越来越多地被用于利用观察性研究来估计治疗效果。我们研究了六种在使用倾向得分匹配或加权时纳入随访变量基线值的方法。这些方法的不同之处在于倾向得分模型中是否包含或排除了随访变量的基线值,是否在匹配样本或加权样本中进行了后续回归调整以调整随访变量的基线值,以及分析是否估算了治疗对随访变量或基线变化的影响。我们使用蒙特卡罗模拟法对 750 种情况进行了模拟。虽然没有哪种分析方法具有一致的优越性能,但我们还是提出了以下建议:首先,在使用加权法且 ATE 为目标估计值时,应使用增强的逆概率加权估计器,或在倾向评分模型中包含随访变量的基线值,然后在回归模型中对随访变量的基线值进行调整。其次,当 ATT 为目标估计值时,无论使用加权还是匹配,都应使用不包括随访变量基线值的倾向评分来分析与基线相比的变化。
{"title":"Propensity Score Analysis With Baseline and Follow-Up Measurements of the Outcome Variable.","authors":"Peter C Austin","doi":"10.1002/pst.2436","DOIUrl":"https://doi.org/10.1002/pst.2436","url":null,"abstract":"<p><p>A common feature in cohort studies is when there is a baseline measurement of the continuous follow-up or outcome variable. Common examples include baseline measurements of physiological characteristics such as blood pressure or heart rate in studies where the outcome is post-baseline measurement of the same variable. Methods incorporating the propensity score are increasingly being used to estimate the effects of treatments using observational studies. We examined six methods for incorporating the baseline value of the follow-up variable when using propensity score matching or weighting. These methods differed according to whether the baseline value of the follow-up variable was included or excluded from the propensity score model, whether subsequent regression adjustment was conducted in the matched or weighted sample to adjust for the baseline value of the follow-up variable, and whether the analysis estimated the effect of treatment on the follow-up variable or on the change from baseline. We used Monte Carlo simulations with 750 scenarios. While no analytic method had uniformly superior performance, we provide the following recommendations: first, when using weighting and the ATE is the target estimand, use an augmented inverse probability weighted estimator or include the baseline value of the follow-up variable in the propensity score model and subsequently adjust for the baseline value of the follow-up variable in a regression model. Second, when the ATT is the target estimand, regardless of whether using weighting or matching, analyze change from baseline using a propensity score that excludes the baseline value of the follow-up variable.</p>","PeriodicalId":19934,"journal":{"name":"Pharmaceutical Statistics","volume":" ","pages":""},"PeriodicalIF":1.3,"publicationDate":"2024-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142140774","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The innovative use of real-world data (RWD) can answer questions that cannot be addressed using data from randomized clinical trials (RCTs). While the sponsors of RCTs have a central database containing all individual patient data (IPD) collected from trials, analysts of RWD face a challenge: regulations on patient privacy make access to IPD from all regions logistically prohibitive. In this research, we propose a double inverse probability weighting (DIPW) approach for the analysis sponsor to estimate the population average treatment effect (PATE) for a target population without the need to access IPD. One probability weighting is for achieving comparable distributions in confounders across treatment groups; another probability weighting is for generalizing the result from a subpopulation of patients who have data on the endpoint to the whole target population. The likelihood expressions for propensity scores and the DIPW estimator of the PATE can be written to only rely on regional summary statistics that do not require IPD. Our approach hinges upon the positivity and conditional independency assumptions, prerequisites to most RWD analysis approaches. Simulations are conducted to compare the performances of the proposed method against a modified meta-analysis and a regular meta-analysis.
{"title":"Generalizing Treatment Effect to a Target Population Without Individual Patient Data in a Real-World Setting.","authors":"Hui Quan, Tong Li, Xun Chen, Gang Li","doi":"10.1002/pst.2435","DOIUrl":"https://doi.org/10.1002/pst.2435","url":null,"abstract":"<p><p>The innovative use of real-world data (RWD) can answer questions that cannot be addressed using data from randomized clinical trials (RCTs). While the sponsors of RCTs have a central database containing all individual patient data (IPD) collected from trials, analysts of RWD face a challenge: regulations on patient privacy make access to IPD from all regions logistically prohibitive. In this research, we propose a double inverse probability weighting (DIPW) approach for the analysis sponsor to estimate the population average treatment effect (PATE) for a target population without the need to access IPD. One probability weighting is for achieving comparable distributions in confounders across treatment groups; another probability weighting is for generalizing the result from a subpopulation of patients who have data on the endpoint to the whole target population. The likelihood expressions for propensity scores and the DIPW estimator of the PATE can be written to only rely on regional summary statistics that do not require IPD. Our approach hinges upon the positivity and conditional independency assumptions, prerequisites to most RWD analysis approaches. Simulations are conducted to compare the performances of the proposed method against a modified meta-analysis and a regular meta-analysis.</p>","PeriodicalId":19934,"journal":{"name":"Pharmaceutical Statistics","volume":" ","pages":""},"PeriodicalIF":1.3,"publicationDate":"2024-09-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142126379","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In a randomized controlled trial with time-to-event endpoint, some commonly used statistical tests to test for various aspects of survival differences, such as survival probability at a fixed time point, survival function up to a specific time point, and restricted mean survival time, may not be directly applicable when external data are leveraged to augment an arm (or both arms) of an RCT. In this paper, we propose a propensity score-integrated approach to extend such tests when external data are leveraged. Simulation studies are conducted to evaluate the operating characteristics of three propensity score-integrated statistical tests, and an illustrative example is given to demonstrate how these proposed procedures can be implemented.
{"title":"A propensity score-integrated approach for leveraging external data in a randomized controlled trial with time-to-event endpoints.","authors":"Wei-Chen Chen, Nelson Lu, Chenguang Wang, Heng Li, Changhong Song, Ram Tiwari, Yunling Xu, Lilly Q Yue","doi":"10.1002/pst.2377","DOIUrl":"10.1002/pst.2377","url":null,"abstract":"<p><p>In a randomized controlled trial with time-to-event endpoint, some commonly used statistical tests to test for various aspects of survival differences, such as survival probability at a fixed time point, survival function up to a specific time point, and restricted mean survival time, may not be directly applicable when external data are leveraged to augment an arm (or both arms) of an RCT. In this paper, we propose a propensity score-integrated approach to extend such tests when external data are leveraged. Simulation studies are conducted to evaluate the operating characteristics of three propensity score-integrated statistical tests, and an illustrative example is given to demonstrate how these proposed procedures can be implemented.</p>","PeriodicalId":19934,"journal":{"name":"Pharmaceutical Statistics","volume":" ","pages":"645-661"},"PeriodicalIF":16.4,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140040067","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}