Pub Date : 2024-12-05DOI: 10.1186/s12874-024-02432-x
Yasin Okkaoglu, Nicky J Welton, Hayley E Jones
Background: Latent class models can be used to estimate diagnostic accuracy without a gold standard test. Early studies often assumed independence between tests given the true disease state, however this can lead to biased estimates when there are inter-test dependencies. Residual correlation plots and chi-squared statistics have been commonly utilized to assess the validity of the conditional independence assumption and, when it does not hold, identify which test pairs are conditionally dependent. We aimed to assess the performance of these tools with a simulation study covering a wide range of scenarios.
Methods: We generated data sets from a model with four tests and a dependence between tests 1 and 2 within the diseased group. We varied sample size, prevalence, covariance, sensitivity and specificity, with 504 combinations of these in total, and 1000 data sets for each combination. We fitted the conditional independence model in a Bayesian framework, and reported absolute bias, coverage, and how often the residual correlation plots, and statistics indicated lack-of-fit globally or for each test pair.
Results: Across all settings, residual correlation plots, pairwise and detected the correct correlated pair of tests only 12.1%, 10.3%, and 10.3% of the time, respectively, but incorrectly suggested dependence between tests 3 and 4 64.9%, 49.7%, and 49.5% of the time. We observed some variation in this across parameter settings, with these tools appearing to perform more as intended when tests 3 and 4 were both much more accurate than tests 1 and 2. Residual correlation plots, and statistics identified a lack of overall fit in 74.3%, 64.5% and 67.5% of models, respectively. The conditional independence model tended to overestimate the sensitivities of the correlated tests (median bias across all scenarios 0.094, 2.5th and 97.5th percentiles -0.003, 0.397) and underestimate prevalence and the specificities of the uncorrelated tests.
Conclusions: Residual correlation plots and chi-squared statistics cannot be relied upon to identify which tests are conditionally dependent, and also have relatively low power to detect lack of overall fit. This is important since failure to account for conditional dependence can lead to highly biased parameter estimates.
{"title":"Detecting departures from the conditional independence assumption in diagnostic latent class models: a simulation study.","authors":"Yasin Okkaoglu, Nicky J Welton, Hayley E Jones","doi":"10.1186/s12874-024-02432-x","DOIUrl":"10.1186/s12874-024-02432-x","url":null,"abstract":"<p><strong>Background: </strong>Latent class models can be used to estimate diagnostic accuracy without a gold standard test. Early studies often assumed independence between tests given the true disease state, however this can lead to biased estimates when there are inter-test dependencies. Residual correlation plots and chi-squared statistics have been commonly utilized to assess the validity of the conditional independence assumption and, when it does not hold, identify which test pairs are conditionally dependent. We aimed to assess the performance of these tools with a simulation study covering a wide range of scenarios.</p><p><strong>Methods: </strong>We generated data sets from a model with four tests and a dependence between tests 1 and 2 within the diseased group. We varied sample size, prevalence, covariance, sensitivity and specificity, with 504 combinations of these in total, and 1000 data sets for each combination. We fitted the conditional independence model in a Bayesian framework, and reported absolute bias, coverage, and how often the residual correlation plots, <math> <msup><mrow><mi>G</mi></mrow> <mn>2</mn></msup> </math> and <math> <msup><mrow><mi>χ</mi></mrow> <mn>2</mn></msup> </math> statistics indicated lack-of-fit globally or for each test pair.</p><p><strong>Results: </strong>Across all settings, residual correlation plots, pairwise <math> <msup><mrow><mi>G</mi></mrow> <mn>2</mn></msup> </math> and <math> <msup><mrow><mi>χ</mi></mrow> <mn>2</mn></msup> </math> detected the correct correlated pair of tests only 12.1%, 10.3%, and 10.3% of the time, respectively, but incorrectly suggested dependence between tests 3 and 4 64.9%, 49.7%, and 49.5% of the time. We observed some variation in this across parameter settings, with these tools appearing to perform more as intended when tests 3 and 4 were both much more accurate than tests 1 and 2. Residual correlation plots, <math> <msup><mrow><mi>G</mi></mrow> <mn>2</mn></msup> </math> and <math> <msup><mrow><mi>χ</mi></mrow> <mn>2</mn></msup> </math> statistics identified a lack of overall fit in 74.3%, 64.5% and 67.5% of models, respectively. The conditional independence model tended to overestimate the sensitivities of the correlated tests (median bias across all scenarios 0.094, 2.5th and 97.5th percentiles -0.003, 0.397) and underestimate prevalence and the specificities of the uncorrelated tests.</p><p><strong>Conclusions: </strong>Residual correlation plots and chi-squared statistics cannot be relied upon to identify which tests are conditionally dependent, and also have relatively low power to detect lack of overall fit. This is important since failure to account for conditional dependence can lead to highly biased parameter estimates.</p>","PeriodicalId":9114,"journal":{"name":"BMC Medical Research Methodology","volume":"24 1","pages":"299"},"PeriodicalIF":3.9,"publicationDate":"2024-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11619692/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142784103","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-12-05DOI: 10.1186/s12874-024-02405-0
Priyanka Mendon, Michael Witsch, Marianne Becker, Aurélie Adamski, Michel Vaillant
Background: Monitoring of somatic development through the assessment of anthropometric variables such as weight, height, and BMI is vital for evaluating the physical development and nutritional status of children. This approach aids in the early identification of somatic developmental disorders, enabling timely medical interventions. It traditionally relies on Z-scores, which compare anthropometric variables with reference standards. In addition to somatic development monitoring, the early detection and management of pediatric and adolescent hypertension are crucial due to potential long-term health risks. However, manual calculations of Z-scores are time-consuming and error-prone, impeding timely interventions for at-risk children. This article introduces an innovative open-code solution for real-time Z-score assessments directly within the electronic data capture platform, Research Electronic Data Capture, (REDCap™), aiming to streamline the monitoring of somatic development in children.
Methods: Leveraging the World Health Organization (WHO) growth standards and National Health and Nutrition Examination Survey (NHANES) references, our approach integrates Z-score computations directly into REDCap, providing a secure and user-friendly environment for healthcare professionals and researchers. We employed Bland-Altman analyses to compare our method with established calculators (Knirps and Growth XP™) using synthetic data values for all variables.
Results: Bland-Altman plots demonstrated strong agreement between our REDCap calculations and the Knirps and Growth XP systems. Z-scores for height, BMI, and blood pressure consistently aligned, affirming the accuracy of our approach across the measurement range.
Conclusion: The integration with REDCap streamlines data collection and analysis, eliminating the need for separate software and data exports. Moreover, our solution uses the World Health Organization (WHO) growth standards and National Health and Nutrition Examination Survey (NHANES) references. This not only ensures calculation accuracy but also enhances its suitability for diverse research contexts. The Bland-Altman analyses establish the reliability of our method, contributing to a more effective approach to child growth and blood pressure monitoring.
背景:通过评估体重、身高和BMI等人体测量变量来监测身体发育对于评估儿童的身体发育和营养状况至关重要。这种方法有助于早期识别躯体发育障碍,从而能够及时进行医疗干预。它传统上依赖于z分数,将人体测量变量与参考标准进行比较。除了身体发育监测外,由于潜在的长期健康风险,儿童和青少年高血压的早期发现和管理至关重要。然而,手工计算z分数既耗时又容易出错,阻碍了对有风险儿童的及时干预。本文介绍了一种创新的开放代码解决方案,直接在电子数据捕获平台Research electronic data capture (REDCap™)中进行实时Z-score评估,旨在简化对儿童身体发育的监测。方法:利用世界卫生组织(WHO)生长标准和国家健康与营养检查调查(NHANES)参考资料,我们的方法将z分数计算直接集成到REDCap中,为医疗保健专业人员和研究人员提供了一个安全且用户友好的环境。我们采用Bland-Altman分析,将我们的方法与现有的计算器(Knirps和Growth XP™)进行比较,使用所有变量的合成数据值。结果:Bland-Altman图显示了我们的REDCap计算与Knirps和Growth XP系统之间的强烈一致性。身高、体重指数和血压的z分数一致,证实了我们的方法在整个测量范围内的准确性。结论:与REDCap的集成简化了数据收集和分析,不再需要单独的软件和数据导出。此外,我们的解决方案采用了世界卫生组织(WHO)生长标准和国家健康与营养检查调查(NHANES)的参考资料。这不仅保证了计算的准确性,而且提高了其对不同研究背景的适用性。Bland-Altman分析建立了我们方法的可靠性,为儿童生长和血压监测提供了更有效的方法。
{"title":"Facilitating comprehensive child health monitoring within REDCap - an open-source code for real-time Z-score assessments.","authors":"Priyanka Mendon, Michael Witsch, Marianne Becker, Aurélie Adamski, Michel Vaillant","doi":"10.1186/s12874-024-02405-0","DOIUrl":"10.1186/s12874-024-02405-0","url":null,"abstract":"<p><strong>Background: </strong>Monitoring of somatic development through the assessment of anthropometric variables such as weight, height, and BMI is vital for evaluating the physical development and nutritional status of children. This approach aids in the early identification of somatic developmental disorders, enabling timely medical interventions. It traditionally relies on Z-scores, which compare anthropometric variables with reference standards. In addition to somatic development monitoring, the early detection and management of pediatric and adolescent hypertension are crucial due to potential long-term health risks. However, manual calculations of Z-scores are time-consuming and error-prone, impeding timely interventions for at-risk children. This article introduces an innovative open-code solution for real-time Z-score assessments directly within the electronic data capture platform, Research Electronic Data Capture, (REDCap™), aiming to streamline the monitoring of somatic development in children.</p><p><strong>Methods: </strong>Leveraging the World Health Organization (WHO) growth standards and National Health and Nutrition Examination Survey (NHANES) references, our approach integrates Z-score computations directly into REDCap, providing a secure and user-friendly environment for healthcare professionals and researchers. We employed Bland-Altman analyses to compare our method with established calculators (Knirps and Growth XP™) using synthetic data values for all variables.</p><p><strong>Results: </strong>Bland-Altman plots demonstrated strong agreement between our REDCap calculations and the Knirps and Growth XP systems. Z-scores for height, BMI, and blood pressure consistently aligned, affirming the accuracy of our approach across the measurement range.</p><p><strong>Conclusion: </strong>The integration with REDCap streamlines data collection and analysis, eliminating the need for separate software and data exports. Moreover, our solution uses the World Health Organization (WHO) growth standards and National Health and Nutrition Examination Survey (NHANES) references. This not only ensures calculation accuracy but also enhances its suitability for diverse research contexts. The Bland-Altman analyses establish the reliability of our method, contributing to a more effective approach to child growth and blood pressure monitoring.</p>","PeriodicalId":9114,"journal":{"name":"BMC Medical Research Methodology","volume":"24 1","pages":"298"},"PeriodicalIF":3.9,"publicationDate":"2024-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11619695/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142784105","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-12-03DOI: 10.1186/s12874-024-02419-8
Isabelle Budin-Ljøsne, Nanna A G Fredheim, Charlotte Alison Jevne, Bojana Milanovic Kleven, Marie Aline Charles, Janine F Felix, Robin Flaig, María Paz García, Alexandra Havdahl, Shahid Islam, Shona M Kerr, Inger Kristine Meder, Lynn Molloy, Susan M B Morton, Costanza Pizzi, Aamnah Rahman, Gonneke Willemsen, Diane Wood, Jennifer R Harris
Background: Longitudinal cohort studies are pivotal to understand how socioeconomic, environmental, biological, and lifestyle factors influence health and disease. The added value of cohort studies increases as they accumulate life course data and expand across generations. Ensuring that participants stay motivated to contribute over decades of follow-up is, however, challenging. Participant engagement and involvement (PEI) aims to secure the long-term commitment of participants and promote researcher-participant interaction. This study explored PEI practices in a selection of pregnancy and birth, twin, and family-based population cohort studies.
Methods: Purposive sampling was used to identify cohorts in Europe, Australia and New Zealand. Fourteen semi-structured digital interviews were conducted with cohort study representatives to explore strategies for participant recruitment, informed consent, communication of general and individual information to participants, data collection, and participant involvement. Experiences, resources allocated to PEI, and reflections on future PEI, were discussed. The interview data were analyzed using a content analysis approach and summary results were reviewed and discussed by the representatives.
Results: The cohort studies used various strategies to recruit participants including support from health professionals and organizations combined with information on the studies' web sites and social media. New approaches such as intra-cohort recruitment, were being investigated. Most cohorts transitioned from paper-based to digital solutions to collect the participants' consent and data. While digital solutions increased efficiency, they also brought new challenges. The studies experimented with the use of participant advisory panels and focus groups to involve participants in making decisions, although their success varied across age and socio-economic background. Most representatives reported PEI resources to be limited and called for more human, technical, educational and financial resources to maximize the positive effects of PEI.
Conclusions: This study of PEI among well-established cohort studies underscores the importance of PEI for project sustainability and highlights key factors to consider in developing PEI. Our analysis shows that knowledge gaps exist regarding which approaches have highest impact on retention rates and are best suited for different participant groups. Research is needed to support the development of best practices for PEI as well as knowledge exchange between cohorts through network building.
{"title":"Participant engagement and involvement in longitudinal cohort studies: qualitative insights from a selection of pregnancy and birth, twin, and family-based population cohort studies.","authors":"Isabelle Budin-Ljøsne, Nanna A G Fredheim, Charlotte Alison Jevne, Bojana Milanovic Kleven, Marie Aline Charles, Janine F Felix, Robin Flaig, María Paz García, Alexandra Havdahl, Shahid Islam, Shona M Kerr, Inger Kristine Meder, Lynn Molloy, Susan M B Morton, Costanza Pizzi, Aamnah Rahman, Gonneke Willemsen, Diane Wood, Jennifer R Harris","doi":"10.1186/s12874-024-02419-8","DOIUrl":"10.1186/s12874-024-02419-8","url":null,"abstract":"<p><strong>Background: </strong>Longitudinal cohort studies are pivotal to understand how socioeconomic, environmental, biological, and lifestyle factors influence health and disease. The added value of cohort studies increases as they accumulate life course data and expand across generations. Ensuring that participants stay motivated to contribute over decades of follow-up is, however, challenging. Participant engagement and involvement (PEI) aims to secure the long-term commitment of participants and promote researcher-participant interaction. This study explored PEI practices in a selection of pregnancy and birth, twin, and family-based population cohort studies.</p><p><strong>Methods: </strong>Purposive sampling was used to identify cohorts in Europe, Australia and New Zealand. Fourteen semi-structured digital interviews were conducted with cohort study representatives to explore strategies for participant recruitment, informed consent, communication of general and individual information to participants, data collection, and participant involvement. Experiences, resources allocated to PEI, and reflections on future PEI, were discussed. The interview data were analyzed using a content analysis approach and summary results were reviewed and discussed by the representatives.</p><p><strong>Results: </strong>The cohort studies used various strategies to recruit participants including support from health professionals and organizations combined with information on the studies' web sites and social media. New approaches such as intra-cohort recruitment, were being investigated. Most cohorts transitioned from paper-based to digital solutions to collect the participants' consent and data. While digital solutions increased efficiency, they also brought new challenges. The studies experimented with the use of participant advisory panels and focus groups to involve participants in making decisions, although their success varied across age and socio-economic background. Most representatives reported PEI resources to be limited and called for more human, technical, educational and financial resources to maximize the positive effects of PEI.</p><p><strong>Conclusions: </strong>This study of PEI among well-established cohort studies underscores the importance of PEI for project sustainability and highlights key factors to consider in developing PEI. Our analysis shows that knowledge gaps exist regarding which approaches have highest impact on retention rates and are best suited for different participant groups. Research is needed to support the development of best practices for PEI as well as knowledge exchange between cohorts through network building.</p>","PeriodicalId":9114,"journal":{"name":"BMC Medical Research Methodology","volume":"24 1","pages":"297"},"PeriodicalIF":3.9,"publicationDate":"2024-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11613753/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142766303","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-12-03DOI: 10.1186/s12874-024-02416-x
Marta Spreafico, Francesca Ieva, Marta Fiocco
Background: This study aims to analyse the effects of reducing Received Dose Intensity (RDI) in chemotherapy treatment for osteosarcoma patients on their survival by using a novel approach. Previous research has highlighted discrepancies between planned and actual RDI, even among patients randomized to the same treatment regimen. To mitigate toxic side effects, treatment adjustments, such as dose reduction or delayed courses, are necessary. Toxicities are therefore risk factors for mortality and predictors of future exposure levels. Toxicity introduces post-assignment confounding when assessing the causal effect of chemotherapy RDI on survival outcomes, a topic of ongoing debate.
Methods: Chemotherapy administration data from BO03 and BO06 Randomized Clinical Trials (RCTs) in ostosarcoma are employed to emulate a target trial with three RDI-based exposure strategies: 1) standard, 2) reduced, and 3) highly-reduced RDI. Investigations are conducted between subgroups of patients characterised by poor or good Histological Responses (HRe), i.e., the strongest known prognostic factor for survival in osteosarcoma. Inverse Probability of Treatment Weighting (IPTW) is first used to transform the original population into a pseudo-population which mimics the target randomized cohort. Then, a Marginal Structural Cox Model with effect modification is employed. Conditional Average Treatment Effects (CATEs) are ultimately measured as the difference between the Restricted Mean Survival Time of reduced/highly-reduced RDI strategy and the standard one. Confidence Intervals for CATEs are obtained using a novel IPTW-based bootstrap procedure.
Results: Significant effect modifications based on HRe were found. Increasing RDI-reductions led to contrasting trends for poor and good responders: the higher the reduction, the better (worsen) was the survival in poor (good) reponders. Due to their intrinsic resistance to chemotherapy, poor reponders could benefit from reduced RDI, with an average gain of 10.2 and 15.4 months at 5-year for reduced and highly-reduced exposures, respectively.
Conclusions: This study introduces a novel approach to (i) comprehensively address the challenges related to the analysis of chemotherapy data, (ii) mitigate the toxicity-treatment-adjustment bias, and (iii) repurpose existing RCT data for retrospective analyses extending beyond the original trials' intended scopes.
背景:本研究旨在通过一种新颖的方法分析降低骨肉瘤患者化疗中接受剂量强度(RDI)对其生存的影响。先前的研究强调了计划和实际RDI之间的差异,即使在随机分配到相同治疗方案的患者中也是如此。为了减轻毒副作用,必须调整治疗,如减少剂量或延迟疗程。因此,毒性是死亡率的危险因素和未来接触水平的预测因素。在评估化疗RDI对生存结果的因果关系时,毒性引入了分配后的混淆,这是一个持续争论的话题。方法:采用BO03和BO06随机临床试验(rct)的骨肉瘤化疗给药数据,模拟三种基于RDI暴露策略的靶试验:1)标准,2)降低RDI, 3)高度降低RDI。研究在以组织学反应(HRe)差或好为特征的患者亚组之间进行,HRe是骨肉瘤患者生存的已知最强预后因素。首先利用处理加权逆概率(IPTW)将原始群体转化为模拟目标随机队列的伪群体。然后,采用效应修正的边际结构Cox模型。条件平均治疗效果(Conditional Average Treatment Effects, CATEs)最终被衡量为减少/高度减少RDI策略与标准策略的限制平均生存时间之间的差异。利用一种新颖的基于iptw的自举方法获得了CATEs的置信区间。结果:以HRe为基础的改良效果显著。rdi减少的增加导致了不良反应和良好反应的不同趋势:减少的越高,不良(良好)反应者的生存越好(越差)。由于对化疗的内在抗性,反应不良的患者可以从减少RDI中获益,减少和高度减少暴露的患者在5年的平均获益分别为10.2和15.4个月。结论:本研究引入了一种新的方法来(i)全面解决与化疗数据分析相关的挑战,(ii)减轻毒性-治疗-调整偏差,以及(iii)重新利用现有的RCT数据进行回顾性分析,扩展到原始试验的预期范围。
{"title":"Causal effect of chemotherapy received dose intensity on survival outcome: a retrospective study in osteosarcoma.","authors":"Marta Spreafico, Francesca Ieva, Marta Fiocco","doi":"10.1186/s12874-024-02416-x","DOIUrl":"10.1186/s12874-024-02416-x","url":null,"abstract":"<p><strong>Background: </strong>This study aims to analyse the effects of reducing Received Dose Intensity (RDI) in chemotherapy treatment for osteosarcoma patients on their survival by using a novel approach. Previous research has highlighted discrepancies between planned and actual RDI, even among patients randomized to the same treatment regimen. To mitigate toxic side effects, treatment adjustments, such as dose reduction or delayed courses, are necessary. Toxicities are therefore risk factors for mortality and predictors of future exposure levels. Toxicity introduces post-assignment confounding when assessing the causal effect of chemotherapy RDI on survival outcomes, a topic of ongoing debate.</p><p><strong>Methods: </strong>Chemotherapy administration data from BO03 and BO06 Randomized Clinical Trials (RCTs) in ostosarcoma are employed to emulate a target trial with three RDI-based exposure strategies: 1) standard, 2) reduced, and 3) highly-reduced RDI. Investigations are conducted between subgroups of patients characterised by poor or good Histological Responses (HRe), i.e., the strongest known prognostic factor for survival in osteosarcoma. Inverse Probability of Treatment Weighting (IPTW) is first used to transform the original population into a pseudo-population which mimics the target randomized cohort. Then, a Marginal Structural Cox Model with effect modification is employed. Conditional Average Treatment Effects (CATEs) are ultimately measured as the difference between the Restricted Mean Survival Time of reduced/highly-reduced RDI strategy and the standard one. Confidence Intervals for CATEs are obtained using a novel IPTW-based bootstrap procedure.</p><p><strong>Results: </strong>Significant effect modifications based on HRe were found. Increasing RDI-reductions led to contrasting trends for poor and good responders: the higher the reduction, the better (worsen) was the survival in poor (good) reponders. Due to their intrinsic resistance to chemotherapy, poor reponders could benefit from reduced RDI, with an average gain of 10.2 and 15.4 months at 5-year for reduced and highly-reduced exposures, respectively.</p><p><strong>Conclusions: </strong>This study introduces a novel approach to (i) comprehensively address the challenges related to the analysis of chemotherapy data, (ii) mitigate the toxicity-treatment-adjustment bias, and (iii) repurpose existing RCT data for retrospective analyses extending beyond the original trials' intended scopes.</p>","PeriodicalId":9114,"journal":{"name":"BMC Medical Research Methodology","volume":"24 1","pages":"296"},"PeriodicalIF":3.9,"publicationDate":"2024-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11613923/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142766298","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-29DOI: 10.1186/s12874-024-02420-1
Amitha Puranik, Peter J Diggle, Maurice R Odiere, Katherine Gass, Stella Kepha, Collins Okoyo, Charles Mwandawiro, Florence Wakesho, Wycliff Omondi, Hadley Matendechero Sultani, Emanuele Giorgi
Background: Soil-transmitted helminthiasis (STH) are a parasitic infection that predominantly affects impoverished regions. Model-based geostatistics (MBG) has been established as a set of modern statistical methods that enable mapping of disease risk in a geographical area of interest. We investigate how the use of remotely sensed covariates can help to improve the predictive inferences on STH prevalence using MBG methods. In particular, we focus on how the covariates impact on the classification of areas into distinct class of STH prevalence.
Methods: This study uses secondary data obtained from a sample of 1551 schools in Kenya, gathered through a combination of longitudinal and cross-sectional surveys. We compare the performance of two geostatistical models: one that does not make use of any spatially referenced covariate; and a second model that uses remotely sensed covariates to assist STH prevalence prediction. We also carry out a simulation study in which we compare the performance of the two models in the classifications of areal units with varying sample sizes and prevalence levels.
Results: The model with covariates generated lower levels of uncertainty and was able to classify 88 more districts into prevalence classes than the model without covariates, which instead left those as "unclassified". The simulation study showed that the model with covariates also yielded a higher proportion of correct classification of at least 40% for all sub-counties.
Conclusion: Covariates can substantially reduce the uncertainty of the predictive inference generated from geostatistical models. Using covariates can thus contribute to the design of more effective STH control strategies by reducing sample sizes without compromising the predictive performance of geostatistical models.
{"title":"Understanding the impact of covariates on the classification of implementation units for soil-transmitted helminths control: a case study from Kenya.","authors":"Amitha Puranik, Peter J Diggle, Maurice R Odiere, Katherine Gass, Stella Kepha, Collins Okoyo, Charles Mwandawiro, Florence Wakesho, Wycliff Omondi, Hadley Matendechero Sultani, Emanuele Giorgi","doi":"10.1186/s12874-024-02420-1","DOIUrl":"10.1186/s12874-024-02420-1","url":null,"abstract":"<p><strong>Background: </strong>Soil-transmitted helminthiasis (STH) are a parasitic infection that predominantly affects impoverished regions. Model-based geostatistics (MBG) has been established as a set of modern statistical methods that enable mapping of disease risk in a geographical area of interest. We investigate how the use of remotely sensed covariates can help to improve the predictive inferences on STH prevalence using MBG methods. In particular, we focus on how the covariates impact on the classification of areas into distinct class of STH prevalence.</p><p><strong>Methods: </strong>This study uses secondary data obtained from a sample of 1551 schools in Kenya, gathered through a combination of longitudinal and cross-sectional surveys. We compare the performance of two geostatistical models: one that does not make use of any spatially referenced covariate; and a second model that uses remotely sensed covariates to assist STH prevalence prediction. We also carry out a simulation study in which we compare the performance of the two models in the classifications of areal units with varying sample sizes and prevalence levels.</p><p><strong>Results: </strong>The model with covariates generated lower levels of uncertainty and was able to classify 88 more districts into prevalence classes than the model without covariates, which instead left those as \"unclassified\". The simulation study showed that the model with covariates also yielded a higher proportion of correct classification of at least 40% for all sub-counties.</p><p><strong>Conclusion: </strong>Covariates can substantially reduce the uncertainty of the predictive inference generated from geostatistical models. Using covariates can thus contribute to the design of more effective STH control strategies by reducing sample sizes without compromising the predictive performance of geostatistical models.</p>","PeriodicalId":9114,"journal":{"name":"BMC Medical Research Methodology","volume":"24 1","pages":"294"},"PeriodicalIF":3.9,"publicationDate":"2024-11-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11606136/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142754683","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-29DOI: 10.1186/s12874-024-02315-1
Abdelrahman M Makram, Randa Elsheikh, Omar M Makram, Nguyen Thanh Van, Nguyen Hai Nam, Nguyen Khoi Quan, Nguyen Tran Minh Duc, Ngoc Quynh Tram Nguyen, Gibson Omwansa Javes, Sara S Elsheikh, Atsuko Imoto, Peter Lee, Norio Ohmagari, Hirotsugu Aiga, Yasuhiko Kamiya, Patricia Takako Endo, Nguyen Tien Huy
A research protocol is a document that outlines the proposed research idea and is submitted to funding agencies, institutions, or journals for approval. Writing a research protocol represents a challenge, particularly for early-career researchers. In this guide, we aim to provide detailed guidance with the key components and offer practical tips for crafting a research protocol in line with the various study designs. Specifically, the structure of a research protocol should contain the following items: (1) a title that is specific, catchy, and impressive within the word limitation; (2) an abstract that briefs the critical points of the study; (3) an introduction highlighting the study context from broad to narrow and defining the knowledge gap; (4) a justification underlining the significance of the proposed study; (5) Specific, Measurable, Attainable, Relevant, and Time-bound (SMART) objective(s) and aim(s); (6) a methodology covering seven sub-items, including [i] study design and settings, [ii] study subjects, [iii] sample size calculation and sampling, [iv] participants recruitment and follow-up, [v] questionnaire development, [vi] potential variables and outcomes, and [vii] data analysis plan; (7) dissemination of the results; (8) ethics and conflict of interests; (9) budgets analysis/ funding disclosure; and (10) references. This guide will give an overview of these steps and provide clear and concise tips on how to successfully draft a scientific protocol. With careful planning and appropriate guidance, it is possible to develop a well-structured and compelling protocol to obtain approval for the conduction of the study or funding from agencies, institutions, or organizations.
{"title":"Tips from an expert panel on the development of a clinical research protocol.","authors":"Abdelrahman M Makram, Randa Elsheikh, Omar M Makram, Nguyen Thanh Van, Nguyen Hai Nam, Nguyen Khoi Quan, Nguyen Tran Minh Duc, Ngoc Quynh Tram Nguyen, Gibson Omwansa Javes, Sara S Elsheikh, Atsuko Imoto, Peter Lee, Norio Ohmagari, Hirotsugu Aiga, Yasuhiko Kamiya, Patricia Takako Endo, Nguyen Tien Huy","doi":"10.1186/s12874-024-02315-1","DOIUrl":"10.1186/s12874-024-02315-1","url":null,"abstract":"<p><p>A research protocol is a document that outlines the proposed research idea and is submitted to funding agencies, institutions, or journals for approval. Writing a research protocol represents a challenge, particularly for early-career researchers. In this guide, we aim to provide detailed guidance with the key components and offer practical tips for crafting a research protocol in line with the various study designs. Specifically, the structure of a research protocol should contain the following items: (1) a title that is specific, catchy, and impressive within the word limitation; (2) an abstract that briefs the critical points of the study; (3) an introduction highlighting the study context from broad to narrow and defining the knowledge gap; (4) a justification underlining the significance of the proposed study; (5) Specific, Measurable, Attainable, Relevant, and Time-bound (SMART) objective(s) and aim(s); (6) a methodology covering seven sub-items, including [i] study design and settings, [ii] study subjects, [iii] sample size calculation and sampling, [iv] participants recruitment and follow-up, [v] questionnaire development, [vi] potential variables and outcomes, and [vii] data analysis plan; (7) dissemination of the results; (8) ethics and conflict of interests; (9) budgets analysis/ funding disclosure; and (10) references. This guide will give an overview of these steps and provide clear and concise tips on how to successfully draft a scientific protocol. With careful planning and appropriate guidance, it is possible to develop a well-structured and compelling protocol to obtain approval for the conduction of the study or funding from agencies, institutions, or organizations.</p>","PeriodicalId":9114,"journal":{"name":"BMC Medical Research Methodology","volume":"24 1","pages":"293"},"PeriodicalIF":3.9,"publicationDate":"2024-11-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11606108/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142754672","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-29DOI: 10.1186/s12874-024-02390-4
George Bouliotis, M Underwood, R Froud
Background: Whether machine learning approaches are superior to classical statistical models for survival analyses, especially in the case of lack of proportionality, is unknown.
Objectives: To compare model performance and predictive accuracy of classic regressions and machine learning approaches using data from the Inspiring Families programme.
Methods: The Inspiring Families programme aims to support members of families with complex issues to return to work. We explored predictors of time to return to work with proportional hazards (Semi-Parametric Cox in Stata) and (Flexible Parametric Parmar-Royston in Stata) against the Survival penalised regression with Elastic Net penalty (scikit-survival), (conditional) Survival Forest algorithm (pySurvival), and (kernel) Survival Support Vector Machine (pySurvival).
Results: At baseline we obtained data on 61 binary variables from all 3161 participants. No model appeared superior, with a low predictive power (concordance index between 0.51 and 0.61). The median time for finding the first job was about 254 days. The top five contributing variables were 'family issues and additional barriers', 'restriction of hours', 'available CV', 'self-employment considered' and 'education'. The Harrell's Concordance index was range from 0.60 (Cox model) to 0.71 (Random Survival Forest) suggesting a better fit for the machine learning approaches. However, the comparison for predicting median time on a selected scenario based showed only minor differences.
Conclusion: Implementing a series of survival models with and without proportional hazards background provides a useful insight as well as better interpretation of the coefficients affected by non-linearities. However, that better fit does not translate to substantially higher predictive power and accuracy from using machine learning approaches. Further tuning of the machine learning algorithms may provide improved results.
{"title":"Predicting the time to get back to work using statistical models and machine learning approaches.","authors":"George Bouliotis, M Underwood, R Froud","doi":"10.1186/s12874-024-02390-4","DOIUrl":"10.1186/s12874-024-02390-4","url":null,"abstract":"<p><strong>Background: </strong>Whether machine learning approaches are superior to classical statistical models for survival analyses, especially in the case of lack of proportionality, is unknown.</p><p><strong>Objectives: </strong>To compare model performance and predictive accuracy of classic regressions and machine learning approaches using data from the Inspiring Families programme.</p><p><strong>Methods: </strong>The Inspiring Families programme aims to support members of families with complex issues to return to work. We explored predictors of time to return to work with proportional hazards (Semi-Parametric Cox in Stata) and (Flexible Parametric Parmar-Royston in Stata) against the Survival penalised regression with Elastic Net penalty (scikit-survival), (conditional) Survival Forest algorithm (pySurvival), and (kernel) Survival Support Vector Machine (pySurvival).</p><p><strong>Results: </strong>At baseline we obtained data on 61 binary variables from all 3161 participants. No model appeared superior, with a low predictive power (concordance index between 0.51 and 0.61). The median time for finding the first job was about 254 days. The top five contributing variables were 'family issues and additional barriers', 'restriction of hours', 'available CV', 'self-employment considered' and 'education'. The Harrell's Concordance index was range from 0.60 (Cox model) to 0.71 (Random Survival Forest) suggesting a better fit for the machine learning approaches. However, the comparison for predicting median time on a selected scenario based showed only minor differences.</p><p><strong>Conclusion: </strong>Implementing a series of survival models with and without proportional hazards background provides a useful insight as well as better interpretation of the coefficients affected by non-linearities. However, that better fit does not translate to substantially higher predictive power and accuracy from using machine learning approaches. Further tuning of the machine learning algorithms may provide improved results.</p>","PeriodicalId":9114,"journal":{"name":"BMC Medical Research Methodology","volume":"24 1","pages":"295"},"PeriodicalIF":3.9,"publicationDate":"2024-11-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11606207/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142754670","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-26DOI: 10.1186/s12874-024-02402-3
Helen Bian, Menglan Pang, Guanbo Wang, Zihang Lu
Background: The hazard ratio of the Cox proportional hazards model is widely used in randomized controlled trials to assess treatment effects. However, two properties of the hazard ratio including the non-collapsibility and built-in selection bias need to be further investigated.
Methods: We conduct simulations to differentiate the non-collapsibility effect and built-in selection bias from the difference between the marginal and the conditional hazard ratio. Meanwhile, we explore the performance of the Cox model with inverse probability of treatment weighting for covariate adjustment when estimating the marginal hazard ratio. The built-in selection bias is further assessed in the period-specific hazard ratio.
Results: The conditional hazard ratio is a biased estimate of the marginal effect due to the non-collapsibility property. In contrast, the hazard ratio estimated from the inverse probability of treatment weighting Cox model provides an unbiased estimate of the true marginal hazard ratio. The built-in selection bias only manifests in the period-specific hazard ratios even when the proportional hazards assumption is satisfied. The Cox model with inverse probability of treatment weighting can be used to account for confounding bias and provide an unbiased effect under the randomized controlled trials setting when the parameter of interest is the marginal effect.
Conclusions: We propose that the period-specific hazard ratios should always be avoided due to the profound effects of built-in selection bias.
{"title":"Non-collapsibility and built-in selection bias of period-specific and conventional hazard ratio in randomized controlled trials.","authors":"Helen Bian, Menglan Pang, Guanbo Wang, Zihang Lu","doi":"10.1186/s12874-024-02402-3","DOIUrl":"10.1186/s12874-024-02402-3","url":null,"abstract":"<p><strong>Background: </strong>The hazard ratio of the Cox proportional hazards model is widely used in randomized controlled trials to assess treatment effects. However, two properties of the hazard ratio including the non-collapsibility and built-in selection bias need to be further investigated.</p><p><strong>Methods: </strong>We conduct simulations to differentiate the non-collapsibility effect and built-in selection bias from the difference between the marginal and the conditional hazard ratio. Meanwhile, we explore the performance of the Cox model with inverse probability of treatment weighting for covariate adjustment when estimating the marginal hazard ratio. The built-in selection bias is further assessed in the period-specific hazard ratio.</p><p><strong>Results: </strong>The conditional hazard ratio is a biased estimate of the marginal effect due to the non-collapsibility property. In contrast, the hazard ratio estimated from the inverse probability of treatment weighting Cox model provides an unbiased estimate of the true marginal hazard ratio. The built-in selection bias only manifests in the period-specific hazard ratios even when the proportional hazards assumption is satisfied. The Cox model with inverse probability of treatment weighting can be used to account for confounding bias and provide an unbiased effect under the randomized controlled trials setting when the parameter of interest is the marginal effect.</p><p><strong>Conclusions: </strong>We propose that the period-specific hazard ratios should always be avoided due to the profound effects of built-in selection bias.</p>","PeriodicalId":9114,"journal":{"name":"BMC Medical Research Methodology","volume":"24 1","pages":"292"},"PeriodicalIF":3.9,"publicationDate":"2024-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11590464/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142726119","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-25DOI: 10.1186/s12874-024-02401-4
Marius Goldkuhle, Caroline Hirsch, Claire Iannizzi, Ana-Mihaela Zorger, Ralf Bender, Elvira C van Dalen, Lars G Hemkens, Ina Monsef, Nina Kreuzberger, Nicole Skoetz
Background: Time-to-event analysis is associated with methodological complexities. Previous research identified flaws in the reporting of time-to-event analyses in randomized trial publications. These hardships impose challenges for meta-analyses of time-to-event outcomes based on aggregate data. We examined the characteristics, reporting and methods of systematic reviews including such analyses.
Methods: Through a systematic search (02/2017-08/2020), we identified 50 Cochrane Reviews with ≥ 1 meta-analysis based on the hazard ratio (HR) and a corresponding random sample (n = 50) from core clinical journals (Medline; 08/02/2021). Data was extracted in duplicate and included outcome definitions, general and time-to-event specific methods and handling of time-to-event relevant trial characteristics.
Results: The included reviews analyzed 217 time-to-event outcomes (Median: 2; IQR 1-2), most frequently overall survival (41%). Outcome definitions were provided for less than half of time-to-event outcomes (48%). Few reviews specified general methods, e.g., included analysis types (intention-to-treat, per protocol) (35%) and adjustment of effect estimates (12%). Sources that review authors used for retrieval of time-to-event summary data from publications varied substantially. Most frequently reported were direct inclusion of HRs (64%) and reference to established guidance without further specification (46%). Study characteristics important to time-to-event analysis, such as variable follow-up, informative censoring or proportional hazards, were rarely reported. If presented, complementary absolute effect estimates calculated based on the pooled HR were incorrectly calculated (14%) or correct but falsely labeled (11%) in several reviews.
Conclusions: Our findings indicate that limitations in reporting of trial time-to-event analyses translate to the review level as well. Inconsistent reporting of meta-analyses of time-to-event outcomes necessitates additional reporting standards.
{"title":"Exploring the characteristics, methods and reporting of systematic reviews with meta-analyses of time-to-event outcomes: a meta-epidemiological study.","authors":"Marius Goldkuhle, Caroline Hirsch, Claire Iannizzi, Ana-Mihaela Zorger, Ralf Bender, Elvira C van Dalen, Lars G Hemkens, Ina Monsef, Nina Kreuzberger, Nicole Skoetz","doi":"10.1186/s12874-024-02401-4","DOIUrl":"10.1186/s12874-024-02401-4","url":null,"abstract":"<p><strong>Background: </strong>Time-to-event analysis is associated with methodological complexities. Previous research identified flaws in the reporting of time-to-event analyses in randomized trial publications. These hardships impose challenges for meta-analyses of time-to-event outcomes based on aggregate data. We examined the characteristics, reporting and methods of systematic reviews including such analyses.</p><p><strong>Methods: </strong>Through a systematic search (02/2017-08/2020), we identified 50 Cochrane Reviews with ≥ 1 meta-analysis based on the hazard ratio (HR) and a corresponding random sample (n = 50) from core clinical journals (Medline; 08/02/2021). Data was extracted in duplicate and included outcome definitions, general and time-to-event specific methods and handling of time-to-event relevant trial characteristics.</p><p><strong>Results: </strong>The included reviews analyzed 217 time-to-event outcomes (Median: 2; IQR 1-2), most frequently overall survival (41%). Outcome definitions were provided for less than half of time-to-event outcomes (48%). Few reviews specified general methods, e.g., included analysis types (intention-to-treat, per protocol) (35%) and adjustment of effect estimates (12%). Sources that review authors used for retrieval of time-to-event summary data from publications varied substantially. Most frequently reported were direct inclusion of HRs (64%) and reference to established guidance without further specification (46%). Study characteristics important to time-to-event analysis, such as variable follow-up, informative censoring or proportional hazards, were rarely reported. If presented, complementary absolute effect estimates calculated based on the pooled HR were incorrectly calculated (14%) or correct but falsely labeled (11%) in several reviews.</p><p><strong>Conclusions: </strong>Our findings indicate that limitations in reporting of trial time-to-event analyses translate to the review level as well. Inconsistent reporting of meta-analyses of time-to-event outcomes necessitates additional reporting standards.</p>","PeriodicalId":9114,"journal":{"name":"BMC Medical Research Methodology","volume":"24 1","pages":"291"},"PeriodicalIF":3.9,"publicationDate":"2024-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11587663/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142715364","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-23DOI: 10.1186/s12874-024-02408-x
Doranne Thomassen, Satrajit Roychoudhury, Cecilie Delphin Amdal, Dries Reynders, Jammbe Z Musoro, Willi Sauerbrei, Els Goetghebeur, Saskia le Cessie
Background: Patient-reported outcomes (PROs) play an increasing role in the evaluation of oncology treatments. At the same time, single-arm trials are commonly included in regulatory approval submissions. Because of the high risk of biases, results from single-arm trials require careful interpretation. This benefits from a clearly defined estimand, or target of estimation. In this case study, we demonstrated how the ICH E9 (R1) estimand framework can be implemented in SATs with PRO endpoints.
Methods: For the global quality of life outcome in a real single-arm lung cancer trial, a range of possible estimands was defined. We focused on the choice of the variable of interest and strategies to deal with intercurrent events (death, treatment discontinuation and disease progression). Statistical methods were described for each estimand and the corresponding results on the trial data were shown.
Results: Each intercurrent event handling strategy resulted in its own estimated mean global quality of life over time, with a specific interpretation, suitable for a corresponding clinical research aim. In the setting of this case study, a 'while alive' strategy for death and a 'treatment policy' strategy for non-terminal intercurrent events were deemed aligned with a descriptive research aim to inform clinicians and patients about expected quality of life after the start of treatment.
Conclusions: The results show that decisions made in the estimand framework are not trivial. Trial results and their interpretation strongly depend on the chosen estimand. The estimand framework provides a structure to match a research question with a clear target of estimation, supporting specific clinical decisions. Adherence to this framework can help improve the quality of data collection, analysis and reporting of PROs in SATs, impacting decision making in clinical practice.
背景:患者报告结果(PROs)在肿瘤治疗评估中发挥着越来越重要的作用。与此同时,单臂试验通常被纳入监管部门的审批申请中。由于存在偏差的高风险,单臂试验的结果需要仔细解读。这得益于明确定义的估算对象或估算目标。在本案例研究中,我们展示了如何将 ICH E9 (R1) 估计指标框架应用于具有 PRO 终点的 SAT:方法:对于一项真实的单臂肺癌试验中的总体生活质量结果,我们定义了一系列可能的估计值。我们重点关注了相关变量的选择以及处理并发症(死亡、治疗中止和疾病进展)的策略。我们介绍了每种估计值的统计方法,并显示了试验数据的相应结果:结果:每种处理并发症的策略都能估算出一段时间内的总体生活质量平均值,并根据相应的临床研究目标做出具体解释。在本案例研究中,针对死亡的 "存活时 "策略和针对非终末期并发症的 "治疗政策 "策略被认为符合描述性研究的目的,即告知临床医生和患者开始治疗后的预期生活质量:研究结果表明,在估计值框架内做出的决定并非微不足道。试验结果及其解释在很大程度上取决于所选择的估计指标。估算指标框架提供了一个结构,使研究问题与明确的估算目标相匹配,从而为具体的临床决策提供支持。遵守这一框架有助于提高 SAT 中 PROs 的数据收集、分析和报告质量,从而影响临床实践中的决策制定。
{"title":"The role of the estimand framework in the analysis of patient-reported outcomes in single-arm trials: a case study in oncology.","authors":"Doranne Thomassen, Satrajit Roychoudhury, Cecilie Delphin Amdal, Dries Reynders, Jammbe Z Musoro, Willi Sauerbrei, Els Goetghebeur, Saskia le Cessie","doi":"10.1186/s12874-024-02408-x","DOIUrl":"10.1186/s12874-024-02408-x","url":null,"abstract":"<p><strong>Background: </strong>Patient-reported outcomes (PROs) play an increasing role in the evaluation of oncology treatments. At the same time, single-arm trials are commonly included in regulatory approval submissions. Because of the high risk of biases, results from single-arm trials require careful interpretation. This benefits from a clearly defined estimand, or target of estimation. In this case study, we demonstrated how the ICH E9 (R1) estimand framework can be implemented in SATs with PRO endpoints.</p><p><strong>Methods: </strong>For the global quality of life outcome in a real single-arm lung cancer trial, a range of possible estimands was defined. We focused on the choice of the variable of interest and strategies to deal with intercurrent events (death, treatment discontinuation and disease progression). Statistical methods were described for each estimand and the corresponding results on the trial data were shown.</p><p><strong>Results: </strong>Each intercurrent event handling strategy resulted in its own estimated mean global quality of life over time, with a specific interpretation, suitable for a corresponding clinical research aim. In the setting of this case study, a 'while alive' strategy for death and a 'treatment policy' strategy for non-terminal intercurrent events were deemed aligned with a descriptive research aim to inform clinicians and patients about expected quality of life after the start of treatment.</p><p><strong>Conclusions: </strong>The results show that decisions made in the estimand framework are not trivial. Trial results and their interpretation strongly depend on the chosen estimand. The estimand framework provides a structure to match a research question with a clear target of estimation, supporting specific clinical decisions. Adherence to this framework can help improve the quality of data collection, analysis and reporting of PROs in SATs, impacting decision making in clinical practice.</p>","PeriodicalId":9114,"journal":{"name":"BMC Medical Research Methodology","volume":"24 1","pages":"290"},"PeriodicalIF":3.9,"publicationDate":"2024-11-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11585159/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142695321","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}