Pub Date : 2024-12-19DOI: 10.1186/s12874-024-02413-0
Mikkel Schou Andersen, Mikkel Seremet Kofoed, Asger Sand Paludan-Müller, Christian Bonde Pedersen, Tiit Mathiesen, Christian Mawrin, Birgitte Brinkmann Olsen, Bo Halle, Frantz Rom Poulsen
Background: Systematic reviews within the field of animal research are becoming more common. However, in animal translational research, issues related to methodological quality and quality of reporting continue to arise, potentially leading to underestimation or overestimation of the effects of interventions or prevent studies from being replicated. The various tools and checklists available to ensure good-quality studies and proper reporting include both unique and/or overlapping items and/or simply lack necessary elements or are too situational to certain conditions or diseases. Currently, there is no tool available, which covers all aspects of animal models, from bench-top activities to animal facilities, hence a new tool is needed. This tool should be designed to be able to assess all kinds of animal studies such as old, new, low quality, high quality, interventional and noninterventional on. It should do this on multiple levels through items on quality of reporting, methodological (technical) quality, and risk of bias, for use in assessing the overall quality of studies involving animal research.
Methods: During a systematic review of meningioma models in animals, we developed a novel unifying tool that can assess all types of animal studies from multiple perspectives. The tool was inspired by the Collaborative Approach to Meta Analysis and Review of Animal Data from Experimental Studies (CAMARADES) checklist, the ARRIVE 2.0 guidelines, and SYRCLE's risk of bias tool, while also incorporating unique items. We used the interrater agreement percentage and Cohen's kappa index to test the interrater agreement between two independent reviewers for the items in the tool.
Results: There was high interrater agreement across all items (92.9%, 95% CI 91.0-94.8). Cohen's kappa index showed quality of reporting had the best mean index of 0.86 (95%-CI 0.78-0.94), methodological quality had a mean index of 0.83 (95%-CI 0.78-0.94) and finally the items from SYRCLE's risk of bias had a mean kappa index of 0.68 (95%-CI 0.57-0.79).
Conclusions: The Critical Appraisal of Methodological (technical) Quality, Quality of Reporting and Risk of Bias in Animal Research (CRIME-Q) tool unifies a broad spectrum of information (both unique items and items inspired by other methods) about the quality of reporting and methodological (technical) quality, and contains items from SYRCLE's risk of bias. The tool is intended for use in assessing overall study quality across multiple domains and items and is not, unlike other tools, restricted to any particular model or study design (whether interventional or noninterventional). It is also easy to apply when designing and conducting animal experiments to ensure proper reporting and design in terms of replicability, transparency, and validity.
背景:动物研究领域的系统综述正变得越来越普遍。然而,在动物转化研究中,与方法质量和报告质量有关的问题继续出现,可能导致对干预措施效果的低估或高估,或阻止研究被复制。可用于确保高质量研究和适当报告的各种工具和检查表包括独特和/或重叠的项目和/或根本缺乏必要的要素,或对某些条件或疾病过于因时制宜。目前,还没有一种工具可以覆盖动物模型的各个方面,从工作台活动到动物设施,因此需要一种新的工具。该工具应设计成能够评估各种动物研究,如旧的、新的、低质量的、高质量的、介入性的和非介入性的。它应该通过报告质量、方法学(技术)质量和偏倚风险等项目在多个层面上进行评估,以用于评估涉及动物研究的总体质量。方法:在对动物脑膜瘤模型的系统回顾中,我们开发了一种新的统一工具,可以从多个角度评估所有类型的动物研究。该工具的灵感来自于《实验研究动物数据荟萃分析和回顾协作方法》(CAMARADES)清单、ARRIVE 2.0指南和sycle的偏倚风险工具,同时也纳入了独特的项目。我们使用仲裁者协议百分比和Cohen的kappa指数来测试两个独立的评论者对工具中项目的仲裁者协议。结果:所有条目的解释一致性很高(92.9%,95% CI 91.0-94.8)。Cohen的kappa指数显示,报告质量的平均指数为0.86 (95%-CI 0.78-0.94),方法学质量的平均指数为0.83 (95%-CI 0.78-0.94),最后来自sycle的偏倚风险项目的平均kappa指数为0.68 (95%-CI 0.57-0.79)。结论:动物研究中方法(技术)质量、报告质量和偏倚风险的关键评估(CRIME-Q)工具统一了关于报告质量和方法(技术)质量的广泛信息(包括独特的项目和受其他方法启发的项目),并包含来自sycle偏倚风险的项目。该工具旨在用于评估跨多个领域和项目的总体研究质量,与其他工具不同,它不限于任何特定的模型或研究设计(无论是介入性的还是非介入性的)。在设计和进行动物实验时也很容易应用,以确保在可复制性、透明度和有效性方面进行适当的报告和设计。
{"title":"CRIME-Q-a unifying tool for critical appraisal of methodological (technical) quality, quality of reporting and risk of bias in animal research.","authors":"Mikkel Schou Andersen, Mikkel Seremet Kofoed, Asger Sand Paludan-Müller, Christian Bonde Pedersen, Tiit Mathiesen, Christian Mawrin, Birgitte Brinkmann Olsen, Bo Halle, Frantz Rom Poulsen","doi":"10.1186/s12874-024-02413-0","DOIUrl":"10.1186/s12874-024-02413-0","url":null,"abstract":"<p><strong>Background: </strong>Systematic reviews within the field of animal research are becoming more common. However, in animal translational research, issues related to methodological quality and quality of reporting continue to arise, potentially leading to underestimation or overestimation of the effects of interventions or prevent studies from being replicated. The various tools and checklists available to ensure good-quality studies and proper reporting include both unique and/or overlapping items and/or simply lack necessary elements or are too situational to certain conditions or diseases. Currently, there is no tool available, which covers all aspects of animal models, from bench-top activities to animal facilities, hence a new tool is needed. This tool should be designed to be able to assess all kinds of animal studies such as old, new, low quality, high quality, interventional and noninterventional on. It should do this on multiple levels through items on quality of reporting, methodological (technical) quality, and risk of bias, for use in assessing the overall quality of studies involving animal research.</p><p><strong>Methods: </strong>During a systematic review of meningioma models in animals, we developed a novel unifying tool that can assess all types of animal studies from multiple perspectives. The tool was inspired by the Collaborative Approach to Meta Analysis and Review of Animal Data from Experimental Studies (CAMARADES) checklist, the ARRIVE 2.0 guidelines, and SYRCLE's risk of bias tool, while also incorporating unique items. We used the interrater agreement percentage and Cohen's kappa index to test the interrater agreement between two independent reviewers for the items in the tool.</p><p><strong>Results: </strong>There was high interrater agreement across all items (92.9%, 95% CI 91.0-94.8). Cohen's kappa index showed quality of reporting had the best mean index of 0.86 (95%-CI 0.78-0.94), methodological quality had a mean index of 0.83 (95%-CI 0.78-0.94) and finally the items from SYRCLE's risk of bias had a mean kappa index of 0.68 (95%-CI 0.57-0.79).</p><p><strong>Conclusions: </strong>The Critical Appraisal of Methodological (technical) Quality, Quality of Reporting and Risk of Bias in Animal Research (CRIME-Q) tool unifies a broad spectrum of information (both unique items and items inspired by other methods) about the quality of reporting and methodological (technical) quality, and contains items from SYRCLE's risk of bias. The tool is intended for use in assessing overall study quality across multiple domains and items and is not, unlike other tools, restricted to any particular model or study design (whether interventional or noninterventional). It is also easy to apply when designing and conducting animal experiments to ensure proper reporting and design in terms of replicability, transparency, and validity.</p>","PeriodicalId":9114,"journal":{"name":"BMC Medical Research Methodology","volume":"24 1","pages":"306"},"PeriodicalIF":3.9,"publicationDate":"2024-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11656974/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142852496","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-12-19DOI: 10.1186/s12874-024-02439-4
Aoife Whiston, K M Kidwell, S O'Reilly, C Walsh, J C Walsh, L Glynn, K Robinson, S Hayes
Background: Physical activity (PA) is often the cornerstone in risk-reduction interventions for the prevention and treatment of many chronic health conditions. PA interventions are inherently multi-dimensional and complex in nature. Thus, study designs used in the evaluation of PA interventions must be adaptive to intervention components and individual capacities. A Sequential Multiple Assignment Randomised Trial (SMART) is a factorial design in a sequential setting used to build effective adaptive interventions. SMARTs represent a relatively new design for PA intervention research. This systematic review aims to examine the state-of-the-art of SMARTs used to develop PA interventions, with a focus on study characteristics, design, and analyses.
Methods: PubMed, Embase, PsychINFO, CENTRAL, and CinAHL were systematically searched through May 2023 for studies wherein PA SMARTs were conducted. Methodological quality was assessed using the Cochrane Risk of Bias 2 Tool.
Results: Twenty studies across a variety of populations - e.g., obesity, chronic pain, and cardiovascular conditions, were included. All PA SMARTs involved two decision stages, with the majority including two initial treatment options. PA interventions most commonly consisted of individual aerobic exercise with strategies such as goal setting, wearable technology, and motivational interviewing also used to promote PA. Variation was observed across tailoring variables and timing of tailoring variables. Non-response strategies primarily involved augmenting and switching treatment options, and for responders to continue with initial treatment options. For analyses, most sample size estimations and outcome analyses accounted for the SMART aims specified. Techniques such as linear mixed models, weighted regressions, and Q-learning regression were frequently used. Risk of bias was high across the majority of included studies.
Conclusions: Individual-based aerobic exercise interventions supported by behaviour change techniques and wearable sensing technology may play a key role in the future development of SMARTs addressing PA intervention development. Clearer rationale for the selection of tailoring variables, timing of tailoring variables, and included measures is essential to advance PA SMART designs. Collaborative efforts from researchers, clinicians, and patients are needed in order to bridge the gap between adaptive research designs and personalised treatment pathways observed in clinical practice.
{"title":"The use of sequential multiple assignment randomized trials (SMARTs) in physical activity interventions: a systematic review.","authors":"Aoife Whiston, K M Kidwell, S O'Reilly, C Walsh, J C Walsh, L Glynn, K Robinson, S Hayes","doi":"10.1186/s12874-024-02439-4","DOIUrl":"10.1186/s12874-024-02439-4","url":null,"abstract":"<p><strong>Background: </strong>Physical activity (PA) is often the cornerstone in risk-reduction interventions for the prevention and treatment of many chronic health conditions. PA interventions are inherently multi-dimensional and complex in nature. Thus, study designs used in the evaluation of PA interventions must be adaptive to intervention components and individual capacities. A Sequential Multiple Assignment Randomised Trial (SMART) is a factorial design in a sequential setting used to build effective adaptive interventions. SMARTs represent a relatively new design for PA intervention research. This systematic review aims to examine the state-of-the-art of SMARTs used to develop PA interventions, with a focus on study characteristics, design, and analyses.</p><p><strong>Methods: </strong>PubMed, Embase, PsychINFO, CENTRAL, and CinAHL were systematically searched through May 2023 for studies wherein PA SMARTs were conducted. Methodological quality was assessed using the Cochrane Risk of Bias 2 Tool.</p><p><strong>Results: </strong>Twenty studies across a variety of populations - e.g., obesity, chronic pain, and cardiovascular conditions, were included. All PA SMARTs involved two decision stages, with the majority including two initial treatment options. PA interventions most commonly consisted of individual aerobic exercise with strategies such as goal setting, wearable technology, and motivational interviewing also used to promote PA. Variation was observed across tailoring variables and timing of tailoring variables. Non-response strategies primarily involved augmenting and switching treatment options, and for responders to continue with initial treatment options. For analyses, most sample size estimations and outcome analyses accounted for the SMART aims specified. Techniques such as linear mixed models, weighted regressions, and Q-learning regression were frequently used. Risk of bias was high across the majority of included studies.</p><p><strong>Conclusions: </strong>Individual-based aerobic exercise interventions supported by behaviour change techniques and wearable sensing technology may play a key role in the future development of SMARTs addressing PA intervention development. Clearer rationale for the selection of tailoring variables, timing of tailoring variables, and included measures is essential to advance PA SMART designs. Collaborative efforts from researchers, clinicians, and patients are needed in order to bridge the gap between adaptive research designs and personalised treatment pathways observed in clinical practice.</p>","PeriodicalId":9114,"journal":{"name":"BMC Medical Research Methodology","volume":"24 1","pages":"308"},"PeriodicalIF":3.9,"publicationDate":"2024-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11658464/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142863446","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-12-19DOI: 10.1186/s12874-024-02426-9
Nicolas Ngo, Pierre Michel, Roch Giorgi
Background: The -metric value is generally used as the importance score of a feature (or a set of features) in a classification context. This study aimed to go further by creating a new methodology for multivariate feature selection for classification, whereby the -metric is associated with a specific search direction (and therefore a specific stopping criterion). As three search directions are used, we effectively created three distinct methods.
Methods: We assessed the performance of our new methodology through a simulation study, comparing them against more conventional methods. Classification performance indicators, number of selected features, stability and execution time were used to evaluate the performance of the methods. We also evaluated how well the proposed methodology selected relevant features for the detection of atrial fibrillation, which is a cardiac arrhythmia.
Results: We found that in the simulation study as well as the detection of AF task, our methods were able to select informative features and maintain a good level of predictive performance; however in a case of strong correlation and large datasets, the -metric based methods were less efficient to exclude non-informative features.
Conclusions: Results highlighted a good combination of both the forward search direction and the -metric as an evaluation function. However, using the backward search direction, the feature selection algorithm could fall into a local optima and can be improved.
{"title":"<ArticleTitle xmlns:ns0=\"http://www.w3.org/1998/Math/MathML\">Multivariate filter methods for feature selection with the <ns0:math><ns0:mrow><ns0:mi>γ</ns0:mi></ns0:mrow> </ns0:math> -metric.","authors":"Nicolas Ngo, Pierre Michel, Roch Giorgi","doi":"10.1186/s12874-024-02426-9","DOIUrl":"10.1186/s12874-024-02426-9","url":null,"abstract":"<p><strong>Background: </strong>The <math><mi>γ</mi></math> -metric value is generally used as the importance score of a feature (or a set of features) in a classification context. This study aimed to go further by creating a new methodology for multivariate feature selection for classification, whereby the <math><mi>γ</mi></math> -metric is associated with a specific search direction (and therefore a specific stopping criterion). As three search directions are used, we effectively created three distinct methods.</p><p><strong>Methods: </strong>We assessed the performance of our new methodology through a simulation study, comparing them against more conventional methods. Classification performance indicators, number of selected features, stability and execution time were used to evaluate the performance of the methods. We also evaluated how well the proposed methodology selected relevant features for the detection of atrial fibrillation, which is a cardiac arrhythmia.</p><p><strong>Results: </strong>We found that in the simulation study as well as the detection of AF task, our methods were able to select informative features and maintain a good level of predictive performance; however in a case of strong correlation and large datasets, the <math><mi>γ</mi></math> -metric based methods were less efficient to exclude non-informative features.</p><p><strong>Conclusions: </strong>Results highlighted a good combination of both the forward search direction and the <math><mi>γ</mi></math> -metric as an evaluation function. However, using the backward search direction, the feature selection algorithm could fall into a local optima and can be improved.</p>","PeriodicalId":9114,"journal":{"name":"BMC Medical Research Methodology","volume":"24 1","pages":"307"},"PeriodicalIF":3.9,"publicationDate":"2024-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11657396/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142863427","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-12-19DOI: 10.1186/s12874-024-02422-z
Ying Hu, Hai Yan, Ming Liu, Jing Gao, Lianhong Xie, Chunyu Zhang, Lili Wei, Yinging Ding, Hong Jiang
Background: Electronic medical records (EMR)-trained machine learning models have the potential in CVD risk prediction by integrating a range of medical data from patients, facilitate timely diagnosis and classification of CVDs. We tested the hypothesis that unsupervised ML approach utilizing EMR could be used to develop a new model for detecting prevalent CVD in clinical settings.
Methods: We included 155,894 patients (aged ≥ 18 years) discharged between January 2014 and July 2022, from Xuhui Hospital, Shanghai, China, including 64,916 CVD cases and 90,979 non-CVD cases. K-means clustering was used to generate the clustering models with k = 2, 4, and 8 as predetermined number of clusters k = 2, 4, and 8. Bayesian theorem was used to estimate the models' predictive accuracy.
Results: The overall predictive accuracy of the 2-, 4-, and 8-classification clustering models in the training set was 0.856, 0.8634, and 0.8506, respectively. Similarly, the predictive accuracy of the 2-, 4-, and 8-classification clustering models in the testing set was 0.8598, 0.8659, and 0.8525, respectively. After reducing from 19 dimensions to 2 dimensions by principal component analysis, significant separation was observed for CVD cases and non-CVD cases in both training and testing sets.
Conclusion: Our findings indicate that the utilization of EMR data can support the development of a robust model for CVD detection through an unsupervised ML approach. Further investigation using longitudinal design is needed to refine the model for its applications in clinical settings.
{"title":"Detecting cardiovascular diseases using unsupervised machine learning clustering based on electronic medical records.","authors":"Ying Hu, Hai Yan, Ming Liu, Jing Gao, Lianhong Xie, Chunyu Zhang, Lili Wei, Yinging Ding, Hong Jiang","doi":"10.1186/s12874-024-02422-z","DOIUrl":"10.1186/s12874-024-02422-z","url":null,"abstract":"<p><strong>Background: </strong>Electronic medical records (EMR)-trained machine learning models have the potential in CVD risk prediction by integrating a range of medical data from patients, facilitate timely diagnosis and classification of CVDs. We tested the hypothesis that unsupervised ML approach utilizing EMR could be used to develop a new model for detecting prevalent CVD in clinical settings.</p><p><strong>Methods: </strong>We included 155,894 patients (aged ≥ 18 years) discharged between January 2014 and July 2022, from Xuhui Hospital, Shanghai, China, including 64,916 CVD cases and 90,979 non-CVD cases. K-means clustering was used to generate the clustering models with k = 2, 4, and 8 as predetermined number of clusters k = 2, 4, and 8. Bayesian theorem was used to estimate the models' predictive accuracy.</p><p><strong>Results: </strong>The overall predictive accuracy of the 2-, 4-, and 8-classification clustering models in the training set was 0.856, 0.8634, and 0.8506, respectively. Similarly, the predictive accuracy of the 2-, 4-, and 8-classification clustering models in the testing set was 0.8598, 0.8659, and 0.8525, respectively. After reducing from 19 dimensions to 2 dimensions by principal component analysis, significant separation was observed for CVD cases and non-CVD cases in both training and testing sets.</p><p><strong>Conclusion: </strong>Our findings indicate that the utilization of EMR data can support the development of a robust model for CVD detection through an unsupervised ML approach. Further investigation using longitudinal design is needed to refine the model for its applications in clinical settings.</p>","PeriodicalId":9114,"journal":{"name":"BMC Medical Research Methodology","volume":"24 1","pages":"309"},"PeriodicalIF":3.9,"publicationDate":"2024-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11658374/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142863432","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-12-18DOI: 10.1186/s12874-024-02435-8
Leonard Roth, Matthias Studer, Emilie Zuercher, Isabelle Peytremann-Bridevaux
Background: In standard Sequence Analysis, similar trajectories are clustered together to create a typology of trajectories, which is then often used to evaluate the association between sequence patterns and covariates inside regression models. The sampling uncertainty, which affects both the derivation of the typology and the associated regressions, is typically ignored in this analysis, an oversight that may lead to wrong statistical conclusions. We propose utilising sampling variation to derive new estimates that further inform on the association of interest.
Methods: We introduce a novel procedure to assess the robustness of regression results obtained from the standard analysis. Bootstrap samples are drawn from the data, and for each bootstrap, a new typology replicating the original one is constructed, followed by the estimation of the corresponding regression models. The bootstrap estimates are then combined using a multilevel modelling framework that mimics a meta-analysis. The fitted values from this multilevel model allow to account for the sampling uncertainty in the inferential analysis. We illustrate the methodology by applying it to the study of healthcare utilisation trajectories in a Swiss cohort of diabetic patients.
Results: The procedure provides robust estimates for an association of interest, along with 95% prediction intervals, representing the range of expected values if the clustering and associated regressions were performed on a new sample from the same underlying distribution. It also identifies central and borderline trajectories within each cluster. Regarding the illustrative application, while there was evidence of an association between regular lipid testing and subsequent healthcare utilisation patterns in the original analysis, this is not supported in the robustness assessment.
Conclusions: Investigating the relationship between trajectory patterns and covariates is of interest in many situations. However, it is a challenging task with potential pitfalls. Our Robustness Assessment of Regression using Cluster Analysis Typologies (RARCAT) may assist in ensuring the robustness of such association studies. The method is applicable wherever clustering is combined with regression analysis, so its relevance goes beyond State Sequence Analysis.
{"title":"Robustness assessment of regressions using cluster analysis typologies: a bootstrap procedure with application in state sequence analysis.","authors":"Leonard Roth, Matthias Studer, Emilie Zuercher, Isabelle Peytremann-Bridevaux","doi":"10.1186/s12874-024-02435-8","DOIUrl":"10.1186/s12874-024-02435-8","url":null,"abstract":"<p><strong>Background: </strong>In standard Sequence Analysis, similar trajectories are clustered together to create a typology of trajectories, which is then often used to evaluate the association between sequence patterns and covariates inside regression models. The sampling uncertainty, which affects both the derivation of the typology and the associated regressions, is typically ignored in this analysis, an oversight that may lead to wrong statistical conclusions. We propose utilising sampling variation to derive new estimates that further inform on the association of interest.</p><p><strong>Methods: </strong>We introduce a novel procedure to assess the robustness of regression results obtained from the standard analysis. Bootstrap samples are drawn from the data, and for each bootstrap, a new typology replicating the original one is constructed, followed by the estimation of the corresponding regression models. The bootstrap estimates are then combined using a multilevel modelling framework that mimics a meta-analysis. The fitted values from this multilevel model allow to account for the sampling uncertainty in the inferential analysis. We illustrate the methodology by applying it to the study of healthcare utilisation trajectories in a Swiss cohort of diabetic patients.</p><p><strong>Results: </strong>The procedure provides robust estimates for an association of interest, along with 95% prediction intervals, representing the range of expected values if the clustering and associated regressions were performed on a new sample from the same underlying distribution. It also identifies central and borderline trajectories within each cluster. Regarding the illustrative application, while there was evidence of an association between regular lipid testing and subsequent healthcare utilisation patterns in the original analysis, this is not supported in the robustness assessment.</p><p><strong>Conclusions: </strong>Investigating the relationship between trajectory patterns and covariates is of interest in many situations. However, it is a challenging task with potential pitfalls. Our Robustness Assessment of Regression using Cluster Analysis Typologies (RARCAT) may assist in ensuring the robustness of such association studies. The method is applicable wherever clustering is combined with regression analysis, so its relevance goes beyond State Sequence Analysis.</p>","PeriodicalId":9114,"journal":{"name":"BMC Medical Research Methodology","volume":"24 1","pages":"303"},"PeriodicalIF":3.9,"publicationDate":"2024-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11654397/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142852570","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-12-18DOI: 10.1186/s12874-024-02414-z
Dawei Yin, Mikaela V Engracia, Matthew K Edema, David C Clarke
Background: A barrier to evidence-informed exercise programming is locating studies of exercise training programs. The purpose of this study was to create a search filter for studies of exercise training programs for the PubMed electronic bibliographic database.
Methods: Candidate search terms were identified from three sources: exercise-relevant MeSH terms and their corresponding Entry terms, word frequency analysis of articles in a gold-standard reference set curated from systematic reviews focused on exercise training, and retrospective searching of articles retrieved in the search filter development and testing steps. These terms were assembled into an exercise training search filter, and its performance was assessed against a basic search string applied to six case studies. Search string performance was measured as sensitivity (relative recall), precision, and number needed to read (NNR). We aimed to achieve relative recall ≥ 85%, and a NNR ≥ 2.
Results: The reference set consisted of 71 articles drawn from six systematic reviews. Sixty-one candidate search terms were evaluated for inclusion, 21 of which were included in the finalized exercise-training search filter. The relative recall of the search filter was 96% for the reference set and the precision mean ± SD was 54 ± 16% across the case studies, with the corresponding NNR = ~ 2. The exercise training search filter consistently outperformed the basic search string.
Conclusion: The exercise training search filter fosters more efficient searches for studies of exercise training programs in the PubMed electronic bibliographic database. This search string may therefore support evidence-informed practice in exercise programming.
{"title":"A PubMed search filter for efficiently retrieving exercise training studies.","authors":"Dawei Yin, Mikaela V Engracia, Matthew K Edema, David C Clarke","doi":"10.1186/s12874-024-02414-z","DOIUrl":"10.1186/s12874-024-02414-z","url":null,"abstract":"<p><strong>Background: </strong>A barrier to evidence-informed exercise programming is locating studies of exercise training programs. The purpose of this study was to create a search filter for studies of exercise training programs for the PubMed electronic bibliographic database.</p><p><strong>Methods: </strong>Candidate search terms were identified from three sources: exercise-relevant MeSH terms and their corresponding Entry terms, word frequency analysis of articles in a gold-standard reference set curated from systematic reviews focused on exercise training, and retrospective searching of articles retrieved in the search filter development and testing steps. These terms were assembled into an exercise training search filter, and its performance was assessed against a basic search string applied to six case studies. Search string performance was measured as sensitivity (relative recall), precision, and number needed to read (NNR). We aimed to achieve relative recall ≥ 85%, and a NNR ≥ 2.</p><p><strong>Results: </strong>The reference set consisted of 71 articles drawn from six systematic reviews. Sixty-one candidate search terms were evaluated for inclusion, 21 of which were included in the finalized exercise-training search filter. The relative recall of the search filter was 96% for the reference set and the precision mean ± SD was 54 ± 16% across the case studies, with the corresponding NNR = ~ 2. The exercise training search filter consistently outperformed the basic search string.</p><p><strong>Conclusion: </strong>The exercise training search filter fosters more efficient searches for studies of exercise training programs in the PubMed electronic bibliographic database. This search string may therefore support evidence-informed practice in exercise programming.</p>","PeriodicalId":9114,"journal":{"name":"BMC Medical Research Methodology","volume":"24 1","pages":"302"},"PeriodicalIF":3.9,"publicationDate":"2024-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11653657/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142851945","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-12-18DOI: 10.1186/s12874-024-02427-8
Pinyan Liu, Han Yuan, Yilin Ning, Bibhas Chakraborty, Nan Liu, Marco Aurélio Peres
Background: Traditional clustering techniques are typically restricted to either continuous or categorical variables. However, most real-world clinical data are mixed type. This study aims to introduce a clustering technique specifically designed for datasets containing both continuous and categorical variables to offer better clustering compatibility, adaptability, and interpretability than other mixed type techniques.
Methods: This paper proposed a modified Gower distance incorporating feature importance as weights to maintain equal contributions between continuous and categorical features. The algorithm (DAFI) was evaluated using five simulated datasets with varying proportions of important features and real-world datasets from the 2011-2014 National Health and Nutrition Examination Survey (NHANES). Effectiveness was demonstrated through comparisons with 13 clustering techniques. Clustering performance was assessed using the adjusted Rand index (ARI) for accuracy in simulation studies and the silhouette score for cohesion and separation in NHANES. Additionally, multivariable logistic regression estimated the association between periodontitis (PD) and cardiovascular diseases (CVDs), adjusting for clusters in NHANES.
Results: In simulation studies, the DAFI-Gower algorithm consistently performs better than baseline methods according to the adjusted Rand index in settings investigated, especially on datasets with more redundant features. In NHANES, 3,760 people were analyzed. DAFI-Gower achieves the highest silhouette score (0.79). Four distinct clusters with diverse health profiles were identified. By incorporating feature importance, we found that cluster formations were more strongly influenced by CVD-related factors. The association between periodontitis and cardiovascular diseases, after adjusting for clusters, reveals significant insights (adjusted OR 1.95, 95% CI 1.50 to 2.55, p = 0.012), highlighting severe periodontitis as a potential risk factor for cardiovascular diseases.
Conclusions: DAFI performed better than classic clustering baselines on both simulated and real-world datasets. It effectively captures cluster characteristics by considering feature importance, which is crucial in clinical settings where many variables may be similar or irrelevant. We envisage that DAFI offers an effective solution for mixed type clustering.
背景:传统的聚类技术通常局限于连续变量或分类变量。然而,大多数现实世界的临床数据是混合型的。本研究旨在引入一种专门为包含连续变量和分类变量的数据集设计的聚类技术,以提供比其他混合类型技术更好的聚类兼容性、适应性和可解释性。方法:本文提出了一种将特征重要度作为权重的改进高尔距离,以保持连续特征和分类特征之间的贡献相等。该算法(DAFI)使用五个具有不同比例重要特征的模拟数据集和来自2011-2014年国家健康与营养检查调查(NHANES)的真实数据集进行评估。通过与13种聚类技术的比较,证明了有效性。在模拟研究中使用调整后的兰德指数(ARI)来评估聚类性能的准确性,在NHANES中使用剪影评分来评估聚类和分离。此外,多变量logistic回归估计了牙周炎(PD)和心血管疾病(cvd)之间的关联,调整了NHANES中的聚类。结果:在模拟研究中,在调查的设置中,根据调整后的Rand指数,DAFI-Gower算法始终优于基线方法,特别是在冗余特征较多的数据集上。NHANES对3760人进行了分析。DAFI-Gower的剪影评分最高(0.79)。确定了四个具有不同健康概况的不同群集。结合特征重要性,我们发现簇的形成更强烈地受到cvd相关因素的影响。在调整聚类后,牙周炎和心血管疾病之间的关联揭示了重要的洞察力(调整OR 1.95, 95% CI 1.50至2.55,p = 0.012),突出了严重的牙周炎是心血管疾病的潜在危险因素。结论:DAFI在模拟和真实数据集上的表现都优于经典聚类基线。它通过考虑特征的重要性来有效地捕获集群特征,这在许多变量可能相似或不相关的临床环境中至关重要。我们设想DAFI为混合型聚类提供了一个有效的解决方案。
{"title":"A modified and weighted Gower distance-based clustering analysis for mixed type data: a simulation and empirical analyses.","authors":"Pinyan Liu, Han Yuan, Yilin Ning, Bibhas Chakraborty, Nan Liu, Marco Aurélio Peres","doi":"10.1186/s12874-024-02427-8","DOIUrl":"10.1186/s12874-024-02427-8","url":null,"abstract":"<p><strong>Background: </strong>Traditional clustering techniques are typically restricted to either continuous or categorical variables. However, most real-world clinical data are mixed type. This study aims to introduce a clustering technique specifically designed for datasets containing both continuous and categorical variables to offer better clustering compatibility, adaptability, and interpretability than other mixed type techniques.</p><p><strong>Methods: </strong>This paper proposed a modified Gower distance incorporating feature importance as weights to maintain equal contributions between continuous and categorical features. The algorithm (DAFI) was evaluated using five simulated datasets with varying proportions of important features and real-world datasets from the 2011-2014 National Health and Nutrition Examination Survey (NHANES). Effectiveness was demonstrated through comparisons with 13 clustering techniques. Clustering performance was assessed using the adjusted Rand index (ARI) for accuracy in simulation studies and the silhouette score for cohesion and separation in NHANES. Additionally, multivariable logistic regression estimated the association between periodontitis (PD) and cardiovascular diseases (CVDs), adjusting for clusters in NHANES.</p><p><strong>Results: </strong>In simulation studies, the DAFI-Gower algorithm consistently performs better than baseline methods according to the adjusted Rand index in settings investigated, especially on datasets with more redundant features. In NHANES, 3,760 people were analyzed. DAFI-Gower achieves the highest silhouette score (0.79). Four distinct clusters with diverse health profiles were identified. By incorporating feature importance, we found that cluster formations were more strongly influenced by CVD-related factors. The association between periodontitis and cardiovascular diseases, after adjusting for clusters, reveals significant insights (adjusted OR 1.95, 95% CI 1.50 to 2.55, p = 0.012), highlighting severe periodontitis as a potential risk factor for cardiovascular diseases.</p><p><strong>Conclusions: </strong>DAFI performed better than classic clustering baselines on both simulated and real-world datasets. It effectively captures cluster characteristics by considering feature importance, which is crucial in clinical settings where many variables may be similar or irrelevant. We envisage that DAFI offers an effective solution for mixed type clustering.</p>","PeriodicalId":9114,"journal":{"name":"BMC Medical Research Methodology","volume":"24 1","pages":"305"},"PeriodicalIF":3.9,"publicationDate":"2024-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11654179/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142852688","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-12-18DOI: 10.1186/s12874-024-02424-x
Maryam Y Garza, Tremaine B Williams, Songthip Ounpraseuth, Zhuopei Hu, Jeannette Lee, Jessica Snowden, Anita C Walden, Alan E Simon, Lori A Devlin, Leslie W Young, Meredith N Zozus
Background: Medical record abstraction (MRA) is a commonly used method for data collection in clinical research, but is prone to error, and the influence of quality control (QC) measures is seldom and inconsistently assessed during the course of a study. We employed a novel, standardized MRA-QC framework as part of an ongoing observational study in an effort to control MRA error rates. In order to assess the effectiveness of our framework, we compared our error rates against traditional MRA studies that had not reported using formalized MRA-QC methods. Thus, the objective of this study was to compare the MRA error rates derived from the literature with the error rates found in a study using MRA as the sole method of data collection that employed an MRA-QC framework.
Methods: A comparison of the error rates derived from MRA-centric studies identified as part of a systematic literature review was conducted against those derived from an MRA-centric study that employed an MRA-QC framework to evaluate the effectiveness of the MRA-QC framework. An inverse variance-weighted meta-analytical method with Freeman-Tukey transformation was used to compute pooled effect size for both the MRA studies identified in the literature and the study that implemented the MRA-QC framework. The level of heterogeneity was assessed using the Q-statistic and Higgins and Thompson's I2 statistic.
Results: The overall error rate from the MRA literature was 6.57%. Error rates for the study using our MRA-QC framework were between 1.04% (optimistic, all-field rate) and 2.57% (conservative, populated-field rate), 4.00-5.53% points less than the observed rate from the literature (p < 0.0001).
Conclusions: Review of the literature indicated that the accuracy associated with MRA varied widely across studies. However, our results demonstrate that, with appropriate training and continuous QC, MRA error rates can be significantly controlled during the course of a clinical research study.
{"title":"Comparing Medical Record Abstraction (MRA) error rates in an observational study to pooled rates identified in the data quality literature.","authors":"Maryam Y Garza, Tremaine B Williams, Songthip Ounpraseuth, Zhuopei Hu, Jeannette Lee, Jessica Snowden, Anita C Walden, Alan E Simon, Lori A Devlin, Leslie W Young, Meredith N Zozus","doi":"10.1186/s12874-024-02424-x","DOIUrl":"10.1186/s12874-024-02424-x","url":null,"abstract":"<p><strong>Background: </strong>Medical record abstraction (MRA) is a commonly used method for data collection in clinical research, but is prone to error, and the influence of quality control (QC) measures is seldom and inconsistently assessed during the course of a study. We employed a novel, standardized MRA-QC framework as part of an ongoing observational study in an effort to control MRA error rates. In order to assess the effectiveness of our framework, we compared our error rates against traditional MRA studies that had not reported using formalized MRA-QC methods. Thus, the objective of this study was to compare the MRA error rates derived from the literature with the error rates found in a study using MRA as the sole method of data collection that employed an MRA-QC framework.</p><p><strong>Methods: </strong>A comparison of the error rates derived from MRA-centric studies identified as part of a systematic literature review was conducted against those derived from an MRA-centric study that employed an MRA-QC framework to evaluate the effectiveness of the MRA-QC framework. An inverse variance-weighted meta-analytical method with Freeman-Tukey transformation was used to compute pooled effect size for both the MRA studies identified in the literature and the study that implemented the MRA-QC framework. The level of heterogeneity was assessed using the Q-statistic and Higgins and Thompson's I<sup>2</sup> statistic.</p><p><strong>Results: </strong>The overall error rate from the MRA literature was 6.57%. Error rates for the study using our MRA-QC framework were between 1.04% (optimistic, all-field rate) and 2.57% (conservative, populated-field rate), 4.00-5.53% points less than the observed rate from the literature (p < 0.0001).</p><p><strong>Conclusions: </strong>Review of the literature indicated that the accuracy associated with MRA varied widely across studies. However, our results demonstrate that, with appropriate training and continuous QC, MRA error rates can be significantly controlled during the course of a clinical research study.</p>","PeriodicalId":9114,"journal":{"name":"BMC Medical Research Methodology","volume":"24 1","pages":"304"},"PeriodicalIF":3.9,"publicationDate":"2024-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11653794/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142852495","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-12-16DOI: 10.1186/s12874-024-02436-7
Loukia M Spineli
Background: Transitivity assumption is the cornerstone of network meta-analysis (NMA). Investigating the plausibility of transitivity can unveil the credibility of NMA results. The commonness of transitivity was examined based on study dissimilarities regarding several study-level aggregate clinical and methodological characteristics reported in the systematic reviews. The present study also demonstrated the disadvantages of using multiple statistical tests to assess transitivity and compared the conclusions drawn from multiple statistical tests with those from the approach of study dissimilarities for transitivity assessment.
Methods: An empirical study was conducted using 209 published systematic reviews with NMA to create a database of study-level aggregate clinical and methodological characteristics found in the tracenma R package. For each systematic review, the network of the primary outcome was considered to create a dataset with extracted study-level aggregate clinical and methodological characteristics reported in the systematic review that may act as effect modifiers. Transitivity was evaluated by calculating study dissimilarities based on the extracted characteristics to provide a measure of overall dissimilarity within and between the observed treatment comparisons. Empirically driven thresholds of low dissimilarity were employed to determine the proportion of datasets with evidence of likely intransitivity. One-way ANOVA and chi-squared test were employed for each characteristic to investigate comparison dissimilarity at a significance level of 5%.
Results: Study dissimilarities covered a wide range of possible values across the datasets. A 'likely concerning' extent of study dissimilarities, both intra-comparison and inter-comparison, dominated the analysed datasets. Using a higher dissimilarity threshold, a 'likely concerning' extent of study dissimilarities persisted for objective outcomes but decreased substantially for subjective outcomes. A likely intransitivity prevailed in all datasets; however, using a higher dissimilarity threshold resulted in few networks with transitivity for semi-objective and subjective outcomes. Statistical tests were feasible in 127 (61%) datasets, yielding conflicting conclusions with the approach of study dissimilarities in many datasets.
Conclusions: Study dissimilarity, manifested from variations in the effect modifiers' distribution across the studies, should be expected and properly quantified. Measuring the overall study dissimilarity between observed comparisons and comparing it with a proper threshold can aid in determining whether concerns of likely intransitivity are warranted.
{"title":"An empirical study on 209 networks of treatments revealed intransitivity to be common and multiple statistical tests suboptimal to assess transitivity.","authors":"Loukia M Spineli","doi":"10.1186/s12874-024-02436-7","DOIUrl":"10.1186/s12874-024-02436-7","url":null,"abstract":"<p><strong>Background: </strong>Transitivity assumption is the cornerstone of network meta-analysis (NMA). Investigating the plausibility of transitivity can unveil the credibility of NMA results. The commonness of transitivity was examined based on study dissimilarities regarding several study-level aggregate clinical and methodological characteristics reported in the systematic reviews. The present study also demonstrated the disadvantages of using multiple statistical tests to assess transitivity and compared the conclusions drawn from multiple statistical tests with those from the approach of study dissimilarities for transitivity assessment.</p><p><strong>Methods: </strong>An empirical study was conducted using 209 published systematic reviews with NMA to create a database of study-level aggregate clinical and methodological characteristics found in the tracenma R package. For each systematic review, the network of the primary outcome was considered to create a dataset with extracted study-level aggregate clinical and methodological characteristics reported in the systematic review that may act as effect modifiers. Transitivity was evaluated by calculating study dissimilarities based on the extracted characteristics to provide a measure of overall dissimilarity within and between the observed treatment comparisons. Empirically driven thresholds of low dissimilarity were employed to determine the proportion of datasets with evidence of likely intransitivity. One-way ANOVA and chi-squared test were employed for each characteristic to investigate comparison dissimilarity at a significance level of 5%.</p><p><strong>Results: </strong>Study dissimilarities covered a wide range of possible values across the datasets. A 'likely concerning' extent of study dissimilarities, both intra-comparison and inter-comparison, dominated the analysed datasets. Using a higher dissimilarity threshold, a 'likely concerning' extent of study dissimilarities persisted for objective outcomes but decreased substantially for subjective outcomes. A likely intransitivity prevailed in all datasets; however, using a higher dissimilarity threshold resulted in few networks with transitivity for semi-objective and subjective outcomes. Statistical tests were feasible in 127 (61%) datasets, yielding conflicting conclusions with the approach of study dissimilarities in many datasets.</p><p><strong>Conclusions: </strong>Study dissimilarity, manifested from variations in the effect modifiers' distribution across the studies, should be expected and properly quantified. Measuring the overall study dissimilarity between observed comparisons and comparing it with a proper threshold can aid in determining whether concerns of likely intransitivity are warranted.</p>","PeriodicalId":9114,"journal":{"name":"BMC Medical Research Methodology","volume":"24 1","pages":"301"},"PeriodicalIF":3.9,"publicationDate":"2024-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11648297/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142833894","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-12-13DOI: 10.1186/s12874-024-02438-5
Abdelrahman M Makram, Randa Elsheikh, Omar M Makram, Nguyen Thanh Van, Nguyen Hai Nam, Nguyen Khoi Quan, Nguyen Tran Minh Duc, Ngoc Quynh Tram Nguyen, Gibson Omwansa Javes, Sara S Elsheikh, Atsuko Imoto, Peter Lee, Norio Ohmagari, Hirotsugu Aiga, Yasuhiko Kamiya, Patricia Takako Endo, Nguyen Tien Huy
{"title":"Correction: Tips from an expert panel on the development of a clinical research protocol.","authors":"Abdelrahman M Makram, Randa Elsheikh, Omar M Makram, Nguyen Thanh Van, Nguyen Hai Nam, Nguyen Khoi Quan, Nguyen Tran Minh Duc, Ngoc Quynh Tram Nguyen, Gibson Omwansa Javes, Sara S Elsheikh, Atsuko Imoto, Peter Lee, Norio Ohmagari, Hirotsugu Aiga, Yasuhiko Kamiya, Patricia Takako Endo, Nguyen Tien Huy","doi":"10.1186/s12874-024-02438-5","DOIUrl":"10.1186/s12874-024-02438-5","url":null,"abstract":"","PeriodicalId":9114,"journal":{"name":"BMC Medical Research Methodology","volume":"24 1","pages":"300"},"PeriodicalIF":3.9,"publicationDate":"2024-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11639107/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142817037","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}