Pub Date : 2025-11-25DOI: 10.1186/s41077-025-00394-8
Claudia Ebm, Cherrelle Smith, Manuela Milani, Mia Karamatsu, Nick Pokrajac, Bernard Dannenberg, Maurizio Cecconi
{"title":"Preparing Italian residents for global medical practice: the role of internationalization in education.","authors":"Claudia Ebm, Cherrelle Smith, Manuela Milani, Mia Karamatsu, Nick Pokrajac, Bernard Dannenberg, Maurizio Cecconi","doi":"10.1186/s41077-025-00394-8","DOIUrl":"10.1186/s41077-025-00394-8","url":null,"abstract":"","PeriodicalId":72108,"journal":{"name":"Advances in simulation (London, England)","volume":"10 1","pages":"61"},"PeriodicalIF":4.7,"publicationDate":"2025-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12645667/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145607754","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-24DOI: 10.1186/s41077-025-00381-z
Clément Buléon, Demian Szyld, Robert Simon, Lon Setnik, Walter J Eppich, Mary Fey, James A Lipshaw, Janice C Palaganas, Jenny W Rudolph
Background: Teaching and learning debriefing and feedback skills-especially to a level of mastery-is challenging without an agreed-upon standard. There are a number of rating scales and rubrics to identify and evaluate debriefing and feedback skills that focus on an entire feedback or debriefing conversation. However, there is no rubric to assess and provide feedback on one of these conversations' most widely used microskills, the Advocacy-Inquiry technique. This study aimed to develop and preliminarily test the Advocacy-Inquiry Rubric (AIR)-a tool designed to support the teaching, coaching, and assessment of Advocacy-Inquiry, a widely used yet challenging debriefing microskill-through an international expert consensus process.
Method: Using a four-round Delphi process, we achieved expert consensus on the behavioral markers of effective and ineffective Advocacy-Inquiry techniques. Thirty-nine experts from 13 countries identified and refined a set of key behavioral anchors for each of Advocacy-Inquiry's five elements: Preview, Observation, Point of View, Inquiry, and Listen. These descriptors were embedded first in a seven-point numeric Behaviorally Anchored Rating Scale, then in a three-point emoji-based version, and finally in a teaching and learning version. The AIR underwent two rounds of usability testing and inter-rater testing of the emoji version. Using an interpretation-use argument approach, evidence was collected for AIR's validity across scoring, generalization, extrapolation, and implication.
Results: The Delphi process established descriptors for each element of Advocacy-Inquiry, categorized by proficiency level (beginner to advanced). Usability testing enhanced the AIR's graphic layout to support both numeric ratings and formative feedback. The AIR was adapted into three tailored versions: a numeric AIR for detailed evaluation and progress tracking, an emoji AIR for peer assessment, and a teaching and learning AIR. Evidence for validity was assessed, highlighting both strengths and gaps.
Conclusion: AIR is an empirical rubric based on expert-derived criteria to support teaching, coaching, and assessing Advocacy-Inquiry microskills. The AIR offers a structured framework for self-, peer-, and mentor-led feedback and assessment to enhance a core skill of facilitators. By anchoring assessments in clear behavioral descriptors, the AIR aims to improve the quality of feedback and debriefing conversations. Future work should focus on rater training, reliability testing, and exploring the AIR's impact on real-world outcomes.
{"title":"The Advocacy-Inquiry Rubric (AIR): a standard to build debriefing and feedback skills.","authors":"Clément Buléon, Demian Szyld, Robert Simon, Lon Setnik, Walter J Eppich, Mary Fey, James A Lipshaw, Janice C Palaganas, Jenny W Rudolph","doi":"10.1186/s41077-025-00381-z","DOIUrl":"10.1186/s41077-025-00381-z","url":null,"abstract":"<p><strong>Background: </strong>Teaching and learning debriefing and feedback skills-especially to a level of mastery-is challenging without an agreed-upon standard. There are a number of rating scales and rubrics to identify and evaluate debriefing and feedback skills that focus on an entire feedback or debriefing conversation. However, there is no rubric to assess and provide feedback on one of these conversations' most widely used microskills, the Advocacy-Inquiry technique. This study aimed to develop and preliminarily test the Advocacy-Inquiry Rubric (AIR)-a tool designed to support the teaching, coaching, and assessment of Advocacy-Inquiry, a widely used yet challenging debriefing microskill-through an international expert consensus process.</p><p><strong>Method: </strong>Using a four-round Delphi process, we achieved expert consensus on the behavioral markers of effective and ineffective Advocacy-Inquiry techniques. Thirty-nine experts from 13 countries identified and refined a set of key behavioral anchors for each of Advocacy-Inquiry's five elements: Preview, Observation, Point of View, Inquiry, and Listen. These descriptors were embedded first in a seven-point numeric Behaviorally Anchored Rating Scale, then in a three-point emoji-based version, and finally in a teaching and learning version. The AIR underwent two rounds of usability testing and inter-rater testing of the emoji version. Using an interpretation-use argument approach, evidence was collected for AIR's validity across scoring, generalization, extrapolation, and implication.</p><p><strong>Results: </strong>The Delphi process established descriptors for each element of Advocacy-Inquiry, categorized by proficiency level (beginner to advanced). Usability testing enhanced the AIR's graphic layout to support both numeric ratings and formative feedback. The AIR was adapted into three tailored versions: a numeric AIR for detailed evaluation and progress tracking, an emoji AIR for peer assessment, and a teaching and learning AIR. Evidence for validity was assessed, highlighting both strengths and gaps.</p><p><strong>Conclusion: </strong>AIR is an empirical rubric based on expert-derived criteria to support teaching, coaching, and assessing Advocacy-Inquiry microskills. The AIR offers a structured framework for self-, peer-, and mentor-led feedback and assessment to enhance a core skill of facilitators. By anchoring assessments in clear behavioral descriptors, the AIR aims to improve the quality of feedback and debriefing conversations. Future work should focus on rater training, reliability testing, and exploring the AIR's impact on real-world outcomes.</p>","PeriodicalId":72108,"journal":{"name":"Advances in simulation (London, England)","volume":"10 1","pages":"60"},"PeriodicalIF":4.7,"publicationDate":"2025-11-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12645724/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145598001","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-24DOI: 10.1186/s41077-025-00386-8
Lotte Cools, Rani Van Schoors, Fien Depaepe, Eline Dancet, Nicolas Delvaux
Background: Simulation-based education is a well-established training technique in medical curricula, also for communication skills. Virtual reality (VR) technology can enhance this form of experience-based learning. How VR interacts with training communication skills for interpersonal and interprofessional medical encounters is, however, unclear. This study investigates how VR influences communication skills and behaviors in patient-student and team encounters in medical undergraduate simulations, in order to make recommendations for VR simulation-based communication skills training (CST).
Methods: We conducted a study with 22 third-year medical students completing a dyadic VR simulation (Smart Collaboration Tutor software). We coded communication skills and behaviors for team and patient-student communication in videorecorded VR simulations. We then analyzed communication patterns and finally developed themes for VR-mediated CST.
Results: Our findings revealed that students preferred the core communication skill of asking questions, informing, and thinking aloud as process communication skills in a VR simulation. Nonverbal and paraverbal behaviors were used with unclear intent. VR negatively impacted the focus of attention and flow of simulation-based communication skills training.
Discussion: Dyadic VR simulations tend to emphasize team and task-oriented communication. Its value for patient-student and relation-oriented communication is unclear. VR influenced conversational turn-taking by altering visual and auditory perceptions. Cognitive load was enhanced, potentially diverting attention from communication goals and observational focus.
Conclusion: Multiuser VR simulation shows certain possibilities for CST in medical undergraduate simulations. Recommendations on the contextual design of VR simulations, however, need to be taken into account to safeguard the focus of attention and flow of CST.
{"title":"Training communication skills in a multiuser medical virtual reality simulation: a qualitative, observational study.","authors":"Lotte Cools, Rani Van Schoors, Fien Depaepe, Eline Dancet, Nicolas Delvaux","doi":"10.1186/s41077-025-00386-8","DOIUrl":"10.1186/s41077-025-00386-8","url":null,"abstract":"<p><strong>Background: </strong>Simulation-based education is a well-established training technique in medical curricula, also for communication skills. Virtual reality (VR) technology can enhance this form of experience-based learning. How VR interacts with training communication skills for interpersonal and interprofessional medical encounters is, however, unclear. This study investigates how VR influences communication skills and behaviors in patient-student and team encounters in medical undergraduate simulations, in order to make recommendations for VR simulation-based communication skills training (CST).</p><p><strong>Methods: </strong>We conducted a study with 22 third-year medical students completing a dyadic VR simulation (Smart Collaboration Tutor software). We coded communication skills and behaviors for team and patient-student communication in videorecorded VR simulations. We then analyzed communication patterns and finally developed themes for VR-mediated CST.</p><p><strong>Results: </strong>Our findings revealed that students preferred the core communication skill of asking questions, informing, and thinking aloud as process communication skills in a VR simulation. Nonverbal and paraverbal behaviors were used with unclear intent. VR negatively impacted the focus of attention and flow of simulation-based communication skills training.</p><p><strong>Discussion: </strong>Dyadic VR simulations tend to emphasize team and task-oriented communication. Its value for patient-student and relation-oriented communication is unclear. VR influenced conversational turn-taking by altering visual and auditory perceptions. Cognitive load was enhanced, potentially diverting attention from communication goals and observational focus.</p><p><strong>Conclusion: </strong>Multiuser VR simulation shows certain possibilities for CST in medical undergraduate simulations. Recommendations on the contextual design of VR simulations, however, need to be taken into account to safeguard the focus of attention and flow of CST.</p>","PeriodicalId":72108,"journal":{"name":"Advances in simulation (London, England)","volume":"10 1","pages":"59"},"PeriodicalIF":4.7,"publicationDate":"2025-11-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12642034/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145598040","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-22DOI: 10.1186/s41077-025-00396-6
Adam Cheng, Yiqun Lin, Gabriel Reedy, Christine Joseph, Samantha Wirkowski, Viviane Mallette, Vikhashni Nagesh, David Krieser, Aaron Calhoun
Background: The increasing use of artificial intelligence (AI) by scholars presents a pressing challenge to healthcare publishing. While legitimate use can potentially accelerate scholarship, unethical approaches also exist, leading to factually inaccurate and biased text that may degrade scholarship. Numerous online AI detection tools exist that provide a percentage score of AI use. These can assist authors and editors in navigating this landscape. In this study, we compared the scores from three AI detection tools (ZeroGPT, PhraslyAI, and Grammarly AI Detector) across five plausible conditions of AI use and evaluated them against human assessments.
Methods: Thirty open access articles published in the journals Advances in Simulation and Simulation in Healthcare prior to 2022 were selected, and the article introductions were extracted. Five experimental conditions were examined, including: (1) 100% human written; (2) human written, light AI editing; (3) human written, heavy AI editing; (4) AI written text from human content; and (5) 100% AI written from article title. The resulting materials were assessed by three open-access AI detection tools and five blinded human raters. Results were summarized descriptively and compared using repeated measures analysis of variance (ANOVA), intraclass correlation coefficients (ICC), and Bland-Altman plots.
Results: The three AI detection tools were able to differentiate between the five test conditions (p < 0.001 for all), but varied significantly in absolute score, with ICC ranging from 0.57 to 0.95, raising concerns regarding overall reliability of these tools. Human scoring was far less consistent, with an overall accuracy of 19%, indistinguishable from chance.
Conclusion: While existing AI detection tools can meaningfully distinguish plausible AI use conditions, reliability across these tools is variable. Human scoring accuracy is uniformly low. Use of AI detection tools by scholars and journal editors may assist in determining potentially unethical use but they should not be relied upon alone at this time.
背景:越来越多的学者使用人工智能(AI)对医疗保健出版提出了紧迫的挑战。虽然合法使用可以潜在地加速学术研究,但不道德的方法也存在,导致事实不准确和有偏见的文本可能会降低学术研究。有许多在线人工智能检测工具可以提供人工智能使用的百分比分数。这些可以帮助作者和编辑导航这一景观。在这项研究中,我们比较了三种人工智能检测工具(ZeroGPT、PhraslyAI和Grammarly AI Detector)在五种可能的人工智能使用条件下的得分,并将它们与人类的评估进行了比较。方法:选取2022年前发表在《Advances in Simulation》和《Simulation in Healthcare》期刊上的开放获取文章30篇,提取文章介绍。研究了五种实验条件,包括:(1)100%人工书写;(2)人工书写,轻AI编辑;(3)人工编写,人工智能大量编辑;(4)人工智能文字来源于人类内容;(5) 100%人工智能从文章标题中编写。所得材料由三种开放获取的人工智能检测工具和五名盲法人类评分员进行评估。对结果进行描述性总结,并使用重复测量方差分析(ANOVA)、类内相关系数(ICC)和Bland-Altman图进行比较。结果:三种人工智能检测工具能够区分五种测试条件(p结论:虽然现有的人工智能检测工具可以有意义地区分合理的人工智能使用条件,但这些工具的可靠性是可变的。人类评分的准确率普遍较低。学者和期刊编辑使用人工智能检测工具可能有助于确定潜在的不道德使用,但目前不应单独依赖它们。
{"title":"Ability of AI detection tools and humans to accurately identify different forms of AI-generated written content.","authors":"Adam Cheng, Yiqun Lin, Gabriel Reedy, Christine Joseph, Samantha Wirkowski, Viviane Mallette, Vikhashni Nagesh, David Krieser, Aaron Calhoun","doi":"10.1186/s41077-025-00396-6","DOIUrl":"10.1186/s41077-025-00396-6","url":null,"abstract":"<p><strong>Background: </strong>The increasing use of artificial intelligence (AI) by scholars presents a pressing challenge to healthcare publishing. While legitimate use can potentially accelerate scholarship, unethical approaches also exist, leading to factually inaccurate and biased text that may degrade scholarship. Numerous online AI detection tools exist that provide a percentage score of AI use. These can assist authors and editors in navigating this landscape. In this study, we compared the scores from three AI detection tools (ZeroGPT, PhraslyAI, and Grammarly AI Detector) across five plausible conditions of AI use and evaluated them against human assessments.</p><p><strong>Methods: </strong>Thirty open access articles published in the journals Advances in Simulation and Simulation in Healthcare prior to 2022 were selected, and the article introductions were extracted. Five experimental conditions were examined, including: (1) 100% human written; (2) human written, light AI editing; (3) human written, heavy AI editing; (4) AI written text from human content; and (5) 100% AI written from article title. The resulting materials were assessed by three open-access AI detection tools and five blinded human raters. Results were summarized descriptively and compared using repeated measures analysis of variance (ANOVA), intraclass correlation coefficients (ICC), and Bland-Altman plots.</p><p><strong>Results: </strong>The three AI detection tools were able to differentiate between the five test conditions (p < 0.001 for all), but varied significantly in absolute score, with ICC ranging from 0.57 to 0.95, raising concerns regarding overall reliability of these tools. Human scoring was far less consistent, with an overall accuracy of 19%, indistinguishable from chance.</p><p><strong>Conclusion: </strong>While existing AI detection tools can meaningfully distinguish plausible AI use conditions, reliability across these tools is variable. Human scoring accuracy is uniformly low. Use of AI detection tools by scholars and journal editors may assist in determining potentially unethical use but they should not be relied upon alone at this time.</p>","PeriodicalId":72108,"journal":{"name":"Advances in simulation (London, England)","volume":" ","pages":"66"},"PeriodicalIF":4.7,"publicationDate":"2025-11-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12752165/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145574975","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-18DOI: 10.1186/s41077-025-00385-9
Miriam Veenhuizen, Andrew O'Malley
Background: Generative artificial intelligence models are being introduced as low-cost tools for creating simulated patient cohorts in undergraduate medical education. Their educational value, however, depends on the extent to which the synthetic populations mirror real-world demographic diversity. We therefore assessed whether two commonly deployed large language models produce patient profiles that reflect the current age, sex, and ethnic composition of the UK.
Methods: GPT-3.5-turbo-0125 and GPT-4-mini-2024-07-18 were each prompted, without demographic steering, to generate 250 UK-based 'patients'. Age was returned directly by the model; sex and ethnicity were inferred from given and family names using a validated census-derived classifier. Observed frequencies for each demographic variable were compared with England and Wales 2021 census expectations by chi-square goodness-of-fit tests.
Results: Both cohorts diverged significantly from census benchmarks (p < 0.0001 for every variable). Age distributions showed an absence of very young and older individuals, with certain middle-aged groups overrepresented (GPT-3.5: χ2(17) = 1310.4, p < 0.0001; GPT4mini: χ2(17) = 1866.1, p < 0.0001). Neither model produced patients younger than 25 years; GPT-3.5 generated no one older than 47 years and GPT-4-mini no one older than 56 years. Gender proportions also differed markedly, skewing heavily toward males (GPT-3.5: χ2(1) = 23.84, p < 0.0001; GPT4mini: χ2(1) = 191.7, p < 0.0001). Male patients constituted 64.7% and 92.8% of the two cohorts. Name diversity was limited: GPT-3.5 yielded 104 unique first-last-name combinations, whereas GPT-4-mini produced only nine. Ethnic profiles were similarly imbalanced, featuring overrepresentation of some groups and complete absence of others (χ2(10) = 42.19, p < 0.0001).
Conclusions: In their default state, the evaluated models create synthetic patient pools that exclude younger, older, female and most minority-ethnic representations. Such demographically narrow outputs threaten to normalise biased clinical expectations and may undermine efforts to prepare students for equitable practice. Baseline auditing of model behaviour is therefore essential, providing a benchmark against which prompt-engineering or data-curation strategies can be evaluated before generative systems are integrated into formal curricula.
{"title":"Demographic biases in AI-generated simulated patient cohorts: a comparative analysis against census benchmarks.","authors":"Miriam Veenhuizen, Andrew O'Malley","doi":"10.1186/s41077-025-00385-9","DOIUrl":"10.1186/s41077-025-00385-9","url":null,"abstract":"<p><strong>Background: </strong>Generative artificial intelligence models are being introduced as low-cost tools for creating simulated patient cohorts in undergraduate medical education. Their educational value, however, depends on the extent to which the synthetic populations mirror real-world demographic diversity. We therefore assessed whether two commonly deployed large language models produce patient profiles that reflect the current age, sex, and ethnic composition of the UK.</p><p><strong>Methods: </strong>GPT-3.5-turbo-0125 and GPT-4-mini-2024-07-18 were each prompted, without demographic steering, to generate 250 UK-based 'patients'. Age was returned directly by the model; sex and ethnicity were inferred from given and family names using a validated census-derived classifier. Observed frequencies for each demographic variable were compared with England and Wales 2021 census expectations by chi-square goodness-of-fit tests.</p><p><strong>Results: </strong>Both cohorts diverged significantly from census benchmarks (p < 0.0001 for every variable). Age distributions showed an absence of very young and older individuals, with certain middle-aged groups overrepresented (GPT-3.5: χ2(17) = 1310.4, p < 0.0001; GPT4mini: χ2(17) = 1866.1, p < 0.0001). Neither model produced patients younger than 25 years; GPT-3.5 generated no one older than 47 years and GPT-4-mini no one older than 56 years. Gender proportions also differed markedly, skewing heavily toward males (GPT-3.5: χ2(1) = 23.84, p < 0.0001; GPT4mini: χ2(1) = 191.7, p < 0.0001). Male patients constituted 64.7% and 92.8% of the two cohorts. Name diversity was limited: GPT-3.5 yielded 104 unique first-last-name combinations, whereas GPT-4-mini produced only nine. Ethnic profiles were similarly imbalanced, featuring overrepresentation of some groups and complete absence of others (χ2(10) = 42.19, p < 0.0001).</p><p><strong>Conclusions: </strong>In their default state, the evaluated models create synthetic patient pools that exclude younger, older, female and most minority-ethnic representations. Such demographically narrow outputs threaten to normalise biased clinical expectations and may undermine efforts to prepare students for equitable practice. Baseline auditing of model behaviour is therefore essential, providing a benchmark against which prompt-engineering or data-curation strategies can be evaluated before generative systems are integrated into formal curricula.</p>","PeriodicalId":72108,"journal":{"name":"Advances in simulation (London, England)","volume":"10 1","pages":"58"},"PeriodicalIF":4.7,"publicationDate":"2025-11-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12625206/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145552096","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-18DOI: 10.1186/s41077-025-00390-y
Steven Bordonaro, Christopher Negro, Karl Neubecker, Eric C Nemec, Suzanne J Rose
Background: A large portion of preventable deaths is a result of uncontrolled bleeding due to a delay in medical intervention. While publicly accessible tourniquets raise the concern of incorrect application by laypeople, tourniquets have proven efficacy and can be effectively applied by bystanders. This systematic review aims to identify if tourniquets applied by laypeople using a basic manikin or tourniquet trainer extremity with little to no training can effectively control bleeding.
Methods: The authors used EBSCOHost to simultaneously search the following databases: Cumulated Index in Nursing and Allied Health Literature (CINAHL) Ultimate, Academic Search Premier, Cochrane Central Register of Controlled Trials, Cochrane Database of Systematic Reviews, and Medical Literature Analysis and Retrieval System Online (MEDLINE) with Full Text. Boolean search strategy included tourniquet AND (layperson OR laypeople) AND ((bleeding AND control) OR (hemorrhage AND control) OR "stop the bleed") NOT surgery. The search was limited to January 1, 2013, to August 31, 2023. Inclusion criteria were layperson participants in peer-reviewed randomized controlled or clinical trials, available in English, that assessed at least one outcome measure related to the efficacy of tourniquet application in a simulated context. Articles including duplicate data and those regarding tourniquet use/efficacy in settings other than prehospital care or bleeding control were excluded. Two independent reviewers selected studies according to prespecified inclusion and exclusion criteria. Risk of bias was assessed using the Cochrane RoB 2 tool.
Results: The initial search identified 83 studies, with 10 retained for inclusion in this review. Two different windlass rod tourniquets and one ratcheting strap tourniquet performed the best in terms of successful application by laypeople. Completing formal bleeding control training increased the average application success rate compared to no prior training. The Layperson Audiovisual Assist Tourniquet was the only audiovisual point-of-care aid that significantly increased the rate of successful applications. Just-in-Time visual cards also increased success rates significantly, showing comparable benefits to manufacturer instructions.
Conclusion: Although some laypeople can successfully place tourniquets without prior training, successful placement rates can be improved with point-of-care aids and formal bleeding control training using a basic manikin or tourniquet trainer extremity.
背景:很大一部分可预防的死亡是由于医疗干预延误而导致出血失控的结果。虽然公众可获得的止血带引起了外行人不正确使用的担忧,但止血带已经被证明是有效的,并且可以由旁观者有效地使用。本系统综述旨在确定外行人使用基本人体模型或止血带训练器进行的止血带是否可以有效地控制出血。方法:作者使用EBSCOHost同时检索以下数据库:护理与相关健康文献累积索引(CINAHL) Ultimate、学术检索Premier、Cochrane对照试验中央注册库、Cochrane系统评价数据库和医学文献分析与检索系统(MEDLINE)全文数据库。布尔搜索策略包括止血带和(外行人或外行人)和((出血和控制)或(出血和控制)或“止血”),而不是手术。搜索范围限于2013年1月1日至2023年8月31日。纳入标准是同行评议的随机对照或临床试验的外行参与者,可获得英文版本,评估至少一项与模拟环境中止血带应用效果相关的结果测量。包括重复数据的文章和关于院前护理或出血控制以外环境中止血带使用/疗效的文章被排除在外。两名独立审稿人根据预先指定的纳入和排除标准选择研究。使用Cochrane RoB 2工具评估偏倚风险。结果:最初的检索确定了83项研究,其中10项保留纳入本综述。两种不同的卷绕杆止血带和一种棘轮带止血带在外行人成功应用方面表现最好。完成正规的出血控制培训与没有事先培训相比,增加了平均应用成功率。外行人视听辅助止血带是唯一的视听护理点辅助,显着增加了成功率的应用。即时可视化卡片也显著提高了成功率,显示出与制造商说明相当的好处。结论:虽然一些外行人可以在没有事先训练的情况下成功放置止血带,但使用基本的人体模型或止血带训练器进行即时护理辅助和正式的出血控制训练可以提高放置成功率。
{"title":"Efficacy of publicly accessible tourniquets: a systematic review of layperson performance utilizing simulation models.","authors":"Steven Bordonaro, Christopher Negro, Karl Neubecker, Eric C Nemec, Suzanne J Rose","doi":"10.1186/s41077-025-00390-y","DOIUrl":"10.1186/s41077-025-00390-y","url":null,"abstract":"<p><strong>Background: </strong>A large portion of preventable deaths is a result of uncontrolled bleeding due to a delay in medical intervention. While publicly accessible tourniquets raise the concern of incorrect application by laypeople, tourniquets have proven efficacy and can be effectively applied by bystanders. This systematic review aims to identify if tourniquets applied by laypeople using a basic manikin or tourniquet trainer extremity with little to no training can effectively control bleeding.</p><p><strong>Methods: </strong>The authors used EBSCOHost to simultaneously search the following databases: Cumulated Index in Nursing and Allied Health Literature (CINAHL) Ultimate, Academic Search Premier, Cochrane Central Register of Controlled Trials, Cochrane Database of Systematic Reviews, and Medical Literature Analysis and Retrieval System Online (MEDLINE) with Full Text. Boolean search strategy included tourniquet AND (layperson OR laypeople) AND ((bleeding AND control) OR (hemorrhage AND control) OR \"stop the bleed\") NOT surgery. The search was limited to January 1, 2013, to August 31, 2023. Inclusion criteria were layperson participants in peer-reviewed randomized controlled or clinical trials, available in English, that assessed at least one outcome measure related to the efficacy of tourniquet application in a simulated context. Articles including duplicate data and those regarding tourniquet use/efficacy in settings other than prehospital care or bleeding control were excluded. Two independent reviewers selected studies according to prespecified inclusion and exclusion criteria. Risk of bias was assessed using the Cochrane RoB 2 tool.</p><p><strong>Results: </strong>The initial search identified 83 studies, with 10 retained for inclusion in this review. Two different windlass rod tourniquets and one ratcheting strap tourniquet performed the best in terms of successful application by laypeople. Completing formal bleeding control training increased the average application success rate compared to no prior training. The Layperson Audiovisual Assist Tourniquet was the only audiovisual point-of-care aid that significantly increased the rate of successful applications. Just-in-Time visual cards also increased success rates significantly, showing comparable benefits to manufacturer instructions.</p><p><strong>Conclusion: </strong>Although some laypeople can successfully place tourniquets without prior training, successful placement rates can be improved with point-of-care aids and formal bleeding control training using a basic manikin or tourniquet trainer extremity.</p>","PeriodicalId":72108,"journal":{"name":"Advances in simulation (London, England)","volume":"10 1","pages":"57"},"PeriodicalIF":4.7,"publicationDate":"2025-11-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12625582/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145552045","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-14DOI: 10.1186/s41077-025-00387-7
Kjetil Torgeirsen, Benjamin Kamala, Estomih Mduma, Florence Salvatory Kalabamu, Robert Moshiro, Doris Østergaard, Jan Terje Kvaløy, Hege Langli Ersdal
Background: Safer Births Bundle of Care (SBBC) is a continuous quality improvement (CQI) program, implemented in 30 facilities in Tanzania, resulting in a 75% reduction in maternal deaths. Simulation training was introduced as a component of the CQI efforts, targeting individual and team skills, focusing on identified clinical needs.
Objective: The aim of this study was to describe the frequency of documented simulation sessions and the number of recurrent participants and associations with changes in maternal death.
Methods: SBBC was a stepped-wedge cluster randomised implementation study in 30 facilities in 5 regions of Tanzania from 2020 through 2023. The SimBegin® facilitator training program was introduced to train facilitators and support implementation of a training cascade. Fifteen selected healthcare workers were trained in three levels of SimBegin® to become facilitators (level 1) and mentors (level 2). Eight were trained to become instructors (level 3). In total, 90 local facilitators were trained to review local clinical data, run simulation sessions, and document in logbooks. Clinical data were collected from patient files by independent data collectors and looped back to the facilities on a weekly basis. Training interventions were planned, conducted, and evaluated based on identified gaps. Output measures were the frequency of simulation sessions, the number of recurring participants, and maternal death within 7 days postpartum the following month.
Results: Overall, 281,165 parturient women were included in this study. The SBBC implementation period was 24-32 months, and 1280 simulation sessions were documented. Maternal deaths declined from 240/100,000 births in the baseline to 60/100,000 after the start of SBBC. There was an association between the frequency of simulation sessions and the reduction in maternal deaths (23% reduction per each unit increase on the log scale, P = 0.0018), and between the number of recurring participants and the reduction in maternal deaths (16% reduction per each unit increase on the log scale, P = 0.0006).
Conclusion: This study documents a significant and clinically relevant association between the frequency of and participation in simulation sessions and the reduction of maternal deaths the following month.
{"title":"Frequency of team simulation and reduction in maternal deaths following Safer Births Bundle of Care implementation-a prospective observational study.","authors":"Kjetil Torgeirsen, Benjamin Kamala, Estomih Mduma, Florence Salvatory Kalabamu, Robert Moshiro, Doris Østergaard, Jan Terje Kvaløy, Hege Langli Ersdal","doi":"10.1186/s41077-025-00387-7","DOIUrl":"10.1186/s41077-025-00387-7","url":null,"abstract":"<p><strong>Background: </strong>Safer Births Bundle of Care (SBBC) is a continuous quality improvement (CQI) program, implemented in 30 facilities in Tanzania, resulting in a 75% reduction in maternal deaths. Simulation training was introduced as a component of the CQI efforts, targeting individual and team skills, focusing on identified clinical needs.</p><p><strong>Objective: </strong>The aim of this study was to describe the frequency of documented simulation sessions and the number of recurrent participants and associations with changes in maternal death.</p><p><strong>Methods: </strong>SBBC was a stepped-wedge cluster randomised implementation study in 30 facilities in 5 regions of Tanzania from 2020 through 2023. The SimBegin® facilitator training program was introduced to train facilitators and support implementation of a training cascade. Fifteen selected healthcare workers were trained in three levels of SimBegin® to become facilitators (level 1) and mentors (level 2). Eight were trained to become instructors (level 3). In total, 90 local facilitators were trained to review local clinical data, run simulation sessions, and document in logbooks. Clinical data were collected from patient files by independent data collectors and looped back to the facilities on a weekly basis. Training interventions were planned, conducted, and evaluated based on identified gaps. Output measures were the frequency of simulation sessions, the number of recurring participants, and maternal death within 7 days postpartum the following month.</p><p><strong>Results: </strong>Overall, 281,165 parturient women were included in this study. The SBBC implementation period was 24-32 months, and 1280 simulation sessions were documented. Maternal deaths declined from 240/100,000 births in the baseline to 60/100,000 after the start of SBBC. There was an association between the frequency of simulation sessions and the reduction in maternal deaths (23% reduction per each unit increase on the log scale, P = 0.0018), and between the number of recurring participants and the reduction in maternal deaths (16% reduction per each unit increase on the log scale, P = 0.0006).</p><p><strong>Conclusion: </strong>This study documents a significant and clinically relevant association between the frequency of and participation in simulation sessions and the reduction of maternal deaths the following month.</p><p><strong>Trial registration: </strong>SBBC main protocol ISRCTN Registry: ISRCTN30541755. Prospectively registered 12.10.2020.</p>","PeriodicalId":72108,"journal":{"name":"Advances in simulation (London, England)","volume":"10 1","pages":"56"},"PeriodicalIF":4.7,"publicationDate":"2025-11-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12619334/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145524946","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-11DOI: 10.1186/s41077-025-00389-5
Pablo F Miranda, Andrea L Araneda, Natalia P Molina, Felipe G Miranda, Christopher Morrison, Marcia A Corvetto, Fernando R Altermatt
Background: The following study aims to determine the learning curve experienced by anesthesia residents when training for an ultrasound-guided popliteal sciatic block and the transference of this training to real patient situations.
Methods: After approval by the ethics committee, eleven first-year anesthesia residents were recruited to participate in a simulation-based training program to perform a single shot in plane popliteal sciatic block. Training consisted of 10 individual sessions, with direct feedback from the instructor, with a specific Laerdal® popliteal sciatic block phantom, lasting one hour and distributed weekly. At the end of each session, the resident's performance was assessed. Residents were videotaped while performing the block, which was to be evaluated using a validated global rating scale (GRS). Additionally, a tracking motion device attached to the operator's hands (Imperial College Surgical Assessment Device, ICSAD) recorded the total distance traveled by both hands (Total path length, TPL), number of movements (NM), and total procedure time (TPT). One week later, the same assessment was done on a real patient.
Results: Ten residents completed the training and the assessments. Median values of GRS scores significantly improved from 15 to 28.3 through the training (p = 0.006). Regarding ICSAD scores, TPT improved from 126 to 63.4 s (p = 0.002), and TPL improved from 11.07 to 9.4 m (p = 0.322). When comparing the last simulated session and the subsequent measurement in an actual patient, median values of GRS, TPL and NM were not different.
Conclusions: This simulation-based training program significantly improved residents' proficiency in an ultrasound-guided popliteal sciatic block. The learning curve plateaued at session 7, and this improvement was transferred to the real patient setting. As expected, residents needed more time for the first block on a real patient than for the last simulated session.
背景:下面的研究旨在确定麻醉住院医师在超声引导下进行腘窝坐骨阻滞训练时所经历的学习曲线,并将这种训练转移到真实的患者情况中。方法:经伦理委员会批准,招募11名一年级麻醉住院医师参加基于模拟的培训计划,进行平面腘窝坐骨阻滞单次射击。训练包括10次单独的训练,由教练直接反馈,使用特定的Laerdal®腘窝坐骨神经阻滞假体,持续1小时,每周进行一次。在每个疗程结束时,对住院医生的表现进行评估。居民们在表演街区时被录像,并使用有效的全球评级量表(GRS)对其进行评估。此外,附着在操作者手上的跟踪运动装置(帝国理工学院手术评估装置,ICSAD)记录了双手移动的总距离(总路径长度,TPL)、运动次数(NM)和总手术时间(TPT)。一周后,对一位真正的病人进行了同样的评估。结果:10名住院医师完成了培训和评估。通过训练,GRS评分中位数由15提高到28.3,差异有统计学意义(p = 0.006)。ICSAD评分方面,TPT从126 s提高到63.4 s (p = 0.002), TPL从11.07 m提高到9.4 m (p = 0.322)。当比较最后一次模拟会话和随后对实际患者的测量时,GRS, TPL和NM的中位数没有差异。结论:这种基于模拟的培训方案显著提高了住院医生在超声引导下腘窝坐骨阻滞的熟练程度。学习曲线在第7期趋于平稳,这种改善被转移到真实的患者环境中。正如预期的那样,住院医生在真实病人身上的第一个街区比最后一个模拟疗程需要更多的时间。临床试验编号:ClinicalTrials.gov,标识符NCT06081790。
{"title":"Simulated-based training for ultrasound-guided popliteal sciatic nerve block: determining the learning curve and transference to real patient.","authors":"Pablo F Miranda, Andrea L Araneda, Natalia P Molina, Felipe G Miranda, Christopher Morrison, Marcia A Corvetto, Fernando R Altermatt","doi":"10.1186/s41077-025-00389-5","DOIUrl":"10.1186/s41077-025-00389-5","url":null,"abstract":"<p><strong>Background: </strong>The following study aims to determine the learning curve experienced by anesthesia residents when training for an ultrasound-guided popliteal sciatic block and the transference of this training to real patient situations.</p><p><strong>Methods: </strong>After approval by the ethics committee, eleven first-year anesthesia residents were recruited to participate in a simulation-based training program to perform a single shot in plane popliteal sciatic block. Training consisted of 10 individual sessions, with direct feedback from the instructor, with a specific Laerdal® popliteal sciatic block phantom, lasting one hour and distributed weekly. At the end of each session, the resident's performance was assessed. Residents were videotaped while performing the block, which was to be evaluated using a validated global rating scale (GRS). Additionally, a tracking motion device attached to the operator's hands (Imperial College Surgical Assessment Device, ICSAD) recorded the total distance traveled by both hands (Total path length, TPL), number of movements (NM), and total procedure time (TPT). One week later, the same assessment was done on a real patient.</p><p><strong>Results: </strong>Ten residents completed the training and the assessments. Median values of GRS scores significantly improved from 15 to 28.3 through the training (p = 0.006). Regarding ICSAD scores, TPT improved from 126 to 63.4 s (p = 0.002), and TPL improved from 11.07 to 9.4 m (p = 0.322). When comparing the last simulated session and the subsequent measurement in an actual patient, median values of GRS, TPL and NM were not different.</p><p><strong>Conclusions: </strong>This simulation-based training program significantly improved residents' proficiency in an ultrasound-guided popliteal sciatic block. The learning curve plateaued at session 7, and this improvement was transferred to the real patient setting. As expected, residents needed more time for the first block on a real patient than for the last simulated session.</p><p><strong>Clinical trial number: </strong>ClinicalTrials.gov, identifier NCT06081790.</p>","PeriodicalId":72108,"journal":{"name":"Advances in simulation (London, England)","volume":"10 1","pages":"55"},"PeriodicalIF":4.7,"publicationDate":"2025-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12607028/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145497493","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-28DOI: 10.1186/s41077-025-00359-x
Aditi Siddharth, Sotiris Mastoridis, Michael Silva, Debbie Aitken, Helen Higham
Background: The acquisition and maintenance of technical skills in surgical specialties has become increasingly challenging for postgraduate trainees, exacerbated by factors such as the shift from traditional apprenticeship models, reduced operative time, and the impact of the COVID-19 pandemic. Virtual reality (VR) simulators offer a promising adjunct to traditional surgical training, though their integration into routine practice remain underexplored.
Objective: This qualitative study investigates the experiences and motivations of general surgical trainees who engaged with a VR laparoscopic simulator as part of a structured training program.
Methods: A case study methodology was chosen to explore the experiences of 22 general surgery trainees using a VR laparoscopic simulator over a period of 3 months. Each of the trainees were adviced to practise a minimum of five repetitions across 25 laparoscopic simulator exercises. The study was designed using Kopta's theory of technical skill learning, focusing on the cognitive phase, where trainees repetitively practised individual steps with feedback. Data collection involved qualitative questionnaires, semi-structured interviews (of seven of the trainees, 8 months later), and quantitative data from the simulator. The qualitative data was analysed using thematic analysis, and descriptive statistical tests were applied to the quantitative data for triangulation.
Results: The study identified key factors influencing trainee engagement, including ease of access, the importance of periodic rather than frequent simulation sessions, Annual Review of Competency Progression (ARCP) overview and the value of setting specific performance goals. The findings suggest that simulation can effectively complement traditional surgical training when incorporated into routine practice, with potential for broader application if barriers such as time constraints and access issues are addressed.
Conclusion: This study contributes to the literature on surgical education by highlighting the need for targeted strategies to enhance the use of simulation as an adjunct alongside more traditional training.
{"title":"Exploring trainee experiences in a structured virtual reality laparoscopic training programme for general surgeons: a longitudinal case study.","authors":"Aditi Siddharth, Sotiris Mastoridis, Michael Silva, Debbie Aitken, Helen Higham","doi":"10.1186/s41077-025-00359-x","DOIUrl":"10.1186/s41077-025-00359-x","url":null,"abstract":"<p><strong>Background: </strong>The acquisition and maintenance of technical skills in surgical specialties has become increasingly challenging for postgraduate trainees, exacerbated by factors such as the shift from traditional apprenticeship models, reduced operative time, and the impact of the COVID-19 pandemic. Virtual reality (VR) simulators offer a promising adjunct to traditional surgical training, though their integration into routine practice remain underexplored.</p><p><strong>Objective: </strong>This qualitative study investigates the experiences and motivations of general surgical trainees who engaged with a VR laparoscopic simulator as part of a structured training program.</p><p><strong>Methods: </strong>A case study methodology was chosen to explore the experiences of 22 general surgery trainees using a VR laparoscopic simulator over a period of 3 months. Each of the trainees were adviced to practise a minimum of five repetitions across 25 laparoscopic simulator exercises. The study was designed using Kopta's theory of technical skill learning, focusing on the cognitive phase, where trainees repetitively practised individual steps with feedback. Data collection involved qualitative questionnaires, semi-structured interviews (of seven of the trainees, 8 months later), and quantitative data from the simulator. The qualitative data was analysed using thematic analysis, and descriptive statistical tests were applied to the quantitative data for triangulation.</p><p><strong>Results: </strong>The study identified key factors influencing trainee engagement, including ease of access, the importance of periodic rather than frequent simulation sessions, Annual Review of Competency Progression (ARCP) overview and the value of setting specific performance goals. The findings suggest that simulation can effectively complement traditional surgical training when incorporated into routine practice, with potential for broader application if barriers such as time constraints and access issues are addressed.</p><p><strong>Conclusion: </strong>This study contributes to the literature on surgical education by highlighting the need for targeted strategies to enhance the use of simulation as an adjunct alongside more traditional training.</p>","PeriodicalId":72108,"journal":{"name":"Advances in simulation (London, England)","volume":"10 1","pages":"54"},"PeriodicalIF":4.7,"publicationDate":"2025-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12570560/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145395700","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-27DOI: 10.1186/s41077-025-00382-y
Timo C Meine, Johannes M Dorl, Anselm A Derda, Nima Mahmoudi, Hans-Jonas Meyer
Background: Angiography simulator training (AST) can help to train important clinical aspects of complex angiography procedures before real patient contact. The aim of the present analysis was to synthesize the results of studies on endovascular interventions performed by interventionalists with and without AST in a meta-analysis.
Methods: A systematic literature research was performed in PubMed, Web-of-Science and CINAHL to identify all relevant studies. Inclusion criteria were original research, English language and comparison of endovascular interventions in procedure time (PT) and fluoroscopy time (FT) performed by interventionalists with and without AST. Study quality was assessed using modified Downs-and-Black-instrument (maximum 8 points). Heterogeneity-analysis (study design and I2) was determined, and fixed- or random-effects model was applied to pool the effect, mean difference (MD), from the individual studies. All analyses were performed two-sided, and the level-of-significance was 0.05.
Results: Overall, 9 studies with 10 datasets and 7774 interventions were included. Study quality was 7 ± 0 for both PT and FT. Heterogeneity was present in the studies on PT (I2 = 61%) and FT (I2 = 99%), and a random-effects model was applied. MD for PT was significant with -2.63 min between the AST-group and control-group among the included studies (p = 0.02). In contrast, MD was not significant with -1.33 min between the AST-group and control-group among the included studies for FT (p = 0.21).
Conclusion: AST translates into an improved PT and similar FT in real interventions compared to conventional training. Angiography simulators offer a valuable, radiation-free alternative and expand training opportunities. Evidence is limited by study heterogeneity.
{"title":"Comparison of fluoroscopy time and procedure time of endovascular interventions with and without prior angiography simulator training: a meta-analysis.","authors":"Timo C Meine, Johannes M Dorl, Anselm A Derda, Nima Mahmoudi, Hans-Jonas Meyer","doi":"10.1186/s41077-025-00382-y","DOIUrl":"10.1186/s41077-025-00382-y","url":null,"abstract":"<p><strong>Background: </strong>Angiography simulator training (AST) can help to train important clinical aspects of complex angiography procedures before real patient contact. The aim of the present analysis was to synthesize the results of studies on endovascular interventions performed by interventionalists with and without AST in a meta-analysis.</p><p><strong>Methods: </strong>A systematic literature research was performed in PubMed, Web-of-Science and CINAHL to identify all relevant studies. Inclusion criteria were original research, English language and comparison of endovascular interventions in procedure time (PT) and fluoroscopy time (FT) performed by interventionalists with and without AST. Study quality was assessed using modified Downs-and-Black-instrument (maximum 8 points). Heterogeneity-analysis (study design and I<sup>2</sup>) was determined, and fixed- or random-effects model was applied to pool the effect, mean difference (MD), from the individual studies. All analyses were performed two-sided, and the level-of-significance was 0.05.</p><p><strong>Results: </strong>Overall, 9 studies with 10 datasets and 7774 interventions were included. Study quality was 7 ± 0 for both PT and FT. Heterogeneity was present in the studies on PT (I<sup>2</sup> = 61%) and FT (I<sup>2</sup> = 99%), and a random-effects model was applied. MD for PT was significant with -2.63 min between the AST-group and control-group among the included studies (p = 0.02). In contrast, MD was not significant with -1.33 min between the AST-group and control-group among the included studies for FT (p = 0.21).</p><p><strong>Conclusion: </strong>AST translates into an improved PT and similar FT in real interventions compared to conventional training. Angiography simulators offer a valuable, radiation-free alternative and expand training opportunities. Evidence is limited by study heterogeneity.</p>","PeriodicalId":72108,"journal":{"name":"Advances in simulation (London, England)","volume":"10 1","pages":"53"},"PeriodicalIF":4.7,"publicationDate":"2025-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12560282/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145379668","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}