首页 > 最新文献

International Journal of Computer Assisted Radiology and Surgery最新文献

英文 中文
AI in radiology and interventions: a structured narrative review of workflow automation, accuracy, and efficiency gains of today and what's coming. 放射学和干预中的人工智能:对工作流自动化、准确性和效率提高的结构化叙述回顾,以及未来的发展趋势。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-11-17 DOI: 10.1007/s11548-025-03547-2
Michael Friebe

Purpose: Artificial intelligence (AI) is rapidly transforming diagnostic and interventional radiology, supported by accelerating regulatory approvals and clinical adoption. Despite progress, integration varies across modalities and procedures. This study is a structured narrative review of four representative workflows-MRI and CT screening, coronary stenting, and liver cryoablation-to quantify automation readiness, accuracy gains, and efficiency improvements. The novelty lies in comparing diagnostic and interventional domains to highlight distinct maturity levels and future opportunities for AI-driven workflow optimization and clinical value creation.

Methods: A structured analysis was performed identifying 43 workflow steps across the four selected procedures. Each step was evaluated for potential automation, accuracy improvement, and ability to provide new clinical insights, considering current availability and projected 2030 maturity. The assessment drew on peer-reviewed literature, FDA approvals, and industry data (2015-2025). A structured taxonomy distinguished between full automation, human-augmented improvements, and novel AI-enabled guidance functions.

Results: Diagnostic imaging showed higher maturity than interventional workflows. Currently, 70% of MRI and 64% of CT steps have available AI solutions, compared to 55% in coronary stenting and 36% in liver cryoablation. By 2030, nearly all steps are expected to be AI-supported. AI achieved up to 94% segmentation accuracy, 95% nodule detection sensitivity, 30-75% scan time reductions, and 30-50% faster reporting. Interventional applications improved catheter navigation, probe placement, and ablation success but still required significant human oversight.

Conclusions: AI has already demonstrated measurable gains in diagnostic accuracy, efficiency, and workflow standardization. Interventional applications are emerging, with future growth expected in guidance, robotics, and real-time optimization. Despite progress, key limitations include algorithm generalizability, clinical interpretability, organizational readiness, and regulatory uncertainty. AI will augment rather than replace human expertise, with collaborative human-AI workflows being essential. Future integration efforts must address interoperability, workforce adaptation, and ethical considerations to ensure safe, equitable, and clinically impactful deployment.

目的:人工智能(AI)在加速监管审批和临床应用的支持下,正在迅速改变诊断和介入放射学。尽管取得了进展,但整合在方式和程序上有所不同。本研究是对四个代表性工作流程(mri和CT筛查、冠状动脉支架植入和肝脏冷冻消融)的结构化叙述性回顾,以量化自动化准备情况、准确性提高和效率提高。其新颖之处在于比较诊断和介入领域,以突出人工智能驱动的工作流程优化和临床价值创造的不同成熟度水平和未来机会。方法:进行结构化分析,确定四个选定程序中的43个工作流程步骤。考虑到目前的可用性和预计2030年的成熟度,每个步骤都评估了潜在的自动化、准确性改进和提供新的临床见解的能力。该评估参考了同行评审文献、FDA批准和行业数据(2015-2025年)。结构化分类法区分了完全自动化、人工增强改进和新颖的人工智能引导功能。结果:诊断性影像学显示比介入性工作流程更成熟。目前,70%的MRI和64%的CT步骤都有可用的人工智能解决方案,相比之下,冠状动脉支架置入和肝脏冷冻消融的人工智能解决方案分别为55%和36%。到2030年,预计几乎所有步骤都将由人工智能支持。人工智能实现了高达94%的分割准确率,95%的结节检测灵敏度,30-75%的扫描时间缩短,30-50%的报告速度加快。介入应用改善了导管导航、探针放置和消融成功率,但仍需要大量的人为监督。结论:人工智能已经在诊断准确性、效率和工作流程标准化方面显示出可衡量的收益。介入应用正在兴起,预计未来将在制导、机器人和实时优化方面增长。尽管取得了进展,但主要的限制包括算法的通用性、临床可解释性、组织准备程度和监管不确定性。人工智能将增强而不是取代人类的专业知识,人类与人工智能的协作工作流程至关重要。未来的整合工作必须解决互操作性、劳动力适应和道德考虑问题,以确保安全、公平和临床有效的部署。
{"title":"AI in radiology and interventions: a structured narrative review of workflow automation, accuracy, and efficiency gains of today and what's coming.","authors":"Michael Friebe","doi":"10.1007/s11548-025-03547-2","DOIUrl":"https://doi.org/10.1007/s11548-025-03547-2","url":null,"abstract":"<p><strong>Purpose: </strong>Artificial intelligence (AI) is rapidly transforming diagnostic and interventional radiology, supported by accelerating regulatory approvals and clinical adoption. Despite progress, integration varies across modalities and procedures. This study is a structured narrative review of four representative workflows-MRI and CT screening, coronary stenting, and liver cryoablation-to quantify automation readiness, accuracy gains, and efficiency improvements. The novelty lies in comparing diagnostic and interventional domains to highlight distinct maturity levels and future opportunities for AI-driven workflow optimization and clinical value creation.</p><p><strong>Methods: </strong>A structured analysis was performed identifying 43 workflow steps across the four selected procedures. Each step was evaluated for potential automation, accuracy improvement, and ability to provide new clinical insights, considering current availability and projected 2030 maturity. The assessment drew on peer-reviewed literature, FDA approvals, and industry data (2015-2025). A structured taxonomy distinguished between full automation, human-augmented improvements, and novel AI-enabled guidance functions.</p><p><strong>Results: </strong>Diagnostic imaging showed higher maturity than interventional workflows. Currently, 70% of MRI and 64% of CT steps have available AI solutions, compared to 55% in coronary stenting and 36% in liver cryoablation. By 2030, nearly all steps are expected to be AI-supported. AI achieved up to 94% segmentation accuracy, 95% nodule detection sensitivity, 30-75% scan time reductions, and 30-50% faster reporting. Interventional applications improved catheter navigation, probe placement, and ablation success but still required significant human oversight.</p><p><strong>Conclusions: </strong>AI has already demonstrated measurable gains in diagnostic accuracy, efficiency, and workflow standardization. Interventional applications are emerging, with future growth expected in guidance, robotics, and real-time optimization. Despite progress, key limitations include algorithm generalizability, clinical interpretability, organizational readiness, and regulatory uncertainty. AI will augment rather than replace human expertise, with collaborative human-AI workflows being essential. Future integration efforts must address interoperability, workforce adaptation, and ethical considerations to ensure safe, equitable, and clinically impactful deployment.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2025-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145543703","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automatic system calibration for orthognathic robot system. 正交机器人系统的自动标定。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-11-12 DOI: 10.1007/s11548-025-03540-9
Qianqian Li, Guoliang Li, Xiaojing Liu, Rui Song

Purpose: System calibration, including hand-robot and robot-world calibration, is an essential step that directly influences the location accuracy of surgical robots. Conventional calibration methods for orthognathic robot systems (ORSs) face significant challenges in handling irregularly shaped end tools, leading to manual intervention and compromised accuracy. Therefore, an automatic method has been proposed to improve the calibration efficiency and accuracy of ORSs.

Methods: The core innovation of the proposed method lies in enabling automation of both pre-intraoperative image registration and robotic hand-eye calibration, via aligning the 3D model of irregularly shaped tools to the preoperative CT space. It can effectively minimize errors caused by manual intervention. firstly, the equations of hand-eye-tool calibration were reconstructed using the preoperative graphic information to define tool endpoints (TEPs). Then, the transformation matrices were solved via a robust optimization method based on least squares. finally, the whole calibration process was completed automatically with robot path planning without human involvement.

Results: A group of simulated robot-assisted orthognathic surgery experiments was performed. The proposed method achieved a calibration error of 1.04 ± 0.54 mm, and the total execution error were reduced to 1.56 ± 0.61 mm.

Conclusion: The experimental results proved that the proposed calibration method could not only automate the calibration process, but also effectively improve the accuracy and stability of the system. It is expected to pave the way for more autonomous and efficient surgical procedures. Also, there are some limitations need to be overcome, including dependency on marker-based tracking and small sample size. Future work will integrate markerless tracking and machine learning for further optimization.

目的:系统标定是直接影响手术机器人定位精度的重要步骤,包括手-机器人标定和机器人世界标定。正齿机器人系统(ors)的传统校准方法在处理不规则形状的端刀时面临重大挑战,导致人工干预和精度降低。为此,提出了一种自动标定方法,以提高遥感卫星的标定效率和精度。方法:该方法的核心创新在于通过将不规则形状工具的三维模型对准术前CT空间,实现术前图像配准和机器人手眼校准的自动化。它可以有效地减少人工干预造成的错误。首先,利用术前图像信息重构手-眼-刀具标定方程,确定刀具端点;然后,采用基于最小二乘的鲁棒优化方法对变换矩阵进行求解。最后,整个标定过程在机器人路径规划的情况下自动完成,无需人工干预。结果:进行了一组模拟机器人辅助正颌手术实验。该方法标定误差为1.04±0.54 mm,总执行误差减小为1.56±0.61 mm。结论:实验结果证明,该标定方法不仅能实现标定过程的自动化,而且能有效提高系统的精度和稳定性。它有望为更自主、更高效的外科手术铺平道路。此外,还有一些限制需要克服,包括依赖于基于标记的跟踪和小样本量。未来的工作将整合无标记跟踪和机器学习以进一步优化。
{"title":"Automatic system calibration for orthognathic robot system.","authors":"Qianqian Li, Guoliang Li, Xiaojing Liu, Rui Song","doi":"10.1007/s11548-025-03540-9","DOIUrl":"https://doi.org/10.1007/s11548-025-03540-9","url":null,"abstract":"<p><strong>Purpose: </strong>System calibration, including hand-robot and robot-world calibration, is an essential step that directly influences the location accuracy of surgical robots. Conventional calibration methods for orthognathic robot systems (ORSs) face significant challenges in handling irregularly shaped end tools, leading to manual intervention and compromised accuracy. Therefore, an automatic method has been proposed to improve the calibration efficiency and accuracy of ORSs.</p><p><strong>Methods: </strong>The core innovation of the proposed method lies in enabling automation of both pre-intraoperative image registration and robotic hand-eye calibration, via aligning the 3D model of irregularly shaped tools to the preoperative CT space. It can effectively minimize errors caused by manual intervention. firstly, the equations of hand-eye-tool calibration were reconstructed using the preoperative graphic information to define tool endpoints (TEPs). Then, the transformation matrices were solved via a robust optimization method based on least squares. finally, the whole calibration process was completed automatically with robot path planning without human involvement.</p><p><strong>Results: </strong>A group of simulated robot-assisted orthognathic surgery experiments was performed. The proposed method achieved a calibration error of 1.04 ± 0.54 mm, and the total execution error were reduced to 1.56 ± 0.61 mm.</p><p><strong>Conclusion: </strong>The experimental results proved that the proposed calibration method could not only automate the calibration process, but also effectively improve the accuracy and stability of the system. It is expected to pave the way for more autonomous and efficient surgical procedures. Also, there are some limitations need to be overcome, including dependency on marker-based tracking and small sample size. Future work will integrate markerless tracking and machine learning for further optimization.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2025-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145496794","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Surgical instrument-tissue interaction recognition with multi-task-attention video transformer. 手术器械-组织交互识别与多任务注意视频转换器。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-11-11 DOI: 10.1007/s11548-025-03546-3
Lennart Maack, Berk Cam, Sarah Latus, Tobias Maurer, Alexander Schlaefer

Purpose: The recognition of surgical instrument-tissue interactions can enhance the surgical workflow analysis, improve automated safety systems and enable skill assessment in minimally invasive surgery. However, current deep learning methods for surgical instrument-tissue interaction recognition often rely on static images or coarse temporal sampling, limiting their ability to capture rapid surgical dynamics. Therefore, this study systematically investigates the impact of incorporating fine-grained temporal context into deep learning models for interaction recognition.

Methods: We conduct extensive experiments with multiple curated video-based datasets to investigate the influence of fine-grained temporal context for the task of instrument-tissue interaction recognition using video transformer with spatio-temporal feature extraction capabilities. Additionally, we propose a multi-task-attention module that utilizes cross-attention and a gating mechanism to improve communication between the subtasks of identifying the surgical instrument, atomic action, and anatomical target.

Results: Our study demonstrates the benefit of utilizing the fine-grained temporal context for recognition of instrument-tissue interactions, with an optimal sampling rate of 6-8 Hz identified for the examined datasets. Furthermore, our proposed MTAM significantly outperforms state-of-the-art multi-task video transformer on the CholecT45-Vid and GraSP-Vid datasets, achieving relative increases of 4.8 % and 5.9 % in surgical instrument-tissue interaction recognition, respectively.

Conclusions: In this work, we demonstrate the benefits of using a fine-grained temporal context rather than static images or coarse temporal context for the task of surgical instrument-tissue interaction recognition. We also show that leveraging cross-attention with spatio-temporal features from various subtasks leads to improved surgical instrument-tissue interaction recognition performance. The project is available at: https://lennart-maack.github.io/InstrTissRec-MTAM .

目的:对手术器械-组织相互作用的识别可以增强手术流程分析,改进自动化安全系统,实现微创手术的技能评估。然而,目前用于手术器械-组织交互识别的深度学习方法通常依赖于静态图像或粗时间采样,限制了它们捕捉快速手术动态的能力。因此,本研究系统地研究了将细粒度时间上下文纳入深度学习模型以进行交互识别的影响。方法:我们对多个基于视频的数据集进行了广泛的实验,以研究细粒度时间背景对使用具有时空特征提取能力的视频转换器进行仪器-组织交互识别任务的影响。此外,我们提出了一个多任务注意模块,该模块利用交叉注意和门控机制来改善识别手术器械、原子作用和解剖目标的子任务之间的沟通。结果:我们的研究证明了利用细粒度时间背景来识别仪器-组织相互作用的好处,为所检查的数据集确定了6-8 Hz的最佳采样率。此外,我们提出的MTAM在CholecT45-Vid和grip - vid数据集上显著优于最先进的多任务视频转换器,在手术器械-组织交互识别方面分别实现了4.8%和5.9%的相对增长。结论:在这项工作中,我们证明了使用细粒度时间背景而不是静态图像或粗糙时间背景进行手术器械-组织相互作用识别的好处。我们还表明,利用来自不同子任务的时空特征的交叉注意可以改善手术器械-组织交互识别性能。该项目可在:https://lennart-maack.github.io/InstrTissRec-MTAM。
{"title":"Surgical instrument-tissue interaction recognition with multi-task-attention video transformer.","authors":"Lennart Maack, Berk Cam, Sarah Latus, Tobias Maurer, Alexander Schlaefer","doi":"10.1007/s11548-025-03546-3","DOIUrl":"https://doi.org/10.1007/s11548-025-03546-3","url":null,"abstract":"<p><strong>Purpose: </strong>The recognition of surgical instrument-tissue interactions can enhance the surgical workflow analysis, improve automated safety systems and enable skill assessment in minimally invasive surgery. However, current deep learning methods for surgical instrument-tissue interaction recognition often rely on static images or coarse temporal sampling, limiting their ability to capture rapid surgical dynamics. Therefore, this study systematically investigates the impact of incorporating fine-grained temporal context into deep learning models for interaction recognition.</p><p><strong>Methods: </strong>We conduct extensive experiments with multiple curated video-based datasets to investigate the influence of fine-grained temporal context for the task of instrument-tissue interaction recognition using video transformer with spatio-temporal feature extraction capabilities. Additionally, we propose a multi-task-attention module that utilizes cross-attention and a gating mechanism to improve communication between the subtasks of identifying the surgical instrument, atomic action, and anatomical target.</p><p><strong>Results: </strong>Our study demonstrates the benefit of utilizing the fine-grained temporal context for recognition of instrument-tissue interactions, with an optimal sampling rate of 6-8 Hz identified for the examined datasets. Furthermore, our proposed MTAM significantly outperforms state-of-the-art multi-task video transformer on the CholecT45-Vid and GraSP-Vid datasets, achieving relative increases of <math><mrow><mn>4.8</mn> <mo>%</mo></mrow> </math> and <math><mrow><mn>5.9</mn> <mo>%</mo></mrow> </math> in surgical instrument-tissue interaction recognition, respectively.</p><p><strong>Conclusions: </strong>In this work, we demonstrate the benefits of using a fine-grained temporal context rather than static images or coarse temporal context for the task of surgical instrument-tissue interaction recognition. We also show that leveraging cross-attention with spatio-temporal features from various subtasks leads to improved surgical instrument-tissue interaction recognition performance. The project is available at: https://lennart-maack.github.io/InstrTissRec-MTAM .</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2025-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145490864","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MR-safe robotic needle driver for real-time MRI-guided minimally invasive procedures: a feasibility study. 核磁共振安全机器人针驱动器用于实时核磁共振引导的微创手术:可行性研究。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-11-08 DOI: 10.1007/s11548-025-03545-4
Atharva Paralikar, Gang Li, Chima Oluigbo, Pavel Yarmolenko, Kevin Cleary, Reza Monfaredi

Purpose: This article reports on the development and feasibility testing of an MR-safe robotic needle driver. The needle driver is pneumatically actuated and designed for automatic insertion and extraction of needles along a straight trajectory within the MRI scanner.

Method: All parts use plastic resins and composite materials to ensure MR-safe operation. A needle could be clamped in the needle carriage using a pneumatically operated clamp. The clamp is designed to be easily attached and detached from the needle driver. Clamps with different opening sizes could accommodate a range of needles from 18 to 22 gauge. To mimic the manual procedure of needle insertion, a pneumatically operated rack-and-pinion mechanism simultaneously translates and rotates the needle carriage along a helical slot. Signal-to-noise ratio (SNR) and 2-D geometric distortion were measured to evaluate the MRI compatibility. Targeting was measured with an electromagnetic tracker. We also evaluated the maximum force that could be generated at the tip of the needle with different clamping pressures using a force sensor.

Results: We recorded the maximum percentage change in SNR for multiple configurations of needle drivers as 6.6% and the maximum geometric distortion at 0.24%. The needle driver's mean positioning accuracy for 105 targets at 50 mm depth was 2.38 ± 1.00 mm in a composite tissue phantom. The angulation error for the straight trajectory was 0.51°, and the mean linear trajectory deviation was statistically negligible. The measured force at the needle tip was 1.17N, 1.6N, and 2.12N at 30, 40, and 50 psi, respectively.

Conclusion: This preliminary study showed that the prototype of our robotic needle driver works as intended for the insertion and extraction of the needle. The driver is MR-safe and serves as a suitable platform for MRI-guided interventions.

目的:本文报道了一种核磁共振安全机器人打针器的研制和可行性测试。针头驱动器是气动驱动的,设计用于沿着MRI扫描仪内的直线轨迹自动插入和取出针头。方法:所有部件均采用塑料树脂和复合材料,确保核磁共振安全操作。可以使用气动钳将针夹在针架中。该夹具的设计是很容易连接和从针驱动器分离。不同开口尺寸的夹子可以容纳18到22号的针。为了模拟人工插针的过程,气动操作的齿条-小齿轮机构同时沿着螺旋槽平移和旋转针架。测量信噪比(SNR)和二维几何畸变来评估MRI兼容性。目标是用电磁跟踪器测量的。我们还使用力传感器评估了不同夹紧压力下针尖可能产生的最大力。结果:我们记录了多种配置的针驱动器的最大信噪比变化百分比为6.6%,最大几何畸变百分比为0.24%。在复合组织模体中,针驱动器对105个50 mm深度目标的平均定位精度为2.38±1.00 mm。直线轨迹的角度误差为0.51°,平均线性轨迹偏差在统计学上可以忽略不计。在30、40和50 psi的压力下,测得针尖处的力分别为1.17、1.6和2.12N。结论:这项初步研究表明,我们的机器人打针器原型在针的插入和拔出方面是预期的。驱动器是核磁共振安全的,可作为核磁共振引导干预的合适平台。
{"title":"MR-safe robotic needle driver for real-time MRI-guided minimally invasive procedures: a feasibility study.","authors":"Atharva Paralikar, Gang Li, Chima Oluigbo, Pavel Yarmolenko, Kevin Cleary, Reza Monfaredi","doi":"10.1007/s11548-025-03545-4","DOIUrl":"https://doi.org/10.1007/s11548-025-03545-4","url":null,"abstract":"<p><strong>Purpose: </strong>This article reports on the development and feasibility testing of an MR-safe robotic needle driver. The needle driver is pneumatically actuated and designed for automatic insertion and extraction of needles along a straight trajectory within the MRI scanner.</p><p><strong>Method: </strong>All parts use plastic resins and composite materials to ensure MR-safe operation. A needle could be clamped in the needle carriage using a pneumatically operated clamp. The clamp is designed to be easily attached and detached from the needle driver. Clamps with different opening sizes could accommodate a range of needles from 18 to 22 gauge. To mimic the manual procedure of needle insertion, a pneumatically operated rack-and-pinion mechanism simultaneously translates and rotates the needle carriage along a helical slot. Signal-to-noise ratio (SNR) and 2-D geometric distortion were measured to evaluate the MRI compatibility. Targeting was measured with an electromagnetic tracker. We also evaluated the maximum force that could be generated at the tip of the needle with different clamping pressures using a force sensor.</p><p><strong>Results: </strong>We recorded the maximum percentage change in SNR for multiple configurations of needle drivers as 6.6% and the maximum geometric distortion at 0.24%. The needle driver's mean positioning accuracy for 105 targets at 50 mm depth was 2.38 ± 1.00 mm in a composite tissue phantom. The angulation error for the straight trajectory was 0.51°, and the mean linear trajectory deviation was statistically negligible. The measured force at the needle tip was 1.17N, 1.6N, and 2.12N at 30, 40, and 50 psi, respectively.</p><p><strong>Conclusion: </strong>This preliminary study showed that the prototype of our robotic needle driver works as intended for the insertion and extraction of the needle. The driver is MR-safe and serves as a suitable platform for MRI-guided interventions.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2025-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145472500","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Toward robust surgical phase recognition via deep ensemble learning. 基于深度集成学习的鲁棒外科相位识别。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-11-08 DOI: 10.1007/s11548-025-03543-6
Flakë Bajraktari, Lina Hauser, Peter P Pott

Purpose: Automatic recognition of surgical workflows is a complex yet essential task of context-aware systems in the operating room. However, achieving high accuracy in phase recognition remains a challenge due to the complexity of surgical procedures. While recent deep learning models have made significant progress, individual models often exhibit limitations-some may excel at capturing spatial features, while others are better at modeling temporal dependencies or handling class imbalance.

Methods: This study investigates the use of ensemble learning to combine the complementary strengths of diverse architectures, aiming to mitigate individual model weaknesses and improve performance in surgical phase recognition using the Cholec80 dataset. A variety of advanced deep learning architectures was integrated into a single ensemble. Models were carefully selected and tuned to ensure diversity, resulting in a final set of 15 unique ensembles. Ensemble strategies were explored to determine the most effective method for combining the distinct models.

Results: The results demonstrated that ensemble learning significantly improved performance. Among the ensemble strategies tested, majority voting achieved the highest F1-score, followed by the proposed artificial neural network StackingNet. Ensembles with high model diversity showed superior performance compared to those with lower diversity. The optimal ensemble configuration integrated top-performing models from different architectures, leading to improvements in accuracy, F1-score, and Jaccard Index by 1.48 %, 3.68 %, and 5.43 %, respectively, compared to the best individual models.

Conclusion: This study demonstrates that ensemble learning can substantially enhance surgical phase recognition by leveraging the complementary strengths of diverse deep learning models. Ensemble size, diversity, and meta-model selection were identified as key factors influencing performance. The resulting improvements translate into clinically meaningful benefits by enabling more reliable context-aware guidance, reducing misclassifications during critical phases, and improving surgeons' trust in artificial intelligence (AI) systems.

目的:手术工作流程的自动识别是手术室环境感知系统的一项复杂而重要的任务。然而,由于手术过程的复杂性,在相位识别中实现高精度仍然是一个挑战。虽然最近的深度学习模型取得了重大进展,但个别模型往往表现出局限性——一些模型可能擅长捕捉空间特征,而另一些模型则更擅长建模时间依赖性或处理类不平衡。方法:本研究探讨了集成学习的使用,以结合不同架构的互补优势,旨在减轻单个模型的弱点,并提高使用Cholec80数据集进行手术相位识别的性能。各种先进的深度学习架构被集成到一个集成中。模特经过精心挑选和调整,以确保多样性,最终形成了15套独特的套装。研究了集成策略,以确定组合不同模型的最有效方法。结果:结果表明,集成学习显著提高成绩。在测试的集成策略中,多数投票获得了最高的f1分,其次是人工神经网络StackingNet。模型多样性高的群落比模型多样性低的群落表现出更好的性能。最优的集成配置集成了来自不同架构的顶级模型,与最佳单个模型相比,其准确性、f1分数和Jaccard指数分别提高了1.48%、3.68%和5.43%。结论:本研究表明,通过利用不同深度学习模型的互补优势,集成学习可以大大增强手术阶段识别。整体规模、多样性和元模型选择是影响绩效的关键因素。通过实现更可靠的上下文感知指导,减少关键阶段的错误分类,以及提高外科医生对人工智能(AI)系统的信任,由此产生的改进转化为临床意义上的益处。
{"title":"Toward robust surgical phase recognition via deep ensemble learning.","authors":"Flakë Bajraktari, Lina Hauser, Peter P Pott","doi":"10.1007/s11548-025-03543-6","DOIUrl":"https://doi.org/10.1007/s11548-025-03543-6","url":null,"abstract":"<p><strong>Purpose: </strong>Automatic recognition of surgical workflows is a complex yet essential task of context-aware systems in the operating room. However, achieving high accuracy in phase recognition remains a challenge due to the complexity of surgical procedures. While recent deep learning models have made significant progress, individual models often exhibit limitations-some may excel at capturing spatial features, while others are better at modeling temporal dependencies or handling class imbalance.</p><p><strong>Methods: </strong>This study investigates the use of ensemble learning to combine the complementary strengths of diverse architectures, aiming to mitigate individual model weaknesses and improve performance in surgical phase recognition using the Cholec80 dataset. A variety of advanced deep learning architectures was integrated into a single ensemble. Models were carefully selected and tuned to ensure diversity, resulting in a final set of 15 unique ensembles. Ensemble strategies were explored to determine the most effective method for combining the distinct models.</p><p><strong>Results: </strong>The results demonstrated that ensemble learning significantly improved performance. Among the ensemble strategies tested, majority voting achieved the highest F1-score, followed by the proposed artificial neural network StackingNet. Ensembles with high model diversity showed superior performance compared to those with lower diversity. The optimal ensemble configuration integrated top-performing models from different architectures, leading to improvements in accuracy, F1-score, and Jaccard Index by 1.48 %, 3.68 %, and 5.43 %, respectively, compared to the best individual models.</p><p><strong>Conclusion: </strong>This study demonstrates that ensemble learning can substantially enhance surgical phase recognition by leveraging the complementary strengths of diverse deep learning models. Ensemble size, diversity, and meta-model selection were identified as key factors influencing performance. The resulting improvements translate into clinically meaningful benefits by enabling more reliable context-aware guidance, reducing misclassifications during critical phases, and improving surgeons' trust in artificial intelligence (AI) systems.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2025-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145477281","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A label-aware diffusion model for weakly supervised deformable registration of multimodal MRI-TRUS in prostate cancer. 用于前列腺癌多模态MRI-TRUS弱监督形变登记的标签感知扩散模型。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-11-06 DOI: 10.1007/s11548-025-03538-3
Zhirong Yao, Jiajun Chen, Tiexiang Wen

Purpose: Prostate cancer is a prevalent malignant tumor in men, and accurate diagnosis and personalized treatment rely on multimodal imaging, such as MRI and TRUS. However, differences in imaging mechanisms and prostate deformation due to ultrasound probe compression pose significant challenges for high-quality registration between the two modalities.

Methods: In this study, we propose a label-aware weakly supervised diffusion model for MRI-TRUS multimodal image registration. First, we align label centroid positions by maximizing the Dice coefficient to correct initial biases. Second, we combine label supervision with a diffusion model to generate high-quality deformation fields. Finally, we incorporate a feature-guided module to better preserve edge structures and improve registration smoothness.

Results: Experiments conducted on the µ-RegPro dataset demonstrate that our method outperforms current state-of-the-art (SOTA) approaches across multiple evaluation metrics. Specifically, it achieves a Dice coefficient of 0.880 and reduces the target registration error (TRE) to 0.940, significantly surpassing unsupervised methods such as VoxelMorph, FSDiffReg, and supervised methods like LocalNet and AutoFuse. The results show that preliminary label centroid alignment effectively enhances the performance of the diffusion-based deformation registration model, reducing the TRE from 3.084 to 0.940. The ablation study demonstrates that the feature-guided diffusion module effectively suppresses deformation field folding, while the label-aware module enhances label alignment. When combined, the proposed framework achieves a favorable balance, substantially improving registration accuracy (Dice = 0.880, TRE = 0.940) with reduced folding (|J|≤0 = 0.134). This method exhibits strong robustness and generalizability in handling large deformations in target regions while preserving details in nontarget regions.

Conclusion: The proposed label-aware weakly supervised diffusion model enables accurate and efficient MRI-TRUS multimodal image registration, offering strong potential for clinical applications such as prostate cancer diagnosis, targeted biopsy, and image-guided navigation.

目的:前列腺癌是男性常见的恶性肿瘤,准确诊断和个性化治疗依赖于MRI、TRUS等多模式影像。然而,成像机制的差异和超声探头压缩导致的前列腺变形对两种方式之间的高质量登记构成了重大挑战。方法:在本研究中,我们提出了一个标签感知的弱监督扩散模型用于MRI-TRUS多模态图像配准。首先,我们通过最大化Dice系数来对齐标签质心位置,以纠正初始偏差。其次,我们将标签监督与扩散模型相结合,生成高质量的变形场。最后,我们加入了一个特征引导模块,以更好地保留边缘结构,提高配准的平稳性。结果:在µ-RegPro数据集上进行的实验表明,我们的方法在多个评估指标上优于当前最先进的(SOTA)方法。具体来说,它实现了0.880的Dice系数,并将目标注册误差(TRE)降低到0.940,显著超过了VoxelMorph、FSDiffReg等无监督方法,以及LocalNet和AutoFuse等有监督方法。结果表明,预标记质心对齐有效地提高了基于扩散的变形配准模型的性能,将TRE从3.084降低到0.940。烧蚀研究表明,特征引导扩散模块有效抑制变形场折叠,而标签感知模块增强标签对齐。当结合使用时,所提出的框架达到了良好的平衡,大大提高了配准精度(Dice = 0.880, TRE = 0.940),减少了折叠(|J|≤0 = 0.134)。该方法具有较强的鲁棒性和泛化性,既能处理目标区域的大变形,又能保留非目标区域的细节。结论:提出的标签感知弱监督扩散模型能够实现准确、高效的MRI-TRUS多模态图像配准,为前列腺癌诊断、靶向活检和图像引导导航等临床应用提供了强大的潜力。
{"title":"A label-aware diffusion model for weakly supervised deformable registration of multimodal MRI-TRUS in prostate cancer.","authors":"Zhirong Yao, Jiajun Chen, Tiexiang Wen","doi":"10.1007/s11548-025-03538-3","DOIUrl":"https://doi.org/10.1007/s11548-025-03538-3","url":null,"abstract":"<p><strong>Purpose: </strong>Prostate cancer is a prevalent malignant tumor in men, and accurate diagnosis and personalized treatment rely on multimodal imaging, such as MRI and TRUS. However, differences in imaging mechanisms and prostate deformation due to ultrasound probe compression pose significant challenges for high-quality registration between the two modalities.</p><p><strong>Methods: </strong>In this study, we propose a label-aware weakly supervised diffusion model for MRI-TRUS multimodal image registration. First, we align label centroid positions by maximizing the Dice coefficient to correct initial biases. Second, we combine label supervision with a diffusion model to generate high-quality deformation fields. Finally, we incorporate a feature-guided module to better preserve edge structures and improve registration smoothness.</p><p><strong>Results: </strong>Experiments conducted on the µ-RegPro dataset demonstrate that our method outperforms current state-of-the-art (SOTA) approaches across multiple evaluation metrics. Specifically, it achieves a Dice coefficient of 0.880 and reduces the target registration error (TRE) to 0.940, significantly surpassing unsupervised methods such as VoxelMorph, FSDiffReg, and supervised methods like LocalNet and AutoFuse. The results show that preliminary label centroid alignment effectively enhances the performance of the diffusion-based deformation registration model, reducing the TRE from 3.084 to 0.940. The ablation study demonstrates that the feature-guided diffusion module effectively suppresses deformation field folding, while the label-aware module enhances label alignment. When combined, the proposed framework achieves a favorable balance, substantially improving registration accuracy (Dice = 0.880, TRE = 0.940) with reduced folding (|J|<sub>≤0</sub> = 0.134). This method exhibits strong robustness and generalizability in handling large deformations in target regions while preserving details in nontarget regions.</p><p><strong>Conclusion: </strong>The proposed label-aware weakly supervised diffusion model enables accurate and efficient MRI-TRUS multimodal image registration, offering strong potential for clinical applications such as prostate cancer diagnosis, targeted biopsy, and image-guided navigation.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2025-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145453739","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Assessing the impact of virtual reality on surgeons' mental models of complex congenital heart cases. 评估虚拟现实技术对外科医生复杂先心病心理模型的影响。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-11-05 DOI: 10.1007/s11548-025-03542-7
Eliot Bethke, Matthew T Bramlet, Bradley P Sutton, James L Evans, Ainsley Hanner, Ashley Tran, Brendan O'Rourke, Nina Soofi, Jennifer R Amos

Purpose: Virtual reality (VR) has attracted attention in healthcare for many promising applications including pre-surgical planning. Currently, there exists a critical gap in comprehension of the impact of VR on physicians' thinking. Self-reported data from surveys and metrics based on confidence and task completion may not yield sufficiently detailed understanding of the complex decision making and cognitive load experienced by surgeons during VR-based pre-surgical planning.

Methods: Our research aims to address the gap in understanding the impact of VR on physicians' mental models through a novel methodology of self-directed think-aloud protocols, offering deeper perspectives into physicians' thought processes within the virtual 3D environment. We performed qualitative analysis of recorded verbalizations and actions in VR in addition to quantitative measures from the NASA task load index (NASA-TLX). Analysis was conducted to identify thematic sequences in VR which influenced clinical decision making when reviewing patient anatomy.

Results: We find a significant increase in reported physician confidence in understanding of the patient anatomy from before VR to after (p = 0.012) and identified several common patterns of 3D exploration of the anatomy in VR. Physicians also reported low cognitive stress on the NASA-TLX.

Conclusion: Our findings indicate VR has value beyond simulating surgery, helping physicians to confirm findings from conventional medical imaging, visualize approaches with detail, and help make complex decisions while mentally preparing for surgery. These findings provide evidence that VR and related 3D visualization are helpful for pre-surgical planning of complex cases.

目的:虚拟现实(VR)在医疗保健中有许多有前途的应用,包括术前计划,引起了人们的注意。目前,在理解VR对医生思维的影响方面存在着一个关键的空白。基于信心和任务完成的调查和指标的自我报告数据可能无法充分详细地了解外科医生在基于vr的术前计划中所经历的复杂决策和认知负荷。方法:我们的研究旨在通过一种新颖的自主思考协议方法来解决VR对医生心理模型影响的理解差距,为医生在虚拟3D环境中的思维过程提供更深入的视角。除了NASA任务负载指数(NASA- tlx)的定量测量外,我们还对VR中记录的语言和动作进行了定性分析。进行了分析,以确定VR中影响临床决策的主题序列,当审查患者解剖时。结果:我们发现,从VR前到VR后,报告的医生对患者解剖结构理解的信心显著增加(p = 0.012),并确定了VR中三维解剖探索的几种常见模式。医生们还报告说,NASA-TLX的认知压力很低。结论:我们的研究结果表明,VR的价值不仅仅是模拟手术,还可以帮助医生确认传统医学成像的结果,详细地可视化方法,并帮助医生在心理上准备手术时做出复杂的决定。这些发现证明了VR和相关的3D可视化有助于复杂病例的术前计划。
{"title":"Assessing the impact of virtual reality on surgeons' mental models of complex congenital heart cases.","authors":"Eliot Bethke, Matthew T Bramlet, Bradley P Sutton, James L Evans, Ainsley Hanner, Ashley Tran, Brendan O'Rourke, Nina Soofi, Jennifer R Amos","doi":"10.1007/s11548-025-03542-7","DOIUrl":"https://doi.org/10.1007/s11548-025-03542-7","url":null,"abstract":"<p><strong>Purpose: </strong>Virtual reality (VR) has attracted attention in healthcare for many promising applications including pre-surgical planning. Currently, there exists a critical gap in comprehension of the impact of VR on physicians' thinking. Self-reported data from surveys and metrics based on confidence and task completion may not yield sufficiently detailed understanding of the complex decision making and cognitive load experienced by surgeons during VR-based pre-surgical planning.</p><p><strong>Methods: </strong>Our research aims to address the gap in understanding the impact of VR on physicians' mental models through a novel methodology of self-directed think-aloud protocols, offering deeper perspectives into physicians' thought processes within the virtual 3D environment. We performed qualitative analysis of recorded verbalizations and actions in VR in addition to quantitative measures from the NASA task load index (NASA-TLX). Analysis was conducted to identify thematic sequences in VR which influenced clinical decision making when reviewing patient anatomy.</p><p><strong>Results: </strong>We find a significant increase in reported physician confidence in understanding of the patient anatomy from before VR to after (p = 0.012) and identified several common patterns of 3D exploration of the anatomy in VR. Physicians also reported low cognitive stress on the NASA-TLX.</p><p><strong>Conclusion: </strong>Our findings indicate VR has value beyond simulating surgery, helping physicians to confirm findings from conventional medical imaging, visualize approaches with detail, and help make complex decisions while mentally preparing for surgery. These findings provide evidence that VR and related 3D visualization are helpful for pre-surgical planning of complex cases.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2025-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145446516","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Large language models with retrieval-augmented generation enhance expert modelling of Bayesian network for clinical decision support. 基于检索增强生成的大型语言模型增强了临床决策支持贝叶斯网络的专家建模。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-11-03 DOI: 10.1007/s11548-025-03524-9
Mario A Cypko, Muhammad Agus Salim, Aditya Kumar, Leonard Berliner, Andreas Dietz, Matthaeus Stoehr, Oliver Amft

Purpose: Bayesian networks (BNs) are valuable for clinical decision support due to their transparency and interpretability. However, BN modelling requires considerable manual effort. This study explores how integrating large language models (LLMs) with retrieval-augmented generation (RAG) can improve BN modelling by increasing efficiency, reducing cognitive workload, and ensuring accuracy.

Methods: We developed a web-based BN modelling service that integrates an LLM-RAG pipeline. A fine-tuned GTE-Large embedding model was employed for knowledge retrieval, optimised through recursive chunking and query expansion. To ensure accurate BN suggestions, we defined a causal structure for medical idioms by unifying existing BN frameworks. GPT-4 and Mixtral 8x7B were used to handle complex data interpretation and to generate modelling suggestions, respectively. A user study with four clinicians assessed usability, retrieval accuracy, and cognitive workload using NASA-TLX. The study demonstrated the system's potential for efficient and clinically relevant BN modelling.

Results: The RAG pipeline improved retrieval accuracy and answer relevance. Recursive chunking with the fine-tuned embedding model GTE-Large achieved the highest retrieval accuracy score (0.9). Query expansion and Hyde optimisation enhanced retrieval accuracy for semantic chunking (0.75 to 0.85). Responses maintained high faithfulness ( 0.9). However, the LLM occasionally failed to adhere to predefined causal structures and medical idioms. All clinicians, regardless of BN experience, created comprehensive models within one hour. Experienced clinicians produced more complex models, but occasionally introduced causality errors, while less experienced users adhered more accurately to predefined structures. The tool reduced cognitive workload (2/7 NASA-TLX) and was described as intuitive, although workflow interruptions and minor technical issues highlighted areas for improvement.

Conclusion: Integrating LLM-RAG into BN modelling enhances efficiency and accuracy. Future work may focus on automated preprocessing, refinements of the user interface, and extending the RAG pipeline with validation steps and external biomedical sources. Generative AI holds promise for expert-driven knowledge modelling.

目的:贝叶斯网络(BNs)因其透明性和可解释性在临床决策支持中具有重要价值。然而,BN建模需要大量的手工工作。本研究探讨了如何将大型语言模型(llm)与检索增强生成(RAG)相结合,通过提高效率、减少认知工作量和确保准确性来改善BN建模。方法:我们开发了一个基于web的BN建模服务,该服务集成了LLM-RAG管道。采用优化后的GTE-Large嵌入模型进行知识检索,并通过递归分块和查询扩展进行优化。为了确保准确的BN建议,我们通过统一现有的BN框架定义了医学习语的因果结构。GPT-4和Mixtral 8x7B分别用于处理复杂的数据解释和生成建模建议。一项由四位临床医生参与的用户研究评估了NASA-TLX的可用性、检索准确性和认知工作量。该研究证明了该系统在高效和临床相关的BN建模方面的潜力。结果:RAG流水线提高了检索准确率和答案相关性。采用微调嵌入模型GTE-Large的递归分块获得了最高的检索准确率得分(0.9)。查询扩展和Hyde优化提高了语义分块的检索精度(0.75到0.85)。应答保持高可信度(≥0.9)。然而,法学硕士偶尔不能坚持预定义的因果结构和医学习语。所有临床医生,无论BN经验如何,都在一小时内创建了全面的模型。经验丰富的临床医生产生了更复杂的模型,但偶尔会引入因果关系错误,而经验不足的用户更准确地遵循预定义的结构。该工具减少了认知工作量(2/7 NASA-TLX),并被描述为直观的,尽管工作流程中断和一些小的技术问题突出了需要改进的领域。结论:将LLM-RAG集成到BN建模中可以提高效率和准确性。未来的工作可能集中在自动化预处理,用户界面的改进,以及通过验证步骤和外部生物医学来源扩展RAG管道。生成式人工智能有望实现专家驱动的知识建模。
{"title":"Large language models with retrieval-augmented generation enhance expert modelling of Bayesian network for clinical decision support.","authors":"Mario A Cypko, Muhammad Agus Salim, Aditya Kumar, Leonard Berliner, Andreas Dietz, Matthaeus Stoehr, Oliver Amft","doi":"10.1007/s11548-025-03524-9","DOIUrl":"https://doi.org/10.1007/s11548-025-03524-9","url":null,"abstract":"<p><strong>Purpose: </strong>Bayesian networks (BNs) are valuable for clinical decision support due to their transparency and interpretability. However, BN modelling requires considerable manual effort. This study explores how integrating large language models (LLMs) with retrieval-augmented generation (RAG) can improve BN modelling by increasing efficiency, reducing cognitive workload, and ensuring accuracy.</p><p><strong>Methods: </strong>We developed a web-based BN modelling service that integrates an LLM-RAG pipeline. A fine-tuned GTE-Large embedding model was employed for knowledge retrieval, optimised through recursive chunking and query expansion. To ensure accurate BN suggestions, we defined a causal structure for medical idioms by unifying existing BN frameworks. GPT-4 and Mixtral 8x7B were used to handle complex data interpretation and to generate modelling suggestions, respectively. A user study with four clinicians assessed usability, retrieval accuracy, and cognitive workload using NASA-TLX. The study demonstrated the system's potential for efficient and clinically relevant BN modelling.</p><p><strong>Results: </strong>The RAG pipeline improved retrieval accuracy and answer relevance. Recursive chunking with the fine-tuned embedding model GTE-Large achieved the highest retrieval accuracy score (0.9). Query expansion and Hyde optimisation enhanced retrieval accuracy for semantic chunking (0.75 to 0.85). Responses maintained high faithfulness ( <math><mo>≥</mo></math> 0.9). However, the LLM occasionally failed to adhere to predefined causal structures and medical idioms. All clinicians, regardless of BN experience, created comprehensive models within one hour. Experienced clinicians produced more complex models, but occasionally introduced causality errors, while less experienced users adhered more accurately to predefined structures. The tool reduced cognitive workload (2/7 NASA-TLX) and was described as intuitive, although workflow interruptions and minor technical issues highlighted areas for improvement.</p><p><strong>Conclusion: </strong>Integrating LLM-RAG into BN modelling enhances efficiency and accuracy. Future work may focus on automated preprocessing, refinements of the user interface, and extending the RAG pipeline with validation steps and external biomedical sources. Generative AI holds promise for expert-driven knowledge modelling.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2025-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145440187","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Correction to: Stereo reconstruction from microscopic images for computer-assisted ophthalmic surgery. 更正:用于计算机辅助眼科手术的显微图像立体重建。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-11-01 DOI: 10.1007/s11548-024-03270-4
Rebekka Peter, Sofia Moreira, Eleonora Tagliabue, Matthias Hillenbrand, Rita G Nunes, Franziska Mathis-Ullrich
{"title":"Correction to: Stereo reconstruction from microscopic images for computer-assisted ophthalmic surgery.","authors":"Rebekka Peter, Sofia Moreira, Eleonora Tagliabue, Matthias Hillenbrand, Rita G Nunes, Franziska Mathis-Ullrich","doi":"10.1007/s11548-024-03270-4","DOIUrl":"10.1007/s11548-024-03270-4","url":null,"abstract":"","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":"2367-2369"},"PeriodicalIF":2.3,"publicationDate":"2025-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12575599/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142717688","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Super-resolution for localizing electrode grids as small, deformable objects during epilepsy surgery using augmented reality headsets. 在癫痫手术期间使用增强现实耳机将电极网格定位为小型可变形物体的超分辨率。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-11-01 Epub Date: 2025-06-19 DOI: 10.1007/s11548-025-03401-5
Hizirwan S Salim, Abdullah Thabit, Sem Hoogteijling, Maryse A van 't Klooster, Theo van Walsum, Maeike Zijlmans, Mohamed Benmahdjoub

Purpose: Epilepsy surgery is a potential curative treatment for people with focal epilepsy. Intraoperative electrocorticogram (ioECoG) recordings from the brain guide neurosurgeons during resection. Accurate localization of epileptic activity and thus the ioECoG grids is critical for successful outcomes. We aim to develop and evaluate the feasibility of a novel method for localizing small, deformable objects using augmented reality (AR) head-mounted displays (HMDs) and artificial intelligence (AI). AR HMDs combine cameras and patient overlay visualization in a compact design.

Methods: We developed an image processing method for the HoloLens 2 to localize a 64-electrode ioECoG grid even when individual electrodes are indistinguishable due to low resolution. The method combines object detection, super-resolution, and pose estimation AI models with stereo triangulation. A synthetic dataset of 90,000 images trained the super-resolution and pose estimation models. The system was tested in a controlled environment against an optical tracker as ground truth. Accuracy was evaluated at distances between 40 and 90 cm.

Results: The system achieved sub-5 mm accuracy in localizing the ioECoG grid at distances shorter than 60 cm. At 40 cm, the accuracy remained below 2 mm, with an average standard deviation of less than 0.5 mm. At 60 cm the method processed on average 24 stereo frames per second.

Conclusion: This study demonstrates the feasibility of localizing small, deformable objects like ioECoG grids using AR HMDs. While results indicate clinically acceptable accuracy, further research is needed to validate the method in clinical environments and assess its impact on surgical precision and outcomes.

目的:癫痫手术是局灶性癫痫的一种潜在治疗方法。术中脑皮质电图(ioECoG)记录指导神经外科医生在切除过程中。癫痫活动的准确定位和脑ecog网格是成功治疗的关键。我们的目标是开发和评估一种利用增强现实(AR)头戴式显示器(hmd)和人工智能(AI)定位小型可变形物体的新方法的可行性。AR头戴式显示器在紧凑的设计中结合了相机和患者覆盖可视化。方法:我们为HoloLens 2开发了一种图像处理方法,即使在单个电极由于低分辨率而无法区分的情况下,也可以定位64电极的ioECoG网格。该方法将目标检测、超分辨率和姿态估计人工智能模型与立体三角测量相结合。一个由9万张图像组成的合成数据集训练了超分辨率和姿态估计模型。该系统在受控环境下与光学跟踪器作为地面真值进行了测试。在距离为40至90厘米之间评估精度。结果:该系统在距离小于60 cm的ioECoG网格定位精度达到了5 mm以下。在40 cm处,精度保持在2 mm以下,平均标准偏差小于0.5 mm。在60厘米处,该方法平均每秒处理24个立体帧。结论:本研究证明了使用AR头显定位小型可变形物体(如ioECoG网格)的可行性。虽然结果表明临床可接受的准确性,但需要进一步的研究来验证该方法在临床环境中的有效性,并评估其对手术精度和结果的影响。
{"title":"Super-resolution for localizing electrode grids as small, deformable objects during epilepsy surgery using augmented reality headsets.","authors":"Hizirwan S Salim, Abdullah Thabit, Sem Hoogteijling, Maryse A van 't Klooster, Theo van Walsum, Maeike Zijlmans, Mohamed Benmahdjoub","doi":"10.1007/s11548-025-03401-5","DOIUrl":"10.1007/s11548-025-03401-5","url":null,"abstract":"<p><strong>Purpose: </strong>Epilepsy surgery is a potential curative treatment for people with focal epilepsy. Intraoperative electrocorticogram (ioECoG) recordings from the brain guide neurosurgeons during resection. Accurate localization of epileptic activity and thus the ioECoG grids is critical for successful outcomes. We aim to develop and evaluate the feasibility of a novel method for localizing small, deformable objects using augmented reality (AR) head-mounted displays (HMDs) and artificial intelligence (AI). AR HMDs combine cameras and patient overlay visualization in a compact design.</p><p><strong>Methods: </strong>We developed an image processing method for the HoloLens 2 to localize a 64-electrode ioECoG grid even when individual electrodes are indistinguishable due to low resolution. The method combines object detection, super-resolution, and pose estimation AI models with stereo triangulation. A synthetic dataset of 90,000 images trained the super-resolution and pose estimation models. The system was tested in a controlled environment against an optical tracker as ground truth. Accuracy was evaluated at distances between 40 and 90 cm.</p><p><strong>Results: </strong>The system achieved sub-5 mm accuracy in localizing the ioECoG grid at distances shorter than 60 cm. At 40 cm, the accuracy remained below 2 mm, with an average standard deviation of less than 0.5 mm. At 60 cm the method processed on average 24 stereo frames per second.</p><p><strong>Conclusion: </strong>This study demonstrates the feasibility of localizing small, deformable objects like ioECoG grids using AR HMDs. While results indicate clinically acceptable accuracy, further research is needed to validate the method in clinical environments and assess its impact on surgical precision and outcomes.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":"2319-2327"},"PeriodicalIF":2.3,"publicationDate":"2025-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12575507/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144327675","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International Journal of Computer Assisted Radiology and Surgery
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1