首页 > 最新文献

International Journal of Computer Assisted Radiology and Surgery最新文献

英文 中文
Weakly supervised pre-training for surgical step recognition using unannotated and heterogeneously labeled videos. 使用无注释和异构标记视频进行手术步骤识别的弱监督预训练。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-02 DOI: 10.1007/s11548-025-03555-2
Sreeram Kamabattula, Kai Chen, Kiran Bhattacharyya

Purpose: Surgical video review is essential for minimally invasive surgical training, but manual annotation of surgical steps is time-consuming and limits scalability. We propose a weakly supervised pre-training framework that leverages unannotated or heterogeneously labeled surgical videos to improve automated surgical step recognition.

Methods: We evaluate three types of weak labels derived from unannotated datasets: (1) surgical phases from the same or other procedures, (2) surgical steps from different procedure types, and (3) intraoperative time progression. Using datasets from four robotic-assisted procedures (sleeve gastrectomy, hysterectomy, cholecystectomy, and radical prostatectomy), we simulate real-world annotation scarcity by varying the proportion of available step annotations ( α 0.25, 0.5, 0.75, 1.0). We benchmark the performance of a 2D CNN model trained with and without weak label pre-training.

Results: Pre-training with surgical phase labels-particularly from the same procedure type (PHASE-WITHIN)-consistently improved step recognition performance, with gains up to 6.4 f1-score points over standard ImageNet-based models under limited annotation conditions ( α = 0.25 on SLG). Cross-procedure step pre-training was beneficial for some procedures, and time-based labels provided moderate gains depending on procedure structure. Label efficiency analysis shows the baseline model would require labeling an additional 30-60 videos at α = 0.25 to match the performance achieved by the best weak-pretraining strategy across procedures.

Conclusion: Weakly supervised pre-training offers a practical strategy to improve surgical step recognition when annotated data is scarce. This approach can support scalable feedback and assessment in surgical training workflows where comprehensive annotations are infeasible.

目的:手术视频回顾是微创手术培训必不可少的,但手工标注手术步骤耗时且限制可扩展性。我们提出了一个弱监督的预训练框架,利用无注释或异构标记的手术视频来提高自动手术步骤识别。方法:我们评估了来自未注释数据集的三种类型的弱标签:(1)来自相同或其他手术的手术阶段,(2)来自不同手术类型的手术步骤,(3)术中时间进展。使用四种机器人辅助手术(袖式胃切除术、子宫切除术、胆囊切除术和根治性前列腺切除术)的数据集,我们通过改变可用步骤注释的比例(α∈0.25,0.5,0.75,1.0)来模拟现实世界的注释稀缺性。我们对使用和不使用弱标签预训练的二维CNN模型的性能进行了基准测试。结果:使用手术阶段标签的预训练-特别是来自同一手术类型(phase - within)-持续提高了步骤识别性能,在有限的注释条件下(在SLG上α = 0.25),与基于标准imagenet的模型相比,增益高达6.4 f1分。跨程序步骤预训练对某些程序是有益的,基于时间的标签根据程序结构提供适度的增益。标签效率分析表明,基线模型将需要在α = 0.25下标记额外的30-60个视频,以匹配最佳弱预训练策略所取得的性能。结论:弱监督预训练在注释数据稀缺的情况下,为提高手术步骤识别提供了一种实用的策略。这种方法可以在外科训练工作流程中支持可扩展的反馈和评估,其中全面的注释是不可行的。
{"title":"Weakly supervised pre-training for surgical step recognition using unannotated and heterogeneously labeled videos.","authors":"Sreeram Kamabattula, Kai Chen, Kiran Bhattacharyya","doi":"10.1007/s11548-025-03555-2","DOIUrl":"https://doi.org/10.1007/s11548-025-03555-2","url":null,"abstract":"<p><strong>Purpose: </strong>Surgical video review is essential for minimally invasive surgical training, but manual annotation of surgical steps is time-consuming and limits scalability. We propose a weakly supervised pre-training framework that leverages unannotated or heterogeneously labeled surgical videos to improve automated surgical step recognition.</p><p><strong>Methods: </strong>We evaluate three types of weak labels derived from unannotated datasets: (1) surgical phases from the same or other procedures, (2) surgical steps from different procedure types, and (3) intraoperative time progression. Using datasets from four robotic-assisted procedures (sleeve gastrectomy, hysterectomy, cholecystectomy, and radical prostatectomy), we simulate real-world annotation scarcity by varying the proportion of available step annotations ( <math><mi>α</mi></math> <math><mo>∈</mo></math> 0.25, 0.5, 0.75, 1.0). We benchmark the performance of a 2D CNN model trained with and without weak label pre-training.</p><p><strong>Results: </strong>Pre-training with surgical phase labels-particularly from the same procedure type (PHASE-WITHIN)-consistently improved step recognition performance, with gains up to 6.4 f1-score points over standard ImageNet-based models under limited annotation conditions ( <math><mi>α</mi></math> = 0.25 on SLG). Cross-procedure step pre-training was beneficial for some procedures, and time-based labels provided moderate gains depending on procedure structure. Label efficiency analysis shows the baseline model would require labeling an additional 30-60 videos at <math><mi>α</mi></math> = 0.25 to match the performance achieved by the best weak-pretraining strategy across procedures.</p><p><strong>Conclusion: </strong>Weakly supervised pre-training offers a practical strategy to improve surgical step recognition when annotated data is scarce. This approach can support scalable feedback and assessment in surgical training workflows where comprehensive annotations are infeasible.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2025-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145656095","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Point cloud registration algorithm using liver vascular skeleton feature with computed tomography and ultrasonography image fusion. 基于肝脏血管骨架特征的点云配准算法与计算机断层和超声图像融合。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-08-23 DOI: 10.1007/s11548-025-03496-w
Satoshi Miura, Masayuki Nakayama, Kexin Xu, Zhang Bo, Ryoko Kuromatsu, Masahito Nakano, Yu Noda, Takumi Kawaguchi

Purpose: Radiofrequency ablation for liver cancer has advanced rapidly. For accurate ultrasound-guided soft-tissue puncture surgery, it is necessary to fuse intraoperative ultrasound images with preoperative computed tomography images. However, the conventional method is difficult to estimate and fuse images accurately. To address this issue, the present study proposes an algorithm for registering cross-source point clouds based on not surface but the geometric features of the vascular point cloud.

Methods: We developed a fusion system that performs cross-source point cloud registration between ultrasound and computed tomography images, extracting the node, skeleton, and geomatic feature of the vascular point cloud. The system completes the fusion process in an average of 14.5 s after acquiring the vascular point clouds via ultrasound.

Results: The experiments were conducted to fuse liver images by the dummy model and the healthy participants, respectively. The results show the proposed method achieved a registration error within 1.4 mm and decreased the target registration error significantly compared to other methods in a liver dummy model registration experiment. Furthermore, the proposed method achieved the averaged RMSE within 2.23 mm in a human liver vascular skeleton.

Conclusion: The study concluded that because the registration method using vascular feature point cloud could realize the rapid and accurate fusion between ultrasound and computed tomography images, the method is useful to apply the real puncture surgery for radiofrequency ablation for liver. In future work, we will evaluate the proposed method by the patients.

目的:射频消融术治疗肝癌进展迅速。超声引导下进行准确的软组织穿刺手术,需要术中超声图像与术前计算机断层图像融合。然而,传统的方法难以准确地估计和融合图像。为了解决这一问题,本研究提出了一种基于血管点云几何特征而非表面特征的交叉源点云配准算法。方法:我们开发了一个融合系统,在超声和计算机断层图像之间进行交叉源点云配准,提取血管点云的节点、骨架和地理特征。该系统在超声获取血管点云后,平均14.5 s完成融合过程。结果:分别对虚拟模型和健康被试进行肝脏图像融合实验。结果表明,在肝脏假人模型配准实验中,与其他方法相比,该方法的配准误差在1.4 mm以内,显著降低了目标配准误差。此外,该方法在人肝脏血管骨架上实现了2.23 mm以内的平均RMSE。结论:基于血管特征点云的配准方法可以实现超声与ct图像的快速准确融合,可用于肝脏射频消融的实际穿刺手术。在未来的工作中,我们将由患者对所提出的方法进行评估。
{"title":"Point cloud registration algorithm using liver vascular skeleton feature with computed tomography and ultrasonography image fusion.","authors":"Satoshi Miura, Masayuki Nakayama, Kexin Xu, Zhang Bo, Ryoko Kuromatsu, Masahito Nakano, Yu Noda, Takumi Kawaguchi","doi":"10.1007/s11548-025-03496-w","DOIUrl":"10.1007/s11548-025-03496-w","url":null,"abstract":"<p><strong>Purpose: </strong>Radiofrequency ablation for liver cancer has advanced rapidly. For accurate ultrasound-guided soft-tissue puncture surgery, it is necessary to fuse intraoperative ultrasound images with preoperative computed tomography images. However, the conventional method is difficult to estimate and fuse images accurately. To address this issue, the present study proposes an algorithm for registering cross-source point clouds based on not surface but the geometric features of the vascular point cloud.</p><p><strong>Methods: </strong>We developed a fusion system that performs cross-source point cloud registration between ultrasound and computed tomography images, extracting the node, skeleton, and geomatic feature of the vascular point cloud. The system completes the fusion process in an average of 14.5 s after acquiring the vascular point clouds via ultrasound.</p><p><strong>Results: </strong>The experiments were conducted to fuse liver images by the dummy model and the healthy participants, respectively. The results show the proposed method achieved a registration error within 1.4 mm and decreased the target registration error significantly compared to other methods in a liver dummy model registration experiment. Furthermore, the proposed method achieved the averaged RMSE within 2.23 mm in a human liver vascular skeleton.</p><p><strong>Conclusion: </strong>The study concluded that because the registration method using vascular feature point cloud could realize the rapid and accurate fusion between ultrasound and computed tomography images, the method is useful to apply the real puncture surgery for radiofrequency ablation for liver. In future work, we will evaluate the proposed method by the patients.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":"2469-2478"},"PeriodicalIF":2.3,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12689734/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144977838","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Synthetic X-Q space learning for diffusion MRI parameter estimation: a pilot study in breast DKI. 用于扩散MRI参数估计的合成X-Q空间学习:乳腺DKI的初步研究。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-11-24 DOI: 10.1007/s11548-025-03550-7
Yoshitaka Masutani, Kousei Konya, Erina Kato, Naoko Mori, Hideki Ota, Shunji Mugikura, Kei Takase, Yuki Ichinoseki

Purpose: For diffusion MRI (dMRI) parameter estimation, machine-learning approaches have shown promising results so far including the synthetic Q-space learning (synQSL) based on regressor training with only synthetic data. In this study, we aimed at the development of a new method named synthetic X-Q space learning (synXQSL) to improve robustness and investigated the basic characteristics.

Methods: For training data, local parameter patterns of 3 × 3 voxels were synthesized by a linear combination of six bases, in which parameters are estimated at the center voxel. We prepared three types of local patterns by choosing the number of bases: flat, linear and quadratic. Then, at each location of 3 × 3 voxels, signal values of the diffusion-weighted image were computed by the signal model equation for diffusional kurtosis imaging and Rician noise simulation. The multi-layer perceptron was used for parameter estimation and was trained for each parameter with various noise levels. The level is controlled by a noise ratio defined as a fraction of the standard deviation in the Rician noise distribution normalized by the average b = 0 signal values. Experiments for visual and quantitative validation were performed with synthetic data, a digital phantom and clinical breast datasets in comparison with the previous methods.

Results: By using synthetic datasets, synXQSL outperformed synQSL in the parameter estimation of noisy data sets. Through the digital phantom experiments, the combination of synXQSL bases yields different results and a quadratic pattern could be the reasonable choice. The clinical data experiments indicate that synXQSL suppresses noises in estimated parameter maps and consequently brings higher contrast.

Conclusion: The basic characteristics of synXQSL were investigated by using various types of datasets. The results indicate that synXQSL with the appropriate choice of bases in training data synthesis has the potential to improve dMRI parameters in noisy datasets.

目的:对于扩散MRI (dMRI)参数估计,机器学习方法迄今为止已经显示出有希望的结果,包括基于仅使用合成数据的回归训练的合成q空间学习(synQSL)。在本研究中,我们旨在开发一种名为合成X-Q空间学习(synXQSL)的新方法来提高鲁棒性,并研究了其基本特征。方法:对于训练数据,通过6个碱基的线性组合合成3 × 3体素的局部参数模式,并在中心体素处估计参数。我们通过选择基底的数量制备了三种类型的局部图案:平面、线性和二次型。然后,在3 × 3体素的每个位置,利用扩散峰度成像信号模型方程计算扩散加权图像的信号值,并进行噪声模拟。多层感知器用于参数估计,并在不同噪声水平下对每个参数进行训练。电平由噪声比控制,噪声比定义为由平均b = 0信号值归一化的噪声分布中标准差的一部分。与之前的方法相比,使用合成数据、数字假体和临床乳房数据集进行视觉和定量验证实验。结果:在使用合成数据集时,synXQSL在噪声数据集参数估计方面优于synQSL。通过数字幻影实验,synXQSL碱基组合产生了不同的结果,二次型模式可能是合理的选择。临床数据实验表明,synXQSL可以抑制估计参数图中的噪声,从而提高对比度。结论:利用不同类型的数据集研究了synXQSL的基本特征。结果表明,在训练数据合成中适当选择碱基的synXQSL具有改善噪声数据集dMRI参数的潜力。
{"title":"Synthetic X-Q space learning for diffusion MRI parameter estimation: a pilot study in breast DKI.","authors":"Yoshitaka Masutani, Kousei Konya, Erina Kato, Naoko Mori, Hideki Ota, Shunji Mugikura, Kei Takase, Yuki Ichinoseki","doi":"10.1007/s11548-025-03550-7","DOIUrl":"10.1007/s11548-025-03550-7","url":null,"abstract":"<p><strong>Purpose: </strong>For diffusion MRI (dMRI) parameter estimation, machine-learning approaches have shown promising results so far including the synthetic Q-space learning (synQSL) based on regressor training with only synthetic data. In this study, we aimed at the development of a new method named synthetic X-Q space learning (synXQSL) to improve robustness and investigated the basic characteristics.</p><p><strong>Methods: </strong>For training data, local parameter patterns of 3 × 3 voxels were synthesized by a linear combination of six bases, in which parameters are estimated at the center voxel. We prepared three types of local patterns by choosing the number of bases: flat, linear and quadratic. Then, at each location of 3 × 3 voxels, signal values of the diffusion-weighted image were computed by the signal model equation for diffusional kurtosis imaging and Rician noise simulation. The multi-layer perceptron was used for parameter estimation and was trained for each parameter with various noise levels. The level is controlled by a noise ratio defined as a fraction of the standard deviation in the Rician noise distribution normalized by the average b = 0 signal values. Experiments for visual and quantitative validation were performed with synthetic data, a digital phantom and clinical breast datasets in comparison with the previous methods.</p><p><strong>Results: </strong>By using synthetic datasets, synXQSL outperformed synQSL in the parameter estimation of noisy data sets. Through the digital phantom experiments, the combination of synXQSL bases yields different results and a quadratic pattern could be the reasonable choice. The clinical data experiments indicate that synXQSL suppresses noises in estimated parameter maps and consequently brings higher contrast.</p><p><strong>Conclusion: </strong>The basic characteristics of synXQSL were investigated by using various types of datasets. The results indicate that synXQSL with the appropriate choice of bases in training data synthesis has the potential to improve dMRI parameters in noisy datasets.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":"2423-2435"},"PeriodicalIF":2.3,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12689713/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145589737","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MHAHF-UNet: a multi-scale hybrid attention hierarchy fusion network for carotid artery segmentation. MHAHF-UNet:用于颈动脉分割的多尺度混合注意层次融合网络。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-06-17 DOI: 10.1007/s11548-025-03449-3
Changshuo Jiang, Lin Gao, Wei Li, Maoyang Zou, Qingxiao Zheng, Xuhua Qiao

Purpose: Carotid plaque is an early manifestation of carotid atherosclerosis, and its accurate segmentation helps to assess cardiovascular disease risk. However, existing carotid artery segmentation algorithms are difficult to accurately capture the structural features of morphologically diverse plaques and lack effective utilization of multilayer features.

Methods: In order to solve the above problems, this paper proposes a multi-scale hybrid attention hierarchical fusion U-network structure (MHAHF-UNet) for segmenting ambiguous plaques in carotid artery images in order to improve the segmentation accuracy for complex structured images. The structure firstly introduces the median-enhanced orthogonal convolution module (MEOConv), which not only effectively suppresses the noise interference in ultrasound images, but also maintains the ability to perceive multi-scale features by combining the median-enhanced ternary channel mechanism and the depth-orthogonal convolution space mechanism. Secondly, it adopts the multi-fusion group convolutional gating module, which realizes the effective integration of shallow detailed features and deep semantic features through the adaptive control strategy of group convolution, and is able to flexibly regulate the transfer weights of features at different levels.

Results: Experiments show that the MHAHF-UNet model achieves a Dice coefficient of 82.46 ± 0.31 % and an IOU of 71.45 ± 0.37 % in the carotid artery segmentation task.

Conclusion: The model is expected to provide strong support for the prevention and treatment of cardiovascular diseases.

目的:颈动脉斑块是颈动脉粥样硬化的早期表现,其准确分割有助于评估心血管疾病的风险。然而,现有的颈动脉分割算法难以准确捕捉形态多样斑块的结构特征,缺乏对多层特征的有效利用。方法:针对上述问题,本文提出了一种多尺度混合关注层次融合u -网络结构(MHAHF-UNet)用于分割颈动脉图像中的模糊斑块,以提高对复杂结构图像的分割精度。该结构首先引入了中值增强正交卷积模块(MEOConv),该模块结合中值增强三元通道机制和深度正交卷积空间机制,不仅有效抑制超声图像中的噪声干扰,而且保持了对多尺度特征的感知能力。其次,采用多融合群卷积门控模块,通过群卷积自适应控制策略实现浅层细节特征与深层语义特征的有效融合,并能灵活调节不同层次特征的传递权值。结果:实验表明,MHAHF-UNet模型在颈动脉分割任务中的Dice系数为82.46±0.31%,IOU为71.45±0.37%。结论:该模型有望为心血管疾病的防治提供有力支持。
{"title":"MHAHF-UNet: a multi-scale hybrid attention hierarchy fusion network for carotid artery segmentation.","authors":"Changshuo Jiang, Lin Gao, Wei Li, Maoyang Zou, Qingxiao Zheng, Xuhua Qiao","doi":"10.1007/s11548-025-03449-3","DOIUrl":"10.1007/s11548-025-03449-3","url":null,"abstract":"<p><strong>Purpose: </strong>Carotid plaque is an early manifestation of carotid atherosclerosis, and its accurate segmentation helps to assess cardiovascular disease risk. However, existing carotid artery segmentation algorithms are difficult to accurately capture the structural features of morphologically diverse plaques and lack effective utilization of multilayer features.</p><p><strong>Methods: </strong>In order to solve the above problems, this paper proposes a multi-scale hybrid attention hierarchical fusion U-network structure (MHAHF-UNet) for segmenting ambiguous plaques in carotid artery images in order to improve the segmentation accuracy for complex structured images. The structure firstly introduces the median-enhanced orthogonal convolution module (MEOConv), which not only effectively suppresses the noise interference in ultrasound images, but also maintains the ability to perceive multi-scale features by combining the median-enhanced ternary channel mechanism and the depth-orthogonal convolution space mechanism. Secondly, it adopts the multi-fusion group convolutional gating module, which realizes the effective integration of shallow detailed features and deep semantic features through the adaptive control strategy of group convolution, and is able to flexibly regulate the transfer weights of features at different levels.</p><p><strong>Results: </strong>Experiments show that the MHAHF-UNet model achieves a Dice coefficient of <math><mrow><mn>82.46</mn> <mo>±</mo> <mn>0.31</mn> <mo>%</mo></mrow> </math> and an IOU of <math><mrow><mn>71.45</mn> <mo>±</mo> <mn>0.37</mn> <mo>%</mo></mrow> </math> in the carotid artery segmentation task.</p><p><strong>Conclusion: </strong>The model is expected to provide strong support for the prevention and treatment of cardiovascular diseases.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":"2541-2551"},"PeriodicalIF":2.3,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144318640","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Augmented reality in pelvic surgery: using an AR-headset as intraoperative radiation-free navigation tool. 增强现实在骨盆手术中的应用:使用ar头显作为术中无辐射导航工具。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-06-26 DOI: 10.1007/s11548-025-03462-6
Vincent K Schenk, Markus A Küper, Maximilian M Menger, Steven C Herath, Tina Histing, Christof K Audretsch

Purpose: The incidence of acetabular and pelvic fractures is rising significantly. Pelvic ring fractures rank as the sixth most common fractures in adults, with the majority occurring in the elderly. Due to complications related to surgical approaches, with rates of up to 31%, there is an increasing demand for minimally invasive surgical techniques. Augmented Reality (AR) has the potential to facilitate spatial orientation by a sophisticated user interface. The aim of this study was to develop an AR-based, radiation-free navigation system for pelvic fractures.

Methods: The Microsoft® HoloLens 2 was used as the AR headset. The Unity® game engine was used for programming. Pelvic models from Sawbones® served as the model. Segmentation was performed using Slicer3D by Slicer Corporation. The symphysis and both anterior superior iliac spines were defined as anatomical reference points. Ten pelvic models were used for testing. A preoperatively defined drill trajectory was displayed to the surgeon. A total of 20 S1 screws and 19 S2 screws were placed using only AR navigation without visual access to the pelvic model. Screw placement was controlled using CT.

Results: The matching process took an average of 3 min and 28 s. 18 out of 20 (90%) S1 screws and 3 out of 20 (15%) S2 screws were placed correctly. In most cases, no perforation occurred. The mean procedure time was 7 min for S1 screws and 5 min for S2 screws.

Conclusion: Proper drilling was achieved by displaying the trajectories via AR, particularly for S1 screws, where a slightly wider drilling corridor was aimed for compared to S2 screws. No registration scan was necessary with our matching method. No intraoperative radiation was required.

目的:髋臼和骨盆骨折的发生率明显上升。骨盆环骨折是成人第六大常见骨折,大多数发生在老年人中。由于与手术入路相关的并发症发生率高达31%,对微创手术技术的需求日益增加。增强现实(AR)具有通过复杂的用户界面促进空间定向的潜力。本研究的目的是开发一种基于ar的无辐射骨盆骨折导航系统。方法:采用Microsoft®HoloLens 2作为AR头显。使用Unity®游戏引擎进行编程。来自Sawbones®的骨盆模型作为模型。使用Slicer Corporation的Slicer3D进行分割。联合和髂前上棘被定义为解剖参考点。采用10个盆腔模型进行试验。将术前定义的钻孔轨迹显示给外科医生。共放置20枚S1螺钉和19枚S2螺钉,仅使用AR导航,不使用视觉进入骨盆模型。CT控制螺钉置入。结果:匹配过程平均耗时3 min 28 s, 20枚S1螺钉中有18枚(90%)正确放置,20枚S2螺钉中有3枚(15%)正确放置。在大多数情况下,没有发生穿孔。S1螺钉的平均手术时间为7分钟,S2螺钉为5分钟。结论:通过AR显示轨迹可以实现适当的钻孔,特别是S1螺钉,与S2螺钉相比,S1螺钉的钻孔通道略宽。我们的匹配方法不需要注册扫描。术中不需要放射治疗。
{"title":"Augmented reality in pelvic surgery: using an AR-headset as intraoperative radiation-free navigation tool.","authors":"Vincent K Schenk, Markus A Küper, Maximilian M Menger, Steven C Herath, Tina Histing, Christof K Audretsch","doi":"10.1007/s11548-025-03462-6","DOIUrl":"10.1007/s11548-025-03462-6","url":null,"abstract":"<p><strong>Purpose: </strong>The incidence of acetabular and pelvic fractures is rising significantly. Pelvic ring fractures rank as the sixth most common fractures in adults, with the majority occurring in the elderly. Due to complications related to surgical approaches, with rates of up to 31%, there is an increasing demand for minimally invasive surgical techniques. Augmented Reality (AR) has the potential to facilitate spatial orientation by a sophisticated user interface. The aim of this study was to develop an AR-based, radiation-free navigation system for pelvic fractures.</p><p><strong>Methods: </strong>The Microsoft® HoloLens 2 was used as the AR headset. The Unity® game engine was used for programming. Pelvic models from Sawbones® served as the model. Segmentation was performed using Slicer3D by Slicer Corporation. The symphysis and both anterior superior iliac spines were defined as anatomical reference points. Ten pelvic models were used for testing. A preoperatively defined drill trajectory was displayed to the surgeon. A total of 20 S1 screws and 19 S2 screws were placed using only AR navigation without visual access to the pelvic model. Screw placement was controlled using CT.</p><p><strong>Results: </strong>The matching process took an average of 3 min and 28 s. 18 out of 20 (90%) S1 screws and 3 out of 20 (15%) S2 screws were placed correctly. In most cases, no perforation occurred. The mean procedure time was 7 min for S1 screws and 5 min for S2 screws.</p><p><strong>Conclusion: </strong>Proper drilling was achieved by displaying the trajectories via AR, particularly for S1 screws, where a slightly wider drilling corridor was aimed for compared to S2 screws. No registration scan was necessary with our matching method. No intraoperative radiation was required.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":"2553-2563"},"PeriodicalIF":2.3,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12689752/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144499154","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards a zero-shot low-latency navigation for open surgery augmented reality applications. 面向开放手术增强现实应用的零射击低延迟导航。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-08-05 DOI: 10.1007/s11548-025-03480-4
Michael Schwimmbeck, Serouj Khajarian, Christopher Auer, Thomas Wittenberg, Stefanie Remmele

Purpose: Augmented reality (AR) enhances surgical navigation by superimposing visible anatomical structures with three-dimensional virtual models using head-mounted displays (HMDs). In particular, interventions such as open liver surgery can benefit from AR navigation, as it aids in identifying and distinguishing tumors and risk structures. However, there is a lack of automatic and markerless methods that are robust against real-world challenges, such as partial occlusion and organ motion.

Methods: We introduce a novel multi-device approach for automatic live navigation in open liver surgery that enhances the visualization and interaction capabilities of a HoloLens 2 HMD through precise and reliable registration using an Intel RealSense RGB-D camera. The intraoperative RGB-D segmentation and the preoperative CT data are utilized to register a virtual liver model to the target anatomy. An AR-prompted Segment Anything Model (SAM) enables robust segmentation of the liver in situ without the need for additional training data. To mitigate algorithmic latency, Double Exponential Smoothing (DES) is applied to forecast registration results.

Results: We conducted a phantom study for open liver surgery, investigating various scenarios of liver motion, viewpoints, and occlusion. The mean registration errors (8.31 mm-18.78 mm TRE) are comparable to those reported in prior work, while our approach demonstrates high success rates even for high occlusion factors and strong motion. Using forecasting, we bypassed the algorithmic latency of 79.8 ms per frame, with median forecasting errors below 2 mms and 1.5 degrees between the quaternions.

Conclusion: To our knowledge, this is the first work to approach markerless in situ visualization by combining a multi-device method with forecasting and a foundation model for segmentation and tracking. This enables a more reliable and precise AR registration of surgical targets with low latency. Our approach can be applied to other surgical applications and AR hardware with minimal effort.

目的:增强现实(AR)通过使用头戴式显示器(hmd)将可见解剖结构与三维虚拟模型叠加来增强手术导航。特别是,开放性肝手术等干预措施可以从AR导航中受益,因为它有助于识别和区分肿瘤和风险结构。然而,缺乏自动和无标记的方法来应对现实世界的挑战,如部分遮挡和器官运动。方法:我们介绍了一种新型的多设备自动实时导航方法,该方法通过使用英特尔RealSense RGB-D相机进行精确可靠的配准,增强了HoloLens 2 HMD的可视化和交互能力。利用术中RGB-D分割和术前CT数据将虚拟肝脏模型注册到目标解剖。ar提示的任何部分模型(SAM)可以在不需要额外训练数据的情况下对肝脏进行原位鲁棒分割。为了减少算法延迟,采用双指数平滑(DES)来预测配准结果。结果:我们进行了一项肝脏开放手术的幻像研究,调查了肝脏运动、视点和闭塞的各种情况。平均配准误差(8.31 mm-18.78 mm TRE)与之前报道的工作相当,而我们的方法即使在高遮挡因素和强运动下也显示出很高的成功率。使用预测,我们绕过了每帧79.8 ms的算法延迟,四元数之间的中位数预测误差低于2 mm和1.5度。结论:据我们所知,这是第一次将多设备方法与预测和分割和跟踪的基础模型相结合来实现无标记的原位可视化。这使得手术目标的AR登记更可靠和精确,延迟更低。我们的方法可以以最小的努力应用于其他外科应用和AR硬件。
{"title":"Towards a zero-shot low-latency navigation for open surgery augmented reality applications.","authors":"Michael Schwimmbeck, Serouj Khajarian, Christopher Auer, Thomas Wittenberg, Stefanie Remmele","doi":"10.1007/s11548-025-03480-4","DOIUrl":"10.1007/s11548-025-03480-4","url":null,"abstract":"<p><strong>Purpose: </strong>Augmented reality (AR) enhances surgical navigation by superimposing visible anatomical structures with three-dimensional virtual models using head-mounted displays (HMDs). In particular, interventions such as open liver surgery can benefit from AR navigation, as it aids in identifying and distinguishing tumors and risk structures. However, there is a lack of automatic and markerless methods that are robust against real-world challenges, such as partial occlusion and organ motion.</p><p><strong>Methods: </strong>We introduce a novel multi-device approach for automatic live navigation in open liver surgery that enhances the visualization and interaction capabilities of a HoloLens 2 HMD through precise and reliable registration using an Intel RealSense RGB-D camera. The intraoperative RGB-D segmentation and the preoperative CT data are utilized to register a virtual liver model to the target anatomy. An AR-prompted Segment Anything Model (SAM) enables robust segmentation of the liver in situ without the need for additional training data. To mitigate algorithmic latency, Double Exponential Smoothing (DES) is applied to forecast registration results.</p><p><strong>Results: </strong>We conducted a phantom study for open liver surgery, investigating various scenarios of liver motion, viewpoints, and occlusion. The mean registration errors (8.31 mm-18.78 mm TRE) are comparable to those reported in prior work, while our approach demonstrates high success rates even for high occlusion factors and strong motion. Using forecasting, we bypassed the algorithmic latency of 79.8 ms per frame, with median forecasting errors below 2 mms and 1.5 degrees between the quaternions.</p><p><strong>Conclusion: </strong>To our knowledge, this is the first work to approach markerless in situ visualization by combining a multi-device method with forecasting and a foundation model for segmentation and tracking. This enables a more reliable and precise AR registration of surgical targets with low latency. Our approach can be applied to other surgical applications and AR hardware with minimal effort.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":"2513-2521"},"PeriodicalIF":2.3,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12689743/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144790679","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Effect of simulator fidelity on skill acquisition and trainee satisfaction in arthroscopic surgery training for novices: a prospective randomized comparative study. 模拟器保真度对新手关节镜手术培训中技能习得和学员满意度的影响:一项前瞻性随机比较研究。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-10-09 DOI: 10.1007/s11548-025-03528-5
Tatsuhiko Kutsuna, Tomofumi Kinoshita, Kazunori Hino, Kunihiko Watamori, Takashi Tsuda, Shintaro Yamaoka, Masaki Takao

Purpose: In recent years, simulator-based training has gained attention as a safe and effective approach for surgical education. In orthopedic surgery, simulators for arthroscopic procedures, which require extensive practice to master, have been developed and shown to benefit clinical performance. This study aimed to evaluate the effectiveness of 2 types of arthroscopy simulators in medical education by assessing medical students without prior surgical training.

Methods: We prospectively and randomly evaluated the effectiveness and satisfaction of arthroscopy simulators using 2 devices: a high-fidelity virtual reality simulator (ArthroSim) and a low-fidelity simulator (AZBOTS). To assess simulator effectiveness, 28 first-year medical students received 90 min of training on either device. Their knee arthroscopy skills, including tasks such as visualizing and probing the medial compartment and cruciate ligaments, were assessed using ArthroSim. Skill acquisition was measured by procedure completion time and the Arthroscopic Surgery Skill Evaluation Tool Global Rating Scale. Additionally, to assess trainee satisfaction, 109 fifth-year medical students completed a 7-point Likert scale questionnaire evaluating spatial judgment, hand-eye coordination, camera navigation, instrument handling, and the overall knee arthroscopy training experience.

Results: No significant differences in skill acquisition were observed between the 2 simulator groups. Likewise, there were no significant differences in total questionnaire scores or in spatial judgment, hand-eye coordination, and camera navigation. However, participants using the high-fidelity simulator reported significantly greater satisfaction with instrument handling (p = 0.024) and the overall knee arthroscopy training experience (p = 0.033).

Conclusions: Although skill acquisition did not differ significantly between the high- and low-fidelity simulators after a single training session, the high-fidelity simulator markedly improved trainee satisfaction, especially in instrument handling and perceived quality of the training experience. These findings suggest that simulator fidelity enhances the educational value of the arthroscopic training for novices emphasizing the role of realistic simulation in early surgical education.

目的:近年来,以模拟器为基础的培训作为一种安全有效的外科教育方式受到重视。在骨科手术中,关节镜手术的模拟器需要大量的实践来掌握,已经开发并显示出有利于临床表现。本研究旨在评估两种类型的关节镜模拟器在医学教育中的有效性,通过评估未受过外科训练的医学生。方法:采用高保真虚拟现实模拟器(ArthroSim)和低保真模拟器(AZBOTS)两种设备,前瞻性和随机评估关节镜模拟器的有效性和满意度。为了评估模拟器的有效性,28名一年级医学生在任何一种设备上接受了90分钟的训练。他们的膝关节镜检查技能,包括观察和探测内侧隔室和交叉韧带等任务,使用ArthroSim进行评估。通过手术完成时间和关节镜手术技能评估工具全球评分量表来测量技能获得。此外,为了评估实习生满意度,109名五年级医学生完成了一份7分李克特量表,评估空间判断、手眼协调、相机导航、仪器处理和膝关节镜整体训练体验。结果:两个模拟器组在技能习得方面无显著差异。同样,在问卷总得分、空间判断、手眼协调和相机导航方面也没有显著差异。然而,使用高保真模拟器的参与者对器械操作(p = 0.024)和整体膝关节镜训练体验(p = 0.033)的满意度显著提高。结论:虽然高保真度模拟器和低保真度模拟器在单次培训后的技能习得没有显著差异,但高保真度模拟器显著提高了受训人员的满意度,特别是在仪器操作和培训体验的感知质量方面。这些研究结果表明,模拟器的逼真度提高了关节镜训练对新手的教育价值,强调了真实模拟在早期外科教育中的作用。
{"title":"Effect of simulator fidelity on skill acquisition and trainee satisfaction in arthroscopic surgery training for novices: a prospective randomized comparative study.","authors":"Tatsuhiko Kutsuna, Tomofumi Kinoshita, Kazunori Hino, Kunihiko Watamori, Takashi Tsuda, Shintaro Yamaoka, Masaki Takao","doi":"10.1007/s11548-025-03528-5","DOIUrl":"10.1007/s11548-025-03528-5","url":null,"abstract":"<p><strong>Purpose: </strong>In recent years, simulator-based training has gained attention as a safe and effective approach for surgical education. In orthopedic surgery, simulators for arthroscopic procedures, which require extensive practice to master, have been developed and shown to benefit clinical performance. This study aimed to evaluate the effectiveness of 2 types of arthroscopy simulators in medical education by assessing medical students without prior surgical training.</p><p><strong>Methods: </strong>We prospectively and randomly evaluated the effectiveness and satisfaction of arthroscopy simulators using 2 devices: a high-fidelity virtual reality simulator (ArthroSim) and a low-fidelity simulator (AZBOTS). To assess simulator effectiveness, 28 first-year medical students received 90 min of training on either device. Their knee arthroscopy skills, including tasks such as visualizing and probing the medial compartment and cruciate ligaments, were assessed using ArthroSim. Skill acquisition was measured by procedure completion time and the Arthroscopic Surgery Skill Evaluation Tool Global Rating Scale. Additionally, to assess trainee satisfaction, 109 fifth-year medical students completed a 7-point Likert scale questionnaire evaluating spatial judgment, hand-eye coordination, camera navigation, instrument handling, and the overall knee arthroscopy training experience.</p><p><strong>Results: </strong>No significant differences in skill acquisition were observed between the 2 simulator groups. Likewise, there were no significant differences in total questionnaire scores or in spatial judgment, hand-eye coordination, and camera navigation. However, participants using the high-fidelity simulator reported significantly greater satisfaction with instrument handling (p = 0.024) and the overall knee arthroscopy training experience (p = 0.033).</p><p><strong>Conclusions: </strong>Although skill acquisition did not differ significantly between the high- and low-fidelity simulators after a single training session, the high-fidelity simulator markedly improved trainee satisfaction, especially in instrument handling and perceived quality of the training experience. These findings suggest that simulator fidelity enhances the educational value of the arthroscopic training for novices emphasizing the role of realistic simulation in early surgical education.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":"2393-2401"},"PeriodicalIF":2.3,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12689780/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145253600","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Benchmarking NousNav: quantifying the spatial accuracy and clinical performance of an affordable, open-source neuronavigation system. 基准测试NousNav:量化空间精度和临床性能的负担得起的,开源的神经导航系统。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-08-15 DOI: 10.1007/s11548-025-03494-y
Colton Barr, Colin Galvin, Parikshit Juvekar, Erickson Torio, Samantha Horvath, Samantha Sadler, Annie Li, Ryan Bardsley, Tina Kapur, Steve Pieper, Sonia Pujol, Sarah Frisken, Gabor Fichtinger, Alexandra Golby

Purpose: NousNav is a low-cost, open-source neuronavigation platform built to address the high costs and resource limitations that hinder access to advanced neurosurgical technologies in low-resource settings. The low-cost and accessibility of the system is made possible using consumer-grade optical tracking and open-source software packages. This study aims to assess the performance of these core enabling technologies by quantifying their spatial accuracy and comparing it to a commercial gold standard.

Methods: A series of experiments were conducted to evaluate the capabilities of the selected hardware and registration infrastructure utilized in NousNav. Each component was tested both in a simulated bench-top environment and clinically across four brain tumor resection cases.

Results: The Optitrack Duo tracker used by NousNav was found to have a mean localization error of 0.8mm (SD 0.4mm). In bench-top phantom testing, NousNav had an average target registration error of 5.0mm (SD 2.3mm) following patient registration. Clinical evaluations revealed a mean distance of 4.2mm (SD 1.5mm) between points reported by NousNav versus those obtained using a commercial neuronavigation system.

Conclusion: These experiments highlight the role of baseline camera tracking performance, tracked instrument calibration, and patient positioning on the spatial performance of NousNav. They also provide an essential benchmark assessment of the system to help inform future clinical use-cases and direct ongoing system development.

目的:NousNav是一个低成本、开源的神经导航平台,旨在解决在低资源环境下阻碍先进神经外科技术获取的高成本和资源限制问题。该系统的低成本和可访问性是通过使用消费级光学跟踪和开源软件包实现的。本研究旨在通过量化其空间精度并将其与商业黄金标准进行比较来评估这些核心使能技术的性能。方法:通过一系列实验来评估NousNav中所选硬件和注册基础设施的性能。每个组件都在模拟的台式环境中进行了测试,并在四个脑肿瘤切除术病例中进行了临床测试。结果:NousNav使用的Optitrack Duo定位器平均定位误差为0.8mm (SD 0.4mm)。在台式幻影测试中,NousNav在患者注册后的平均目标注册误差为5.0mm (SD 2.3mm)。临床评估显示,与商用神经导航系统相比,NousNav报告的点之间的平均距离为4.2mm (SD 1.5mm)。结论:这些实验突出了基线摄像机跟踪性能、跟踪仪器校准和患者定位对NousNav空间性能的影响。它们还提供了系统的基本基准评估,以帮助告知未来的临床用例和指导正在进行的系统开发。
{"title":"Benchmarking NousNav: quantifying the spatial accuracy and clinical performance of an affordable, open-source neuronavigation system.","authors":"Colton Barr, Colin Galvin, Parikshit Juvekar, Erickson Torio, Samantha Horvath, Samantha Sadler, Annie Li, Ryan Bardsley, Tina Kapur, Steve Pieper, Sonia Pujol, Sarah Frisken, Gabor Fichtinger, Alexandra Golby","doi":"10.1007/s11548-025-03494-y","DOIUrl":"10.1007/s11548-025-03494-y","url":null,"abstract":"<p><strong>Purpose: </strong>NousNav is a low-cost, open-source neuronavigation platform built to address the high costs and resource limitations that hinder access to advanced neurosurgical technologies in low-resource settings. The low-cost and accessibility of the system is made possible using consumer-grade optical tracking and open-source software packages. This study aims to assess the performance of these core enabling technologies by quantifying their spatial accuracy and comparing it to a commercial gold standard.</p><p><strong>Methods: </strong>A series of experiments were conducted to evaluate the capabilities of the selected hardware and registration infrastructure utilized in NousNav. Each component was tested both in a simulated bench-top environment and clinically across four brain tumor resection cases.</p><p><strong>Results: </strong>The Optitrack Duo tracker used by NousNav was found to have a mean localization error of 0.8mm (SD 0.4mm). In bench-top phantom testing, NousNav had an average target registration error of 5.0mm (SD 2.3mm) following patient registration. Clinical evaluations revealed a mean distance of 4.2mm (SD 1.5mm) between points reported by NousNav versus those obtained using a commercial neuronavigation system.</p><p><strong>Conclusion: </strong>These experiments highlight the role of baseline camera tracking performance, tracked instrument calibration, and patient positioning on the spatial performance of NousNav. They also provide an essential benchmark assessment of the system to help inform future clinical use-cases and direct ongoing system development.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":"2529-2540"},"PeriodicalIF":2.3,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12469177/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144856940","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dynamic multi-scale deep learning with mixture of experts for differentiating iNPH and PSP using MRI. 基于专家混合的动态多尺度深度学习在MRI上区分iNPH和PSP。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-11-18 DOI: 10.1007/s11548-025-03537-4
Fubuki Sawa, Daisuke Fujita, Kenichi Shimada, Hideo Aihara, Toshiyuki Uehara, Yutaka Koide, Ryota Kawasaki, Kazunari Ishii, Syoji Kobashi

Purpose: Distinguishing idiopathic normal pressure hydrocephalus (iNPH) from progressive supranuclear palsy (PSP) presents a clinical challenge due to overlapping clinical symptoms such as gait disturbances and cognitive decline. This study presents a novel multi-scale deep learning framework that integrates global and local magnetic resonance imaging (MRI) features using a mixture of experts (MoE) mechanism, enhancing diagnostic accuracy and minimizing interobserver variability.

Methods: The proposed framework combines a 3D convolutional neural network (CNN) for capturing global volumetric features with a 2.5D recurrent CNN focusing on disease-specific regions of interest (ROIs), including the lateral ventricles, high convexity sulci, midbrain, and Sylvian fissures. The MoE mechanism dynamically weights global and local features, optimizing the classification process. Model performance was assessed using stratified fivefold cross-validation on T1-weighted MRI from 118 patients (53 iNPH, 65 PSP) to ensure balanced class distributions across training folds.

Results: The MoE model using ResNet-34 achieved an accuracy of 0.983 (95% CI 0.875-1.000), a recall of 0.985 (95% CI 0.750-1.000), a precision of 0.986 (95% CI 0.769-1.000), and an area under the curve (AUC) of 1.000 (95% CI 1.000-1.000), outperforming traditional morphological markers and single-branch deep learning models. The MoE mechanism allowed adaptive weighting of global and local features, contributing to both improved robustness and interpretability. Grad-CAM visualizations highlighted disease-specific regions, demonstrating that the model focused on relevant features in both successful and failure modes of the 3D CNN expert for iNPH and PSP.

Conclusion: The dynamic integration of global and local MRI features through the MoE framework offers a powerful, robust, and interpretable tool for differentiating iNPH from PSP. This approach reduces reliance on subjective visual assessments and has the potential for broader clinical application through dataset expansion and multicenter validation.

目的:区分特发性正常压力脑积水(iNPH)和进行性核上性麻痹(PSP)是一项临床挑战,因为它们有重叠的临床症状,如步态障碍和认知能力下降。本研究提出了一种新的多尺度深度学习框架,该框架使用混合专家(MoE)机制集成了全局和局部磁共振成像(MRI)特征,提高了诊断准确性并最大限度地减少了观察者之间的可变性。方法:提出的框架结合了用于捕获全局体积特征的3D卷积神经网络(CNN)和专注于疾病特异性感兴趣区域(roi)的2.5D复发CNN,包括侧脑室、高凸沟、中脑和Sylvian裂缝。MoE机制动态加权全局和局部特征,优化分类过程。通过118例患者(53例iNPH, 65例PSP)的t1加权MRI分层五重交叉验证来评估模型的性能,以确保训练折叠间的类别分布平衡。结果:基于ResNet-34的MoE模型准确率为0.983 (95% CI 0.875-1.000),召回率为0.985 (95% CI 0.750-1.000),精密度为0.986 (95% CI 0.769-1.000),曲线下面积(AUC)为1.000 (95% CI 1.000-1.000),优于传统形态学标记和单分支深度学习模型。MoE机制允许对全局和局部特征进行自适应加权,有助于提高鲁棒性和可解释性。Grad-CAM可视化突出了疾病特异性区域,表明该模型在iNPH和PSP的3D CNN专家的成功和失败模式中都关注了相关特征。结论:通过MoE框架动态整合整体和局部MRI特征为区分iNPH和PSP提供了一个强大、可靠和可解释的工具。这种方法减少了对主观视觉评估的依赖,并且通过数据集扩展和多中心验证具有更广泛的临床应用潜力。
{"title":"Dynamic multi-scale deep learning with mixture of experts for differentiating iNPH and PSP using MRI.","authors":"Fubuki Sawa, Daisuke Fujita, Kenichi Shimada, Hideo Aihara, Toshiyuki Uehara, Yutaka Koide, Ryota Kawasaki, Kazunari Ishii, Syoji Kobashi","doi":"10.1007/s11548-025-03537-4","DOIUrl":"10.1007/s11548-025-03537-4","url":null,"abstract":"<p><strong>Purpose: </strong>Distinguishing idiopathic normal pressure hydrocephalus (iNPH) from progressive supranuclear palsy (PSP) presents a clinical challenge due to overlapping clinical symptoms such as gait disturbances and cognitive decline. This study presents a novel multi-scale deep learning framework that integrates global and local magnetic resonance imaging (MRI) features using a mixture of experts (MoE) mechanism, enhancing diagnostic accuracy and minimizing interobserver variability.</p><p><strong>Methods: </strong>The proposed framework combines a 3D convolutional neural network (CNN) for capturing global volumetric features with a 2.5D recurrent CNN focusing on disease-specific regions of interest (ROIs), including the lateral ventricles, high convexity sulci, midbrain, and Sylvian fissures. The MoE mechanism dynamically weights global and local features, optimizing the classification process. Model performance was assessed using stratified fivefold cross-validation on T1-weighted MRI from 118 patients (53 iNPH, 65 PSP) to ensure balanced class distributions across training folds.</p><p><strong>Results: </strong>The MoE model using ResNet-34 achieved an accuracy of 0.983 (95% CI 0.875-1.000), a recall of 0.985 (95% CI 0.750-1.000), a precision of 0.986 (95% CI 0.769-1.000), and an area under the curve (AUC) of 1.000 (95% CI 1.000-1.000), outperforming traditional morphological markers and single-branch deep learning models. The MoE mechanism allowed adaptive weighting of global and local features, contributing to both improved robustness and interpretability. Grad-CAM visualizations highlighted disease-specific regions, demonstrating that the model focused on relevant features in both successful and failure modes of the 3D CNN expert for iNPH and PSP.</p><p><strong>Conclusion: </strong>The dynamic integration of global and local MRI features through the MoE framework offers a powerful, robust, and interpretable tool for differentiating iNPH from PSP. This approach reduces reliance on subjective visual assessments and has the potential for broader clinical application through dataset expansion and multicenter validation.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":"2413-2422"},"PeriodicalIF":2.3,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145551804","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Correction to: Micro-robotic percutaneous targeting of type II endoleaks in the angio-suite. 更正:微型机器人经皮定位血管套房中的 II 型内漏。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 DOI: 10.1007/s11548-024-03271-3
Gerlig Widmann, Johannes Deeg, Andreas Frech, Josef Klocker, Gudrun Feuchtner, Martin Freund
{"title":"Correction to: Micro-robotic percutaneous targeting of type II endoleaks in the angio-suite.","authors":"Gerlig Widmann, Johannes Deeg, Andreas Frech, Josef Klocker, Gudrun Feuchtner, Martin Freund","doi":"10.1007/s11548-024-03271-3","DOIUrl":"10.1007/s11548-024-03271-3","url":null,"abstract":"","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":"2565"},"PeriodicalIF":2.3,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12689722/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142590891","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International Journal of Computer Assisted Radiology and Surgery
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1