首页 > 最新文献

International Journal of Computer Assisted Radiology and Surgery最新文献

英文 中文
Investigating workload and usability of remote magnetic navigation for catheter ablation. 研究远程磁导航用于导管消融的工作量和可用性。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-15 DOI: 10.1007/s11548-025-03558-z
Florian Heemeyer, Leonardo E Guido Lopez, Miguel E Jáuregui Abularach, Beatriz Sanz Verdejo, Quentin Boehler, Oliver Brinkmann, José L Merino, Bradley J Nelson

Purpose: Robotic systems for catheter ablation have been in clinical use for many years. While their impact on the clinical outcome and procedure times is well studied, aspects like usability and operator workload have received limited attention in the literature. Reduced workload and stress levels benefit the operator's mental and physical health, and can also lower the risk of errors and ultimately improve patient safety. The aim of this study is to investigate the workload and usability of remote magnetic navigation compared to conventional manual navigation.

Methods: We performed a user study with eight electrophysiologists. Each participant performed identical in-vitro navigation tasks replicating those found in pulmonary vein isolation using both manual and magnetic navigation. Magnetic navigation experiments were performed using the Navion, a mobile electromagnetic navigation system.

Results: Magnetic navigation significantly improved usability (p < 0.02) and workload (p < 0.01) compared to manual navigation, measured using the System Usability Scale (magnetic: 85.6 ± 9.3 vs. manual: 75.0 ± 17.8) and NASA Task Load Index (magnetic: 72.4 ± 13.5 vs. manual: 45.8 ± 16.7). Additionally, task completion times were shorter (p < 0.01) with magnetic navigation (284.6 ± 80.7 s) compared to manual navigation (411.0 ± 123.7 s).

Conclusion: The findings of this study suggest that remote magnetic navigation using the Navion significantly improves operator experiences in terms of workload and usability, reinforcing the case for wider adoption of well-designed robotic systems in cardiac electrophysiology labs.

目的:机器人导管消融系统已在临床应用多年。虽然它们对临床结果和手术时间的影响已经得到了很好的研究,但诸如可用性和操作员工作量等方面在文献中受到的关注有限。减少工作量和压力水平有利于操作人员的身心健康,还可以降低出错的风险,最终提高患者的安全。本研究的目的是研究远程磁导航与传统手动导航相比的工作量和可用性。方法:我们与8位电生理学家进行了一项用户研究。每个参与者都执行了相同的体外导航任务,复制了在肺静脉隔离中使用手动和磁导航的任务。利用移动电磁导航系统Navion进行了磁导航实验。结论:本研究的结果表明,使用navon的远程磁导航在工作量和可用性方面显着改善了操作员的体验,加强了在心脏电生理实验室中广泛采用精心设计的机器人系统的案例。
{"title":"Investigating workload and usability of remote magnetic navigation for catheter ablation.","authors":"Florian Heemeyer, Leonardo E Guido Lopez, Miguel E Jáuregui Abularach, Beatriz Sanz Verdejo, Quentin Boehler, Oliver Brinkmann, José L Merino, Bradley J Nelson","doi":"10.1007/s11548-025-03558-z","DOIUrl":"https://doi.org/10.1007/s11548-025-03558-z","url":null,"abstract":"<p><strong>Purpose: </strong>Robotic systems for catheter ablation have been in clinical use for many years. While their impact on the clinical outcome and procedure times is well studied, aspects like usability and operator workload have received limited attention in the literature. Reduced workload and stress levels benefit the operator's mental and physical health, and can also lower the risk of errors and ultimately improve patient safety. The aim of this study is to investigate the workload and usability of remote magnetic navigation compared to conventional manual navigation.</p><p><strong>Methods: </strong>We performed a user study with eight electrophysiologists. Each participant performed identical in-vitro navigation tasks replicating those found in pulmonary vein isolation using both manual and magnetic navigation. Magnetic navigation experiments were performed using the Navion, a mobile electromagnetic navigation system.</p><p><strong>Results: </strong>Magnetic navigation significantly improved usability (p < 0.02) and workload (p < 0.01) compared to manual navigation, measured using the System Usability Scale (magnetic: 85.6 ± 9.3 vs. manual: 75.0 ± 17.8) and NASA Task Load Index (magnetic: 72.4 ± 13.5 vs. manual: 45.8 ± 16.7). Additionally, task completion times were shorter (p < 0.01) with magnetic navigation (284.6 ± 80.7 s) compared to manual navigation (411.0 ± 123.7 s).</p><p><strong>Conclusion: </strong>The findings of this study suggest that remote magnetic navigation using the Navion significantly improves operator experiences in terms of workload and usability, reinforcing the case for wider adoption of well-designed robotic systems in cardiac electrophysiology labs.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2025-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145764582","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Reshapeit: reliable shape interaction with implicit template for medical anatomy reconstruction. shapeit:可靠的形状交互与隐式模板用于医学解剖重建。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-13 DOI: 10.1007/s11548-025-03557-0
Minghui Zhang, Yun Gu

Purpose: Shape modeling of volumetric medical images plays a crucial role in quantitative analysis and surgical planning for computer-aided diagnosis. However, automatic shape reconstruction from deep learning models often suffers from limited image resolution and the lack of shape prior constraints. This study aims to address these challenges by developing a method that enables reliable and accurate anatomical shape modeling in the continuous space.

Methods: We present the Reliable Shape Interaction with Implicit Template (ReShapeIT) network, which represents anatomical structures using continuous implicit fields rather than discrete voxel grids. The approach combines a category-specific implicit template field with a deformation network to encode anatomical shapes from training shapes. In addition, a Template Interaction Module (TIM) is designed to refine test cases by aligning learned template shapes with instance-specific latent codes.

Results: We evaluated ReShapeIT on three anatomical datasets-Liver, Pancreas, and Lung Lobe. The proposed method outperforms state-of-the-art approaches in 3D shape reconstruction, achieving Chamfer Distance/Earth Mover's Distance scores of 0.225/0.318 for Liver, 0.125/0.067 for Pancreas, and 0.414/0.098 for Lung Lobe.

Conclusion: ReShapeIT provides a reliable and generalizable solution for implicit anatomical shape modeling by leveraging shared template priors and instance-level deformations. The implementation is publicly available at: https://github.com/EndoluminalSurgicalVision-IMR/ReShapeIT .

目的:体积医学图像的形状建模在计算机辅助诊断的定量分析和手术计划中起着至关重要的作用。然而,深度学习模型的自动形状重建经常受到图像分辨率有限和缺乏形状先验约束的影响。本研究旨在通过开发一种在连续空间中实现可靠和准确解剖形状建模的方法来解决这些挑战。方法:我们提出了可靠的形状交互隐式模板(ReShapeIT)网络,该网络使用连续隐式场而不是离散体素网格来表示解剖结构。该方法将特定类别的隐式模板域与变形网络相结合,从训练形状中编码解剖形状。此外,还设计了一个模板交互模块(TIM),通过将学习到的模板形状与实例特定的潜在代码对齐来改进测试用例。结果:我们在三个解剖数据集——肝脏、胰腺和肺叶上评估了repeit。该方法在三维形状重建方面优于最先进的方法,肝脏的Chamfer Distance/Earth Mover’s Distance评分为0.225/0.318,胰腺为0.125/0.067,肺叶为0.414/0.098。结论:通过利用共享模板先验和实例级变形,ReShapeIT为隐式解剖形状建模提供了可靠且可推广的解决方案。该实现可在:https://github.com/EndoluminalSurgicalVision-IMR/ReShapeIT上公开获得。
{"title":"Reshapeit: reliable shape interaction with implicit template for medical anatomy reconstruction.","authors":"Minghui Zhang, Yun Gu","doi":"10.1007/s11548-025-03557-0","DOIUrl":"https://doi.org/10.1007/s11548-025-03557-0","url":null,"abstract":"<p><strong>Purpose: </strong>Shape modeling of volumetric medical images plays a crucial role in quantitative analysis and surgical planning for computer-aided diagnosis. However, automatic shape reconstruction from deep learning models often suffers from limited image resolution and the lack of shape prior constraints. This study aims to address these challenges by developing a method that enables reliable and accurate anatomical shape modeling in the continuous space.</p><p><strong>Methods: </strong>We present the Reliable Shape Interaction with Implicit Template (ReShapeIT) network, which represents anatomical structures using continuous implicit fields rather than discrete voxel grids. The approach combines a category-specific implicit template field with a deformation network to encode anatomical shapes from training shapes. In addition, a Template Interaction Module (TIM) is designed to refine test cases by aligning learned template shapes with instance-specific latent codes.</p><p><strong>Results: </strong>We evaluated ReShapeIT on three anatomical datasets-Liver, Pancreas, and Lung Lobe. The proposed method outperforms state-of-the-art approaches in 3D shape reconstruction, achieving Chamfer Distance/Earth Mover's Distance scores of 0.225/0.318 for Liver, 0.125/0.067 for Pancreas, and 0.414/0.098 for Lung Lobe.</p><p><strong>Conclusion: </strong>ReShapeIT provides a reliable and generalizable solution for implicit anatomical shape modeling by leveraging shared template priors and instance-level deformations. The implementation is publicly available at: https://github.com/EndoluminalSurgicalVision-IMR/ReShapeIT .</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2025-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145745328","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Machine learning-based treatment outcome prediction in head and neck cancer using integrated noninvasive diagnostics. 基于机器学习的综合无创诊断头颈癌治疗结果预测。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-08 DOI: 10.1007/s11548-025-03539-2
Melda Yeghaian, Stefano Trebeschi, Marina Herrero-Huertas, Francisco Javier Mendoza Ferradás, Paula Bos, Maarten J A van Alphen, Marcel A J van Gerven, Regina G H Beets-Tan, Zuhir Bodalal, Lilly-Ann van der Velden

Purpose: Accurate prediction of treatment outcomes is crucial for personalized treatment in head and neck squamous cell carcinoma (HNSCC). Beyond one-year survival, assessing long-term enteral nutrition dependence is essential for optimizing patient counseling and resource allocation. This preliminary study aimed to predict one-year survival and feeding tube dependence in surgically treated HNSCC patients using classical machine learning.

Methods: This proof-of-principle retrospective study included 558 surgically treated HNSCC patients. Baseline clinical data, routine blood markers, and MRI-based radiomic features were collected before treatment. Additional postsurgical treatments within one year were also recorded. Random forest classifiers were trained to predict one-year survival and feeding tube dependence. Model explainability was assessed using Shapley Additive exPlanation (SHAP) values.

Results: Using tenfold stratified cross-validation, clinical data showed the highest predictive performance for survival (AUC = 0.75 ± 0.10; p < 0.001). Blood (AUC = 0.67 ± 0.17; p = 0.001) and imaging (AUC = 0.68 ± 0.16; p = 0.26) showed moderate performance, and multimodal integration did not improve predictions (AUC = 0.68 ± 0.16; p = 0.38). For feeding tube dependence, all modalities had low predictive power (AUC ≤ 0.66; p > 0.05). However, postsurgical treatment information outperformed all other modalities (AUC = 0.67 ± 0.07; p = 0.002), but had the lowest predictive value for survival (AUC = 0.57 ± 0.11; p = 0.08).

Conclusion: Clinical data appeared to be the strongest predictor of one-year survival in surgically treated HNSCC, although overall predictive performance was moderate. Postsurgical treatment information played a key role in predicting tube feeding dependence. While multimodal integration did not enhance overall model performance, it showed modest gains for weaker individual modalities, suggesting potential complementarity that warrants further investigation.

目的:准确预测治疗结果对头颈部鳞状细胞癌(HNSCC)的个性化治疗至关重要。超过一年的生存期,评估长期肠内营养依赖对于优化患者咨询和资源分配至关重要。这项初步研究旨在使用经典机器学习预测手术治疗的HNSCC患者的一年生存率和饲管依赖性。方法:这项原理证明回顾性研究包括558例手术治疗的HNSCC患者。治疗前收集基线临床数据、常规血液标志物和基于mri的放射学特征。术后一年内的其他治疗情况也有记录。随机森林分类器被训练来预测一年生存率和饲管依赖性。采用Shapley加性解释(SHAP)值评估模型的可解释性。结果:采用十倍分层交叉验证,临床数据显示最高的生存预测性能(AUC = 0.75±0.10;p 0.05)。然而,术后治疗信息优于所有其他方式(AUC = 0.67±0.07;p = 0.002),但对生存的预测值最低(AUC = 0.57±0.11;p = 0.08)。结论:临床数据似乎是手术治疗的HNSCC一年生存率的最强预测指标,尽管总体预测性能一般。术后治疗信息是预测管饲依赖的关键因素。虽然多模态集成并没有提高整体模型的性能,但对于较弱的单个模态,它显示出适度的增益,这表明潜在的互补性值得进一步研究。
{"title":"Machine learning-based treatment outcome prediction in head and neck cancer using integrated noninvasive diagnostics.","authors":"Melda Yeghaian, Stefano Trebeschi, Marina Herrero-Huertas, Francisco Javier Mendoza Ferradás, Paula Bos, Maarten J A van Alphen, Marcel A J van Gerven, Regina G H Beets-Tan, Zuhir Bodalal, Lilly-Ann van der Velden","doi":"10.1007/s11548-025-03539-2","DOIUrl":"https://doi.org/10.1007/s11548-025-03539-2","url":null,"abstract":"<p><strong>Purpose: </strong>Accurate prediction of treatment outcomes is crucial for personalized treatment in head and neck squamous cell carcinoma (HNSCC). Beyond one-year survival, assessing long-term enteral nutrition dependence is essential for optimizing patient counseling and resource allocation. This preliminary study aimed to predict one-year survival and feeding tube dependence in surgically treated HNSCC patients using classical machine learning.</p><p><strong>Methods: </strong>This proof-of-principle retrospective study included 558 surgically treated HNSCC patients. Baseline clinical data, routine blood markers, and MRI-based radiomic features were collected before treatment. Additional postsurgical treatments within one year were also recorded. Random forest classifiers were trained to predict one-year survival and feeding tube dependence. Model explainability was assessed using Shapley Additive exPlanation (SHAP) values.</p><p><strong>Results: </strong>Using tenfold stratified cross-validation, clinical data showed the highest predictive performance for survival (AUC = 0.75 ± 0.10; p < 0.001). Blood (AUC = 0.67 ± 0.17; p = 0.001) and imaging (AUC = 0.68 ± 0.16; p = 0.26) showed moderate performance, and multimodal integration did not improve predictions (AUC = 0.68 ± 0.16; p = 0.38). For feeding tube dependence, all modalities had low predictive power (AUC ≤ 0.66; p > 0.05). However, postsurgical treatment information outperformed all other modalities (AUC = 0.67 ± 0.07; p = 0.002), but had the lowest predictive value for survival (AUC = 0.57 ± 0.11; p = 0.08).</p><p><strong>Conclusion: </strong>Clinical data appeared to be the strongest predictor of one-year survival in surgically treated HNSCC, although overall predictive performance was moderate. Postsurgical treatment information played a key role in predicting tube feeding dependence. While multimodal integration did not enhance overall model performance, it showed modest gains for weaker individual modalities, suggesting potential complementarity that warrants further investigation.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2025-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145702875","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Glioblastoma survival prediction through MRI and clinical data integration with transfer learning. 通过MRI和临床数据整合迁移学习预测胶质母细胞瘤的生存。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-04 DOI: 10.1007/s11548-025-03548-1
A Marasi, D Milesi, D Aquino, F M Doniselli, R Pascuzzo, M Grisoli, A Redaelli, E De Momi

Purpose: Accurate prediction of overall survival (OS) in glioblastoma patients is critical for advancing personalized treatments and improving clinical trial design. Conventional radiomics approaches rely on manually engineered features, which limit their ability to capture complex, high-dimensional imaging patterns. This study employs a deep learning architecture to process MRI data for automated glioma segmentation and feature extraction, leveraging high-level representations from the encoder's latent space.

Methods: Multimodal MRI data from the BraTS2020 dataset and a proprietary dataset from Fondazione IRCCS Istituto Neurologico Carlo Besta (Milan, Italy) were processed independently using a U-Net-like model pre-trained on BraTS2018 and fine-tuned on BraTS2020. Features extracted from the encoder's latent space represented hierarchical imaging patterns. These features were combined with clinical variable (patient's age) and reduced via principal component analysis (PCA) to enhance computational efficiency. Machine learning classifiers-including random forest, XGBoost, and a fully connected neural network-were trained on the reduced feature vectors for OS classification.

Results: In the four-modality BraTS4CH setting, the multi-layer perceptron achieved the best performance (F1 = 0.71, AUC = 0.74, accuracy = 0.71). When limited to two modalities on BraTS2020 (BraTS2CH), MLP again led (F1 = 0.67, AUC = 0.70, accuracy = 0.67). On the IRCCS Besta two-modality cohort (Besta2CH), XGBoost produced the highest F1-score and accuracy (F1 = 0.65, accuracy = 0.66), while MLP obtained the top AUC (0.70). These results are competitive with-and in some metrics exceed-state-of-the-art reports, demonstrating the robustness and scalability of our automated framework relative to traditional radiomics and AI-driven approaches.

Conclusion: Integrating encoder-derived features from multimodal MRI data with clinical variables offers a scalable and effective approach for OS prediction in glioblastoma patients. This study demonstrates the potential of deep learning to address traditional radiomics limitations, paving the way for more precise and personalized prognostic tools.

目的:准确预测胶质母细胞瘤患者的总生存期(OS)对于推进个性化治疗和改进临床试验设计至关重要。传统的放射组学方法依赖于人工设计的特征,这限制了它们捕捉复杂、高维成像模式的能力。本研究采用深度学习架构来处理MRI数据,以实现自动胶质瘤分割和特征提取,利用编码器潜在空间的高级表示。方法:使用在BraTS2018上预训练并在BraTS2020上微调的类似u - net的模型,对来自BraTS2020数据集和来自Fondazione IRCCS Istituto Neurologico Carlo Besta (Milan, Italy)的专有数据集的多模态MRI数据进行独立处理。从编码器的潜在空间中提取的特征表示分层成像模式。这些特征与临床变量(患者年龄)相结合,并通过主成分分析(PCA)减少以提高计算效率。机器学习分类器——包括随机森林、XGBoost和一个完全连接的神经网络——在简化的特征向量上进行训练,用于OS分类。结果:在四模态BraTS4CH设置下,多层感知器的性能最佳(F1 = 0.71, AUC = 0.74,准确率= 0.71)。当限于BraTS2020 (BraTS2CH)的两种模式时,MLP再次领先(F1 = 0.67, AUC = 0.70,准确率= 0.67)。在IRCCS best双模队列(Besta2CH)中,XGBoost获得最高的F1得分和准确率(F1 = 0.65,准确率= 0.66),而MLP获得最高的AUC(0.70)。这些结果与最先进的报告具有竞争力,在某些指标上甚至超过了最先进的报告,证明了我们的自动化框架相对于传统放射组学和人工智能驱动方法的稳健性和可扩展性。结论:将多模态MRI数据的编码器衍生特征与临床变量相结合,为胶质母细胞瘤患者的OS预测提供了一种可扩展且有效的方法。这项研究展示了深度学习解决传统放射组学限制的潜力,为更精确和个性化的预后工具铺平了道路。
{"title":"Glioblastoma survival prediction through MRI and clinical data integration with transfer learning.","authors":"A Marasi, D Milesi, D Aquino, F M Doniselli, R Pascuzzo, M Grisoli, A Redaelli, E De Momi","doi":"10.1007/s11548-025-03548-1","DOIUrl":"https://doi.org/10.1007/s11548-025-03548-1","url":null,"abstract":"<p><strong>Purpose: </strong>Accurate prediction of overall survival (OS) in glioblastoma patients is critical for advancing personalized treatments and improving clinical trial design. Conventional radiomics approaches rely on manually engineered features, which limit their ability to capture complex, high-dimensional imaging patterns. This study employs a deep learning architecture to process MRI data for automated glioma segmentation and feature extraction, leveraging high-level representations from the encoder's latent space.</p><p><strong>Methods: </strong>Multimodal MRI data from the BraTS2020 dataset and a proprietary dataset from Fondazione IRCCS Istituto Neurologico Carlo Besta (Milan, Italy) were processed independently using a U-Net-like model pre-trained on BraTS2018 and fine-tuned on BraTS2020. Features extracted from the encoder's latent space represented hierarchical imaging patterns. These features were combined with clinical variable (patient's age) and reduced via principal component analysis (PCA) to enhance computational efficiency. Machine learning classifiers-including random forest, XGBoost, and a fully connected neural network-were trained on the reduced feature vectors for OS classification.</p><p><strong>Results: </strong>In the four-modality BraTS4CH setting, the multi-layer perceptron achieved the best performance (F1 = 0.71, AUC = 0.74, accuracy = 0.71). When limited to two modalities on BraTS2020 (BraTS2CH), MLP again led (F1 = 0.67, AUC = 0.70, accuracy = 0.67). On the IRCCS Besta two-modality cohort (Besta2CH), XGBoost produced the highest F1-score and accuracy (F1 = 0.65, accuracy = 0.66), while MLP obtained the top AUC (0.70). These results are competitive with-and in some metrics exceed-state-of-the-art reports, demonstrating the robustness and scalability of our automated framework relative to traditional radiomics and AI-driven approaches.</p><p><strong>Conclusion: </strong>Integrating encoder-derived features from multimodal MRI data with clinical variables offers a scalable and effective approach for OS prediction in glioblastoma patients. This study demonstrates the potential of deep learning to address traditional radiomics limitations, paving the way for more precise and personalized prognostic tools.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2025-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145670828","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
In-depth characterization of a laparoscopic radical prostatectomy procedure based on surgical process modeling. 基于手术过程建模的腹腔镜根治性前列腺切除术的深入表征。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-03 DOI: 10.1007/s11548-025-03552-5
Nuno S Rodrigues, Pedro Morais, Lukas R Buschle, Estevão Lima, João L Vilaça

Purpose: Minimally invasive surgical approaches are currently the standard of care for men with prostate cancer, presenting higher rates of erectile function preservation. With these laparoscopic techniques, there is an increasing amount of data and information available. Adaptive systems can play an important role, acting as an intelligent information filter, assuring that all the available information can become useful for the procedure and not overwhelming for the surgeon. Standardizing and structuring the surgical workflow are key requirements for such smart assistants to recognize the different surgical steps through context information about the environment. This work aims to do a detailed characterization of a laparoscopic radical prostatectomy procedure, focusing on the formalization of medical expert knowledge, via surgical process modeling.

Methods: Data were acquired manually, via online and offline observation, and discussion with medical experts. A total of 14 procedures were observed. Both manual laparoscopic radical prostatectomy and robot-assisted laparoscopic prostatectomy were studied. The derived SPM focuses only on the intraoperatory part of the procedure, with constant feedback from the endoscopic camera. For surgery observation, a dedicated Excel template was developed.

Results: The final model is represented in a descriptive and numerical format, combining task description with a workflow diagram arrangement for ease of interpretation. Practical applications of the generated surgical process model are exemplified with the creation of activation trees for surgical phase identification. Anatomical structures are reported for each phase, distinguishing between visible and inferable ones. Additionally, the surgeons involved are identified, surgical instruments, and actions performed in each phase. A total of 11 phases were identified and characterized. Average surgery duration is 87 min.

Conclusion: The generated surgical process model is a first step toward the development of a context-aware surgical assistant and can potentially be used as a roadmap by other research teams, operating room managers and surgical teams.

目的:微创手术入路是目前前列腺癌患者的标准治疗方法,具有较高的勃起功能保留率。有了这些腹腔镜技术,有越来越多的数据和信息可用。自适应系统可以发挥重要作用,充当智能信息过滤器,确保所有可用信息对手术有用,而不会让外科医生不知所措。手术工作流程的标准化和结构化是智能助手通过环境上下文信息识别不同手术步骤的关键要求。本工作旨在通过手术过程建模,对腹腔镜根治性前列腺切除术进行详细描述,重点关注医学专家知识的形式化。方法:人工采集资料,通过线上、线下观察,并与医学专家讨论。共观察了14个手术过程。对人工腹腔镜根治性前列腺切除术和机器人辅助腹腔镜前列腺切除术进行了研究。导出的SPM仅关注术中部分的过程,并从内窥镜相机不断反馈。为手术观察,开发了专用的Excel模板。结果:最终模型以描述性和数字格式表示,将任务描述与易于解释的工作流图安排相结合。通过创建用于手术阶段识别的激活树,举例说明了生成的手术过程模型的实际应用。解剖结构报告了每个阶段,区分可见的和可推断的。此外,在每个阶段确定涉及的外科医生、手术器械和操作。共鉴定和表征了11个相。平均手术时间为87分钟。结论:所生成的手术过程模型是开发情境感知手术助手的第一步,可能被其他研究团队、手术室管理人员和手术团队用作路线图。
{"title":"In-depth characterization of a laparoscopic radical prostatectomy procedure based on surgical process modeling.","authors":"Nuno S Rodrigues, Pedro Morais, Lukas R Buschle, Estevão Lima, João L Vilaça","doi":"10.1007/s11548-025-03552-5","DOIUrl":"https://doi.org/10.1007/s11548-025-03552-5","url":null,"abstract":"<p><strong>Purpose: </strong>Minimally invasive surgical approaches are currently the standard of care for men with prostate cancer, presenting higher rates of erectile function preservation. With these laparoscopic techniques, there is an increasing amount of data and information available. Adaptive systems can play an important role, acting as an intelligent information filter, assuring that all the available information can become useful for the procedure and not overwhelming for the surgeon. Standardizing and structuring the surgical workflow are key requirements for such smart assistants to recognize the different surgical steps through context information about the environment. This work aims to do a detailed characterization of a laparoscopic radical prostatectomy procedure, focusing on the formalization of medical expert knowledge, via surgical process modeling.</p><p><strong>Methods: </strong>Data were acquired manually, via online and offline observation, and discussion with medical experts. A total of 14 procedures were observed. Both manual laparoscopic radical prostatectomy and robot-assisted laparoscopic prostatectomy were studied. The derived SPM focuses only on the intraoperatory part of the procedure, with constant feedback from the endoscopic camera. For surgery observation, a dedicated Excel template was developed.</p><p><strong>Results: </strong>The final model is represented in a descriptive and numerical format, combining task description with a workflow diagram arrangement for ease of interpretation. Practical applications of the generated surgical process model are exemplified with the creation of activation trees for surgical phase identification. Anatomical structures are reported for each phase, distinguishing between visible and inferable ones. Additionally, the surgeons involved are identified, surgical instruments, and actions performed in each phase. A total of 11 phases were identified and characterized. Average surgery duration is 87 min.</p><p><strong>Conclusion: </strong>The generated surgical process model is a first step toward the development of a context-aware surgical assistant and can potentially be used as a roadmap by other research teams, operating room managers and surgical teams.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2025-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145670873","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Weakly supervised pre-training for surgical step recognition using unannotated and heterogeneously labeled videos. 使用无注释和异构标记视频进行手术步骤识别的弱监督预训练。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-02 DOI: 10.1007/s11548-025-03555-2
Sreeram Kamabattula, Kai Chen, Kiran Bhattacharyya

Purpose: Surgical video review is essential for minimally invasive surgical training, but manual annotation of surgical steps is time-consuming and limits scalability. We propose a weakly supervised pre-training framework that leverages unannotated or heterogeneously labeled surgical videos to improve automated surgical step recognition.

Methods: We evaluate three types of weak labels derived from unannotated datasets: (1) surgical phases from the same or other procedures, (2) surgical steps from different procedure types, and (3) intraoperative time progression. Using datasets from four robotic-assisted procedures (sleeve gastrectomy, hysterectomy, cholecystectomy, and radical prostatectomy), we simulate real-world annotation scarcity by varying the proportion of available step annotations ( α 0.25, 0.5, 0.75, 1.0). We benchmark the performance of a 2D CNN model trained with and without weak label pre-training.

Results: Pre-training with surgical phase labels-particularly from the same procedure type (PHASE-WITHIN)-consistently improved step recognition performance, with gains up to 6.4 f1-score points over standard ImageNet-based models under limited annotation conditions ( α = 0.25 on SLG). Cross-procedure step pre-training was beneficial for some procedures, and time-based labels provided moderate gains depending on procedure structure. Label efficiency analysis shows the baseline model would require labeling an additional 30-60 videos at α = 0.25 to match the performance achieved by the best weak-pretraining strategy across procedures.

Conclusion: Weakly supervised pre-training offers a practical strategy to improve surgical step recognition when annotated data is scarce. This approach can support scalable feedback and assessment in surgical training workflows where comprehensive annotations are infeasible.

目的:手术视频回顾是微创手术培训必不可少的,但手工标注手术步骤耗时且限制可扩展性。我们提出了一个弱监督的预训练框架,利用无注释或异构标记的手术视频来提高自动手术步骤识别。方法:我们评估了来自未注释数据集的三种类型的弱标签:(1)来自相同或其他手术的手术阶段,(2)来自不同手术类型的手术步骤,(3)术中时间进展。使用四种机器人辅助手术(袖式胃切除术、子宫切除术、胆囊切除术和根治性前列腺切除术)的数据集,我们通过改变可用步骤注释的比例(α∈0.25,0.5,0.75,1.0)来模拟现实世界的注释稀缺性。我们对使用和不使用弱标签预训练的二维CNN模型的性能进行了基准测试。结果:使用手术阶段标签的预训练-特别是来自同一手术类型(phase - within)-持续提高了步骤识别性能,在有限的注释条件下(在SLG上α = 0.25),与基于标准imagenet的模型相比,增益高达6.4 f1分。跨程序步骤预训练对某些程序是有益的,基于时间的标签根据程序结构提供适度的增益。标签效率分析表明,基线模型将需要在α = 0.25下标记额外的30-60个视频,以匹配最佳弱预训练策略所取得的性能。结论:弱监督预训练在注释数据稀缺的情况下,为提高手术步骤识别提供了一种实用的策略。这种方法可以在外科训练工作流程中支持可扩展的反馈和评估,其中全面的注释是不可行的。
{"title":"Weakly supervised pre-training for surgical step recognition using unannotated and heterogeneously labeled videos.","authors":"Sreeram Kamabattula, Kai Chen, Kiran Bhattacharyya","doi":"10.1007/s11548-025-03555-2","DOIUrl":"https://doi.org/10.1007/s11548-025-03555-2","url":null,"abstract":"<p><strong>Purpose: </strong>Surgical video review is essential for minimally invasive surgical training, but manual annotation of surgical steps is time-consuming and limits scalability. We propose a weakly supervised pre-training framework that leverages unannotated or heterogeneously labeled surgical videos to improve automated surgical step recognition.</p><p><strong>Methods: </strong>We evaluate three types of weak labels derived from unannotated datasets: (1) surgical phases from the same or other procedures, (2) surgical steps from different procedure types, and (3) intraoperative time progression. Using datasets from four robotic-assisted procedures (sleeve gastrectomy, hysterectomy, cholecystectomy, and radical prostatectomy), we simulate real-world annotation scarcity by varying the proportion of available step annotations ( <math><mi>α</mi></math> <math><mo>∈</mo></math> 0.25, 0.5, 0.75, 1.0). We benchmark the performance of a 2D CNN model trained with and without weak label pre-training.</p><p><strong>Results: </strong>Pre-training with surgical phase labels-particularly from the same procedure type (PHASE-WITHIN)-consistently improved step recognition performance, with gains up to 6.4 f1-score points over standard ImageNet-based models under limited annotation conditions ( <math><mi>α</mi></math> = 0.25 on SLG). Cross-procedure step pre-training was beneficial for some procedures, and time-based labels provided moderate gains depending on procedure structure. Label efficiency analysis shows the baseline model would require labeling an additional 30-60 videos at <math><mi>α</mi></math> = 0.25 to match the performance achieved by the best weak-pretraining strategy across procedures.</p><p><strong>Conclusion: </strong>Weakly supervised pre-training offers a practical strategy to improve surgical step recognition when annotated data is scarce. This approach can support scalable feedback and assessment in surgical training workflows where comprehensive annotations are infeasible.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":""},"PeriodicalIF":2.3,"publicationDate":"2025-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145656095","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Point cloud registration algorithm using liver vascular skeleton feature with computed tomography and ultrasonography image fusion. 基于肝脏血管骨架特征的点云配准算法与计算机断层和超声图像融合。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-08-23 DOI: 10.1007/s11548-025-03496-w
Satoshi Miura, Masayuki Nakayama, Kexin Xu, Zhang Bo, Ryoko Kuromatsu, Masahito Nakano, Yu Noda, Takumi Kawaguchi

Purpose: Radiofrequency ablation for liver cancer has advanced rapidly. For accurate ultrasound-guided soft-tissue puncture surgery, it is necessary to fuse intraoperative ultrasound images with preoperative computed tomography images. However, the conventional method is difficult to estimate and fuse images accurately. To address this issue, the present study proposes an algorithm for registering cross-source point clouds based on not surface but the geometric features of the vascular point cloud.

Methods: We developed a fusion system that performs cross-source point cloud registration between ultrasound and computed tomography images, extracting the node, skeleton, and geomatic feature of the vascular point cloud. The system completes the fusion process in an average of 14.5 s after acquiring the vascular point clouds via ultrasound.

Results: The experiments were conducted to fuse liver images by the dummy model and the healthy participants, respectively. The results show the proposed method achieved a registration error within 1.4 mm and decreased the target registration error significantly compared to other methods in a liver dummy model registration experiment. Furthermore, the proposed method achieved the averaged RMSE within 2.23 mm in a human liver vascular skeleton.

Conclusion: The study concluded that because the registration method using vascular feature point cloud could realize the rapid and accurate fusion between ultrasound and computed tomography images, the method is useful to apply the real puncture surgery for radiofrequency ablation for liver. In future work, we will evaluate the proposed method by the patients.

目的:射频消融术治疗肝癌进展迅速。超声引导下进行准确的软组织穿刺手术,需要术中超声图像与术前计算机断层图像融合。然而,传统的方法难以准确地估计和融合图像。为了解决这一问题,本研究提出了一种基于血管点云几何特征而非表面特征的交叉源点云配准算法。方法:我们开发了一个融合系统,在超声和计算机断层图像之间进行交叉源点云配准,提取血管点云的节点、骨架和地理特征。该系统在超声获取血管点云后,平均14.5 s完成融合过程。结果:分别对虚拟模型和健康被试进行肝脏图像融合实验。结果表明,在肝脏假人模型配准实验中,与其他方法相比,该方法的配准误差在1.4 mm以内,显著降低了目标配准误差。此外,该方法在人肝脏血管骨架上实现了2.23 mm以内的平均RMSE。结论:基于血管特征点云的配准方法可以实现超声与ct图像的快速准确融合,可用于肝脏射频消融的实际穿刺手术。在未来的工作中,我们将由患者对所提出的方法进行评估。
{"title":"Point cloud registration algorithm using liver vascular skeleton feature with computed tomography and ultrasonography image fusion.","authors":"Satoshi Miura, Masayuki Nakayama, Kexin Xu, Zhang Bo, Ryoko Kuromatsu, Masahito Nakano, Yu Noda, Takumi Kawaguchi","doi":"10.1007/s11548-025-03496-w","DOIUrl":"10.1007/s11548-025-03496-w","url":null,"abstract":"<p><strong>Purpose: </strong>Radiofrequency ablation for liver cancer has advanced rapidly. For accurate ultrasound-guided soft-tissue puncture surgery, it is necessary to fuse intraoperative ultrasound images with preoperative computed tomography images. However, the conventional method is difficult to estimate and fuse images accurately. To address this issue, the present study proposes an algorithm for registering cross-source point clouds based on not surface but the geometric features of the vascular point cloud.</p><p><strong>Methods: </strong>We developed a fusion system that performs cross-source point cloud registration between ultrasound and computed tomography images, extracting the node, skeleton, and geomatic feature of the vascular point cloud. The system completes the fusion process in an average of 14.5 s after acquiring the vascular point clouds via ultrasound.</p><p><strong>Results: </strong>The experiments were conducted to fuse liver images by the dummy model and the healthy participants, respectively. The results show the proposed method achieved a registration error within 1.4 mm and decreased the target registration error significantly compared to other methods in a liver dummy model registration experiment. Furthermore, the proposed method achieved the averaged RMSE within 2.23 mm in a human liver vascular skeleton.</p><p><strong>Conclusion: </strong>The study concluded that because the registration method using vascular feature point cloud could realize the rapid and accurate fusion between ultrasound and computed tomography images, the method is useful to apply the real puncture surgery for radiofrequency ablation for liver. In future work, we will evaluate the proposed method by the patients.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":"2469-2478"},"PeriodicalIF":2.3,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12689734/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144977838","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Synthetic X-Q space learning for diffusion MRI parameter estimation: a pilot study in breast DKI. 用于扩散MRI参数估计的合成X-Q空间学习:乳腺DKI的初步研究。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-11-24 DOI: 10.1007/s11548-025-03550-7
Yoshitaka Masutani, Kousei Konya, Erina Kato, Naoko Mori, Hideki Ota, Shunji Mugikura, Kei Takase, Yuki Ichinoseki

Purpose: For diffusion MRI (dMRI) parameter estimation, machine-learning approaches have shown promising results so far including the synthetic Q-space learning (synQSL) based on regressor training with only synthetic data. In this study, we aimed at the development of a new method named synthetic X-Q space learning (synXQSL) to improve robustness and investigated the basic characteristics.

Methods: For training data, local parameter patterns of 3 × 3 voxels were synthesized by a linear combination of six bases, in which parameters are estimated at the center voxel. We prepared three types of local patterns by choosing the number of bases: flat, linear and quadratic. Then, at each location of 3 × 3 voxels, signal values of the diffusion-weighted image were computed by the signal model equation for diffusional kurtosis imaging and Rician noise simulation. The multi-layer perceptron was used for parameter estimation and was trained for each parameter with various noise levels. The level is controlled by a noise ratio defined as a fraction of the standard deviation in the Rician noise distribution normalized by the average b = 0 signal values. Experiments for visual and quantitative validation were performed with synthetic data, a digital phantom and clinical breast datasets in comparison with the previous methods.

Results: By using synthetic datasets, synXQSL outperformed synQSL in the parameter estimation of noisy data sets. Through the digital phantom experiments, the combination of synXQSL bases yields different results and a quadratic pattern could be the reasonable choice. The clinical data experiments indicate that synXQSL suppresses noises in estimated parameter maps and consequently brings higher contrast.

Conclusion: The basic characteristics of synXQSL were investigated by using various types of datasets. The results indicate that synXQSL with the appropriate choice of bases in training data synthesis has the potential to improve dMRI parameters in noisy datasets.

目的:对于扩散MRI (dMRI)参数估计,机器学习方法迄今为止已经显示出有希望的结果,包括基于仅使用合成数据的回归训练的合成q空间学习(synQSL)。在本研究中,我们旨在开发一种名为合成X-Q空间学习(synXQSL)的新方法来提高鲁棒性,并研究了其基本特征。方法:对于训练数据,通过6个碱基的线性组合合成3 × 3体素的局部参数模式,并在中心体素处估计参数。我们通过选择基底的数量制备了三种类型的局部图案:平面、线性和二次型。然后,在3 × 3体素的每个位置,利用扩散峰度成像信号模型方程计算扩散加权图像的信号值,并进行噪声模拟。多层感知器用于参数估计,并在不同噪声水平下对每个参数进行训练。电平由噪声比控制,噪声比定义为由平均b = 0信号值归一化的噪声分布中标准差的一部分。与之前的方法相比,使用合成数据、数字假体和临床乳房数据集进行视觉和定量验证实验。结果:在使用合成数据集时,synXQSL在噪声数据集参数估计方面优于synQSL。通过数字幻影实验,synXQSL碱基组合产生了不同的结果,二次型模式可能是合理的选择。临床数据实验表明,synXQSL可以抑制估计参数图中的噪声,从而提高对比度。结论:利用不同类型的数据集研究了synXQSL的基本特征。结果表明,在训练数据合成中适当选择碱基的synXQSL具有改善噪声数据集dMRI参数的潜力。
{"title":"Synthetic X-Q space learning for diffusion MRI parameter estimation: a pilot study in breast DKI.","authors":"Yoshitaka Masutani, Kousei Konya, Erina Kato, Naoko Mori, Hideki Ota, Shunji Mugikura, Kei Takase, Yuki Ichinoseki","doi":"10.1007/s11548-025-03550-7","DOIUrl":"10.1007/s11548-025-03550-7","url":null,"abstract":"<p><strong>Purpose: </strong>For diffusion MRI (dMRI) parameter estimation, machine-learning approaches have shown promising results so far including the synthetic Q-space learning (synQSL) based on regressor training with only synthetic data. In this study, we aimed at the development of a new method named synthetic X-Q space learning (synXQSL) to improve robustness and investigated the basic characteristics.</p><p><strong>Methods: </strong>For training data, local parameter patterns of 3 × 3 voxels were synthesized by a linear combination of six bases, in which parameters are estimated at the center voxel. We prepared three types of local patterns by choosing the number of bases: flat, linear and quadratic. Then, at each location of 3 × 3 voxels, signal values of the diffusion-weighted image were computed by the signal model equation for diffusional kurtosis imaging and Rician noise simulation. The multi-layer perceptron was used for parameter estimation and was trained for each parameter with various noise levels. The level is controlled by a noise ratio defined as a fraction of the standard deviation in the Rician noise distribution normalized by the average b = 0 signal values. Experiments for visual and quantitative validation were performed with synthetic data, a digital phantom and clinical breast datasets in comparison with the previous methods.</p><p><strong>Results: </strong>By using synthetic datasets, synXQSL outperformed synQSL in the parameter estimation of noisy data sets. Through the digital phantom experiments, the combination of synXQSL bases yields different results and a quadratic pattern could be the reasonable choice. The clinical data experiments indicate that synXQSL suppresses noises in estimated parameter maps and consequently brings higher contrast.</p><p><strong>Conclusion: </strong>The basic characteristics of synXQSL were investigated by using various types of datasets. The results indicate that synXQSL with the appropriate choice of bases in training data synthesis has the potential to improve dMRI parameters in noisy datasets.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":"2423-2435"},"PeriodicalIF":2.3,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12689713/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145589737","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MHAHF-UNet: a multi-scale hybrid attention hierarchy fusion network for carotid artery segmentation. MHAHF-UNet:用于颈动脉分割的多尺度混合注意层次融合网络。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-06-17 DOI: 10.1007/s11548-025-03449-3
Changshuo Jiang, Lin Gao, Wei Li, Maoyang Zou, Qingxiao Zheng, Xuhua Qiao

Purpose: Carotid plaque is an early manifestation of carotid atherosclerosis, and its accurate segmentation helps to assess cardiovascular disease risk. However, existing carotid artery segmentation algorithms are difficult to accurately capture the structural features of morphologically diverse plaques and lack effective utilization of multilayer features.

Methods: In order to solve the above problems, this paper proposes a multi-scale hybrid attention hierarchical fusion U-network structure (MHAHF-UNet) for segmenting ambiguous plaques in carotid artery images in order to improve the segmentation accuracy for complex structured images. The structure firstly introduces the median-enhanced orthogonal convolution module (MEOConv), which not only effectively suppresses the noise interference in ultrasound images, but also maintains the ability to perceive multi-scale features by combining the median-enhanced ternary channel mechanism and the depth-orthogonal convolution space mechanism. Secondly, it adopts the multi-fusion group convolutional gating module, which realizes the effective integration of shallow detailed features and deep semantic features through the adaptive control strategy of group convolution, and is able to flexibly regulate the transfer weights of features at different levels.

Results: Experiments show that the MHAHF-UNet model achieves a Dice coefficient of 82.46 ± 0.31 % and an IOU of 71.45 ± 0.37 % in the carotid artery segmentation task.

Conclusion: The model is expected to provide strong support for the prevention and treatment of cardiovascular diseases.

目的:颈动脉斑块是颈动脉粥样硬化的早期表现,其准确分割有助于评估心血管疾病的风险。然而,现有的颈动脉分割算法难以准确捕捉形态多样斑块的结构特征,缺乏对多层特征的有效利用。方法:针对上述问题,本文提出了一种多尺度混合关注层次融合u -网络结构(MHAHF-UNet)用于分割颈动脉图像中的模糊斑块,以提高对复杂结构图像的分割精度。该结构首先引入了中值增强正交卷积模块(MEOConv),该模块结合中值增强三元通道机制和深度正交卷积空间机制,不仅有效抑制超声图像中的噪声干扰,而且保持了对多尺度特征的感知能力。其次,采用多融合群卷积门控模块,通过群卷积自适应控制策略实现浅层细节特征与深层语义特征的有效融合,并能灵活调节不同层次特征的传递权值。结果:实验表明,MHAHF-UNet模型在颈动脉分割任务中的Dice系数为82.46±0.31%,IOU为71.45±0.37%。结论:该模型有望为心血管疾病的防治提供有力支持。
{"title":"MHAHF-UNet: a multi-scale hybrid attention hierarchy fusion network for carotid artery segmentation.","authors":"Changshuo Jiang, Lin Gao, Wei Li, Maoyang Zou, Qingxiao Zheng, Xuhua Qiao","doi":"10.1007/s11548-025-03449-3","DOIUrl":"10.1007/s11548-025-03449-3","url":null,"abstract":"<p><strong>Purpose: </strong>Carotid plaque is an early manifestation of carotid atherosclerosis, and its accurate segmentation helps to assess cardiovascular disease risk. However, existing carotid artery segmentation algorithms are difficult to accurately capture the structural features of morphologically diverse plaques and lack effective utilization of multilayer features.</p><p><strong>Methods: </strong>In order to solve the above problems, this paper proposes a multi-scale hybrid attention hierarchical fusion U-network structure (MHAHF-UNet) for segmenting ambiguous plaques in carotid artery images in order to improve the segmentation accuracy for complex structured images. The structure firstly introduces the median-enhanced orthogonal convolution module (MEOConv), which not only effectively suppresses the noise interference in ultrasound images, but also maintains the ability to perceive multi-scale features by combining the median-enhanced ternary channel mechanism and the depth-orthogonal convolution space mechanism. Secondly, it adopts the multi-fusion group convolutional gating module, which realizes the effective integration of shallow detailed features and deep semantic features through the adaptive control strategy of group convolution, and is able to flexibly regulate the transfer weights of features at different levels.</p><p><strong>Results: </strong>Experiments show that the MHAHF-UNet model achieves a Dice coefficient of <math><mrow><mn>82.46</mn> <mo>±</mo> <mn>0.31</mn> <mo>%</mo></mrow> </math> and an IOU of <math><mrow><mn>71.45</mn> <mo>±</mo> <mn>0.37</mn> <mo>%</mo></mrow> </math> in the carotid artery segmentation task.</p><p><strong>Conclusion: </strong>The model is expected to provide strong support for the prevention and treatment of cardiovascular diseases.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":"2541-2551"},"PeriodicalIF":2.3,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144318640","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Augmented reality in pelvic surgery: using an AR-headset as intraoperative radiation-free navigation tool. 增强现实在骨盆手术中的应用:使用ar头显作为术中无辐射导航工具。
IF 2.3 3区 医学 Q3 ENGINEERING, BIOMEDICAL Pub Date : 2025-12-01 Epub Date: 2025-06-26 DOI: 10.1007/s11548-025-03462-6
Vincent K Schenk, Markus A Küper, Maximilian M Menger, Steven C Herath, Tina Histing, Christof K Audretsch

Purpose: The incidence of acetabular and pelvic fractures is rising significantly. Pelvic ring fractures rank as the sixth most common fractures in adults, with the majority occurring in the elderly. Due to complications related to surgical approaches, with rates of up to 31%, there is an increasing demand for minimally invasive surgical techniques. Augmented Reality (AR) has the potential to facilitate spatial orientation by a sophisticated user interface. The aim of this study was to develop an AR-based, radiation-free navigation system for pelvic fractures.

Methods: The Microsoft® HoloLens 2 was used as the AR headset. The Unity® game engine was used for programming. Pelvic models from Sawbones® served as the model. Segmentation was performed using Slicer3D by Slicer Corporation. The symphysis and both anterior superior iliac spines were defined as anatomical reference points. Ten pelvic models were used for testing. A preoperatively defined drill trajectory was displayed to the surgeon. A total of 20 S1 screws and 19 S2 screws were placed using only AR navigation without visual access to the pelvic model. Screw placement was controlled using CT.

Results: The matching process took an average of 3 min and 28 s. 18 out of 20 (90%) S1 screws and 3 out of 20 (15%) S2 screws were placed correctly. In most cases, no perforation occurred. The mean procedure time was 7 min for S1 screws and 5 min for S2 screws.

Conclusion: Proper drilling was achieved by displaying the trajectories via AR, particularly for S1 screws, where a slightly wider drilling corridor was aimed for compared to S2 screws. No registration scan was necessary with our matching method. No intraoperative radiation was required.

目的:髋臼和骨盆骨折的发生率明显上升。骨盆环骨折是成人第六大常见骨折,大多数发生在老年人中。由于与手术入路相关的并发症发生率高达31%,对微创手术技术的需求日益增加。增强现实(AR)具有通过复杂的用户界面促进空间定向的潜力。本研究的目的是开发一种基于ar的无辐射骨盆骨折导航系统。方法:采用Microsoft®HoloLens 2作为AR头显。使用Unity®游戏引擎进行编程。来自Sawbones®的骨盆模型作为模型。使用Slicer Corporation的Slicer3D进行分割。联合和髂前上棘被定义为解剖参考点。采用10个盆腔模型进行试验。将术前定义的钻孔轨迹显示给外科医生。共放置20枚S1螺钉和19枚S2螺钉,仅使用AR导航,不使用视觉进入骨盆模型。CT控制螺钉置入。结果:匹配过程平均耗时3 min 28 s, 20枚S1螺钉中有18枚(90%)正确放置,20枚S2螺钉中有3枚(15%)正确放置。在大多数情况下,没有发生穿孔。S1螺钉的平均手术时间为7分钟,S2螺钉为5分钟。结论:通过AR显示轨迹可以实现适当的钻孔,特别是S1螺钉,与S2螺钉相比,S1螺钉的钻孔通道略宽。我们的匹配方法不需要注册扫描。术中不需要放射治疗。
{"title":"Augmented reality in pelvic surgery: using an AR-headset as intraoperative radiation-free navigation tool.","authors":"Vincent K Schenk, Markus A Küper, Maximilian M Menger, Steven C Herath, Tina Histing, Christof K Audretsch","doi":"10.1007/s11548-025-03462-6","DOIUrl":"10.1007/s11548-025-03462-6","url":null,"abstract":"<p><strong>Purpose: </strong>The incidence of acetabular and pelvic fractures is rising significantly. Pelvic ring fractures rank as the sixth most common fractures in adults, with the majority occurring in the elderly. Due to complications related to surgical approaches, with rates of up to 31%, there is an increasing demand for minimally invasive surgical techniques. Augmented Reality (AR) has the potential to facilitate spatial orientation by a sophisticated user interface. The aim of this study was to develop an AR-based, radiation-free navigation system for pelvic fractures.</p><p><strong>Methods: </strong>The Microsoft® HoloLens 2 was used as the AR headset. The Unity® game engine was used for programming. Pelvic models from Sawbones® served as the model. Segmentation was performed using Slicer3D by Slicer Corporation. The symphysis and both anterior superior iliac spines were defined as anatomical reference points. Ten pelvic models were used for testing. A preoperatively defined drill trajectory was displayed to the surgeon. A total of 20 S1 screws and 19 S2 screws were placed using only AR navigation without visual access to the pelvic model. Screw placement was controlled using CT.</p><p><strong>Results: </strong>The matching process took an average of 3 min and 28 s. 18 out of 20 (90%) S1 screws and 3 out of 20 (15%) S2 screws were placed correctly. In most cases, no perforation occurred. The mean procedure time was 7 min for S1 screws and 5 min for S2 screws.</p><p><strong>Conclusion: </strong>Proper drilling was achieved by displaying the trajectories via AR, particularly for S1 screws, where a slightly wider drilling corridor was aimed for compared to S2 screws. No registration scan was necessary with our matching method. No intraoperative radiation was required.</p>","PeriodicalId":51251,"journal":{"name":"International Journal of Computer Assisted Radiology and Surgery","volume":" ","pages":"2553-2563"},"PeriodicalIF":2.3,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12689752/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144499154","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International Journal of Computer Assisted Radiology and Surgery
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1