首页 > 最新文献

Computer methods and programs in biomedicine update最新文献

英文 中文
Semi-supervised learning and integration of multi-sequence MR-images for carotid vessel wall and plaque segmentation 半监督学习和多序列磁共振图像集成用于颈动脉血管壁和斑块分割
Pub Date : 2026-06-01 Epub Date: 2025-12-31 DOI: 10.1016/j.cmpbup.2025.100230
Marie-Christine Pali , Christina Schwaiger , Malik Galijasevic , Valentin K. Ladenhauf , Stephanie Mangesius , Elke R. Gizewski
The analysis of carotid arteries, particularly plaques, in multi-sequence Magnetic Resonance Imaging (MRI) data is crucial for assessing the risk of atherosclerosis and ischemic stroke. In order to evaluate metrics and radiomic features, quantifying the state of atherosclerosis, accurate segmentation is important. However, the complex morphology of plaques and the scarcity of labeled data poses significant challenges. In this work, we address these problems and propose a semi-supervised deep learning-based approach designed to effectively integrate multi-sequence MRI data for the segmentation of carotid artery vessel wall and plaque. The proposed algorithm consists of two networks: a coarse localization model identifies the region of interest guided by some prior knowledge on the position and number of carotid arteries, followed by a fine segmentation model for precise delineation of vessel walls and plaques. To effectively integrate complementary information across different MRI sequences, we investigate different fusion strategies and introduce a multi-level multi-sequence version of U-Net architecture. To address the challenges of limited labeled data and the complexity of carotid artery MRI, we propose a semi-supervised approach that enforces consistency under various input transformations. Our approach is evaluated on 52 patients with arteriosclerosis, each with five MRI sequences. Comprehensive experiments demonstrate the effectiveness of our approach and emphasize the role of fusion point selection in U-Net-based architectures. To validate the accuracy of our results, we also include an expert-based assessment of model performance. Our findings highlight the potential of fusion strategies and semi-supervised learning for improving carotid artery segmentation in data-limited MRI applications.
在多序列磁共振成像(MRI)数据中分析颈动脉,特别是斑块,对于评估动脉粥样硬化和缺血性中风的风险至关重要。为了评估指标和放射学特征,量化动脉粥样硬化的状态,准确的分割是很重要的。然而,斑块的复杂形态和标记数据的稀缺性构成了重大挑战。在这项工作中,我们解决了这些问题,并提出了一种基于半监督深度学习的方法,旨在有效地整合多序列MRI数据,用于分割颈动脉血管壁和斑块。该算法由两个网络组成:一个粗定位模型通过对颈动脉位置和数量的先验知识来识别感兴趣的区域,然后是一个精细分割模型来精确描绘血管壁和斑块。为了有效地整合不同MRI序列的互补信息,我们研究了不同的融合策略,并引入了多级多序列版本的U-Net架构。为了解决有限的标记数据和颈动脉MRI的复杂性的挑战,我们提出了一种半监督的方法,在各种输入变换下强制一致性。我们的方法在52例动脉硬化患者中进行了评估,每个患者有5个MRI序列。综合实验证明了该方法的有效性,并强调了融合点选择在基于u - net的体系结构中的作用。为了验证结果的准确性,我们还包括基于专家的模型性能评估。我们的研究结果强调了融合策略和半监督学习在数据有限的MRI应用中改善颈动脉分割的潜力。
{"title":"Semi-supervised learning and integration of multi-sequence MR-images for carotid vessel wall and plaque segmentation","authors":"Marie-Christine Pali ,&nbsp;Christina Schwaiger ,&nbsp;Malik Galijasevic ,&nbsp;Valentin K. Ladenhauf ,&nbsp;Stephanie Mangesius ,&nbsp;Elke R. Gizewski","doi":"10.1016/j.cmpbup.2025.100230","DOIUrl":"10.1016/j.cmpbup.2025.100230","url":null,"abstract":"<div><div>The analysis of carotid arteries, particularly plaques, in multi-sequence Magnetic Resonance Imaging (MRI) data is crucial for assessing the risk of atherosclerosis and ischemic stroke. In order to evaluate metrics and radiomic features, quantifying the state of atherosclerosis, accurate segmentation is important. However, the complex morphology of plaques and the scarcity of labeled data poses significant challenges. In this work, we address these problems and propose a semi-supervised deep learning-based approach designed to effectively integrate multi-sequence MRI data for the segmentation of carotid artery vessel wall and plaque. The proposed algorithm consists of two networks: a coarse localization model identifies the region of interest guided by some prior knowledge on the position and number of carotid arteries, followed by a fine segmentation model for precise delineation of vessel walls and plaques. To effectively integrate complementary information across different MRI sequences, we investigate different fusion strategies and introduce a multi-level multi-sequence version of U-Net architecture. To address the challenges of limited labeled data and the complexity of carotid artery MRI, we propose a semi-supervised approach that enforces consistency under various input transformations. Our approach is evaluated on 52 patients with arteriosclerosis, each with five MRI sequences. Comprehensive experiments demonstrate the effectiveness of our approach and emphasize the role of fusion point selection in U-Net-based architectures. To validate the accuracy of our results, we also include an expert-based assessment of model performance. Our findings highlight the potential of fusion strategies and semi-supervised learning for improving carotid artery segmentation in data-limited MRI applications.</div></div>","PeriodicalId":72670,"journal":{"name":"Computer methods and programs in biomedicine update","volume":"9 ","pages":"Article 100230"},"PeriodicalIF":0.0,"publicationDate":"2026-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145940016","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Reassessment of pelvic radiographic measurements for delivery prediction using machine learning 利用机器学习对骨盆x线测量进行产程预测的重新评估
Pub Date : 2026-06-01 Epub Date: 2026-01-03 DOI: 10.1016/j.cmpbup.2026.100231
Ayano Suemori , Jota Maki , Hikaru Ooba , Hikari Nakato , Keiichi Oishi , Tomohiro Mitoma , Sakurako Mishima , Akiko Ohira , Satoe Kirino , Eriko Eto , Hisashi Masuyama

Background and Objective

Pelvimetry has historically shown limitations in diagnosing cephalopelvic disproportion, yet recent evidence suggests potential predictive value. This study uses artificial intelligence to reassess pelvimetry's utility in predicting cesarean section.

Methods

This single-center, retrospective case-control study included pregnant women at 37 weeks 0 days and 41 weeks 6 days of gestation, who underwent pelvic radiography for suspected cephalopelvic disproportion from January 2015 to August 2023. Pelvic radiographic measurements were obtained using the Guthmann-Sussmann method. Maternal characteristics, ultrasound examination data, and pelvimetric measurements were extracted from electronic medical records as potential predictors of delivery outcomes. In this study, the input data were analyzed using four machine learning models: Light Gradient Boosting Machine, Random Forest, Extreme Gradient Boosting, and Category Boosting. The primary outcome was the hierarchical importance of pelvic measurements in the predictive models.

Results

Analysis included 355 participants. The strongest predictors were the differences between (1) the obstetric conjugate and biparietal diameter and (2) the interspinous diameter and biparietal diameter. The receiver operating characteristic curve for each model was Light Gradient Boosting Machine 0.74, Random Forest 0.85, Extreme Gradient Boosting 0.83, and Category Boosting 0.82.

Conclusions

We developed high-performance machine learning models demonstrating that pelvimetric measurements— particularly, the differences between the obstetric conjugate and biparietal diameter, and between the interspinous diameter and biparietal diameter —combined with maternal and ultrasound factors, are strong predictors of cesarean section. The model’s ability to capture nonlinear associations may enhance predictive accuracy, and reassessing pelvimetric values could support delivery planning in clinical settings.
背景与目的骨盆测量在诊断头骨盆比例失调方面一直存在局限性,但最近的证据表明其具有潜在的预测价值。本研究使用人工智能重新评估骨盆测量在预测剖宫产中的效用。方法本研究为单中心、回顾性病例对照研究,纳入2015年1月至2023年8月期间因疑似头骨盆比例失调接受盆腔x线检查的妊娠37周0天和41周6天孕妇。盆腔x线测量采用Guthmann-Sussmann方法。从电子病历中提取产妇特征、超声检查数据和骨盆测量数据,作为分娩结局的潜在预测因素。在本研究中,使用四种机器学习模型对输入数据进行分析:光梯度增强机、随机森林、极端梯度增强和类别增强。主要结果是预测模型中骨盆测量的等级重要性。结果共纳入355名参与者。最强的预测因子是(1)产科共轭物和双顶骨直径和(2)棘间直径和双顶骨直径之间的差异。各模型的接收者工作特征曲线分别为光梯度增强机0.74、随机森林0.85、极端梯度增强0.83、类别增强0.82。结论:我们开发了高性能的机器学习模型,证明骨盆测量-特别是产科结合部和双顶叶直径之间的差异,以及棘间直径和双顶叶直径之间的差异-结合母体和超声因素,是剖宫产的有力预测因素。该模型捕捉非线性关联的能力可以提高预测的准确性,重新评估骨盆测量值可以支持临床环境中的分娩计划。
{"title":"Reassessment of pelvic radiographic measurements for delivery prediction using machine learning","authors":"Ayano Suemori ,&nbsp;Jota Maki ,&nbsp;Hikaru Ooba ,&nbsp;Hikari Nakato ,&nbsp;Keiichi Oishi ,&nbsp;Tomohiro Mitoma ,&nbsp;Sakurako Mishima ,&nbsp;Akiko Ohira ,&nbsp;Satoe Kirino ,&nbsp;Eriko Eto ,&nbsp;Hisashi Masuyama","doi":"10.1016/j.cmpbup.2026.100231","DOIUrl":"10.1016/j.cmpbup.2026.100231","url":null,"abstract":"<div><h3>Background and Objective</h3><div>Pelvimetry has historically shown limitations in diagnosing cephalopelvic disproportion, yet recent evidence suggests potential predictive value. This study uses artificial intelligence to reassess pelvimetry's utility in predicting cesarean section.</div></div><div><h3>Methods</h3><div>This single-center, retrospective case-control study included pregnant women at 37 weeks 0 days and 41 weeks 6 days of gestation, who underwent pelvic radiography for suspected cephalopelvic disproportion from January 2015 to August 2023. Pelvic radiographic measurements were obtained using the Guthmann-Sussmann method. Maternal characteristics, ultrasound examination data, and pelvimetric measurements were extracted from electronic medical records as potential predictors of delivery outcomes. In this study, the input data were analyzed using four machine learning models: Light Gradient Boosting Machine, Random Forest, Extreme Gradient Boosting, and Category Boosting. The primary outcome was the hierarchical importance of pelvic measurements in the predictive models.</div></div><div><h3>Results</h3><div>Analysis included 355 participants. The strongest predictors were the differences between (1) the obstetric conjugate and biparietal diameter and (2) the interspinous diameter and biparietal diameter. The receiver operating characteristic curve for each model was Light Gradient Boosting Machine 0.74, Random Forest 0.85, Extreme Gradient Boosting 0.83, and Category Boosting 0.82.</div></div><div><h3>Conclusions</h3><div>We developed high-performance machine learning models demonstrating that pelvimetric measurements— particularly, the differences between the obstetric conjugate and biparietal diameter, and between the interspinous diameter and biparietal diameter —combined with maternal and ultrasound factors, are strong predictors of cesarean section. The model’s ability to capture nonlinear associations may enhance predictive accuracy, and reassessing pelvimetric values could support delivery planning in clinical settings.</div></div>","PeriodicalId":72670,"journal":{"name":"Computer methods and programs in biomedicine update","volume":"9 ","pages":"Article 100231"},"PeriodicalIF":0.0,"publicationDate":"2026-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146037809","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Predictive analysis of clinical features for HPV status in oropharynx squamous cell carcinoma: A machine learning approach with explainability 口咽鳞癌 HPV 状态的临床特征预测分析:一种具有可解释性的机器学习方法
Pub Date : 2025-01-01 Epub Date: 2024-12-18 DOI: 10.1016/j.cmpbup.2024.100170
Emily Diaz Badilla , Ignasi Cos , Claudio Sampieri , Berta Alegre , Isabel Vilaseca , Simone Balocco , Petia Radeva

Background and Objective:

Oropharynx Squamous Cell Carcinoma (OPSCC) linked to Human Papillomavirus (HPV) exhibits a more favorable prognosis than other squamous cell carcinomas of the upper aerodigestive tract. Finding reliable non-invasive detection methods of this prognostic entity is key to propose appropriate therapeutic decisions. This study aims to provide a comprehensive method based on pre-treatment clinical data for predicting the patient’s HPV status over a large OPSCC patient cohort and employing explainability techniques to interpret the significance and effects of the features.

Materials and Methods:

We employed the RADCURE dataset clinical information to train six Machine Learning algorithms, evaluating them via cross-validation for grid search hyper-parameter tuning and feature selection as well as a final performance measurement on a 20% sample test set. For explainability, SHAP and LIME were used to identify the most relevant relationships and their effect on the predictive model. Furthermore, additional publicly available datasets were scrutinized to compare outcomes and assess the method’s generalization across diverse feature sets and populations.

Results:

The best model yielded an AUC of 0.85, a sensitivity of 0.83, and a specificity of 0.75 over the testing set. The explainability analysis highlighted the remarkable significance of specific clinical attributes, in particular the oropharynx subsite tumor location and the patient’s smoking history. The contribution of each variable to the prediction was substantiated by creating a 95% confidence intervals of model coefficients by means of a 10,000 sample bootstrap and by analyzing top contributors across the best-performing models.

Conclusions:

The combination of specific clinical factors typically collected for OPSCC patients, such as smoking habits and the tumor oropharynx sub-location, along with the ML models hereby presented, can by themselves provide an informed analysis of the HPV status, and of proper use of data science techniques to explain it. Future work should focus on adding other data modalities such as CT scans to enhance performance and to uncover new relations, thus aiding medical practitioners in diagnosing OPSCC more accurately.
背景与目的:与人乳头状瘤病毒(HPV)相关的口咽鳞状细胞癌(OPSCC)表现出比其他上呼吸道鳞状细胞癌更好的预后。寻找可靠的非侵入性检测方法是提出适当治疗决策的关键。本研究旨在提供一种基于治疗前临床数据的综合方法来预测患者的HPV状态,并采用可解释性技术来解释这些特征的意义和影响。材料和方法:我们使用RADCURE数据集临床信息来训练六种机器学习算法,通过网格搜索超参数调整和特征选择的交叉验证来评估它们,并在20%的样本测试集上进行最终性能测量。为了可解释性,我们使用SHAP和LIME来确定最相关的关系及其对预测模型的影响。此外,还仔细检查了其他公开可用的数据集,以比较结果并评估该方法在不同特征集和人群中的泛化性。结果:最佳模型在测试集上的AUC为0.85,灵敏度为0.83,特异性为0.75。可解释性分析强调了特定临床属性的显著意义,特别是口咽部亚位肿瘤的位置和患者的吸烟史。每个变量对预测的贡献是通过创建模型系数的95%置信区间来证实的,方法是通过10,000个样本的自举,并通过分析表现最好的模型中的顶级贡献者。结论:结合OPSCC患者通常收集的特定临床因素,如吸烟习惯和肿瘤口咽亚位,以及本文提出的ML模型,可以单独提供对HPV状态的知情分析,并正确使用数据科学技术来解释它。未来的工作应侧重于增加其他数据模式,如CT扫描,以提高性能和发现新的关系,从而帮助医生更准确地诊断OPSCC。
{"title":"Predictive analysis of clinical features for HPV status in oropharynx squamous cell carcinoma: A machine learning approach with explainability","authors":"Emily Diaz Badilla ,&nbsp;Ignasi Cos ,&nbsp;Claudio Sampieri ,&nbsp;Berta Alegre ,&nbsp;Isabel Vilaseca ,&nbsp;Simone Balocco ,&nbsp;Petia Radeva","doi":"10.1016/j.cmpbup.2024.100170","DOIUrl":"10.1016/j.cmpbup.2024.100170","url":null,"abstract":"<div><h3>Background and Objective:</h3><div>Oropharynx Squamous Cell Carcinoma (OPSCC) linked to Human Papillomavirus (HPV) exhibits a more favorable prognosis than other squamous cell carcinomas of the upper aerodigestive tract. Finding reliable non-invasive detection methods of this prognostic entity is key to propose appropriate therapeutic decisions. This study aims to provide a comprehensive method based on pre-treatment clinical data for predicting the patient’s HPV status over a large OPSCC patient cohort and employing explainability techniques to interpret the significance and effects of the features.</div></div><div><h3>Materials and Methods:</h3><div>We employed the RADCURE dataset clinical information to train six Machine Learning algorithms, evaluating them via cross-validation for grid search hyper-parameter tuning and feature selection as well as a final performance measurement on a 20% sample test set. For explainability, SHAP and LIME were used to identify the most relevant relationships and their effect on the predictive model. Furthermore, additional publicly available datasets were scrutinized to compare outcomes and assess the method’s generalization across diverse feature sets and populations.</div></div><div><h3>Results:</h3><div>The best model yielded an AUC of 0.85, a sensitivity of 0.83, and a specificity of 0.75 over the testing set. The explainability analysis highlighted the remarkable significance of specific clinical attributes, in particular the oropharynx subsite tumor location and the patient’s smoking history. The contribution of each variable to the prediction was substantiated by creating a 95% confidence intervals of model coefficients by means of a 10,000 sample bootstrap and by analyzing top contributors across the best-performing models.</div></div><div><h3>Conclusions:</h3><div>The combination of specific clinical factors typically collected for OPSCC patients, such as smoking habits and the tumor oropharynx sub-location, along with the ML models hereby presented, can by themselves provide an informed analysis of the HPV status, and of proper use of data science techniques to explain it. Future work should focus on adding other data modalities such as CT scans to enhance performance and to uncover new relations, thus aiding medical practitioners in diagnosing OPSCC more accurately.</div></div>","PeriodicalId":72670,"journal":{"name":"Computer methods and programs in biomedicine update","volume":"7 ","pages":"Article 100170"},"PeriodicalIF":0.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143180353","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Independence on the lead of the identification of the ventricular depolarization in the electrocardiogram in wearable devices 独立于可穿戴设备中心电图心室去极化的识别
Pub Date : 2025-01-01 Epub Date: 2025-06-02 DOI: 10.1016/j.cmpbup.2025.100196
Noemi Giordano, Silvia Cannone, Gabriella Balestra, Marco Knaflitz

Goal

The home monitoring of cardiac time intervals reduces hospitalization and mortality of cardiovascular patients. However, a reliable time reference in the electrocardiogram is necessary. Nevertheless, the use of different single leads, typical of wearable devices, impacts the repeatability of the time reference and thus the accuracy of the time-dependent parameters. This work proposes a simple approach to detect the peak and onset of the ventricular depolarization, and demonstrates its lead independence, which makes it suitable for wearable devices even with non-standard leads.

Methods

Our method grounds on an energy-based approach, which we applied on a) a publicly available dataset with standard 12-lead recordings; b) a proof-of-concept dataset including a custom precordial non-standard lead implemented on a wearable device.

Results

Compared against the Pan-Tompkins algorithm, our method reduced the absolute error between each lead and the first standard lead by 26 % to 64 % for the peak, and by 70 % to 82 % for the onset detection. The achieved consistency across leads is compatible with clinical monitoring. The computational time was also reduced by 65 % to 96 %, making the algorithm suitable for use on microcontroller-based wearable devices.

Conclusions

The proposed method enables the identification of a stable reference of the ventricular depolarization regardless of the choice of the lead. The presented results open to the implementation on wearable devices for chronic disease monitoring purposes.
目的心脏时间间隔的家庭监测可降低心血管患者的住院率和死亡率。然而,在心电图中有一个可靠的时间参考是必要的。然而,使用不同的单引线(典型的可穿戴设备)会影响时间参考的可重复性,从而影响时间相关参数的准确性。这项工作提出了一种简单的方法来检测心室去极化的峰值和开始,并证明了它的引线独立性,这使得它适用于可穿戴设备,即使是非标准引线。该方法基于基于能量的方法,我们将其应用于a)具有标准12导联记录的公开可用数据集;B)概念验证数据集,包括在可穿戴设备上实现的自定义pre - cordial非标准lead。结果与Pan-Tompkins算法相比,该方法将每条导联与第一标准导联的峰值绝对误差降低了26% ~ 64%,在起始检测方面降低了70% ~ 82%。导联间的一致性与临床监测一致。计算时间也减少了65%至96%,使该算法适用于基于微控制器的可穿戴设备。结论无论选择何种导联,该方法都能确定稳定的心室去极化参考。提出的结果对用于慢性疾病监测目的的可穿戴设备的实施开放。
{"title":"Independence on the lead of the identification of the ventricular depolarization in the electrocardiogram in wearable devices","authors":"Noemi Giordano,&nbsp;Silvia Cannone,&nbsp;Gabriella Balestra,&nbsp;Marco Knaflitz","doi":"10.1016/j.cmpbup.2025.100196","DOIUrl":"10.1016/j.cmpbup.2025.100196","url":null,"abstract":"<div><h3>Goal</h3><div>The home monitoring of cardiac time intervals reduces hospitalization and mortality of cardiovascular patients. However, a reliable time reference in the electrocardiogram is necessary. Nevertheless, the use of different single leads, typical of wearable devices, impacts the repeatability of the time reference and thus the accuracy of the time-dependent parameters. This work proposes a simple approach to detect the peak and onset of the ventricular depolarization, and demonstrates its lead independence, which makes it suitable for wearable devices even with non-standard leads.</div></div><div><h3>Methods</h3><div>Our method grounds on an energy-based approach, which we applied on a) a publicly available dataset with standard 12-lead recordings; b) a proof-of-concept dataset including a custom precordial non-standard lead implemented on a wearable device.</div></div><div><h3>Results</h3><div>Compared against the Pan-Tompkins algorithm, our method reduced the absolute error between each lead and the first standard lead by 26 % to 64 % for the peak, and by 70 % to 82 % for the onset detection. The achieved consistency across leads is compatible with clinical monitoring. The computational time was also reduced by 65 % to 96 %, making the algorithm suitable for use on microcontroller-based wearable devices.</div></div><div><h3>Conclusions</h3><div>The proposed method enables the identification of a stable reference of the ventricular depolarization regardless of the choice of the lead. The presented results open to the implementation on wearable devices for chronic disease monitoring purposes.</div></div>","PeriodicalId":72670,"journal":{"name":"Computer methods and programs in biomedicine update","volume":"8 ","pages":"Article 100196"},"PeriodicalIF":0.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144230049","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multiscale guided attention network for optic disc segmentation of retinal images 视网膜图像视盘分割的多尺度引导注意网络
Pub Date : 2025-01-01 Epub Date: 2025-01-16 DOI: 10.1016/j.cmpbup.2025.100180
A Z M Ehtesham Chowdhury , Andrew Mehnert , Graham Mann , William H. Morgan , Ferdous Sohel
Optic disc (OD) segmentation from retinal images is crucial for diagnosing, assessing, and tracking the progression of several sight-threatening diseases. This paper presents a deep machine-learning method for semantically segmenting OD from retinal images. The method is named multiscale guided attention network (MSGANet-OD), comprising encoders for extracting multiscale features and decoders for constructing segmentation maps from the extracted features. The decoder also includes a guided attention module that incorporates features related to structural, contextual, and illumination information to segment OD. A custom loss function is proposed to retain the optic disc's geometrical shape (i.e., elliptical) constraint and to alleviate the blood vessels' influence in the overlapping region between the OD and vessels. MSGANet-OD was trained and tested on an in-house clinical color retinal image dataset captured during ophthalmodynamometry as well as on several publicly available color fundus image datasets, e.g., DRISHTI-GS, RIM-ONE-r3, and REFUGE1. Experimental results show that MSGANet-OD achieved superior OD segmentation performance from ophthalmodynamometry images compared to widely used segmentation methods. Our method also achieved competitive results compared to state-of-the-art OD segmentation methods on public datasets. The proposed method can be used in automated systems to quantitatively assess optic nerve head abnormalities (e.g., glaucoma, optic disc neuropathy) and vascular changes in the OD region.
从视网膜图像中分割视盘(OD)对于诊断、评估和跟踪几种威胁视力的疾病的进展至关重要。本文介绍了一种从视网膜图像中语义分割视盘的深度机器学习方法。该方法被命名为多尺度引导注意力网络(MSGANet-OD),包括用于提取多尺度特征的编码器和用于从提取的特征中构建分割图的解码器。解码器还包括一个引导注意力模块,该模块结合与结构、上下文和光照信息相关的特征来分割 OD。为了保留视盘的几何形状(即椭圆形)约束,并减轻血管在视盘和血管重叠区域的影响,提出了一种自定义损失函数。MSGANet-OD 在眼动力测定时捕获的内部临床彩色视网膜图像数据集以及几个公开的彩色眼底图像数据集(如 DRISHTI-GS、RIM-ONE-r3 和 REFUGE1)上进行了训练和测试。实验结果表明,与广泛使用的眼动力计图像分割方法相比,MSGANet-OD 的外径分割性能更为出色。与公共数据集上最先进的外径分割方法相比,我们的方法也取得了具有竞争力的结果。所提出的方法可用于自动系统,定量评估视神经头异常(如青光眼、视盘神经病变)和外径区域的血管变化。
{"title":"Multiscale guided attention network for optic disc segmentation of retinal images","authors":"A Z M Ehtesham Chowdhury ,&nbsp;Andrew Mehnert ,&nbsp;Graham Mann ,&nbsp;William H. Morgan ,&nbsp;Ferdous Sohel","doi":"10.1016/j.cmpbup.2025.100180","DOIUrl":"10.1016/j.cmpbup.2025.100180","url":null,"abstract":"<div><div>Optic disc (OD) segmentation from retinal images is crucial for diagnosing, assessing, and tracking the progression of several sight-threatening diseases. This paper presents a deep machine-learning method for semantically segmenting OD from retinal images. The method is named multiscale guided attention network (MSGANet-OD), comprising encoders for extracting multiscale features and decoders for constructing segmentation maps from the extracted features. The decoder also includes a guided attention module that incorporates features related to structural, contextual, and illumination information to segment OD. A custom loss function is proposed to retain the optic disc's geometrical shape (i.e., elliptical) constraint and to alleviate the blood vessels' influence in the overlapping region between the OD and vessels. MSGANet-OD was trained and tested on an in-house clinical color retinal image dataset captured during ophthalmodynamometry as well as on several publicly available color fundus image datasets, e.g., DRISHTI-GS, RIM-ONE-r3, and REFUGE1. Experimental results show that MSGANet-OD achieved superior OD segmentation performance from ophthalmodynamometry images compared to widely used segmentation methods. Our method also achieved competitive results compared to state-of-the-art OD segmentation methods on public datasets. The proposed method can be used in automated systems to quantitatively assess optic nerve head abnormalities (e.g., glaucoma, optic disc neuropathy) and vascular changes in the OD region.</div></div>","PeriodicalId":72670,"journal":{"name":"Computer methods and programs in biomedicine update","volume":"7 ","pages":"Article 100180"},"PeriodicalIF":0.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143179430","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Robust lung segmentation in Chest X-ray images using modified U-Net with deeper network and residual blocks 基于改进U-Net的胸部x线图像鲁棒肺分割
Pub Date : 2025-01-01 Epub Date: 2025-07-31 DOI: 10.1016/j.cmpbup.2025.100211
Wiley Tam , Paul Babyn , Javad Alirezaie
Lung diseases remain a leading cause of mortality worldwide, as evidenced by statistics from the World Health Organization (WHO). The limited availability of radiologists to interpret Chest X-ray (CXR) images for diagnosing common lung conditions poses a significant challenge, often resulting in delayed diagnosis and treatment. In response, Computer-Aided Diagnostic (CAD) tools can be used to potentially streamline and expedite the diagnostic process. Recently, deep learning techniques have gained prominence in the automated analysis of CXR images, particularly in segmenting lung regions as a critical preliminary step. This study aims to develop and evaluate a lung segmentation model based on a modified U-Net architecture. The architecture leverages techniques such as transfer learning with DenseNet201 as a feature extractor alongside dilated convolutions and residual blocks. An ablation study was conducted to evaluate these architectural components, along with additional elements like augmented data, alternative backbones, and attention mechanisms. Numerous and extensive experiments were performed on two publicly available datasets, the Montgomery County (MC) and Shenzhen Hospital (SH) datasets, to validate the efficacy of these techniques on segmentation performance. Outperforming other state-of-the-art methods on the MC dataset, the proposed model achieved a Jaccard Index (IoU) of 97.77 and a Dice Similarity Coefficient (DSC) of 98.87. These results represent a significant improvement over the baseline U-Net, with gains of 3.37% and 1.75% in IoU and DSC, respectively. These findings highlight the importance of architectural enhancements in deep learning-based lung segmentation models, contributing to more efficient, accurate, and reliable CAD systems for lung disease assessment.
世界卫生组织(世卫组织)的统计数据证明,肺部疾病仍然是全世界死亡的主要原因。放射科医生解释胸部x光片(CXR)图像以诊断常见肺部疾病的有限可用性构成了重大挑战,经常导致诊断和治疗延迟。因此,可以使用计算机辅助诊断(CAD)工具来简化和加快诊断过程。最近,深度学习技术在CXR图像的自动分析中获得了突出的地位,特别是在分割肺区域作为关键的初步步骤方面。本研究旨在开发和评估一个基于改进U-Net架构的肺分割模型。该架构利用了DenseNet201的迁移学习等技术,作为扩展卷积和残差块的特征提取器。我们进行了一项消融研究,以评估这些架构组件,以及其他元素,如增强数据、替代主干和注意机制。在蒙哥马利县(MC)和深圳医院(SH)两个公开可用的数据集上进行了大量广泛的实验,以验证这些技术在分割性能方面的有效性。该模型在MC数据集上的表现优于其他最先进的方法,其Jaccard Index (IoU)为97.77,Dice Similarity Coefficient (DSC)为98.87。这些结果表明,与基线U-Net相比,IoU和DSC分别增加了3.37%和1.75%。这些发现强调了基于深度学习的肺分割模型的架构增强的重要性,有助于更有效、准确和可靠的肺部疾病评估CAD系统。
{"title":"Robust lung segmentation in Chest X-ray images using modified U-Net with deeper network and residual blocks","authors":"Wiley Tam ,&nbsp;Paul Babyn ,&nbsp;Javad Alirezaie","doi":"10.1016/j.cmpbup.2025.100211","DOIUrl":"10.1016/j.cmpbup.2025.100211","url":null,"abstract":"<div><div>Lung diseases remain a leading cause of mortality worldwide, as evidenced by statistics from the World Health Organization (WHO). The limited availability of radiologists to interpret Chest X-ray (CXR) images for diagnosing common lung conditions poses a significant challenge, often resulting in delayed diagnosis and treatment. In response, Computer-Aided Diagnostic (CAD) tools can be used to potentially streamline and expedite the diagnostic process. Recently, deep learning techniques have gained prominence in the automated analysis of CXR images, particularly in segmenting lung regions as a critical preliminary step. This study aims to develop and evaluate a lung segmentation model based on a modified U-Net architecture. The architecture leverages techniques such as transfer learning with DenseNet201 as a feature extractor alongside dilated convolutions and residual blocks. An ablation study was conducted to evaluate these architectural components, along with additional elements like augmented data, alternative backbones, and attention mechanisms. Numerous and extensive experiments were performed on two publicly available datasets, the Montgomery County (MC) and Shenzhen Hospital (SH) datasets, to validate the efficacy of these techniques on segmentation performance. Outperforming other state-of-the-art methods on the MC dataset, the proposed model achieved a Jaccard Index (IoU) of 97.77 and a Dice Similarity Coefficient (DSC) of 98.87. These results represent a significant improvement over the baseline U-Net, with gains of 3.37% and 1.75% in IoU and DSC, respectively. These findings highlight the importance of architectural enhancements in deep learning-based lung segmentation models, contributing to more efficient, accurate, and reliable CAD systems for lung disease assessment.</div></div>","PeriodicalId":72670,"journal":{"name":"Computer methods and programs in biomedicine update","volume":"8 ","pages":"Article 100211"},"PeriodicalIF":0.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144780982","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Acoustic cues for person identification using cough sounds 用咳嗽声作为识别人的声学线索
Pub Date : 2025-01-01 Epub Date: 2025-06-02 DOI: 10.1016/j.cmpbup.2025.100195
Van-Thuan Tran, Ting-Hao You, Wei-Ho Tsai

Objectives

This study presents an improved approach to person identification (PID) using nonverbal vocalizations, focusing specifically on cough sounds as a biometric modality. While recent works have demonstrated the feasibility of cough-based PID (CPID), most report accuracies around 80–90 % and could face limitations in terms of model efficiency, generalization, or robustness. Our objective is to advance CPID performance through compact model design and more effective training strategies.

Methods

We collected a custom dataset from 19 subjects and developed a lightweight yet effective deep learning framework for CPID. The proposed architecture, CoughCueNet, is a convolutional recurrent neural network designed to capture both spatial and temporal patterns in cough sounds. The training process incorporates a hybrid loss function that combines supervised contrastive (SC) learning and cross-entropy (CE) loss to enhance feature discrimination. We systematically evaluated multiple acoustic representations, including MFCCs and spectrograms, to identify the most suitable features. We also applied data augmentation for robustness and investigated cross-modal transferability by testing speech-trained models on cough data.

Results

Our CPID system achieved a mean identification accuracy of 97.18 %. Training the proposed CoughCueNet using a hybrid SC+CE loss function consistently improved model generalization and robustness. It outperformed the same network and larger-capacity networks (i.e., VGG16 and ResNet50) trained with CE loss alone, which achieved accuracies around 90 %. Among the tested features, MFCCs yielded superior identification performance over spectrograms. Experiments with speech-trained models tested on coughs revealed limited cross-vocal transferability, emphasizing the need for cough-specific models.

Conclusion

This work advances the state of cough-based PID by demonstrating that high-accuracy identification is achievable using compact models and hybrid training strategies. It establishes cough sounds as a practical and distinctive biometric modality, with promising applications in security, user authentication, and health monitoring, particularly in environments where speech-based systems are less reliable or infeasible.
目的:本研究提出了一种改进的非语言发声方法来识别人(PID),特别关注咳嗽声音作为一种生物识别模式。虽然最近的研究已经证明了基于咳嗽的PID (CPID)的可行性,但大多数报告的准确率约为80 - 90%,并且在模型效率、泛化或鲁棒性方面可能面临限制。我们的目标是通过紧凑的模型设计和更有效的培训策略来提高CPID的表现。方法我们收集了来自19名受试者的自定义数据集,并开发了一个轻量级但有效的CPID深度学习框架。提出的架构CoughCueNet是一个卷积循环神经网络,旨在捕捉咳嗽声音的空间和时间模式。训练过程结合了一个混合损失函数,该函数结合了监督对比(SC)学习和交叉熵(CE)损失来增强特征识别。我们系统地评估了多种声学表征,包括mfc和频谱图,以确定最合适的特征。我们还应用数据增强来增强鲁棒性,并通过测试咳嗽数据上的语音训练模型来研究跨模态可移植性。结果CPID系统的平均识别准确率为97.18%。使用混合SC+CE损失函数训练所提出的CoughCueNet,不断提高模型的泛化和鲁棒性。它优于单独使用CE损失训练的相同网络和更大容量网络(即VGG16和ResNet50),其准确率达到90%左右。在测试的特征中,MFCCs产生了优于谱图的识别性能。用经过语言训练的模型对咳嗽进行的实验显示,跨声音可移植性有限,这强调了对特定咳嗽模型的需求。本研究通过证明使用紧凑模型和混合训练策略可以实现高精度的识别,从而提高了基于咳嗽的PID的状态。它将咳嗽声音确立为一种实用而独特的生物识别方式,在安全、用户身份验证和健康监测方面具有很好的应用前景,特别是在基于语音的系统不太可靠或不可行的环境中。
{"title":"Acoustic cues for person identification using cough sounds","authors":"Van-Thuan Tran,&nbsp;Ting-Hao You,&nbsp;Wei-Ho Tsai","doi":"10.1016/j.cmpbup.2025.100195","DOIUrl":"10.1016/j.cmpbup.2025.100195","url":null,"abstract":"<div><h3>Objectives</h3><div>This study presents an improved approach to person identification (PID) using nonverbal vocalizations, focusing specifically on cough sounds as a biometric modality. While recent works have demonstrated the feasibility of cough-based PID (CPID), most report accuracies around 80–90 % and could face limitations in terms of model efficiency, generalization, or robustness. Our objective is to advance CPID performance through compact model design and more effective training strategies.</div></div><div><h3>Methods</h3><div>We collected a custom dataset from 19 subjects and developed a lightweight yet effective deep learning framework for CPID. The proposed architecture, CoughCueNet, is a convolutional recurrent neural network designed to capture both spatial and temporal patterns in cough sounds. The training process incorporates a hybrid loss function that combines supervised contrastive (SC) learning and cross-entropy (CE) loss to enhance feature discrimination. We systematically evaluated multiple acoustic representations, including MFCCs and spectrograms, to identify the most suitable features. We also applied data augmentation for robustness and investigated cross-modal transferability by testing speech-trained models on cough data.</div></div><div><h3>Results</h3><div>Our CPID system achieved a mean identification accuracy of 97.18 %. Training the proposed CoughCueNet using a hybrid SC+CE loss function consistently improved model generalization and robustness. It outperformed the same network and larger-capacity networks (i.e., VGG16 and ResNet50) trained with CE loss alone, which achieved accuracies around 90 %. Among the tested features, MFCCs yielded superior identification performance over spectrograms. Experiments with speech-trained models tested on coughs revealed limited cross-vocal transferability, emphasizing the need for cough-specific models.</div></div><div><h3>Conclusion</h3><div>This work advances the state of cough-based PID by demonstrating that high-accuracy identification is achievable using compact models and hybrid training strategies. It establishes cough sounds as a practical and distinctive biometric modality, with promising applications in security, user authentication, and health monitoring, particularly in environments where speech-based systems are less reliable or infeasible.</div></div>","PeriodicalId":72670,"journal":{"name":"Computer methods and programs in biomedicine update","volume":"8 ","pages":"Article 100195"},"PeriodicalIF":0.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144230050","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing stroke prediction models: A mixing of data augmentation and transfer learning for small-scale dataset in machine learning 增强中风预测模型:机器学习中小规模数据集的数据增强和迁移学习的混合
Pub Date : 2025-01-01 Epub Date: 2025-06-21 DOI: 10.1016/j.cmpbup.2025.100198
Imam Tahyudin , Ade Nurhopipah , Ades Tikaningsih , Puji Lestari , Yaya Suryana , Edi Winarko , Eko Winarto , Nazwan Haza , Hidetaka Nambo
Machine learning is a powerful technique for analysing datasets and making data-driven recommendations. However, in general, the performance of machine learning in recognising patterns is proportional to the size of the dataset. On the other hand, in some cases, such as in the medical field, providing an instance of a dataset takes a lot of work and budget. Therefore, additional data acquisition techniques are needed to increase data size and improve model quality.
This study applied Data Augmentation and Transfer Learning to solve small-scale dataset problems in analyzing stroke patient information in The Banyumas Regional General Hospital (RSUD Banyumas). The information is utilized to predict the patient's status when discharged from the hospital. The research compared the prediction accuracy from three solutions: Data Augmentation, Transfer Learning, and the mixing of both methods. The classification models employed in this study were four algorithms: Random Forest, Support Vector Machine, Gradient Boosting, and Extreme Gradient Boosting. We implemented the Synthetic Minority Over-sampling Technique for Nominal and Continuous to generate the artificial dataset. In the Transfer Learning process, we used a benchmark stroke dataset with a different target than ours, so we labelled it based on the nearest neighbours of the original dataset. Applying Data Augmentation in this study is a good decision because it leads to better performance than using only the original dataset. However, implementing the Transfer Learning technique does not give a satisfying result for XGBoost and SVM. Mixing Data Augmentation and Transfer Learning provides the best performance with accuracy and recall, both 0.813, the precision of 0.853497, and the F-1 score of 0.826628 given by the Random Forest model. The research can contribute significantly to developing better classification models so physicians can obtain more accurate information and help treat stroke cases more effectively and efficiently.
机器学习是分析数据集和提出数据驱动建议的强大技术。然而,一般来说,机器学习在识别模式方面的表现与数据集的大小成正比。另一方面,在某些情况下,例如在医疗领域,提供数据集的实例需要大量的工作和预算。因此,需要额外的数据采集技术来增加数据大小和提高模型质量。本研究应用数据增强和迁移学习来解决Banyumas地区总医院(RSUD Banyumas)中风患者信息分析中的小规模数据集问题。这些信息被用来预测病人出院时的状态。该研究比较了三种解决方案的预测精度:数据增强、迁移学习和两种方法的混合。本研究使用的分类模型有四种算法:随机森林、支持向量机、梯度增强和极端梯度增强。我们实现了对标称和连续的合成少数派过采样技术来生成人工数据集。在迁移学习过程中,我们使用了一个与我们的目标不同的基准笔画数据集,因此我们基于原始数据集的最近邻居对其进行标记。在本研究中应用数据增强是一个很好的决定,因为它比只使用原始数据集带来更好的性能。然而,迁移学习技术的实现并没有给XGBoost和SVM带来令人满意的结果。混合数据增强和迁移学习的准确率和召回率均为0.813,精度为0.853497,随机森林模型给出的F-1分数为0.826628。这项研究可以为开发更好的分类模型做出重大贡献,这样医生就可以获得更准确的信息,并帮助更有效地治疗中风病例。
{"title":"Enhancing stroke prediction models: A mixing of data augmentation and transfer learning for small-scale dataset in machine learning","authors":"Imam Tahyudin ,&nbsp;Ade Nurhopipah ,&nbsp;Ades Tikaningsih ,&nbsp;Puji Lestari ,&nbsp;Yaya Suryana ,&nbsp;Edi Winarko ,&nbsp;Eko Winarto ,&nbsp;Nazwan Haza ,&nbsp;Hidetaka Nambo","doi":"10.1016/j.cmpbup.2025.100198","DOIUrl":"10.1016/j.cmpbup.2025.100198","url":null,"abstract":"<div><div>Machine learning is a powerful technique for analysing datasets and making data-driven recommendations. However, in general, the performance of machine learning in recognising patterns is proportional to the size of the dataset. On the other hand, in some cases, such as in the medical field, providing an instance of a dataset takes a lot of work and budget. Therefore, additional data acquisition techniques are needed to increase data size and improve model quality.</div><div>This study applied Data Augmentation and Transfer Learning to solve small-scale dataset problems in analyzing stroke patient information in The Banyumas Regional General Hospital (RSUD Banyumas). The information is utilized to predict the patient's status when discharged from the hospital. The research compared the prediction accuracy from three solutions: Data Augmentation, Transfer Learning, and the mixing of both methods. The classification models employed in this study were four algorithms: Random Forest, Support Vector Machine, Gradient Boosting, and Extreme Gradient Boosting. We implemented the Synthetic Minority Over-sampling Technique for Nominal and Continuous to generate the artificial dataset. In the Transfer Learning process, we used a benchmark stroke dataset with a different target than ours, so we labelled it based on the nearest neighbours of the original dataset. Applying Data Augmentation in this study is a good decision because it leads to better performance than using only the original dataset. However, implementing the Transfer Learning technique does not give a satisfying result for XGBoost and SVM. Mixing Data Augmentation and Transfer Learning provides the best performance with accuracy and recall, both 0.813, the precision of 0.853497, and the F-1 score of 0.826628 given by the Random Forest model. The research can contribute significantly to developing better classification models so physicians can obtain more accurate information and help treat stroke cases more effectively and efficiently.</div></div>","PeriodicalId":72670,"journal":{"name":"Computer methods and programs in biomedicine update","volume":"8 ","pages":"Article 100198"},"PeriodicalIF":0.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144500858","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Resectograms: Planning liver surgery with real-time occlusion-free visualization of virtual resections 切除图:通过虚拟切除的实时无闭塞可视化规划肝脏手术
Pub Date : 2025-01-01 Epub Date: 2025-02-23 DOI: 10.1016/j.cmpbup.2025.100186
Ruoyan Meng , Davit Aghayan , Egidijus Pelanis , Bjørn Edwin , Faouzi Alaya Cheikh , Rafael Palomar

Background and Objective:

Visualization of virtual resections plays a central role in computer-assisted liver surgery planning. However, the intricate liver anatomical information often results in occlusions and visualization information clutter, which can lead to inaccuracies in virtual resections. To overcome these challenges, we introduce Resectograms, which are planar (2D) representations of virtual resections enabling the visualization of information associated with the surgical plan.

Methods:

Resectograms are computed in real-time and displayed as additional 2D views showing anatomical, functional, and risk-associated information extracted from the 3D virtual resection as this is modified during planning, offering surgeons an occlusion-free visualization of the virtual resection during surgery planning. To further improve functionality, we explored three flattening methods: fixed-shape, Least Squares Conformal Maps, and As-Rigid-As-Possible, to generate these 2D views. Additionally, we optimized GPU memory usage by downsampling texture objects, ensuring errors remain within acceptable limits as defined by surgeons.

Results:

We evaluated Resectograms with experienced surgeons (n = 4, 9-15 years) and assessed 2D flattening methods with computer and biomedical scientists (n = 11) through visual experiments. Surgeons found Resectograms valuable for enhancing surgical planning effectiveness and accuracy. Among flattening methods, Least Squares Conformal Maps and As-Rigid-As-Possible techniques demonstrated similarly low distortion levels, superior to the fixed-shape approach. Our analysis of texture object downsampling revealed effectiveness for liver and tumor segmentations, but less so for vessel segmentations.

Conclusions:

This paper presents Resectograms, a novel method for visualizing liver virtual resection plans in 2D, offering an intuitive, occlusion-free representation computable in real-time. Resectograms incorporate multiple information layers, providing comprehensive data for liver surgery planning. We enhanced the visualization through improved 3D-to-2D orientation mapping and distortion-minimizing parameterization algorithms. This research contributes to advancing liver surgery planning tools by offering a more accessible and informative visualization method. The code repository for this work is available at: https://github.com/ALive-research/Slicer-Liver.
背景与目的:虚拟切除的可视化在计算机辅助肝脏手术计划中起着核心作用。然而,复杂的肝脏解剖信息往往导致闭塞和可视化信息混乱,从而导致虚拟切除的不准确性。为了克服这些挑战,我们引入了切除图,它是虚拟切除的平面(2D)表示,使与手术计划相关的信息可视化。方法:实时计算切除图,并显示为额外的2D视图,显示从3D虚拟切除中提取的解剖,功能和风险相关信息,因为在计划期间对其进行了修改,为外科医生在手术计划期间提供无闭塞的虚拟切除可视化。为了进一步改进功能,我们探索了三种扁平化方法:固定形状、最小二乘共形映射和尽可能刚性映射,以生成这些2D视图。此外,我们通过降低纹理对象的采样来优化GPU内存使用,确保误差保持在外科医生定义的可接受范围内。结果:我们与经验丰富的外科医生(n = 4, 9-15年)评估了Resectograms,并与计算机和生物医学科学家(n = 11)通过视觉实验评估了二维平坦化方法。外科医生发现切除图对提高手术计划的有效性和准确性很有价值。在平坦化方法中,最小二乘共形映射和尽可能刚性技术显示出类似的低失真水平,优于固定形状方法。我们对纹理对象下采样的分析显示了对肝脏和肿瘤分割的有效性,但对血管分割的效果较差。结论:本文提出了一种新的方法,用于在2D中可视化肝脏虚拟切除计划,提供直观,无阻塞的实时计算表示。切除图包含多个信息层,为肝脏手术计划提供全面的数据。我们通过改进的3d到2d方向映射和最小化扭曲参数化算法增强了可视化。本研究通过提供一种更方便和信息丰富的可视化方法,有助于推进肝脏手术计划工具。这项工作的代码存储库可从:https://github.com/ALive-research/Slicer-Liver获得。
{"title":"Resectograms: Planning liver surgery with real-time occlusion-free visualization of virtual resections","authors":"Ruoyan Meng ,&nbsp;Davit Aghayan ,&nbsp;Egidijus Pelanis ,&nbsp;Bjørn Edwin ,&nbsp;Faouzi Alaya Cheikh ,&nbsp;Rafael Palomar","doi":"10.1016/j.cmpbup.2025.100186","DOIUrl":"10.1016/j.cmpbup.2025.100186","url":null,"abstract":"<div><h3>Background and Objective:</h3><div>Visualization of virtual resections plays a central role in computer-assisted liver surgery planning. However, the intricate liver anatomical information often results in occlusions and visualization information clutter, which can lead to inaccuracies in virtual resections. To overcome these challenges, we introduce <em>Resectograms</em>, which are planar (2D) representations of virtual resections enabling the visualization of information associated with the surgical plan.</div></div><div><h3>Methods:</h3><div>Resectograms are computed in real-time and displayed as additional 2D views showing anatomical, functional, and risk-associated information extracted from the 3D virtual resection as this is modified during planning, offering surgeons an occlusion-free visualization of the virtual resection during surgery planning. To further improve functionality, we explored three flattening methods: fixed-shape, Least Squares Conformal Maps, and As-Rigid-As-Possible, to generate these 2D views. Additionally, we optimized GPU memory usage by downsampling texture objects, ensuring errors remain within acceptable limits as defined by surgeons.</div></div><div><h3>Results:</h3><div>We evaluated Resectograms with experienced surgeons (n = 4, 9-15 years) and assessed 2D flattening methods with computer and biomedical scientists (n = 11) through visual experiments. Surgeons found Resectograms valuable for enhancing surgical planning effectiveness and accuracy. Among flattening methods, Least Squares Conformal Maps and As-Rigid-As-Possible techniques demonstrated similarly low distortion levels, superior to the fixed-shape approach. Our analysis of texture object downsampling revealed effectiveness for liver and tumor segmentations, but less so for vessel segmentations.</div></div><div><h3>Conclusions:</h3><div>This paper presents Resectograms, a novel method for visualizing liver virtual resection plans in 2D, offering an intuitive, occlusion-free representation computable in real-time. Resectograms incorporate multiple information layers, providing comprehensive data for liver surgery planning. We enhanced the visualization through improved 3D-to-2D orientation mapping and distortion-minimizing parameterization algorithms. This research contributes to advancing liver surgery planning tools by offering a more accessible and informative visualization method. The code repository for this work is available at: <span><span>https://github.com/ALive-research/Slicer-Liver</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":72670,"journal":{"name":"Computer methods and programs in biomedicine update","volume":"7 ","pages":"Article 100186"},"PeriodicalIF":0.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143518926","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
GLAAM and GLAAI: Pioneering attention models for robust automated cataract detection GLAAM和GLAAI:用于稳健自动白内障检测的开创性注意力模型
Pub Date : 2025-01-01 Epub Date: 2025-02-22 DOI: 10.1016/j.cmpbup.2025.100182
Deepak Kumar , Chaman Verma , Zoltán Illés

Background and Objective:

Early detection of eye diseases, especially cataracts, is essential for preventing vision impairment. Accurate and cost-effective cataract diagnosis often requires advanced methods. This study proposes novel deep learning models that integrate global and local attention mechanisms into MobileNet and InceptionV3 architectures to improve cataract detection from fundus images.

Methods:

Two deep learning models, Global–Local Attention Augmented MobileNet (GLAAM) and Global–Local Attention Augmented InceptionV3 (GLAAI), were developed to enhance the analysis of fundus images. The models incorporate a combined attention mechanism to effectively capture deteriorated regions in retinal images. Data augmentation techniques were employed to prevent overfitting during training and testing on two cataract datasets. Additionally, Grad-CAM visualizations were used to increase interpretability by highlighting key regions influencing predictions.

Results:

The GLAAM model achieved a balanced accuracy of 97.08%, an average precision of 97.11%, and an F1-score of 97.12% on the retinal dataset. Grad-CAM visualizations confirmed the models’ ability to identify crucial cataract-related regions in fundus images.

Conclusion:

This study demonstrates a significant advancement in cataract diagnosis using deep learning, with GLAAM and GLAAI models exhibiting strong diagnostic performance. These models have the potential to enhance diagnostic tools and improve patient care by offering a cost-effective and accurate solution for cataract detection, suitable for integration into clinical settings.
背景和目的:早期发现眼疾,尤其是白内障,对于预防视力损伤至关重要。准确且经济高效的白内障诊断通常需要先进的方法。本研究提出了新颖的深度学习模型,将全局和局部注意力机制整合到 MobileNet 和 InceptionV3 架构中,以改进眼底图像的白内障检测。这些模型结合了联合注意力机制,可有效捕捉视网膜图像中的恶化区域。在两个白内障数据集的训练和测试过程中,采用了数据增强技术来防止过度拟合。结果:在视网膜数据集上,GLAAM 模型的均衡准确率达到了 97.08%,平均精确率达到了 97.11%,F1 分数达到了 97.12%。Grad-CAM 可视化证实了模型识别眼底图像中关键白内障相关区域的能力。这些模型具有增强诊断工具和改善患者护理的潜力,可为白内障检测提供具有成本效益的准确解决方案,适合集成到临床环境中。
{"title":"GLAAM and GLAAI: Pioneering attention models for robust automated cataract detection","authors":"Deepak Kumar ,&nbsp;Chaman Verma ,&nbsp;Zoltán Illés","doi":"10.1016/j.cmpbup.2025.100182","DOIUrl":"10.1016/j.cmpbup.2025.100182","url":null,"abstract":"<div><h3>Background and Objective:</h3><div>Early detection of eye diseases, especially cataracts, is essential for preventing vision impairment. Accurate and cost-effective cataract diagnosis often requires advanced methods. This study proposes novel deep learning models that integrate global and local attention mechanisms into MobileNet and InceptionV3 architectures to improve cataract detection from fundus images.</div></div><div><h3>Methods:</h3><div>Two deep learning models, Global–Local Attention Augmented MobileNet (GLAAM) and Global–Local Attention Augmented InceptionV3 (GLAAI), were developed to enhance the analysis of fundus images. The models incorporate a combined attention mechanism to effectively capture deteriorated regions in retinal images. Data augmentation techniques were employed to prevent overfitting during training and testing on two cataract datasets. Additionally, Grad-CAM visualizations were used to increase interpretability by highlighting key regions influencing predictions.</div></div><div><h3>Results:</h3><div>The GLAAM model achieved a balanced accuracy of 97.08%, an average precision of 97.11%, and an F1-score of 97.12% on the retinal dataset. Grad-CAM visualizations confirmed the models’ ability to identify crucial cataract-related regions in fundus images.</div></div><div><h3>Conclusion:</h3><div>This study demonstrates a significant advancement in cataract diagnosis using deep learning, with GLAAM and GLAAI models exhibiting strong diagnostic performance. These models have the potential to enhance diagnostic tools and improve patient care by offering a cost-effective and accurate solution for cataract detection, suitable for integration into clinical settings.</div></div>","PeriodicalId":72670,"journal":{"name":"Computer methods and programs in biomedicine update","volume":"7 ","pages":"Article 100182"},"PeriodicalIF":0.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143474173","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Computer methods and programs in biomedicine update
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1