首页 > 最新文献

Intelligence-based medicine最新文献

英文 中文
Dual-stage deep learning framework for brain tumor classification and localization using multimodal MRI scans 使用多模态MRI扫描进行脑肿瘤分类和定位的双阶段深度学习框架
Pub Date : 2026-03-01 Epub Date: 2026-02-18 DOI: 10.1016/j.ibmed.2026.100361
Deependra Rastogi , Prashant Johri , Sumit Singh Dhanda , Anand Singh , Suman Avdhesh Yadav , Arfat Ahmad Khan , Seifedine Kadry
<div><h3>Background</h3><div>Brain tumor classification and localization are essential for accurate diagnosis and effective treatment planning. With the increasing use of MRI in clinical workflows, automated tools have become crucial for assisting radiologists in providing fast and reliable analysis. Traditional approaches for tumor classification and segmentation are often time-consuming and subjective, underscoring the need for deep learning-based frameworks that can enhance diagnostic accuracy and efficiency.</div></div><div><h3>Method</h3><div>This study used the TCGA-GBM dataset from The Cancer Imaging Archive (TCIA) to assess a dual-stage deep learning framework combining a classification model with a RESUNET for segmentation. TCGA-GBM, part of the broader TCGA project, focuses on Glioblastoma Multiforme (GBM), a highly aggressive brain cancer. The dataset includes 3929 images, with 2556 non-tumor (class 0) and 1373 tumor (class 1) samples. The framework incorporated three convolutional neural network (CNN) architectures—MobileNet, NASNetMobile, and ResNet101—each enhanced with a Transfer Learning Layer for classification, followed by a RESUNET network for tumor localization. Transfer learning enabled the use of pre-trained weights, improving convergence speed and generalization. MobileNet offered a lightweight, efficient solution; NASNetMobile provided a strong balance between accuracy and computational cost; and ResNet101 delivered deeper feature extraction for higher precision. The RESUNET architecture, combining U-Net and residual learning, accurately segmented tumor regions, enabling effective integration of classification and localization within a unified framework.</div></div><div><h3>Results</h3><div>In the classification stage, the models achieved average accuracies of 0.9600 for MobileNet + Transfer Learning Layer, 0.9700 for NASNetMobile + Transfer Learning Layer, and 0.9500 for ResNet101 + Transfer Learning Layer, demonstrating their effectiveness in categorizing brain tumors. Performance was further evaluated using precision, recall, and F1 scores for both classes. MobileNet + Transfer Learning Layer achieved 0.95 precision, 0.99 recall, and 0.97 F1 for class 0, and 0.98, 0.92, and 0.95 for class 1. NASNetMobile + Transfer Learning Layer achieved 0.96, 0.99, and 0.98 for class 0, and 0.99, 0.93, and 0.96 for class 1. ResNet101 + Transfer Learning Layer achieved 0.97, 0.96, and 0.96 for class 0, and 0.93, 0.94, and 0.94 for class 1. For tumor localization, the RESUNET segmentation network accurately delineated tumor regions across all classification models.</div></div><div><h3>Conclusions</h3><div>The proposed dual-stage deep learning framework effectively automates both classification and localization of brain tumors from MRI scans. The results demonstrate strong performance across all architectures, with NASNetMobile + RESUNET achieving the most balanced combination of precision and efficiency. Compared to baseline CNNs and pr
脑肿瘤的分类和定位对于准确诊断和有效的治疗计划至关重要。随着MRI在临床工作流程中的使用越来越多,自动化工具对于协助放射科医生提供快速可靠的分析已经变得至关重要。传统的肿瘤分类和分割方法往往耗时且主观,因此需要基于深度学习的框架来提高诊断的准确性和效率。方法本研究使用来自美国癌症影像档案(TCIA)的TCGA-GBM数据集来评估一种结合分类模型和RESUNET进行分割的双阶段深度学习框架。TCGA-GBM是更广泛的TCGA项目的一部分,专注于多形性胶质母细胞瘤(GBM),一种高度侵袭性的脑癌。该数据集包括3929张图像,其中非肿瘤(0类)样本2556张,肿瘤(1类)样本1373张。该框架结合了三个卷积神经网络(CNN)架构——mobilenet、NASNetMobile和resnet101——每个都增强了一个用于分类的迁移学习层,然后是一个用于肿瘤定位的RESUNET网络。迁移学习可以使用预训练的权重,提高收敛速度和泛化。MobileNet提供了一种轻量级、高效的解决方案;NASNetMobile在准确性和计算成本之间提供了强有力的平衡;ResNet101提供了更深入的特征提取,精度更高。RESUNET架构结合U-Net和残差学习,对肿瘤区域进行了精确的分割,在统一的框架内实现了分类和定位的有效整合。结果在分类阶段,模型对MobileNet +迁移学习层的平均准确率为0.9600,对NASNetMobile +迁移学习层的平均准确率为0.9700,对ResNet101 +迁移学习层的平均准确率为0.9500,显示了模型对脑肿瘤分类的有效性。使用两个类别的准确率、召回率和F1分数进一步评估性能。MobileNet + Transfer Learning Layer对于class 0的准确率为0.95,召回率为0.99,F1为0.97;对于class 1的准确率为0.98,召回率为0.92,F1为0.95。NASNetMobile +迁移学习层在第0类中达到0.96、0.99和0.98,在第1类中达到0.99、0.93和0.96。ResNet101 +迁移学习层在第0类上实现了0.97、0.96和0.96,在第1类上实现了0.93、0.94和0.94。对于肿瘤定位,RESUNET分割网络在所有分类模型中准确地描绘了肿瘤区域。结论所提出的双阶段深度学习框架能够有效地实现脑肿瘤MRI扫描分类和定位的自动化。结果表明,NASNetMobile + RESUNET在所有架构中都具有强大的性能,实现了精度和效率的最平衡组合。与基线cnn和先前的混合模型相比,我们的框架实现了卓越的分类性能和更准确的肿瘤定位,突出了其方法学和实用优势。该框架有望集成到实时临床工作流程中。未来的研究将致力于将该方法扩展到其他肿瘤类型和MRI模式,优化其实时部署,并验证其在不同临床数据集中的普遍性。
{"title":"Dual-stage deep learning framework for brain tumor classification and localization using multimodal MRI scans","authors":"Deependra Rastogi ,&nbsp;Prashant Johri ,&nbsp;Sumit Singh Dhanda ,&nbsp;Anand Singh ,&nbsp;Suman Avdhesh Yadav ,&nbsp;Arfat Ahmad Khan ,&nbsp;Seifedine Kadry","doi":"10.1016/j.ibmed.2026.100361","DOIUrl":"10.1016/j.ibmed.2026.100361","url":null,"abstract":"&lt;div&gt;&lt;h3&gt;Background&lt;/h3&gt;&lt;div&gt;Brain tumor classification and localization are essential for accurate diagnosis and effective treatment planning. With the increasing use of MRI in clinical workflows, automated tools have become crucial for assisting radiologists in providing fast and reliable analysis. Traditional approaches for tumor classification and segmentation are often time-consuming and subjective, underscoring the need for deep learning-based frameworks that can enhance diagnostic accuracy and efficiency.&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;h3&gt;Method&lt;/h3&gt;&lt;div&gt;This study used the TCGA-GBM dataset from The Cancer Imaging Archive (TCIA) to assess a dual-stage deep learning framework combining a classification model with a RESUNET for segmentation. TCGA-GBM, part of the broader TCGA project, focuses on Glioblastoma Multiforme (GBM), a highly aggressive brain cancer. The dataset includes 3929 images, with 2556 non-tumor (class 0) and 1373 tumor (class 1) samples. The framework incorporated three convolutional neural network (CNN) architectures—MobileNet, NASNetMobile, and ResNet101—each enhanced with a Transfer Learning Layer for classification, followed by a RESUNET network for tumor localization. Transfer learning enabled the use of pre-trained weights, improving convergence speed and generalization. MobileNet offered a lightweight, efficient solution; NASNetMobile provided a strong balance between accuracy and computational cost; and ResNet101 delivered deeper feature extraction for higher precision. The RESUNET architecture, combining U-Net and residual learning, accurately segmented tumor regions, enabling effective integration of classification and localization within a unified framework.&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;h3&gt;Results&lt;/h3&gt;&lt;div&gt;In the classification stage, the models achieved average accuracies of 0.9600 for MobileNet + Transfer Learning Layer, 0.9700 for NASNetMobile + Transfer Learning Layer, and 0.9500 for ResNet101 + Transfer Learning Layer, demonstrating their effectiveness in categorizing brain tumors. Performance was further evaluated using precision, recall, and F1 scores for both classes. MobileNet + Transfer Learning Layer achieved 0.95 precision, 0.99 recall, and 0.97 F1 for class 0, and 0.98, 0.92, and 0.95 for class 1. NASNetMobile + Transfer Learning Layer achieved 0.96, 0.99, and 0.98 for class 0, and 0.99, 0.93, and 0.96 for class 1. ResNet101 + Transfer Learning Layer achieved 0.97, 0.96, and 0.96 for class 0, and 0.93, 0.94, and 0.94 for class 1. For tumor localization, the RESUNET segmentation network accurately delineated tumor regions across all classification models.&lt;/div&gt;&lt;/div&gt;&lt;div&gt;&lt;h3&gt;Conclusions&lt;/h3&gt;&lt;div&gt;The proposed dual-stage deep learning framework effectively automates both classification and localization of brain tumors from MRI scans. The results demonstrate strong performance across all architectures, with NASNetMobile + RESUNET achieving the most balanced combination of precision and efficiency. Compared to baseline CNNs and pr","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"13 ","pages":"Article 100361"},"PeriodicalIF":0.0,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147396425","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Diffusion models vs. DCGANs for class-imbalanced lung cancer CT classification: A comparative study 扩散模型与dcgan在分级不平衡肺癌CT分型中的比较研究
Pub Date : 2026-03-01 Epub Date: 2025-12-22 DOI: 10.1016/j.ibmed.2025.100336
Masoud Tabibian, Tahereh Razmpour, Rajib Saha
Effective lung cancer detection from CT scans remains critically challenged by class imbalance where benign and normal cases are underrepresented, leading to biased machine learning models with reduced sensitivity for minority classes and potentially missed diagnoses in cancer screening applications. We present a comprehensive comparative analysis of Diffusion Models and Deep Convolutional Generative Adversarial Networks (DCGANs), both incorporating modern architectural enhancements including spectral normalization, self-attention mechanisms, and conditional generation, for addressing class imbalance in lung cancer CT classification. Using the IQ-OTH/NCCD dataset comprising 1097 CT images across normal, benign, and malignant categories with statistical validation across 10 independent runs, we evaluated both approaches through quantitative image quality metrics (Fréchet Inception Distance, Kullback-Leibler divergence, Kernel Inception Distance, and Inception Score) and downstream classification performance. While Diffusion models consistently outperformed DCGANs across most image quality measures, the clinical significance was confirmed through task-based validation. Both generative approaches successfully addressed class imbalance: DCGAN-augmented datasets achieved overall accuracy of 0.9760 ± 0.0116 with benign recall improvement from 0.833 to 0.933, while Diffusion-augmented datasets reached superior performance of 0.9959 ± 0.0068 with perfect benign recall (1.000 ± 0.000). Critically for cancer screening where false negatives carry severe consequences, Diffusion maintained the highest malignant detection sensitivity (0.997 ± 0.008) with substantially lower performance variance, demonstrating more consistent synthetic data quality. These findings establish that while both modern architectures can mitigate class imbalance, Diffusion models' superior recall performance and lower variability position them as the preferred approach for high-stakes clinical applications, demonstrating that ultimate validation must prioritize downstream clinical task performance over image quality metrics alone.
CT扫描中肺癌的有效检测仍然受到类别不平衡的严重挑战,其中良性和正常病例的代表性不足,导致机器学习模型对少数类别的敏感性降低,并可能在癌症筛查应用中遗漏诊断。我们对扩散模型和深度卷积生成对抗网络(dcgan)进行了全面的比较分析,两者都采用了现代架构增强,包括谱归一化、自注意机制和条件生成,以解决肺癌CT分类中的类别不平衡问题。使用IQ-OTH/NCCD数据集,包括1097张CT图像,包括正常、良性和恶性类别,并在10次独立运行中进行统计验证,我们通过定量图像质量指标(fr起始距离、Kullback-Leibler散度、核起始距离和起始分数)和下游分类性能来评估这两种方法。虽然扩散模型在大多数图像质量测量中始终优于dcgan,但通过基于任务的验证证实了临床意义。两种生成方法都成功地解决了类不平衡问题:dcgan增强数据集的总体准确率为0.9760±0.0116,良性召回率从0.833提高到0.933,而扩散增强数据集的总体准确率为0.9959±0.0068,良性召回率为1.000±0.000。对于假阴性会带来严重后果的癌症筛查来说,Diffusion保持了最高的恶性检测灵敏度(0.997±0.008),性能差异显著降低,显示出更一致的合成数据质量。这些发现表明,虽然这两种现代架构都可以缓解类失衡,但扩散模型优越的召回性能和较低的可变性使其成为高风险临床应用的首选方法,这表明最终验证必须优先考虑下游临床任务性能,而不是单独的图像质量指标。
{"title":"Diffusion models vs. DCGANs for class-imbalanced lung cancer CT classification: A comparative study","authors":"Masoud Tabibian,&nbsp;Tahereh Razmpour,&nbsp;Rajib Saha","doi":"10.1016/j.ibmed.2025.100336","DOIUrl":"10.1016/j.ibmed.2025.100336","url":null,"abstract":"<div><div>Effective lung cancer detection from CT scans remains critically challenged by class imbalance where benign and normal cases are underrepresented, leading to biased machine learning models with reduced sensitivity for minority classes and potentially missed diagnoses in cancer screening applications. We present a comprehensive comparative analysis of Diffusion Models and Deep Convolutional Generative Adversarial Networks (DCGANs), both incorporating modern architectural enhancements including spectral normalization, self-attention mechanisms, and conditional generation, for addressing class imbalance in lung cancer CT classification. Using the IQ-OTH/NCCD dataset comprising 1097 CT images across normal, benign, and malignant categories with statistical validation across 10 independent runs, we evaluated both approaches through quantitative image quality metrics (Fréchet Inception Distance, Kullback-Leibler divergence, Kernel Inception Distance, and Inception Score) and downstream classification performance. While Diffusion models consistently outperformed DCGANs across most image quality measures, the clinical significance was confirmed through task-based validation. Both generative approaches successfully addressed class imbalance: DCGAN-augmented datasets achieved overall accuracy of 0.9760 ± 0.0116 with benign recall improvement from 0.833 to 0.933, while Diffusion-augmented datasets reached superior performance of 0.9959 ± 0.0068 with perfect benign recall (1.000 ± 0.000). Critically for cancer screening where false negatives carry severe consequences, Diffusion maintained the highest malignant detection sensitivity (0.997 ± 0.008) with substantially lower performance variance, demonstrating more consistent synthetic data quality. These findings establish that while both modern architectures can mitigate class imbalance, Diffusion models' superior recall performance and lower variability position them as the preferred approach for high-stakes clinical applications, demonstrating that ultimate validation must prioritize downstream clinical task performance over image quality metrics alone.</div></div>","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"13 ","pages":"Article 100336"},"PeriodicalIF":0.0,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145926916","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TransFusion-BCNet: A transformer-driven multi-modal fusion and explainable deep learning framework for breast cancer diagnosis 输血- bcnet:用于乳腺癌诊断的转换器驱动的多模式融合和可解释的深度学习框架
Pub Date : 2026-03-01 Epub Date: 2026-01-12 DOI: 10.1016/j.ibmed.2026.100346
Ahmed Kateb Jumaah Al-Nussairi , Saleem Malik , Yasser Taha Alzubaidi , S Gopal Krishna Patro , Kasim Sakran Abass , Iman Basheti , Mohammad Khishe
Most AI breast cancer detection systems use single-modality imaging algorithms, limiting clinical reliability. Early and accurate detection improves therapy and mortality. These challenges are addressed by Transformer-driven multi-modal fusion and explainable deep learning system, TransFusion-BCNet for breast cancer diagnosis. The framework consists of three parts. The TriFusion-Transformer (TriFT) performs three-tier fusion: intra-modality fusion across multiple mammogram views and imaging sources, inter-modality fusion combining mammogram, ultrasound, MRI, and clinical features, and decision-level fusion for robust outcome prediction. TriFT detects complicated connections across heterogeneous modalities, unlike classical fusion. Second, we present the FusionAttribution Map (FAMap), a dual-level interpretability mechanism that generates imaging data region-level saliency maps and modality-level contribution scores to evaluate input source influence. This openness helps clinicians understand where and which modality drives predictions. Third, the MetaFusion Optimizer (MFO) adjusts fusion weights, network depth, and learning parameters via evolutionary search and gradient-based fine-tuning. Traditional optimizers lack model generalization and training stability. This staged technique improved both. TransFusion-BCNet outperforms CNN–Transformer hybrids with 99.4 % accuracy, 99.0 % precision, 99.2 % recall, and 99.1 % F1-score in extensive CBIS-DDSM,BUSI, TCGA-BRCA and RIDER Breast MRI datasets. With TriFT, FAMap, and MFO, TransFusion-BCNet provides a robust, transparent, and clinically interpretable diagnostic framework, improving AI in breast cancer screening and decision assistance.
大多数人工智能乳腺癌检测系统使用单模态成像算法,限制了临床可靠性。早期和准确的检测可以改善治疗和死亡率。这些挑战由transformer驱动的多模态融合和可解释的深度学习系统——用于乳腺癌诊断的输血- bcnet来解决。该框架由三部分组成。TriFT进行三层融合:跨多个乳房x光片视图和成像源的模态内融合,结合乳房x光片、超声、MRI和临床特征的模态间融合,以及用于稳健结果预测的决策级融合。与传统的融合不同,TriFT可以检测跨异构模式的复杂连接。其次,我们提出了FusionAttribution Map (FAMap),这是一种双层可解释性机制,可生成成像数据区域级显著性图和模式级贡献分数,以评估输入源的影响。这种开放性有助于临床医生了解在哪里以及哪种模式驱动预测。第三,MetaFusion Optimizer (MFO)通过进化搜索和基于梯度的微调来调整融合权重、网络深度和学习参数。传统的优化算法缺乏模型泛化和训练稳定性。这种分阶段的技术改善了这两方面。在广泛的CBIS-DDSM、BUSI、TCGA-BRCA和RIDER乳腺MRI数据集中,输血- bcnet的准确率为99.4%,精密度为99.0%,召回率为99.2%,f1评分为99.1%,优于CNN-Transformer混合方案。通过TriFT、FAMap和MFO,输血- bcnet提供了一个强大、透明和临床可解释的诊断框架,改善了人工智能在乳腺癌筛查和决策辅助方面的应用。
{"title":"TransFusion-BCNet: A transformer-driven multi-modal fusion and explainable deep learning framework for breast cancer diagnosis","authors":"Ahmed Kateb Jumaah Al-Nussairi ,&nbsp;Saleem Malik ,&nbsp;Yasser Taha Alzubaidi ,&nbsp;S Gopal Krishna Patro ,&nbsp;Kasim Sakran Abass ,&nbsp;Iman Basheti ,&nbsp;Mohammad Khishe","doi":"10.1016/j.ibmed.2026.100346","DOIUrl":"10.1016/j.ibmed.2026.100346","url":null,"abstract":"<div><div>Most AI breast cancer detection systems use single-modality imaging algorithms, limiting clinical reliability. Early and accurate detection improves therapy and mortality. These challenges are addressed by Transformer-driven multi-modal fusion and explainable deep learning system, TransFusion-BCNet for breast cancer diagnosis. The framework consists of three parts. The TriFusion-Transformer (TriFT) performs three-tier fusion: intra-modality fusion across multiple mammogram views and imaging sources, inter-modality fusion combining mammogram, ultrasound, MRI, and clinical features, and decision-level fusion for robust outcome prediction. TriFT detects complicated connections across heterogeneous modalities, unlike classical fusion. Second, we present the FusionAttribution Map (FAMap), a dual-level interpretability mechanism that generates imaging data region-level saliency maps and modality-level contribution scores to evaluate input source influence. This openness helps clinicians understand where and which modality drives predictions. Third, the MetaFusion Optimizer (MFO) adjusts fusion weights, network depth, and learning parameters via evolutionary search and gradient-based fine-tuning. Traditional optimizers lack model generalization and training stability. This staged technique improved both. TransFusion-BCNet outperforms CNN–Transformer hybrids with 99.4 % accuracy, 99.0 % precision, 99.2 % recall, and 99.1 % F1-score in extensive CBIS-DDSM,BUSI, TCGA-BRCA and RIDER Breast MRI datasets. With TriFT, FAMap, and MFO, TransFusion-BCNet provides a robust, transparent, and clinically interpretable diagnostic framework, improving AI in breast cancer screening and decision assistance.</div></div>","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"13 ","pages":"Article 100346"},"PeriodicalIF":0.0,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146077478","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MoE-Net: A deep ensemble framework optimized by genetic algorithm for nasal bone fracture detection on lateral X-ray images MoE-Net:基于遗传算法优化的鼻骨骨折侧位x线图像深度集成框架
Pub Date : 2026-03-01 Epub Date: 2026-02-07 DOI: 10.1016/j.ibmed.2026.100359
Mobin Mehrpour , Seyed Abolghasem Mirroshandel , Tahereh Mortezaei , Zahra Dalili Kajan
Reliable identification of nasal bone fractures on lateral radiographs is a persistent clinical challenge, largely due to the subtle radiographic presentation of fractures, variability and noise inherent in routine imaging, and the limited availability of expert-annotated datasets. To overcome these obstacles, we present MoE-Net, a precision-oriented Mixture-of-Experts Network designed to enhance automated nasal bone fracture detection. The proposed approach systematically evaluates a pool of ten pre-trained convolutional neural network (CNN) and transformer-based architectures, from which the three most accurate and error-resilient models, InceptionResNetV2, DenseNet121, and Xception, are selected as specialized experts. Each expert model undergoes targeted fine-tuning, while a genetic algorithm optimizes their contribution weights within the ensemble to maximize predictive performance. MoE-Net demonstrates substantial performance gains over individual models and existing handcrafted-feature-based methods. On the test dataset, the framework achieves an accuracy of 91.70%, sensitivity of 91.70%, specificity of 89.95%, F2-score of 91.69%, Matthews Correlation Coefficient of 83.48%, and an area under the ROC curve of 91.99%. This performance reflects a clinically meaningful balance between minimizing false negatives, critical for preventing missed diagnoses, and controlling false positives to avoid unnecessary interventions. These findings support the clinical applicability of MoE-Net as a robust, high-performance decision-support tool for nasal bone fracture detection. The study highlights the advantages of precision-driven ensemble strategies in medical imaging and underscores their potential to improve diagnostic accuracy and contribute to more efficient patient care.
在侧位x线片上可靠地识别鼻骨骨折是一个持续的临床挑战,主要是由于骨折的细微x线表现,常规影像学固有的变异性和噪声,以及专家注释数据集的有限可用性。为了克服这些障碍,我们提出了MoE-Net,一个精确导向的混合专家网络,旨在增强自动鼻骨骨折检测。提出的方法系统地评估了10个预训练的卷积神经网络(CNN)和基于变压器的架构,从中选择了三个最准确和最具容错性的模型,InceptionResNetV2, DenseNet121和Xception作为专业专家。每个专家模型经过有针对性的微调,而遗传算法优化他们在集成中的贡献权重,以最大限度地提高预测性能。与单个模型和现有的基于手工特征的方法相比,MoE-Net展示了显著的性能提升。在测试数据集上,该框架的准确率为91.70%,灵敏度为91.70%,特异性为89.95%,f2评分为91.69%,马修斯相关系数为83.48%,ROC曲线下面积为91.99%。这种表现反映了在减少假阴性和控制假阳性之间的临床意义平衡,这对预防漏诊至关重要,避免不必要的干预。这些发现支持了MoE-Net作为一种强大的、高性能的鼻骨骨折检测决策支持工具的临床适用性。该研究强调了精确驱动的集成策略在医学成像中的优势,并强调了它们在提高诊断准确性和促进更有效的患者护理方面的潜力。
{"title":"MoE-Net: A deep ensemble framework optimized by genetic algorithm for nasal bone fracture detection on lateral X-ray images","authors":"Mobin Mehrpour ,&nbsp;Seyed Abolghasem Mirroshandel ,&nbsp;Tahereh Mortezaei ,&nbsp;Zahra Dalili Kajan","doi":"10.1016/j.ibmed.2026.100359","DOIUrl":"10.1016/j.ibmed.2026.100359","url":null,"abstract":"<div><div>Reliable identification of nasal bone fractures on lateral radiographs is a persistent clinical challenge, largely due to the subtle radiographic presentation of fractures, variability and noise inherent in routine imaging, and the limited availability of expert-annotated datasets. To overcome these obstacles, we present MoE-Net, a precision-oriented Mixture-of-Experts Network designed to enhance automated nasal bone fracture detection. The proposed approach systematically evaluates a pool of ten pre-trained convolutional neural network (CNN) and transformer-based architectures, from which the three most accurate and error-resilient models, InceptionResNetV2, DenseNet121, and Xception, are selected as specialized experts. Each expert model undergoes targeted fine-tuning, while a genetic algorithm optimizes their contribution weights within the ensemble to maximize predictive performance. MoE-Net demonstrates substantial performance gains over individual models and existing handcrafted-feature-based methods. On the test dataset, the framework achieves an accuracy of 91.70%, sensitivity of 91.70%, specificity of 89.95%, F2-score of 91.69%, Matthews Correlation Coefficient of 83.48%, and an area under the ROC curve of 91.99%. This performance reflects a clinically meaningful balance between minimizing false negatives, critical for preventing missed diagnoses, and controlling false positives to avoid unnecessary interventions. These findings support the clinical applicability of MoE-Net as a robust, high-performance decision-support tool for nasal bone fracture detection. The study highlights the advantages of precision-driven ensemble strategies in medical imaging and underscores their potential to improve diagnostic accuracy and contribute to more efficient patient care.</div></div>","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"13 ","pages":"Article 100359"},"PeriodicalIF":0.0,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146188803","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing risk prediction for diabetes, hypertension, and heart disease using SMOTE-ENN balancing with PCA and gradient boosting in healthcare AI 在医疗保健AI中使用SMOTE-ENN平衡与PCA和梯度增强来增强糖尿病、高血压和心脏病的风险预测
Pub Date : 2026-03-01 Epub Date: 2026-01-08 DOI: 10.1016/j.ibmed.2025.100339
Tapon Paul , Md Assaduzzaman , Nafiz Fahad , Md Jakir Hossen
Predicting chronic disease remains a crucial problem, particularly in low-resource environments where accurate and timely predictions are of utmost importance. Machine learning methods do not adequately generalize, encounter data imbalance and computation issues when used. In this study, an improved model that makes predictions using Synthetic minority over-sampling technique and edited nearest neighbor balancing techniques and Principal Component Analysis and Gradient Boosting to predict chronic diseases based on readily available clinical profiles is proposed. The dataset used in this study was retrieved from provide precise source: Kaggle dataset. It contains over 1500 anonymized patient records and includes 15 features such as demographic, lifestyle, and clinical measures. Standardized encoding labels, and adjusting classes using SMOTE-ENN have been completed before the PCA was conducted to improve computation speed and reduce overfitting. Decision Tree, Random Forest, LightGBM, XGBoost have been used for comparison to suggest the best performing model seen to be Gradient Boosting. PCA is performed, and the Gradient Boosting approach produces better results. Precision measures how often the classification system is correct when making a positive test result, while recall is determined using a contingency table, and the F1 score, the possibility of modeling outcomes out of 100 trials. The model proposed in the experiment provides the following outputs: Accuracy (CV: 99.33 %, CI: 98.90 %–99.50 %), Precision (CV: 99 %, CI: 98 %–99.5 %), Recall (CV: 99 %, CI: 98 %–99.5 %), F1-Score (CV: 99 %, CI: 98 %–99.5 %). The model's performance was evaluated using cross-validation, yielding an accuracy of 98.90 %. The classifying system performance is specified by the ROC-AUC ranking. It outperforms the model making indefinite projections; its ROC-AUC value is greater than 0.99. The suggested model is a robust, interpretable, and high-precision approach for the early detection of chronic conditions. Therefore, the suggested machine learning system can deliver a considerable promise with respect to creating patient-oriented outcomes.
预测慢性病仍然是一个关键问题,特别是在资源匮乏的环境中,准确和及时的预测至关重要。机器学习方法泛化不充分,使用时会遇到数据不平衡和计算问题。在本研究中,提出了一种改进的模型,该模型使用合成少数过采样技术和编辑最近邻平衡技术以及主成分分析和梯度增强来预测基于现有临床资料的慢性疾病。本研究使用的数据集检索自提供精确来源:Kaggle数据集。它包含1500多名匿名患者记录,包括15个特征,如人口统计、生活方式和临床措施。为了提高计算速度和减少过拟合,在PCA之前已经完成了标准化的编码标签和使用SMOTE-ENN进行类调整。决策树,随机森林,LightGBM, XGBoost已经被用于比较,以表明表现最好的模型是梯度增强。进行了主成分分析,结果表明梯度增强方法效果较好。精确度衡量的是分类系统在产生阳性测试结果时的正确频率,而召回率是使用列联表和F1分数来确定的,F1分数是100次试验中建模结果的可能性。实验中提出的模型提供了以下输出:准确率(CV: 99.33%, CI: 98.90% - 99.50%),精度(CV: 99%, CI: 98% - 99.5%),召回率(CV: 99%, CI: 98% - 99.5%), F1-Score (CV: 99%, CI: 98% - 99.5%)。使用交叉验证对模型的性能进行了评估,准确度为98.90%。分类系统的性能由ROC-AUC排序指定。它优于模型进行不确定的预测;ROC-AUC值大于0.99。所建议的模型是一种稳健的、可解释的、高精度的慢性疾病早期检测方法。因此,建议的机器学习系统在创造以患者为导向的结果方面可以提供相当大的承诺。
{"title":"Enhancing risk prediction for diabetes, hypertension, and heart disease using SMOTE-ENN balancing with PCA and gradient boosting in healthcare AI","authors":"Tapon Paul ,&nbsp;Md Assaduzzaman ,&nbsp;Nafiz Fahad ,&nbsp;Md Jakir Hossen","doi":"10.1016/j.ibmed.2025.100339","DOIUrl":"10.1016/j.ibmed.2025.100339","url":null,"abstract":"<div><div>Predicting chronic disease remains a crucial problem, particularly in low-resource environments where accurate and timely predictions are of utmost importance. Machine learning methods do not adequately generalize, encounter data imbalance and computation issues when used. In this study, an improved model that makes predictions using Synthetic minority over-sampling technique and edited nearest neighbor balancing techniques and Principal Component Analysis and Gradient Boosting to predict chronic diseases based on readily available clinical profiles is proposed. The dataset used in this study was retrieved from provide precise source: Kaggle dataset. It contains over 1500 anonymized patient records and includes 15 features such as demographic, lifestyle, and clinical measures. Standardized encoding labels, and adjusting classes using SMOTE-ENN have been completed before the PCA was conducted to improve computation speed and reduce overfitting. Decision Tree, Random Forest, LightGBM, XGBoost have been used for comparison to suggest the best performing model seen to be Gradient Boosting. PCA is performed, and the Gradient Boosting approach produces better results. Precision measures how often the classification system is correct when making a positive test result, while recall is determined using a contingency table, and the F1 score, the possibility of modeling outcomes out of 100 trials. The model proposed in the experiment provides the following outputs: Accuracy (CV: 99.33 %, CI: 98.90 %–99.50 %), Precision (CV: 99 %, CI: 98 %–99.5 %), Recall (CV: 99 %, CI: 98 %–99.5 %), F1-Score (CV: 99 %, CI: 98 %–99.5 %). The model's performance was evaluated using cross-validation, yielding an accuracy of 98.90 %. The classifying system performance is specified by the ROC-AUC ranking. It outperforms the model making indefinite projections; its ROC-AUC value is greater than 0.99. The suggested model is a robust, interpretable, and high-precision approach for the early detection of chronic conditions. Therefore, the suggested machine learning system can deliver a considerable promise with respect to creating patient-oriented outcomes.</div></div>","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"13 ","pages":"Article 100339"},"PeriodicalIF":0.0,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146037648","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing AI-based oral decision support systems: Hybrid image processing for detecting impacted maxillary canines in orthopantomograms 增强基于人工智能的口腔决策支持系统:混合图像处理在矫形断层摄影中检测上颌埋伏齿
Pub Date : 2026-03-01 Epub Date: 2026-01-07 DOI: 10.1016/j.ibmed.2026.100345
Roopitha C.H , Veena Mayya , V. Sivakumar , Vathsala Patil , Divya Pai , Pavithra Varchas
Timely identification of impacted canines is crucial for preventing complications such as root resorption, misalignment, and damage to neighboring teeth. This research evaluates image processing techniques for AI-based oral decision support systems using orthopantomogram (OPG) images to determine effective preprocessing methods to improve deep learning (DL) model performance. The approach involved preprocessing OPG images followed by object detection using YOLOv5. Preprocessing techniques included median filtering, Gaussian blur, CLAHE, image sharpening, histogram stretching, and CLAHE combined with image sharpening. Performance was evaluated using precision, recall, F1 score, mAP, and IoU metrics. The CLAHE-sharpening combination achieved superior performance with precision of 0.934, recall of 0.932, F1 score of 0.931, mAP of 0.948, and IoU of 0.741, significantly outperforming unprocessed images. Grad-CAM visualizations confirmed that preprocessing enabled the model to identify relevant regions effectively. This study emphasizes the importance of preprocessing methods in improving the diagnostic accuracy in dental radiography for improved treatment planning.
及时识别患牙对于预防并发症如牙根吸收、牙齿错位和对邻近牙齿的损害至关重要。本研究评估了基于人工智能的口头决策支持系统的图像处理技术,使用正体层图(OPG)图像来确定有效的预处理方法,以提高深度学习(DL)模型的性能。该方法包括预处理OPG图像,然后使用YOLOv5进行目标检测。预处理技术包括中值滤波、高斯模糊、CLAHE、图像锐化、直方图拉伸、CLAHE与图像锐化相结合。使用精度、召回率、F1分数、mAP和IoU指标评估性能。clahe -锐化组合的精度为0.934,召回率为0.932,F1评分为0.931,mAP为0.948,IoU为0.741,明显优于未经处理的图像。Grad-CAM可视化证实,预处理使模型能够有效地识别相关区域。本研究强调了预处理方法在提高牙科x线摄影诊断准确性方面的重要性,以改善治疗计划。
{"title":"Enhancing AI-based oral decision support systems: Hybrid image processing for detecting impacted maxillary canines in orthopantomograms","authors":"Roopitha C.H ,&nbsp;Veena Mayya ,&nbsp;V. Sivakumar ,&nbsp;Vathsala Patil ,&nbsp;Divya Pai ,&nbsp;Pavithra Varchas","doi":"10.1016/j.ibmed.2026.100345","DOIUrl":"10.1016/j.ibmed.2026.100345","url":null,"abstract":"<div><div>Timely identification of impacted canines is crucial for preventing complications such as root resorption, misalignment, and damage to neighboring teeth. This research evaluates image processing techniques for AI-based oral decision support systems using orthopantomogram (OPG) images to determine effective preprocessing methods to improve deep learning (DL) model performance. The approach involved preprocessing OPG images followed by object detection using YOLOv5. Preprocessing techniques included median filtering, Gaussian blur, CLAHE, image sharpening, histogram stretching, and CLAHE combined with image sharpening. Performance was evaluated using precision, recall, F1 score, mAP, and IoU metrics. The CLAHE-sharpening combination achieved superior performance with precision of 0.934, recall of 0.932, F1 score of 0.931, mAP of 0.948, and IoU of 0.741, significantly outperforming unprocessed images. Grad-CAM visualizations confirmed that preprocessing enabled the model to identify relevant regions effectively. This study emphasizes the importance of preprocessing methods in improving the diagnostic accuracy in dental radiography for improved treatment planning.</div></div>","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"13 ","pages":"Article 100345"},"PeriodicalIF":0.0,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145926930","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Explainable AI for breast cancer detection: Biglycan biomarker classification with transfer learning 用于乳腺癌检测的可解释人工智能:Biglycan生物标志物分类与迁移学习
Pub Date : 2026-03-01 Epub Date: 2026-01-01 DOI: 10.1016/j.ibmed.2025.100340
Md. Mominul Islam , Naime Akter , Md. Assaduzzaman , Md. Monir Hossain Shimul , Rahmatul Kabir Rasel Sarker

Background

Breast cancer is a leading global cause of cancer-related mortality, where early diagnosis is essential for improved survival outcomes. Although deep learning has shown strong performance in pathology image classification, many models remain difficult to interpret, which limits clinical trust and practical adoption. This study aimed to develop an explainable deep learning framework for classifying breast tissue based on Biglycan (BGN) biomarker expression.

Methods

Immunohistochemical photomicrographs from the publicly available Biglycan Breast Cancer Dataset were processed and classified into cancerous and healthy categories. Three transfer learning models, EfficientNet-B0, DenseNet-161, and ResNet-50, were fine-tuned using ImageNet pre-trained weights under a unified training setting with the Adam optimizer (learning rate = 0.001, batch size = 32, epochs = 50). Model performance was evaluated using accuracy, precision, recall, and F1-score. To enhance interpretability, Gradient-weighted Class Activation Mapping (Grad-CAM) was applied to visualize tissue regions that contributed most to the model's predictions.

Results

EfficientNet-B0 achieved the best overall performance with 99 % accuracy, precision of 0.97, recall of 1.00 for the cancerous class, and an F1-score of 0.99. Grad-CAM heatmaps indicated that the model focused on diagnostically relevant tissue regions associated with strong BGN-related staining patterns. The best-performing model was integrated into a lightweight web-based application to enable real-time image upload and prediction.

Conclusion

This study presents an explainable transfer learning approach for BGN-driven breast tissue classification with strong performance on the Biglycan dataset. The integration of Grad-CAM provides region-level visual explanations that improve transparency, while the web deployment demonstrates a practical pathway for accessible decision support in digital pathology.
乳腺癌是全球癌症相关死亡的主要原因,早期诊断对于改善生存结果至关重要。尽管深度学习在病理图像分类方面表现出色,但许多模型仍然难以解释,这限制了临床信任和实际应用。本研究旨在开发一个可解释的深度学习框架,用于基于Biglycan (BGN)生物标志物表达对乳腺组织进行分类。方法对公开的Biglycan乳腺癌数据集的免疫组化显微照片进行处理,并将其分为癌变和健康两类。使用ImageNet预训练权值,在Adam优化器的统一训练设置下(学习率= 0.001,批大小= 32,epoch = 50),对effentnet - b0、DenseNet-161和ResNet-50三个迁移学习模型进行了微调。使用准确性、精密度、召回率和f1评分来评估模型的性能。为了提高可解释性,应用梯度加权类激活映射(Grad-CAM)来可视化对模型预测贡献最大的组织区域。结果fficientnet - b0的准确率为99%,精密度为0.97,查全率为1.00,f1评分为0.99。Grad-CAM热图显示,该模型专注于与bgn相关的强染色模式相关的诊断相关组织区域。性能最好的模型被集成到一个轻量级的基于web的应用程序中,以实现实时图像上传和预测。本研究提出了一种可解释的迁移学习方法,用于bgnn驱动的乳腺组织分类,在Biglycan数据集上具有较强的性能。Grad-CAM的集成提供了区域级的可视化解释,提高了透明度,而网络部署展示了数字病理学中可访问的决策支持的实用途径。
{"title":"Explainable AI for breast cancer detection: Biglycan biomarker classification with transfer learning","authors":"Md. Mominul Islam ,&nbsp;Naime Akter ,&nbsp;Md. Assaduzzaman ,&nbsp;Md. Monir Hossain Shimul ,&nbsp;Rahmatul Kabir Rasel Sarker","doi":"10.1016/j.ibmed.2025.100340","DOIUrl":"10.1016/j.ibmed.2025.100340","url":null,"abstract":"<div><h3>Background</h3><div>Breast cancer is a leading global cause of cancer-related mortality, where early diagnosis is essential for improved survival outcomes. Although deep learning has shown strong performance in pathology image classification, many models remain difficult to interpret, which limits clinical trust and practical adoption. This study aimed to develop an explainable deep learning framework for classifying breast tissue based on Biglycan (BGN) biomarker expression.</div></div><div><h3>Methods</h3><div>Immunohistochemical photomicrographs from the publicly available Biglycan Breast Cancer Dataset were processed and classified into cancerous and healthy categories. Three transfer learning models, EfficientNet-B0, DenseNet-161, and ResNet-50, were fine-tuned using ImageNet pre-trained weights under a unified training setting with the Adam optimizer (learning rate = 0.001, batch size = 32, epochs = 50). Model performance was evaluated using accuracy, precision, recall, and F1-score. To enhance interpretability, Gradient-weighted Class Activation Mapping (Grad-CAM) was applied to visualize tissue regions that contributed most to the model's predictions.</div></div><div><h3>Results</h3><div>EfficientNet-B0 achieved the best overall performance with 99 % accuracy, precision of 0.97, recall of 1.00 for the cancerous class, and an F1-score of 0.99. Grad-CAM heatmaps indicated that the model focused on diagnostically relevant tissue regions associated with strong BGN-related staining patterns. The best-performing model was integrated into a lightweight web-based application to enable real-time image upload and prediction.</div></div><div><h3>Conclusion</h3><div>This study presents an explainable transfer learning approach for BGN-driven breast tissue classification with strong performance on the Biglycan dataset. The integration of Grad-CAM provides region-level visual explanations that improve transparency, while the web deployment demonstrates a practical pathway for accessible decision support in digital pathology.</div></div>","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"13 ","pages":"Article 100340"},"PeriodicalIF":0.0,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145926928","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automated detection of HER2 gene copy number in breast cancer using deep learning techniques 利用深度学习技术自动检测乳腺癌中HER2基因拷贝数
Pub Date : 2026-03-01 Epub Date: 2025-12-18 DOI: 10.1016/j.ibmed.2025.100333
Terisara Micaraseth , Shanop Shuangshoti , Sakdina Prommaouan , Somruetai Shuangshoti , Rizwan Ullah , Gridsada Phanomchoeng
Accurate evaluation of HER2 gene amplification is critical for guiding breast cancer treatment decisions. This study proposes a deep learning-based diagnostic system for analyzing Dual In Situ Hybridization (DISH) images to support HER2 status assessment. The system integrates two models— YOLOv11-seg for cell detection and YOLOv11 object detection models for HER2 and CEP17 signal quantification—into a unified pipeline. High-resolution whole-slide images were preprocessed and annotated to train the models, which were then embedded into a standalone application designed for clinical environments. Upon uploading TIFF format images, the application performs automated cell detection, red/black signal analysis, and HER2/CEP17 ratio computation. Experimental results demonstrated an accuracy 95.24 % for the best identification and mean deviations of 6.08 % (CEP17) and 12.78 % (HER2) compared with manual counting. Statistical analyses confirm high consistency, particularly in red signal detection. Clinical feedback under scores the system's ease of use, accuracy, and potential to reduce diagnostic burden. The proposed approach demonstrates strong feasibility for routine adoption in pathology workflows.
准确评估HER2基因扩增对指导乳腺癌治疗决策至关重要。本研究提出了一种基于深度学习的诊断系统,用于分析双原位杂交(DISH)图像,以支持HER2状态评估。该系统将用于细胞检测的YOLOv11-seg和用于HER2和CEP17信号量化的YOLOv11目标检测模型集成到一个统一的管道中。对高分辨率的整张幻灯片图像进行预处理和注释以训练模型,然后将其嵌入为临床环境设计的独立应用程序中。在上传TIFF格式的图像后,应用程序执行自动细胞检测,红/黑信号分析和HER2/CEP17比率计算。实验结果表明,与人工计数相比,该方法的最佳识别准确率为95.24%,平均偏差为6.08% (CEP17)和12.78% (HER2)。统计分析证实了高一致性,特别是在红色信号检测方面。临床反馈评分系统的易用性、准确性和减少诊断负担的潜力。所提出的方法证明了在病理工作流程中常规采用的强大可行性。
{"title":"Automated detection of HER2 gene copy number in breast cancer using deep learning techniques","authors":"Terisara Micaraseth ,&nbsp;Shanop Shuangshoti ,&nbsp;Sakdina Prommaouan ,&nbsp;Somruetai Shuangshoti ,&nbsp;Rizwan Ullah ,&nbsp;Gridsada Phanomchoeng","doi":"10.1016/j.ibmed.2025.100333","DOIUrl":"10.1016/j.ibmed.2025.100333","url":null,"abstract":"<div><div>Accurate evaluation of HER2 gene amplification is critical for guiding breast cancer treatment decisions. This study proposes a deep learning-based diagnostic system for analyzing Dual In Situ Hybridization (DISH) images to support HER2 status assessment. The system integrates two models— YOLOv11-seg for cell detection and YOLOv11 object detection models for HER2 and CEP17 signal quantification—into a unified pipeline. High-resolution whole-slide images were preprocessed and annotated to train the models, which were then embedded into a standalone application designed for clinical environments. Upon uploading TIFF format images, the application performs automated cell detection, red/black signal analysis, and HER2/CEP17 ratio computation. Experimental results demonstrated an accuracy 95.24 % for the best identification and mean deviations of 6.08 % (CEP17) and 12.78 % (HER2) compared with manual counting. Statistical analyses confirm high consistency, particularly in red signal detection. Clinical feedback under scores the system's ease of use, accuracy, and potential to reduce diagnostic burden. The proposed approach demonstrates strong feasibility for routine adoption in pathology workflows.</div></div>","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"13 ","pages":"Article 100333"},"PeriodicalIF":0.0,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145926924","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A machine learning–based method for supporting the diagnosis of eosinophilic granulomatosis with polyangiitis: Development and evaluation 一种基于机器学习的支持诊断嗜酸性肉芽肿病合并多血管炎的方法:发展和评估
Pub Date : 2026-03-01 Epub Date: 2025-11-28 DOI: 10.1016/j.ibmed.2025.100317
Yosra Bouattour , Mohamed Hédi Maâloul , Zouhir Bahloul , Sameh Marzouk

Background/objectives

Eosinophilic granulomatosis with polyangiitis (EGPA), formerly Churg-Strauss syndrome, is a rare systemic vasculitis often diagnosed late due to its heterogeneous presentation, leading to severe complications—particularly cardiac involvement, a major cause of morbidity and mortality. We developed EGPA-ML, an artificial intelligence (AI)-based tool using supervised machine learning (ML), to support early and accurate EGPA diagnosis, especially in non-specialized settings.

Methods

A retrospective cohort of patients evaluated for suspected vasculitis at Hedi Chaker Hospital, Sfax, Tunisia, from 1997 to 2023 (nearly three decades), provided 1904 clinical, biological, and histological features. After data cleaning, standardization, and feature selection, 56 key features were retained. Patients were classified as {EGPA} or {NOT_EGPA} per the 2022 ACR/EULAR criteria, with expert consensus (κ = 0.85). Multiple supervised ML algorithms were evaluated via 10-fold cross-validation. The best model was integrated into EGPA-ML, a Java-based clinical decision support system. Performance was assessed on an independent dataset of n = 280 key features, with reference classification {EGPA}/{NOT_EGPA} validated by experts (κ = 0.89).

Results

On the test and evaluation dataset, EGPA-ML achieved a recall of 0.992, precision of 0.869, and F1-score of 0.926. Feature importance analysis identified asthma and eosinophil count as top predictors (36.5 % each), followed by ANCA status, vascular purpura, and histological vasculitis.

Conclusions

EGPA-ML is a high-performance, interpretable, and adaptive tool based on supervised ML, supporting timely EGPA diagnosis. It represents a practical advancement for clinical decision-making in rare diseases, particularly in internal medicine, pulmonology, and cardiology.
背景/目的嗜酸性肉芽肿病合并多血管炎(EGPA),以前称为Churg-Strauss综合征,是一种罕见的全身性血管炎,由于其异质表现,通常诊断较晚,导致严重的并发症,特别是心脏受累,是发病率和死亡率的主要原因。我们开发了EGPA-ML,这是一种基于人工智能(AI)的工具,使用监督机器学习(ML)来支持早期和准确的EGPA诊断,特别是在非专业环境中。方法对1997年至2023年(近30年)在突尼斯斯法克斯Hedi Chaker医院接受疑似血管炎评估的患者进行回顾性队列分析,提供了1904项临床、生物学和组织学特征。经过数据清理、标准化和特征选择,保留了56个关键特征。根据2022年ACR/EULAR标准将患者分为{EGPA}或{NOT_EGPA},专家共识(κ = 0.85)。通过10倍交叉验证评估多个监督ML算法。将最佳模型集成到基于java的临床决策支持系统EGPA-ML中。在n = 280个关键特征的独立数据集上进行性能评估,参考分类{EGPA}/{NOT_EGPA}经过专家验证(κ = 0.89)。结果在测试和评价数据集中,EGPA-ML的召回率为0.992,精密度为0.869,f1得分为0.926。特征重要性分析发现哮喘和嗜酸性粒细胞计数是最重要的预测因子(各占36.5%),其次是ANCA状态、血管性紫癜和组织学血管炎。结论segpa -ML是一种基于监督式ML的高性能、可解释性和自适应的工具,支持EGPA的及时诊断。它代表了罕见病临床决策的实际进步,特别是在内科、肺脏学和心脏病学方面。
{"title":"A machine learning–based method for supporting the diagnosis of eosinophilic granulomatosis with polyangiitis: Development and evaluation","authors":"Yosra Bouattour ,&nbsp;Mohamed Hédi Maâloul ,&nbsp;Zouhir Bahloul ,&nbsp;Sameh Marzouk","doi":"10.1016/j.ibmed.2025.100317","DOIUrl":"10.1016/j.ibmed.2025.100317","url":null,"abstract":"<div><h3>Background/objectives</h3><div>Eosinophilic granulomatosis with polyangiitis (EGPA), formerly Churg-Strauss syndrome, is a rare systemic vasculitis often diagnosed late due to its heterogeneous presentation, leading to severe complications—particularly cardiac involvement, a major cause of morbidity and mortality. We developed EGPA-ML, an artificial intelligence (AI)-based tool using supervised machine learning (ML), to support early and accurate EGPA diagnosis, especially in non-specialized settings.</div></div><div><h3>Methods</h3><div>A retrospective cohort of patients evaluated for suspected vasculitis at Hedi Chaker Hospital, Sfax, Tunisia, from 1997 to 2023 (nearly three decades), provided 1904 clinical, biological, and histological features. After data cleaning, standardization, and feature selection, 56 key features were retained. Patients were classified as {EGPA} or {NOT_EGPA} per the 2022 ACR/EULAR criteria, with expert consensus (κ = 0.85). Multiple supervised ML algorithms were evaluated via 10-fold cross-validation. The best model was integrated into EGPA-ML, a Java-based clinical decision support system. Performance was assessed on an independent dataset of n = 280 key features, with reference classification {EGPA}/{NOT_EGPA} validated by experts (κ = 0.89).</div></div><div><h3>Results</h3><div>On the test and evaluation dataset, EGPA-ML achieved a recall of 0.992, precision of 0.869, and F1-score of 0.926. Feature importance analysis identified asthma and eosinophil count as top predictors (36.5 % each), followed by ANCA status, vascular purpura, and histological vasculitis.</div></div><div><h3>Conclusions</h3><div>EGPA-ML is a high-performance, interpretable, and adaptive tool based on supervised ML, supporting timely EGPA diagnosis. It represents a practical advancement for clinical decision-making in rare diseases, particularly in internal medicine, pulmonology, and cardiology.</div></div>","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"13 ","pages":"Article 100317"},"PeriodicalIF":0.0,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145926989","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sparse representation of high-density retinal signal in time-frequency domain to support diagnosis in psychiatric disorders 高密度视网膜信号的时频稀疏表示支持精神疾病的诊断
Pub Date : 2026-03-01 Epub Date: 2026-02-28 DOI: 10.1016/j.ibmed.2026.100367
Salomon Massoda , Claudia Émond , Vincent Tellier , Hossein Kourkchi , Hind Rakkay , Alexandre Sasseville , George Stoica , Albert Chau , Stuart Coupland , Claude Hariton

Background

We integrated the features used to generate retinal signals, as well as the measured outcome, in our high-density signal concept. In this work, we expand to time-frequency signal analysis processes, and present a novel approach based upon specifically designed frequency patterns.

Methods

A multicentre clinical study was conducted to support prediction models able to accurately differentiate two serious mental health conditions: schizophrenia and type 1 bipolar disorder. Time-frequency domain features were processed using sparse representation (SR) and Principal Component Analysis (PCA). Commonly used mother wavelets have been chosen, as well as specific mother wavelets designed from reference retinal signal patterns. Wavelet coefficients and spectral entropies were selected as retinal signal features in developing the prediction models.

Results

Mappings of predictors in the time-frequency domain demonstrate that retinal signal regions that were not yet investigated include meaningful classifiers. With the implementation of SR combined with time-frequency retinal signal analysis, we developed classification models achieving meaningful level of performances. Cross-validation analyses with 100 replicates reached the highest training performance (99% mean accuracy) with SVM analysis using Gaus8 predictors. In testing, cross-validation highest performance was obtained with the Ridge logistic regression using the Gaus2 predictors (91% mean accuracy). In prediction models using ERGW or ERGWB wavelet coefficients as predictors, cross-validation reached highest testing performance (89% mean accuracy) with LASSO logistic regression model when using ERGWB real even wavelet coefficients as predictors.

Conclusion

Our approach leads to a more refined multimodal construct, able to differentiate subtle information within biosignatures in complex neuropsychiatric disorders. One distinctive improvement is the design and application of pattern-adapted wavelets derived from control-group retinal signals, to be used as deciphering tools for biosignature extraction in pathological conditions.
在我们的高密度信号概念中,我们整合了用于生成视网膜信号的特征以及测量结果。在这项工作中,我们扩展到时频信号分析过程,并提出了一种基于专门设计的频率模式的新方法。方法通过一项多中心临床研究,支持能够准确区分精神分裂症和1型双相情感障碍两种严重精神健康状况的预测模型。采用稀疏表示(SR)和主成分分析(PCA)处理时频域特征。选择了常用的母小波,并根据参考视网膜信号模式设计了特定的母小波。在建立预测模型时,选取小波系数和谱熵作为视网膜信号特征。结果预测因子在时频域中的标记表明,尚未研究的视网膜信号区域包括有意义的分类器。通过SR与时频视网膜信号分析相结合的实现,我们开发的分类模型达到了有意义的性能水平。使用高斯8预测因子的支持向量机分析,100个重复的交叉验证分析达到了最高的训练性能(99%的平均准确率)。在测试中,使用ga2预测因子的Ridge逻辑回归获得了交叉验证的最高性能(平均准确率为91%)。在使用ERGW或ERGWB小波系数作为预测因子的预测模型中,当使用ERGWB实偶小波系数作为预测因子时,LASSO逻辑回归模型的交叉验证达到了最高的测试性能(89%的平均准确率)。我们的方法导致了一个更精细的多模态结构,能够在复杂的神经精神疾病的生物特征中区分细微的信息。一个显著的改进是设计和应用的模式自适应小波从对照组视网膜信号,用于在病理条件下提取生物特征的解密工具。
{"title":"Sparse representation of high-density retinal signal in time-frequency domain to support diagnosis in psychiatric disorders","authors":"Salomon Massoda ,&nbsp;Claudia Émond ,&nbsp;Vincent Tellier ,&nbsp;Hossein Kourkchi ,&nbsp;Hind Rakkay ,&nbsp;Alexandre Sasseville ,&nbsp;George Stoica ,&nbsp;Albert Chau ,&nbsp;Stuart Coupland ,&nbsp;Claude Hariton","doi":"10.1016/j.ibmed.2026.100367","DOIUrl":"10.1016/j.ibmed.2026.100367","url":null,"abstract":"<div><h3>Background</h3><div>We integrated the features used to generate retinal signals, as well as the measured outcome, in our high-density signal concept. In this work, we expand to time-frequency signal analysis processes, and present a novel approach based upon specifically designed frequency patterns.</div></div><div><h3>Methods</h3><div>A multicentre clinical study was conducted to support prediction models able to accurately differentiate two serious mental health conditions: schizophrenia and type 1 bipolar disorder. Time-frequency domain features were processed using sparse representation (SR) and Principal Component Analysis (PCA). Commonly used mother wavelets have been chosen, as well as specific mother wavelets designed from reference retinal signal patterns. Wavelet coefficients and spectral entropies were selected as retinal signal features in developing the prediction models.</div></div><div><h3>Results</h3><div>Mappings of predictors in the time-frequency domain demonstrate that retinal signal regions that were not yet investigated include meaningful classifiers. With the implementation of SR combined with time-frequency retinal signal analysis, we developed classification models achieving meaningful level of performances. Cross-validation analyses with 100 replicates reached the highest training performance (99% mean accuracy) with SVM analysis using Gaus8 predictors. In testing, cross-validation highest performance was obtained with the Ridge logistic regression using the Gaus2 predictors (91% mean accuracy). In prediction models using ERGW or ERGWB wavelet coefficients as predictors, cross-validation reached highest testing performance (89% mean accuracy) with LASSO logistic regression model when using ERGWB real even wavelet coefficients as predictors.</div></div><div><h3>Conclusion</h3><div>Our approach leads to a more refined multimodal construct, able to differentiate subtle information within biosignatures in complex neuropsychiatric disorders. One distinctive improvement is the design and application of pattern-adapted wavelets derived from control-group retinal signals, to be used as deciphering tools for biosignature extraction in pathological conditions.</div></div>","PeriodicalId":73399,"journal":{"name":"Intelligence-based medicine","volume":"13 ","pages":"Article 100367"},"PeriodicalIF":0.0,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147396420","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Intelligence-based medicine
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1