首页 > 最新文献

International Journal of Imaging Systems and Technology最新文献

英文 中文
Comprehensive Experimentation of Pretrained Models on Slice-Based Classification of Interstitial Lung Disease Patterns 基于切片的间质性肺疾病模式的预训练模型综合实验
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-10-30 DOI: 10.1002/ima.70232
Hakan Buyukpatpat, Ebru Akcapinar Sezer, Mehmet Serdar Guzel

Interstitial Lung Diseases (ILD) are typically progressive diseases characterized by poor prognosis due to the inflammation and fibrosis affecting lung tissue. ILD is diagnosed through the identification of specific patterns or combinations of patterns that occur in various regions of the lung. This study employs High-Resolution Computed Tomography (HRCT) scans from the MedGIFT database to classify the patterns causing ILD on a slice-based. To achieve this, the pretrained models and a base Convolutional Neural Network (CNN) are utilized to provide a slice-based classification of ILD patterns in five, six, and seven classes. Four different pretrained models, namely VGG, DenseNet, MobileNet, and EfficientNet, are employed, and the performance impact of two training strategies, namely transfer learning and fine-tuning, is also evaluated. In the study, the effects of four different input resolution types on classification performance were investigated. The features extracted from the pretrained models and a base CNN are classified using a fully connected Artificial Neural Network classifier. The classification performance was further examined using two data augmentation methods for the most successful model and input resolution types. With the EfficientNetB0 pretrained model, classification results of five, six, and seven classes are obtained as 98.070%, 90.819%, and 87.781% F-score, respectively. Additionally, the computational costs and time complexity of all model combinations are analyzed, and their characteristics are comparatively discussed. This study contributes to the limited body of research on slice-based classification and advances clinical practice by facilitating the automatic detection of patterns on HRCT slices as a preprocessing step. Furthermore, the MedGIFT database is systematically analyzed in terms of slice and Region of Interest numbers across different pattern types, offering meaningful insights to support and guide its use in future research.

间质性肺病(ILD)是一种典型的进行性疾病,其特点是由于肺组织的炎症和纤维化而导致预后不良。ILD的诊断是通过识别发生在肺不同区域的特定模式或模式的组合。本研究采用MedGIFT数据库中的高分辨率计算机断层扫描(HRCT),以切片为基础对导致ILD的模式进行分类。为了实现这一点,利用预训练模型和基本卷积神经网络(CNN)提供基于切片的ILD模式分类,分为五类、六类和七类。采用了VGG、DenseNet、MobileNet和EfficientNet四种不同的预训练模型,并评估了迁移学习和微调两种训练策略对性能的影响。在研究中,研究了四种不同的输入分辨率类型对分类性能的影响。从预训练模型和基础CNN中提取的特征使用全连接人工神经网络分类器进行分类。使用两种数据增强方法对最成功的模型和输入分辨率类型进行分类性能进一步检验。使用effentnetb0预训练模型,5类、6类和7类的分类结果f值分别为98.070%、90.819%和87.781%。此外,还分析了各种模型组合的计算成本和时间复杂度,并比较讨论了它们的特点。本研究通过促进HRCT切片模式的自动检测作为预处理步骤,为有限的基于切片的分类研究做出了贡献,并推进了临床实践。此外,根据不同模式类型的切片和感兴趣区域数对MedGIFT数据库进行了系统分析,为支持和指导其在未来研究中的使用提供了有意义的见解。
{"title":"Comprehensive Experimentation of Pretrained Models on Slice-Based Classification of Interstitial Lung Disease Patterns","authors":"Hakan Buyukpatpat,&nbsp;Ebru Akcapinar Sezer,&nbsp;Mehmet Serdar Guzel","doi":"10.1002/ima.70232","DOIUrl":"https://doi.org/10.1002/ima.70232","url":null,"abstract":"<div>\u0000 \u0000 <p>Interstitial Lung Diseases (ILD) are typically progressive diseases characterized by poor prognosis due to the inflammation and fibrosis affecting lung tissue. ILD is diagnosed through the identification of specific patterns or combinations of patterns that occur in various regions of the lung. This study employs High-Resolution Computed Tomography (HRCT) scans from the MedGIFT database to classify the patterns causing ILD on a slice-based. To achieve this, the pretrained models and a base Convolutional Neural Network (CNN) are utilized to provide a slice-based classification of ILD patterns in five, six, and seven classes. Four different pretrained models, namely VGG, DenseNet, MobileNet, and EfficientNet, are employed, and the performance impact of two training strategies, namely transfer learning and fine-tuning, is also evaluated. In the study, the effects of four different input resolution types on classification performance were investigated. The features extracted from the pretrained models and a base CNN are classified using a fully connected Artificial Neural Network classifier. The classification performance was further examined using two data augmentation methods for the most successful model and input resolution types. With the EfficientNetB0 pretrained model, classification results of five, six, and seven classes are obtained as 98.070%, 90.819%, and 87.781% F-score, respectively. Additionally, the computational costs and time complexity of all model combinations are analyzed, and their characteristics are comparatively discussed. This study contributes to the limited body of research on slice-based classification and advances clinical practice by facilitating the automatic detection of patterns on HRCT slices as a preprocessing step. Furthermore, the MedGIFT database is systematically analyzed in terms of slice and Region of Interest numbers across different pattern types, offering meaningful insights to support and guide its use in future research.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145406873","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Non-Invasive Diabetes Detection Through Human Breath Using Hybrid Octave-CenterNet Neural Network With DenseNet-77 Model 基于DenseNet-77模型的混合八度-中心神经网络无创呼吸检测糖尿病
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-10-29 DOI: 10.1002/ima.70237
R. Meena, S. Vinu, J. Omana

Diabetes Mellitus (DM), including Type 1 and Type 2, is a metabolic disorder caused by defects in insulin secretion or action. Non-invasive detection is more critical because invasive methods often lack data and have reduced accuracy, leading to poorer machine learning performance. This research proposes a new Octave-CenterNet with DenseNet-77 framework for efficient detection and classification of diabetes from Volatile Organic Compounds (VOCs). The method combines a rapid discrete curvelet transform with wrapping to capture prominent features quickly, uses octave convolution to preserve high and low-frequency patterns and enrich representations, employs CenterNet to detect acetone as a major biomarker, and leverages DenseNet-77 for gradient-efficient classification. Willow sled catkin optimization adaptively fine-tunes hyperparameters to further enhance performance. The model effectively distinguishes healthy individuals from diabetic patients and differentiates between Type 1 and Type 2 diabetes. Experimental results demonstrate excellent performance with 98.7% accuracy, 98% precision, 99.7% recall, and 99.34% F1 score, validating its robustness. Overall, this end-to-end, noise-resistant, and computationally efficient framework offers a technically advanced and practical solution for non-invasive diabetic detection.

糖尿病(DM),包括1型和2型,是一种由胰岛素分泌或作用缺陷引起的代谢紊乱。非侵入性检测更为关键,因为侵入性方法通常缺乏数据且准确性降低,导致机器学习性能较差。本研究提出了一种新的基于DenseNet-77框架的Octave-CenterNet,用于从挥发性有机化合物(VOCs)中高效检测和分类糖尿病。该方法结合了快速离散曲线变换和包裹来快速捕获突出特征,使用八度卷积来保留高低频模式并丰富表征,使用CenterNet来检测丙酮作为主要生物标志物,并利用DenseNet-77进行梯度高效分类。柳橇柳絮优化自适应微调超参数,进一步提高性能。该模型有效地区分了健康个体和糖尿病患者,并区分了1型和2型糖尿病。实验结果表明,该方法具有98.7%的正确率、98%的精密度、99.7%的召回率和99.34%的F1分数,验证了其鲁棒性。总的来说,这种端到端、抗噪声和计算效率高的框架为非侵入性糖尿病检测提供了技术先进和实用的解决方案。
{"title":"Non-Invasive Diabetes Detection Through Human Breath Using Hybrid Octave-CenterNet Neural Network With DenseNet-77 Model","authors":"R. Meena,&nbsp;S. Vinu,&nbsp;J. Omana","doi":"10.1002/ima.70237","DOIUrl":"https://doi.org/10.1002/ima.70237","url":null,"abstract":"<div>\u0000 \u0000 <p>Diabetes Mellitus (DM), including Type 1 and Type 2, is a metabolic disorder caused by defects in insulin secretion or action. Non-invasive detection is more critical because invasive methods often lack data and have reduced accuracy, leading to poorer machine learning performance. This research proposes a new Octave-CenterNet with DenseNet-77 framework for efficient detection and classification of diabetes from Volatile Organic Compounds (VOCs). The method combines a rapid discrete curvelet transform with wrapping to capture prominent features quickly, uses octave convolution to preserve high and low-frequency patterns and enrich representations, employs CenterNet to detect acetone as a major biomarker, and leverages DenseNet-77 for gradient-efficient classification. Willow sled catkin optimization adaptively fine-tunes hyperparameters to further enhance performance. The model effectively distinguishes healthy individuals from diabetic patients and differentiates between Type 1 and Type 2 diabetes. Experimental results demonstrate excellent performance with 98.7% accuracy, 98% precision, 99.7% recall, and 99.34% F1 score, validating its robustness. Overall, this end-to-end, noise-resistant, and computationally efficient framework offers a technically advanced and practical solution for non-invasive diabetic detection.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145407463","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multimodal Radiomics and Deep Learning Integration for Bone Health Assessment in Postmenopausal Women via Dental Radiographs: Development of an Interpretable Nomogram 多模态放射组学和深度学习集成用于绝经后妇女通过牙科x线片进行骨骼健康评估:可解释Nomogram发展
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-10-28 DOI: 10.1002/ima.70239
Zhengxia Hu, Xiaodong Wang, Hai Lan

To develop and validate a multimodal machine learning model for opportunistic osteoporosis screening in postmenopausal women using dental periapical radiographs. This retrospective multicenter study analyzed 3885 periapical radiographs paired with DEXA-derived T-scores from postmenopausal women. Clinical, handcrafted radiomic, and deep features were extracted, resulting in a fused feature set. Radiomic features (n = 215) followed Image Biomarker Standardization Initiative (IBSI) guidelines, and deep features (n = 128) were derived from a novel attention-based autoencoder. Feature harmonization used ComBat adjustment; reliability was ensured by intra-class correlation coefficient (ICC) filtering (ICC ≥ 0.80). Dimensionality was reduced via Pearson correlation and LASSO regression. Four classifiers—logistic regression, random forest, multilayer perceptron, and XGBoost—were trained and evaluated across stratified training, internal, and external test sets. A logistic regression model was selected for clinical translation and nomogram development. Decision curve analysis assessed clinical utility. XGBoost achieved the highest classification performance using the fused feature set, with an internal AUC of 94.6% and external AUC of 93.7%. Logistic regression maintained strong performance (external AUC = 91.3%) and facilitated nomogram construction. Deep and radiomic features independently outperformed clinical-only models, confirming their predictive strength. SHAP analysis identified DEXA T-score, age, vitamin D, and selected radiomic/deep features as key contributors. Calibration curves and Hosmer–Lemeshow test (p = 0.492) confirmed model reliability. Decision curve analysis showed meaningful net clinical benefit across decision thresholds. Dental periapical radiographs can be leveraged for accurate, non-invasive osteoporosis screening in postmenopausal women. The proposed model demonstrates high accuracy, generalizability, and interpretability, offering a scalable solution for integration into dental practice.

开发并验证一种多模态机器学习模型,用于绝经后妇女根尖周x线片的机会性骨质疏松症筛查。这项回顾性多中心研究分析了绝经后妇女的3885张根尖周围x线片和dexa衍生的t评分。提取临床、手工制作的放射学和深度特征,形成融合的特征集。放射学特征(n = 215)遵循图像生物标志物标准化倡议(IBSI)指南,深度特征(n = 128)来自一种新型的基于注意力的自编码器。特征协调使用战斗调整;通过类内相关系数(ICC)滤波(ICC≥0.80)保证信度。通过Pearson相关和LASSO回归降低维度。四个分类器——逻辑回归、随机森林、多层感知器和xgboost——在分层训练、内部和外部测试集上进行了训练和评估。选择逻辑回归模型进行临床翻译和nomogram发展。决策曲线分析评估临床效用。使用融合特征集,XGBoost实现了最高的分类性能,内部AUC为94.6%,外部AUC为93.7%。Logistic回归保持了较强的表现(外部AUC = 91.3%),并促进了nomogram构建。深度和放射学特征独立优于临床模型,证实了它们的预测强度。SHAP分析确定DEXA t评分、年龄、维生素D和选定的放射学/深部特征是关键因素。校正曲线和Hosmer-Lemeshow检验(p = 0.492)证实了模型的可靠性。决策曲线分析显示有意义的净临床效益跨越决策阈值。牙科根尖周x线片可用于绝经后妇女的准确、非侵入性骨质疏松症筛查。该模型具有较高的准确性、通用性和可解释性,为整合到牙科实践中提供了可扩展的解决方案。
{"title":"Multimodal Radiomics and Deep Learning Integration for Bone Health Assessment in Postmenopausal Women via Dental Radiographs: Development of an Interpretable Nomogram","authors":"Zhengxia Hu,&nbsp;Xiaodong Wang,&nbsp;Hai Lan","doi":"10.1002/ima.70239","DOIUrl":"https://doi.org/10.1002/ima.70239","url":null,"abstract":"<div>\u0000 \u0000 <p>To develop and validate a multimodal machine learning model for opportunistic osteoporosis screening in postmenopausal women using dental periapical radiographs. This retrospective multicenter study analyzed 3885 periapical radiographs paired with DEXA-derived <i>T</i>-scores from postmenopausal women. Clinical, handcrafted radiomic, and deep features were extracted, resulting in a fused feature set. Radiomic features (<i>n</i> = 215) followed Image Biomarker Standardization Initiative (IBSI) guidelines, and deep features (<i>n</i> = 128) were derived from a novel attention-based autoencoder. Feature harmonization used ComBat adjustment; reliability was ensured by intra-class correlation coefficient (ICC) filtering (ICC ≥ 0.80). Dimensionality was reduced via Pearson correlation and LASSO regression. Four classifiers—logistic regression, random forest, multilayer perceptron, and XGBoost—were trained and evaluated across stratified training, internal, and external test sets. A logistic regression model was selected for clinical translation and nomogram development. Decision curve analysis assessed clinical utility. XGBoost achieved the highest classification performance using the fused feature set, with an internal AUC of 94.6% and external AUC of 93.7%. Logistic regression maintained strong performance (external AUC = 91.3%) and facilitated nomogram construction. Deep and radiomic features independently outperformed clinical-only models, confirming their predictive strength. SHAP analysis identified DEXA <i>T</i>-score, age, vitamin D, and selected radiomic/deep features as key contributors. Calibration curves and Hosmer–Lemeshow test (<i>p</i> = 0.492) confirmed model reliability. Decision curve analysis showed meaningful net clinical benefit across decision thresholds. Dental periapical radiographs can be leveraged for accurate, non-invasive osteoporosis screening in postmenopausal women. The proposed model demonstrates high accuracy, generalizability, and interpretability, offering a scalable solution for integration into dental practice.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145406702","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Feature Reconstruction-Guided Multi-Scale Attention Network for Non-Significant Lung Nodule Detection 特征重构引导的多尺度注意网络在非显著性肺结节检测中的应用
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-10-28 DOI: 10.1002/ima.70235
Huiqing Xu, Wei Li, Junfang Tu, Lvchen Cao

Lung cancer remains the leading cause of cancer-related incidence and mortality worldwide. Early detection of lung nodules is crucial for significantly reducing the risk of lung cancer. However, due to the high similarity in CT image features between lung nodules and surrounding normal tissues, nodules are often missed or misidentified during the detection process. Moreover, the diverse types and morphologies of nodules further complicate the development of a unified detection approach. To address these challenges, this study proposes a novel Feature Reconstruction-guided Multi-Scale Attention Network (FRMANet). Specifically, a Refined Feature Reconstruction Module is designed to effectively suppress redundant information while preserving essential feature representations of nodules, ensuring high sensitivity and enhanced representation capability for nodule regions during feature extraction. Additionally, a Multi-scale Feature Enhancement Attention mechanism is introduced, which utilizes an attention-based fusion strategy across multiple scales to fully capture discriminative features of nodules with varying sizes and shapes. Experimental results on the LUNA16 dataset demonstrate that the proposed FRMANet achieves superior detection performance, with a mAP of 0.894 and an F1 score of 0.923, outperforming existing state-of-the-art methods.

肺癌仍然是全球癌症相关发病率和死亡率的主要原因。早期发现肺结节对于显著降低肺癌风险至关重要。然而,由于肺结节与周围正常组织的CT图像特征高度相似,在检测过程中经常被遗漏或误认。此外,结节的不同类型和形态进一步复杂化了统一检测方法的发展。为了解决这些挑战,本研究提出了一种新的特征重构引导的多尺度注意力网络(FRMANet)。具体而言,设计了一个精细化特征重构模块,在保留结节基本特征表示的同时有效地抑制冗余信息,确保特征提取过程中对结节区域的高灵敏度和增强的表示能力。此外,介绍了一种多尺度特征增强注意机制,该机制利用基于注意的多尺度融合策略,充分捕获不同大小和形状的结节的判别特征。在LUNA16数据集上的实验结果表明,本文提出的FRMANet具有优越的检测性能,mAP为0.894,F1分数为0.923,优于现有的先进方法。
{"title":"Feature Reconstruction-Guided Multi-Scale Attention Network for Non-Significant Lung Nodule Detection","authors":"Huiqing Xu,&nbsp;Wei Li,&nbsp;Junfang Tu,&nbsp;Lvchen Cao","doi":"10.1002/ima.70235","DOIUrl":"https://doi.org/10.1002/ima.70235","url":null,"abstract":"<div>\u0000 \u0000 <p>Lung cancer remains the leading cause of cancer-related incidence and mortality worldwide. Early detection of lung nodules is crucial for significantly reducing the risk of lung cancer. However, due to the high similarity in CT image features between lung nodules and surrounding normal tissues, nodules are often missed or misidentified during the detection process. Moreover, the diverse types and morphologies of nodules further complicate the development of a unified detection approach. To address these challenges, this study proposes a novel Feature Reconstruction-guided Multi-Scale Attention Network (FRMANet). Specifically, a Refined Feature Reconstruction Module is designed to effectively suppress redundant information while preserving essential feature representations of nodules, ensuring high sensitivity and enhanced representation capability for nodule regions during feature extraction. Additionally, a Multi-scale Feature Enhancement Attention mechanism is introduced, which utilizes an attention-based fusion strategy across multiple scales to fully capture discriminative features of nodules with varying sizes and shapes. Experimental results on the LUNA16 dataset demonstrate that the proposed FRMANet achieves superior detection performance, with a mAP of 0.894 and an F1 score of 0.923, outperforming existing state-of-the-art methods.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145406806","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Radiomic Feature-Based Prediction of Primary Cancer Origins in Brain Metastases Using Machine Learning 基于放射学特征的脑转移癌起源机器学习预测
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-10-28 DOI: 10.1002/ima.70234
Dilek Betül Sarıdede, Sevim Cengiz

Identifying the primary tumor origin is a critical factor in determining treatment strategies for brain metastases, which remain a major challenge in clinical practice. Traditional diagnostic methods rely on invasive procedures, which may be limited by sampling errors. In this study, a dataset of 200 patients with brain metastases originating from six different cancer types (breast, gastrointestinal, small cell lung, melanoma, non-small cell lung, and renal cell carcinoma) was included. Radiomic features were extracted from different magnetic resonance images (MRI) and selected using the Kruskal–Wallis test, correlation analysis, and ElasticNet regression. Machine learning models, including support vector machine, logistic regression, and random forest, were trained and evaluated using cross-validation and unseen test sets to predict the primary origins of metastatic brain tumors. Our results demonstrate that radiomic features can significantly enhance classification accuracy, with AUC values reaching 0.98 in distinguishing between specific cancer types. Additionally, survival analysis revealed significant differences in survival probabilities across primary tumor types. This study utilizes a larger, single-center cohort and a standardized MRI protocol, applying rigorous feature selection and multiple machine learning classifiers to enhance the robustness and clinical relevance of radiomic predictions. Our findings support the potential of radiomics as a non-invasive tool for metastatic tumor prediction and prognostic assessment, paving the way for improved personalized treatment strategies. Radiomic features extracted from MRI images can significantly enhance the prediction of the main origin of the metastatic tumor types in the brain, thereby informing treatment decisions and prognostic assessments.

确定原发肿瘤的起源是确定脑转移治疗策略的关键因素,这在临床实践中仍然是一个主要挑战。传统的诊断方法依赖于侵入性程序,这可能受到抽样误差的限制。在这项研究中,纳入了200例源自6种不同癌症类型(乳腺癌、胃肠道癌、小细胞肺癌、黑色素瘤、非小细胞肺癌和肾细胞癌)的脑转移患者的数据集。从不同的磁共振图像(MRI)中提取放射学特征,并使用Kruskal-Wallis检验、相关分析和ElasticNet回归进行选择。机器学习模型,包括支持向量机、逻辑回归和随机森林,使用交叉验证和未见测试集进行训练和评估,以预测转移性脑肿瘤的主要起源。我们的研究结果表明,放射学特征可以显著提高分类精度,在区分特定癌症类型时,AUC值达到0.98。此外,生存分析显示不同原发肿瘤类型的生存率存在显著差异。本研究采用更大的单中心队列和标准化的MRI方案,应用严格的特征选择和多个机器学习分类器来增强放射学预测的稳健性和临床相关性。我们的研究结果支持放射组学作为转移性肿瘤预测和预后评估的非侵入性工具的潜力,为改进个性化治疗策略铺平了道路。从MRI图像中提取的放射学特征可以显著增强对脑转移性肿瘤主要来源的预测,从而为治疗决策和预后评估提供信息。
{"title":"Radiomic Feature-Based Prediction of Primary Cancer Origins in Brain Metastases Using Machine Learning","authors":"Dilek Betül Sarıdede,&nbsp;Sevim Cengiz","doi":"10.1002/ima.70234","DOIUrl":"https://doi.org/10.1002/ima.70234","url":null,"abstract":"<div>\u0000 \u0000 <p>Identifying the primary tumor origin is a critical factor in determining treatment strategies for brain metastases, which remain a major challenge in clinical practice. Traditional diagnostic methods rely on invasive procedures, which may be limited by sampling errors. In this study, a dataset of 200 patients with brain metastases originating from six different cancer types (breast, gastrointestinal, small cell lung, melanoma, non-small cell lung, and renal cell carcinoma) was included. Radiomic features were extracted from different magnetic resonance images (MRI) and selected using the Kruskal–Wallis test, correlation analysis, and ElasticNet regression. Machine learning models, including support vector machine, logistic regression, and random forest, were trained and evaluated using cross-validation and unseen test sets to predict the primary origins of metastatic brain tumors. Our results demonstrate that radiomic features can significantly enhance classification accuracy, with AUC values reaching 0.98 in distinguishing between specific cancer types. Additionally, survival analysis revealed significant differences in survival probabilities across primary tumor types. This study utilizes a larger, single-center cohort and a standardized MRI protocol, applying rigorous feature selection and multiple machine learning classifiers to enhance the robustness and clinical relevance of radiomic predictions. Our findings support the potential of radiomics as a non-invasive tool for metastatic tumor prediction and prognostic assessment, paving the way for improved personalized treatment strategies. Radiomic features extracted from MRI images can significantly enhance the prediction of the main origin of the metastatic tumor types in the brain, thereby informing treatment decisions and prognostic assessments.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145406805","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ViTCXRResNet: Harnessing Explainable Artificial Intelligence in Medical Imaging—Chest X-Ray-Based Patients Demographic Prediction 在医学成像中利用可解释的人工智能——基于胸部x光的患者人口统计预测
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-10-27 DOI: 10.1002/ima.70233
Sugirdha Ranganathan, Kirubhasini Srinivasan, Sriramakrishnan Pathmanaban, Kalaiselvi Thiruvenkadam

Patient demographic prediction involves estimating age, gender, ethnicity, and other personal characteristics using X-rays. This can help in personalized medicine and improved healthcare outcomes. It can assist in automated diagnosis for some diseases that exhibit age and gender-specific prevalence. It can also help in forensic science to identify individuals when demographic information is missing. Insights from deep learning can verify the gender and age of self-reported individuals through chest X-rays (CXRs). In this proposed work, we have deployed an artificial intelligence (AI) enabled model which focuses on two tasks: gender classification and age prediction from CXRs. For gender classification, the model combines ResNet-50 (CNN) and Vision Transformer (ViT) to leverage both local feature extraction and global contextual understanding for predicting gender and is called ViTCXRResNet. The model was trained and validated on an Amazon Web Services (SPR) dataset of 10702 images, split with an 80–20 ratio, which was evaluated with classification metrics to determine the model's behavior. For age prediction, extracted features from ResNet-50 were used with dimensionality reduction through principal component analysis (PCA). A fully connected feedforward neural network was trained on the reduced feature set to predict age. The classification and regression model achieves accuracy results of 93.46% for gender classification and 0.86 for the R2 score for age prediction on the SPR dataset. For visual interpretation, explainable AI (Gradient-weighted Class Activation Mapping) was utilized to visualize and find out which parts of the image are prioritized for classifying gender. The proposed model yields high classification accuracy in gender detection and significant accuracy in age prediction. The model shows competitive accuracy compared to existing methods. Further, the demographic prediction stability of the model was proven on two different ethnic groups, such as the Japanese Society of Radiological Technology (JSRT) and Montgomery (USA) datasets.

患者人口统计预测包括使用x射线估计年龄、性别、种族和其他个人特征。这有助于个性化医疗和改善医疗保健结果。它可以帮助对某些表现出年龄和性别特定患病率的疾病进行自动诊断。它还可以帮助法医科学在人口统计信息缺失的情况下识别个人。来自深度学习的见解可以通过胸部x光片(cxr)验证自我报告的个体的性别和年龄。在这项工作中,我们部署了一个人工智能(AI)支持的模型,该模型专注于两项任务:性别分类和来自cxr的年龄预测。对于性别分类,该模型结合了ResNet-50 (CNN)和Vision Transformer (ViT),利用局部特征提取和全局上下文理解来预测性别,称为ViTCXRResNet。该模型在Amazon Web Services (SPR)的10702张图像数据集上进行训练和验证,以80-20的比例进行分割,并使用分类指标进行评估,以确定模型的行为。年龄预测采用ResNet-50提取的特征,并通过主成分分析(PCA)进行降维。在约简特征集上训练全连接前馈神经网络进行年龄预测。该分类回归模型在SPR数据集上对性别分类的准确率为93.46%,对年龄预测的R2评分为0.86。对于视觉解释,使用可解释的AI(梯度加权类激活映射)来可视化并找出图像的哪些部分优先用于分类性别。该模型在性别检测方面具有较高的分类准确率,在年龄预测方面具有显著的准确率。与现有方法相比,该模型具有相当的准确性。此外,在日本放射技术学会(JSRT)和Montgomery(美国)两个不同的族群数据集上验证了该模型的人口统计学预测稳定性。
{"title":"ViTCXRResNet: Harnessing Explainable Artificial Intelligence in Medical Imaging—Chest X-Ray-Based Patients Demographic Prediction","authors":"Sugirdha Ranganathan,&nbsp;Kirubhasini Srinivasan,&nbsp;Sriramakrishnan Pathmanaban,&nbsp;Kalaiselvi Thiruvenkadam","doi":"10.1002/ima.70233","DOIUrl":"https://doi.org/10.1002/ima.70233","url":null,"abstract":"<div>\u0000 \u0000 <p>Patient demographic prediction involves estimating age, gender, ethnicity, and other personal characteristics using X-rays. This can help in personalized medicine and improved healthcare outcomes. It can assist in automated diagnosis for some diseases that exhibit age and gender-specific prevalence. It can also help in forensic science to identify individuals when demographic information is missing. Insights from deep learning can verify the gender and age of self-reported individuals through chest X-rays (CXRs). In this proposed work, we have deployed an artificial intelligence (AI) enabled model which focuses on two tasks: gender classification and age prediction from CXRs. For gender classification, the model combines ResNet-50 (CNN) and Vision Transformer (ViT) to leverage both local feature extraction and global contextual understanding for predicting gender and is called ViTCXRResNet. The model was trained and validated on an Amazon Web Services (SPR) dataset of 10702 images, split with an 80–20 ratio, which was evaluated with classification metrics to determine the model's behavior. For age prediction, extracted features from ResNet-50 were used with dimensionality reduction through principal component analysis (PCA). A fully connected feedforward neural network was trained on the reduced feature set to predict age. The classification and regression model achieves accuracy results of 93.46% for gender classification and 0.86 for the <i>R</i><sup>2</sup> score for age prediction on the SPR dataset. For visual interpretation, explainable AI (Gradient-weighted Class Activation Mapping) was utilized to visualize and find out which parts of the image are prioritized for classifying gender. The proposed model yields high classification accuracy in gender detection and significant accuracy in age prediction. The model shows competitive accuracy compared to existing methods. Further, the demographic prediction stability of the model was proven on two different ethnic groups, such as the Japanese Society of Radiological Technology (JSRT) and Montgomery (USA) datasets.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145406549","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DenseNet201SA++: Enhanced Melanoma Recognition in Dermoscopy Images via Soft Attention Guided Feature Learning densenet201sa++:基于软注意引导特征学习的皮肤镜图像黑色素瘤识别
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-10-24 DOI: 10.1002/ima.70236
Shuangshuang Hu, Xiaomei Xu

As the first line of defense in the human immune system, the skin is highly susceptible to environmental toxins. Melanoma, the most lethal type of skin cancer, is characterized by high mortality and a strong tendency for metastasis. It can sometimes originate from pre-existing nevi, particularly dysplastic nevi. Early identification is crucial for improving patient survival rates. However, traditional skin lesion detection faces challenges due to image quality limitations, dataset imperfections, and the complexity of lesion features. This study proposes the DenseNet201SA++ model, which uses image augmentation techniques and the soft attention mechanism to optimize dermoscopy image quality and automatically capture critical features. Experiments on the HAM10000 dataset with 10,015 dermoscopic images, focusing on binary classification (melanoma vs. nevus), show that the DenseNet201SA++ model achieves significant performance gains, with improvements in precision, recall, F1-score, and accuracy of at least 7.2%, 14.7%, 12.7%, and 14.7% compared to baseline networks. The proposed soft attention-guided feature fusion in DenseNet201SA++ addresses feature redundancy in traditional attention mechanisms, achieving superior performance in distinguishing Mel from Nv, while the DenseNet201 backbone shows distinct advantages. Ablation studies confirm the significant role of data augmentation. The integrated DenseNet201SA++ model achieves robust results with precision, recall, F1-score, and accuracy all reaching 0.983, complemented by an AUC of 0.993. These metrics demonstrate the model's exceptional balance between discriminative power and generalization capability, validating the effectiveness of our proposed architecture.

作为人体免疫系统的第一道防线,皮肤极易受到环境毒素的影响。黑色素瘤是最致命的一种皮肤癌,其特点是死亡率高,有很强的转移倾向。它有时可能源于已有的痣,特别是发育不良的痣。早期识别对提高患者存活率至关重要。然而,由于图像质量的限制、数据集的不完善以及病变特征的复杂性,传统的皮肤病变检测面临着挑战。本研究提出了densenet201sa++模型,该模型利用图像增强技术和软注意机制优化皮肤镜图像质量并自动捕获关键特征。在带有10015张皮肤镜图像的HAM10000数据集上进行的实验,重点是二元分类(黑色素瘤与痣),结果表明,与基线网络相比,densenet201sa+ +模型取得了显著的性能提升,精度、召回率、f1评分和准确率分别提高了7.2%、14.7%、12.7%和14.7%。提出的软注意引导特征融合在densenet201sa++中解决了传统注意机制中的特征冗余,在区分Mel和Nv方面取得了优异的性能,而DenseNet201骨干网则表现出明显的优势。消融研究证实了数据增强的重要作用。集成的DenseNet201SA++模型具有鲁棒性,精密度、召回率、f1分数和准确度均达到0.983,AUC为0.993。这些指标证明了模型在判别能力和泛化能力之间的卓越平衡,验证了我们提出的架构的有效性。
{"title":"DenseNet201SA++: Enhanced Melanoma Recognition in Dermoscopy Images via Soft Attention Guided Feature Learning","authors":"Shuangshuang Hu,&nbsp;Xiaomei Xu","doi":"10.1002/ima.70236","DOIUrl":"https://doi.org/10.1002/ima.70236","url":null,"abstract":"<div>\u0000 \u0000 <p>As the first line of defense in the human immune system, the skin is highly susceptible to environmental toxins. Melanoma, the most lethal type of skin cancer, is characterized by high mortality and a strong tendency for metastasis. It can sometimes originate from pre-existing nevi, particularly dysplastic nevi. Early identification is crucial for improving patient survival rates. However, traditional skin lesion detection faces challenges due to image quality limitations, dataset imperfections, and the complexity of lesion features. This study proposes the DenseNet201SA++ model, which uses image augmentation techniques and the soft attention mechanism to optimize dermoscopy image quality and automatically capture critical features. Experiments on the HAM10000 dataset with 10,015 dermoscopic images, focusing on binary classification (melanoma vs. nevus), show that the DenseNet201SA++ model achieves significant performance gains, with improvements in precision, recall, F1-score, and accuracy of at least 7.2%, 14.7%, 12.7%, and 14.7% compared to baseline networks. The proposed soft attention-guided feature fusion in DenseNet201SA++ addresses feature redundancy in traditional attention mechanisms, achieving superior performance in distinguishing Mel from Nv, while the DenseNet201 backbone shows distinct advantages. Ablation studies confirm the significant role of data augmentation. The integrated DenseNet201SA++ model achieves robust results with precision, recall, F1-score, and accuracy all reaching 0.983, complemented by an AUC of 0.993. These metrics demonstrate the model's exceptional balance between discriminative power and generalization capability, validating the effectiveness of our proposed architecture.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145367145","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Machine Learning Framework for Classification of COVID-19 Variants Using K-mer Based DNA Sequencing 基于K-mer的DNA测序的COVID-19变体分类机器学习框架
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-10-22 DOI: 10.1002/ima.70231
Sunil Kumar, Sanjay Raju, Biswajit Bhowmik

Accurate classification of viral DNA sequences is essential for tracking mutations, understanding viral evolution, and enabling timely public health responses. Traditional alignment-based methods are often computationally intensive and less effective for highly mutating viruses. This article presents a machine learning framework for classifying DNA sequences of COVID-19 variants using K-mer-based tokenization and vectorization techniques inspired by Natural Language Processing (NLP). DNA sequences corresponding to Alpha, Beta, Gamma, and Omicron variants are obtained from the Global Initiative on Sharing All Influenza Data (GISAID) database and encoded into feature vectors. Multiple classifiers, including Extra Trees, Random Forest, Support Vector Classifier (SVC), Decision Tree, Logistic Regression, Naive Bayes, K-Nearest Neighbor (KNN), Ridge Classifier, Stochastic Gradient Descent (SGD), and XGBoost, are evaluated based on accuracy, precision, recall, and F1-score. The Extra Trees model achieved the highest accuracy of 93.10% ±$$ pm $$ 0.42, followed by Random Forest with 92.60% ±$$ pm $$ 0.38, both demonstrating robust and balanced performance. Statistical significance tests confirmed the robustness of the results. The results validate the effectiveness of K-mer-based encoding combined with traditional machine learning models in classifying COVID-19 variants, offering a scalable and efficient solution for genomic surveillance.

病毒DNA序列的准确分类对于跟踪突变、了解病毒进化和及时作出公共卫生反应至关重要。传统的基于比对的方法通常是计算密集型的,并且对高度变异的病毒不太有效。本文提出了一种机器学习框架,用于使用受自然语言处理(NLP)启发的基于k -mer的标记化和矢量化技术对COVID-19变体的DNA序列进行分类。Alpha、Beta、Gamma和Omicron变体对应的DNA序列从共享所有流感数据全球倡议(GISAID)数据库中获得,并编码为特征向量。多个分类器,包括额外树,随机森林,支持向量分类器(SVC),决策树,逻辑回归,朴素贝叶斯,k -最近邻(KNN),山脊分类器,随机梯度下降(SGD)和XGBoost,基于准确性,精密度,召回率和一级分数进行评估。Extra Trees模型达到了93.10的最高精度% ± $$ pm $$ 0.42, followed by Random Forest with 92.60% ± $$ pm $$ 0.38, both demonstrating robust and balanced performance. Statistical significance tests confirmed the robustness of the results. The results validate the effectiveness of K-mer-based encoding combined with traditional machine learning models in classifying COVID-19 variants, offering a scalable and efficient solution for genomic surveillance.
{"title":"Machine Learning Framework for Classification of COVID-19 Variants Using K-mer Based DNA Sequencing","authors":"Sunil Kumar,&nbsp;Sanjay Raju,&nbsp;Biswajit Bhowmik","doi":"10.1002/ima.70231","DOIUrl":"https://doi.org/10.1002/ima.70231","url":null,"abstract":"<div>\u0000 \u0000 <p>Accurate classification of viral DNA sequences is essential for tracking mutations, understanding viral evolution, and enabling timely public health responses. Traditional alignment-based methods are often computationally intensive and less effective for highly mutating viruses. This article presents a machine learning framework for classifying DNA sequences of COVID-19 variants using K-mer-based tokenization and vectorization techniques inspired by Natural Language Processing (NLP). DNA sequences corresponding to Alpha, Beta, Gamma, and Omicron variants are obtained from the Global Initiative on Sharing All Influenza Data (GISAID) database and encoded into feature vectors. Multiple classifiers, including Extra Trees, Random Forest, Support Vector Classifier (SVC), Decision Tree, Logistic Regression, Naive Bayes, K-Nearest Neighbor (KNN), Ridge Classifier, Stochastic Gradient Descent (SGD), and XGBoost, are evaluated based on accuracy, precision, recall, and F1-score. The Extra Trees model achieved the highest accuracy of 93.10% <span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 <mo>±</mo>\u0000 </mrow>\u0000 <annotation>$$ pm $$</annotation>\u0000 </semantics></math> 0.42, followed by Random Forest with 92.60% <span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 <mo>±</mo>\u0000 </mrow>\u0000 <annotation>$$ pm $$</annotation>\u0000 </semantics></math> 0.38, both demonstrating robust and balanced performance. Statistical significance tests confirmed the robustness of the results. The results validate the effectiveness of K-mer-based encoding combined with traditional machine learning models in classifying COVID-19 variants, offering a scalable and efficient solution for genomic surveillance.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-10-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145366787","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
M3IF-(SWT-TVC): Multi-Modal Medical Image Fusion via Weighted Energy, Contrast in the SWT Domain, and Total Variation Minimization With Chambolle's Algorithm M3IF-(SWT- tvc):基于加权能量、SWT域对比度和Chambolle算法的总变异最小化的多模态医学图像融合
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-10-18 DOI: 10.1002/ima.70222
Prabhishek Singh, Manoj Diwakar

The multi-modal medical image fusion (M3IF) combines the required and important information from different medical imaging modalities (computed tomography [CT], magnetic resonance imaging (MRI), positron emission tomography [PET], and single photon emission computed tomography [SPECT]) to provide single informative image. M3IF provides enhanced patient diagnosis, and precise treatment planning. This paper proposes a hybrid M3IF where input medical images are decomposed using stationary wavelet transform (SWT) into low-frequency components (LFCs) and high-frequency components (HFCs). The LFCs and HFCs are fused using energy- and contrast-based metrics. And later reconstruction is performed using inverse SWT (ISWT). The total variation minimization (TVM) using Chambolle's algorithm is applied as a post-refinement operation to reduce noise and preserves the fine details. In this paper, the proposed methodology is termed as M3IF-(SWT-TVC), Here, the acronym TVC is the combination of TVM using Chambolle's algorithm. TVM refinement process is an iterative approach, with the fusion outcomes of M3IF-(SWT-TVC) assessed over a predefined 100 iterations. The TVM, and SWT are blended to balance smoothness and structural details. The final fusion results obtained through M3IF-(SWT-TVC) are evaluated against several prominent non-traditional methods. Based on both visual quality and quantitative metric analysis, it is observed that M3IF-(SWT-TVC) outperforms all the methods used for comparison.

多模态医学图像融合(M3IF)将不同医学成像模式(计算机断层扫描[CT]、磁共振成像(MRI)、正电子发射断层扫描[PET]和单光子发射计算机断层扫描[SPECT])所需的重要信息结合起来,提供单一的信息图像。M3IF提供增强的患者诊断和精确的治疗计划。本文提出了一种混合M3IF算法,利用平稳小波变换(SWT)将输入医学图像分解为低频分量(lfc)和高频分量(hfc)。使用基于能量和对比度的度量来融合低碳化合物和氢氟化合物。然后使用逆SWT (ISWT)进行重建。采用Chambolle算法的总变异最小化(total variation minimization, TVM)作为后精运算,既能降低噪声,又能保留图像的细节。本文提出的方法称为M3IF-(SWT-TVC),这里的缩写TVC是使用Chambolle算法的TVM组合。TVM细化过程是一种迭代方法,在预定义的100次迭代中评估M3IF-(SWT-TVC)的融合结果。TVM和SWT混合,以平衡平滑和结构细节。通过M3IF-(SWT-TVC)获得的最终融合结果与几种突出的非传统方法进行了评估。基于视觉质量和定量度量分析,观察到M3IF-(SWT-TVC)优于所有用于比较的方法。
{"title":"M3IF-(SWT-TVC): Multi-Modal Medical Image Fusion via Weighted Energy, Contrast in the SWT Domain, and Total Variation Minimization With Chambolle's Algorithm","authors":"Prabhishek Singh,&nbsp;Manoj Diwakar","doi":"10.1002/ima.70222","DOIUrl":"https://doi.org/10.1002/ima.70222","url":null,"abstract":"<div>\u0000 \u0000 <p>The multi-modal medical image fusion (M3IF) combines the required and important information from different medical imaging modalities (computed tomography [CT], magnetic resonance imaging (MRI), positron emission tomography [PET], and single photon emission computed tomography [SPECT]) to provide single informative image. M3IF provides enhanced patient diagnosis, and precise treatment planning. This paper proposes a hybrid <i>M3IF</i> where input medical images are decomposed using stationary wavelet transform (SWT) into low-frequency components (LFCs) and high-frequency components (HFCs). The LFCs and HFCs are fused using energy- and contrast-based metrics. And later reconstruction is performed using inverse SWT (ISWT). The total variation minimization (TVM) using Chambolle's algorithm is applied as a post-refinement operation to reduce noise and preserves the fine details. In this paper, the proposed methodology is termed as M3IF-(SWT-TVC), Here, the acronym TVC is the combination of TVM using Chambolle's algorithm. TVM refinement process is an iterative approach, with the fusion outcomes of M3IF-(SWT-TVC) assessed over a predefined 100 iterations. The TVM, and SWT are blended to balance smoothness and structural details. The final fusion results obtained through M3IF-(SWT-TVC) are evaluated against several prominent non-traditional methods. Based on both visual quality and quantitative metric analysis, it is observed that M3IF-(SWT-TVC) outperforms all the methods used for comparison.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-10-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145317671","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimizing Skin Cancer Classification With ResNet-18: A Scalable Approach With 3D Total Body Photography (3D-TBP) 使用ResNet-18优化皮肤癌分类:3D全身摄影(3D- tbp)的可扩展方法
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-10-16 DOI: 10.1002/ima.70224
Javed Rashid, Turke Althobaiti, Alina Shabbir, Muhammad Shoaib Saleem, Muhammad Faheem

Skin cancer, particularly melanoma, remains a major public health challenge because of its rising incidence and mortality rates. Traditional methods of diagnosis, like dermoscopy and biopsies, are invasive, time-consuming, and highly dependent on clinical experience. Furthermore, previous research has predominantly focused on 2D dermoscopic images, which do not capture important volumetric information required for the proper evaluation of the injury. This work introduces a new deep learning architecture based on the ResNet-18 model, augmented by transfer learning, for binary classification of malignant and benign skin lesions. The model is trained on the ISIC 2024 3D Total Body Photography dataset and uses pre-trained ImageNet weights to enable effective feature extraction. To counter the dataset's natural class imbalance and minimize overfitting, the model uses sophisticated data augmentation and oversampling methods. The suggested model boasts a staggering classification accuracy of 99.82%, surpassing many other 2D-based alternatives. The utilization of 3D-TBP offers a strong diagnostic benefit by allowing volumetric lesion analysis, retaining spatial and depth features usually lost in the conventional 2D images. The findings validate the clinical feasibility of the method, presenting a scalable, noninvasive, and very accurate early detection and diagnosis of melanoma using 3D skin imaging.

皮肤癌,特别是黑色素瘤,由于其发病率和死亡率不断上升,仍然是一个重大的公共卫生挑战。传统的诊断方法,如皮肤镜检查和活检,是侵入性的,耗时的,并且高度依赖于临床经验。此外,以前的研究主要集中在2D皮肤镜图像上,这并不能捕捉到正确评估损伤所需的重要体积信息。这项工作引入了一种基于ResNet-18模型的新的深度学习架构,通过迁移学习增强,用于恶性和良性皮肤病变的二元分类。该模型在ISIC 2024 3D全身摄影数据集上进行训练,并使用预训练的ImageNet权值进行有效的特征提取。为了对抗数据集的自然类不平衡并最小化过拟合,该模型使用了复杂的数据增强和过采样方法。所建议的模型拥有惊人的99.82%的分类准确率,超过了许多其他基于2d的替代品。3D-TBP的应用提供了强大的诊断优势,允许对病变进行体积分析,保留了传统2D图像中通常丢失的空间和深度特征。研究结果验证了该方法的临床可行性,通过3D皮肤成像提供了可扩展的、无创的、非常准确的黑色素瘤早期检测和诊断。
{"title":"Optimizing Skin Cancer Classification With ResNet-18: A Scalable Approach With 3D Total Body Photography (3D-TBP)","authors":"Javed Rashid,&nbsp;Turke Althobaiti,&nbsp;Alina Shabbir,&nbsp;Muhammad Shoaib Saleem,&nbsp;Muhammad Faheem","doi":"10.1002/ima.70224","DOIUrl":"https://doi.org/10.1002/ima.70224","url":null,"abstract":"<p>Skin cancer, particularly melanoma, remains a major public health challenge because of its rising incidence and mortality rates. Traditional methods of diagnosis, like dermoscopy and biopsies, are invasive, time-consuming, and highly dependent on clinical experience. Furthermore, previous research has predominantly focused on 2D dermoscopic images, which do not capture important volumetric information required for the proper evaluation of the injury. This work introduces a new deep learning architecture based on the ResNet-18 model, augmented by transfer learning, for binary classification of malignant and benign skin lesions. The model is trained on the ISIC 2024 3D Total Body Photography dataset and uses pre-trained ImageNet weights to enable effective feature extraction. To counter the dataset's natural class imbalance and minimize overfitting, the model uses sophisticated data augmentation and oversampling methods. The suggested model boasts a staggering classification accuracy of 99.82%, surpassing many other 2D-based alternatives. The utilization of 3D-TBP offers a strong diagnostic benefit by allowing volumetric lesion analysis, retaining spatial and depth features usually lost in the conventional 2D images. The findings validate the clinical feasibility of the method, presenting a scalable, noninvasive, and very accurate early detection and diagnosis of melanoma using 3D skin imaging.</p>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ima.70224","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145317266","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International Journal of Imaging Systems and Technology
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1