首页 > 最新文献

International Journal of Imaging Systems and Technology最新文献

英文 中文
Respiratory Differencing: Enhancing Pulmonary Thermal Ablation Evaluation Through Pre- and Intraoperative Image Fusion 呼吸差异:通过术前和术中图像融合增强肺热消融评估
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-10-31 DOI: 10.1002/ima.70242
Wan Li, Wei Li, Hengmo Rong, Yutao Rao, Hui Tang, Yudong Zhang, Feng Wang

CT-guided thermal ablation is increasingly being used for the treatment of lung cancer; however, follow-up studies indicate that physicians' subjective intraoperative assessments often overestimate ablation success, potentially leading to incomplete treatment. To address this limitation, we developed Respiratory Differencing, a CT image-based intraoperative assistance system designed to improve ablation evaluation. The system first segments tumor regions in preoperative CT scans and then applies a multistage registration strategy to align them with intra- or postoperative CT/CBCT images, compensating for respiratory motion and treatment-induced anatomical changes. The system provides two key outputs. First, differential images are generated by subtracting the registered preoperative scan from the intraoperative scan, enabling direct visualization and quantitative comparison of pre- and posttreatment regions. These registered images, together with tumor masks, allow physicians to assess the spatial relationship between tumor and ablation zones—even when the tumor is no longer visible in postablation scans. Second, the system computes a quantitative Ablation Effectiveness Scale (AES) that measures the spatial discrepancy between the tumor region and the ablation zone, offering an objective index of treatment adequacy. By accounting for complex pulmonary deformations and integrating pre- and intraoperative data, this system enhances quality control in ablation procedures. In a retrospective study of 35 clinical cases, Respiratory Differencing significantly outperformed conventional subjective assessment in detecting under-ablation during or immediately after treatment, underscoring its potential to improve intraoperative decision-making and patient outcomes.

ct引导下的热消融越来越多地被用于肺癌的治疗;然而,随访研究表明,医生的主观术中评估往往高估消融成功,可能导致治疗不完全。为了解决这一局限性,我们开发了Respiratory differentiation,这是一种基于CT图像的术中辅助系统,旨在改善消融评估。该系统首先在术前CT扫描中分割肿瘤区域,然后应用多阶段配准策略将其与术中或术后CT/CBCT图像对齐,补偿呼吸运动和治疗引起的解剖变化。系统提供两个关键输出。首先,通过从术中扫描中减去术前扫描记录生成差分图像,可以直接可视化和定量比较治疗前和治疗后区域。这些注册图像,连同肿瘤掩膜,允许医生评估肿瘤和消融区域之间的空间关系——即使在消融后扫描中肿瘤不再可见。其次,系统计算定量消融有效性量表(AES),测量肿瘤区域与消融区域之间的空间差异,提供治疗充分性的客观指标。通过考虑复杂的肺部变形并整合术前和术中数据,该系统提高了消融过程的质量控制。在一项对35例临床病例的回顾性研究中,呼吸差异在治疗期间或治疗后立即检测消融不足方面明显优于传统的主观评估,强调了其改善术中决策和患者预后的潜力。
{"title":"Respiratory Differencing: Enhancing Pulmonary Thermal Ablation Evaluation Through Pre- and Intraoperative Image Fusion","authors":"Wan Li,&nbsp;Wei Li,&nbsp;Hengmo Rong,&nbsp;Yutao Rao,&nbsp;Hui Tang,&nbsp;Yudong Zhang,&nbsp;Feng Wang","doi":"10.1002/ima.70242","DOIUrl":"https://doi.org/10.1002/ima.70242","url":null,"abstract":"<div>\u0000 \u0000 <p>CT-guided thermal ablation is increasingly being used for the treatment of lung cancer; however, follow-up studies indicate that physicians' subjective intraoperative assessments often overestimate ablation success, potentially leading to incomplete treatment. To address this limitation, we developed <i>Respiratory Differencing</i>, a CT image-based intraoperative assistance system designed to improve ablation evaluation. The system first segments tumor regions in preoperative CT scans and then applies a multistage registration strategy to align them with intra- or postoperative CT/CBCT images, compensating for respiratory motion and treatment-induced anatomical changes. The system provides two key outputs. First, differential images are generated by subtracting the registered preoperative scan from the intraoperative scan, enabling direct visualization and quantitative comparison of pre- and posttreatment regions. These registered images, together with tumor masks, allow physicians to assess the spatial relationship between tumor and ablation zones—even when the tumor is no longer visible in postablation scans. Second, the system computes a quantitative Ablation Effectiveness Scale (AES) that measures the spatial discrepancy between the tumor region and the ablation zone, offering an objective index of treatment adequacy. By accounting for complex pulmonary deformations and integrating pre- and intraoperative data, this system enhances quality control in ablation procedures. In a retrospective study of 35 clinical cases, <i>Respiratory Differencing</i> significantly outperformed conventional subjective assessment in detecting under-ablation during or immediately after treatment, underscoring its potential to improve intraoperative decision-making and patient outcomes.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145407406","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adaptive Binary Focal Loss: Enhancing Radiograph Image Classification With Balanced Specificity and Sensitivity 自适应二元焦丢失:增强x线图像分类与平衡的特异性和敏感性
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-10-31 DOI: 10.1002/ima.70238
Gokaramaiah Thota, Nagaraju Karinagappa, Sathya Babu Korra
<div> <p>Convolutional neural networks (CNN) are widely used to classify radiograph images. Musculoskeletal disorders (MSD) of the upper extremity (which comprises upper body parts such as the shoulder, elbow, wrist, and hand, allowing movement, strength and fine motor skills). However, their performance is often limited by class imbalance and the presence of hard samples. Although approaches like ensemble models, capsule networks and regularised CNNs in groups can address these issues, they require substantial computational resources. The adoption of loss function does not require additional computational overhead. Focal loss prioritises hard samples (samples that are not easy to classify); it simultaneously suppresses the gradients for easy samples, which affects learning. This can reduce accuracy and create an imbalance between sensitivity and specificity, which is an undesirable outcome in medical diagnostics. To overcome these limitations, adaptive binary focal loss (ABFL) is proposed here, which combines the strengths of binary cross-entropy and focal loss to achieve balanced learning between easy and hard samples. A balance parameter, <span></span><math> <semantics> <mrow> <mi>λ</mi> </mrow> <annotation>$$ lambda $$</annotation> </semantics></math>, is introduced to adaptively weigh the contributions of binary cross-entropy and focal loss. This approach is further extended to multi-class classification tasks through the proposed adaptive categorical focal loss (ACFL). In addition, a procedure is introduced to automatically tune the three key hyperparameters <span></span><math> <semantics> <mrow> <mi>λ</mi> </mrow> <annotation>$$ lambda $$</annotation> </semantics></math>, <span></span><math> <semantics> <mrow> <mi>γ</mi> </mrow> <annotation>$$ gamma $$</annotation> </semantics></math> and <span></span><math> <semantics> <mrow> <mi>β</mi> </mrow> <annotation>$$ beta $$</annotation> </semantics></math> based on the characteristics of the dataset. This eliminates the need for manual intervention. ABFL and ACFL are compared with seven existing loss functions using DenseNet-169 and Inception-v3 on musculoskeletal radiograph images (MURA), a digital database for screening mammography (DDSM) and a garbage classification dataset. Compared to focal loss, Cohen's kappa score performance improved by 33.70% in ABFL on the MURA finger dataset. Similarly, ACFL achieved improvements of 58.07% and 20.23% on the DDSM and garbage datasets, respectively, while maintaining balanced sensitivity and specificity. These results show the robustness and effectiveness of both ABFL and ACFL in handling clas
卷积神经网络(CNN)被广泛用于x线图像分类。上肢的肌肉骨骼疾病(包括肩部、肘部、手腕和手等上肢部位,允许运动、力量和精细运动技能)。然而,它们的性能往往受到类别不平衡和硬样本存在的限制。尽管像集成模型、胶囊网络和正则化cnn组这样的方法可以解决这些问题,但它们需要大量的计算资源。采用损失函数不需要额外的计算开销。焦点丢失优先考虑硬样本(不容易分类的样本);它同时抑制了简单样本的梯度,这会影响学习。这可能会降低准确性,并造成敏感性和特异性之间的不平衡,这是医疗诊断中不希望看到的结果。为了克服这些局限性,本文提出了自适应二元焦点损失算法(ABFL),该算法结合了二元交叉熵和焦点损失的优点,实现了简单和困难样本之间的平衡学习。引入平衡参数λ $$ lambda $$自适应地衡量二元交叉熵和焦损的贡献。该方法通过提出的自适应分类焦丢失(ACFL)进一步扩展到多类分类任务。此外,还介绍了一种基于数据集特征自动调整三个关键超参数λ $$ lambda $$、γ $$ gamma $$和β $$ beta $$的过程。这消除了人工干预的需要。ABFL和ACFL与现有的7种损失函数进行比较,使用DenseNet-169和Inception-v3对肌肉骨骼x线照片(MURA)、乳房x线摄影筛查数字数据库(DDSM)和垃圾分类数据集进行比较。与失焦组相比,Cohen的kappa评分提高了33.70分% in ABFL on the MURA finger dataset. Similarly, ACFL achieved improvements of 58.07% and 20.23% on the DDSM and garbage datasets, respectively, while maintaining balanced sensitivity and specificity. These results show the robustness and effectiveness of both ABFL and ACFL in handling class imbalance and hard samples in CNN-based classification.
{"title":"Adaptive Binary Focal Loss: Enhancing Radiograph Image Classification With Balanced Specificity and Sensitivity","authors":"Gokaramaiah Thota,&nbsp;Nagaraju Karinagappa,&nbsp;Sathya Babu Korra","doi":"10.1002/ima.70238","DOIUrl":"https://doi.org/10.1002/ima.70238","url":null,"abstract":"&lt;div&gt;\u0000 \u0000 &lt;p&gt;Convolutional neural networks (CNN) are widely used to classify radiograph images. Musculoskeletal disorders (MSD) of the upper extremity (which comprises upper body parts such as the shoulder, elbow, wrist, and hand, allowing movement, strength and fine motor skills). However, their performance is often limited by class imbalance and the presence of hard samples. Although approaches like ensemble models, capsule networks and regularised CNNs in groups can address these issues, they require substantial computational resources. The adoption of loss function does not require additional computational overhead. Focal loss prioritises hard samples (samples that are not easy to classify); it simultaneously suppresses the gradients for easy samples, which affects learning. This can reduce accuracy and create an imbalance between sensitivity and specificity, which is an undesirable outcome in medical diagnostics. To overcome these limitations, adaptive binary focal loss (ABFL) is proposed here, which combines the strengths of binary cross-entropy and focal loss to achieve balanced learning between easy and hard samples. A balance parameter, &lt;span&gt;&lt;/span&gt;&lt;math&gt;\u0000 &lt;semantics&gt;\u0000 &lt;mrow&gt;\u0000 &lt;mi&gt;λ&lt;/mi&gt;\u0000 &lt;/mrow&gt;\u0000 &lt;annotation&gt;$$ lambda $$&lt;/annotation&gt;\u0000 &lt;/semantics&gt;&lt;/math&gt;, is introduced to adaptively weigh the contributions of binary cross-entropy and focal loss. This approach is further extended to multi-class classification tasks through the proposed adaptive categorical focal loss (ACFL). In addition, a procedure is introduced to automatically tune the three key hyperparameters &lt;span&gt;&lt;/span&gt;&lt;math&gt;\u0000 &lt;semantics&gt;\u0000 &lt;mrow&gt;\u0000 &lt;mi&gt;λ&lt;/mi&gt;\u0000 &lt;/mrow&gt;\u0000 &lt;annotation&gt;$$ lambda $$&lt;/annotation&gt;\u0000 &lt;/semantics&gt;&lt;/math&gt;, &lt;span&gt;&lt;/span&gt;&lt;math&gt;\u0000 &lt;semantics&gt;\u0000 &lt;mrow&gt;\u0000 &lt;mi&gt;γ&lt;/mi&gt;\u0000 &lt;/mrow&gt;\u0000 &lt;annotation&gt;$$ gamma $$&lt;/annotation&gt;\u0000 &lt;/semantics&gt;&lt;/math&gt; and &lt;span&gt;&lt;/span&gt;&lt;math&gt;\u0000 &lt;semantics&gt;\u0000 &lt;mrow&gt;\u0000 &lt;mi&gt;β&lt;/mi&gt;\u0000 &lt;/mrow&gt;\u0000 &lt;annotation&gt;$$ beta $$&lt;/annotation&gt;\u0000 &lt;/semantics&gt;&lt;/math&gt; based on the characteristics of the dataset. This eliminates the need for manual intervention. ABFL and ACFL are compared with seven existing loss functions using DenseNet-169 and Inception-v3 on musculoskeletal radiograph images (MURA), a digital database for screening mammography (DDSM) and a garbage classification dataset. Compared to focal loss, Cohen's kappa score performance improved by 33.70% in ABFL on the MURA finger dataset. Similarly, ACFL achieved improvements of 58.07% and 20.23% on the DDSM and garbage datasets, respectively, while maintaining balanced sensitivity and specificity. These results show the robustness and effectiveness of both ABFL and ACFL in handling clas","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145407323","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Smart Lung Cancer Monitoring and Prediction via Optimized Feature Selection and Enhanced Deep CNNs Using Improved GoogleNet and Modified VGG19 基于改进GoogleNet和改进VGG19的优化特征选择和增强深度cnn的智能肺癌监测与预测
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-10-30 DOI: 10.1002/ima.70241
Shaik Bhasha, B. N. Jagadesh

Breathing regulation is centralized in the lungs, ensuring that all body cells receive oxygen. Simultaneously, they filter the air to stop pathogens and unnecessary chemicals from entering the body. Specialized defence systems shield the lungs in the human body. Nevertheless, they are insufficient to remove the chance of developing several lung-related diseases. Inflammation, infections, or even more severe conditions can affect the lungs, potentially leading to the development of a malignant tumor. An efficient healthcare monitoring system is used in the present research to identify people who are at a high risk of developing lung cancer and to provide early treatments to prevent long-term problems. To effectively predict lung cancer, a combined deep-learning intelligence system is proposed in this research. The correct features are selected using the Relief and Improved Least Absolute Shrinkage and Selection Operator (LASSO) techniques, reducing the feature dimension by 42.7%. Then, the deep learning-based improved GoogleNet model is used to extract deep discriminative features of 1024 dimensions, which enhances the feature representation capacity. Finally, the deep learning-based modified VGG19 model predicts lung cancer disease. On the UCI machine learning repository dataset (276 samples), the proposed model achieves 98.85% accuracy, with a 98.78% precision, 98.82% recall, 98.75% F1-score, and 96.52% AUC. The proposed combined intelligent system outperformed cutting-edge techniques regarding feature extraction and lung cancer prediction accuracy, showing a 4%–6% improvement over existing models. The developed smart and intelligent model is sustainable, saving computational resources by approximately 17% and human resources by reducing manual diagnosis time by over 60%. It is safe and helps medical professionals make accurate and timely decisions about lung cancer detection.

呼吸调节集中在肺部,确保身体所有细胞都能获得氧气。同时,它们过滤空气,阻止病原体和不必要的化学物质进入人体。专门的防御系统保护着人体的肺部。然而,它们不足以消除发生几种肺部相关疾病的机会。炎症、感染,甚至更严重的情况都会影响肺部,可能导致恶性肿瘤的发展。本研究采用了一套高效的健康监测系统来识别肺癌高危人群,并提供早期治疗以预防长期问题。为了有效地预测肺癌,本研究提出了一种结合深度学习的智能系统。使用浮雕和改进的最小绝对收缩和选择算子(LASSO)技术选择正确的特征,将特征维度减少42.7%。然后,利用基于深度学习的改进GoogleNet模型提取1024维深度判别特征,增强特征表示能力;最后,基于深度学习的改进VGG19模型预测肺癌疾病。在UCI机器学习存储库数据集(276个样本)上,该模型的准确率为98.85%,精度为98.78%,召回率为98.82%,f1分数为98.75%,AUC为96.52%。所提出的组合智能系统在特征提取和肺癌预测精度方面优于尖端技术,比现有模型提高4%-6%。开发的智能智能模型是可持续的,通过减少60%以上的人工诊断时间,节省了约17%的计算资源和人力资源。它是安全的,可以帮助医疗专业人员对肺癌检测做出准确及时的决定。
{"title":"Smart Lung Cancer Monitoring and Prediction via Optimized Feature Selection and Enhanced Deep CNNs Using Improved GoogleNet and Modified VGG19","authors":"Shaik Bhasha,&nbsp;B. N. Jagadesh","doi":"10.1002/ima.70241","DOIUrl":"https://doi.org/10.1002/ima.70241","url":null,"abstract":"<div>\u0000 \u0000 <p>Breathing regulation is centralized in the lungs, ensuring that all body cells receive oxygen. Simultaneously, they filter the air to stop pathogens and unnecessary chemicals from entering the body. Specialized defence systems shield the lungs in the human body. Nevertheless, they are insufficient to remove the chance of developing several lung-related diseases. Inflammation, infections, or even more severe conditions can affect the lungs, potentially leading to the development of a malignant tumor. An efficient healthcare monitoring system is used in the present research to identify people who are at a high risk of developing lung cancer and to provide early treatments to prevent long-term problems. To effectively predict lung cancer, a combined deep-learning intelligence system is proposed in this research. The correct features are selected using the Relief and Improved Least Absolute Shrinkage and Selection Operator (LASSO) techniques, reducing the feature dimension by 42.7%. Then, the deep learning-based improved GoogleNet model is used to extract deep discriminative features of 1024 dimensions, which enhances the feature representation capacity. Finally, the deep learning-based modified VGG19 model predicts lung cancer disease. On the UCI machine learning repository dataset (276 samples), the proposed model achieves 98.85% accuracy, with a 98.78% precision, 98.82% recall, 98.75% F1-score, and 96.52% AUC. The proposed combined intelligent system outperformed cutting-edge techniques regarding feature extraction and lung cancer prediction accuracy, showing a 4%–6% improvement over existing models. The developed smart and intelligent model is sustainable, saving computational resources by approximately 17% and human resources by reducing manual diagnosis time by over 60%. It is safe and helps medical professionals make accurate and timely decisions about lung cancer detection.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145406872","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Comprehensive Experimentation of Pretrained Models on Slice-Based Classification of Interstitial Lung Disease Patterns 基于切片的间质性肺疾病模式的预训练模型综合实验
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-10-30 DOI: 10.1002/ima.70232
Hakan Buyukpatpat, Ebru Akcapinar Sezer, Mehmet Serdar Guzel

Interstitial Lung Diseases (ILD) are typically progressive diseases characterized by poor prognosis due to the inflammation and fibrosis affecting lung tissue. ILD is diagnosed through the identification of specific patterns or combinations of patterns that occur in various regions of the lung. This study employs High-Resolution Computed Tomography (HRCT) scans from the MedGIFT database to classify the patterns causing ILD on a slice-based. To achieve this, the pretrained models and a base Convolutional Neural Network (CNN) are utilized to provide a slice-based classification of ILD patterns in five, six, and seven classes. Four different pretrained models, namely VGG, DenseNet, MobileNet, and EfficientNet, are employed, and the performance impact of two training strategies, namely transfer learning and fine-tuning, is also evaluated. In the study, the effects of four different input resolution types on classification performance were investigated. The features extracted from the pretrained models and a base CNN are classified using a fully connected Artificial Neural Network classifier. The classification performance was further examined using two data augmentation methods for the most successful model and input resolution types. With the EfficientNetB0 pretrained model, classification results of five, six, and seven classes are obtained as 98.070%, 90.819%, and 87.781% F-score, respectively. Additionally, the computational costs and time complexity of all model combinations are analyzed, and their characteristics are comparatively discussed. This study contributes to the limited body of research on slice-based classification and advances clinical practice by facilitating the automatic detection of patterns on HRCT slices as a preprocessing step. Furthermore, the MedGIFT database is systematically analyzed in terms of slice and Region of Interest numbers across different pattern types, offering meaningful insights to support and guide its use in future research.

间质性肺病(ILD)是一种典型的进行性疾病,其特点是由于肺组织的炎症和纤维化而导致预后不良。ILD的诊断是通过识别发生在肺不同区域的特定模式或模式的组合。本研究采用MedGIFT数据库中的高分辨率计算机断层扫描(HRCT),以切片为基础对导致ILD的模式进行分类。为了实现这一点,利用预训练模型和基本卷积神经网络(CNN)提供基于切片的ILD模式分类,分为五类、六类和七类。采用了VGG、DenseNet、MobileNet和EfficientNet四种不同的预训练模型,并评估了迁移学习和微调两种训练策略对性能的影响。在研究中,研究了四种不同的输入分辨率类型对分类性能的影响。从预训练模型和基础CNN中提取的特征使用全连接人工神经网络分类器进行分类。使用两种数据增强方法对最成功的模型和输入分辨率类型进行分类性能进一步检验。使用effentnetb0预训练模型,5类、6类和7类的分类结果f值分别为98.070%、90.819%和87.781%。此外,还分析了各种模型组合的计算成本和时间复杂度,并比较讨论了它们的特点。本研究通过促进HRCT切片模式的自动检测作为预处理步骤,为有限的基于切片的分类研究做出了贡献,并推进了临床实践。此外,根据不同模式类型的切片和感兴趣区域数对MedGIFT数据库进行了系统分析,为支持和指导其在未来研究中的使用提供了有意义的见解。
{"title":"Comprehensive Experimentation of Pretrained Models on Slice-Based Classification of Interstitial Lung Disease Patterns","authors":"Hakan Buyukpatpat,&nbsp;Ebru Akcapinar Sezer,&nbsp;Mehmet Serdar Guzel","doi":"10.1002/ima.70232","DOIUrl":"https://doi.org/10.1002/ima.70232","url":null,"abstract":"<div>\u0000 \u0000 <p>Interstitial Lung Diseases (ILD) are typically progressive diseases characterized by poor prognosis due to the inflammation and fibrosis affecting lung tissue. ILD is diagnosed through the identification of specific patterns or combinations of patterns that occur in various regions of the lung. This study employs High-Resolution Computed Tomography (HRCT) scans from the MedGIFT database to classify the patterns causing ILD on a slice-based. To achieve this, the pretrained models and a base Convolutional Neural Network (CNN) are utilized to provide a slice-based classification of ILD patterns in five, six, and seven classes. Four different pretrained models, namely VGG, DenseNet, MobileNet, and EfficientNet, are employed, and the performance impact of two training strategies, namely transfer learning and fine-tuning, is also evaluated. In the study, the effects of four different input resolution types on classification performance were investigated. The features extracted from the pretrained models and a base CNN are classified using a fully connected Artificial Neural Network classifier. The classification performance was further examined using two data augmentation methods for the most successful model and input resolution types. With the EfficientNetB0 pretrained model, classification results of five, six, and seven classes are obtained as 98.070%, 90.819%, and 87.781% F-score, respectively. Additionally, the computational costs and time complexity of all model combinations are analyzed, and their characteristics are comparatively discussed. This study contributes to the limited body of research on slice-based classification and advances clinical practice by facilitating the automatic detection of patterns on HRCT slices as a preprocessing step. Furthermore, the MedGIFT database is systematically analyzed in terms of slice and Region of Interest numbers across different pattern types, offering meaningful insights to support and guide its use in future research.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145406873","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Non-Invasive Diabetes Detection Through Human Breath Using Hybrid Octave-CenterNet Neural Network With DenseNet-77 Model 基于DenseNet-77模型的混合八度-中心神经网络无创呼吸检测糖尿病
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-10-29 DOI: 10.1002/ima.70237
R. Meena, S. Vinu, J. Omana

Diabetes Mellitus (DM), including Type 1 and Type 2, is a metabolic disorder caused by defects in insulin secretion or action. Non-invasive detection is more critical because invasive methods often lack data and have reduced accuracy, leading to poorer machine learning performance. This research proposes a new Octave-CenterNet with DenseNet-77 framework for efficient detection and classification of diabetes from Volatile Organic Compounds (VOCs). The method combines a rapid discrete curvelet transform with wrapping to capture prominent features quickly, uses octave convolution to preserve high and low-frequency patterns and enrich representations, employs CenterNet to detect acetone as a major biomarker, and leverages DenseNet-77 for gradient-efficient classification. Willow sled catkin optimization adaptively fine-tunes hyperparameters to further enhance performance. The model effectively distinguishes healthy individuals from diabetic patients and differentiates between Type 1 and Type 2 diabetes. Experimental results demonstrate excellent performance with 98.7% accuracy, 98% precision, 99.7% recall, and 99.34% F1 score, validating its robustness. Overall, this end-to-end, noise-resistant, and computationally efficient framework offers a technically advanced and practical solution for non-invasive diabetic detection.

糖尿病(DM),包括1型和2型,是一种由胰岛素分泌或作用缺陷引起的代谢紊乱。非侵入性检测更为关键,因为侵入性方法通常缺乏数据且准确性降低,导致机器学习性能较差。本研究提出了一种新的基于DenseNet-77框架的Octave-CenterNet,用于从挥发性有机化合物(VOCs)中高效检测和分类糖尿病。该方法结合了快速离散曲线变换和包裹来快速捕获突出特征,使用八度卷积来保留高低频模式并丰富表征,使用CenterNet来检测丙酮作为主要生物标志物,并利用DenseNet-77进行梯度高效分类。柳橇柳絮优化自适应微调超参数,进一步提高性能。该模型有效地区分了健康个体和糖尿病患者,并区分了1型和2型糖尿病。实验结果表明,该方法具有98.7%的正确率、98%的精密度、99.7%的召回率和99.34%的F1分数,验证了其鲁棒性。总的来说,这种端到端、抗噪声和计算效率高的框架为非侵入性糖尿病检测提供了技术先进和实用的解决方案。
{"title":"Non-Invasive Diabetes Detection Through Human Breath Using Hybrid Octave-CenterNet Neural Network With DenseNet-77 Model","authors":"R. Meena,&nbsp;S. Vinu,&nbsp;J. Omana","doi":"10.1002/ima.70237","DOIUrl":"https://doi.org/10.1002/ima.70237","url":null,"abstract":"<div>\u0000 \u0000 <p>Diabetes Mellitus (DM), including Type 1 and Type 2, is a metabolic disorder caused by defects in insulin secretion or action. Non-invasive detection is more critical because invasive methods often lack data and have reduced accuracy, leading to poorer machine learning performance. This research proposes a new Octave-CenterNet with DenseNet-77 framework for efficient detection and classification of diabetes from Volatile Organic Compounds (VOCs). The method combines a rapid discrete curvelet transform with wrapping to capture prominent features quickly, uses octave convolution to preserve high and low-frequency patterns and enrich representations, employs CenterNet to detect acetone as a major biomarker, and leverages DenseNet-77 for gradient-efficient classification. Willow sled catkin optimization adaptively fine-tunes hyperparameters to further enhance performance. The model effectively distinguishes healthy individuals from diabetic patients and differentiates between Type 1 and Type 2 diabetes. Experimental results demonstrate excellent performance with 98.7% accuracy, 98% precision, 99.7% recall, and 99.34% F1 score, validating its robustness. Overall, this end-to-end, noise-resistant, and computationally efficient framework offers a technically advanced and practical solution for non-invasive diabetic detection.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145407463","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multimodal Radiomics and Deep Learning Integration for Bone Health Assessment in Postmenopausal Women via Dental Radiographs: Development of an Interpretable Nomogram 多模态放射组学和深度学习集成用于绝经后妇女通过牙科x线片进行骨骼健康评估:可解释Nomogram发展
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-10-28 DOI: 10.1002/ima.70239
Zhengxia Hu, Xiaodong Wang, Hai Lan

To develop and validate a multimodal machine learning model for opportunistic osteoporosis screening in postmenopausal women using dental periapical radiographs. This retrospective multicenter study analyzed 3885 periapical radiographs paired with DEXA-derived T-scores from postmenopausal women. Clinical, handcrafted radiomic, and deep features were extracted, resulting in a fused feature set. Radiomic features (n = 215) followed Image Biomarker Standardization Initiative (IBSI) guidelines, and deep features (n = 128) were derived from a novel attention-based autoencoder. Feature harmonization used ComBat adjustment; reliability was ensured by intra-class correlation coefficient (ICC) filtering (ICC ≥ 0.80). Dimensionality was reduced via Pearson correlation and LASSO regression. Four classifiers—logistic regression, random forest, multilayer perceptron, and XGBoost—were trained and evaluated across stratified training, internal, and external test sets. A logistic regression model was selected for clinical translation and nomogram development. Decision curve analysis assessed clinical utility. XGBoost achieved the highest classification performance using the fused feature set, with an internal AUC of 94.6% and external AUC of 93.7%. Logistic regression maintained strong performance (external AUC = 91.3%) and facilitated nomogram construction. Deep and radiomic features independently outperformed clinical-only models, confirming their predictive strength. SHAP analysis identified DEXA T-score, age, vitamin D, and selected radiomic/deep features as key contributors. Calibration curves and Hosmer–Lemeshow test (p = 0.492) confirmed model reliability. Decision curve analysis showed meaningful net clinical benefit across decision thresholds. Dental periapical radiographs can be leveraged for accurate, non-invasive osteoporosis screening in postmenopausal women. The proposed model demonstrates high accuracy, generalizability, and interpretability, offering a scalable solution for integration into dental practice.

开发并验证一种多模态机器学习模型,用于绝经后妇女根尖周x线片的机会性骨质疏松症筛查。这项回顾性多中心研究分析了绝经后妇女的3885张根尖周围x线片和dexa衍生的t评分。提取临床、手工制作的放射学和深度特征,形成融合的特征集。放射学特征(n = 215)遵循图像生物标志物标准化倡议(IBSI)指南,深度特征(n = 128)来自一种新型的基于注意力的自编码器。特征协调使用战斗调整;通过类内相关系数(ICC)滤波(ICC≥0.80)保证信度。通过Pearson相关和LASSO回归降低维度。四个分类器——逻辑回归、随机森林、多层感知器和xgboost——在分层训练、内部和外部测试集上进行了训练和评估。选择逻辑回归模型进行临床翻译和nomogram发展。决策曲线分析评估临床效用。使用融合特征集,XGBoost实现了最高的分类性能,内部AUC为94.6%,外部AUC为93.7%。Logistic回归保持了较强的表现(外部AUC = 91.3%),并促进了nomogram构建。深度和放射学特征独立优于临床模型,证实了它们的预测强度。SHAP分析确定DEXA t评分、年龄、维生素D和选定的放射学/深部特征是关键因素。校正曲线和Hosmer-Lemeshow检验(p = 0.492)证实了模型的可靠性。决策曲线分析显示有意义的净临床效益跨越决策阈值。牙科根尖周x线片可用于绝经后妇女的准确、非侵入性骨质疏松症筛查。该模型具有较高的准确性、通用性和可解释性,为整合到牙科实践中提供了可扩展的解决方案。
{"title":"Multimodal Radiomics and Deep Learning Integration for Bone Health Assessment in Postmenopausal Women via Dental Radiographs: Development of an Interpretable Nomogram","authors":"Zhengxia Hu,&nbsp;Xiaodong Wang,&nbsp;Hai Lan","doi":"10.1002/ima.70239","DOIUrl":"https://doi.org/10.1002/ima.70239","url":null,"abstract":"<div>\u0000 \u0000 <p>To develop and validate a multimodal machine learning model for opportunistic osteoporosis screening in postmenopausal women using dental periapical radiographs. This retrospective multicenter study analyzed 3885 periapical radiographs paired with DEXA-derived <i>T</i>-scores from postmenopausal women. Clinical, handcrafted radiomic, and deep features were extracted, resulting in a fused feature set. Radiomic features (<i>n</i> = 215) followed Image Biomarker Standardization Initiative (IBSI) guidelines, and deep features (<i>n</i> = 128) were derived from a novel attention-based autoencoder. Feature harmonization used ComBat adjustment; reliability was ensured by intra-class correlation coefficient (ICC) filtering (ICC ≥ 0.80). Dimensionality was reduced via Pearson correlation and LASSO regression. Four classifiers—logistic regression, random forest, multilayer perceptron, and XGBoost—were trained and evaluated across stratified training, internal, and external test sets. A logistic regression model was selected for clinical translation and nomogram development. Decision curve analysis assessed clinical utility. XGBoost achieved the highest classification performance using the fused feature set, with an internal AUC of 94.6% and external AUC of 93.7%. Logistic regression maintained strong performance (external AUC = 91.3%) and facilitated nomogram construction. Deep and radiomic features independently outperformed clinical-only models, confirming their predictive strength. SHAP analysis identified DEXA <i>T</i>-score, age, vitamin D, and selected radiomic/deep features as key contributors. Calibration curves and Hosmer–Lemeshow test (<i>p</i> = 0.492) confirmed model reliability. Decision curve analysis showed meaningful net clinical benefit across decision thresholds. Dental periapical radiographs can be leveraged for accurate, non-invasive osteoporosis screening in postmenopausal women. The proposed model demonstrates high accuracy, generalizability, and interpretability, offering a scalable solution for integration into dental practice.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145406702","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Feature Reconstruction-Guided Multi-Scale Attention Network for Non-Significant Lung Nodule Detection 特征重构引导的多尺度注意网络在非显著性肺结节检测中的应用
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-10-28 DOI: 10.1002/ima.70235
Huiqing Xu, Wei Li, Junfang Tu, Lvchen Cao

Lung cancer remains the leading cause of cancer-related incidence and mortality worldwide. Early detection of lung nodules is crucial for significantly reducing the risk of lung cancer. However, due to the high similarity in CT image features between lung nodules and surrounding normal tissues, nodules are often missed or misidentified during the detection process. Moreover, the diverse types and morphologies of nodules further complicate the development of a unified detection approach. To address these challenges, this study proposes a novel Feature Reconstruction-guided Multi-Scale Attention Network (FRMANet). Specifically, a Refined Feature Reconstruction Module is designed to effectively suppress redundant information while preserving essential feature representations of nodules, ensuring high sensitivity and enhanced representation capability for nodule regions during feature extraction. Additionally, a Multi-scale Feature Enhancement Attention mechanism is introduced, which utilizes an attention-based fusion strategy across multiple scales to fully capture discriminative features of nodules with varying sizes and shapes. Experimental results on the LUNA16 dataset demonstrate that the proposed FRMANet achieves superior detection performance, with a mAP of 0.894 and an F1 score of 0.923, outperforming existing state-of-the-art methods.

肺癌仍然是全球癌症相关发病率和死亡率的主要原因。早期发现肺结节对于显著降低肺癌风险至关重要。然而,由于肺结节与周围正常组织的CT图像特征高度相似,在检测过程中经常被遗漏或误认。此外,结节的不同类型和形态进一步复杂化了统一检测方法的发展。为了解决这些挑战,本研究提出了一种新的特征重构引导的多尺度注意力网络(FRMANet)。具体而言,设计了一个精细化特征重构模块,在保留结节基本特征表示的同时有效地抑制冗余信息,确保特征提取过程中对结节区域的高灵敏度和增强的表示能力。此外,介绍了一种多尺度特征增强注意机制,该机制利用基于注意的多尺度融合策略,充分捕获不同大小和形状的结节的判别特征。在LUNA16数据集上的实验结果表明,本文提出的FRMANet具有优越的检测性能,mAP为0.894,F1分数为0.923,优于现有的先进方法。
{"title":"Feature Reconstruction-Guided Multi-Scale Attention Network for Non-Significant Lung Nodule Detection","authors":"Huiqing Xu,&nbsp;Wei Li,&nbsp;Junfang Tu,&nbsp;Lvchen Cao","doi":"10.1002/ima.70235","DOIUrl":"https://doi.org/10.1002/ima.70235","url":null,"abstract":"<div>\u0000 \u0000 <p>Lung cancer remains the leading cause of cancer-related incidence and mortality worldwide. Early detection of lung nodules is crucial for significantly reducing the risk of lung cancer. However, due to the high similarity in CT image features between lung nodules and surrounding normal tissues, nodules are often missed or misidentified during the detection process. Moreover, the diverse types and morphologies of nodules further complicate the development of a unified detection approach. To address these challenges, this study proposes a novel Feature Reconstruction-guided Multi-Scale Attention Network (FRMANet). Specifically, a Refined Feature Reconstruction Module is designed to effectively suppress redundant information while preserving essential feature representations of nodules, ensuring high sensitivity and enhanced representation capability for nodule regions during feature extraction. Additionally, a Multi-scale Feature Enhancement Attention mechanism is introduced, which utilizes an attention-based fusion strategy across multiple scales to fully capture discriminative features of nodules with varying sizes and shapes. Experimental results on the LUNA16 dataset demonstrate that the proposed FRMANet achieves superior detection performance, with a mAP of 0.894 and an F1 score of 0.923, outperforming existing state-of-the-art methods.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145406806","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Radiomic Feature-Based Prediction of Primary Cancer Origins in Brain Metastases Using Machine Learning 基于放射学特征的脑转移癌起源机器学习预测
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-10-28 DOI: 10.1002/ima.70234
Dilek Betül Sarıdede, Sevim Cengiz

Identifying the primary tumor origin is a critical factor in determining treatment strategies for brain metastases, which remain a major challenge in clinical practice. Traditional diagnostic methods rely on invasive procedures, which may be limited by sampling errors. In this study, a dataset of 200 patients with brain metastases originating from six different cancer types (breast, gastrointestinal, small cell lung, melanoma, non-small cell lung, and renal cell carcinoma) was included. Radiomic features were extracted from different magnetic resonance images (MRI) and selected using the Kruskal–Wallis test, correlation analysis, and ElasticNet regression. Machine learning models, including support vector machine, logistic regression, and random forest, were trained and evaluated using cross-validation and unseen test sets to predict the primary origins of metastatic brain tumors. Our results demonstrate that radiomic features can significantly enhance classification accuracy, with AUC values reaching 0.98 in distinguishing between specific cancer types. Additionally, survival analysis revealed significant differences in survival probabilities across primary tumor types. This study utilizes a larger, single-center cohort and a standardized MRI protocol, applying rigorous feature selection and multiple machine learning classifiers to enhance the robustness and clinical relevance of radiomic predictions. Our findings support the potential of radiomics as a non-invasive tool for metastatic tumor prediction and prognostic assessment, paving the way for improved personalized treatment strategies. Radiomic features extracted from MRI images can significantly enhance the prediction of the main origin of the metastatic tumor types in the brain, thereby informing treatment decisions and prognostic assessments.

确定原发肿瘤的起源是确定脑转移治疗策略的关键因素,这在临床实践中仍然是一个主要挑战。传统的诊断方法依赖于侵入性程序,这可能受到抽样误差的限制。在这项研究中,纳入了200例源自6种不同癌症类型(乳腺癌、胃肠道癌、小细胞肺癌、黑色素瘤、非小细胞肺癌和肾细胞癌)的脑转移患者的数据集。从不同的磁共振图像(MRI)中提取放射学特征,并使用Kruskal-Wallis检验、相关分析和ElasticNet回归进行选择。机器学习模型,包括支持向量机、逻辑回归和随机森林,使用交叉验证和未见测试集进行训练和评估,以预测转移性脑肿瘤的主要起源。我们的研究结果表明,放射学特征可以显著提高分类精度,在区分特定癌症类型时,AUC值达到0.98。此外,生存分析显示不同原发肿瘤类型的生存率存在显著差异。本研究采用更大的单中心队列和标准化的MRI方案,应用严格的特征选择和多个机器学习分类器来增强放射学预测的稳健性和临床相关性。我们的研究结果支持放射组学作为转移性肿瘤预测和预后评估的非侵入性工具的潜力,为改进个性化治疗策略铺平了道路。从MRI图像中提取的放射学特征可以显著增强对脑转移性肿瘤主要来源的预测,从而为治疗决策和预后评估提供信息。
{"title":"Radiomic Feature-Based Prediction of Primary Cancer Origins in Brain Metastases Using Machine Learning","authors":"Dilek Betül Sarıdede,&nbsp;Sevim Cengiz","doi":"10.1002/ima.70234","DOIUrl":"https://doi.org/10.1002/ima.70234","url":null,"abstract":"<div>\u0000 \u0000 <p>Identifying the primary tumor origin is a critical factor in determining treatment strategies for brain metastases, which remain a major challenge in clinical practice. Traditional diagnostic methods rely on invasive procedures, which may be limited by sampling errors. In this study, a dataset of 200 patients with brain metastases originating from six different cancer types (breast, gastrointestinal, small cell lung, melanoma, non-small cell lung, and renal cell carcinoma) was included. Radiomic features were extracted from different magnetic resonance images (MRI) and selected using the Kruskal–Wallis test, correlation analysis, and ElasticNet regression. Machine learning models, including support vector machine, logistic regression, and random forest, were trained and evaluated using cross-validation and unseen test sets to predict the primary origins of metastatic brain tumors. Our results demonstrate that radiomic features can significantly enhance classification accuracy, with AUC values reaching 0.98 in distinguishing between specific cancer types. Additionally, survival analysis revealed significant differences in survival probabilities across primary tumor types. This study utilizes a larger, single-center cohort and a standardized MRI protocol, applying rigorous feature selection and multiple machine learning classifiers to enhance the robustness and clinical relevance of radiomic predictions. Our findings support the potential of radiomics as a non-invasive tool for metastatic tumor prediction and prognostic assessment, paving the way for improved personalized treatment strategies. Radiomic features extracted from MRI images can significantly enhance the prediction of the main origin of the metastatic tumor types in the brain, thereby informing treatment decisions and prognostic assessments.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145406805","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ViTCXRResNet: Harnessing Explainable Artificial Intelligence in Medical Imaging—Chest X-Ray-Based Patients Demographic Prediction 在医学成像中利用可解释的人工智能——基于胸部x光的患者人口统计预测
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-10-27 DOI: 10.1002/ima.70233
Sugirdha Ranganathan, Kirubhasini Srinivasan, Sriramakrishnan Pathmanaban, Kalaiselvi Thiruvenkadam

Patient demographic prediction involves estimating age, gender, ethnicity, and other personal characteristics using X-rays. This can help in personalized medicine and improved healthcare outcomes. It can assist in automated diagnosis for some diseases that exhibit age and gender-specific prevalence. It can also help in forensic science to identify individuals when demographic information is missing. Insights from deep learning can verify the gender and age of self-reported individuals through chest X-rays (CXRs). In this proposed work, we have deployed an artificial intelligence (AI) enabled model which focuses on two tasks: gender classification and age prediction from CXRs. For gender classification, the model combines ResNet-50 (CNN) and Vision Transformer (ViT) to leverage both local feature extraction and global contextual understanding for predicting gender and is called ViTCXRResNet. The model was trained and validated on an Amazon Web Services (SPR) dataset of 10702 images, split with an 80–20 ratio, which was evaluated with classification metrics to determine the model's behavior. For age prediction, extracted features from ResNet-50 were used with dimensionality reduction through principal component analysis (PCA). A fully connected feedforward neural network was trained on the reduced feature set to predict age. The classification and regression model achieves accuracy results of 93.46% for gender classification and 0.86 for the R2 score for age prediction on the SPR dataset. For visual interpretation, explainable AI (Gradient-weighted Class Activation Mapping) was utilized to visualize and find out which parts of the image are prioritized for classifying gender. The proposed model yields high classification accuracy in gender detection and significant accuracy in age prediction. The model shows competitive accuracy compared to existing methods. Further, the demographic prediction stability of the model was proven on two different ethnic groups, such as the Japanese Society of Radiological Technology (JSRT) and Montgomery (USA) datasets.

患者人口统计预测包括使用x射线估计年龄、性别、种族和其他个人特征。这有助于个性化医疗和改善医疗保健结果。它可以帮助对某些表现出年龄和性别特定患病率的疾病进行自动诊断。它还可以帮助法医科学在人口统计信息缺失的情况下识别个人。来自深度学习的见解可以通过胸部x光片(cxr)验证自我报告的个体的性别和年龄。在这项工作中,我们部署了一个人工智能(AI)支持的模型,该模型专注于两项任务:性别分类和来自cxr的年龄预测。对于性别分类,该模型结合了ResNet-50 (CNN)和Vision Transformer (ViT),利用局部特征提取和全局上下文理解来预测性别,称为ViTCXRResNet。该模型在Amazon Web Services (SPR)的10702张图像数据集上进行训练和验证,以80-20的比例进行分割,并使用分类指标进行评估,以确定模型的行为。年龄预测采用ResNet-50提取的特征,并通过主成分分析(PCA)进行降维。在约简特征集上训练全连接前馈神经网络进行年龄预测。该分类回归模型在SPR数据集上对性别分类的准确率为93.46%,对年龄预测的R2评分为0.86。对于视觉解释,使用可解释的AI(梯度加权类激活映射)来可视化并找出图像的哪些部分优先用于分类性别。该模型在性别检测方面具有较高的分类准确率,在年龄预测方面具有显著的准确率。与现有方法相比,该模型具有相当的准确性。此外,在日本放射技术学会(JSRT)和Montgomery(美国)两个不同的族群数据集上验证了该模型的人口统计学预测稳定性。
{"title":"ViTCXRResNet: Harnessing Explainable Artificial Intelligence in Medical Imaging—Chest X-Ray-Based Patients Demographic Prediction","authors":"Sugirdha Ranganathan,&nbsp;Kirubhasini Srinivasan,&nbsp;Sriramakrishnan Pathmanaban,&nbsp;Kalaiselvi Thiruvenkadam","doi":"10.1002/ima.70233","DOIUrl":"https://doi.org/10.1002/ima.70233","url":null,"abstract":"<div>\u0000 \u0000 <p>Patient demographic prediction involves estimating age, gender, ethnicity, and other personal characteristics using X-rays. This can help in personalized medicine and improved healthcare outcomes. It can assist in automated diagnosis for some diseases that exhibit age and gender-specific prevalence. It can also help in forensic science to identify individuals when demographic information is missing. Insights from deep learning can verify the gender and age of self-reported individuals through chest X-rays (CXRs). In this proposed work, we have deployed an artificial intelligence (AI) enabled model which focuses on two tasks: gender classification and age prediction from CXRs. For gender classification, the model combines ResNet-50 (CNN) and Vision Transformer (ViT) to leverage both local feature extraction and global contextual understanding for predicting gender and is called ViTCXRResNet. The model was trained and validated on an Amazon Web Services (SPR) dataset of 10702 images, split with an 80–20 ratio, which was evaluated with classification metrics to determine the model's behavior. For age prediction, extracted features from ResNet-50 were used with dimensionality reduction through principal component analysis (PCA). A fully connected feedforward neural network was trained on the reduced feature set to predict age. The classification and regression model achieves accuracy results of 93.46% for gender classification and 0.86 for the <i>R</i><sup>2</sup> score for age prediction on the SPR dataset. For visual interpretation, explainable AI (Gradient-weighted Class Activation Mapping) was utilized to visualize and find out which parts of the image are prioritized for classifying gender. The proposed model yields high classification accuracy in gender detection and significant accuracy in age prediction. The model shows competitive accuracy compared to existing methods. Further, the demographic prediction stability of the model was proven on two different ethnic groups, such as the Japanese Society of Radiological Technology (JSRT) and Montgomery (USA) datasets.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-10-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145406549","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DenseNet201SA++: Enhanced Melanoma Recognition in Dermoscopy Images via Soft Attention Guided Feature Learning densenet201sa++:基于软注意引导特征学习的皮肤镜图像黑色素瘤识别
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-10-24 DOI: 10.1002/ima.70236
Shuangshuang Hu, Xiaomei Xu

As the first line of defense in the human immune system, the skin is highly susceptible to environmental toxins. Melanoma, the most lethal type of skin cancer, is characterized by high mortality and a strong tendency for metastasis. It can sometimes originate from pre-existing nevi, particularly dysplastic nevi. Early identification is crucial for improving patient survival rates. However, traditional skin lesion detection faces challenges due to image quality limitations, dataset imperfections, and the complexity of lesion features. This study proposes the DenseNet201SA++ model, which uses image augmentation techniques and the soft attention mechanism to optimize dermoscopy image quality and automatically capture critical features. Experiments on the HAM10000 dataset with 10,015 dermoscopic images, focusing on binary classification (melanoma vs. nevus), show that the DenseNet201SA++ model achieves significant performance gains, with improvements in precision, recall, F1-score, and accuracy of at least 7.2%, 14.7%, 12.7%, and 14.7% compared to baseline networks. The proposed soft attention-guided feature fusion in DenseNet201SA++ addresses feature redundancy in traditional attention mechanisms, achieving superior performance in distinguishing Mel from Nv, while the DenseNet201 backbone shows distinct advantages. Ablation studies confirm the significant role of data augmentation. The integrated DenseNet201SA++ model achieves robust results with precision, recall, F1-score, and accuracy all reaching 0.983, complemented by an AUC of 0.993. These metrics demonstrate the model's exceptional balance between discriminative power and generalization capability, validating the effectiveness of our proposed architecture.

作为人体免疫系统的第一道防线,皮肤极易受到环境毒素的影响。黑色素瘤是最致命的一种皮肤癌,其特点是死亡率高,有很强的转移倾向。它有时可能源于已有的痣,特别是发育不良的痣。早期识别对提高患者存活率至关重要。然而,由于图像质量的限制、数据集的不完善以及病变特征的复杂性,传统的皮肤病变检测面临着挑战。本研究提出了densenet201sa++模型,该模型利用图像增强技术和软注意机制优化皮肤镜图像质量并自动捕获关键特征。在带有10015张皮肤镜图像的HAM10000数据集上进行的实验,重点是二元分类(黑色素瘤与痣),结果表明,与基线网络相比,densenet201sa+ +模型取得了显著的性能提升,精度、召回率、f1评分和准确率分别提高了7.2%、14.7%、12.7%和14.7%。提出的软注意引导特征融合在densenet201sa++中解决了传统注意机制中的特征冗余,在区分Mel和Nv方面取得了优异的性能,而DenseNet201骨干网则表现出明显的优势。消融研究证实了数据增强的重要作用。集成的DenseNet201SA++模型具有鲁棒性,精密度、召回率、f1分数和准确度均达到0.983,AUC为0.993。这些指标证明了模型在判别能力和泛化能力之间的卓越平衡,验证了我们提出的架构的有效性。
{"title":"DenseNet201SA++: Enhanced Melanoma Recognition in Dermoscopy Images via Soft Attention Guided Feature Learning","authors":"Shuangshuang Hu,&nbsp;Xiaomei Xu","doi":"10.1002/ima.70236","DOIUrl":"https://doi.org/10.1002/ima.70236","url":null,"abstract":"<div>\u0000 \u0000 <p>As the first line of defense in the human immune system, the skin is highly susceptible to environmental toxins. Melanoma, the most lethal type of skin cancer, is characterized by high mortality and a strong tendency for metastasis. It can sometimes originate from pre-existing nevi, particularly dysplastic nevi. Early identification is crucial for improving patient survival rates. However, traditional skin lesion detection faces challenges due to image quality limitations, dataset imperfections, and the complexity of lesion features. This study proposes the DenseNet201SA++ model, which uses image augmentation techniques and the soft attention mechanism to optimize dermoscopy image quality and automatically capture critical features. Experiments on the HAM10000 dataset with 10,015 dermoscopic images, focusing on binary classification (melanoma vs. nevus), show that the DenseNet201SA++ model achieves significant performance gains, with improvements in precision, recall, F1-score, and accuracy of at least 7.2%, 14.7%, 12.7%, and 14.7% compared to baseline networks. The proposed soft attention-guided feature fusion in DenseNet201SA++ addresses feature redundancy in traditional attention mechanisms, achieving superior performance in distinguishing Mel from Nv, while the DenseNet201 backbone shows distinct advantages. Ablation studies confirm the significant role of data augmentation. The integrated DenseNet201SA++ model achieves robust results with precision, recall, F1-score, and accuracy all reaching 0.983, complemented by an AUC of 0.993. These metrics demonstrate the model's exceptional balance between discriminative power and generalization capability, validating the effectiveness of our proposed architecture.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145367145","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International Journal of Imaging Systems and Technology
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1