首页 > 最新文献

International Journal of Imaging Systems and Technology最新文献

英文 中文
Adaptive Evidential Fusion of Light–Dark Features With Multi-Scan Mamba for Automated Macular Edema Diagnosis 自适应证据融合的光-暗特征与多扫描曼巴自动诊断黄斑水肿
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-09 DOI: 10.1002/ima.70265
Yiming Zhang, Hongqing Zhu, Tianwei Qian, Tong Hou, Ning Chen, Xun Xu, Bingcang Huang

Macular edema is a retinal disorder that can lead to significant vision loss, underscoring the need for accurate and intelligent automated diagnosis. However, its subtle manifestations in color fundus photography (CFP) pose considerable challenges for conventional deep learning models. In this work, we propose a novel diagnostic framework that integrates Dempster–Shafer (D–S) evidence theory—a principled approach for uncertainty quantification and multi-source information fusion—with the advanced Mamba architecture. The proposed method employs a dual-branch network to selectively enhance and extract discriminative features from both bright and dark regions of fundus images. These features are dynamically aligned and fused via an Adaptive Multi-Branch Feature Synthesis (AMFS) module. To model long-range dependencies and aggregate complementary information from multiple scanning views, we introduce a multi-scan Mamba module, whose outputs are further fused using a principled D–S evidence mechanism. This synergistic integration of mathematical theory and deep learning not only reduces information redundancy but also enables confidence-aware automated decision-making. Extensive experiments on three retinal image datasets—IDRiD, Messidor, and a proprietary clinical cohort—demonstrate that our framework consistently outperforms state-of-the-art methods in terms of accuracy, F1-score, and robustness. These results highlight the promise of combining evidence theory with modern deep learning for challenging medical image analysis tasks.

黄斑水肿是一种视网膜疾病,可导致严重的视力丧失,强调需要准确和智能的自动诊断。然而,它在彩色眼底摄影(CFP)中的细微表现对传统的深度学习模型提出了相当大的挑战。在这项工作中,我们提出了一个新的诊断框架,该框架将Dempster-Shafer (D-S)证据理论(一种用于不确定性量化和多源信息融合的原则性方法)与先进的Mamba架构集成在一起。该方法采用双分支网络对眼底图像的明暗区域进行选择性增强和特征提取。这些特征通过自适应多分支特征合成(AMFS)模块动态对齐和融合。为了对远程依赖关系进行建模,并从多个扫描视图中汇总互补信息,我们引入了一个多扫描Mamba模块,其输出使用原则性D-S证据机制进一步融合。这种数学理论和深度学习的协同集成不仅减少了信息冗余,而且还实现了自信感知的自动决策。在三个视网膜图像数据集(idrid、Messidor和一个专有的临床队列)上进行的广泛实验表明,我们的框架在准确性、f1评分和稳健性方面始终优于最先进的方法。这些结果突出了将证据理论与现代深度学习相结合以解决具有挑战性的医学图像分析任务的前景。
{"title":"Adaptive Evidential Fusion of Light–Dark Features With Multi-Scan Mamba for Automated Macular Edema Diagnosis","authors":"Yiming Zhang,&nbsp;Hongqing Zhu,&nbsp;Tianwei Qian,&nbsp;Tong Hou,&nbsp;Ning Chen,&nbsp;Xun Xu,&nbsp;Bingcang Huang","doi":"10.1002/ima.70265","DOIUrl":"https://doi.org/10.1002/ima.70265","url":null,"abstract":"<div>\u0000 \u0000 <p>Macular edema is a retinal disorder that can lead to significant vision loss, underscoring the need for accurate and intelligent automated diagnosis. However, its subtle manifestations in color fundus photography (CFP) pose considerable challenges for conventional deep learning models. In this work, we propose a novel diagnostic framework that integrates Dempster–Shafer (D–S) evidence theory—a principled approach for uncertainty quantification and multi-source information fusion—with the advanced Mamba architecture. The proposed method employs a dual-branch network to selectively enhance and extract discriminative features from both bright and dark regions of fundus images. These features are dynamically aligned and fused via an Adaptive Multi-Branch Feature Synthesis (AMFS) module. To model long-range dependencies and aggregate complementary information from multiple scanning views, we introduce a multi-scan Mamba module, whose outputs are further fused using a principled D–S evidence mechanism. This synergistic integration of mathematical theory and deep learning not only reduces information redundancy but also enables confidence-aware automated decision-making. Extensive experiments on three retinal image datasets—IDRiD, Messidor, and a proprietary clinical cohort—demonstrate that our framework consistently outperforms state-of-the-art methods in terms of accuracy, F1-score, and robustness. These results highlight the promise of combining evidence theory with modern deep learning for challenging medical image analysis tasks.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"36 1","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145750501","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Task-Specific Electroencephlogram Analysis: A Novel ICA and Dynamic Multi-Stage Clustering Approach for Neural Signal Processing 任务特异性脑电图分析:一种新的ICA和动态多阶段聚类方法用于神经信号处理
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-11-29 DOI: 10.1002/ima.70260
C. Kaviyazhiny, P. Shanthi Bala, R. Priyadharshini, S. Ajeeth

Electroencephalogram (EEG) plays a vital role in identifying the neural activities and human behaviors. EEG is a high-temporal-resolution signal that is contaminated and influenced by various external factors and artifacts. It is very difficult to identify and analyze the EEG signal when it is contaminated by artifacts. The impact of artifacts in EEG signals became an open challenge in EEG signal processing. Even though many techniques and methods are available to remove the artifacts in EEG, analyzing the contaminated EEG signals remains a significant challenge. The signal is polluted by different artifacts like eye blink, muscle activities, environmental interfaces, and so on. Compared with other artifacts and noises, task-specific artifacts such as motor movement and motor imagery have a greater impact on EEG signals due to their direct inference on cognitive signals, and isolating task-specific artifacts from the EEG signal is difficult. To overcome these challenges, a novel hybrid task-specific Dynamic Multistage k-Means Clustering algorithm (TS-DMKC) has been proposed that detects and extracts the processed EEG signal, which will be helpful for all the EEG applications. Initially, the motor movement and motor imagery artifacts are distinguished using the Fast ICA by applying the filtering threshold. Then, a multistage k-means clustering algorithm is employed to isolate the artifacts in different clusters dynamically, and the cluster effectiveness is calculated by using the silhouette score. This proposed novel hybrid algorithm will be useful in various EEG applications like neuroscience, medical diagnosis, brain–computer interface, authentication, brain signal analysis, and gaming. The experimental results demonstrate that the proposed algorithm TS-DMKC achieved a classification accuracy of 94.2% using the Physionet motor movement/imagery dataset, reducing task-dependent variability and offering more robust and reliable EEG systems.

脑电图(EEG)在识别神经活动和人类行为方面起着至关重要的作用。脑电图是一种高时间分辨率的信号,受到各种外界因素和人为因素的污染和影响。当脑电信号受到人为干扰时,对其进行识别和分析是非常困难的。脑电信号中伪影的影响成为脑电信号处理中的一个开放性难题。尽管有许多技术和方法可以去除脑电信号中的伪影,但分析受污染的脑电信号仍然是一个重大挑战。信号会受到不同的干扰,比如眨眼、肌肉活动、环境界面等等。与其他伪像和噪声相比,特定任务伪像(如运动运动和运动意象)对脑电信号的影响更大,因为它们直接对认知信号进行推断,并且很难从脑电信号中分离出特定任务伪像。为了克服这些挑战,本文提出了一种新的针对特定任务的混合动态多阶段k-均值聚类算法(TS-DMKC)来检测和提取处理后的脑电信号,这将有助于脑电信号的各种应用。首先,使用Fast ICA通过应用滤波阈值来区分运动运动和运动图像伪影。然后,采用多阶段k-means聚类算法动态分离不同聚类中的伪影,并利用轮廓分数计算聚类有效性;该算法可应用于神经科学、医学诊断、脑机接口、身份验证、脑信号分析、游戏等领域。实验结果表明,所提出的算法TS-DMKC在使用Physionet运动/图像数据集的情况下,达到了94.2%的分类准确率,减少了任务相关的可变性,提供了更加鲁棒和可靠的脑电系统。
{"title":"Task-Specific Electroencephlogram Analysis: A Novel ICA and Dynamic Multi-Stage Clustering Approach for Neural Signal Processing","authors":"C. Kaviyazhiny,&nbsp;P. Shanthi Bala,&nbsp;R. Priyadharshini,&nbsp;S. Ajeeth","doi":"10.1002/ima.70260","DOIUrl":"https://doi.org/10.1002/ima.70260","url":null,"abstract":"<div>\u0000 \u0000 <p>Electroencephalogram (EEG) plays a vital role in identifying the neural activities and human behaviors. EEG is a high-temporal-resolution signal that is contaminated and influenced by various external factors and artifacts. It is very difficult to identify and analyze the EEG signal when it is contaminated by artifacts. The impact of artifacts in EEG signals became an open challenge in EEG signal processing. Even though many techniques and methods are available to remove the artifacts in EEG, analyzing the contaminated EEG signals remains a significant challenge. The signal is polluted by different artifacts like eye blink, muscle activities, environmental interfaces, and so on. Compared with other artifacts and noises, task-specific artifacts such as motor movement and motor imagery have a greater impact on EEG signals due to their direct inference on cognitive signals, and isolating task-specific artifacts from the EEG signal is difficult. To overcome these challenges, a novel hybrid task-specific Dynamic Multistage <i>k</i>-Means Clustering algorithm (TS-DMKC) has been proposed that detects and extracts the processed EEG signal, which will be helpful for all the EEG applications. Initially, the motor movement and motor imagery artifacts are distinguished using the Fast ICA by applying the filtering threshold. Then, a multistage <i>k</i>-means clustering algorithm is employed to isolate the artifacts in different clusters dynamically, and the cluster effectiveness is calculated by using the silhouette score. This proposed novel hybrid algorithm will be useful in various EEG applications like neuroscience, medical diagnosis, brain–computer interface, authentication, brain signal analysis, and gaming. The experimental results demonstrate that the proposed algorithm TS-DMKC achieved a classification accuracy of 94.2% using the Physionet motor movement/imagery dataset, reducing task-dependent variability and offering more robust and reliable EEG systems.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"36 1","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-11-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145619201","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HPPE-Unet: A Boundary-Enhanced U-Shaped Network With Hybrid Attention and Multi-Scale Feature Fusion for Small-Sample Medical Image Segmentation hpe - unet:一种混合关注和多尺度特征融合的边界增强u型网络用于小样本医学图像分割
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-11-20 DOI: 10.1002/ima.70256
Changhao Sun, Tong Liu, Mujun Zang, Shusen Zhou, Chanjuan Liu, Qingjun Wang

Accurate diagnosis of diseases relies on the subjective experience of physicians. Automatic image segmentation technology can assist doctors in quickly locating lesion areas, thereby improving diagnostic efficiency. In the field of medical image segmentation, the current mainstream models are mostly variants of Swin-Unet, which perform well on large-sample datasets. However, due to the window constraints of the Swin-Transformer architecture, these models often face performance limitations when processing small-sample datasets. In contrast, convolutional-based Unet variant models exhibit better segmentation performance under small-sample conditions but demonstrate poorer segmentation results for medical images with blurred boundaries. To tackle these challenges, this paper introduces a novel boundary information-enhanced architecture, the Hybrid-Parallel-Attention and Pyramid-Edge-Extraction Enhanced UNet (HPPE-Unet). Based on an encoder-decoder architecture of convolutional neural networks, the model incorporates a multi-scale feature extraction strategy and an optimized skip-connection mechanism. It integrates two key modules: the Pyramid Edge Extraction (PEE) module and the Hybrid Parallel Attention (HP) module. The PEE module is applied at each stage of the encoder and decoder, significantly enhancing the model's ability to capture subtle boundary structures through multi-scale feature fusion and boundary reinforcement mechanisms. The HP employs a residual parallel structure combining spatial (SA) and channel attention (CA) blocks, effectively bridging the semantic gap between features in the encoding and decoding stages and addressing the issue of information loss in skip connections between the encoder and decoder. The proposed method was evaluated on an aortic dissection dataset provided by a tertiary hospital. The results show that the method achieved a Dice similarity coefficient (DSC) of 97.65% and a Mean intersection over union (Miou) of 97.92% in the segmentation task. On the public Automated Cardiac Diagnostic Challenge (ACDC) dataset, the DSC reached 90.39%. These results demonstrate that the proposed method holds significant practical value for clinical disease diagnosis.

疾病的准确诊断依赖于医师的主观经验。自动图像分割技术可以帮助医生快速定位病变区域,从而提高诊断效率。在医学图像分割领域,目前主流的模型大多是swan - unet的变体,在大样本数据集上表现良好。然而,由于swing - transformer架构的窗口约束,这些模型在处理小样本数据集时经常面临性能限制。相比之下,基于卷积的Unet变体模型在小样本条件下表现出更好的分割性能,但在边界模糊的医学图像中表现出较差的分割效果。为了解决这些问题,本文提出了一种新的边界信息增强架构——混合并行注意和金字塔边缘提取增强UNet (hpe - UNet)。该模型基于卷积神经网络的编码器-解码器架构,结合了多尺度特征提取策略和优化的跳跃连接机制。它集成了两个关键模块:金字塔边缘提取(PEE)模块和混合并行注意(HP)模块。在编码器和解码器的每个阶段都应用了PEE模块,通过多尺度特征融合和边界强化机制,显著增强了模型捕捉细微边界结构的能力。HP采用结合空间(SA)和信道注意(CA)块的残差并行结构,有效地弥合了编码和解码阶段特征之间的语义差距,并解决了编码器和解码器之间的跳过连接中的信息丢失问题。该方法在一家三级医院提供的主动脉夹层数据集上进行了评估。结果表明,该方法在分割任务中获得了97.65%的Dice similarity coefficient (DSC)和97.92%的Mean intersection over union (Miou)。在公共自动心脏诊断挑战(ACDC)数据集上,DSC达到90.39%。结果表明,该方法对临床疾病诊断具有重要的实用价值。
{"title":"HPPE-Unet: A Boundary-Enhanced U-Shaped Network With Hybrid Attention and Multi-Scale Feature Fusion for Small-Sample Medical Image Segmentation","authors":"Changhao Sun,&nbsp;Tong Liu,&nbsp;Mujun Zang,&nbsp;Shusen Zhou,&nbsp;Chanjuan Liu,&nbsp;Qingjun Wang","doi":"10.1002/ima.70256","DOIUrl":"https://doi.org/10.1002/ima.70256","url":null,"abstract":"<div>\u0000 \u0000 <p>Accurate diagnosis of diseases relies on the subjective experience of physicians. Automatic image segmentation technology can assist doctors in quickly locating lesion areas, thereby improving diagnostic efficiency. In the field of medical image segmentation, the current mainstream models are mostly variants of Swin-Unet, which perform well on large-sample datasets. However, due to the window constraints of the Swin-Transformer architecture, these models often face performance limitations when processing small-sample datasets. In contrast, convolutional-based Unet variant models exhibit better segmentation performance under small-sample conditions but demonstrate poorer segmentation results for medical images with blurred boundaries. To tackle these challenges, this paper introduces a novel boundary information-enhanced architecture, the Hybrid-Parallel-Attention and Pyramid-Edge-Extraction Enhanced UNet (HPPE-Unet). Based on an encoder-decoder architecture of convolutional neural networks, the model incorporates a multi-scale feature extraction strategy and an optimized skip-connection mechanism. It integrates two key modules: the Pyramid Edge Extraction (PEE) module and the Hybrid Parallel Attention (HP) module. The PEE module is applied at each stage of the encoder and decoder, significantly enhancing the model's ability to capture subtle boundary structures through multi-scale feature fusion and boundary reinforcement mechanisms. The HP employs a residual parallel structure combining spatial (SA) and channel attention (CA) blocks, effectively bridging the semantic gap between features in the encoding and decoding stages and addressing the issue of information loss in skip connections between the encoder and decoder. The proposed method was evaluated on an aortic dissection dataset provided by a tertiary hospital. The results show that the method achieved a Dice similarity coefficient (DSC) of 97.65% and a Mean intersection over union (Miou) of 97.92% in the segmentation task. On the public Automated Cardiac Diagnostic Challenge (ACDC) dataset, the DSC reached 90.39%. These results demonstrate that the proposed method holds significant practical value for clinical disease diagnosis.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145572471","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Uncertainty-Aware Graph Self-Training for Autism Spectrum Disorder Classification in Multiple Centers 自闭症谱系障碍多中心分类的不确定性感知图自训练
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-11-20 DOI: 10.1002/ima.70250
Qianhui Yang, Jun Wang, Jiale Dun, Juncheng Li, Jun Shi

Classical self-training methods for graph convolutional networks (GCNs) assume that both labeled and unlabeled data follow the identical distribution. However, these works do not work well when they are used in medical applications such as classifying autism spectrum disorder (ASD) in multiple centers, where the unlabeled samples from different imaging centers have varying distribution shifts from the labeled samples. To this end, we propose uncertainty-aware graph self-training (UA-GST) by extending graph self-training to a new situation, in which the labeled data come from one imaging center and the unlabeled data come from several other imaging centers. Specifically, an uncertainty-aware mechanism is proposed to select unlabeled centers and adversarial domain adaptation is introduced into graph self-training to reduce domain shift between centers. With the progression of self-training, more pseudo-labeled test samples are gradually included in the training set, and a final model is finally trained on all the labeled and pseudo-labeled samples. Considering the over-confidence issue of the classifier, an evidential GCN is further proposed to estimate the uncertainty of the pseudo-labels using Dempster–Shafer (D–S) evidence theory. It is evaluated on the Autism Brain Imaging Data Exchange (ABIDE). Experimental results verified the effectiveness of the proposed method in classifying ASD in multiple centers, outperforming existing state-of-the-art methods.

经典的图卷积网络(GCNs)自训练方法假设标记数据和未标记数据遵循相同的分布。然而,当这些工作被用于医学应用时,如在多个中心对自闭症谱系障碍(ASD)进行分类时,这些工作就不能很好地发挥作用,因为来自不同成像中心的未标记样本与标记样本的分布变化不同。为此,我们提出了不确定性感知图自训练(UA-GST),将图自训练扩展到一种新的情况,即标记数据来自一个成像中心,未标记数据来自其他几个成像中心。具体而言,提出了一种不确定性感知机制来选择未标记的中心,并在图自训练中引入对抗域适应来减少中心之间的域移动。随着自我训练的进行,越来越多的伪标记测试样本逐渐被纳入训练集,最后在所有标记和伪标记样本上训练出最终的模型。考虑到分类器的过度置信度问题,利用Dempster-Shafer (D-S)证据理论,进一步提出了一种估计伪标签不确定性的证据GCN。在自闭症脑成像数据交换(Autism Brain Imaging Data Exchange,简称ABIDE)上进行评估。实验结果验证了该方法在多个中心的ASD分类中的有效性,优于现有的最先进的方法。
{"title":"Uncertainty-Aware Graph Self-Training for Autism Spectrum Disorder Classification in Multiple Centers","authors":"Qianhui Yang,&nbsp;Jun Wang,&nbsp;Jiale Dun,&nbsp;Juncheng Li,&nbsp;Jun Shi","doi":"10.1002/ima.70250","DOIUrl":"https://doi.org/10.1002/ima.70250","url":null,"abstract":"<div>\u0000 \u0000 <p>Classical self-training methods for graph convolutional networks (GCNs) assume that both labeled and unlabeled data follow the identical distribution. However, these works do not work well when they are used in medical applications such as classifying autism spectrum disorder (ASD) in multiple centers, where the unlabeled samples from different imaging centers have varying distribution shifts from the labeled samples. To this end, we propose uncertainty-aware graph self-training (UA-GST) by extending graph self-training to a new situation, in which the labeled data come from one imaging center and the unlabeled data come from several other imaging centers. Specifically, an uncertainty-aware mechanism is proposed to select unlabeled centers and adversarial domain adaptation is introduced into graph self-training to reduce domain shift between centers. With the progression of self-training, more pseudo-labeled test samples are gradually included in the training set, and a final model is finally trained on all the labeled and pseudo-labeled samples. Considering the over-confidence issue of the classifier, an evidential GCN is further proposed to estimate the uncertainty of the pseudo-labels using Dempster–Shafer (D–S) evidence theory. It is evaluated on the Autism Brain Imaging Data Exchange (ABIDE). Experimental results verified the effectiveness of the proposed method in classifying ASD in multiple centers, outperforming existing state-of-the-art methods.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145572607","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Self-Supervised DRL With Twin Critics: A Novel Framework for Glioblastoma Survival Prognostication 双重批评的自我监督DRL:胶质母细胞瘤生存预测的新框架
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-11-20 DOI: 10.1002/ima.70258
M. Renugadevi, K. Ramkumar, N. Raju, K. Adalarasu, Surya Prasath, K. Narasimhan

The accurate survival prediction in glioblastoma patients remains a major challenge, largely owing to the considerable variability among brain tumor subtypes and their clinical profiles. This research introduces an innovative SimCLR-Twin Critic DDPG framework that integrates self-supervised representation learning with deep reinforcement learning for enhanced prognostic prediction. Tumor subregion segmentations were independently obtained using four advanced models namely nnU-Net, VNet, SwinUNETR, and UNet. Based on the performance of these segmentation models, deep features were extracted from the nnU-Net segmented tumor subregions and combined with handcrafted radiomics features. The combined features were subjected to the feature selection process using BorutaShap, followed by self-supervised pretraining through SimCLR to improve feature representation and generalization. The optimized features were then fed into a Twin Critic DDPG agent designed for regression-based survival time predictions. The proposed method outperformed existing techniques on the BraTS 2020 dataset, achieving the lowest MSE of 31 062.36 and the highest C-index of 0.87 in overall survival prediction. To further confirm robustness and generalizability, the framework was externally validated on the BraTS 2019 dataset, where it achieved a comparable MSE of 31 156.56 and a C-index of 0.86 using fused features. Additionally, LIME-based local interpretability provided clinically relevant explanations for individual predictions, thereby enhancing trust in the AI-driven system. This study highlights the SimCLR-Twin Critic DDPG framework as a robust and interpretable solution for accurate survival prediction in glioblastoma patients.

胶质母细胞瘤患者的准确生存预测仍然是一个主要的挑战,主要是由于脑肿瘤亚型及其临床特征的相当大的变异性。本研究引入了一个创新的SimCLR-Twin Critic DDPG框架,该框架集成了自监督表示学习和深度强化学习,以增强预后预测。采用nnU-Net、VNet、SwinUNETR和UNet四种高级模型独立获得肿瘤亚区分割。基于这些分割模型的性能,从nnU-Net分割的肿瘤亚区域中提取深度特征,并与手工制作的放射组学特征相结合。组合后的特征采用BorutaShap进行特征选择,然后采用SimCLR进行自监督预训练,提高特征表征和泛化能力。然后将优化的特征输入Twin Critic DDPG代理,该代理设计用于基于回归的生存时间预测。该方法在BraTS 2020数据集上优于现有技术,在总体生存预测中实现了最低的MSE为31 062.36,最高的c指数为0.87。为了进一步确认稳健性和泛化性,该框架在BraTS 2019数据集上进行了外部验证,其中使用融合特征实现了可比较的MSE为31 156.56和c指数为0.86。此外,基于lime的局部可解释性为个体预测提供了临床相关的解释,从而增强了对人工智能驱动系统的信任。该研究强调了SimCLR-Twin Critic DDPG框架作为胶质母细胞瘤患者准确生存预测的可靠且可解释的解决方案。
{"title":"Self-Supervised DRL With Twin Critics: A Novel Framework for Glioblastoma Survival Prognostication","authors":"M. Renugadevi,&nbsp;K. Ramkumar,&nbsp;N. Raju,&nbsp;K. Adalarasu,&nbsp;Surya Prasath,&nbsp;K. Narasimhan","doi":"10.1002/ima.70258","DOIUrl":"https://doi.org/10.1002/ima.70258","url":null,"abstract":"<div>\u0000 \u0000 <p>The accurate survival prediction in glioblastoma patients remains a major challenge, largely owing to the considerable variability among brain tumor subtypes and their clinical profiles. This research introduces an innovative SimCLR-Twin Critic DDPG framework that integrates self-supervised representation learning with deep reinforcement learning for enhanced prognostic prediction. Tumor subregion segmentations were independently obtained using four advanced models namely nnU-Net, VNet, SwinUNETR, and UNet. Based on the performance of these segmentation models, deep features were extracted from the nnU-Net segmented tumor subregions and combined with handcrafted radiomics features. The combined features were subjected to the feature selection process using BorutaShap, followed by self-supervised pretraining through SimCLR to improve feature representation and generalization. The optimized features were then fed into a Twin Critic DDPG agent designed for regression-based survival time predictions. The proposed method outperformed existing techniques on the BraTS 2020 dataset, achieving the lowest MSE of 31 062.36 and the highest C-index of 0.87 in overall survival prediction. To further confirm robustness and generalizability, the framework was externally validated on the BraTS 2019 dataset, where it achieved a comparable MSE of 31 156.56 and a C-index of 0.86 using fused features. Additionally, LIME-based local interpretability provided clinically relevant explanations for individual predictions, thereby enhancing trust in the AI-driven system. This study highlights the SimCLR-Twin Critic DDPG framework as a robust and interpretable solution for accurate survival prediction in glioblastoma patients.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145581098","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-Objectives Optimization (MOO)-Based NAS Approach 基于多目标优化(MOO)的NAS方法
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-11-15 DOI: 10.1002/ima.70240
Abass Sana, Kaoutar Senhaji, Amir Nakib

This research dives into creating optimized neural network architectures tailored for image classification and medical image segmentation, with a particular emphasis on analyzing lung nodules using the LIDC-IDRI dataset. We introduce a hybrid approach to neural architecture search (NAS) that fuses convolutional neural networks (CNNs), transformer-based models, and custom-built graph architectures. Our evolutionary strategies employ genetic operations like crossover and mutation to gradually enhance these architectures, while the NSGA-III algorithm helps us navigate the tricky balance of multiple conflicting objectives. The goal of our method is to boost performance metrics such as the F1 score and classification accuracy, all while keeping an eye on minimizing computational complexity, which we measure in terms of FLOPS, parameter count, and inference time. Our experiments demonstrate competitive performance on the CIFAR-10 and CIFAR-100 datasets, with promising results on the LIDC-IDRI segmentation task. This work aims to push the boundaries of automated model design, contributing to more efficient and effective deep learning architectures in both general and medical imaging contexts.

本研究深入研究了为图像分类和医学图像分割量身定制的优化神经网络架构,特别强调了使用LIDC-IDRI数据集分析肺结节。我们介绍了一种神经架构搜索(NAS)的混合方法,它融合了卷积神经网络(cnn)、基于变压器的模型和定制的图架构。我们的进化策略采用交叉和突变等遗传操作来逐步增强这些结构,而NSGA-III算法则帮助我们在多个相互冲突的目标之间找到棘手的平衡。我们方法的目标是提高F1分数和分类准确性等性能指标,同时关注最小化计算复杂性,我们用FLOPS、参数计数和推理时间来衡量计算复杂性。我们的实验在CIFAR-10和CIFAR-100数据集上展示了具有竞争力的性能,在LIDC-IDRI分割任务上取得了令人鼓舞的结果。这项工作旨在推动自动化模型设计的边界,为通用和医学成像环境中更高效和有效的深度学习架构做出贡献。
{"title":"Multi-Objectives Optimization (MOO)-Based NAS Approach","authors":"Abass Sana,&nbsp;Kaoutar Senhaji,&nbsp;Amir Nakib","doi":"10.1002/ima.70240","DOIUrl":"https://doi.org/10.1002/ima.70240","url":null,"abstract":"<div>\u0000 \u0000 <p>This research dives into creating optimized neural network architectures tailored for image classification and medical image segmentation, with a particular emphasis on analyzing lung nodules using the LIDC-IDRI dataset. We introduce a hybrid approach to neural architecture search (NAS) that fuses convolutional neural networks (CNNs), transformer-based models, and custom-built graph architectures. Our evolutionary strategies employ genetic operations like crossover and mutation to gradually enhance these architectures, while the NSGA-III algorithm helps us navigate the tricky balance of multiple conflicting objectives. The goal of our method is to boost performance metrics such as the F1 score and classification accuracy, all while keeping an eye on minimizing computational complexity, which we measure in terms of FLOPS, parameter count, and inference time. Our experiments demonstrate competitive performance on the CIFAR-10 and CIFAR-100 datasets, with promising results on the LIDC-IDRI segmentation task. This work aims to push the boundaries of automated model design, contributing to more efficient and effective deep learning architectures in both general and medical imaging contexts.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145530079","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improved Brain Tumor Segmentation With SegFormer: A Transformer-Based Architecture for Cross-Dataset Generalization 用SegFormer改进脑肿瘤分割:一种基于变换的跨数据集泛化架构
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-11-15 DOI: 10.1002/ima.70257
Alaa Altheneyan, Mohammed Zakariah, Abeer Alnuaim, Aseel Alhadlaq

This research work aims to improve brain tumor segmentation using a transformer architecture known as SegFormer. It is essential to segment brain tumors correctly for surgical planning and tumor progression assessment. Many existing segmentation techniques have problems in visualizing the shapes and structures of a brain tumor; they include issues of overfitting and underfitting. The SegFormer model is developed to change transformers employed in natural language processing to overcome such difficulties. Moreover, the SegFormer model improves accuracy by adequately integrating the image patches, hence improving the structures of the tumors. Accurate segmentation of brain tumors is crucial in clinical neuro-oncology, diagnosis, treatment planning, and follow-up care. Here, we present a novel deep-learning-based SegFormer framework for brain tumor segmentation that includes hierarchical patch embedding and transformer-based self-attention to consider both the global and local tumor features simultaneously. In comparison to CNN-based and hybrid models, our model is much more accurate in segmentation and has a lower computational cost. Our initial attempt to apply the proposed model on the BraTS2020 benchmark dataset, with a Dice score of 0.9425, exceeded the state-of-the-art models, including RMTF-Net, HMNet, and CBAM TransUnet. To additionally confirm generalization, we applied the model to the BraTS2021 dataset (Dice: 0.92123), the dataset of ischemic stroke lesion segmentation in ISLES (Dice: 0.874), and the TCGA dataset (Dice: 0.860). These findings indicate the soundness and applicability of our structure in a variety of clinical imaging conditions. We make the following contributions: (1) we combine transformer-based self-attention and lightweight decoding to segment the image accurately, (2) we achieve higher results on BraTS2020 than state-of-the-art models, and (3) we also validate the obtained results on a large set of datasets, which confirms the good generalization. These results establish SegFormer as a promising foundation for the next generation of brain tumor and lesion segmentation tasks in clinical imaging.

这项研究工作旨在使用一种称为SegFormer的变压器架构来改善脑肿瘤的分割。正确分割脑肿瘤对于手术计划和肿瘤进展评估至关重要。许多现有的分割技术在可视化脑肿瘤的形状和结构方面存在问题;它们包括过拟合和欠拟合的问题。SegFormer模型是为了改变自然语言处理中使用的变压器来克服这些困难而开发的。此外,SegFormer模型通过充分整合图像斑块来提高准确性,从而改善肿瘤的结构。脑肿瘤的准确分割对临床神经肿瘤学、诊断、治疗计划和随访至关重要。在这里,我们提出了一种新的基于深度学习的脑肿瘤分割SegFormer框架,该框架包括分层补丁嵌入和基于变压器的自关注,同时考虑全局和局部肿瘤特征。与基于cnn的模型和混合模型相比,我们的模型在分割上更加准确,并且计算成本更低。我们最初尝试在BraTS2020基准数据集上应用所提出的模型,其Dice得分为0.9425,超过了最先进的模型,包括RMTF-Net, HMNet和CBAM TransUnet。为了进一步确认泛化,我们将该模型应用于BraTS2021数据集(Dice: 0.92123)、ISLES中缺血性脑卒中病变分割数据集(Dice: 0.874)和TCGA数据集(Dice: 0.860)。这些发现表明我们的结构在各种临床成像条件下的可靠性和适用性。我们做出了以下贡献:(1)我们结合了基于变压器的自关注和轻量级解码来准确分割图像;(2)我们在BraTS2020上获得了比最先进的模型更高的结果;(3)我们还在大量数据集上验证了所获得的结果,这证实了良好的一般化。这些结果奠定了SegFormer作为下一代脑肿瘤和病变分割任务在临床成像中的前景基础。
{"title":"Improved Brain Tumor Segmentation With SegFormer: A Transformer-Based Architecture for Cross-Dataset Generalization","authors":"Alaa Altheneyan,&nbsp;Mohammed Zakariah,&nbsp;Abeer Alnuaim,&nbsp;Aseel Alhadlaq","doi":"10.1002/ima.70257","DOIUrl":"https://doi.org/10.1002/ima.70257","url":null,"abstract":"<div>\u0000 \u0000 <p>This research work aims to improve brain tumor segmentation using a transformer architecture known as SegFormer. It is essential to segment brain tumors correctly for surgical planning and tumor progression assessment. Many existing segmentation techniques have problems in visualizing the shapes and structures of a brain tumor; they include issues of overfitting and underfitting. The SegFormer model is developed to change transformers employed in natural language processing to overcome such difficulties. Moreover, the SegFormer model improves accuracy by adequately integrating the image patches, hence improving the structures of the tumors. Accurate segmentation of brain tumors is crucial in clinical neuro-oncology, diagnosis, treatment planning, and follow-up care. Here, we present a novel deep-learning-based SegFormer framework for brain tumor segmentation that includes hierarchical patch embedding and transformer-based self-attention to consider both the global and local tumor features simultaneously. In comparison to CNN-based and hybrid models, our model is much more accurate in segmentation and has a lower computational cost. Our initial attempt to apply the proposed model on the BraTS2020 benchmark dataset, with a Dice score of 0.9425, exceeded the state-of-the-art models, including RMTF-Net, HMNet, and CBAM TransUnet. To additionally confirm generalization, we applied the model to the BraTS2021 dataset (Dice: 0.92123), the dataset of ischemic stroke lesion segmentation in ISLES (Dice: 0.874), and the TCGA dataset (Dice: 0.860). These findings indicate the soundness and applicability of our structure in a variety of clinical imaging conditions. We make the following contributions: (1) we combine transformer-based self-attention and lightweight decoding to segment the image accurately, (2) we achieve higher results on BraTS2020 than state-of-the-art models, and (3) we also validate the obtained results on a large set of datasets, which confirms the good generalization. These results establish SegFormer as a promising foundation for the next generation of brain tumor and lesion segmentation tasks in clinical imaging.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145530187","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SkinSegNet: An Advanced Encoder–Decoder for Skin Lesion Segmentation Using Channel-Aware and Spatial Cross-Scale Attention SkinSegNet:一种先进的编码器-解码器,用于使用通道感知和空间跨尺度注意的皮肤病变分割
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-11-15 DOI: 10.1002/ima.70255
P. Nithin, G. Muthulakshmi

To achieve precise and efficient skin cancer segmentation, an innovative SkinSegNet architecture is proposed. Inspired by U-Net, SkinSegNet uses an encoder–decoder architecture incorporating advanced feature extraction and attention mechanisms. The encoder utilizes convolutional blocks and pooling attention (PA) layers to downsample feature maps and focus on significant regions. At the bottleneck, the proposed feature aggregation block integrates channel-aware multihead attention (CAMA) and Adaptive Spatial Cross-Scale Attention (ASCA) modules. These modules let the model capture channel relationships and complex spatial dependencies for accurate segmentation. The decoder reconstructs the segmentation mask through continuous upsampling and skips connections, ensuring fine-grained spatial detail is preserved. Experiments are carried out on benchmark datasets namely ISIC 2016, ISIC 2017, and ISIC 2018 demonstrating SkinSegNet's superior performance in skin lesion segmentation, achieving state-of-the-art accuracy of 95.12%, 94.01%, and 95.04%, respectively. Furthermore, cross-dataset experiments show that SkinSegNet performs well on ISIC datasets, demonstrating its strong generalization ability.

为了实现精确高效的皮肤癌分割,提出了一种创新的SkinSegNet架构。受U-Net的启发,SkinSegNet使用了一种编码器-解码器架构,结合了先进的特征提取和注意机制。编码器利用卷积块和集中注意力(PA)层对特征图进行下采样并聚焦于重要区域。在瓶颈处,所提出的特征聚合块集成了信道感知多头注意(CAMA)和自适应空间跨尺度注意(ASCA)模块。这些模块使模型能够捕获通道关系和复杂的空间依赖关系,从而实现准确的分割。解码器通过连续上采样重建分割掩码,并跳过连接,确保保留细粒度的空间细节。在ISIC 2016、ISIC 2017和ISIC 2018的基准数据集上进行了实验,证明了SkinSegNet在皮肤病变分割方面的卓越性能,分别达到了95.12%、94.01%和95.04%的准确率。此外,跨数据集实验表明,SkinSegNet在ISIC数据集上表现良好,展示了其强大的泛化能力。
{"title":"SkinSegNet: An Advanced Encoder–Decoder for Skin Lesion Segmentation Using Channel-Aware and Spatial Cross-Scale Attention","authors":"P. Nithin,&nbsp;G. Muthulakshmi","doi":"10.1002/ima.70255","DOIUrl":"https://doi.org/10.1002/ima.70255","url":null,"abstract":"<div>\u0000 \u0000 <p>To achieve precise and efficient skin cancer segmentation, an innovative SkinSegNet architecture is proposed. Inspired by U-Net, SkinSegNet uses an encoder–decoder architecture incorporating advanced feature extraction and attention mechanisms. The encoder utilizes convolutional blocks and pooling attention (PA) layers to downsample feature maps and focus on significant regions. At the bottleneck, the proposed feature aggregation block integrates channel-aware multihead attention (CAMA) and Adaptive Spatial Cross-Scale Attention (ASCA) modules. These modules let the model capture channel relationships and complex spatial dependencies for accurate segmentation. The decoder reconstructs the segmentation mask through continuous upsampling and skips connections, ensuring fine-grained spatial detail is preserved. Experiments are carried out on benchmark datasets namely ISIC 2016, ISIC 2017, and ISIC 2018 demonstrating SkinSegNet's superior performance in skin lesion segmentation, achieving state-of-the-art accuracy of 95.12%, 94.01%, and 95.04%, respectively. Furthermore, cross-dataset experiments show that SkinSegNet performs well on ISIC datasets, demonstrating its strong generalization ability.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145530188","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
High-Resolution Visualization of the Human Semicircular Canals Using Optical Coherence Tomography 利用光学相干断层成像对人体半规管进行高分辨率可视化
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-11-15 DOI: 10.1002/ima.70259
Elke Loos, Lore Kerkhofs, Raymond Van de Berg, Christian Desloovere, Tristan Putzeys, Nicolas Verhaert

The semicircular canals are essential sensors of the vestibular organ. Current clinical in vivo imaging techniques lack the resolution needed to visualize critical vestibular microstructures, such as the membranous labyrinth and crista ampullaris. Optical coherence tomography (OCT) has emerged as a promising imaging technique, offering high-resolution, cross-sectional visualization of inner ear structures, particularly within the cochlea. In this study, we assess the feasibility of OCT for imaging the vestibular system in human temporal bones and present the first comprehensive OCT atlas of the human vestibular organ confirmed on micro-CT imaging. This atlas may support future research by providing detailed anatomical references. This technique holds promise for future diagnostic applications and for improving the accuracy of surgical procedures, such as vestibular implant surgery.

半规管是前庭器官的重要感受器。目前的临床活体成像技术缺乏可视化关键前庭显微结构所需的分辨率,如膜迷路和壶腹嵴。光学相干断层扫描(OCT)已成为一种有前途的成像技术,提供内耳结构,特别是耳蜗内的高分辨率,横断面可视化。在这项研究中,我们评估了OCT成像人类颞骨前庭系统的可行性,并提出了第一个经显微ct成像证实的人类前庭器官的全面OCT图谱。该图谱可以通过提供详细的解剖学参考来支持未来的研究。这项技术为未来的诊断应用和提高外科手术的准确性,如前庭植入手术,带来了希望。
{"title":"High-Resolution Visualization of the Human Semicircular Canals Using Optical Coherence Tomography","authors":"Elke Loos,&nbsp;Lore Kerkhofs,&nbsp;Raymond Van de Berg,&nbsp;Christian Desloovere,&nbsp;Tristan Putzeys,&nbsp;Nicolas Verhaert","doi":"10.1002/ima.70259","DOIUrl":"https://doi.org/10.1002/ima.70259","url":null,"abstract":"<p>The semicircular canals are essential sensors of the vestibular organ. Current clinical in vivo imaging techniques lack the resolution needed to visualize critical vestibular microstructures, such as the membranous labyrinth and crista ampullaris. Optical coherence tomography (OCT) has emerged as a promising imaging technique, offering high-resolution, cross-sectional visualization of inner ear structures, particularly within the cochlea. In this study, we assess the feasibility of OCT for imaging the vestibular system in human temporal bones and present the first comprehensive OCT atlas of the human vestibular organ confirmed on micro-CT imaging. This atlas may support future research by providing detailed anatomical references. This technique holds promise for future diagnostic applications and for improving the accuracy of surgical procedures, such as vestibular implant surgery.</p>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ima.70259","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145530189","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimized Hybrid CNN Framework for Enhanced Tumor Classification in Breast Cancer Diagnosis 优化混合CNN框架增强乳腺癌诊断中的肿瘤分类
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-11-11 DOI: 10.1002/ima.70252
Shumaila Batool, Saima Zainab, Muhammad Usman, Juhua Pu

Convolutional neural networks (CNNs) have augmented conventional approaches in medical imaging by improving tumor detection and classification efficacy. To enable oncologists to diagnose abnormalities promptly, this research proposes an innovative classification framework for breast cancer diagnosis. It integrates an improved optimization method with a hybridized CNN architecture. In this article, a custom CNN, feed-forward and backpropagation have been implemented. The scaled conjugate algorithm is employed in the feed-forward paradigm, yielding a formidable accuracy of 99.1%. On the other hand, backpropagation implements stochastic gradient descent and exhibits a remarkable accuracy rate of 97.3%. Additionally, by integrating the grey wolf optimization (GWO) algorithm with the Backpropagation Neural Network (BPNN), model performance is enhanced by optimizing parameters and accuracy to 100%. Furthermore, the custom CNN achieves an incredible 98% accuracy by utilizing the Adam optimizer in conjunction with the ReduceLROnPlateau approach. Statistical analysis utilizing Analysis of Variance (ANOVA) and Honestly Significant Difference (HSD) tests has demonstrated that the suggested hybrid model improves detection accuracy and reliability. These results highlight the adaptability and effectiveness of various optimization techniques in enhancing the performance of neural network models on a range of demanding tasks related to machine learning and pattern recognition.

卷积神经网络(cnn)通过提高肿瘤检测和分类效率,增强了传统医学成像方法。为了使肿瘤学家能够及时诊断异常,本研究提出了一种创新的乳腺癌诊断分类框架。它将一种改进的优化方法与一种杂交CNN架构相结合。在本文中,实现了自定义CNN,前馈和反向传播。前馈模式采用缩放共轭算法,精度高达99.1%。另一方面,反向传播实现了随机梯度下降,准确率达到97.3%。此外,将灰狼优化(GWO)算法与反向传播神经网络(BPNN)相结合,优化参数,提高模型性能,准确率达到100%。此外,通过使用Adam优化器和ReduceLROnPlateau方法,自定义CNN达到了令人难以置信的98%的准确率。利用方差分析(ANOVA)和诚实显著差异(HSD)检验的统计分析表明,建议的混合模型提高了检测的准确性和可靠性。这些结果突出了各种优化技术在提高神经网络模型在与机器学习和模式识别相关的一系列苛刻任务上的性能方面的适应性和有效性。
{"title":"Optimized Hybrid CNN Framework for Enhanced Tumor Classification in Breast Cancer Diagnosis","authors":"Shumaila Batool,&nbsp;Saima Zainab,&nbsp;Muhammad Usman,&nbsp;Juhua Pu","doi":"10.1002/ima.70252","DOIUrl":"https://doi.org/10.1002/ima.70252","url":null,"abstract":"<div>\u0000 \u0000 <p>Convolutional neural networks (CNNs) have augmented conventional approaches in medical imaging by improving tumor detection and classification efficacy. To enable oncologists to diagnose abnormalities promptly, this research proposes an innovative classification framework for breast cancer diagnosis. It integrates an improved optimization method with a hybridized CNN architecture. In this article, a custom CNN, feed-forward and backpropagation have been implemented. The scaled conjugate algorithm is employed in the feed-forward paradigm, yielding a formidable accuracy of 99.1%. On the other hand, backpropagation implements stochastic gradient descent and exhibits a remarkable accuracy rate of 97.3%. Additionally, by integrating the grey wolf optimization (GWO) algorithm with the Backpropagation Neural Network (BPNN), model performance is enhanced by optimizing parameters and accuracy to 100%. Furthermore, the custom CNN achieves an incredible 98% accuracy by utilizing the Adam optimizer in conjunction with the ReduceLROnPlateau approach. Statistical analysis utilizing Analysis of Variance (ANOVA) and Honestly Significant Difference (HSD) tests has demonstrated that the suggested hybrid model improves detection accuracy and reliability. These results highlight the adaptability and effectiveness of various optimization techniques in enhancing the performance of neural network models on a range of demanding tasks related to machine learning and pattern recognition.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 6","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145529987","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International Journal of Imaging Systems and Technology
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1