首页 > 最新文献

International Journal of Imaging Systems and Technology最新文献

英文 中文
Research on Multi-Objective Optimization of Medical Image Segmentation Based on Frequency Domain Decoupling and Dual Attention Mechanism 基于频域解耦和双注意机制的医学图像分割多目标优化研究
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-08-27 DOI: 10.1002/ima.70186
Xiaoling Zhou, Shili Wu, Yalu Qiao, Yongkun Guo, Chao Qian, Xinyou Zhang

Medical image segmentation faces the challenge of balancing multiscale anatomical structure modeling and computational efficiency. To address this issue, this paper proposes a “Frequency-Attentive Multi-Hierarchical Network for Medical Image Segmentation” (FreqAtt-MultHier-Net), aiming to achieve synergistic optimization of accuracy, efficiency, and robustness. The core innovations of this paper include: A dual-frequency block (DFB), which decouples high-frequency (detail) and low-frequency (semantic) features through a learnable channel splitting mechanism, and enhances multiscale representations through cross-frequency interaction and dynamic calibration. A multiscale dual-attention fusion block (MSDAFB), which couples channel-spatial dual attention with multi-kernel convolutions to suppress background noise and strengthen local–global contextual fusion. A lightweight ConvMixer module that replaces Transformers with sublinear computational complexity to achieve efficient long-range dependency modeling. In tasks involving cell contour, cell nucleus, lung cancer, skin cancer, liver tumor segmentation and retinal vessel segmentation Task, our model achieves dice similarity coefficients (DSCs) of 95.64%, 92.74%, 83.63%, 85.96%, 85.86% and 84.26%, respectively, while reducing parameter count (25.48 M) and computational cost (5.84 G FLOPs) by 75.9%–84.9% compared to Transformer-based architectures. Ablation experiments validate the independent contributions of each module, with frequency-domain decoupling improving high-frequency detail retention by 18.8% and lightweight design reducing FLOPs by 78.3%. FreqAtt-MultHier-Net provides a high-precision, low-redundancy general solution for medical image segmentation, with potential for low-power clinical deployment. The code is available at the following URL: https://github.com/wu501-CPU/FreqAtt-MultHier-UNet.

医学图像分割面临着平衡多尺度解剖结构建模和计算效率的挑战。针对这一问题,本文提出了一种“频率关注的多层医学图像分割网络”(FreqAtt-MultHier-Net),旨在实现精度、效率和鲁棒性的协同优化。本文的核心创新包括:双频块(DFB),通过可学习的信道分裂机制将高频(细节)和低频(语义)特征解耦,并通过跨频交互和动态校准增强多尺度表示。一种多尺度双注意融合块(MSDAFB),将通道-空间双注意与多核卷积相结合,抑制背景噪声,增强局部-全局上下文融合。一个轻量级的ConvMixer模块,它取代了具有次线性计算复杂性的transformer,以实现高效的远程依赖建模。在涉及细胞轮廓、细胞核、肺癌、皮肤癌、肝肿瘤分割和视网膜血管分割的任务中,我们的模型分别实现了95.64%、92.74%、83.63%、85.96%、85.86%和84.26%的骰子相似系数(dsc),与基于变压器的架构相比,参数计数(25.48 M)和计算成本(5.84 G FLOPs)减少了75.9%-84.9%。烧蚀实验验证了每个模块的独立贡献,频域解耦将高频细节保留率提高了18.8%,轻量化设计将FLOPs降低了78.3%。freqatt - multitier - net为医学图像分割提供了高精度、低冗余的通用解决方案,具有低功耗临床部署的潜力。代码可从以下URL获得:https://github.com/wu501-CPU/FreqAtt-MultHier-UNet。
{"title":"Research on Multi-Objective Optimization of Medical Image Segmentation Based on Frequency Domain Decoupling and Dual Attention Mechanism","authors":"Xiaoling Zhou,&nbsp;Shili Wu,&nbsp;Yalu Qiao,&nbsp;Yongkun Guo,&nbsp;Chao Qian,&nbsp;Xinyou Zhang","doi":"10.1002/ima.70186","DOIUrl":"https://doi.org/10.1002/ima.70186","url":null,"abstract":"<div>\u0000 \u0000 <p>Medical image segmentation faces the challenge of balancing multiscale anatomical structure modeling and computational efficiency. To address this issue, this paper proposes a “Frequency-Attentive Multi-Hierarchical Network for Medical Image Segmentation” (FreqAtt-MultHier-Net), aiming to achieve synergistic optimization of accuracy, efficiency, and robustness. The core innovations of this paper include: A dual-frequency block (DFB), which decouples high-frequency (detail) and low-frequency (semantic) features through a learnable channel splitting mechanism, and enhances multiscale representations through cross-frequency interaction and dynamic calibration. A multiscale dual-attention fusion block (MSDAFB), which couples channel-spatial dual attention with multi-kernel convolutions to suppress background noise and strengthen local–global contextual fusion. A lightweight ConvMixer module that replaces Transformers with sublinear computational complexity to achieve efficient long-range dependency modeling. In tasks involving cell contour, cell nucleus, lung cancer, skin cancer, liver tumor segmentation and retinal vessel segmentation Task, our model achieves dice similarity coefficients (DSCs) of 95.64%, 92.74%, 83.63%, 85.96%, 85.86% and 84.26%, respectively, while reducing parameter count (25.48 M) and computational cost (5.84 G FLOPs) by 75.9%–84.9% compared to Transformer-based architectures. Ablation experiments validate the independent contributions of each module, with frequency-domain decoupling improving high-frequency detail retention by 18.8% and lightweight design reducing FLOPs by 78.3%. FreqAtt-MultHier-Net provides a high-precision, low-redundancy general solution for medical image segmentation, with potential for low-power clinical deployment. The code is available at the following URL: https://github.com/wu501-CPU/FreqAtt-MultHier-UNet.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 5","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144910384","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
VMDUnet: Advancing Glioma Segmentation Integrating With Mamba and Dual Cross-Attention VMDUnet:整合曼巴和双重交叉注意的神经胶质瘤分割
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-08-25 DOI: 10.1002/ima.70187
Zhuo Chen, Yisong Wang, Fangfang Gou

Gliomas are the most common type of primary brain tumor, characterized by their diffuse invasiveness and origin within the central nervous system. Manual identification and segmentation of tumor regions in MRI is a time-consuming and subjective process, and may negatively impact diagnostic accuracy because the heterogeneity and infiltrative pattern of glioma are complex. To address these problems, we propose an automated glioma segmentation approach named IADSG (Intelligent Assistant Diagnosis System for Glioma), based on our novel VMDUnet architecture. Our method incorporates Contrast Limited Adaptive Histogram Equalization (CLAHE) as a preprocessing step to enhance image contrast and quality. Moreover, we use data augmentation techniques to improve the generalization and adaptability to complex clinical images of the model. Crucially, the integration of a Mamba module and a dual cross-attention mechanism enables the model to effectively balance segmentation accuracy with computational efficiency. Experimental results show that our approach achieves a segmentation accuracy of 0.7769 DSC on the internal glioma dataset and 0.9117 DSC on the public BraTS dataset, outperforming existing segmentation methods on both benchmarks. This approach reduces the time and effort involved in manual segmentation, reduces the probabilities of misdiagnosis, and provides robust support for the diagnosis and treatment to be accurately conducted. Our code is available at https://github.com/CarioAo/VMDUnet.

胶质瘤是最常见的原发性脑肿瘤类型,其特点是弥漫性侵袭性和起源于中枢神经系统。由于胶质瘤的异质性和浸润模式复杂,MRI对肿瘤区域的人工识别和分割是一个耗时且主观的过程,可能会对诊断准确性产生负面影响。为了解决这些问题,我们提出了一种基于VMDUnet架构的神经胶质瘤自动分割方法,命名为IADSG (Intelligent Assistant Diagnosis System for glioma)。我们的方法采用对比度限制自适应直方图均衡化(CLAHE)作为预处理步骤,以提高图像的对比度和质量。此外,我们使用数据增强技术来提高模型的泛化和对复杂临床图像的适应性。至关重要的是,曼巴模块和双重交叉注意机制的集成使该模型能够有效地平衡分割精度和计算效率。实验结果表明,该方法在胶质瘤内部数据集上的分割精度为0.7769 DSC,在BraTS公共数据集上的分割精度为0.9117 DSC,在这两个基准上都优于现有的分割方法。该方法减少了人工分割的时间和精力,降低了误诊的概率,为准确进行诊断和治疗提供了强有力的支持。我们的代码可在https://github.com/CarioAo/VMDUnet上获得。
{"title":"VMDUnet: Advancing Glioma Segmentation Integrating With Mamba and Dual Cross-Attention","authors":"Zhuo Chen,&nbsp;Yisong Wang,&nbsp;Fangfang Gou","doi":"10.1002/ima.70187","DOIUrl":"https://doi.org/10.1002/ima.70187","url":null,"abstract":"<div>\u0000 \u0000 <p>Gliomas are the most common type of primary brain tumor, characterized by their diffuse invasiveness and origin within the central nervous system. Manual identification and segmentation of tumor regions in MRI is a time-consuming and subjective process, and may negatively impact diagnostic accuracy because the heterogeneity and infiltrative pattern of glioma are complex. To address these problems, we propose an automated glioma segmentation approach named IADSG (Intelligent Assistant Diagnosis System for Glioma), based on our novel VMDUnet architecture. Our method incorporates Contrast Limited Adaptive Histogram Equalization (CLAHE) as a preprocessing step to enhance image contrast and quality. Moreover, we use data augmentation techniques to improve the generalization and adaptability to complex clinical images of the model. Crucially, the integration of a Mamba module and a dual cross-attention mechanism enables the model to effectively balance segmentation accuracy with computational efficiency. Experimental results show that our approach achieves a segmentation accuracy of 0.7769 DSC on the internal glioma dataset and 0.9117 DSC on the public BraTS dataset, outperforming existing segmentation methods on both benchmarks. This approach reduces the time and effort involved in manual segmentation, reduces the probabilities of misdiagnosis, and provides robust support for the diagnosis and treatment to be accurately conducted. Our code is available at https://github.com/CarioAo/VMDUnet.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 5","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-08-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144894408","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Learning for Differentiating Active and Inactive Multiple Sclerosis Plaques: A Comparative Analysis of MRI-Based Classification Models 深度学习用于区分活跃和非活跃的多发性硬化斑块:基于mri的分类模型的比较分析
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-08-23 DOI: 10.1002/ima.70188
Mohammad Amin Shahram, Mostafa Robatjazi, Atefeh Rostami, Vahid Shahmaei, Ramin Shahrayini, Mohammad Salari

Multiple sclerosis (MS) is a chronic inflammatory disease-causing neurological disability, particularly in young adults. Magnetic resonance imaging (MRI) is the most effective tool for detecting MS plaques, but contrast-enhanced imaging involves potential risks, including toxicity and increased imaging time. Previous methods for differentiating plaque types, such as texture analysis and manual feature extraction, face challenges such as limited datasets and poor generalizability. This study aims to develop and compare deep learning-based methods, specifically convolutional neural networks (CNNs), to classify MS lesion types using non-contrast MRI, aiming to improve clinical applicability and reduce reliance on contrast agents. This study involved 106 multiple sclerosis (MS) patients from two MRI centers. A total of 3410 lesions were analyzed, including 1408 active and 2002 inactive lesions. MRI images, including T1-weighted imaging with gadolinium contrast (T1 + Gd(, T1, Fluid-Attenuated Inversion Recovery (FLAIR), and T2 sequences, were acquired. The segmented lesions were converted into 2D slices and resampled to 128 × 128 pixels for deep learning input. Data augmentation and normalization were applied to improve model generalizability. A custom CNN model was developed and compared with four pre-trained models (ResNet50, VGG16, DenseNet121, and EfficientNetB0) using fivefold cross-validation to evaluate model performance. Performance metrics including accuracy, sensitivity, specificity, and AUC were used. The custom CNN achieved 90.15% accuracy and 94.67% AUC in FLAIR, outperforming pre-trained models. DenseNet121 showed competitive results with 88.23% accuracy and 92.86% AUC in FLAIR. Non-contrast sequences (T1, T2, and FLAIR) combined with deep learning provided promising results, reducing reliance on contrast agents. The custom CNN model excelled in classifying MS lesions across multiple MRI sequences, offering improved diagnostic accuracy and patient safety. Custom models for specialized datasets can enhance clinical outcomes, demonstrating the potential of deep learning in MS diagnosis. These findings suggest that deep learning models can be replaced with contrast agents in routine practice. Future research may explore combining CNNs with clinical features to enhance performance and interpretability.

多发性硬化症(MS)是一种慢性炎症性疾病,引起神经功能障碍,特别是在年轻人中。磁共振成像(MRI)是检测MS斑块最有效的工具,但对比增强成像存在潜在风险,包括毒性和增加成像时间。以前用于区分斑块类型的方法,如纹理分析和人工特征提取,面临着数据集有限和泛化能力差等挑战。本研究旨在开发和比较基于深度学习的方法,特别是卷积神经网络(cnn),利用非对比MRI对MS病变类型进行分类,旨在提高临床适用性,减少对对比剂的依赖。这项研究涉及来自两个MRI中心的106名多发性硬化症(MS)患者。共分析了3410个病变,其中1408个为活动性病变,2002个为非活动性病变。MRI图像包括钆造影剂T1加权成像(T1 + Gd), T1,流体衰减反转恢复(FLAIR)和T2序列。将分割后的病灶转换成二维切片,重新采样到128 × 128像素,作为深度学习输入。采用数据增强和归一化方法提高模型的泛化能力。开发了自定义CNN模型,并将其与四个预训练模型(ResNet50、VGG16、DenseNet121和EfficientNetB0)进行了五倍交叉验证,以评估模型的性能。使用的性能指标包括准确性、灵敏度、特异性和AUC。自定义CNN在FLAIR上的准确率为90.15%,AUC为94.67%,优于预训练模型。DenseNet121在FLAIR上的准确度为88.23%,AUC为92.86%。非对比序列(T1、T2和FLAIR)结合深度学习提供了令人满意的结果,减少了对对比剂的依赖。自定义CNN模型擅长在多个MRI序列中对MS病变进行分类,提高了诊断准确性和患者安全性。专门数据集的定制模型可以提高临床结果,展示了深度学习在MS诊断中的潜力。这些发现表明,在日常实践中,深度学习模型可以被造影剂取代。未来的研究可能会探索将cnn与临床特征结合起来,以提高性能和可解释性。
{"title":"Deep Learning for Differentiating Active and Inactive Multiple Sclerosis Plaques: A Comparative Analysis of MRI-Based Classification Models","authors":"Mohammad Amin Shahram,&nbsp;Mostafa Robatjazi,&nbsp;Atefeh Rostami,&nbsp;Vahid Shahmaei,&nbsp;Ramin Shahrayini,&nbsp;Mohammad Salari","doi":"10.1002/ima.70188","DOIUrl":"https://doi.org/10.1002/ima.70188","url":null,"abstract":"<div>\u0000 \u0000 <p>Multiple sclerosis (MS) is a chronic inflammatory disease-causing neurological disability, particularly in young adults. Magnetic resonance imaging (MRI) is the most effective tool for detecting MS plaques, but contrast-enhanced imaging involves potential risks, including toxicity and increased imaging time. Previous methods for differentiating plaque types, such as texture analysis and manual feature extraction, face challenges such as limited datasets and poor generalizability. This study aims to develop and compare deep learning-based methods, specifically convolutional neural networks (CNNs), to classify MS lesion types using non-contrast MRI, aiming to improve clinical applicability and reduce reliance on contrast agents. This study involved 106 multiple sclerosis (MS) patients from two MRI centers. A total of 3410 lesions were analyzed, including 1408 active and 2002 inactive lesions. MRI images, including T1-weighted imaging with gadolinium contrast (T1 + Gd(, T1, Fluid-Attenuated Inversion Recovery (FLAIR), and T2 sequences, were acquired. The segmented lesions were converted into 2D slices and resampled to 128 × 128 pixels for deep learning input. Data augmentation and normalization were applied to improve model generalizability. A custom CNN model was developed and compared with four pre-trained models (ResNet50, VGG16, DenseNet121, and EfficientNetB0) using fivefold cross-validation to evaluate model performance. Performance metrics including accuracy, sensitivity, specificity, and AUC were used. The custom CNN achieved 90.15% accuracy and 94.67% AUC in FLAIR, outperforming pre-trained models. DenseNet121 showed competitive results with 88.23% accuracy and 92.86% AUC in FLAIR. Non-contrast sequences (T1, T2, and FLAIR) combined with deep learning provided promising results, reducing reliance on contrast agents. The custom CNN model excelled in classifying MS lesions across multiple MRI sequences, offering improved diagnostic accuracy and patient safety. Custom models for specialized datasets can enhance clinical outcomes, demonstrating the potential of deep learning in MS diagnosis. These findings suggest that deep learning models can be replaced with contrast agents in routine practice. Future research may explore combining CNNs with clinical features to enhance performance and interpretability.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 5","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-08-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144892482","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MSAF: A Cardiac 3D Image Segmentation Network Based on Multiscale Collaborative Attention and Multiscale Feature Fusion 基于多尺度协同关注和多尺度特征融合的心脏三维图像分割网络
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-08-21 DOI: 10.1002/ima.70184
Guodong Zhang, He Li, Wanying Xie, Bin Yang, Zhaoxuan Gong, Wei Guo, Ronghui Ju

Accurate segmentation of cardiac structures is essential for clinical diagnosis and treatment of cardiovascular diseases. Existing Transformer-based cardiac segmentation methods mostly rely on single-scale token-wise attention mechanisms that emphasize global feature modeling, but they lack sufficient sensitivity to local spatial structures, such as myocardial boundaries in cardiac 3D images, resulting in ineffective multiscale feature capturing and a loss of local spatial details, thereby negatively impacting the accuracy of cardiac anatomical segmentation. To address the above issues, this paper proposes a cardiac 3D image segmentation network named MSAF, which integrates Multiscale Collaborative Attention (MSCA) and Multiscale Feature Fusion (MSFF) modules to enhance the multiscale feature perception capability at both microscopic and macroscopic levels, thereby improving segmentation accuracy for complex cardiac structures. Within the MSCA module, a Collaborative Attention (CoA) module combined with hierarchical residual-like connections is designed, enabling the model to effectively capture interactive information across spatial and channel dimensions at various receptive fields and facilitating finer-grained feature extraction. In the MSFF module, a gradient-based feature importance weighting mechanism dynamically adjusts feature contributions from different hierarchical levels, effectively fusing high-level abstract semantic information with low-level spatial details, thereby enhancing cross-scale feature representation and optimizing both global completeness and local boundary precision in segmentation results. Experimental validation of MSAF was conducted on four publicly available medical image segmentation datasets, including ACDC, FlARE21, and MM-WHS (MRI and CT modalities), yielding average Dice values of 93.27%, 88.16%, 92.23%, and 91.22%, respectively. These experimental results demonstrate the effectiveness of MSAF in segmenting detailed cardiac structures.

心脏结构的准确分割对于心血管疾病的临床诊断和治疗至关重要。现有的基于变压器的心脏分割方法大多依赖于强调全局特征建模的单尺度标记注意机制,但对局部空间结构(如心脏三维图像中的心肌边界)缺乏足够的敏感性,导致多尺度特征捕获无效,局部空间细节丢失,从而影响心脏解剖分割的准确性。针对上述问题,本文提出了一种心脏三维图像分割网络MSAF,该网络集成了多尺度协同关注(MSCA)和多尺度特征融合(MSFF)模块,从微观和宏观两个层面增强了心脏三维图像的多尺度特征感知能力,从而提高了心脏复杂结构的分割精度。在MSCA模块中,设计了一个协同注意(CoA)模块,结合分层残差连接,使模型能够在不同的接受域上有效地捕获跨空间和通道维度的交互信息,并促进更细粒度的特征提取。在MSFF模块中,基于梯度的特征重要性加权机制动态调整不同层次的特征贡献,有效融合高层次抽象语义信息和低层次空间细节信息,增强跨尺度特征表示,优化分割结果的全局完整性和局部边界精度。在ACDC、FlARE21和MM-WHS (MRI和CT模式)4个公开的医学图像分割数据集上对MSAF进行了实验验证,平均Dice值分别为93.27%、88.16%、92.23%和91.22%。这些实验结果证明了MSAF分割心脏详细结构的有效性。
{"title":"MSAF: A Cardiac 3D Image Segmentation Network Based on Multiscale Collaborative Attention and Multiscale Feature Fusion","authors":"Guodong Zhang,&nbsp;He Li,&nbsp;Wanying Xie,&nbsp;Bin Yang,&nbsp;Zhaoxuan Gong,&nbsp;Wei Guo,&nbsp;Ronghui Ju","doi":"10.1002/ima.70184","DOIUrl":"https://doi.org/10.1002/ima.70184","url":null,"abstract":"<div>\u0000 \u0000 <p>Accurate segmentation of cardiac structures is essential for clinical diagnosis and treatment of cardiovascular diseases. Existing Transformer-based cardiac segmentation methods mostly rely on single-scale token-wise attention mechanisms that emphasize global feature modeling, but they lack sufficient sensitivity to local spatial structures, such as myocardial boundaries in cardiac 3D images, resulting in ineffective multiscale feature capturing and a loss of local spatial details, thereby negatively impacting the accuracy of cardiac anatomical segmentation. To address the above issues, this paper proposes a cardiac 3D image segmentation network named MSAF, which integrates Multiscale Collaborative Attention (MSCA) and Multiscale Feature Fusion (MSFF) modules to enhance the multiscale feature perception capability at both microscopic and macroscopic levels, thereby improving segmentation accuracy for complex cardiac structures. Within the MSCA module, a Collaborative Attention (CoA) module combined with hierarchical residual-like connections is designed, enabling the model to effectively capture interactive information across spatial and channel dimensions at various receptive fields and facilitating finer-grained feature extraction. In the MSFF module, a gradient-based feature importance weighting mechanism dynamically adjusts feature contributions from different hierarchical levels, effectively fusing high-level abstract semantic information with low-level spatial details, thereby enhancing cross-scale feature representation and optimizing both global completeness and local boundary precision in segmentation results. Experimental validation of MSAF was conducted on four publicly available medical image segmentation datasets, including ACDC, FlARE21, and MM-WHS (MRI and CT modalities), yielding average Dice values of 93.27%, 88.16%, 92.23%, and 91.22%, respectively. These experimental results demonstrate the effectiveness of MSAF in segmenting detailed cardiac structures.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 5","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-08-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144885087","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient Melanoma Detection Using Pixel Intensity-Based Masking and Intensity-Weighted Binary Cross-Entropy 基于像素强度的掩蔽和强度加权二元交叉熵的高效黑色素瘤检测
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-08-19 DOI: 10.1002/ima.70179
Asaad Ahmed, Guangmin Sun, Mohamed Saadeldin, Anas Bilal, Yu Li, Musa Osman, Shouki A. Ebad

Melanoma, the deadliest form of skin cancer, requires accurate and timely detection to improve survival rates and treatment outcomes. Deep learning has shown significant potential in automating melanoma detection; however, existing methods face challenges such as irrelevant background information in dermoscopic images and class imbalance in melanoma datasets, which hinder diagnostic performance. To address these challenges, this paper introduces two complementary contributions: Pixel Intensity-Based Masking (PIBM) and Intensity-Weighted Binary Cross-Entropy (IW-BCE). PIBM is a novel preprocessing technique that dynamically identifies and masks low-priority regions in dermoscopic images based on pixel intensity values. By preserving high-intensity lesion regions and suppressing irrelevant background artifacts, PIBM reduces computational complexity and enhances the model's focus on diagnostically critical features, all without requiring ground truth annotations or pixel-level labeling. Additionally, IW-BCE, a custom loss function, is designed to handle class imbalance by dynamically adjusting the contribution of each class during training. By assigning higher weights to the minority class (malignant lesions), IW-BCE enhances the model's sensitivity, reduces false negatives, and improves recall, an essential metric in medical diagnostics. The proposed framework integrates PIBM and IW-BCE into a deep-learning pipeline for melanoma detection. Evaluations on benchmark datasets demonstrate that the combined approach achieves superior performance compared to traditional methods in terms of accuracy, sensitivity, and computational efficiency. Specifically, the proposed method achieves a higher recall and F1-score, highlighting its ability to address the critical limitations of existing systems. This work offers a robust and clinically relevant solution for real-time melanoma detection, paving the way for improved early diagnosis and patient outcomes.

黑色素瘤是最致命的皮肤癌,需要准确及时的检测来提高生存率和治疗效果。深度学习在自动化黑色素瘤检测方面显示出巨大的潜力;然而,现有的方法面临着诸如皮肤镜图像背景信息不相关和黑色素瘤数据集分类不平衡等挑战,这些都阻碍了诊断性能。为了解决这些挑战,本文介绍了两个互补的贡献:基于像素强度的掩蔽(PIBM)和强度加权二进制交叉熵(IW-BCE)。PIBM是一种基于像素强度值动态识别和掩盖皮肤镜图像中低优先级区域的新型预处理技术。通过保留高强度病变区域和抑制不相关的背景伪影,PIBM降低了计算复杂性,增强了模型对诊断关键特征的关注,所有这些都不需要基础真值注释或像素级标记。此外,自定义损失函数IW-BCE通过在训练期间动态调整每个类的贡献来处理类不平衡。通过给少数类别(恶性病变)分配更高的权重,IW-BCE提高了模型的灵敏度,减少了假阴性,提高了召回率,这是医疗诊断中的一个重要指标。该框架将PIBM和IW-BCE集成到黑色素瘤检测的深度学习管道中。对基准数据集的评估表明,与传统方法相比,该组合方法在准确性、灵敏度和计算效率方面取得了更好的性能。具体来说,所提出的方法实现了更高的召回率和f1分,突出了其解决现有系统的关键限制的能力。这项工作为实时黑色素瘤检测提供了一个强大的临床相关解决方案,为改善早期诊断和患者预后铺平了道路。
{"title":"Efficient Melanoma Detection Using Pixel Intensity-Based Masking and Intensity-Weighted Binary Cross-Entropy","authors":"Asaad Ahmed,&nbsp;Guangmin Sun,&nbsp;Mohamed Saadeldin,&nbsp;Anas Bilal,&nbsp;Yu Li,&nbsp;Musa Osman,&nbsp;Shouki A. Ebad","doi":"10.1002/ima.70179","DOIUrl":"https://doi.org/10.1002/ima.70179","url":null,"abstract":"<div>\u0000 \u0000 <p>Melanoma, the deadliest form of skin cancer, requires accurate and timely detection to improve survival rates and treatment outcomes. Deep learning has shown significant potential in automating melanoma detection; however, existing methods face challenges such as irrelevant background information in dermoscopic images and class imbalance in melanoma datasets, which hinder diagnostic performance. To address these challenges, this paper introduces two complementary contributions: Pixel Intensity-Based Masking (PIBM) and Intensity-Weighted Binary Cross-Entropy (IW-BCE). PIBM is a novel preprocessing technique that dynamically identifies and masks low-priority regions in dermoscopic images based on pixel intensity values. By preserving high-intensity lesion regions and suppressing irrelevant background artifacts, PIBM reduces computational complexity and enhances the model's focus on diagnostically critical features, all without requiring ground truth annotations or pixel-level labeling. Additionally, IW-BCE, a custom loss function, is designed to handle class imbalance by dynamically adjusting the contribution of each class during training. By assigning higher weights to the minority class (malignant lesions), IW-BCE enhances the model's sensitivity, reduces false negatives, and improves recall, an essential metric in medical diagnostics. The proposed framework integrates PIBM and IW-BCE into a deep-learning pipeline for melanoma detection. Evaluations on benchmark datasets demonstrate that the combined approach achieves superior performance compared to traditional methods in terms of accuracy, sensitivity, and computational efficiency. Specifically, the proposed method achieves a higher recall and F1-score, highlighting its ability to address the critical limitations of existing systems. This work offers a robust and clinically relevant solution for real-time melanoma detection, paving the way for improved early diagnosis and patient outcomes.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 5","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144869631","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
GlaucoDiff: A Framework for Generating Balanced Glaucoma Fundus Images and Improving Diagnostic Performance GlaucoDiff:生成平衡的青光眼眼底图像和提高诊断性能的框架
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-08-18 DOI: 10.1002/ima.70185
Caisheng Liao, Yuki Todo, Jiashu Zhang, Zheng Tang

Glaucoma is a leading cause of irreversible blindness, and early diagnosis is critical. While retinal fundus images are commonly used for screening, AI-based diagnostic models face challenges such as data scarcity, class imbalance, and limited image diversity. To address this, we introduce GlaucoDiff, a diffusion-based image synthesis framework designed to generate clinically meaningful glaucoma fundus images. It employs a two-stage training strategy and integrates a multimodal large language model as an automated quality filter to ensure clinical relevance. Experiments on the JustRAIGS dataset show that GlaucoDiff outperforms commercial generators such as DALL-E 3 and Keling, achieving better image quality and diversity (FID: 109.8; SWD: 222.2). When synthetic images were used to augment the training set of a vision transformer classifier, sensitivity improved consistently from 0.8182 with only real data to 0.8615 with 10% synthetic images, and further to 0.8788 with 50%. However, as the proportion of synthetic data increased, other important metrics such as specificity, accuracy, and AUC began to decline compared to the results with 10% synthetic data. This finding suggests that although more synthetic images can enhance the model's ability to detect positive cases, too much synthetic data may reduce overall classification performance. These results demonstrate the practical value of GlaucoDiff in alleviating data imbalance and improving diagnostic accuracy for AI-assisted glaucoma screening.

青光眼是不可逆失明的主要原因,早期诊断至关重要。虽然视网膜眼底图像通常用于筛查,但基于人工智能的诊断模型面临着数据稀缺、类别不平衡和图像多样性有限等挑战。为了解决这个问题,我们引入了GlaucoDiff,这是一个基于弥散的图像合成框架,旨在生成具有临床意义的青光眼眼底图像。它采用两阶段训练策略,并集成了多模态大语言模型作为自动质量过滤器,以确保临床相关性。在JustRAIGS数据集上的实验表明,GlaucoDiff优于商用生成器,如DALL-E 3和Keling,实现了更好的图像质量和多样性(FID: 109.8;社署:222.2)。当使用合成图像增强视觉变换分类器的训练集时,灵敏度从仅真实数据时的0.8182持续提高到10%合成图像时的0.8615,再到50%合成图像时的0.8788。然而,随着合成数据比例的增加,与10%合成数据的结果相比,特异性、准确性和AUC等其他重要指标开始下降。这一发现表明,虽然更多的合成图像可以增强模型检测阳性病例的能力,但过多的合成数据可能会降低整体分类性能。这些结果证明了GlaucoDiff在缓解数据不平衡和提高人工智能辅助青光眼筛查诊断准确性方面的实用价值。
{"title":"GlaucoDiff: A Framework for Generating Balanced Glaucoma Fundus Images and Improving Diagnostic Performance","authors":"Caisheng Liao,&nbsp;Yuki Todo,&nbsp;Jiashu Zhang,&nbsp;Zheng Tang","doi":"10.1002/ima.70185","DOIUrl":"https://doi.org/10.1002/ima.70185","url":null,"abstract":"<div>\u0000 \u0000 <p>Glaucoma is a leading cause of irreversible blindness, and early diagnosis is critical. While retinal fundus images are commonly used for screening, AI-based diagnostic models face challenges such as data scarcity, class imbalance, and limited image diversity. To address this, we introduce GlaucoDiff, a diffusion-based image synthesis framework designed to generate clinically meaningful glaucoma fundus images. It employs a two-stage training strategy and integrates a multimodal large language model as an automated quality filter to ensure clinical relevance. Experiments on the JustRAIGS dataset show that GlaucoDiff outperforms commercial generators such as DALL-E 3 and Keling, achieving better image quality and diversity (FID: 109.8; SWD: 222.2). When synthetic images were used to augment the training set of a vision transformer classifier, sensitivity improved consistently from 0.8182 with only real data to 0.8615 with 10% synthetic images, and further to 0.8788 with 50%. However, as the proportion of synthetic data increased, other important metrics such as specificity, accuracy, and AUC began to decline compared to the results with 10% synthetic data. This finding suggests that although more synthetic images can enhance the model's ability to detect positive cases, too much synthetic data may reduce overall classification performance. These results demonstrate the practical value of GlaucoDiff in alleviating data imbalance and improving diagnostic accuracy for AI-assisted glaucoma screening.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 5","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144861895","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Novel Multimodal Medical Image Fusion Method Based on Detail Enhancement and Dual-Branch Feature Fusion 一种基于细节增强和双分支特征融合的多模态医学图像融合方法
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-08-15 DOI: 10.1002/ima.70181
Kun Zhang, Hui Yuan, Zhongwei Zhang, PengPeng Sun

Multimodal medical image fusion integrates effective information from different modal images and integrates salient and complementary features, which can more comprehensively describe the condition of lesions and make medical diagnosis results more reliable. This paper proposes a multimodal medical image fusion method based on image detail enhancement and dual-branch feature fusion (DEDF). First, the source images are preprocessed by guided filtering to enhance important details and improve the fusion and visualization effects. Then, local extreme maps are used as guides to smooth the source images. Finally, a DEDF mechanism based on guided filtering and bilateral filtering is established to obtain multiscale bright and dark feature maps, as well as base images of different modalities, which are fused to obtain a more comprehensive medical image and improve the accuracy of medical diagnosis results. Extensive experiments, compared qualitatively and quantitatively with various state-of-the-art medical image fusion methods, validate the superior fusion performance and effectiveness of the proposed method.

多模态医学图像融合融合了不同模态图像的有效信息,融合了显著特征和互补特征,可以更全面地描述病变的状况,使医学诊断结果更加可靠。提出了一种基于图像细节增强和双分支特征融合(DEDF)的多模态医学图像融合方法。首先,对源图像进行引导滤波预处理,增强重要细节,提高融合效果和可视化效果;然后,利用局部极值图作为导线对源图像进行平滑处理。最后,建立基于引导滤波和双边滤波的DEDF机制,获得多尺度明暗特征图,以及不同模态的基础图像,融合得到更全面的医学图像,提高医学诊断结果的准确性。大量的实验,定性和定量地比较了各种最先进的医学图像融合方法,验证了该方法优越的融合性能和有效性。
{"title":"A Novel Multimodal Medical Image Fusion Method Based on Detail Enhancement and Dual-Branch Feature Fusion","authors":"Kun Zhang,&nbsp;Hui Yuan,&nbsp;Zhongwei Zhang,&nbsp;PengPeng Sun","doi":"10.1002/ima.70181","DOIUrl":"https://doi.org/10.1002/ima.70181","url":null,"abstract":"<div>\u0000 \u0000 <p>Multimodal medical image fusion integrates effective information from different modal images and integrates salient and complementary features, which can more comprehensively describe the condition of lesions and make medical diagnosis results more reliable. This paper proposes a multimodal medical image fusion method based on image detail enhancement and dual-branch feature fusion (DEDF). First, the source images are preprocessed by guided filtering to enhance important details and improve the fusion and visualization effects. Then, local extreme maps are used as guides to smooth the source images. Finally, a DEDF mechanism based on guided filtering and bilateral filtering is established to obtain multiscale bright and dark feature maps, as well as base images of different modalities, which are fused to obtain a more comprehensive medical image and improve the accuracy of medical diagnosis results. Extensive experiments, compared qualitatively and quantitatively with various state-of-the-art medical image fusion methods, validate the superior fusion performance and effectiveness of the proposed method.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 5","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-08-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144853690","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
BATU: A Workflow for Multi-Network Ensemble Learning in Cross-Dataset Generalization of Skin Lesion Analysis 皮肤损伤分析跨数据集泛化中的多网络集成学习工作流
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-08-15 DOI: 10.1002/ima.70183
Ömer Faruk Söylemez

The development of computer vision systems for dermatological diagnosis is often hindered by dataset heterogeneity, including differences in image quality, labeling strategies, and patient demographics. In this study, we examine how such heterogeneity affects the generalization ability of computer vision models across three public dermatology image datasets. We trained five different deep learning models on each dataset separately and evaluated their performance in both intra-dataset and cross-dataset settings. To further investigate robustness, we conducted multi-source domain generalization experiments by training models on combinations of two datasets and testing on the third unseen dataset. We observed a significant drop in performance during cross-dataset evaluations. To address this, we applied various ensemble learning methods by combining the predictions from the individual models. Our results demonstrate that ensemble approaches consistently outperform individual models, achieving accuracy improvements exceeding 4% in many cases. These findings highlight the potential of ensemble learning to address challenges related to dataset variability in dermatological image analysis.

用于皮肤病诊断的计算机视觉系统的发展经常受到数据集异质性的阻碍,包括图像质量、标记策略和患者人口统计数据的差异。在这项研究中,我们研究了这种异质性如何影响计算机视觉模型在三个公共皮肤病学图像数据集上的泛化能力。我们在每个数据集上分别训练了五种不同的深度学习模型,并评估了它们在数据集内和跨数据集设置下的性能。为了进一步研究鲁棒性,我们通过在两个数据集的组合上训练模型并在第三个未见数据集上进行测试,进行了多源领域泛化实验。在跨数据集评估期间,我们观察到性能显著下降。为了解决这个问题,我们通过组合来自各个模型的预测,应用了各种集成学习方法。我们的结果表明,集成方法始终优于单个模型,在许多情况下实现了超过4%的精度改进。这些发现突出了集成学习在解决皮肤病学图像分析中与数据集变异性相关的挑战方面的潜力。
{"title":"BATU: A Workflow for Multi-Network Ensemble Learning in Cross-Dataset Generalization of Skin Lesion Analysis","authors":"Ömer Faruk Söylemez","doi":"10.1002/ima.70183","DOIUrl":"https://doi.org/10.1002/ima.70183","url":null,"abstract":"<div>\u0000 \u0000 <p>The development of computer vision systems for dermatological diagnosis is often hindered by dataset heterogeneity, including differences in image quality, labeling strategies, and patient demographics. In this study, we examine how such heterogeneity affects the generalization ability of computer vision models across three public dermatology image datasets. We trained five different deep learning models on each dataset separately and evaluated their performance in both intra-dataset and cross-dataset settings. To further investigate robustness, we conducted multi-source domain generalization experiments by training models on combinations of two datasets and testing on the third unseen dataset. We observed a significant drop in performance during cross-dataset evaluations. To address this, we applied various ensemble learning methods by combining the predictions from the individual models. Our results demonstrate that ensemble approaches consistently outperform individual models, achieving accuracy improvements exceeding 4% in many cases. These findings highlight the potential of ensemble learning to address challenges related to dataset variability in dermatological image analysis.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 5","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-08-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144853691","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards Real Time Alzheimer's Diagnosis: A PSO-GA-Driven Deep Learning Solution for Telemedicine 面向实时阿尔茨海默病诊断:pso - ga驱动的远程医疗深度学习解决方案
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-08-14 DOI: 10.1002/ima.70180
Anupam Kumar, Faiyaz Ahmad, Bashir Alam

Alzheimer's disease (AD) is a progressive neurodegenerative disorder characterized by cognitive decline and structural brain deterioration, with its global prevalence projected to exceed 125 million by 2050. Early and accurate diagnosis—particularly the differentiation of mild cognitive impairment (MCI) from normal aging—is critical for effective intervention; yet it remains challenging due to subtle anatomical changes and high-dimensional imaging data. This study presents a telehealth-compatible computer-aided diagnosis (CAD) framework for multi-class AD classification using structural MRI (sMRI) images from the publicly available ADNI dataset. The framework integrates transfer learning with DenseNet121 (pre-trained on RadImageNet) for deep feature extraction and employs a hybrid bio-inspired particle swarm optimization–genetic algorithm (PSO-GA) for feature selection and dimensionality reduction. This optimized pipeline reduces the original high-dimensional feature space to 16 key features, improving classification accuracy from 88.48% to 99.78% using AdaBoost. The proposed PSO-GA-DenseNet framework delivers a lightweight, scalable solution suitable for remote diagnostic settings. Compared to existing state-of-the-art models, it offers enhanced computational efficiency and robust cross-site adaptability. Future research will focus on improving generalizability across imaging modalities and incorporating longitudinal data to enable real-time, cross-modal, and large-scale deployment in clinical and telehealth environments.

阿尔茨海默病(AD)是一种进行性神经退行性疾病,其特征是认知能力下降和大脑结构性退化,到2050年,其全球患病率预计将超过1.25亿。早期和准确的诊断-特别是轻度认知障碍(MCI)与正常衰老的区分-是有效干预的关键;然而,由于细微的解剖变化和高维成像数据,它仍然具有挑战性。本研究提出了一个远程医疗兼容的计算机辅助诊断(CAD)框架,用于使用来自公开可用的ADNI数据集的结构MRI (sMRI)图像进行AD多类别分类。该框架将迁移学习与DenseNet121(在RadImageNet上预训练)集成在一起进行深度特征提取,并采用混合生物灵感粒子群优化-遗传算法(PSO-GA)进行特征选择和降维。优化后的管道将原始的高维特征空间减少到16个关键特征,使用AdaBoost将分类准确率从88.48%提高到99.78%。提出的PSO-GA-DenseNet框架提供了适合远程诊断设置的轻量级,可扩展的解决方案。与现有的最先进的模型相比,它提供了更高的计算效率和强大的跨站点适应性。未来的研究将侧重于提高跨成像模式的通用性,并结合纵向数据,以便在临床和远程医疗环境中实现实时、跨模式和大规模部署。
{"title":"Towards Real Time Alzheimer's Diagnosis: A PSO-GA-Driven Deep Learning Solution for Telemedicine","authors":"Anupam Kumar,&nbsp;Faiyaz Ahmad,&nbsp;Bashir Alam","doi":"10.1002/ima.70180","DOIUrl":"https://doi.org/10.1002/ima.70180","url":null,"abstract":"<div>\u0000 \u0000 <p>Alzheimer's disease (AD) is a progressive neurodegenerative disorder characterized by cognitive decline and structural brain deterioration, with its global prevalence projected to exceed 125 million by 2050. Early and accurate diagnosis—particularly the differentiation of mild cognitive impairment (MCI) from normal aging—is critical for effective intervention; yet it remains challenging due to subtle anatomical changes and high-dimensional imaging data. This study presents a telehealth-compatible computer-aided diagnosis (CAD) framework for multi-class AD classification using structural MRI (sMRI) images from the publicly available ADNI dataset. The framework integrates transfer learning with DenseNet121 (pre-trained on RadImageNet) for deep feature extraction and employs a hybrid bio-inspired particle swarm optimization–genetic algorithm (PSO-GA) for feature selection and dimensionality reduction. This optimized pipeline reduces the original high-dimensional feature space to 16 key features, improving classification accuracy from 88.48% to 99.78% using AdaBoost. The proposed PSO-GA-DenseNet framework delivers a lightweight, scalable solution suitable for remote diagnostic settings. Compared to existing state-of-the-art models, it offers enhanced computational efficiency and robust cross-site adaptability. Future research will focus on improving generalizability across imaging modalities and incorporating longitudinal data to enable real-time, cross-modal, and large-scale deployment in clinical and telehealth environments.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 5","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144833299","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Bifocusing-Based Microwave Imaging Under Background Conductivity Mismatch 背景电导率失配下基于双聚焦的微波成像
IF 2.5 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-08-13 DOI: 10.1002/ima.70182
Janghoon Jeong, Seong-Ho Son

Microwave imaging is a promising technique for non-invasive diagnostics in areas such as medical imaging, remote sensing, and subsurface exploration. Its performance, however, strongly depends on accurate knowledge of background medium properties, particularly electrical conductivity. In practice, these parameters are often uncertain or mismatched, leading to signal degradation and inaccurate reconstructions. This study investigates the impact of conductivity mismatch on the bifocusing method (BFM), a qualitative microwave imaging algorithm, and proposes an improved version incorporating attenuation compensation. Simulations show that conventional BFM fails to identify objects when the assumed conductivity is significantly higher than the actual value. To resolve this, we modify the Green's function by introducing a compensation term based on the attenuation constant, restoring the incident field amplitude. The improved method enables successful object recovery even under severe mismatch. Quantitative evaluation using the Jaccard similarity index confirms improved localization accuracy. This approach enhances the robustness of microwave imaging and shows promise for medical diagnostics in highly attenuating biological tissues.

微波成像在医学成像、遥感和地下探测等领域是非侵入性诊断技术。然而,它的性能在很大程度上取决于对背景介质特性,特别是导电性的准确了解。实际上,这些参数往往是不确定的或不匹配的,导致信号退化和不准确的重建。本研究探讨了电导率失配对定性微波成像算法双聚焦法(BFM)的影响,并提出了一种包含衰减补偿的改进版本。仿真结果表明,当假设电导率显著高于实际电导率时,传统的BFM无法识别目标。为了解决这个问题,我们通过引入基于衰减常数的补偿项来修改格林函数,恢复入射场振幅。改进的方法即使在严重不匹配的情况下也能成功地恢复对象。使用Jaccard相似性指数的定量评价证实了定位精度的提高。该方法增强了微波成像的鲁棒性,并为高度衰减的生物组织的医学诊断显示了希望。
{"title":"Bifocusing-Based Microwave Imaging Under Background Conductivity Mismatch","authors":"Janghoon Jeong,&nbsp;Seong-Ho Son","doi":"10.1002/ima.70182","DOIUrl":"https://doi.org/10.1002/ima.70182","url":null,"abstract":"<div>\u0000 \u0000 <p>Microwave imaging is a promising technique for non-invasive diagnostics in areas such as medical imaging, remote sensing, and subsurface exploration. Its performance, however, strongly depends on accurate knowledge of background medium properties, particularly electrical conductivity. In practice, these parameters are often uncertain or mismatched, leading to signal degradation and inaccurate reconstructions. This study investigates the impact of conductivity mismatch on the bifocusing method (BFM), a qualitative microwave imaging algorithm, and proposes an improved version incorporating attenuation compensation. Simulations show that conventional BFM fails to identify objects when the assumed conductivity is significantly higher than the actual value. To resolve this, we modify the Green's function by introducing a compensation term based on the attenuation constant, restoring the incident field amplitude. The improved method enables successful object recovery even under severe mismatch. Quantitative evaluation using the Jaccard similarity index confirms improved localization accuracy. This approach enhances the robustness of microwave imaging and shows promise for medical diagnostics in highly attenuating biological tissues.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 5","pages":""},"PeriodicalIF":2.5,"publicationDate":"2025-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144833228","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International Journal of Imaging Systems and Technology
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1