首页 > 最新文献

International Journal of Imaging Systems and Technology最新文献

英文 中文
CasUNeXt: A Cascaded Transformer With Intra- and Inter-Scale Information for Medical Image Segmentation CasUNeXt:用于医学图像分割的具有尺度内和尺度间信息的级联变换器
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-09-21 DOI: 10.1002/ima.23184
Junding Sun, Xiaopeng Zheng, Xiaosheng Wu, Chaosheng Tang, Shuihua Wang, Yudong Zhang

Due to the Transformer's ability to capture long-range dependencies through Self-Attention, it has shown immense potential in medical image segmentation. However, it lacks the capability to model local relationships between pixels. Therefore, many previous approaches embedded the Transformer into the CNN encoder. However, current methods often fall short in modeling the relationships between multi-scale features, specifically the spatial correspondence between features at different scales. This limitation can result in the ineffective capture of scale differences for each object and the loss of features for small targets. Furthermore, due to the high complexity of the Transformer, it is challenging to integrate local and global information within the same scale effectively. To address these limitations, we propose a novel backbone network called CasUNeXt, which features three appealing design elements: (1) We use the idea of cascade to redesign the way CNN and Transformer are combined to enhance modeling the unique interrelationships between multi-scale information. (2) We design a Cascaded Scale-wise Transformer Module capable of cross-scale interactions. It not only strengthens feature extraction within a single scale but also models interactions between different scales. (3) We overhaul the multi-head Channel Attention mechanism to enable it to model context information in feature maps from multiple perspectives within the channel dimension. These design features collectively enable CasUNeXt to better integrate local and global information and capture relationships between multi-scale features, thereby improving the performance of medical image segmentation. Through experimental comparisons on various benchmark datasets, our CasUNeXt method exhibits outstanding performance in medical image segmentation tasks, surpassing the current state-of-the-art methods.

由于 Transformer 能够通过自我关注捕捉长距离依赖关系,因此在医学图像分割方面显示出巨大的潜力。然而,它缺乏对像素间局部关系建模的能力。因此,以前的许多方法都将变换器嵌入到 CNN 编码器中。然而,目前的方法往往无法模拟多尺度特征之间的关系,特别是不同尺度特征之间的空间对应关系。这种局限性会导致无法有效捕捉每个物体的尺度差异,以及丢失小目标的特征。此外,由于变换器的高复杂性,在同一尺度内有效整合局部和全局信息也是一项挑战。为了解决这些局限性,我们提出了一种名为 CasUNeXt 的新型骨干网络,它具有三个吸引人的设计元素:(1) 我们利用级联的思想重新设计了 CNN 和 Transformer 的组合方式,以加强对多尺度信息之间独特相互关系的建模。(2) 我们设计了一个能够进行跨尺度交互的级联尺度变换器模块。它不仅能加强单一尺度内的特征提取,还能模拟不同尺度之间的交互。(3) 我们彻底改变了多头通道关注机制,使其能够在通道维度内从多个角度对特征图中的上下文信息进行建模。这些设计特点使 CasUNeXt 能够更好地整合局部和全局信息,捕捉多尺度特征之间的关系,从而提高医学图像分割的性能。通过在各种基准数据集上的实验比较,我们的 CasUNeXt 方法在医学图像分割任务中表现出卓越的性能,超越了目前最先进的方法。
{"title":"CasUNeXt: A Cascaded Transformer With Intra- and Inter-Scale Information for Medical Image Segmentation","authors":"Junding Sun,&nbsp;Xiaopeng Zheng,&nbsp;Xiaosheng Wu,&nbsp;Chaosheng Tang,&nbsp;Shuihua Wang,&nbsp;Yudong Zhang","doi":"10.1002/ima.23184","DOIUrl":"https://doi.org/10.1002/ima.23184","url":null,"abstract":"<p>Due to the Transformer's ability to capture long-range dependencies through Self-Attention, it has shown immense potential in medical image segmentation. However, it lacks the capability to model local relationships between pixels. Therefore, many previous approaches embedded the Transformer into the CNN encoder. However, current methods often fall short in modeling the relationships between multi-scale features, specifically the spatial correspondence between features at different scales. This limitation can result in the ineffective capture of scale differences for each object and the loss of features for small targets. Furthermore, due to the high complexity of the Transformer, it is challenging to integrate local and global information within the same scale effectively. To address these limitations, we propose a novel backbone network called CasUNeXt, which features three appealing design elements: (1) We use the idea of cascade to redesign the way CNN and Transformer are combined to enhance modeling the unique interrelationships between multi-scale information. (2) We design a Cascaded Scale-wise Transformer Module capable of cross-scale interactions. It not only strengthens feature extraction within a single scale but also models interactions between different scales. (3) We overhaul the multi-head Channel Attention mechanism to enable it to model context information in feature maps from multiple perspectives within the channel dimension. These design features collectively enable CasUNeXt to better integrate local and global information and capture relationships between multi-scale features, thereby improving the performance of medical image segmentation. Through experimental comparisons on various benchmark datasets, our CasUNeXt method exhibits outstanding performance in medical image segmentation tasks, surpassing the current state-of-the-art methods.</p>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ima.23184","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142276596","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ConvNext Mixer-Based Encoder Decoder Method for Nuclei Segmentation in Histopathology Images 基于 ConvNext 混合器的编码器解码器方法用于组织病理学图像中的细胞核分割
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-09-21 DOI: 10.1002/ima.23181
Hüseyin Firat, Hüseyin Üzen, Davut Hanbay, Abdulkadir Şengür

Histopathology, vital in diagnosing medical conditions, especially in cancer research, relies on analyzing histopathology images (HIs). Nuclei segmentation, a key task, involves precisely identifying cell nuclei boundaries. Manual segmentation by pathologists is time-consuming, prompting the need for robust automated methods. Challenges in segmentation arise from HI complexities, necessitating advanced techniques. Recent advancements in deep learning, particularly Convolutional Neural Networks (CNNs), have transformed nuclei segmentation. This study emphasizes feature extraction, introducing the ConvNext Mixer-based Encoder-Decoder (CNM-ED) model. Unlike traditional CNN based models, the proposed CNM-ED model enables the extraction of spatial and long context features to address the inherent complexities of histopathology images. This method leverages a multi-path strategy using a traditional CNN architecture as well as different paths focused on obtaining customized long context features using the ConvNext Mixer block structure that combines ConvMixer and ConvNext blocks. The fusion of these diverse features in the final segmentation output enables improved accuracy and performance, surpassing existing state-of-the-art segmentation models. Moreover, our multi-level feature extraction strategy is more effective than models using self-attention mechanisms such as SwinUnet and TransUnet, which have been frequently used in recent years. Experimental studies were conducted using five different datasets (TNBC, MoNuSeg, CoNSeP, CPM17, and CryoNuSeg) to analyze the performance of the proposed CNM-ED model. Comparisons were made with various CNN based models in the literature using evaluation metrics such as accuracy, AJI, macro F1 score, macro intersection over union, macro precision, and macro recall. It was observed that the proposed CNM-ED model achieved highly successful results across all metrics. Through comparisons with state-art-of models from the literature, the proposed CNM-ED model stands out as a promising advancement in nuclei segmentation, addressing the intricacies of histopathological images. The model demonstrates enhanced diagnostic capabilities and holds the potential for significant progress in medical research.

组织病理学是诊断疾病,尤其是癌症研究的重要依据,它依赖于对组织病理学图像(HIs)的分析。细胞核分割是一项关键任务,包括精确识别细胞核边界。病理学家手动分割非常耗时,因此需要强大的自动方法。HI 的复杂性给分割带来了挑战,因此需要先进的技术。深度学习,尤其是卷积神经网络(CNN)的最新进展改变了细胞核分割的方式。本研究强调特征提取,引入了基于 ConvNext 混合器的编码器-解码器(CNM-ED)模型。与传统的基于 CNN 的模型不同,所提出的 CNM-ED 模型能够提取空间和长上下文特征,以解决组织病理学图像固有的复杂性问题。该方法利用传统 CNN 架构的多路径策略,以及使用 ConvMixer 和 ConvNext 块相结合的 ConvNext 混合器块结构获取定制长上下文特征的不同路径。在最终的分割输出中融合这些不同的特征,可以提高准确性和性能,超越现有的最先进分割模型。此外,我们的多层次特征提取策略比近年来经常使用的 SwinUnet 和 TransUnet 等使用自我关注机制的模型更加有效。我们使用五个不同的数据集(TNBC、MoNuSeg、CoNSeP、CPM17 和 CryoNuSeg)进行了实验研究,以分析所提出的 CNM-ED 模型的性能。使用准确率、AJI、宏 F1 分数、宏交集大于联合、宏精确度和宏召回率等评价指标与文献中各种基于 CNN 的模型进行了比较。结果表明,所提出的 CNM-ED 模型在所有指标上都取得了非常成功的结果。通过与文献中的早期模型进行比较,所提出的 CNM-ED 模型在细胞核分割方面取得了巨大进步,解决了组织病理学图像的复杂性问题。该模型增强了诊断能力,有望在医学研究领域取得重大进展。
{"title":"ConvNext Mixer-Based Encoder Decoder Method for Nuclei Segmentation in Histopathology Images","authors":"Hüseyin Firat,&nbsp;Hüseyin Üzen,&nbsp;Davut Hanbay,&nbsp;Abdulkadir Şengür","doi":"10.1002/ima.23181","DOIUrl":"https://doi.org/10.1002/ima.23181","url":null,"abstract":"<p>Histopathology, vital in diagnosing medical conditions, especially in cancer research, relies on analyzing histopathology images (HIs). Nuclei segmentation, a key task, involves precisely identifying cell nuclei boundaries. Manual segmentation by pathologists is time-consuming, prompting the need for robust automated methods. Challenges in segmentation arise from HI complexities, necessitating advanced techniques. Recent advancements in deep learning, particularly Convolutional Neural Networks (CNNs), have transformed nuclei segmentation. This study emphasizes feature extraction, introducing the ConvNext Mixer-based Encoder-Decoder (CNM-ED) model. Unlike traditional CNN based models, the proposed CNM-ED model enables the extraction of spatial and long context features to address the inherent complexities of histopathology images. This method leverages a multi-path strategy using a traditional CNN architecture as well as different paths focused on obtaining customized long context features using the ConvNext Mixer block structure that combines ConvMixer and ConvNext blocks. The fusion of these diverse features in the final segmentation output enables improved accuracy and performance, surpassing existing state-of-the-art segmentation models. Moreover, our multi-level feature extraction strategy is more effective than models using self-attention mechanisms such as SwinUnet and TransUnet, which have been frequently used in recent years. Experimental studies were conducted using five different datasets (TNBC, MoNuSeg, CoNSeP, CPM17, and CryoNuSeg) to analyze the performance of the proposed CNM-ED model. Comparisons were made with various CNN based models in the literature using evaluation metrics such as accuracy, AJI, macro F1 score, macro intersection over union, macro precision, and macro recall. It was observed that the proposed CNM-ED model achieved highly successful results across all metrics. Through comparisons with state-art-of models from the literature, the proposed CNM-ED model stands out as a promising advancement in nuclei segmentation, addressing the intricacies of histopathological images. The model demonstrates enhanced diagnostic capabilities and holds the potential for significant progress in medical research.</p>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ima.23181","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142276595","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhanced Deformation Vector Field Generation With Diffusion Models and Mamba-Based Network for Registration Performance Enhancement 利用扩散模型和基于 Mamba 的网络生成增强型形变矢量场,以提高注册性能
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-09-19 DOI: 10.1002/ima.23171
Zengan Huang, Shan Gao, Xiaxia Yu, Liangjia Zhu, Yi Gao

Recent advancements in deformable image registration (DIR) have seen the emergence of supervised and unsupervised deep learning techniques. However, supervised methods are limited by the quality of deformation vector fields (DVFs), while unsupervised approaches often yield suboptimal results due to their reliance on indirect dissimilarity metrics. Moreover, both methods struggle to effectively model long-range dependencies. This study proposes a novel DIR method that integrates the advantages of supervised and unsupervised learning and tackle issues related to long-range dependencies, thereby improving registration results. Specifically, we propose a DVF generation diffusion model to enhance DVFs diversity, which could be used to facilitate the integration of supervised and unsupervised learning approaches. This fusion allows the method to leverage the benefits of both paradigms. Furthermore, a multi-scale frequency-weighted denoising module is integrated to enhance DVFs generation quality and improve the registration accuracy. Additionally, we propose a novel MambaReg network that adeptly manages long-range dependencies, further optimizing registration outcomes. Experimental evaluation of four public data sets demonstrates that our method outperforms several state-of-the-art techniques based on either supervised or unsupervised learning. Qualitative and quantitative comparisons highlight the superior performance of our approach.

可变形图像配准(DIR)领域的最新进展见证了有监督和无监督深度学习技术的出现。然而,有监督的方法受限于形变向量场(DVF)的质量,而无监督的方法由于依赖于间接的不相似度指标,往往会产生次优结果。此外,这两种方法都难以有效地模拟长程依赖关系。本研究提出了一种新颖的 DIR 方法,它整合了有监督学习和无监督学习的优势,解决了与长距离依赖性相关的问题,从而改善了注册结果。具体来说,我们提出了一种 DVF 生成扩散模型,以增强 DVF 的多样性,从而促进监督和非监督学习方法的融合。这种融合使该方法能够充分利用两种范例的优势。此外,我们还集成了一个多尺度频率加权去噪模块,以提高 DVFs 的生成质量和配准精度。此外,我们还提出了一种新颖的 MambaReg 网络,它能有效管理长距离依赖关系,进一步优化配准结果。对四个公共数据集的实验评估表明,我们的方法优于几种基于监督或非监督学习的先进技术。定性和定量比较凸显了我们方法的卓越性能。
{"title":"Enhanced Deformation Vector Field Generation With Diffusion Models and Mamba-Based Network for Registration Performance Enhancement","authors":"Zengan Huang,&nbsp;Shan Gao,&nbsp;Xiaxia Yu,&nbsp;Liangjia Zhu,&nbsp;Yi Gao","doi":"10.1002/ima.23171","DOIUrl":"https://doi.org/10.1002/ima.23171","url":null,"abstract":"<p>Recent advancements in deformable image registration (DIR) have seen the emergence of supervised and unsupervised deep learning techniques. However, supervised methods are limited by the quality of deformation vector fields (DVFs), while unsupervised approaches often yield suboptimal results due to their reliance on indirect dissimilarity metrics. Moreover, both methods struggle to effectively model long-range dependencies. This study proposes a novel DIR method that integrates the advantages of supervised and unsupervised learning and tackle issues related to long-range dependencies, thereby improving registration results. Specifically, we propose a DVF generation diffusion model to enhance DVFs diversity, which could be used to facilitate the integration of supervised and unsupervised learning approaches. This fusion allows the method to leverage the benefits of both paradigms. Furthermore, a multi-scale frequency-weighted denoising module is integrated to enhance DVFs generation quality and improve the registration accuracy. Additionally, we propose a novel MambaReg network that adeptly manages long-range dependencies, further optimizing registration outcomes. Experimental evaluation of four public data sets demonstrates that our method outperforms several state-of-the-art techniques based on either supervised or unsupervised learning. Qualitative and quantitative comparisons highlight the superior performance of our approach.</p>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ima.23171","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142273012","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Hybrid Convolutional Neural Network Model for the Classification of Multi-Class Skin Cancer 用于多类皮肤癌分类的混合卷积神经网络模型
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-09-19 DOI: 10.1002/ima.23180
Ahmet Nusret Toprak, Ibrahim Aruk

Skin cancer is a significant public health issue, making accurate and early diagnosis crucial. This study proposes a novel and efficient hybrid deep-learning model for accurate skin cancer diagnosis. The model first employs DeepLabV3+ for precise segmentation of skin lesions in dermoscopic images. Feature extraction is then carried out using three pretrained models: MobileNetV2, EfficientNetB0, and DenseNet201 to ensure balanced performance and robust feature learning. These extracted features are then concatenated, and the ReliefF algorithm is employed to select the most relevant features. Finally, obtained features are classified into eight categories: actinic keratosis, basal cell carcinoma, benign keratosis, dermatofibroma, melanoma, melanocytic nevus, squamous cell carcinoma, and vascular lesion using the kNN algorithm. The proposed model achieves an F1 score of 93.49% and an accuracy of 94.42% on the ISIC-2019 dataset, surpassing the best individual model, EfficientNetB0, by 1.20%. Furthermore, the evaluation of the PH2 dataset yielded an F1 score of 94.43% and an accuracy of 94.44%, confirming its generalizability. These findings signify the potential of the proposed model as an expedient, accurate, and valuable tool for early skin cancer detection. They also indicate combining different CNN models achieves superior results over the results obtained from individual models.

皮肤癌是一个重大的公共卫生问题,因此准确和早期诊断至关重要。本研究提出了一种新颖高效的混合深度学习模型,用于准确诊断皮肤癌。该模型首先使用 DeepLabV3+ 对皮肤镜图像中的皮肤病变进行精确分割。然后使用三个预训练模型进行特征提取:MobileNetV2、EfficientNetB0 和 DenseNet201,以确保均衡的性能和稳健的特征学习。然后将这些提取的特征串联起来,并使用 ReliefF 算法来选择最相关的特征。最后,利用 kNN 算法将获得的特征分为八类:光化性角化病、基底细胞癌、良性角化病、皮肤纤维瘤、黑色素瘤、黑素细胞痣、鳞状细胞癌和血管病变。所提出的模型在 ISIC-2019 数据集上的 F1 得分为 93.49%,准确率为 94.42%,比最佳个体模型 EfficientNetB0 高出 1.20%。此外,对 PH2 数据集的评估得出了 94.43% 的 F1 分数和 94.44% 的准确率,证实了其通用性。这些研究结果表明,所提出的模型有潜力成为一种快速、准确、有价值的早期皮肤癌检测工具。这些研究结果还表明,将不同的 CNN 模型结合在一起可获得优于单个模型的结果。
{"title":"A Hybrid Convolutional Neural Network Model for the Classification of Multi-Class Skin Cancer","authors":"Ahmet Nusret Toprak,&nbsp;Ibrahim Aruk","doi":"10.1002/ima.23180","DOIUrl":"https://doi.org/10.1002/ima.23180","url":null,"abstract":"<p>Skin cancer is a significant public health issue, making accurate and early diagnosis crucial. This study proposes a novel and efficient hybrid deep-learning model for accurate skin cancer diagnosis. The model first employs DeepLabV3+ for precise segmentation of skin lesions in dermoscopic images. Feature extraction is then carried out using three pretrained models: MobileNetV2, EfficientNetB0, and DenseNet201 to ensure balanced performance and robust feature learning. These extracted features are then concatenated, and the ReliefF algorithm is employed to select the most relevant features. Finally, obtained features are classified into eight categories: actinic keratosis, basal cell carcinoma, benign keratosis, dermatofibroma, melanoma, melanocytic nevus, squamous cell carcinoma, and vascular lesion using the kNN algorithm. The proposed model achieves an <i>F</i>1 score of 93.49% and an accuracy of 94.42% on the ISIC-2019 dataset, surpassing the best individual model, EfficientNetB0, by 1.20%. Furthermore, the evaluation of the PH2 dataset yielded an <i>F</i>1 score of 94.43% and an accuracy of 94.44%, confirming its generalizability. These findings signify the potential of the proposed model as an expedient, accurate, and valuable tool for early skin cancer detection. They also indicate combining different CNN models achieves superior results over the results obtained from individual models.</p>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ima.23180","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142273010","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deciphering the Complexities of COVID-19-Related Cardiac Complications: Enhancing Classification Accuracy With an Advanced Deep Learning Framework 解密 COVID-19 相关心脏并发症的复杂性:利用先进的深度学习框架提高分类准确性
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-09-19 DOI: 10.1002/ima.23189
Narjes Benameur, Ameni Sassi, Wael Ouarda, Ramzi Mahmoudi, Younes Arous, Mazin Abed Mohammed, Chokri ben Amar, Salam Labidi, Halima Mahjoubi

The literature has widely described the interaction between cardiac complications and COVID-19. However, the diagnosis of cardiac complications caused by COVID-19 using Computed Tomography (CT) images remains a challenge due to the diverse clinical manifestations. To address this issue, this study proposes a novel configuration of Convolutional Neural Network (CNN) for detecting cardiac complications derived from COVID-19 using CT images. The main contribution of this work lies in the use of CNN techniques in combination with Long Short-Term Memory (LSTM) for cardiac complication detection. To explore two-class classification (COVID-19 without cardiac complications vs. COVID-19 with cardiac complications), 10 650 CT images were collected from COVID-19 patients with and without myocardial infarction, myocarditis, and arrhythmia. The information was annotated by two radiology specialists. A total of 0.926 was found to be the accuracy, 0.84 was the recall, 0.82 was the precision, 0.82 was the F1-score, and 0.830 was the g-mean of the suggested model. These results show that the suggested approach can identify cardiac problems from COVID-19 in CT scans. Patients with COVID-19 may benefit from the proposed LSTM-CNN architecture's enhanced ability to identify cardiac problems.

文献广泛描述了心脏并发症与 COVID-19 之间的相互作用。然而,由于临床表现多种多样,使用计算机断层扫描(CT)图像诊断 COVID-19 引起的心脏并发症仍是一项挑战。为解决这一问题,本研究提出了一种新型卷积神经网络(CNN)配置,用于利用 CT 图像检测 COVID-19 引起的心脏并发症。这项工作的主要贡献在于将 CNN 技术与长短期记忆(LSTM)相结合,用于心脏并发症检测。为了探索两类分类(无心脏并发症的 COVID-19 与有心脏并发症的 COVID-19),研究人员从 COVID-19 患者中收集了 10 650 张 CT 图像,这些患者有的患有心肌梗塞,有的没有,有的患有心肌炎,有的患有心律失常。信息由两名放射科专家进行注释。结果发现,建议模型的准确度为 0.926,召回率为 0.84,精确度为 0.82,F1 分数为 0.82,g 均值为 0.830。这些结果表明,所建议的方法可以从 CT 扫描中的 COVID-19 识别心脏问题。建议的 LSTM-CNN 架构增强了识别心脏问题的能力,COVID-19 患者可从中受益。
{"title":"Deciphering the Complexities of COVID-19-Related Cardiac Complications: Enhancing Classification Accuracy With an Advanced Deep Learning Framework","authors":"Narjes Benameur,&nbsp;Ameni Sassi,&nbsp;Wael Ouarda,&nbsp;Ramzi Mahmoudi,&nbsp;Younes Arous,&nbsp;Mazin Abed Mohammed,&nbsp;Chokri ben Amar,&nbsp;Salam Labidi,&nbsp;Halima Mahjoubi","doi":"10.1002/ima.23189","DOIUrl":"https://doi.org/10.1002/ima.23189","url":null,"abstract":"<div>\u0000 \u0000 <p>The literature has widely described the interaction between cardiac complications and COVID-19. However, the diagnosis of cardiac complications caused by COVID-19 using Computed Tomography (CT) images remains a challenge due to the diverse clinical manifestations. To address this issue, this study proposes a novel configuration of Convolutional Neural Network (CNN) for detecting cardiac complications derived from COVID-19 using CT images. The main contribution of this work lies in the use of CNN techniques in combination with Long Short-Term Memory (LSTM) for cardiac complication detection. To explore two-class classification (COVID-19 without cardiac complications vs. COVID-19 with cardiac complications), 10 650 CT images were collected from COVID-19 patients with and without myocardial infarction, myocarditis, and arrhythmia. The information was annotated by two radiology specialists. A total of 0.926 was found to be the accuracy, 0.84 was the recall, 0.82 was the precision, 0.82 was the <i>F</i><sub>1</sub>-score, and 0.830 was the g-mean of the suggested model. These results show that the suggested approach can identify cardiac problems from COVID-19 in CT scans. Patients with COVID-19 may benefit from the proposed LSTM-CNN architecture's enhanced ability to identify cardiac problems.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142273011","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Intervertebral Cervical Disc Intensity (IVCDI) Detection and Classification on MRI Scans Using Deep Learning Methods 使用深度学习方法对磁共振成像扫描进行颈椎椎间盘强度(IVCDI)检测和分类
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-09-16 DOI: 10.1002/ima.23174
M. Fatih Erkoc, Hasan Ulutas, M. Emin Sahin

Radiologists manually interpret magnetic resonance imaging (MRI) scans for the detection of intervertebral cervical disc degeneration, which are often obtained in a primary care or emergency hospital context. The ability of computer models to work with pathological findings and aid in the first interpretation of medical imaging tests is widely acknowledged. Deep learning methods, which are commonly employed today in the diagnosis or detection of many diseases, show great promise in this area. For the detection and segmentation of intervertebral cervical disc intensity, we propose a Mask-RCNN-based deep learning algorithm in this study. The provided approach begins by creating an original dataset using MRI scans that were collected from Yozgat Bozok University. The senior radiologist labels the data, and three classes of intensity are chosen for the classification (low, intermediate, and high). Two alternative network backbones are used in the study, and as a consequence of the training for the Mask R-CNN algorithm, 98.14% and 96.72% mean average precision (mAP) values are obtained with the ResNet50 and ResNet101 architectures, respectively. Utilizing the five-fold cross-validation approach, the study is conducted. This study also applied the Faster R-CNN method, achieving a mAP value of 85.2%. According to the author's knowledge, no study has yet been conducted to apply deep learning algorithms to detect intervertebral cervical disc intensity in a patient population with cervical intervertebral disc degeneration. By ensuring accurate MRI image interpretation and effectively supplying supplementary diagnostic information to provide accuracy and consistency in radiological diagnosis, the proposed method is proving to be a highly useful tool for radiologists.

放射科医生通常在基层医疗机构或医院急诊室通过人工解读磁共振成像(MRI)扫描来检测颈椎间盘变性。计算机模型能够处理病理结果并辅助医学影像检测的首次解读,这一点已得到广泛认可。深度学习方法如今已被广泛应用于多种疾病的诊断或检测,在这一领域大有可为。针对颈椎间盘强度的检测和分割,我们在本研究中提出了一种基于 Mask-RCNN 的深度学习算法。所提供的方法首先使用从尤兹加特博佐克大学收集的核磁共振扫描数据创建原始数据集。资深放射科医生对数据进行标记,并选择三个强度等级(低、中、高)进行分类。研究中使用了两种可供选择的网络骨干,作为掩码 R-CNN 算法的训练结果,ResNet50 和 ResNet101 架构分别获得了 98.14% 和 96.72% 的平均精度 (mAP) 值。研究采用了五倍交叉验证方法。本研究还应用了 Faster R-CNN 方法,获得了 85.2% 的 mAP 值。据笔者所知,目前还没有研究应用深度学习算法检测颈椎间盘退变患者群体的颈椎间盘强度。通过确保核磁共振图像解读的准确性,并有效提供补充诊断信息,从而提供放射诊断的准确性和一致性,所提出的方法被证明是放射科医生的一个非常有用的工具。
{"title":"Intervertebral Cervical Disc Intensity (IVCDI) Detection and Classification on MRI Scans Using Deep Learning Methods","authors":"M. Fatih Erkoc,&nbsp;Hasan Ulutas,&nbsp;M. Emin Sahin","doi":"10.1002/ima.23174","DOIUrl":"https://doi.org/10.1002/ima.23174","url":null,"abstract":"<p>Radiologists manually interpret magnetic resonance imaging (MRI) scans for the detection of intervertebral cervical disc degeneration, which are often obtained in a primary care or emergency hospital context. The ability of computer models to work with pathological findings and aid in the first interpretation of medical imaging tests is widely acknowledged. Deep learning methods, which are commonly employed today in the diagnosis or detection of many diseases, show great promise in this area. For the detection and segmentation of intervertebral cervical disc intensity, we propose a Mask-RCNN-based deep learning algorithm in this study. The provided approach begins by creating an original dataset using MRI scans that were collected from Yozgat Bozok University. The senior radiologist labels the data, and three classes of intensity are chosen for the classification (low, intermediate, and high). Two alternative network backbones are used in the study, and as a consequence of the training for the Mask R-CNN algorithm, 98.14% and 96.72% mean average precision (mAP) values are obtained with the ResNet50 and ResNet101 architectures, respectively. Utilizing the five-fold cross-validation approach, the study is conducted. This study also applied the Faster R-CNN method, achieving a mAP value of 85.2%. According to the author's knowledge, no study has yet been conducted to apply deep learning algorithms to detect intervertebral cervical disc intensity in a patient population with cervical intervertebral disc degeneration. By ensuring accurate MRI image interpretation and effectively supplying supplementary diagnostic information to provide accuracy and consistency in radiological diagnosis, the proposed method is proving to be a highly useful tool for radiologists.</p>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ima.23174","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142234992","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient-Residual Net—A Hybrid Neural Network for Automated Brain Tumor Detection 用于自动脑肿瘤检测的高效-残余网络--混合神经网络
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-09-14 DOI: 10.1002/ima.23170
Jainy Sachdeva, Deepanshu Sharma, Chirag Kamal Ahuja, Arnavdeep Singh

A multiscale feature fusion of Efficient-Residual Net is proposed for classifying tumors on brain Magnetic resonance images with solid or cystic masses, inadequate borders, unpredictable cystic and necrotic regions, and variable heterogeneity. Therefore, in this research, Efficient-Residual Net is proposed by efficaciously amalgamating features of two Deep Convolution Neural Networks—ResNet50 and EffficientNetB0. The skip connections in ResNet50 have reduced the chances of vanishing gradient and overfitting problems considerably thus learning of a higher number of features from input MR images. In addition, EffficientNetB0 uses a compound scaling coefficient for uniformly scaling the dimensions of the network such as depth, width, and resolution. The hybrid model has improved classification results on brain tumors with similar morphology and is tested on three internet repository datasets, namely, Kaggle, BraTS 2018, BraTS 2021, and real-time dataset from Post Graduate Institute of Medical Education and Research (PGIMER), Chandigarh. It is observed that the proposed system delivers an overall accuracy of 96.40%, 97.59%, 97.75%, and 97.99% on the four datasets, respectively. The proposed hybrid methodology has given assuring results of 98%–99% of other statistical such parameters as precision, negatively predicted values, and F1 score. The cloud-based web page is also created using the Django framework in Python programming language for accurate prediction and classification of different types of brain tumors.

针对脑磁共振图像上的肿瘤,如实性或囊性肿块、边界不清、不可预测的囊性和坏死区域以及不同的异质性,提出了Efficient-Residual Net的多尺度特征融合。因此,在这项研究中,通过有效地合并两个深度卷积神经网络--ResNet50 和 EffficientNetB0 的特征,提出了 Efficient-Residual Net。ResNet50 中的跳转连接大大降低了梯度消失和过拟合问题的发生几率,因此能从输入的 MR 图像中学习到更多的特征。此外,EffficientNetB0 还使用了复合缩放系数来统一缩放网络的深度、宽度和分辨率等维度。混合模型改善了对形态相似的脑肿瘤的分类结果,并在三个互联网存储数据集(即 Kaggle、BraTS 2018、BraTS 2021 和来自昌迪加尔医学教育与研究研究生院(PGIMER)的实时数据集)上进行了测试。据观察,拟议系统在四个数据集上的总体准确率分别为 96.40%、97.59%、97.75% 和 97.99%。所提出的混合方法在精确度、负预测值和 F1 分数等其他统计参数上也取得了 98%-99% 的可靠结果。此外,还使用 Python 编程语言中的 Django 框架创建了基于云的网页,用于准确预测和分类不同类型的脑肿瘤。
{"title":"Efficient-Residual Net—A Hybrid Neural Network for Automated Brain Tumor Detection","authors":"Jainy Sachdeva,&nbsp;Deepanshu Sharma,&nbsp;Chirag Kamal Ahuja,&nbsp;Arnavdeep Singh","doi":"10.1002/ima.23170","DOIUrl":"https://doi.org/10.1002/ima.23170","url":null,"abstract":"<div>\u0000 \u0000 <p>A multiscale feature fusion of Efficient-Residual Net is proposed for classifying tumors on brain Magnetic resonance images with solid or cystic masses, inadequate borders, unpredictable cystic and necrotic regions, and variable heterogeneity. Therefore, in this research, Efficient-Residual Net is proposed by efficaciously amalgamating features of two Deep Convolution Neural Networks—ResNet50 and EffficientNetB0. The skip connections in ResNet50 have reduced the chances of vanishing gradient and overfitting problems considerably thus learning of a higher number of features from input MR images. In addition, EffficientNetB0 uses a compound scaling coefficient for uniformly scaling the dimensions of the network such as depth, width, and resolution. The hybrid model has improved classification results on brain tumors with similar morphology and is tested on three internet repository datasets, namely, Kaggle, BraTS 2018, BraTS 2021, and real-time dataset from Post Graduate Institute of Medical Education and Research (PGIMER), Chandigarh. It is observed that the proposed system delivers an overall accuracy of 96.40%, 97.59%, 97.75%, and 97.99% on the four datasets, respectively. The proposed hybrid methodology has given assuring results of 98%–99% of other statistical such parameters as precision, negatively predicted values, and F1 score. The cloud-based web page is also created using the Django framework in Python programming language for accurate prediction and classification of different types of brain tumors.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142233227","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dermo-Optimizer: Skin Lesion Classification Using Information-Theoretic Deep Feature Fusion and Entropy-Controlled Binary Bat Optimization 皮肤优化器:利用信息论深度特征融合和熵控制二元蝙蝠优化进行皮肤病变分类
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-09-14 DOI: 10.1002/ima.23172
Tallha Akram, Anas Alsuhaibani, Muhammad Attique Khan, Sajid Ullah Khan, Syed Rameez Naqvi, Mohsin Bilal

Increases in the prevalence of melanoma, the most lethal form of skin cancer, have been observed over the last few decades. However, the likelihood of a longer life span for individuals is considerably improved with early detection of this malignant illness. Even though the field of computer vision has attained a certain level of success, there is still a degree of ambiguity that represents an unresolved research challenge. In the initial phase of this study, the primary objective is to improve the information derived from input features by combining multiple deep models with the proposed Information-theoretic feature fusion method. Subsequently, in the second phase, the study aims to decrease the redundant and noisy information through down-sampling using the proposed entropy-controlled binary bat selection algorithm. The proposed methodology effectively maintains the integrity of the original feature space, resulting in the creation of highly distinctive feature information. In order to obtain the desired set of features, three contemporary deep models are employed via transfer learning: Inception-Resnet V2, DenseNet-201, and Nasnet Mobile. By combining feature fusion and selection techniques, we may effectively fuse a significant amount of information into the feature vector and subsequently remove any redundant feature information. The effectiveness of the proposed methodology is supported by an evaluation conducted on three well-known dermoscopic datasets, specifically PH2$$ {mathrm{PH}}^2 $$, ISIC-2016, and ISIC-2017. In order to validate the proposed approach, several performance indicators are taken into account, such as accuracy, sensitivity, specificity, false negative rate (FNR), false positive rate (FPR), and F1-score. The accuracies obtained for all datasets utilizing the proposed methodology are 99.05%, 96.26%, and 95.71%, respectively.

黑色素瘤是最致命的皮肤癌,在过去几十年中,黑色素瘤的发病率不断上升。然而,如果能及早发现这种恶性疾病,就能大大延长患者的寿命。尽管计算机视觉领域已经取得了一定的成就,但仍然存在一定程度的模糊性,这是一个尚未解决的研究难题。在本研究的初始阶段,主要目标是通过将多个深度模型与所提出的信息论特征融合方法相结合,改进从输入特征中获得的信息。随后,在第二阶段,本研究旨在利用所提出的熵控制二元蝙蝠选择算法,通过向下采样来减少冗余和噪声信息。所提出的方法有效地保持了原始特征空间的完整性,从而创建了高度独特的特征信息。为了获得所需的特征集,我们通过迁移学习采用了三种当代深度模型:Inception-Resnet V2、DenseNet-201 和 Nasnet Mobile。通过将特征融合与选择技术相结合,我们可以有效地将大量信息融合到特征向量中,并随后去除任何冗余特征信息。我们在三个著名的皮肤镜数据集(特别是 PH 2 $$ {mathrm{PH}}^2 $$、ISIC-2016 和 ISIC-2017)上进行了评估,证明了所提方法的有效性。为了验证所提出的方法,考虑了几个性能指标,如准确度、灵敏度、特异性、假阴性率(FNR)、假阳性率(FPR)和 F1 分数。采用所提方法的所有数据集的准确率分别为 99.05%、96.26% 和 95.71%。
{"title":"Dermo-Optimizer: Skin Lesion Classification Using Information-Theoretic Deep Feature Fusion and Entropy-Controlled Binary Bat Optimization","authors":"Tallha Akram,&nbsp;Anas Alsuhaibani,&nbsp;Muhammad Attique Khan,&nbsp;Sajid Ullah Khan,&nbsp;Syed Rameez Naqvi,&nbsp;Mohsin Bilal","doi":"10.1002/ima.23172","DOIUrl":"https://doi.org/10.1002/ima.23172","url":null,"abstract":"<div>\u0000 \u0000 <p>Increases in the prevalence of melanoma, the most lethal form of skin cancer, have been observed over the last few decades. However, the likelihood of a longer life span for individuals is considerably improved with early detection of this malignant illness. Even though the field of computer vision has attained a certain level of success, there is still a degree of ambiguity that represents an unresolved research challenge. In the initial phase of this study, the primary objective is to improve the information derived from input features by combining multiple deep models with the proposed Information-theoretic feature fusion method. Subsequently, in the second phase, the study aims to decrease the redundant and noisy information through down-sampling using the proposed entropy-controlled binary bat selection algorithm. The proposed methodology effectively maintains the integrity of the original feature space, resulting in the creation of highly distinctive feature information. In order to obtain the desired set of features, three contemporary deep models are employed via transfer learning: Inception-Resnet V2, DenseNet-201, and Nasnet Mobile. By combining feature fusion and selection techniques, we may effectively fuse a significant amount of information into the feature vector and subsequently remove any redundant feature information. The effectiveness of the proposed methodology is supported by an evaluation conducted on three well-known dermoscopic datasets, specifically <span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 <msup>\u0000 <mi>PH</mi>\u0000 <mn>2</mn>\u0000 </msup>\u0000 </mrow>\u0000 <annotation>$$ {mathrm{PH}}^2 $$</annotation>\u0000 </semantics></math>, ISIC-2016, and ISIC-2017. In order to validate the proposed approach, several performance indicators are taken into account, such as accuracy, sensitivity, specificity, false negative rate (FNR), false positive rate (FPR), and F1-score. The accuracies obtained for all datasets utilizing the proposed methodology are 99.05%, 96.26%, and 95.71%, respectively.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142233204","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Novel Perceptual Constrained cycleGAN With Attention Mechanisms for Unsupervised MR-to-CT Synthesis 用于无监督 MR-CT 合成的具有注意机制的新型感知受限循环基因组学模型
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-09-14 DOI: 10.1002/ima.23169
Ruiming Zhu, Xinliang Liu, Mingrui Li, Wei Qian, Yueyang Teng

Radiotherapy treatment planning (RTP) requires both magnetic resonance (MR) and computed tomography (CT) modalities. However, conducting separate MR and CT scans for patients leads to misalignment, increased radiation exposure, and higher costs. To address these challenges and mitigate the limitations of supervised synthesis methods, we propose a novel unsupervised perceptual attention image synthesis model based on cycleGAN (PA-cycleGAN). The innovation of PA-cycleGAN lies in its model structure, which incorporates dynamic feature encoding and deep feature extraction to improve the understanding of image structure and contextual information. To ensure the visual authenticity of the synthetic images, we design a hybrid loss function that incorporates perceptual constraints using high-level features extracted by deep neural networks. Our PA-cycleGAN achieves notable results, with an average peak signal-to-noise ratio (PSNR) of 28.06, structural similarity (SSIM) of 0.95, and mean absolute error (MAE) of 46.90 on a pelvic dataset. Additionally, we validate the generalization of our method by conducting experiments on an additional head dataset. These experiments demonstrate that PA-cycleGAN consistently outperforms other state-of-the-art methods in both quantitative metrics and image synthesis quality.

放疗治疗计划(RTP)需要磁共振(MR)和计算机断层扫描(CT)两种模式。然而,对患者分别进行磁共振和 CT 扫描会导致对位错误、辐照增加和成本上升。为了应对这些挑战并减少监督合成方法的局限性,我们提出了一种基于循环广义注视模型(PA-cycleGAN)的新型无监督感知注视图像合成模型。PA-cycleGAN 的创新之处在于其模型结构,它结合了动态特征编码和深度特征提取,以提高对图像结构和上下文信息的理解。为了确保合成图像的视觉真实性,我们设计了一种混合损失函数,利用深度神经网络提取的高级特征,将感知约束条件纳入其中。我们的 PA-cycleGAN 取得了显著的成果,在骨盆数据集上的平均峰值信噪比(PSNR)为 28.06,结构相似度(SSIM)为 0.95,平均绝对误差(MAE)为 46.90。此外,我们还在另外一个头部数据集上进行了实验,验证了我们方法的通用性。这些实验表明,PA-cycleGAN 在定量指标和图像合成质量方面始终优于其他最先进的方法。
{"title":"A Novel Perceptual Constrained cycleGAN With Attention Mechanisms for Unsupervised MR-to-CT Synthesis","authors":"Ruiming Zhu,&nbsp;Xinliang Liu,&nbsp;Mingrui Li,&nbsp;Wei Qian,&nbsp;Yueyang Teng","doi":"10.1002/ima.23169","DOIUrl":"https://doi.org/10.1002/ima.23169","url":null,"abstract":"<div>\u0000 \u0000 <p>Radiotherapy treatment planning (RTP) requires both magnetic resonance (MR) and computed tomography (CT) modalities. However, conducting separate MR and CT scans for patients leads to misalignment, increased radiation exposure, and higher costs. To address these challenges and mitigate the limitations of supervised synthesis methods, we propose a novel unsupervised perceptual attention image synthesis model based on cycleGAN (PA-cycleGAN). The innovation of PA-cycleGAN lies in its model structure, which incorporates dynamic feature encoding and deep feature extraction to improve the understanding of image structure and contextual information. To ensure the visual authenticity of the synthetic images, we design a hybrid loss function that incorporates perceptual constraints using high-level features extracted by deep neural networks. Our PA-cycleGAN achieves notable results, with an average peak signal-to-noise ratio (PSNR) of 28.06, structural similarity (SSIM) of 0.95, and mean absolute error (MAE) of 46.90 on a pelvic dataset. Additionally, we validate the generalization of our method by conducting experiments on an additional head dataset. These experiments demonstrate that PA-cycleGAN consistently outperforms other state-of-the-art methods in both quantitative metrics and image synthesis quality.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142233208","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SSANet—Novel Residual Network for Computer-Aided Diagnosis of Pulmonary Nodules in Chest Computed Tomography 用于胸部计算机断层扫描肺结节计算机辅助诊断的 SSANet-Novel 残差网络
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-09-14 DOI: 10.1002/ima.23176
Yu Gu, Jiaqi Liu, Lidong Yang, Baohua Zhang, Jing Wang, Xiaoqi Lu, Jianjun Li, Xin Liu, Dahua Yu, Ying Zhao, Siyuan Tang, Qun He

The manifestations of early lung cancer in medical imaging often appear as pulmonary nodules, which can be classified as benign or malignant. In recent years, there has been a gradual application of deep learning-based computer-aided diagnosis technology to assist in the diagnosis of lung nodules. This study introduces a novel three-dimensional (3D) residual network called SSANet, which integrates split-based convolution, shuffle attention, and a novel activation function. The aim is to enhance the accuracy of distinguishing between benign and malignant lung nodules using convolutional neural networks (CNNs) and alleviate the burden on doctors when interpreting the images. To fully extract pulmonary nodule information from chest CT images, the original residual network is expanded into a 3D CNN structure. Additionally, a 3D split-based convolutional operation (SPConv) is designed and integrated into the feature extraction module to reduce redundancy in feature maps and improve network inference speed. In the SSABlock part of the proposed network, ACON (Activated or Not) function is also introduced. The proposed SSANet also incorporates an attention module to capture critical characteristics of lung nodules. During the training process, the PolyLoss function is utilized. Once SSANet generates the diagnosis result, a heatmap displays using Score-CAM is employed to evaluate whether the network accurately identifies the location of lung nodules. In the final test set, the proposed network achieves an accuracy of 89.13%, an F1-score of 84.85%, and a G-mean of 86.20%. These metrics represent improvements of 5.43%, 5.98%, and 4.09%, respectively, compared with the original base network. The experimental results align with those of previous studies on pulmonary nodule diagnosis networks, confirming the reliability and clinical applicability of the diagnostic outcomes.

早期肺癌在医学影像中的表现往往是肺部结节,可分为良性和恶性。近年来,基于深度学习的计算机辅助诊断技术逐渐应用于肺结节的辅助诊断。本研究介绍了一种名为 SSANet 的新型三维(3D)残差网络,它集成了基于分裂的卷积、洗牌注意和新型激活函数。其目的是利用卷积神经网络(CNN)提高区分肺结节良性和恶性的准确性,并减轻医生判读图像的负担。为了从胸部 CT 图像中充分提取肺结节信息,原始残差网络被扩展为三维卷积神经网络结构。此外,还设计了一种基于三维分裂的卷积运算(SPConv),并将其集成到特征提取模块中,以减少特征图中的冗余,提高网络推理速度。在拟议网络的 SSABlock 部分,还引入了 ACON(激活或未激活)功能。拟议的 SSANet 还加入了注意力模块,以捕捉肺结节的关键特征。在训练过程中,使用了 PolyLoss 函数。SSANet 生成诊断结果后,将使用 Score-CAM 进行热图显示,以评估网络是否能准确识别肺结节的位置。在最终测试集中,建议的网络达到了 89.13% 的准确率、84.85% 的 F1 分数和 86.20% 的 G 平均值。与原始基础网络相比,这些指标分别提高了 5.43%、5.98% 和 4.09%。实验结果与之前关于肺结节诊断网络的研究结果一致,证实了诊断结果的可靠性和临床适用性。
{"title":"SSANet—Novel Residual Network for Computer-Aided Diagnosis of Pulmonary Nodules in Chest Computed Tomography","authors":"Yu Gu,&nbsp;Jiaqi Liu,&nbsp;Lidong Yang,&nbsp;Baohua Zhang,&nbsp;Jing Wang,&nbsp;Xiaoqi Lu,&nbsp;Jianjun Li,&nbsp;Xin Liu,&nbsp;Dahua Yu,&nbsp;Ying Zhao,&nbsp;Siyuan Tang,&nbsp;Qun He","doi":"10.1002/ima.23176","DOIUrl":"https://doi.org/10.1002/ima.23176","url":null,"abstract":"<div>\u0000 \u0000 <p>The manifestations of early lung cancer in medical imaging often appear as pulmonary nodules, which can be classified as benign or malignant. In recent years, there has been a gradual application of deep learning-based computer-aided diagnosis technology to assist in the diagnosis of lung nodules. This study introduces a novel three-dimensional (3D) residual network called SSANet, which integrates split-based convolution, shuffle attention, and a novel activation function. The aim is to enhance the accuracy of distinguishing between benign and malignant lung nodules using convolutional neural networks (CNNs) and alleviate the burden on doctors when interpreting the images. To fully extract pulmonary nodule information from chest CT images, the original residual network is expanded into a 3D CNN structure. Additionally, a 3D split-based convolutional operation (SPConv) is designed and integrated into the feature extraction module to reduce redundancy in feature maps and improve network inference speed. In the SSABlock part of the proposed network, ACON (Activated or Not) function is also introduced. The proposed SSANet also incorporates an attention module to capture critical characteristics of lung nodules. During the training process, the PolyLoss function is utilized. Once SSANet generates the diagnosis result, a heatmap displays using Score-CAM is employed to evaluate whether the network accurately identifies the location of lung nodules. In the final test set, the proposed network achieves an accuracy of 89.13%, an F1-score of 84.85%, and a G-mean of 86.20%. These metrics represent improvements of 5.43%, 5.98%, and 4.09%, respectively, compared with the original base network. The experimental results align with those of previous studies on pulmonary nodule diagnosis networks, confirming the reliability and clinical applicability of the diagnostic outcomes.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142233214","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International Journal of Imaging Systems and Technology
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1