首页 > 最新文献

International Journal of Imaging Systems and Technology最新文献

英文 中文
A Hybrid Convolutional Neural Network Model for the Classification of Multi-Class Skin Cancer 用于多类皮肤癌分类的混合卷积神经网络模型
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-09-19 DOI: 10.1002/ima.23180
Ahmet Nusret Toprak, Ibrahim Aruk

Skin cancer is a significant public health issue, making accurate and early diagnosis crucial. This study proposes a novel and efficient hybrid deep-learning model for accurate skin cancer diagnosis. The model first employs DeepLabV3+ for precise segmentation of skin lesions in dermoscopic images. Feature extraction is then carried out using three pretrained models: MobileNetV2, EfficientNetB0, and DenseNet201 to ensure balanced performance and robust feature learning. These extracted features are then concatenated, and the ReliefF algorithm is employed to select the most relevant features. Finally, obtained features are classified into eight categories: actinic keratosis, basal cell carcinoma, benign keratosis, dermatofibroma, melanoma, melanocytic nevus, squamous cell carcinoma, and vascular lesion using the kNN algorithm. The proposed model achieves an F1 score of 93.49% and an accuracy of 94.42% on the ISIC-2019 dataset, surpassing the best individual model, EfficientNetB0, by 1.20%. Furthermore, the evaluation of the PH2 dataset yielded an F1 score of 94.43% and an accuracy of 94.44%, confirming its generalizability. These findings signify the potential of the proposed model as an expedient, accurate, and valuable tool for early skin cancer detection. They also indicate combining different CNN models achieves superior results over the results obtained from individual models.

皮肤癌是一个重大的公共卫生问题,因此准确和早期诊断至关重要。本研究提出了一种新颖高效的混合深度学习模型,用于准确诊断皮肤癌。该模型首先使用 DeepLabV3+ 对皮肤镜图像中的皮肤病变进行精确分割。然后使用三个预训练模型进行特征提取:MobileNetV2、EfficientNetB0 和 DenseNet201,以确保均衡的性能和稳健的特征学习。然后将这些提取的特征串联起来,并使用 ReliefF 算法来选择最相关的特征。最后,利用 kNN 算法将获得的特征分为八类:光化性角化病、基底细胞癌、良性角化病、皮肤纤维瘤、黑色素瘤、黑素细胞痣、鳞状细胞癌和血管病变。所提出的模型在 ISIC-2019 数据集上的 F1 得分为 93.49%,准确率为 94.42%,比最佳个体模型 EfficientNetB0 高出 1.20%。此外,对 PH2 数据集的评估得出了 94.43% 的 F1 分数和 94.44% 的准确率,证实了其通用性。这些研究结果表明,所提出的模型有潜力成为一种快速、准确、有价值的早期皮肤癌检测工具。这些研究结果还表明,将不同的 CNN 模型结合在一起可获得优于单个模型的结果。
{"title":"A Hybrid Convolutional Neural Network Model for the Classification of Multi-Class Skin Cancer","authors":"Ahmet Nusret Toprak,&nbsp;Ibrahim Aruk","doi":"10.1002/ima.23180","DOIUrl":"https://doi.org/10.1002/ima.23180","url":null,"abstract":"<p>Skin cancer is a significant public health issue, making accurate and early diagnosis crucial. This study proposes a novel and efficient hybrid deep-learning model for accurate skin cancer diagnosis. The model first employs DeepLabV3+ for precise segmentation of skin lesions in dermoscopic images. Feature extraction is then carried out using three pretrained models: MobileNetV2, EfficientNetB0, and DenseNet201 to ensure balanced performance and robust feature learning. These extracted features are then concatenated, and the ReliefF algorithm is employed to select the most relevant features. Finally, obtained features are classified into eight categories: actinic keratosis, basal cell carcinoma, benign keratosis, dermatofibroma, melanoma, melanocytic nevus, squamous cell carcinoma, and vascular lesion using the kNN algorithm. The proposed model achieves an <i>F</i>1 score of 93.49% and an accuracy of 94.42% on the ISIC-2019 dataset, surpassing the best individual model, EfficientNetB0, by 1.20%. Furthermore, the evaluation of the PH2 dataset yielded an <i>F</i>1 score of 94.43% and an accuracy of 94.44%, confirming its generalizability. These findings signify the potential of the proposed model as an expedient, accurate, and valuable tool for early skin cancer detection. They also indicate combining different CNN models achieves superior results over the results obtained from individual models.</p>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ima.23180","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142273010","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deciphering the Complexities of COVID-19-Related Cardiac Complications: Enhancing Classification Accuracy With an Advanced Deep Learning Framework 解密 COVID-19 相关心脏并发症的复杂性:利用先进的深度学习框架提高分类准确性
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-09-19 DOI: 10.1002/ima.23189
Narjes Benameur, Ameni Sassi, Wael Ouarda, Ramzi Mahmoudi, Younes Arous, Mazin Abed Mohammed, Chokri ben Amar, Salam Labidi, Halima Mahjoubi

The literature has widely described the interaction between cardiac complications and COVID-19. However, the diagnosis of cardiac complications caused by COVID-19 using Computed Tomography (CT) images remains a challenge due to the diverse clinical manifestations. To address this issue, this study proposes a novel configuration of Convolutional Neural Network (CNN) for detecting cardiac complications derived from COVID-19 using CT images. The main contribution of this work lies in the use of CNN techniques in combination with Long Short-Term Memory (LSTM) for cardiac complication detection. To explore two-class classification (COVID-19 without cardiac complications vs. COVID-19 with cardiac complications), 10 650 CT images were collected from COVID-19 patients with and without myocardial infarction, myocarditis, and arrhythmia. The information was annotated by two radiology specialists. A total of 0.926 was found to be the accuracy, 0.84 was the recall, 0.82 was the precision, 0.82 was the F1-score, and 0.830 was the g-mean of the suggested model. These results show that the suggested approach can identify cardiac problems from COVID-19 in CT scans. Patients with COVID-19 may benefit from the proposed LSTM-CNN architecture's enhanced ability to identify cardiac problems.

文献广泛描述了心脏并发症与 COVID-19 之间的相互作用。然而,由于临床表现多种多样,使用计算机断层扫描(CT)图像诊断 COVID-19 引起的心脏并发症仍是一项挑战。为解决这一问题,本研究提出了一种新型卷积神经网络(CNN)配置,用于利用 CT 图像检测 COVID-19 引起的心脏并发症。这项工作的主要贡献在于将 CNN 技术与长短期记忆(LSTM)相结合,用于心脏并发症检测。为了探索两类分类(无心脏并发症的 COVID-19 与有心脏并发症的 COVID-19),研究人员从 COVID-19 患者中收集了 10 650 张 CT 图像,这些患者有的患有心肌梗塞,有的没有,有的患有心肌炎,有的患有心律失常。信息由两名放射科专家进行注释。结果发现,建议模型的准确度为 0.926,召回率为 0.84,精确度为 0.82,F1 分数为 0.82,g 均值为 0.830。这些结果表明,所建议的方法可以从 CT 扫描中的 COVID-19 识别心脏问题。建议的 LSTM-CNN 架构增强了识别心脏问题的能力,COVID-19 患者可从中受益。
{"title":"Deciphering the Complexities of COVID-19-Related Cardiac Complications: Enhancing Classification Accuracy With an Advanced Deep Learning Framework","authors":"Narjes Benameur,&nbsp;Ameni Sassi,&nbsp;Wael Ouarda,&nbsp;Ramzi Mahmoudi,&nbsp;Younes Arous,&nbsp;Mazin Abed Mohammed,&nbsp;Chokri ben Amar,&nbsp;Salam Labidi,&nbsp;Halima Mahjoubi","doi":"10.1002/ima.23189","DOIUrl":"https://doi.org/10.1002/ima.23189","url":null,"abstract":"<div>\u0000 \u0000 <p>The literature has widely described the interaction between cardiac complications and COVID-19. However, the diagnosis of cardiac complications caused by COVID-19 using Computed Tomography (CT) images remains a challenge due to the diverse clinical manifestations. To address this issue, this study proposes a novel configuration of Convolutional Neural Network (CNN) for detecting cardiac complications derived from COVID-19 using CT images. The main contribution of this work lies in the use of CNN techniques in combination with Long Short-Term Memory (LSTM) for cardiac complication detection. To explore two-class classification (COVID-19 without cardiac complications vs. COVID-19 with cardiac complications), 10 650 CT images were collected from COVID-19 patients with and without myocardial infarction, myocarditis, and arrhythmia. The information was annotated by two radiology specialists. A total of 0.926 was found to be the accuracy, 0.84 was the recall, 0.82 was the precision, 0.82 was the <i>F</i><sub>1</sub>-score, and 0.830 was the g-mean of the suggested model. These results show that the suggested approach can identify cardiac problems from COVID-19 in CT scans. Patients with COVID-19 may benefit from the proposed LSTM-CNN architecture's enhanced ability to identify cardiac problems.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142273011","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Intervertebral Cervical Disc Intensity (IVCDI) Detection and Classification on MRI Scans Using Deep Learning Methods 使用深度学习方法对磁共振成像扫描进行颈椎椎间盘强度(IVCDI)检测和分类
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-09-16 DOI: 10.1002/ima.23174
M. Fatih Erkoc, Hasan Ulutas, M. Emin Sahin

Radiologists manually interpret magnetic resonance imaging (MRI) scans for the detection of intervertebral cervical disc degeneration, which are often obtained in a primary care or emergency hospital context. The ability of computer models to work with pathological findings and aid in the first interpretation of medical imaging tests is widely acknowledged. Deep learning methods, which are commonly employed today in the diagnosis or detection of many diseases, show great promise in this area. For the detection and segmentation of intervertebral cervical disc intensity, we propose a Mask-RCNN-based deep learning algorithm in this study. The provided approach begins by creating an original dataset using MRI scans that were collected from Yozgat Bozok University. The senior radiologist labels the data, and three classes of intensity are chosen for the classification (low, intermediate, and high). Two alternative network backbones are used in the study, and as a consequence of the training for the Mask R-CNN algorithm, 98.14% and 96.72% mean average precision (mAP) values are obtained with the ResNet50 and ResNet101 architectures, respectively. Utilizing the five-fold cross-validation approach, the study is conducted. This study also applied the Faster R-CNN method, achieving a mAP value of 85.2%. According to the author's knowledge, no study has yet been conducted to apply deep learning algorithms to detect intervertebral cervical disc intensity in a patient population with cervical intervertebral disc degeneration. By ensuring accurate MRI image interpretation and effectively supplying supplementary diagnostic information to provide accuracy and consistency in radiological diagnosis, the proposed method is proving to be a highly useful tool for radiologists.

放射科医生通常在基层医疗机构或医院急诊室通过人工解读磁共振成像(MRI)扫描来检测颈椎间盘变性。计算机模型能够处理病理结果并辅助医学影像检测的首次解读,这一点已得到广泛认可。深度学习方法如今已被广泛应用于多种疾病的诊断或检测,在这一领域大有可为。针对颈椎间盘强度的检测和分割,我们在本研究中提出了一种基于 Mask-RCNN 的深度学习算法。所提供的方法首先使用从尤兹加特博佐克大学收集的核磁共振扫描数据创建原始数据集。资深放射科医生对数据进行标记,并选择三个强度等级(低、中、高)进行分类。研究中使用了两种可供选择的网络骨干,作为掩码 R-CNN 算法的训练结果,ResNet50 和 ResNet101 架构分别获得了 98.14% 和 96.72% 的平均精度 (mAP) 值。研究采用了五倍交叉验证方法。本研究还应用了 Faster R-CNN 方法,获得了 85.2% 的 mAP 值。据笔者所知,目前还没有研究应用深度学习算法检测颈椎间盘退变患者群体的颈椎间盘强度。通过确保核磁共振图像解读的准确性,并有效提供补充诊断信息,从而提供放射诊断的准确性和一致性,所提出的方法被证明是放射科医生的一个非常有用的工具。
{"title":"Intervertebral Cervical Disc Intensity (IVCDI) Detection and Classification on MRI Scans Using Deep Learning Methods","authors":"M. Fatih Erkoc,&nbsp;Hasan Ulutas,&nbsp;M. Emin Sahin","doi":"10.1002/ima.23174","DOIUrl":"https://doi.org/10.1002/ima.23174","url":null,"abstract":"<p>Radiologists manually interpret magnetic resonance imaging (MRI) scans for the detection of intervertebral cervical disc degeneration, which are often obtained in a primary care or emergency hospital context. The ability of computer models to work with pathological findings and aid in the first interpretation of medical imaging tests is widely acknowledged. Deep learning methods, which are commonly employed today in the diagnosis or detection of many diseases, show great promise in this area. For the detection and segmentation of intervertebral cervical disc intensity, we propose a Mask-RCNN-based deep learning algorithm in this study. The provided approach begins by creating an original dataset using MRI scans that were collected from Yozgat Bozok University. The senior radiologist labels the data, and three classes of intensity are chosen for the classification (low, intermediate, and high). Two alternative network backbones are used in the study, and as a consequence of the training for the Mask R-CNN algorithm, 98.14% and 96.72% mean average precision (mAP) values are obtained with the ResNet50 and ResNet101 architectures, respectively. Utilizing the five-fold cross-validation approach, the study is conducted. This study also applied the Faster R-CNN method, achieving a mAP value of 85.2%. According to the author's knowledge, no study has yet been conducted to apply deep learning algorithms to detect intervertebral cervical disc intensity in a patient population with cervical intervertebral disc degeneration. By ensuring accurate MRI image interpretation and effectively supplying supplementary diagnostic information to provide accuracy and consistency in radiological diagnosis, the proposed method is proving to be a highly useful tool for radiologists.</p>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ima.23174","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142234992","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient-Residual Net—A Hybrid Neural Network for Automated Brain Tumor Detection 用于自动脑肿瘤检测的高效-残余网络--混合神经网络
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-09-14 DOI: 10.1002/ima.23170
Jainy Sachdeva, Deepanshu Sharma, Chirag Kamal Ahuja, Arnavdeep Singh

A multiscale feature fusion of Efficient-Residual Net is proposed for classifying tumors on brain Magnetic resonance images with solid or cystic masses, inadequate borders, unpredictable cystic and necrotic regions, and variable heterogeneity. Therefore, in this research, Efficient-Residual Net is proposed by efficaciously amalgamating features of two Deep Convolution Neural Networks—ResNet50 and EffficientNetB0. The skip connections in ResNet50 have reduced the chances of vanishing gradient and overfitting problems considerably thus learning of a higher number of features from input MR images. In addition, EffficientNetB0 uses a compound scaling coefficient for uniformly scaling the dimensions of the network such as depth, width, and resolution. The hybrid model has improved classification results on brain tumors with similar morphology and is tested on three internet repository datasets, namely, Kaggle, BraTS 2018, BraTS 2021, and real-time dataset from Post Graduate Institute of Medical Education and Research (PGIMER), Chandigarh. It is observed that the proposed system delivers an overall accuracy of 96.40%, 97.59%, 97.75%, and 97.99% on the four datasets, respectively. The proposed hybrid methodology has given assuring results of 98%–99% of other statistical such parameters as precision, negatively predicted values, and F1 score. The cloud-based web page is also created using the Django framework in Python programming language for accurate prediction and classification of different types of brain tumors.

针对脑磁共振图像上的肿瘤,如实性或囊性肿块、边界不清、不可预测的囊性和坏死区域以及不同的异质性,提出了Efficient-Residual Net的多尺度特征融合。因此,在这项研究中,通过有效地合并两个深度卷积神经网络--ResNet50 和 EffficientNetB0 的特征,提出了 Efficient-Residual Net。ResNet50 中的跳转连接大大降低了梯度消失和过拟合问题的发生几率,因此能从输入的 MR 图像中学习到更多的特征。此外,EffficientNetB0 还使用了复合缩放系数来统一缩放网络的深度、宽度和分辨率等维度。混合模型改善了对形态相似的脑肿瘤的分类结果,并在三个互联网存储数据集(即 Kaggle、BraTS 2018、BraTS 2021 和来自昌迪加尔医学教育与研究研究生院(PGIMER)的实时数据集)上进行了测试。据观察,拟议系统在四个数据集上的总体准确率分别为 96.40%、97.59%、97.75% 和 97.99%。所提出的混合方法在精确度、负预测值和 F1 分数等其他统计参数上也取得了 98%-99% 的可靠结果。此外,还使用 Python 编程语言中的 Django 框架创建了基于云的网页,用于准确预测和分类不同类型的脑肿瘤。
{"title":"Efficient-Residual Net—A Hybrid Neural Network for Automated Brain Tumor Detection","authors":"Jainy Sachdeva,&nbsp;Deepanshu Sharma,&nbsp;Chirag Kamal Ahuja,&nbsp;Arnavdeep Singh","doi":"10.1002/ima.23170","DOIUrl":"https://doi.org/10.1002/ima.23170","url":null,"abstract":"<div>\u0000 \u0000 <p>A multiscale feature fusion of Efficient-Residual Net is proposed for classifying tumors on brain Magnetic resonance images with solid or cystic masses, inadequate borders, unpredictable cystic and necrotic regions, and variable heterogeneity. Therefore, in this research, Efficient-Residual Net is proposed by efficaciously amalgamating features of two Deep Convolution Neural Networks—ResNet50 and EffficientNetB0. The skip connections in ResNet50 have reduced the chances of vanishing gradient and overfitting problems considerably thus learning of a higher number of features from input MR images. In addition, EffficientNetB0 uses a compound scaling coefficient for uniformly scaling the dimensions of the network such as depth, width, and resolution. The hybrid model has improved classification results on brain tumors with similar morphology and is tested on three internet repository datasets, namely, Kaggle, BraTS 2018, BraTS 2021, and real-time dataset from Post Graduate Institute of Medical Education and Research (PGIMER), Chandigarh. It is observed that the proposed system delivers an overall accuracy of 96.40%, 97.59%, 97.75%, and 97.99% on the four datasets, respectively. The proposed hybrid methodology has given assuring results of 98%–99% of other statistical such parameters as precision, negatively predicted values, and F1 score. The cloud-based web page is also created using the Django framework in Python programming language for accurate prediction and classification of different types of brain tumors.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142233227","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dermo-Optimizer: Skin Lesion Classification Using Information-Theoretic Deep Feature Fusion and Entropy-Controlled Binary Bat Optimization 皮肤优化器:利用信息论深度特征融合和熵控制二元蝙蝠优化进行皮肤病变分类
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-09-14 DOI: 10.1002/ima.23172
Tallha Akram, Anas Alsuhaibani, Muhammad Attique Khan, Sajid Ullah Khan, Syed Rameez Naqvi, Mohsin Bilal

Increases in the prevalence of melanoma, the most lethal form of skin cancer, have been observed over the last few decades. However, the likelihood of a longer life span for individuals is considerably improved with early detection of this malignant illness. Even though the field of computer vision has attained a certain level of success, there is still a degree of ambiguity that represents an unresolved research challenge. In the initial phase of this study, the primary objective is to improve the information derived from input features by combining multiple deep models with the proposed Information-theoretic feature fusion method. Subsequently, in the second phase, the study aims to decrease the redundant and noisy information through down-sampling using the proposed entropy-controlled binary bat selection algorithm. The proposed methodology effectively maintains the integrity of the original feature space, resulting in the creation of highly distinctive feature information. In order to obtain the desired set of features, three contemporary deep models are employed via transfer learning: Inception-Resnet V2, DenseNet-201, and Nasnet Mobile. By combining feature fusion and selection techniques, we may effectively fuse a significant amount of information into the feature vector and subsequently remove any redundant feature information. The effectiveness of the proposed methodology is supported by an evaluation conducted on three well-known dermoscopic datasets, specifically PH2$$ {mathrm{PH}}^2 $$, ISIC-2016, and ISIC-2017. In order to validate the proposed approach, several performance indicators are taken into account, such as accuracy, sensitivity, specificity, false negative rate (FNR), false positive rate (FPR), and F1-score. The accuracies obtained for all datasets utilizing the proposed methodology are 99.05%, 96.26%, and 95.71%, respectively.

黑色素瘤是最致命的皮肤癌,在过去几十年中,黑色素瘤的发病率不断上升。然而,如果能及早发现这种恶性疾病,就能大大延长患者的寿命。尽管计算机视觉领域已经取得了一定的成就,但仍然存在一定程度的模糊性,这是一个尚未解决的研究难题。在本研究的初始阶段,主要目标是通过将多个深度模型与所提出的信息论特征融合方法相结合,改进从输入特征中获得的信息。随后,在第二阶段,本研究旨在利用所提出的熵控制二元蝙蝠选择算法,通过向下采样来减少冗余和噪声信息。所提出的方法有效地保持了原始特征空间的完整性,从而创建了高度独特的特征信息。为了获得所需的特征集,我们通过迁移学习采用了三种当代深度模型:Inception-Resnet V2、DenseNet-201 和 Nasnet Mobile。通过将特征融合与选择技术相结合,我们可以有效地将大量信息融合到特征向量中,并随后去除任何冗余特征信息。我们在三个著名的皮肤镜数据集(特别是 PH 2 $$ {mathrm{PH}}^2 $$、ISIC-2016 和 ISIC-2017)上进行了评估,证明了所提方法的有效性。为了验证所提出的方法,考虑了几个性能指标,如准确度、灵敏度、特异性、假阴性率(FNR)、假阳性率(FPR)和 F1 分数。采用所提方法的所有数据集的准确率分别为 99.05%、96.26% 和 95.71%。
{"title":"Dermo-Optimizer: Skin Lesion Classification Using Information-Theoretic Deep Feature Fusion and Entropy-Controlled Binary Bat Optimization","authors":"Tallha Akram,&nbsp;Anas Alsuhaibani,&nbsp;Muhammad Attique Khan,&nbsp;Sajid Ullah Khan,&nbsp;Syed Rameez Naqvi,&nbsp;Mohsin Bilal","doi":"10.1002/ima.23172","DOIUrl":"https://doi.org/10.1002/ima.23172","url":null,"abstract":"<div>\u0000 \u0000 <p>Increases in the prevalence of melanoma, the most lethal form of skin cancer, have been observed over the last few decades. However, the likelihood of a longer life span for individuals is considerably improved with early detection of this malignant illness. Even though the field of computer vision has attained a certain level of success, there is still a degree of ambiguity that represents an unresolved research challenge. In the initial phase of this study, the primary objective is to improve the information derived from input features by combining multiple deep models with the proposed Information-theoretic feature fusion method. Subsequently, in the second phase, the study aims to decrease the redundant and noisy information through down-sampling using the proposed entropy-controlled binary bat selection algorithm. The proposed methodology effectively maintains the integrity of the original feature space, resulting in the creation of highly distinctive feature information. In order to obtain the desired set of features, three contemporary deep models are employed via transfer learning: Inception-Resnet V2, DenseNet-201, and Nasnet Mobile. By combining feature fusion and selection techniques, we may effectively fuse a significant amount of information into the feature vector and subsequently remove any redundant feature information. The effectiveness of the proposed methodology is supported by an evaluation conducted on three well-known dermoscopic datasets, specifically <span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 <msup>\u0000 <mi>PH</mi>\u0000 <mn>2</mn>\u0000 </msup>\u0000 </mrow>\u0000 <annotation>$$ {mathrm{PH}}^2 $$</annotation>\u0000 </semantics></math>, ISIC-2016, and ISIC-2017. In order to validate the proposed approach, several performance indicators are taken into account, such as accuracy, sensitivity, specificity, false negative rate (FNR), false positive rate (FPR), and F1-score. The accuracies obtained for all datasets utilizing the proposed methodology are 99.05%, 96.26%, and 95.71%, respectively.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142233204","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Novel Perceptual Constrained cycleGAN With Attention Mechanisms for Unsupervised MR-to-CT Synthesis 用于无监督 MR-CT 合成的具有注意机制的新型感知受限循环基因组学模型
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-09-14 DOI: 10.1002/ima.23169
Ruiming Zhu, Xinliang Liu, Mingrui Li, Wei Qian, Yueyang Teng

Radiotherapy treatment planning (RTP) requires both magnetic resonance (MR) and computed tomography (CT) modalities. However, conducting separate MR and CT scans for patients leads to misalignment, increased radiation exposure, and higher costs. To address these challenges and mitigate the limitations of supervised synthesis methods, we propose a novel unsupervised perceptual attention image synthesis model based on cycleGAN (PA-cycleGAN). The innovation of PA-cycleGAN lies in its model structure, which incorporates dynamic feature encoding and deep feature extraction to improve the understanding of image structure and contextual information. To ensure the visual authenticity of the synthetic images, we design a hybrid loss function that incorporates perceptual constraints using high-level features extracted by deep neural networks. Our PA-cycleGAN achieves notable results, with an average peak signal-to-noise ratio (PSNR) of 28.06, structural similarity (SSIM) of 0.95, and mean absolute error (MAE) of 46.90 on a pelvic dataset. Additionally, we validate the generalization of our method by conducting experiments on an additional head dataset. These experiments demonstrate that PA-cycleGAN consistently outperforms other state-of-the-art methods in both quantitative metrics and image synthesis quality.

放疗治疗计划(RTP)需要磁共振(MR)和计算机断层扫描(CT)两种模式。然而,对患者分别进行磁共振和 CT 扫描会导致对位错误、辐照增加和成本上升。为了应对这些挑战并减少监督合成方法的局限性,我们提出了一种基于循环广义注视模型(PA-cycleGAN)的新型无监督感知注视图像合成模型。PA-cycleGAN 的创新之处在于其模型结构,它结合了动态特征编码和深度特征提取,以提高对图像结构和上下文信息的理解。为了确保合成图像的视觉真实性,我们设计了一种混合损失函数,利用深度神经网络提取的高级特征,将感知约束条件纳入其中。我们的 PA-cycleGAN 取得了显著的成果,在骨盆数据集上的平均峰值信噪比(PSNR)为 28.06,结构相似度(SSIM)为 0.95,平均绝对误差(MAE)为 46.90。此外,我们还在另外一个头部数据集上进行了实验,验证了我们方法的通用性。这些实验表明,PA-cycleGAN 在定量指标和图像合成质量方面始终优于其他最先进的方法。
{"title":"A Novel Perceptual Constrained cycleGAN With Attention Mechanisms for Unsupervised MR-to-CT Synthesis","authors":"Ruiming Zhu,&nbsp;Xinliang Liu,&nbsp;Mingrui Li,&nbsp;Wei Qian,&nbsp;Yueyang Teng","doi":"10.1002/ima.23169","DOIUrl":"https://doi.org/10.1002/ima.23169","url":null,"abstract":"<div>\u0000 \u0000 <p>Radiotherapy treatment planning (RTP) requires both magnetic resonance (MR) and computed tomography (CT) modalities. However, conducting separate MR and CT scans for patients leads to misalignment, increased radiation exposure, and higher costs. To address these challenges and mitigate the limitations of supervised synthesis methods, we propose a novel unsupervised perceptual attention image synthesis model based on cycleGAN (PA-cycleGAN). The innovation of PA-cycleGAN lies in its model structure, which incorporates dynamic feature encoding and deep feature extraction to improve the understanding of image structure and contextual information. To ensure the visual authenticity of the synthetic images, we design a hybrid loss function that incorporates perceptual constraints using high-level features extracted by deep neural networks. Our PA-cycleGAN achieves notable results, with an average peak signal-to-noise ratio (PSNR) of 28.06, structural similarity (SSIM) of 0.95, and mean absolute error (MAE) of 46.90 on a pelvic dataset. Additionally, we validate the generalization of our method by conducting experiments on an additional head dataset. These experiments demonstrate that PA-cycleGAN consistently outperforms other state-of-the-art methods in both quantitative metrics and image synthesis quality.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142233208","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SSANet—Novel Residual Network for Computer-Aided Diagnosis of Pulmonary Nodules in Chest Computed Tomography 用于胸部计算机断层扫描肺结节计算机辅助诊断的 SSANet-Novel 残差网络
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-09-14 DOI: 10.1002/ima.23176
Yu Gu, Jiaqi Liu, Lidong Yang, Baohua Zhang, Jing Wang, Xiaoqi Lu, Jianjun Li, Xin Liu, Dahua Yu, Ying Zhao, Siyuan Tang, Qun He

The manifestations of early lung cancer in medical imaging often appear as pulmonary nodules, which can be classified as benign or malignant. In recent years, there has been a gradual application of deep learning-based computer-aided diagnosis technology to assist in the diagnosis of lung nodules. This study introduces a novel three-dimensional (3D) residual network called SSANet, which integrates split-based convolution, shuffle attention, and a novel activation function. The aim is to enhance the accuracy of distinguishing between benign and malignant lung nodules using convolutional neural networks (CNNs) and alleviate the burden on doctors when interpreting the images. To fully extract pulmonary nodule information from chest CT images, the original residual network is expanded into a 3D CNN structure. Additionally, a 3D split-based convolutional operation (SPConv) is designed and integrated into the feature extraction module to reduce redundancy in feature maps and improve network inference speed. In the SSABlock part of the proposed network, ACON (Activated or Not) function is also introduced. The proposed SSANet also incorporates an attention module to capture critical characteristics of lung nodules. During the training process, the PolyLoss function is utilized. Once SSANet generates the diagnosis result, a heatmap displays using Score-CAM is employed to evaluate whether the network accurately identifies the location of lung nodules. In the final test set, the proposed network achieves an accuracy of 89.13%, an F1-score of 84.85%, and a G-mean of 86.20%. These metrics represent improvements of 5.43%, 5.98%, and 4.09%, respectively, compared with the original base network. The experimental results align with those of previous studies on pulmonary nodule diagnosis networks, confirming the reliability and clinical applicability of the diagnostic outcomes.

早期肺癌在医学影像中的表现往往是肺部结节,可分为良性和恶性。近年来,基于深度学习的计算机辅助诊断技术逐渐应用于肺结节的辅助诊断。本研究介绍了一种名为 SSANet 的新型三维(3D)残差网络,它集成了基于分裂的卷积、洗牌注意和新型激活函数。其目的是利用卷积神经网络(CNN)提高区分肺结节良性和恶性的准确性,并减轻医生判读图像的负担。为了从胸部 CT 图像中充分提取肺结节信息,原始残差网络被扩展为三维卷积神经网络结构。此外,还设计了一种基于三维分裂的卷积运算(SPConv),并将其集成到特征提取模块中,以减少特征图中的冗余,提高网络推理速度。在拟议网络的 SSABlock 部分,还引入了 ACON(激活或未激活)功能。拟议的 SSANet 还加入了注意力模块,以捕捉肺结节的关键特征。在训练过程中,使用了 PolyLoss 函数。SSANet 生成诊断结果后,将使用 Score-CAM 进行热图显示,以评估网络是否能准确识别肺结节的位置。在最终测试集中,建议的网络达到了 89.13% 的准确率、84.85% 的 F1 分数和 86.20% 的 G 平均值。与原始基础网络相比,这些指标分别提高了 5.43%、5.98% 和 4.09%。实验结果与之前关于肺结节诊断网络的研究结果一致,证实了诊断结果的可靠性和临床适用性。
{"title":"SSANet—Novel Residual Network for Computer-Aided Diagnosis of Pulmonary Nodules in Chest Computed Tomography","authors":"Yu Gu,&nbsp;Jiaqi Liu,&nbsp;Lidong Yang,&nbsp;Baohua Zhang,&nbsp;Jing Wang,&nbsp;Xiaoqi Lu,&nbsp;Jianjun Li,&nbsp;Xin Liu,&nbsp;Dahua Yu,&nbsp;Ying Zhao,&nbsp;Siyuan Tang,&nbsp;Qun He","doi":"10.1002/ima.23176","DOIUrl":"https://doi.org/10.1002/ima.23176","url":null,"abstract":"<div>\u0000 \u0000 <p>The manifestations of early lung cancer in medical imaging often appear as pulmonary nodules, which can be classified as benign or malignant. In recent years, there has been a gradual application of deep learning-based computer-aided diagnosis technology to assist in the diagnosis of lung nodules. This study introduces a novel three-dimensional (3D) residual network called SSANet, which integrates split-based convolution, shuffle attention, and a novel activation function. The aim is to enhance the accuracy of distinguishing between benign and malignant lung nodules using convolutional neural networks (CNNs) and alleviate the burden on doctors when interpreting the images. To fully extract pulmonary nodule information from chest CT images, the original residual network is expanded into a 3D CNN structure. Additionally, a 3D split-based convolutional operation (SPConv) is designed and integrated into the feature extraction module to reduce redundancy in feature maps and improve network inference speed. In the SSABlock part of the proposed network, ACON (Activated or Not) function is also introduced. The proposed SSANet also incorporates an attention module to capture critical characteristics of lung nodules. During the training process, the PolyLoss function is utilized. Once SSANet generates the diagnosis result, a heatmap displays using Score-CAM is employed to evaluate whether the network accurately identifies the location of lung nodules. In the final test set, the proposed network achieves an accuracy of 89.13%, an F1-score of 84.85%, and a G-mean of 86.20%. These metrics represent improvements of 5.43%, 5.98%, and 4.09%, respectively, compared with the original base network. The experimental results align with those of previous studies on pulmonary nodule diagnosis networks, confirming the reliability and clinical applicability of the diagnostic outcomes.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142233214","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Novel Dual Attention Approach for DNN Based Automated Diabetic Retinopathy Grading 基于 DNN 的糖尿病视网膜病变自动分级的新型双重关注方法
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-09-13 DOI: 10.1002/ima.23175
Tareque Bashar Ovi, Nomaiya Bashree, Hussain Nyeem, Md Abdul Wahed, Faiaz Hasanuzzaman Rhythm, Ayat Subah Alam

Diabetic retinopathy (DR) poses a serious threat to vision, emphasising the need for early detection. Manual analysis of fundus images, though common, is error-prone and time-intensive. Existing automated diagnostic methods lack precision, particularly in the early stages of DR. This paper introduces the Soft Convolutional Block Attention Module-based Network (Soft-CBAMNet), a deep learning network designed for severity detection, which features Soft-CBAM attention to capture complex features from fundus images. The proposed network integrates both the convolutional block attention module (CBAM) and the soft-attention components, ensuring simultaneous processing of input features. Following this, attention maps undergo a max-pooling operation, and refined features are concatenated before passing through a dropout layer with a dropout rate of 50%. Experimental results on the APTOS dataset demonstrate the superior performance of Soft-CBAMNet, achieving an accuracy of 85.4% in multiclass DR grading. The proposed architecture has shown strong robustness and general feature learning capability, achieving a mean AUC of 0.81 on the IDRID dataset. Soft-CBAMNet's dynamic feature extraction capability across all classes is further justified by the inspection of intermediate feature maps. The model excels in identifying all stages of DR with increased precision, surpassing contemporary approaches. Soft-CBAMNet presents a significant advancement in DR diagnosis, offering improved accuracy and efficiency for timely intervention.

糖尿病视网膜病变(DR)对视力构成严重威胁,因此需要及早发现。人工分析眼底图像虽然常见,但容易出错且耗费时间。现有的自动诊断方法缺乏精确性,尤其是在 DR 的早期阶段。本文介绍了基于软卷积块注意模块的网络(Soft-CBAMNet),这是一种专为严重程度检测而设计的深度学习网络,它采用软卷积块注意模块来捕捉眼底图像中的复杂特征。该网络集成了卷积块注意力模块(CBAM)和软注意力组件,确保同时处理输入特征。随后,注意力图经过最大池化运算,在通过滤除层(滤除率为 50%)之前将精炼的特征串联起来。在 APTOS 数据集上的实验结果证明了 Soft-CBAMNet 的卓越性能,它在多类 DR 分级中的准确率达到了 85.4%。所提出的架构具有很强的鲁棒性和通用特征学习能力,在 IDRID 数据集上的平均 AUC 达到了 0.81。对中间特征图的检查进一步证明了 Soft-CBAMNet 跨所有类别的动态特征提取能力。该模型在识别 DR 的各个阶段时都表现出色,而且精确度更高,超越了其他同类方法。Soft-CBAMNet 在 DR 诊断方面取得了重大进展,提高了及时干预的准确性和效率。
{"title":"A Novel Dual Attention Approach for DNN Based Automated Diabetic Retinopathy Grading","authors":"Tareque Bashar Ovi,&nbsp;Nomaiya Bashree,&nbsp;Hussain Nyeem,&nbsp;Md Abdul Wahed,&nbsp;Faiaz Hasanuzzaman Rhythm,&nbsp;Ayat Subah Alam","doi":"10.1002/ima.23175","DOIUrl":"https://doi.org/10.1002/ima.23175","url":null,"abstract":"<div>\u0000 \u0000 <p>Diabetic retinopathy (DR) poses a serious threat to vision, emphasising the need for early detection. Manual analysis of fundus images, though common, is error-prone and time-intensive. Existing automated diagnostic methods lack precision, particularly in the early stages of DR. This paper introduces the Soft Convolutional Block Attention Module-based Network (Soft-CBAMNet), a deep learning network designed for severity detection, which features Soft-CBAM attention to capture complex features from fundus images. The proposed network integrates both the convolutional block attention module (CBAM) and the soft-attention components, ensuring simultaneous processing of input features. Following this, attention maps undergo a max-pooling operation, and refined features are concatenated before passing through a dropout layer with a dropout rate of 50%. Experimental results on the APTOS dataset demonstrate the superior performance of Soft-CBAMNet, achieving an accuracy of 85.4% in multiclass DR grading. The proposed architecture has shown strong robustness and general feature learning capability, achieving a mean AUC of 0.81 on the IDRID dataset. Soft-CBAMNet's dynamic feature extraction capability across all classes is further justified by the inspection of intermediate feature maps. The model excels in identifying all stages of DR with increased precision, surpassing contemporary approaches. Soft-CBAMNet presents a significant advancement in DR diagnosis, offering improved accuracy and efficiency for timely intervention.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142230984","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Lightweight Deep Learning Model Optimization for Medical Image Analysis 用于医学图像分析的轻量级深度学习模型优化
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-09-13 DOI: 10.1002/ima.23173
Zahraa Al-Milaji, Hayder Yousif

Medical image labeling requires specialized knowledge; hence, the solution to the challenge of medical image classification lies in efficiently utilizing the few labeled samples to create a high-performance model. Building a high-performance model requires a complicated convolutional neural network (CNN) model with numerous parameters to be trained which makes the test quite expensive. In this paper, we propose optimizing a lightweight deep learning model with only five convolutional layers using the particle swarm optimization (PSO) algorithm to find the best number of kernel filters for each convolutional layer. For colored red, green, and blue (RGB) images acquired from different data sources, we suggest using stain separation using color deconvolution and horizontal and vertical flipping to produce new versions that can concentrate the representation of the images on structures and patterns. To mitigate the effect of training with incorrectly or uncertainly labeled images, grades of disease could have small variances, we apply a second-pass training excluding uncertain data. With a small number of parameters and higher accuracy, the proposed lightweight deep learning model optimization (LDLMO) algorithm shows strong resilience and generalization ability compared with most recent research on four MedMNIST datasets (RetinaMNIST, BreastMNIST, DermMNIST, and OCTMNIST), Medical-MNIST, and brain tumor MRI datasets.

医学图像标注需要专业知识;因此,解决医学图像分类难题的方法在于有效利用为数不多的标注样本来创建高性能模型。建立一个高性能模型需要一个复杂的卷积神经网络(CNN)模型,并需要训练大量参数,这使得测试成本相当高昂。在本文中,我们建议使用粒子群优化(PSO)算法优化仅有五个卷积层的轻量级深度学习模型,为每个卷积层找到最佳的内核过滤器数量。对于从不同数据源获取的彩色红、绿、蓝(RGB)图像,我们建议使用色彩解卷积和水平与垂直翻转来进行污点分离,从而生成新的版本,将图像的表示集中在结构和模式上。为了减轻使用不正确或不确定标记的图像进行训练所带来的影响,我们采用了排除不确定数据的二次训练。所提出的轻量级深度学习模型优化(LDLMO)算法参数少、准确率高,在四个MedMNIST数据集(RetinaMNIST、BreastMNIST、DermMNIST和OCTMNIST)、Medical-MNIST和脑肿瘤MRI数据集上,与最新研究相比,显示出很强的弹性和泛化能力。
{"title":"Lightweight Deep Learning Model Optimization for Medical Image Analysis","authors":"Zahraa Al-Milaji,&nbsp;Hayder Yousif","doi":"10.1002/ima.23173","DOIUrl":"https://doi.org/10.1002/ima.23173","url":null,"abstract":"<div>\u0000 \u0000 <p>Medical image labeling requires specialized knowledge; hence, the solution to the challenge of medical image classification lies in efficiently utilizing the few labeled samples to create a high-performance model. Building a high-performance model requires a complicated convolutional neural network (CNN) model with numerous parameters to be trained which makes the test quite expensive. In this paper, we propose optimizing a lightweight deep learning model with only five convolutional layers using the particle swarm optimization (PSO) algorithm to find the best number of kernel filters for each convolutional layer. For colored red, green, and blue (RGB) images acquired from different data sources, we suggest using stain separation using color deconvolution and horizontal and vertical flipping to produce new versions that can concentrate the representation of the images on structures and patterns. To mitigate the effect of training with incorrectly or uncertainly labeled images, grades of disease could have small variances, we apply a second-pass training excluding uncertain data. With a small number of parameters and higher accuracy, the proposed lightweight deep learning model optimization (LDLMO) algorithm shows strong resilience and generalization ability compared with most recent research on four MedMNIST datasets (RetinaMNIST, BreastMNIST, DermMNIST, and OCTMNIST), Medical-MNIST, and brain tumor MRI datasets.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142230985","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Dynamic Multi-Output Convolutional Neural Network for Skin Lesion Classification 用于皮肤病变分类的动态多输出卷积神经网络
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-09-09 DOI: 10.1002/ima.23164
Yingyue Zhou, Junfei Guo, Hanmin Yao, Jiaqi Zhao, Xiaoxia Li, Jiamin Qin, Shuangli Liu

Skin cancer is a pressing global health issue, with high incidence and mortality rates. Convolutional neural network (CNN) models have been proven to be effective in improving performance in skin lesion image classification and reducing the medical burden. However, the inherent class imbalance in training data caused by the difficulty of collecting dermoscopy images leads to categorical overfitting, which still limits the effectiveness of these data-driven models in recognizing few-shot categories. To address this challenge, we propose a dynamic multi-output convolutional neural network (DMO-CNN) model that incorporates exit nodes into the standard CNN structure and includes feature refinement layers (FRLs) and an adaptive output scheduling (AOS) module. This model improves feature representation ability through multi-scale sub-feature maps and reduces the inter-layer dependencies during backpropagation. The FRLs ensure efficient and low-loss down-sampling, while the AOS module utilizes a trainable layer selection mechanism to refocus the model's attention on few-shot lesion categories. Additionally, a novel correction factor loss is introduced to supervise and promote AOS convergence. Our evaluation of the DMO-CNN model on the HAM10000 dataset demonstrates its effectiveness in multi-class skin lesion classification and its superior performance in recognizing few-shot categories. Despite utilizing a very simple VGG structure as the sole backbone structure, DMO-CNN achieved impressive performance of 0.885 in BACC and 0.983 in weighted AUC. These results are comparable to those of the ensemble model that won the ISIC 2018 challenge, highlighting the strong potential of DMO-CNN in dealing with few-shot skin lesion data.

皮肤癌是一个紧迫的全球健康问题,发病率和死亡率都很高。卷积神经网络(CNN)模型已被证明能有效提高皮肤病变图像分类的性能,减轻医疗负担。然而,由于皮肤镜图像难以收集,训练数据中固有的类别不平衡导致了分类过拟合,这仍然限制了这些数据驱动模型在识别少数几个类别时的有效性。为了应对这一挑战,我们提出了一种动态多输出卷积神经网络(DMO-CNN)模型,它将退出节点纳入标准 CNN 结构,并包含特征细化层(FRL)和自适应输出调度(AOS)模块。该模型通过多尺度子特征图提高了特征表示能力,并减少了反向传播过程中的层间依赖性。FRLs 可确保高效、低损耗的下采样,而 AOS 模块则利用可训练的层选择机制,将模型的注意力重新集中到少数病变类别上。此外,还引入了一种新的校正因子损失,以监督和促进 AOS 的收敛。我们在 HAM10000 数据集上对 DMO-CNN 模型进行了评估,结果表明该模型在多类皮损分类中非常有效,而且在识别少量皮损类别方面表现出色。尽管 DMO-CNN 采用了非常简单的 VGG 结构作为唯一的骨干结构,但其 BACC 和加权 AUC 分别达到了令人印象深刻的 0.885 和 0.983。这些结果与赢得 ISIC 2018 挑战赛的集合模型不相上下,凸显了 DMO-CNN 在处理少量皮损数据方面的强大潜力。
{"title":"A Dynamic Multi-Output Convolutional Neural Network for Skin Lesion Classification","authors":"Yingyue Zhou,&nbsp;Junfei Guo,&nbsp;Hanmin Yao,&nbsp;Jiaqi Zhao,&nbsp;Xiaoxia Li,&nbsp;Jiamin Qin,&nbsp;Shuangli Liu","doi":"10.1002/ima.23164","DOIUrl":"https://doi.org/10.1002/ima.23164","url":null,"abstract":"<div>\u0000 \u0000 <p>Skin cancer is a pressing global health issue, with high incidence and mortality rates. Convolutional neural network (CNN) models have been proven to be effective in improving performance in skin lesion image classification and reducing the medical burden. However, the inherent class imbalance in training data caused by the difficulty of collecting dermoscopy images leads to categorical overfitting, which still limits the effectiveness of these data-driven models in recognizing few-shot categories. To address this challenge, we propose a dynamic multi-output convolutional neural network (DMO-CNN) model that incorporates exit nodes into the standard CNN structure and includes feature refinement layers (FRLs) and an adaptive output scheduling (AOS) module. This model improves feature representation ability through multi-scale sub-feature maps and reduces the inter-layer dependencies during backpropagation. The FRLs ensure efficient and low-loss down-sampling, while the AOS module utilizes a trainable layer selection mechanism to refocus the model's attention on few-shot lesion categories. Additionally, a novel correction factor loss is introduced to supervise and promote AOS convergence. Our evaluation of the DMO-CNN model on the HAM10000 dataset demonstrates its effectiveness in multi-class skin lesion classification and its superior performance in recognizing few-shot categories. Despite utilizing a very simple VGG structure as the sole backbone structure, DMO-CNN achieved impressive performance of 0.885 in BACC and 0.983 in weighted AUC. These results are comparable to those of the ensemble model that won the ISIC 2018 challenge, highlighting the strong potential of DMO-CNN in dealing with few-shot skin lesion data.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142165288","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International Journal of Imaging Systems and Technology
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1