首页 > 最新文献

International Journal of Imaging Systems and Technology最新文献

英文 中文
Application of Non-Immersive Virtual Reality in Cerebral Palsy Children: A Systematic Review 非沉浸式虚拟现实技术在脑瘫儿童中的应用:系统回顾
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-08-27 DOI: 10.1002/ima.23162
Agrawal Luckykumar Dwarkadas, Viswanath Talasila, Sheeraz Kirmani, Rama Krishna Challa, K. G. Srinivasa

Cerebral palsy (CP) is a movement disorder caused by brain damage. Virtual reality (VR) can improve motor function and daily life activities in CP patients. This systematic review examines the use of non-immersive VR in treating CP children. The objective is to evaluate the effectiveness of non-immersive VR in rehabilitating CP children as a standalone intervention or in combination with traditional therapy. The review follows the PRISMA guidelines and includes a comprehensive search of five bibliographic databases. Two reviewers independently assess the search results, evaluate full-text publications, and extract relevant data. The outcomes were described by the International Classification of Functioning, Disability, and Health for Children and Youth (ICF-CY) framework. A total of 20 English-language studies published between January 2013 and January 2023 are included in the review based on predefined inclusion and exclusion criteria. The findings demonstrate that non-immersive VR, when used in conjunction with traditional therapy, yields positive effects on body structure and function (hand function, grip strength, and upper extremity function), activity (motor function, activities of daily life [ADL], balance), and participation (caretakers' assessment, usability, motivation, and user satisfaction) in CP children. Moreover, non-immersive VR alone is found to be more efficient than traditional therapy in improving manual dexterity in CP children. The non-immersive VR can be effective in rehabilitating CP children, and the review concludes by recommending future research with larger sample sizes and randomized trials to investigate further the potential benefits of non-immersive VR in this population.

脑性瘫痪(CP)是一种由脑损伤引起的运动障碍。虚拟现实(VR)可以改善 CP 患者的运动功能和日常生活活动。本系统综述研究了非沉浸式 VR 在治疗 CP 儿童中的应用。目的是评估非沉浸式 VR 作为一种独立的干预措施或与传统疗法相结合对 CP 儿童康复的有效性。综述遵循 PRISMA 指南,包括对五个文献数据库的全面检索。两名审稿人独立评估搜索结果、评价全文出版物并提取相关数据。研究结果采用《国际儿童和青少年功能、残疾和健康分类》(ICF-CY)框架进行描述。根据预定义的纳入和排除标准,本综述共纳入了 2013 年 1 月至 2023 年 1 月间发表的 20 篇英文研究。研究结果表明,非沉浸式 VR 与传统疗法结合使用时,可对 CP 儿童的身体结构和功能(手部功能、握力和上肢功能)、活动(运动功能、日常生活活动 [ADL]、平衡)和参与(护理人员评估、可用性、动机和用户满意度)产生积极影响。此外,与传统疗法相比,非沉浸式虚拟现实疗法在改善 CP 儿童的手部灵活性方面更为有效。非沉浸式虚拟现实技术可有效帮助 CP 儿童康复,综述最后建议今后开展样本量更大的研究和随机试验,以进一步调查非沉浸式虚拟现实技术在这一人群中的潜在益处。
{"title":"Application of Non-Immersive Virtual Reality in Cerebral Palsy Children: A Systematic Review","authors":"Agrawal Luckykumar Dwarkadas,&nbsp;Viswanath Talasila,&nbsp;Sheeraz Kirmani,&nbsp;Rama Krishna Challa,&nbsp;K. G. Srinivasa","doi":"10.1002/ima.23162","DOIUrl":"https://doi.org/10.1002/ima.23162","url":null,"abstract":"<div>\u0000 \u0000 <p>Cerebral palsy (CP) is a movement disorder caused by brain damage. Virtual reality (VR) can improve motor function and daily life activities in CP patients. This systematic review examines the use of non-immersive VR in treating CP children. The objective is to evaluate the effectiveness of non-immersive VR in rehabilitating CP children as a standalone intervention or in combination with traditional therapy. The review follows the PRISMA guidelines and includes a comprehensive search of five bibliographic databases. Two reviewers independently assess the search results, evaluate full-text publications, and extract relevant data. The outcomes were described by the International Classification of Functioning, Disability, and Health for Children and Youth (ICF-CY) framework. A total of 20 English-language studies published between January 2013 and January 2023 are included in the review based on predefined inclusion and exclusion criteria. The findings demonstrate that non-immersive VR, when used in conjunction with traditional therapy, yields positive effects on body structure and function (hand function, grip strength, and upper extremity function), activity (motor function, activities of daily life [ADL], balance), and participation (caretakers' assessment, usability, motivation, and user satisfaction) in CP children. Moreover, non-immersive VR alone is found to be more efficient than traditional therapy in improving manual dexterity in CP children. The non-immersive VR can be effective in rehabilitating CP children, and the review concludes by recommending future research with larger sample sizes and randomized trials to investigate further the potential benefits of non-immersive VR in this population.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142077830","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Smartphone App to Detect Pathological Myopia Using Spatial Attention and Squeeze-Excitation Network as a Classifier and Segmentation Encoder 利用空间注意力和挤压-激发网络作为分类器和分割编码器检测病理性近视的智能手机应用程序
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-08-27 DOI: 10.1002/ima.23157
Sarvat Ali, Shital Raut

Pathological myopia (PM) is a worldwide visual health concern that can cause irreversible vision impairment. It affects up to 20 crore population, causing social and economic burdens. Initial screening of PM using computer-aided diagnosis (CAD) can prevent loss of time and finances for intricate treatments later on. Current research works utilizes complex models that are too resource-intensive or lack explanations behind the categorizations. To emphasize the significance of artificial intelligence for the ophthalmic usage and address the limitations of the current studies, we have designed a mobile-compatible application for smartphone users to detect PM. For this purpose, we have developed a lightweight model, using the enhanced MobileNetV3 architecture integrated with spatial attention (SA) and squeeze-excitation (SE) modules to effectively capture lesion location and channel features. To demonstrate its robustness, the model is tested against three heterogeneous datasets namely PALM, RFMID, and ODIR reporting the area under curve (AUC) score of 0.9983, 0.95, and 0.94, respectively. In order to support PM categorization and demonstrate its correlation with the associated lesions, we have segmented different forms of PM lesion atrophy, which gave us intersection over union (IOU) scores of 0.96 and fscore of 0.97 using the same SA+SE inclusive MobileNetV3 as an encoder. This lesion segmentation can aid ophthalmologists in further analysis and treatment. The optimized and explainable model version is calibrated to develop the smartphone application, which can identify fundus image as PM or normal vision. This app is appropriate for ophthalmologists seeking second opinions or by rural general practitioners to refer PM cases to specialists.

病理性近视(PM)是全球关注的视觉健康问题,可造成不可逆转的视力损害。它影响着多达 20 亿人口,造成社会和经济负担。使用计算机辅助诊断(CAD)对病理性近视进行初步筛查,可避免因日后复杂的治疗而浪费时间和金钱。目前的研究工作使用的复杂模型过于耗费资源,或缺乏分类背后的解释。为了强调人工智能在眼科应用中的重要性并解决当前研究的局限性,我们为智能手机用户设计了一款兼容移动设备的应用程序来检测 PM。为此,我们开发了一个轻量级模型,使用集成了空间注意力(SA)和挤压激励(SE)模块的增强型 MobileNetV3 架构,以有效捕捉病变位置和通道特征。为了证明该模型的鲁棒性,我们对 PALM、RFMID 和 ODIR 这三个异构数据集进行了测试,结果显示曲线下面积(AUC)分别为 0.9983、0.95 和 0.94。为了支持 PM 分类并证明其与相关病变的相关性,我们对不同形式的 PM 病变萎缩进行了分割,使用相同的 SA+SE 包括 MobileNetV3 作为编码器,得到了 0.96 的交集大于联合(IOU)分数和 0.97 的 fscore 分数。这种病变分割可以帮助眼科医生进行进一步的分析和治疗。经过校准的优化可解释模型版本可用于开发智能手机应用程序,该应用程序可将眼底图像识别为 PM 或正常视力。该应用程序适用于寻求第二意见的眼科医生,或农村全科医生将 PM 病例转介给专科医生。
{"title":"Smartphone App to Detect Pathological Myopia Using Spatial Attention and Squeeze-Excitation Network as a Classifier and Segmentation Encoder","authors":"Sarvat Ali,&nbsp;Shital Raut","doi":"10.1002/ima.23157","DOIUrl":"https://doi.org/10.1002/ima.23157","url":null,"abstract":"<div>\u0000 \u0000 <p>Pathological myopia (PM) is a worldwide visual health concern that can cause irreversible vision impairment. It affects up to 20 crore population, causing social and economic burdens. Initial screening of PM using computer-aided diagnosis (CAD) can prevent loss of time and finances for intricate treatments later on. Current research works utilizes complex models that are too resource-intensive or lack explanations behind the categorizations. To emphasize the significance of artificial intelligence for the ophthalmic usage and address the limitations of the current studies, we have designed a mobile-compatible application for smartphone users to detect PM. For this purpose, we have developed a lightweight model, using the enhanced MobileNetV3 architecture integrated with spatial attention (SA) and squeeze-excitation (SE) modules to effectively capture lesion location and channel features. To demonstrate its robustness, the model is tested against three heterogeneous datasets namely PALM, RFMID, and ODIR reporting the area under curve (AUC) score of 0.9983, 0.95, and 0.94, respectively. In order to support PM categorization and demonstrate its correlation with the associated lesions, we have segmented different forms of PM lesion atrophy, which gave us intersection over union (IOU) scores of 0.96 and fscore of 0.97 using the same SA+SE inclusive MobileNetV3 as an encoder. This lesion segmentation can aid ophthalmologists in further analysis and treatment. The optimized and explainable model version is calibrated to develop the smartphone application, which can identify fundus image as PM or normal vision. This app is appropriate for ophthalmologists seeking second opinions or by rural general practitioners to refer PM cases to specialists.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142077831","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
VTCNet: A Feature Fusion DL Model Based on CNN and ViT for the Classification of Cervical Cells VTCNet:基于 CNN 和 ViT 的特征融合 DL 模型用于颈椎细胞分类
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-08-27 DOI: 10.1002/ima.23161
Mingzhe Li, Ningfeng Que, Juanhua Zhang, Pingfang Du, Yin Dai

Cervical cancer is a common malignancy worldwide with high incidence and mortality rates in underdeveloped countries. The Pap smear test, widely used for early detection of cervical cancer, aims to minimize missed diagnoses, which sometimes results in higher false-positive rates. To enhance manual screening practices, computer-aided diagnosis (CAD) systems based on machine learning (ML) and deep learning (DL) for classifying cervical Pap cells have been extensively researched. In our study, we introduced a DL-based method named VTCNet for the task of cervical cell classification. Our approach combines CNN-SPPF and ViT components, integrating modules like Focus and SeparableC3, to capture more potential information, extract local and global features, and merge them to enhance classification performance. We evaluated our method on the public SIPaKMeD dataset, achieving accuracies, precision, recall, and F1 scores of 97.16%, 97.22%, 97.19%, and 97.18%, respectively. We also conducted additional experiments on the Herlev dataset, where our results outperformed previous methods. The VTCNet method achieved higher classification accuracy than traditional ML or shallow DL models through this integration. Related codes: https://github.com/Camellia-0892/VTCNet/tree/main.

宫颈癌是全球常见的恶性肿瘤,在不发达国家的发病率和死亡率都很高。巴氏涂片检查被广泛用于宫颈癌的早期检测,其目的是最大限度地减少漏诊,而漏诊有时会导致较高的假阳性率。为了提高人工筛查的效率,基于机器学习(ML)和深度学习(DL)的计算机辅助诊断(CAD)系统已被广泛研究,用于对宫颈巴氏细胞进行分类。在我们的研究中,我们针对宫颈细胞分类任务引入了一种基于深度学习的方法,名为 VTCNet。我们的方法结合了 CNN-SPPF 和 ViT 组件,集成了 Focus 和 SeparableC3 等模块,以捕获更多潜在信息,提取局部和全局特征,并将它们合并以提高分类性能。我们在公开的 SIPaKMeD 数据集上评估了我们的方法,其准确率、精确度、召回率和 F1 分数分别达到了 97.16%、97.22%、97.19% 和 97.18%。我们还在 Herlev 数据集上进行了额外的实验,结果优于之前的方法。通过这种集成,VTCNet 方法比传统的 ML 或浅层 DL 模型获得了更高的分类准确率。相关代码:https://github.com/Camellia-0892/VTCNet/tree/main.
{"title":"VTCNet: A Feature Fusion DL Model Based on CNN and ViT for the Classification of Cervical Cells","authors":"Mingzhe Li,&nbsp;Ningfeng Que,&nbsp;Juanhua Zhang,&nbsp;Pingfang Du,&nbsp;Yin Dai","doi":"10.1002/ima.23161","DOIUrl":"https://doi.org/10.1002/ima.23161","url":null,"abstract":"<div>\u0000 \u0000 <p>Cervical cancer is a common malignancy worldwide with high incidence and mortality rates in underdeveloped countries. The Pap smear test, widely used for early detection of cervical cancer, aims to minimize missed diagnoses, which sometimes results in higher false-positive rates. To enhance manual screening practices, computer-aided diagnosis (CAD) systems based on machine learning (ML) and deep learning (DL) for classifying cervical Pap cells have been extensively researched. In our study, we introduced a DL-based method named VTCNet for the task of cervical cell classification. Our approach combines CNN-SPPF and ViT components, integrating modules like Focus and SeparableC3, to capture more potential information, extract local and global features, and merge them to enhance classification performance. We evaluated our method on the public SIPaKMeD dataset, achieving accuracies, precision, recall, and F1 scores of 97.16%, 97.22%, 97.19%, and 97.18%, respectively. We also conducted additional experiments on the Herlev dataset, where our results outperformed previous methods. The VTCNet method achieved higher classification accuracy than traditional ML or shallow DL models through this integration. Related codes: https://github.com/Camellia-0892/VTCNet/tree/main.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142077841","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multimodal Neuroimaging Fusion for Alzheimer's Disease: An Image Colorization Approach With Mobile Vision Transformer 阿尔茨海默病的多模态神经成像融合:利用移动视觉转换器的图像着色方法
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-08-26 DOI: 10.1002/ima.23158
Modupe Odusami, Robertas Damasevicius, Egle Milieskaite-Belousoviene, Rytis Maskeliunas

Multimodal neuroimaging, combining data from different sources, has shown promise in the classification of the Alzheimer's disease (AD) stage. Existing multimodal neuroimaging fusion methods exhibit certain limitations, which require advancements to enhance their objective performance, sensitivity, and specificity for AD classification. This study uses the use of a Pareto-optimal cosine color map to enhance classification performance and visual clarity of fused images. A mobile vision transformer (ViT) model, incorporating the swish activation function, is introduced for effective feature extraction and classification. Fused images from the Alzheimer's Disease Neuroimaging Initiative (ADNI), the Whole Brain Atlas (AANLIB), and Open Access Series of Imaging Studies (OASIS) datasets, obtained through optimized transposed convolution, are utilized for model training, while evaluation is achieved using images that have not been fused from the same databases. The proposed model demonstrates high accuracy in AD classification across different datasets, achieving 98.76% accuracy for Early Mild Cognitive Impairment (EMCI) versus LMCI, 98.65% for Late Mild Cognitive Impairment (LMCI) versus AD, 98.60% for EMCI versus AD, and 99.25% for AD versus Cognitive Normal (CN) in the ADNI dataset. Similarly, on OASIS and AANLIB, the precision of the AD versus CN classification is 99.50% and 96.00%, respectively. Evaluation metrics showcase the model's precision, recall, and F1 score for various binary classifications, emphasizing its robust performance.

多模态神经成像结合了不同来源的数据,在阿尔茨海默病(AD)分期分类方面大有可为。现有的多模态神经成像融合方法存在一定的局限性,需要改进以提高其客观性能、灵敏度和特异性。本研究利用帕累托最优余弦色彩图来提高融合图像的分类性能和视觉清晰度。研究还引入了一个移动视觉转换器(ViT)模型,该模型结合了swish激活函数,可有效提取特征并进行分类。模型训练使用的融合图像来自阿尔茨海默病神经成像计划(ADNI)、全脑图集(AANLIB)和开放存取成像研究系列(OASIS)数据集,这些数据集是通过优化的转置卷积获得的,而评估则使用未从相同数据库中融合的图像进行。所提出的模型在不同数据集上的AD分类准确率都很高,在ADNI数据集中,早期轻度认知障碍(EMCI)与LMCI的分类准确率为98.76%,晚期轻度认知障碍(LMCI)与AD的分类准确率为98.65%,EMCI与AD的分类准确率为98.60%,AD与认知正常(CN)的分类准确率为99.25%。同样,在 OASIS 和 AANLIB 数据集中,AD 与 CN 分类的精确度分别为 99.50% 和 96.00%。评估指标展示了该模型在各种二元分类中的精确度、召回率和 F1 分数,强调了其强大的性能。
{"title":"Multimodal Neuroimaging Fusion for Alzheimer's Disease: An Image Colorization Approach With Mobile Vision Transformer","authors":"Modupe Odusami,&nbsp;Robertas Damasevicius,&nbsp;Egle Milieskaite-Belousoviene,&nbsp;Rytis Maskeliunas","doi":"10.1002/ima.23158","DOIUrl":"https://doi.org/10.1002/ima.23158","url":null,"abstract":"<div>\u0000 \u0000 <p>Multimodal neuroimaging, combining data from different sources, has shown promise in the classification of the Alzheimer's disease (AD) stage. Existing multimodal neuroimaging fusion methods exhibit certain limitations, which require advancements to enhance their objective performance, sensitivity, and specificity for AD classification. This study uses the use of a Pareto-optimal cosine color map to enhance classification performance and visual clarity of fused images. A mobile vision transformer (ViT) model, incorporating the swish activation function, is introduced for effective feature extraction and classification. Fused images from the Alzheimer's Disease Neuroimaging Initiative (ADNI), the Whole Brain Atlas (AANLIB), and Open Access Series of Imaging Studies (OASIS) datasets, obtained through optimized transposed convolution, are utilized for model training, while evaluation is achieved using images that have not been fused from the same databases. The proposed model demonstrates high accuracy in AD classification across different datasets, achieving 98.76% accuracy for Early Mild Cognitive Impairment (EMCI) versus LMCI, 98.65% for Late Mild Cognitive Impairment (LMCI) versus AD, 98.60% for EMCI versus AD, and 99.25% for AD versus Cognitive Normal (CN) in the ADNI dataset. Similarly, on OASIS and AANLIB, the precision of the AD versus CN classification is 99.50% and 96.00%, respectively. Evaluation metrics showcase the model's precision, recall, and F1 score for various binary classifications, emphasizing its robust performance.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-08-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142077964","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Novel Classification Approach for Retinal Disease Using Improved Gannet Optimization-Based Capsule DenseNet 使用基于改进型 Gannet 优化胶囊致密网的新型视网膜疾病分类方法
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-08-22 DOI: 10.1002/ima.23156
S. Venkatesan, M. Kempanna, J. Nagaraja, A. Bhuvanesh

An unusual condition of the eye called diabetic retinopathy affects the human retina and is brought on by the blood's constant rise in insulin levels. Loss of vision is the result. Diabetic retinopathy can be improved by receiving an early diagnosis to prevent further damage. A cost-effective method of accumulating medical treatments is through appropriate DR screening. In this work, deep learning framework is introduced for the accurate classification of retinal diseases. The proposed method processes retinal fundus images obtained from databases, addressing noise and artifacts through an improved median filter (ImMF). It leverages the UNet++ model for precise segmentation of the disease-affected regions. UNet++ enhances feature extraction through cross-stage connections, improving segmentation results. The segmented images are then fed as input to the improved gannet optimization-based capsule DenseNet (IG-CDNet) for retinal disease classification. The hybrid capsule DenseNet (CDNet) classifies disease and is optimized using the improved gannet optimization algorithm to boost classification accuracy. Finally, the accuracy and dice score values achieved are 0.9917 and 0.9652 on the APTOS-2019 dataset.

一种名为糖尿病视网膜病变的眼部异常症状会影响人的视网膜,它是由血液中不断升高的胰岛素水平引起的。其结果是视力丧失。糖尿病视网膜病变可以通过早期诊断得到改善,以防止进一步的损害。通过适当的糖尿病视网膜病变筛查,是积累医疗手段的一种经济有效的方法。在这项工作中,引入了深度学习框架,用于对视网膜疾病进行准确分类。所提出的方法处理从数据库中获取的视网膜眼底图像,通过改进的中值滤波器(ImMF)处理噪声和伪影。它利用 UNet++ 模型对受疾病影响的区域进行精确分割。UNet++ 通过跨阶段连接增强了特征提取,从而改善了分割结果。分割后的图像作为输入输入到基于改进甘网优化的胶囊 DenseNet(IG-CDNet)中,用于视网膜疾病分类。混合胶囊 DenseNet(CDNet)对疾病进行分类,并使用改进的甘网优化算法进行优化,以提高分类准确性。最后,在 APTOS-2019 数据集上取得的准确率和骰子分值分别为 0.9917 和 0.9652。
{"title":"A Novel Classification Approach for Retinal Disease Using Improved Gannet Optimization-Based Capsule DenseNet","authors":"S. Venkatesan,&nbsp;M. Kempanna,&nbsp;J. Nagaraja,&nbsp;A. Bhuvanesh","doi":"10.1002/ima.23156","DOIUrl":"https://doi.org/10.1002/ima.23156","url":null,"abstract":"<div>\u0000 \u0000 <p>An unusual condition of the eye called diabetic retinopathy affects the human retina and is brought on by the blood's constant rise in insulin levels. Loss of vision is the result. Diabetic retinopathy can be improved by receiving an early diagnosis to prevent further damage. A cost-effective method of accumulating medical treatments is through appropriate DR screening. In this work, deep learning framework is introduced for the accurate classification of retinal diseases. The proposed method processes retinal fundus images obtained from databases, addressing noise and artifacts through an improved median filter (ImMF). It leverages the UNet++ model for precise segmentation of the disease-affected regions. UNet++ enhances feature extraction through cross-stage connections, improving segmentation results. The segmented images are then fed as input to the improved gannet optimization-based capsule DenseNet (IG-CDNet) for retinal disease classification. The hybrid capsule DenseNet (CDNet) classifies disease and is optimized using the improved gannet optimization algorithm to boost classification accuracy. Finally, the accuracy and dice score values achieved are 0.9917 and 0.9652 on the APTOS-2019 dataset.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142041540","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improving COVID-19 Detection Through Cooperative Deep-Learning Pipeline for Lung Semantic Segmentation in Medical Imaging 通过合作深度学习管道改进 COVID-19 检测,实现医学影像中的肺部语义分割
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-08-14 DOI: 10.1002/ima.23129
Youssef Mourdi, Hanane Allioui, Mohamed Sadgal

The global impact of COVID-19 has resulted in millions of individuals being afflicted, with a staggering mortality toll of over 16 000 over a span of 2 years. The dearth of resources and diagnostic techniques has had an impact on both emerging and wealthy nations. In response to this, researchers from the domains of engineering and medicine are using deep learning methods to create automated algorithms for detecting COVID-19. This work included the development and comparison of a collaborative deep-learning model for the identification of COVID-19 using CT scan images, in comparison to previous deep learning-based methods. The model underwent an ablation study using publicly accessible COVID-19 CT imaging datasets, with encouraging outcomes. The suggested model might aid doctors and academics in devising tools to expedite the process of determining the optimal therapeutic approach for health professionals, hence reducing the risk of potential problems.

COVID-19 的全球影响已导致数百万人患病,两年内死亡人数超过 16 000 人,令人震惊。资源和诊断技术的匮乏对新兴国家和富裕国家都造成了影响。为此,来自工程和医学领域的研究人员正在利用深度学习方法创建检测 COVID-19 的自动算法。这项工作包括开发和比较一个协作式深度学习模型,用于利用 CT 扫描图像识别 COVID-19,并与之前基于深度学习的方法进行比较。该模型使用可公开访问的 COVID-19 CT 成像数据集进行了消融研究,结果令人鼓舞。建议的模型可以帮助医生和学者设计工具,加快医疗专业人员确定最佳治疗方法的过程,从而降低潜在问题的风险。
{"title":"Improving COVID-19 Detection Through Cooperative Deep-Learning Pipeline for Lung Semantic Segmentation in Medical Imaging","authors":"Youssef Mourdi,&nbsp;Hanane Allioui,&nbsp;Mohamed Sadgal","doi":"10.1002/ima.23129","DOIUrl":"https://doi.org/10.1002/ima.23129","url":null,"abstract":"<div>\u0000 \u0000 <p>The global impact of COVID-19 has resulted in millions of individuals being afflicted, with a staggering mortality toll of over 16 000 over a span of 2 years. The dearth of resources and diagnostic techniques has had an impact on both emerging and wealthy nations. In response to this, researchers from the domains of engineering and medicine are using deep learning methods to create automated algorithms for detecting COVID-19. This work included the development and comparison of a collaborative deep-learning model for the identification of COVID-19 using CT scan images, in comparison to previous deep learning-based methods. The model underwent an ablation study using publicly accessible COVID-19 CT imaging datasets, with encouraging outcomes. The suggested model might aid doctors and academics in devising tools to expedite the process of determining the optimal therapeutic approach for health professionals, hence reducing the risk of potential problems.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141980489","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Breaking Barriers in Cancer Diagnosis: Super-Light Compact Convolution Transformer for Colon and Lung Cancer Detection 打破癌症诊断障碍:用于结肠癌和肺癌检测的超轻型紧凑卷积变压器
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-08-12 DOI: 10.1002/ima.23154
Ritesh Maurya, Nageshwar Nath Pandey, Mohan Karnati, Geet Sahu

According to the World Health Organization, lung and colon cancers are known for their high mortality rates which necessitate the diagnosis of these cancers at an early stage. However, the limited availability of data such as histopathology images used for diagnosis of these cancers, poses a significant challenge while developing computer-aided detection system. This makes it necessary to keep a check on the number of parameters in the artificial intelligence (AI) model used for the detection of these cancers considering the limited availability of the data. In this work, a customised compact and efficient convolution transformer architecture, termed, C3-Transformer has been proposed for the diagnosis of colon and lung cancers using histopathological images. The proposed C3-Transformer relies on convolutional tokenisation and sequence pooling approach to keep a check on the number of parameters and to combine the advantage of convolution neural network with the advantages of transformer model. The novelty of the proposed method lies in efficient classification of colon and lung cancers using the proposed C3-Transformer architecture. The performance of the proposed method has been evaluated on the ‘LC25000’ dataset. Experimental results shows that the proposed method has been able to achieve average classification accuracy, precision and recall value of 99.30%, 0.9941 and 0.9950, in classifying the five different classes of colon and lung cancer with only 0.0316 million parameters. Thus, the present computer-aided detection system developed using proposed C3-Transformer can efficiently detect the colon and lung cancers using histopathology images with high detection accuracy.

世界卫生组织指出,肺癌和结肠癌的死亡率很高,因此必须在早期阶段对这些癌症进行诊断。然而,用于诊断这些癌症的组织病理学图像等数据的可用性有限,这给计算机辅助检测系统的开发带来了巨大挑战。因此,考虑到数据的有限性,有必要对用于检测这些癌症的人工智能(AI)模型中的参数数量进行检查。在这项工作中,提出了一种定制的紧凑高效卷积变换器架构,称为 C3 变换器,用于使用组织病理学图像诊断结肠癌和肺癌。所提出的 C3-Transformer 依靠卷积标记化和序列池方法来控制参数数量,并将卷积神经网络的优势与变压器模型的优势结合起来。拟议方法的新颖之处在于利用拟议的 C3 变换器架构对结肠癌和肺癌进行高效分类。我们在 "LC25000 "数据集上对所提方法的性能进行了评估。实验结果表明,在对五种不同类别的结肠癌和肺癌进行分类时,拟议方法仅用了 0.0316 万个参数,就实现了 99.30%、0.9941 和 0.9950 的平均分类准确率、精确度和召回值。因此,利用 C3 变换器开发的本计算机辅助检测系统可以利用组织病理学图像有效地检测出结肠癌和肺癌,并具有较高的检测精度。
{"title":"Breaking Barriers in Cancer Diagnosis: Super-Light Compact Convolution Transformer for Colon and Lung Cancer Detection","authors":"Ritesh Maurya,&nbsp;Nageshwar Nath Pandey,&nbsp;Mohan Karnati,&nbsp;Geet Sahu","doi":"10.1002/ima.23154","DOIUrl":"https://doi.org/10.1002/ima.23154","url":null,"abstract":"<div>\u0000 \u0000 <p>According to the World Health Organization, lung and colon cancers are known for their high mortality rates which necessitate the diagnosis of these cancers at an early stage. However, the limited availability of data such as histopathology images used for diagnosis of these cancers, poses a significant challenge while developing computer-aided detection system. This makes it necessary to keep a check on the number of parameters in the artificial intelligence (AI) model used for the detection of these cancers considering the limited availability of the data. In this work, a customised compact and efficient convolution transformer architecture, termed, C3-Transformer has been proposed for the diagnosis of colon and lung cancers using histopathological images. The proposed C3-Transformer relies on convolutional tokenisation and sequence pooling approach to keep a check on the number of parameters and to combine the advantage of convolution neural network with the advantages of transformer model. The novelty of the proposed method lies in efficient classification of colon and lung cancers using the proposed C3-Transformer architecture. The performance of the proposed method has been evaluated on the ‘LC25000’ dataset. Experimental results shows that the proposed method has been able to achieve average classification accuracy, precision and recall value of 99.30%, 0.9941 and 0.9950, in classifying the five different classes of colon and lung cancer with only 0.0316 million parameters. Thus, the present computer-aided detection system developed using proposed C3-Transformer can efficiently detect the colon and lung cancers using histopathology images with high detection accuracy.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-08-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141967902","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multimodal Fusion for Enhanced Semantic Segmentation in Brain Tumor Imaging: Integrating Deep Learning and Guided Filtering Via Advanced 3D Semantic Segmentation Architectures 多模态融合增强脑肿瘤成像中的语义分割:通过先进的三维语义分割架构整合深度学习和引导式过滤技术
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-08-12 DOI: 10.1002/ima.23152
Abbadullah .H Saleh, Ümit Atila, Oğuzhan Menemencioğlu

Brain tumor segmentation is paramount in medical diagnostics. This study presents a multistage segmentation model consisting of two main steps. First, the fusion of magnetic resonance imaging (MRI) modalities creates new and more effective tumor imaging modalities. Second, the semantic segmentation of the original and fused modalities, utilizing various modified architectures of the U-Net model. In the first step, a residual network with multi-scale backbone architecture (Res2Net) and guided filter are employed for pixel-by-pixel image fusion tasks without requiring any training or learning process. This method captures both detailed and base elements from the multimodal images to produce better and more informative fused images that significantly enhance the segmentation process. Many fusion scenarios were performed and analyzed, revealing that the best fusion results are attained when combining T2-weighted (T2) with fluid-attenuated inversion recovery (FLAIR) and T1-weighted contrast-enhanced (T1CE) with FLAIR modalities. In the second step, several models, including the U-Net and its many modifications (adding attention layers, residual connections, and depthwise separable connections), are trained using both the original and fused modalities. Further, a “Model Selection-based” fusion of these individual models is also considered for more enhancement. In the preprocessing step, the images are resized by cropping them to decrease the pixel count and minimize background interference. Experiments utilizing the brain tumor segmentation (BraTS) 2020 dataset were performed to verify the efficiency and accuracy of the proposed methodology. The “Model Selection-based” fusion model achieved an average Dice score of 88.4%, an individual score of 91.1% for the whole tumor (WT) class, an average sensitivity score of 86.26%, and a specificity score of 91.7%. These results prove the robustness and high performance of the proposed methodology compared to other state-of-the-art methods.

脑肿瘤分割在医学诊断中至关重要。本研究提出了一种由两个主要步骤组成的多阶段分割模型。首先,融合磁共振成像(MRI)模式,创建新的、更有效的肿瘤成像模式。其次,利用 U-Net 模型的各种改进架构,对原始模式和融合模式进行语义分割。第一步,采用具有多尺度骨干架构的残差网络(Res2Net)和引导滤波器来完成逐像素图像融合任务,无需任何训练或学习过程。这种方法可以捕捉多模态图像中的细节和基本元素,生成更好、信息量更大的融合图像,从而显著增强分割过程。我们对许多融合情况进行了分析,结果表明,将 T2 加权(T2)与流体衰减反转恢复(FLAIR)和 T1 加权对比增强(T1CE)与 FLAIR 模式相结合,可获得最佳融合效果。第二步,使用原始模态和融合模态训练多个模型,包括 U-Net 及其多种修改(增加注意力层、残差连接和深度可分离连接)。此外,还考虑对这些单独的模型进行 "基于模型选择 "的融合,以进一步增强效果。在预处理步骤中,通过裁剪调整图像大小,以减少像素数量并最大限度地降低背景干扰。利用脑肿瘤分割(BraTS)2020 数据集进行了实验,以验证所提方法的效率和准确性。基于模型选择的 "融合模型的平均骰子得分率为 88.4%,整个肿瘤(WT)类的单个得分率为 91.1%,平均灵敏度得分率为 86.26%,特异性得分率为 91.7%。与其他最先进的方法相比,这些结果证明了所提出方法的稳健性和高性能。
{"title":"Multimodal Fusion for Enhanced Semantic Segmentation in Brain Tumor Imaging: Integrating Deep Learning and Guided Filtering Via Advanced 3D Semantic Segmentation Architectures","authors":"Abbadullah .H Saleh,&nbsp;Ümit Atila,&nbsp;Oğuzhan Menemencioğlu","doi":"10.1002/ima.23152","DOIUrl":"https://doi.org/10.1002/ima.23152","url":null,"abstract":"<p>Brain tumor segmentation is paramount in medical diagnostics. This study presents a multistage segmentation model consisting of two main steps. First, the fusion of magnetic resonance imaging (MRI) modalities creates new and more effective tumor imaging modalities. Second, the semantic segmentation of the original and fused modalities, utilizing various modified architectures of the U-Net model. In the first step, a residual network with multi-scale backbone architecture (Res2Net) and guided filter are employed for pixel-by-pixel image fusion tasks without requiring any training or learning process. This method captures both detailed and base elements from the multimodal images to produce better and more informative fused images that significantly enhance the segmentation process. Many fusion scenarios were performed and analyzed, revealing that the best fusion results are attained when combining T2-weighted (T2) with fluid-attenuated inversion recovery (FLAIR) and T1-weighted contrast-enhanced (T1CE) with FLAIR modalities. In the second step, several models, including the U-Net and its many modifications (adding attention layers, residual connections, and depthwise separable connections), are trained using both the original and fused modalities. Further, a “Model Selection-based” fusion of these individual models is also considered for more enhancement. In the preprocessing step, the images are resized by cropping them to decrease the pixel count and minimize background interference. Experiments utilizing the brain tumor segmentation (BraTS) 2020 dataset were performed to verify the efficiency and accuracy of the proposed methodology. The “Model Selection-based” fusion model achieved an average Dice score of 88.4%, an individual score of 91.1% for the whole tumor (WT) class, an average sensitivity score of 86.26%, and a specificity score of 91.7%. These results prove the robustness and high performance of the proposed methodology compared to other state-of-the-art methods.</p>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-08-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ima.23152","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141980407","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Correction to “Multi-Branch Sustainable Convolutional Neural Network for Disease Classification” 对 "用于疾病分类的多分支可持续卷积神经网络 "的更正
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-08-06 DOI: 10.1002/ima.23153

M. Naz, M. A. Shah, H. A. Khattak, et al., “Multi-Branch Sustainable Convolutional Neural Network for Disease Classification,” International Journal of Imaging Systems and Technology 33, no. 5 (2023): 1621–1633, https://doi.org/10.1002/ima.22884.

The affiliation of Hafiz Tayyab Rauf should be: Independent Researcher, UK. The correct author list and affiliations appear below.

Maria Naz1 | Munam Ali Shah1 | Hasan Ali Khattak2 | Abdul Wahid2,3 | Muhammad Nabeel Asghar4 | Hafiz Tayyab Rauf5 | Muhammad Attique Khan6 | Zoobia Ameer7

1Department of Computer Science, COMSATS University Islamabad, Islamabad, Pakistan

2School of Electrical Engineering & Computer Science (SEECS), National University of Sciences and Technology (NUST), Islamabad, Pakistan

3School of Computer Science, University of Birmingham, Dubai, United Arab Emirates

4Department of Computer Science, Bahauddin Zakariya University, Multan, Pakistan

5Independent Researcher, UK

6HITEC University Taxila, Taxila, Pakistan

7Shaheed Benazir Bhutto Women University Peshawar, Peshawar, Pakistan

We apologize for this error.

M.M.Naz、M.A.Shah、H.A.Khattak 等人,"用于疾病分类的多分支可持续卷积神经网络",《国际成像系统与技术杂志》第 33 期,第 5 号(2023 年):"用于疾病分类的多分支可持续卷积神经网络"。5 (2023):1621-1633, https://doi.org/10.1002/ima.22884.The Hafiz Tayyab Rauf 的隶属关系应为:英国独立研究员。正确的作者名单和单位如下。Maria Naz1 | Munam Ali Shah1 | Hasan Ali Khattak2 | Abdul Wahid2,3 | Muhammad Nabeel Asghar4 | Hafiz Tayyab Rauf5 | Muhammad Attique Khan6 | Zoobia Ameer71Department of Computer Science, COMSATS University Islamabad, Islamabad, Pakistan2School of Electrical Engineering &;计算机科学 (SEECS), National University of Sciences and Technology (NUST), Islamabad, Pakistan3School of Computer Science, University of Birmingham, Dubai, United Arab Emirates4Department of Computer Science, Bahauddin Zakariya University, Multan, Pakistan5Independent Researcher, UK6HITEC University Taxila, Taxila, Pakistan7Shaed Benazir Bhutto Women University Peshawar, Peshawar, PakistanWe apologize for this error.
{"title":"Correction to “Multi-Branch Sustainable Convolutional Neural Network for Disease Classification”","authors":"","doi":"10.1002/ima.23153","DOIUrl":"10.1002/ima.23153","url":null,"abstract":"<p>M. Naz, M. A. Shah, H. A. Khattak, et al., “Multi-Branch Sustainable Convolutional Neural Network for Disease Classification,” <i>International Journal of Imaging Systems and Technology</i> 33, no. 5 (2023): 1621–1633, https://doi.org/10.1002/ima.22884.</p><p>The affiliation of Hafiz Tayyab Rauf should be: Independent Researcher, UK. The correct author list and affiliations appear below.</p><p>Maria Naz<sup>1</sup> | Munam Ali Shah<sup>1</sup> | Hasan Ali Khattak<sup>2</sup> | Abdul Wahid<sup>2,3</sup> | Muhammad Nabeel Asghar<sup>4</sup> | Hafiz Tayyab Rauf<sup>5</sup> | Muhammad Attique Khan<sup>6</sup> | Zoobia Ameer<sup>7</sup></p><p><sup>1</sup>Department of Computer Science, COMSATS University Islamabad, Islamabad, Pakistan</p><p><sup>2</sup>School of Electrical Engineering &amp; Computer Science (SEECS), National University of Sciences and Technology (NUST), Islamabad, Pakistan</p><p><sup>3</sup>School of Computer Science, University of Birmingham, Dubai, United Arab Emirates</p><p><sup>4</sup>Department of Computer Science, Bahauddin Zakariya University, Multan, Pakistan</p><p><sup>5</sup>Independent Researcher, UK</p><p><sup>6</sup>HITEC University Taxila, Taxila, Pakistan</p><p><sup>7</sup>Shaheed Benazir Bhutto Women University Peshawar, Peshawar, Pakistan</p><p>We apologize for this error.</p>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ima.23153","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141935110","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ADT-UNet: An Innovative Algorithm for Glioma Segmentation in MR Images ADT-UNet:磁共振图像中胶质瘤分割的创新算法
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-08-01 DOI: 10.1002/ima.23150
Liu Zhipeng, Wu Jiawei, Jing Ye, Xuefeng Bian, Wu Qiwei, Rui Li, Yinxing Zhu

The precise delineation of glioma tumors is of paramount importance for surgical and radiotherapy planning. Presently, the primary drawbacks associated with the manual segmentation approach are its laboriousness and inefficiency. In order to tackle these challenges, a deep learning-based automatic segmentation technique was introduced to enhance the efficiency of the segmentation process. We proposed ADT-UNet, an innovative algorithm for segmenting glioma tumors in MR images. ADT-UNet leveraged attention-dense blocks and Transformer as its foundational elements. It extended the U-Net framework by incorporating the dense connection structure and attention mechanism. Additionally, a Transformer structure was introduced at the end of the encoder. Furthermore, a novel attention-guided multi-scale feature fusion module was integrated into the decoder. To enhance network stability during training, a loss function was devised that combines Dice loss and binary cross-entropy loss, effectively guiding the network optimization process. On the test set, the DSC was 0.933, the IOU was 0.878, the PPV was 0.942, and the Sen was 0.938. Ablation experiments conclusively demonstrated that the inclusion of all the three proposed modules led to enhanced segmentation accuracy within the model. The most favorable outcomes were observed when all the three modules were employed simultaneously. The proposed methodology exhibited substantial competitiveness across various evaluation indices, with the three additional modules synergistically complementing each other to collectively enhance the segmentation accuracy of the model. Consequently, it is anticipated that this method will serve as a robust tool for assisting clinicians in auxiliary diagnosis and contribute to the advancement of medical intelligence technology.

精确划分胶质瘤肿瘤对手术和放疗计划至关重要。目前,人工分割方法的主要缺点是费力和效率低下。为了应对这些挑战,我们引入了基于深度学习的自动分割技术,以提高分割过程的效率。我们提出了一种创新算法 ADT-UNet,用于分割磁共振图像中的胶质瘤肿瘤。ADT-UNet 利用注意力密集块和 Transformer 作为其基础元素。它结合了密集连接结构和注意力机制,扩展了 U-Net 框架。此外,还在编码器末端引入了 Transformer 结构。此外,解码器中还集成了一个新颖的注意力引导的多尺度特征融合模块。为了增强训练过程中的网络稳定性,设计了一种结合了骰子损失和二元交叉熵损失的损失函数,有效地指导了网络优化过程。在测试集上,DSC 为 0.933,IOU 为 0.878,PPV 为 0.942,Sen 为 0.938。消融实验最终证明,在模型中加入所有三个建议的模块可提高分割准确性。当同时使用所有三个模块时,观察到了最有利的结果。所提出的方法在各种评价指标上都表现出了很强的竞争力,三个附加模块协同互补,共同提高了模型的分割准确性。因此,该方法有望成为辅助临床医生进行辅助诊断的有力工具,并为医学智能技术的发展做出贡献。
{"title":"ADT-UNet: An Innovative Algorithm for Glioma Segmentation in MR Images","authors":"Liu Zhipeng,&nbsp;Wu Jiawei,&nbsp;Jing Ye,&nbsp;Xuefeng Bian,&nbsp;Wu Qiwei,&nbsp;Rui Li,&nbsp;Yinxing Zhu","doi":"10.1002/ima.23150","DOIUrl":"10.1002/ima.23150","url":null,"abstract":"<div>\u0000 \u0000 <p>The precise delineation of glioma tumors is of paramount importance for surgical and radiotherapy planning. Presently, the primary drawbacks associated with the manual segmentation approach are its laboriousness and inefficiency. In order to tackle these challenges, a deep learning-based automatic segmentation technique was introduced to enhance the efficiency of the segmentation process. We proposed ADT-UNet, an innovative algorithm for segmenting glioma tumors in MR images. ADT-UNet leveraged attention-dense blocks and Transformer as its foundational elements. It extended the U-Net framework by incorporating the dense connection structure and attention mechanism. Additionally, a Transformer structure was introduced at the end of the encoder. Furthermore, a novel attention-guided multi-scale feature fusion module was integrated into the decoder. To enhance network stability during training, a loss function was devised that combines Dice loss and binary cross-entropy loss, effectively guiding the network optimization process. On the test set, the DSC was 0.933, the IOU was 0.878, the PPV was 0.942, and the Sen was 0.938. Ablation experiments conclusively demonstrated that the inclusion of all the three proposed modules led to enhanced segmentation accuracy within the model. The most favorable outcomes were observed when all the three modules were employed simultaneously. The proposed methodology exhibited substantial competitiveness across various evaluation indices, with the three additional modules synergistically complementing each other to collectively enhance the segmentation accuracy of the model. Consequently, it is anticipated that this method will serve as a robust tool for assisting clinicians in auxiliary diagnosis and contribute to the advancement of medical intelligence technology.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 5","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141886667","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International Journal of Imaging Systems and Technology
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1