首页 > 最新文献

International Journal of Imaging Systems and Technology最新文献

英文 中文
A Novel Edge-Enhanced Networks for Optic Disc and Optic Cup Segmentation 一种新的视盘和视杯分割边缘增强网络
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-12-20 DOI: 10.1002/ima.70019
Mingtao Liu, Yunyu Wang, Yuxuan Li, Shunbo Hu, Guodong Wang, Jing Wang

Optic disc and optic cup segmentation plays a key role in early diagnosis of glaucoma which is a serious eye disease that can cause damage to the optic nerve, retina, and may cause permanent blindness. Deep learning-based models are used to improve the efficiency and accuracy of fundus image segmentation. However, most approaches currently still have limitations in accurately segmenting optic disc and optic cup, which suffer from the lack of feature abstraction representation and blurring of segmentation in edge regions. This paper proposes a novel edge enhancement network called EE-TransUNet to tackle this challenge. It incorporates the Cascaded Convolutional Fusion block before each decoder layer. This enhances the abstract representation of features and preserves the information of the original features, thereby improving the model's nonlinear fitting ability. Additionally, the Channel Shuffling Multiple Expansion Fusion block is incorporated into the skip connections of the model. This block enhances the network's ability to perceive and characterize image features, thereby improving segmentation accuracy at the edges of the optic cup and optic disc. We validate the effectiveness of the method by conducting experiments on three publicly available datasets, RIM-ONE-v3, REFUGUE and DRISHTI-GS. The Dice coefficients on the test set are 0.871, 0.9056, 0.9068 for the optic cup region and 0.9721, 0.967, 0.9774 for the optic disc region, respectively. The proposed method achieves competitive results compared to other state-of-the-art methods. Our code is available at: https://github.com/wangyunyuwyy/EE-TransUNet.

青光眼是一种严重的眼病,可引起视神经、视网膜损伤,并可能导致永久性失明。视盘和视杯分割在青光眼的早期诊断中起着关键作用。利用基于深度学习的模型提高眼底图像分割的效率和准确性。然而,目前大多数方法在准确分割视盘和视杯方面仍然存在局限性,缺乏特征抽象表示和边缘区域分割模糊。本文提出了一种新的边缘增强网络EE-TransUNet来解决这一挑战。它在每个解码器层之前合并了级联卷积融合块。这样既增强了特征的抽象表示,又保留了原始特征的信息,从而提高了模型的非线性拟合能力。此外,通道洗牌多重扩展融合块被纳入该模型的跳过连接。该块增强了网络感知和表征图像特征的能力,从而提高了视杯和视盘边缘的分割精度。我们通过在RIM-ONE-v3、refuge和DRISHTI-GS三个公开数据集上进行实验来验证该方法的有效性。测试集上的Dice系数分别为:视杯区0.871、0.9056、0.9068,视盘区0.9721、0.967、0.9774。与其他最先进的方法相比,所提出的方法取得了具有竞争力的结果。我们的代码可在:https://github.com/wangyunyuwyy/EE-TransUNet。
{"title":"A Novel Edge-Enhanced Networks for Optic Disc and Optic Cup Segmentation","authors":"Mingtao Liu,&nbsp;Yunyu Wang,&nbsp;Yuxuan Li,&nbsp;Shunbo Hu,&nbsp;Guodong Wang,&nbsp;Jing Wang","doi":"10.1002/ima.70019","DOIUrl":"https://doi.org/10.1002/ima.70019","url":null,"abstract":"<div>\u0000 \u0000 <p>Optic disc and optic cup segmentation plays a key role in early diagnosis of glaucoma which is a serious eye disease that can cause damage to the optic nerve, retina, and may cause permanent blindness. Deep learning-based models are used to improve the efficiency and accuracy of fundus image segmentation. However, most approaches currently still have limitations in accurately segmenting optic disc and optic cup, which suffer from the lack of feature abstraction representation and blurring of segmentation in edge regions. This paper proposes a novel edge enhancement network called EE-TransUNet to tackle this challenge. It incorporates the Cascaded Convolutional Fusion block before each decoder layer. This enhances the abstract representation of features and preserves the information of the original features, thereby improving the model's nonlinear fitting ability. Additionally, the Channel Shuffling Multiple Expansion Fusion block is incorporated into the skip connections of the model. This block enhances the network's ability to perceive and characterize image features, thereby improving segmentation accuracy at the edges of the optic cup and optic disc. We validate the effectiveness of the method by conducting experiments on three publicly available datasets, RIM-ONE-v3, REFUGUE and DRISHTI-GS. The Dice coefficients on the test set are 0.871, 0.9056, 0.9068 for the optic cup region and 0.9721, 0.967, 0.9774 for the optic disc region, respectively. The proposed method achieves competitive results compared to other state-of-the-art methods. Our code is available at: https://github.com/wangyunyuwyy/EE-TransUNet.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142868846","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Interactive Pulmonary Lobe Segmentation in CT Images Based on Oriented Derivative of Stick Filter and Surface Fitting Model 基于棒状滤波器定向导数和表面拟合模型的CT图像交互式肺叶分割
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-12-19 DOI: 10.1002/ima.70011
Yuanyuan Peng, Jiawei Liao, Xuemei Xu, Zixu Zhang, Siqiang Zhu

Automated approaches for pulmonary lobe segmentation frequently encounter difficulties when applied to clinically significant cases, primarily stemming from factors such as incomplete and blurred pulmonary fissures, unpredictable pathological deformation, indistinguishable pulmonary arteries and veins, and severe damage to the lung trachea. To address these challenges, an interactive and intuitive approach utilizing an oriented derivative of stick (ODoS) filter and a surface fitting model is proposed to effectively extract and repair incomplete pulmonary fissures for accurate lung lobe segmentation in computed tomography (CT) images. First, an ODoS filter was employed in a two-dimensional (2D) space to enhance the visibility of pulmonary fissures using a triple-stick template to match the curvilinear structures across various orientations. Second, a three-dimensional (3D) post-processing pipeline based on a direction partition and integration approach was implemented for the initial detection of pulmonary fissures. Third, a coarse-to-fine segmentation strategy is utilized to eliminate extraneous clutter and rectify missed pulmonary fissures, thereby generating accurate pulmonary fissure segmentation. Finally, considering that pulmonary fissures serve as physical boundaries of the lung lobes, a multi-projection technique and surface fitting model were combined to generate a comprehensive fissure surface for pulmonary lobe segmentation. To assess the effectiveness of our approach, we actively participated in an internationally recognized lung lobe segmentation challenge known as LObe and Lung Analysis 2011 (LOLA11), which encompasses 55 CT scans. The validity of the proposed methodology was confirmed by its successful application to a publicly accessible challenge dataset. Overall, our method achieved an average intersection over union (IoU) of 0.913 for lung lobe segmentation, ranking seventh among all participants so far. Furthermore, experimental outcomes demonstrated excellent performance compared with other methods, as evidenced by both visual examination and quantitative evaluation.

肺叶分割的自动化方法在应用于有临床意义的病例时经常遇到困难,主要是由于肺裂隙不完整和模糊、不可预测的病理变形、难以区分的肺动脉和静脉以及肺气管的严重损伤等因素。为了解决这些挑战,提出了一种交互式和直观的方法,利用定向棒导数(ODoS)滤波器和表面拟合模型有效地提取和修复不完整的肺裂隙,以便在计算机断层扫描(CT)图像中进行准确的肺叶分割。首先,在二维(2D)空间中使用ODoS滤波器,使用三棒模板来匹配不同方向的曲线结构,以增强肺裂缝的可见性。其次,实现了基于方向划分积分的三维后处理流水线,用于肺裂隙的初始检测。第三,采用从粗到精的分割策略,剔除多余的杂波,对漏失的肺裂隙进行校正,得到准确的肺裂隙分割。最后,考虑到肺裂隙作为肺叶的物理边界,结合多投影技术和表面拟合模型,生成综合的肺叶分割裂隙面。为了评估我们方法的有效性,我们积极参与了国际公认的肺叶分割挑战,称为肺叶和肺分析2011 (LOLA11),其中包括55个CT扫描。所提出方法的有效性通过其成功应用于可公开访问的挑战数据集得到了证实。总体而言,我们的方法对肺叶分割的平均IoU为0.913,在所有参与者中排名第七。此外,与其他方法相比,实验结果显示出优异的性能,无论是视觉检查还是定量评价都证明了这一点。
{"title":"Interactive Pulmonary Lobe Segmentation in CT Images Based on Oriented Derivative of Stick Filter and Surface Fitting Model","authors":"Yuanyuan Peng,&nbsp;Jiawei Liao,&nbsp;Xuemei Xu,&nbsp;Zixu Zhang,&nbsp;Siqiang Zhu","doi":"10.1002/ima.70011","DOIUrl":"https://doi.org/10.1002/ima.70011","url":null,"abstract":"<div>\u0000 \u0000 <p>Automated approaches for pulmonary lobe segmentation frequently encounter difficulties when applied to clinically significant cases, primarily stemming from factors such as incomplete and blurred pulmonary fissures, unpredictable pathological deformation, indistinguishable pulmonary arteries and veins, and severe damage to the lung trachea. To address these challenges, an interactive and intuitive approach utilizing an oriented derivative of stick (ODoS) filter and a surface fitting model is proposed to effectively extract and repair incomplete pulmonary fissures for accurate lung lobe segmentation in computed tomography (CT) images. First, an ODoS filter was employed in a two-dimensional (2D) space to enhance the visibility of pulmonary fissures using a triple-stick template to match the curvilinear structures across various orientations. Second, a three-dimensional (3D) post-processing pipeline based on a direction partition and integration approach was implemented for the initial detection of pulmonary fissures. Third, a coarse-to-fine segmentation strategy is utilized to eliminate extraneous clutter and rectify missed pulmonary fissures, thereby generating accurate pulmonary fissure segmentation. Finally, considering that pulmonary fissures serve as physical boundaries of the lung lobes, a multi-projection technique and surface fitting model were combined to generate a comprehensive fissure surface for pulmonary lobe segmentation. To assess the effectiveness of our approach, we actively participated in an internationally recognized lung lobe segmentation challenge known as LObe and Lung Analysis 2011 (LOLA11), which encompasses 55 CT scans. The validity of the proposed methodology was confirmed by its successful application to a publicly accessible challenge dataset. Overall, our method achieved an average intersection over union (IoU) of 0.913 for lung lobe segmentation, ranking seventh among all participants so far. Furthermore, experimental outcomes demonstrated excellent performance compared with other methods, as evidenced by both visual examination and quantitative evaluation.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142861739","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Relation Explore Convolutional Block Attention Module for Skin Lesion Classification 用于皮肤病变分类的卷积块注意力模块的关系探索
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-12-19 DOI: 10.1002/ima.70002
Qichen Su, Haza Nuzly Abdull Hamed, Dazhuo Zhou

Skin cancer remains a significant global health concern, demanding accurate and efficient diagnostic solutions. Despite advances in convolutional neural networks for computer vision, automated skin lesion diagnosis remains challenging due to the small lesion region in images and limited inter-class variation. Accurate classification depends on precise lesion localization and recognition of fine-grained visual differences. To address these challenges, this paper proposes an enhancement to the Convolutional Block Attention Module, referred to as Relation Explore Convolutional Block Attention Module. This enhancement improves upon the existing module by utilizing multiple combinations of pooling-based attentions, enabling the model to better learn and leverage complex interactions during training. Extensive experiments are conducted to investigate the performance of skin lesion diagnosis when integrating Relation Explore Convolutional Block Attention Module with ResNet50 at different stages. The best-performing model achieves outstanding classification results on the publicly available HAM10000 dataset, with an Accuracy of 97.63%, Precision of 88.98%, Sensitivity of 82.86%, Specificity of 97.65%, and F1-score of 85.46%, using fivefold cross-validation. The high performance of this model, alongside the clear interpretability provided by its attention maps, builds trust in automated systems. This trust empowers clinicians to make well-informed decisions, significantly enhancing the potential for improved patient outcomes.

皮肤癌仍然是全球关注的重大健康问题,需要准确高效的诊断解决方案。尽管用于计算机视觉的卷积神经网络取得了进展,但由于图像中的病变区域较小,且类间差异有限,因此自动皮肤病变诊断仍具有挑战性。准确的分类取决于精确的病变定位和对细粒度视觉差异的识别。为了应对这些挑战,本文提出了一种卷积块注意力模块的增强方法,称为 "关系探索卷积块注意力模块"。该增强模块利用基于集合的注意力的多种组合来改进现有模块,从而使模型在训练过程中更好地学习和利用复杂的交互。我们进行了广泛的实验,研究了在不同阶段将关系探索卷积块注意力模块与 ResNet50 集成后的皮损诊断性能。表现最好的模型在公开的 HAM10000 数据集上取得了出色的分类结果,使用五重交叉验证,准确率为 97.63%,精确率为 88.98%,灵敏度为 82.86%,特异性为 97.65%,F1 分数为 85.46%。该模型的高性能以及其注意力图提供的清晰可解释性,建立了人们对自动化系统的信任。这种信任使临床医生能够在充分知情的情况下做出决定,从而大大提高了改善患者预后的潜力。
{"title":"Relation Explore Convolutional Block Attention Module for Skin Lesion Classification","authors":"Qichen Su,&nbsp;Haza Nuzly Abdull Hamed,&nbsp;Dazhuo Zhou","doi":"10.1002/ima.70002","DOIUrl":"https://doi.org/10.1002/ima.70002","url":null,"abstract":"<div>\u0000 \u0000 <p>Skin cancer remains a significant global health concern, demanding accurate and efficient diagnostic solutions. Despite advances in convolutional neural networks for computer vision, automated skin lesion diagnosis remains challenging due to the small lesion region in images and limited inter-class variation. Accurate classification depends on precise lesion localization and recognition of fine-grained visual differences. To address these challenges, this paper proposes an enhancement to the Convolutional Block Attention Module, referred to as Relation Explore Convolutional Block Attention Module. This enhancement improves upon the existing module by utilizing multiple combinations of pooling-based attentions, enabling the model to better learn and leverage complex interactions during training. Extensive experiments are conducted to investigate the performance of skin lesion diagnosis when integrating Relation Explore Convolutional Block Attention Module with ResNet50 at different stages. The best-performing model achieves outstanding classification results on the publicly available HAM10000 dataset, with an Accuracy of 97.63%, Precision of 88.98%, Sensitivity of 82.86%, Specificity of 97.65%, and F1-score of 85.46%, using fivefold cross-validation. The high performance of this model, alongside the clear interpretability provided by its attention maps, builds trust in automated systems. This trust empowers clinicians to make well-informed decisions, significantly enhancing the potential for improved patient outcomes.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142861678","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Microaneurysm Detection With Multiscale Attention and Trident RPN 利用多尺度注意力和三叉戟 RPN 检测微动脉瘤
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-12-19 DOI: 10.1002/ima.70015
Jiawen Lin, Shilin Liu, Meiyan Mao, Susu Chen

Diabetic retinopathy (DR) is the most serious and common complication of diabetes. Microaneurysm (MA) detection is of great importance for DR screening by providing the earliest indicator of presence of DR. Extremely small size of MAs, low color contrast in fundus images, and the interference from blood vessels and other lesions with similar characteristics make MA detection still challenging. In this paper, a novel two-stage MA detector with multiscale attention and trident Region proposal network (RPN) is proposed. A scale selection pyramid network based on the attention mechanism is established to improve detection performance on the small objects by reducing the gradient inconsistency between low and high level features. Meanwhile, a trident RPN with three-branch parallel feature enhance head is designed to promote more distinguishing learning, further reducing the misrecognition. The proposed method is validated on IDRiD, e-ophtha, and ROC datasets with the average scores of 0.516, 0.646, and 0.245, respectively, achieving the best or nearly optimal performance compared to the state-of-the-arts. Besides, the proposed MA detector illustrates a more balanced performance on the three datasets, showing strong generalization.

糖尿病视网膜病变(DR)是糖尿病最严重、最常见的并发症。微动脉瘤(micro动脉瘤,MA)的检测对于DR的筛查具有重要意义,它能提供DR存在的最早指标。由于MA的体积极小,眼底图像颜色对比度较低,再加上血管等具有相似特征病变的干扰,使得MA的检测仍然具有挑战性。提出了一种基于多尺度注意力和三叉戟区域建议网络(RPN)的两级MA检测器。建立了一种基于注意机制的尺度选择金字塔网络,通过减少高低阶特征之间的梯度不一致性,提高对小目标的检测性能。同时,设计了具有三分支并行特征增强头的三叉戟RPN,提高了学习的可识别性,进一步减少了误识别。在IDRiD、e-ophtha和ROC数据集上进行了验证,平均得分分别为0.516、0.646和0.245,与目前的方法相比,该方法达到了最佳或接近最佳的性能。此外,本文提出的MA检测器在三个数据集上表现出更均衡的性能,具有较强的泛化能力。
{"title":"Microaneurysm Detection With Multiscale Attention and Trident RPN","authors":"Jiawen Lin,&nbsp;Shilin Liu,&nbsp;Meiyan Mao,&nbsp;Susu Chen","doi":"10.1002/ima.70015","DOIUrl":"https://doi.org/10.1002/ima.70015","url":null,"abstract":"<div>\u0000 \u0000 <p>Diabetic retinopathy (DR) is the most serious and common complication of diabetes. Microaneurysm (MA) detection is of great importance for DR screening by providing the earliest indicator of presence of DR. Extremely small size of MAs, low color contrast in fundus images, and the interference from blood vessels and other lesions with similar characteristics make MA detection still challenging. In this paper, a novel two-stage MA detector with multiscale attention and trident Region proposal network (RPN) is proposed. A scale selection pyramid network based on the attention mechanism is established to improve detection performance on the small objects by reducing the gradient inconsistency between low and high level features. Meanwhile, a trident RPN with three-branch parallel feature enhance head is designed to promote more distinguishing learning, further reducing the misrecognition. The proposed method is validated on IDRiD, e-ophtha, and ROC datasets with the average scores of 0.516, 0.646, and 0.245, respectively, achieving the best or nearly optimal performance compared to the state-of-the-arts. Besides, the proposed MA detector illustrates a more balanced performance on the three datasets, showing strong generalization.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142861740","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
C-TUnet: A CNN-Transformer Architecture-Based Ultrasound Breast Image Classification Network C-TUnet:基于CNN-Transformer架构的超声乳房图像分类网络
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-12-17 DOI: 10.1002/ima.70014
Ying Wu, Faming Li, Bo Xu

Ultrasound breast image classification plays a crucial role in the early detection of breast cancer, particularly in differentiating benign from malignant lesions. Traditional methods face limitations in feature extraction and global information capture, often resulting in lower accuracy for complex and noisy ultrasound images. This paper introduces a novel ultrasound breast image classification network, C-TUnet, which combines a convolutional neural network (CNN) with a Transformer architecture. In this model, the CNN module initially extracts key features from ultrasound images, followed by the Transformer module, which captures global context information to enhance classification accuracy. Experimental results demonstrate that the proposed model achieves excellent classification performance on public datasets, showing clear advantages over traditional methods. Our analysis confirms the effectiveness of combining CNN and Transformer modules—a strategy that not only boosts the accuracy and robustness of ultrasound breast image classification but also offers a reliable tool for clinical diagnostics, holding substantial potential for real-world application.

超声乳腺图像分类对乳腺癌的早期发现,特别是对良恶性病变的鉴别具有至关重要的作用。传统方法在特征提取和全局信息捕获方面存在局限性,往往导致复杂和噪声超声图像精度较低。本文介绍了一种将卷积神经网络(CNN)与Transformer结构相结合的新型超声乳房图像分类网络C-TUnet。在该模型中,CNN模块首先从超声图像中提取关键特征,然后是Transformer模块,它捕获全局上下文信息以提高分类精度。实验结果表明,该模型在公共数据集上取得了优异的分类性能,与传统方法相比具有明显的优势。我们的分析证实了CNN和Transformer模块相结合的有效性——这一策略不仅提高了超声乳房图像分类的准确性和稳健性,而且为临床诊断提供了可靠的工具,具有实际应用的巨大潜力。
{"title":"C-TUnet: A CNN-Transformer Architecture-Based Ultrasound Breast Image Classification Network","authors":"Ying Wu,&nbsp;Faming Li,&nbsp;Bo Xu","doi":"10.1002/ima.70014","DOIUrl":"https://doi.org/10.1002/ima.70014","url":null,"abstract":"<div>\u0000 \u0000 <p>Ultrasound breast image classification plays a crucial role in the early detection of breast cancer, particularly in differentiating benign from malignant lesions. Traditional methods face limitations in feature extraction and global information capture, often resulting in lower accuracy for complex and noisy ultrasound images. This paper introduces a novel ultrasound breast image classification network, C-TUnet, which combines a convolutional neural network (CNN) with a Transformer architecture. In this model, the CNN module initially extracts key features from ultrasound images, followed by the Transformer module, which captures global context information to enhance classification accuracy. Experimental results demonstrate that the proposed model achieves excellent classification performance on public datasets, showing clear advantages over traditional methods. Our analysis confirms the effectiveness of combining CNN and Transformer modules—a strategy that not only boosts the accuracy and robustness of ultrasound breast image classification but also offers a reliable tool for clinical diagnostics, holding substantial potential for real-world application.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142861490","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
YOLOv8 Outperforms Traditional CNN Models in Mammography Classification: Insights From a Multi-Institutional Dataset YOLOv8 在乳腺 X 射线摄影分类中的表现优于传统 CNN 模型:来自多机构数据集的启示
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-12-16 DOI: 10.1002/ima.70008
Erfan AkbarnezhadSany, Hossein EntezariZarch, Mohammad AlipoorKermani, Baharak Shahin, Mohsen Cheki, Aida Karami, Samaneh Zahedi, Zahra AhmadPour, Sadegh Ahmadi-Mazhin, Ali Rahimnezhad, Sahar Sayfollahi, Salar Bijari, Melika Shojaee, Seyed Masoud Rezaeijo

This study evaluates the efficacy of four deep learning methods—YOLOv8, VGG16, ResNet101, and EfficientNet—for classifying mammography images into normal, benign, and malignant categories using a large-scale, multi-institutional dataset. Each dataset was divided into training and testing groups with an 80%/20% split, ensuring that all examinations from the same patient were consistently allocated to the same split. The training set for the malignant class contained 10 220 images, the benign class 6086 images, and the normal class 8526 images. For testing, the malignant class had 1441 images, the benign class 1124 images, and the normal class 1881 images. All models were fine-tuned using transfer learning and standardized to 224 × 224 pixels with data augmentation techniques to improve robustness. Among the models, YOLOv8 demonstrated the highest performance, achieving an AUC of 93.33% for the training dataset and 91% for the testing dataset. It also exhibited superior accuracy (91.82% training, 86.68% testing), F1-score (91.11% training, 84.86% testing), and specificity (95.80% training, 93.32% testing). ResNet101, VGG16, and EfficientNet also performed well, with ResNet101 achieving an AUC of 91.67% (training) and 90.00% (testing). Grad-CAM visualizations were used to identify the regions most influential in model decision-making. This multi-model evaluation highlights YOLOv8's potential for accurately classifying mammograms, while demonstrating that all models contribute valuable insights for improving breast cancer detection. Future clinical trials will focus on refining these models to assist healthcare professionals in delivering accurate and timely diagnoses.

本研究评估了四种深度学习方法(yolov8、VGG16、ResNet101和efficientnet)在使用大规模、多机构数据集将乳房x线摄影图像分为正常、良性和恶性类别方面的效果。每个数据集被分成训练组和测试组,分成80%/20%,确保来自同一患者的所有检查都一致地分配到同一组。恶性类的训练集包含10 220张图像,良性类的训练集包含6086张图像,正常类的训练集包含8526张图像。恶性分类有1441张,良性分类有1124张,正常分类有1881张。所有模型都使用迁移学习进行微调,并使用数据增强技术将其标准化到224 × 224像素,以提高鲁棒性。在这些模型中,YOLOv8表现出了最高的性能,训练数据集的AUC为93.33%,测试数据集的AUC为91%。准确率(训练组91.82%,测试组86.68%)、f1评分(训练组91.11%,测试组84.86%)和特异性(训练组95.80%,测试组93.32%)均较优。ResNet101、VGG16和EfficientNet也表现良好,其中ResNet101的AUC为91.67%(训练)和90.00%(测试)。使用Grad-CAM可视化来识别对模型决策影响最大的区域。这一多模型评估突出了YOLOv8在准确分类乳房x线照片方面的潜力,同时表明所有模型都为提高乳腺癌检测提供了有价值的见解。未来的临床试验将侧重于改进这些模型,以帮助医疗保健专业人员提供准确和及时的诊断。
{"title":"YOLOv8 Outperforms Traditional CNN Models in Mammography Classification: Insights From a Multi-Institutional Dataset","authors":"Erfan AkbarnezhadSany,&nbsp;Hossein EntezariZarch,&nbsp;Mohammad AlipoorKermani,&nbsp;Baharak Shahin,&nbsp;Mohsen Cheki,&nbsp;Aida Karami,&nbsp;Samaneh Zahedi,&nbsp;Zahra AhmadPour,&nbsp;Sadegh Ahmadi-Mazhin,&nbsp;Ali Rahimnezhad,&nbsp;Sahar Sayfollahi,&nbsp;Salar Bijari,&nbsp;Melika Shojaee,&nbsp;Seyed Masoud Rezaeijo","doi":"10.1002/ima.70008","DOIUrl":"https://doi.org/10.1002/ima.70008","url":null,"abstract":"<div>\u0000 \u0000 <p>This study evaluates the efficacy of four deep learning methods—YOLOv8, VGG16, ResNet101, and EfficientNet—for classifying mammography images into normal, benign, and malignant categories using a large-scale, multi-institutional dataset. Each dataset was divided into training and testing groups with an 80%/20% split, ensuring that all examinations from the same patient were consistently allocated to the same split. The training set for the malignant class contained 10 220 images, the benign class 6086 images, and the normal class 8526 images. For testing, the malignant class had 1441 images, the benign class 1124 images, and the normal class 1881 images. All models were fine-tuned using transfer learning and standardized to 224 × 224 pixels with data augmentation techniques to improve robustness. Among the models, YOLOv8 demonstrated the highest performance, achieving an AUC of 93.33% for the training dataset and 91% for the testing dataset. It also exhibited superior accuracy (91.82% training, 86.68% testing), F1-score (91.11% training, 84.86% testing), and specificity (95.80% training, 93.32% testing). ResNet101, VGG16, and EfficientNet also performed well, with ResNet101 achieving an AUC of 91.67% (training) and 90.00% (testing). Grad-CAM visualizations were used to identify the regions most influential in model decision-making. This multi-model evaluation highlights YOLOv8's potential for accurately classifying mammograms, while demonstrating that all models contribute valuable insights for improving breast cancer detection. Future clinical trials will focus on refining these models to assist healthcare professionals in delivering accurate and timely diagnoses.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142861358","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DBE-Net: A Dual-Branch Boundary Enhancement Network for Pathological Image Segmentation DBE-Net:用于病理图像分割的双分支边界增强网络
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-12-16 DOI: 10.1002/ima.70017
Zefeng Liu, Zhenyu Liu

Pathological image segmentation provides support for the accurate assessment of lesion area by precisely segmenting various tissues and cellular structures in pathological images. Due to the unclear boundaries between targets and backgrounds, as well as the information loss during upsampling and downsampling operations, it remains a challenging task to identify boundary details, especially in differentiating between adjacent tissues, minor lesions, or clustered cell nuclei. In this paper, a Dual-branch Boundary Enhancement Network (DBE-Net) is proposed to improve the sensitivity of the model to the boundary. Firstly, the proposed method includes a main task and an auxiliary task. The main task focuses on segmenting the target object and the auxiliary task is dedicated to extracting boundary information. Secondly, a feature processing architecture is established which includes three modules: Feature Preservation (FP), Feature Fusion (FF), and Hybrid Attention Fusion (HAF) module. The FP module and the FF module are used to provide original information for the encoder and fuse information from every layer of the decoder. The HAF is introduced to replace the skip connections between the encoder and decoder. Finally, a boundary-dependent loss function is designed to simultaneously optimize both tasks for the dual-branch network. The proposed loss function enhances the dependence of the main task on the boundary information supplied by the auxiliary task. The proposed method has been validated on three datasets, including Glas, CoCaHis, and CoNSep dataset.

病理图像分割通过对病理图像中的各种组织和细胞结构进行精确分割,为准确评估病变区域提供支持。由于目标和背景之间的边界不明确,以及在上采样和下采样操作过程中的信息丢失,识别边界细节仍然是一项具有挑战性的任务,特别是在区分邻近组织,小病变或聚集的细胞核时。为了提高模型对边界的敏感性,本文提出了一种双分支边界增强网络(DBE-Net)。首先,该方法包括一个主任务和一个辅助任务。主要任务是对目标物体进行分割,辅助任务是提取边界信息。其次,建立了特征处理体系,该体系包括特征保持(FP)、特征融合(FF)和混合注意融合(HAF)三个模块;FP模块和FF模块为编码器提供原始信息,并融合来自解码器各层的信息。引入HAF来取代编码器和解码器之间的跳线连接。最后,设计了一个边界相关的损失函数来同时优化双分支网络的两个任务。所提出的损失函数增强了主任务对辅助任务提供的边界信息的依赖性。该方法在Glas、CoCaHis和CoNSep三个数据集上进行了验证。
{"title":"DBE-Net: A Dual-Branch Boundary Enhancement Network for Pathological Image Segmentation","authors":"Zefeng Liu,&nbsp;Zhenyu Liu","doi":"10.1002/ima.70017","DOIUrl":"https://doi.org/10.1002/ima.70017","url":null,"abstract":"<div>\u0000 \u0000 <p>Pathological image segmentation provides support for the accurate assessment of lesion area by precisely segmenting various tissues and cellular structures in pathological images. Due to the unclear boundaries between targets and backgrounds, as well as the information loss during upsampling and downsampling operations, it remains a challenging task to identify boundary details, especially in differentiating between adjacent tissues, minor lesions, or clustered cell nuclei. In this paper, a Dual-branch Boundary Enhancement Network (DBE-Net) is proposed to improve the sensitivity of the model to the boundary. Firstly, the proposed method includes a main task and an auxiliary task. The main task focuses on segmenting the target object and the auxiliary task is dedicated to extracting boundary information. Secondly, a feature processing architecture is established which includes three modules: Feature Preservation (FP), Feature Fusion (FF), and Hybrid Attention Fusion (HAF) module. The FP module and the FF module are used to provide original information for the encoder and fuse information from every layer of the decoder. The HAF is introduced to replace the skip connections between the encoder and decoder. Finally, a boundary-dependent loss function is designed to simultaneously optimize both tasks for the dual-branch network. The proposed loss function enhances the dependence of the main task on the boundary information supplied by the auxiliary task. The proposed method has been validated on three datasets, including Glas, CoCaHis, and CoNSep dataset.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142861359","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Depthwise Residual Network for Knee Meniscus Segmentation From Magnetic Resonance Imaging 基于磁共振图像的膝关节半月板深度残差网络分割
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-12-16 DOI: 10.1002/ima.70006
Anita Thengade, A. M. Rajurkar, Sanjay N. Talbar

The menisci within the knee are essential for various anatomical functions, including load-bearing, joint stability, cartilage protection, shock absorption, and lubrication. Magnetic resonance imaging (MRI) provides highly detailed images of internal organs and soft tissues, which are indispensable for physicians and radiologists assessing the meniscus. Given the multitude of images in each MRI sequence and diverse MRI data, the segmentation of the meniscus presents considerable challenges through image processing methods. The region-specific characteristics of the meniscus can vary from one image to another within the sequence. Consequently, achieving automatic and accurate segmentation of meniscus in knee MRI images is a crucial step in meniscus analysis. This paper introduces the “UNet with depthwise residual network” (DR-UNet), a depthwise convolutional neural network, designed specifically for meniscus segmentation in MRI images. The proposed architecture significantly improves the accuracy of meniscus segmentation compared to different segmentation networks. The training and testing phases utilized fat suppression turbo-spin-echo (FS TSE) MRI sequences collected from 100 distinct knee joints using a Siemens 3 Tesla MRI machine. Additionally, we employed data augmentation techniques to expand the dataset strategically, addressing the challenge of a substantial training dataset requirement. The DR-UNet model demonstrated impressive meniscus segmentation performance, achieving a Dice similarity coefficient range of 0.743–0.9646 and a Jaccard index range of 0.653–0.869, thereby showcasing its advanced segmentation capabilities.

膝关节内的半月板对各种解剖功能至关重要,包括承重、关节稳定性、软骨保护、减震和润滑。磁共振成像(MRI)提供了内部器官和软组织的高度详细的图像,这是医生和放射科医生评估半月板不可缺少的。考虑到每个MRI序列中图像的数量和MRI数据的多样性,通过图像处理方法对半月板的分割提出了相当大的挑战。在序列中,半月板的区域特异性特征可以从一个图像到另一个图像变化。因此,在膝关节MRI图像中实现半月板的自动准确分割是半月板分析的关键一步。本文介绍了一种专门用于MRI图像半月板分割的深度卷积神经网络——“带深度残差网络的UNet”(DR-UNet)。与其他分割网络相比,该结构显著提高了半月板分割的精度。训练和测试阶段使用西门子3特斯拉核磁共振成像仪从100个不同的膝关节收集脂肪抑制涡轮自旋回波(FS TSE) MRI序列。此外,我们采用数据增强技术来战略性地扩展数据集,解决了大量训练数据集需求的挑战。DR-UNet模型显示了令人印象深刻的半月板分割性能,Dice相似系数范围为0.743-0.9646,Jaccard指数范围为0.653-0.869,显示了其先进的分割能力。
{"title":"Deep Depthwise Residual Network for Knee Meniscus Segmentation From Magnetic Resonance Imaging","authors":"Anita Thengade,&nbsp;A. M. Rajurkar,&nbsp;Sanjay N. Talbar","doi":"10.1002/ima.70006","DOIUrl":"https://doi.org/10.1002/ima.70006","url":null,"abstract":"<div>\u0000 \u0000 <p>The menisci within the knee are essential for various anatomical functions, including load-bearing, joint stability, cartilage protection, shock absorption, and lubrication. Magnetic resonance imaging (MRI) provides highly detailed images of internal organs and soft tissues, which are indispensable for physicians and radiologists assessing the meniscus. Given the multitude of images in each MRI sequence and diverse MRI data, the segmentation of the meniscus presents considerable challenges through image processing methods. The region-specific characteristics of the meniscus can vary from one image to another within the sequence. Consequently, achieving automatic and accurate segmentation of meniscus in knee MRI images is a crucial step in meniscus analysis. This paper introduces the “UNet with depthwise residual network” (DR-UNet), a depthwise convolutional neural network, designed specifically for meniscus segmentation in MRI images. The proposed architecture significantly improves the accuracy of meniscus segmentation compared to different segmentation networks. The training and testing phases utilized fat suppression turbo-spin-echo (FS TSE) MRI sequences collected from 100 distinct knee joints using a Siemens 3 Tesla MRI machine. Additionally, we employed data augmentation techniques to expand the dataset strategically, addressing the challenge of a substantial training dataset requirement. The DR-UNet model demonstrated impressive meniscus segmentation performance, achieving a Dice similarity coefficient range of 0.743–0.9646 and a Jaccard index range of 0.653–0.869, thereby showcasing its advanced segmentation capabilities.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142861357","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Synthesizing Images With Annotations for Medical Image Segmentation Using Diffusion Probabilistic Model 基于扩散概率模型的医学图像分割中带注释的图像合成
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-12-14 DOI: 10.1002/ima.70007
Zengan Huang, Qinzhu Yang, Mu Tian, Yi Gao

To alleviate the burden of manual annotation, there are numerous excellent segmentation models for images segmentation being developed. However, the performance of these data-driven segmentation models is frequently constrained by the availability of samples sizes of pair medical images and segmentation annotations. Therefore, to address this challenge, this study introduces the medical image segmentation augmentation diffusion model (MEDSAD). MEDSAD solves the problem of annotation scarcity by utilizing a given simple annotation to generate paired medical images. To improve stability, we used the traditional diffusion model for this study. To exert better control over the texture synthesis in the medical images generated by MEDSAD, the texture style injection (TSI) mechanism is introduced. Additionally, we propose the feature frequency domain attention (FFDA) module to mitigate the adverse effects of high-frequency noise during generation. The efficacy of MEDSAD is substantiated through the validation of three distinct medical segmentation tasks encompassing magnetic resonance (MR) and ultrasound (US) imaging modalities, focusing on the segmentation of breast tumors, brain tumors, and nerve structures. The findings demonstrate the MEDSAD model's proficiency in synthesizing medical image pairs based on provided annotations, thereby facilitating a notable augmentation in performance for subsequent segmentation tasks. Moreover, the improvement in performance becomes greater as the quantity of synthetic available data samples increases. This underscores the robust generalization capability and efficacy intrinsic to the MEDSAD model, potentially offering avenues for future explorations in data-driven model training and research.

为了减轻手工标注的负担,人们开发了许多优秀的图像分割模型。然而,这些数据驱动的分割模型的性能经常受到成对医学图像的样本大小和分割注释的可用性的限制。因此,为了解决这一挑战,本研究引入了医学图像分割增强扩散模型(MEDSAD)。MEDSAD通过使用给定的简单注释来生成成对的医学图像,从而解决了注释稀缺性的问题。为了提高稳定性,我们使用传统的扩散模型进行研究。为了更好地控制MEDSAD生成的医学图像的纹理合成,引入了纹理样式注入(TSI)机制。此外,我们提出了特征频域注意(FFDA)模块,以减轻高频噪声在生成过程中的不利影响。MEDSAD的疗效通过三种不同的医学分割任务的验证得到证实,包括磁共振(MR)和超声(US)成像模式,重点是乳腺肿瘤、脑肿瘤和神经结构的分割。研究结果表明,MEDSAD模型能够熟练地根据提供的注释合成医学图像对,从而显著提高了后续分割任务的性能。此外,随着合成可用数据样本数量的增加,性能的提高也会越来越大。这强调了MEDSAD模型固有的强大泛化能力和有效性,为数据驱动模型训练和研究的未来探索提供了潜在的途径。
{"title":"Synthesizing Images With Annotations for Medical Image Segmentation Using Diffusion Probabilistic Model","authors":"Zengan Huang,&nbsp;Qinzhu Yang,&nbsp;Mu Tian,&nbsp;Yi Gao","doi":"10.1002/ima.70007","DOIUrl":"https://doi.org/10.1002/ima.70007","url":null,"abstract":"<p>To alleviate the burden of manual annotation, there are numerous excellent segmentation models for images segmentation being developed. However, the performance of these data-driven segmentation models is frequently constrained by the availability of samples sizes of pair medical images and segmentation annotations. Therefore, to address this challenge, this study introduces the medical image segmentation augmentation diffusion model (MEDSAD). MEDSAD solves the problem of annotation scarcity by utilizing a given simple annotation to generate paired medical images. To improve stability, we used the traditional diffusion model for this study. To exert better control over the texture synthesis in the medical images generated by MEDSAD, the texture style injection (TSI) mechanism is introduced. Additionally, we propose the feature frequency domain attention (FFDA) module to mitigate the adverse effects of high-frequency noise during generation. The efficacy of MEDSAD is substantiated through the validation of three distinct medical segmentation tasks encompassing magnetic resonance (MR) and ultrasound (US) imaging modalities, focusing on the segmentation of breast tumors, brain tumors, and nerve structures. The findings demonstrate the MEDSAD model's proficiency in synthesizing medical image pairs based on provided annotations, thereby facilitating a notable augmentation in performance for subsequent segmentation tasks. Moreover, the improvement in performance becomes greater as the quantity of synthetic available data samples increases. This underscores the robust generalization capability and efficacy intrinsic to the MEDSAD model, potentially offering avenues for future explorations in data-driven model training and research.</p>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-12-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/ima.70007","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142861090","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Three-Step Automated Segmentation Method for Early Cervical Cancer MRI Images Based on Deep Learning 基于深度学习的早期宫颈癌MRI图像三步自动分割方法
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-12-13 DOI: 10.1002/ima.23207
Liu Xiong, Chunxia Chen, Yongping Lin, Zhiyu Song, Jialin Su

Tumor detection and segmentation are essential for cervical cancer (CC) treatment and diagnosis. This study presents a model that segmented the tumor, uterus, and vagina based on deep learning automatically on magnetic resonance imaging (MRI) images of patients with CC. The tumor detection dataset consists of 68 CC patients' diffusion-weighted magnetic resonance imaging (DWI) images. The segmented dataset consists of 73 CC patients' T2-weighted imaging (T2WI) images. First, the three clear images of the patient's DWI images are detected using a single-shot multibox detector (SSD). Second, the serial number of the clearest image is obtained by scores, while the corresponding T2WI image with the same serial number is selected. Third, the selected images are segmented by employing the semantic segmentation (U-Net) model with the squeeze-and-excitation (SE) block and attention gate (SE-ATT-Unet). Three segmentation models are implemented to automatically segment the tumor, uterus, and vagina separately by adding different attention mechanisms at different locations. The target detection accuracy of the model is 92.32%, and the selection accuracy is 90.9%. The dice similarity coefficient (DSC) on the tumor is 92.20%, pixel accuracy (PA) is 93.08%, and the mean Hausdorff distance (HD) is 3.41 mm. The DSC on the uterus is 93.63%, PA is 91.75%, and the mean HD is 9.79 mm. The DSC on the vagina is 75.70%, PA is 85.46%, and the mean HD is 10.52 mm. The results show that the proposed method accurately selects images for segmentation, and the SE-ATT-Unet is effective in segmenting different regions on MRI images.

肿瘤的检测和分割对于宫颈癌的治疗和诊断至关重要。本研究提出了一种基于深度学习的CC患者磁共振成像(MRI)图像自动分割肿瘤、子宫和阴道的模型,肿瘤检测数据集由68张CC患者弥散加权磁共振成像(DWI)图像组成。分割的数据集由73例CC患者的T2WI图像组成。首先,使用单镜头多盒检测器(SSD)检测患者DWI图像的三张清晰图像。其次,通过评分获得最清晰图像的序列号,同时选择具有相同序列号的相应T2WI图像。第三,采用语义分割(U-Net)模型,结合挤压-激励(SE)块和注意门(SE- att - unet)对所选图像进行分割。实现了三种分割模型,通过在不同位置添加不同的注意机制,分别对肿瘤、子宫和阴道进行自动分割。模型的目标检测准确率为92.32%,选择准确率为90.9%。肿瘤上的骰子相似系数(DSC)为92.20%,像素精度(PA)为93.08%,平均Hausdorff距离(HD)为3.41 mm。子宫DSC为93.63%,PA为91.75%,平均HD为9.79 mm。阴道DSC为75.70%,PA为85.46%,平均HD为10.52 mm。结果表明,该方法可以准确地选择图像进行分割,se - at - unet对MRI图像的不同区域进行分割是有效的。
{"title":"A Three-Step Automated Segmentation Method for Early Cervical Cancer MRI Images Based on Deep Learning","authors":"Liu Xiong,&nbsp;Chunxia Chen,&nbsp;Yongping Lin,&nbsp;Zhiyu Song,&nbsp;Jialin Su","doi":"10.1002/ima.23207","DOIUrl":"https://doi.org/10.1002/ima.23207","url":null,"abstract":"<div>\u0000 \u0000 <p>Tumor detection and segmentation are essential for cervical cancer (CC) treatment and diagnosis. This study presents a model that segmented the tumor, uterus, and vagina based on deep learning automatically on magnetic resonance imaging (MRI) images of patients with CC. The tumor detection dataset consists of 68 CC patients' diffusion-weighted magnetic resonance imaging (DWI) images. The segmented dataset consists of 73 CC patients' T2-weighted imaging (T2WI) images. First, the three clear images of the patient's DWI images are detected using a single-shot multibox detector (SSD). Second, the serial number of the clearest image is obtained by scores, while the corresponding T2WI image with the same serial number is selected. Third, the selected images are segmented by employing the semantic segmentation (U-Net) model with the squeeze-and-excitation (SE) block and attention gate (SE-ATT-Unet). Three segmentation models are implemented to automatically segment the tumor, uterus, and vagina separately by adding different attention mechanisms at different locations. The target detection accuracy of the model is 92.32%, and the selection accuracy is 90.9%. The dice similarity coefficient (DSC) on the tumor is 92.20%, pixel accuracy (PA) is 93.08%, and the mean Hausdorff distance (HD) is 3.41 mm. The DSC on the uterus is 93.63%, PA is 91.75%, and the mean HD is 9.79 mm. The DSC on the vagina is 75.70%, PA is 85.46%, and the mean HD is 10.52 mm. The results show that the proposed method accurately selects images for segmentation, and the SE-ATT-Unet is effective in segmenting different regions on MRI images.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"35 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142861301","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International Journal of Imaging Systems and Technology
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1