首页 > 最新文献

International Journal of Imaging Systems and Technology最新文献

英文 中文
CATNet: A Cross Attention and Texture-Aware Network for Polyp Segmentation CATNet:用于息肉分割的交叉注意和纹理感知网络
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-11-16 DOI: 10.1002/ima.23220
Zhifang Deng, Yangdong Wu

Polyp segmentation is a challenging task, as some polyps exhibit similar textures to surrounding tissues, making them difficult to distinguish. Therefore, we present a parallel cross-attention and texture-aware network to address this challenging task. CATNet incorporates the parallel cross-attention mechanism, Residual Feature Fusion Module, and texture-aware module. Initially, polyp images undergo processing in our backbone network to extract multi-level polyp features. Subsequently, the parallel cross-attention mechanism sequentially captures channel and spatial dependencies across multi-scale polyp features, thereby yielding enhanced representations. These enhanced representations are then input into multiple texture-aware modules, which facilitate polyp segmentation by accentuating subtle textural disparities between polyps and the background. Finally, the Residual Feature Fusion module integrates the segmentation results with the previous layer of enhanced representations. This process serves to eliminate background noise and enhance intricate details. We assess the efficacy of our proposed method across five distinct polyp datasets. On three unseen datasets, CVC-300, CVC-ColonDB, and ETIS. We achieve mDice scores of 0.916, 0.817, and 0.777, respectively. Experimental results unequivocally demonstrate the superior performance of our approach over current models. The proposed CATNet addresses the challenges posed by textural similarities, setting a benchmark for future advancements in automated polyp detection and segmentation.

息肉分割是一项具有挑战性的任务,因为有些息肉的纹理与周围组织相似,因此难以区分。因此,我们提出了一种并行交叉注意和纹理感知网络来解决这一具有挑战性的任务。CATNet 包含并行交叉注意机制、残留特征融合模块和纹理感知模块。首先,息肉图像在主干网络中进行处理,提取多层次息肉特征。随后,并行交叉注意机制会依次捕捉多尺度息肉特征之间的通道和空间依赖关系,从而生成增强表征。这些增强表征随后被输入多个纹理感知模块,通过突出息肉与背景之间微妙的纹理差异,促进息肉分割。最后,残差特征融合模块将分割结果与前一层增强表征进行整合。这一过程可消除背景噪音,增强复杂细节。我们在五个不同的息肉数据集上评估了我们提出的方法的有效性。在三个未见数据集(CVC-300、CVC-ColonDB 和 ETIS)上,我们的 mDice 得分均为 0.5。我们的 mDice 分数分别达到 0.916、0.817 和 0.777。实验结果清楚地表明,我们的方法比现有模型性能更优越。提出的 CATNet 解决了纹理相似性带来的挑战,为未来息肉自动检测和分割的进步树立了标杆。
{"title":"CATNet: A Cross Attention and Texture-Aware Network for Polyp Segmentation","authors":"Zhifang Deng,&nbsp;Yangdong Wu","doi":"10.1002/ima.23220","DOIUrl":"https://doi.org/10.1002/ima.23220","url":null,"abstract":"<div>\u0000 \u0000 <p>Polyp segmentation is a challenging task, as some polyps exhibit similar textures to surrounding tissues, making them difficult to distinguish. Therefore, we present a parallel cross-attention and texture-aware network to address this challenging task. CATNet incorporates the parallel cross-attention mechanism, Residual Feature Fusion Module, and texture-aware module. Initially, polyp images undergo processing in our backbone network to extract multi-level polyp features. Subsequently, the parallel cross-attention mechanism sequentially captures channel and spatial dependencies across multi-scale polyp features, thereby yielding enhanced representations. These enhanced representations are then input into multiple texture-aware modules, which facilitate polyp segmentation by accentuating subtle textural disparities between polyps and the background. Finally, the Residual Feature Fusion module integrates the segmentation results with the previous layer of enhanced representations. This process serves to eliminate background noise and enhance intricate details. We assess the efficacy of our proposed method across five distinct polyp datasets. On three unseen datasets, CVC-300, CVC-ColonDB, and ETIS. We achieve mDice scores of 0.916, 0.817, and 0.777, respectively. Experimental results unequivocally demonstrate the superior performance of our approach over current models. The proposed CATNet addresses the challenges posed by textural similarities, setting a benchmark for future advancements in automated polyp detection and segmentation.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 6","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142664822","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Predicting the Early Detection of Breast Cancer Using Hybrid Machine Learning Systems and Thermographic Imaging 利用混合机器学习系统和热成像技术预测乳腺癌的早期发现
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-11-16 DOI: 10.1002/ima.23211
Mohammad Mehdi Hosseini, Zahra Mosahebeh, Somenath Chakraborty, Abdorreza Alavi Gharahbagh

Breast cancer is a leading cause of mortality among women, emphasizing the critical need for precise early detection and prognosis. However, conventional methods often struggle to differentiate precancerous lesions or tailor treatments effectively. Thermal imaging, capturing subtle temperature variations, presents a promising avenue for non-invasive cancer detection. While some studies explore thermography for breast cancer detection, integrating it with advanced machine learning for early diagnosis and personalized prediction remains relatively unexplored. This study proposes a novel hybrid machine learning system (HMLS) incorporating deep autoencoder techniques for automated early detection and prognostic stratification of breast cancer patients. By exploiting the temporal dynamics of thermographic data, this approach offers a more comprehensive analysis than static single-frame approaches. Data processing involves splitting the dataset for training and testing. A predominant infrared image was selected, and matrix factorization was applied to capture temperature changes over time. Integration of convex factor analysis and bell-curve membership function embedding for dimensionality reduction and feature extraction. The autoencoder deep neural network further reduces dimensionality. HMLS model development included feature selection and optimization of survival prediction algorithms through cross-validation. Model performance was assessed using accuracy and F-measure metrics. HMLS, integrating clinical data, achieved 81.6% accuracy, surpassing 77.6% using only convex-NMF. The best classifier attained 83.2% accuracy on test data. This study demonstrates the effectiveness of thermographic imaging and HMLS for accurate early detection and personalized prediction of breast cancer. The proposed framework holds promise for enhancing patient care and potentially reducing mortality rates.

乳腺癌是导致妇女死亡的主要原因之一,因此强调精确的早期检测和预后至关重要。然而,传统方法往往难以区分癌前病变或有效地调整治疗方法。热成像技术能捕捉微妙的温度变化,是一种很有前景的非侵入性癌症检测方法。虽然一些研究探讨了热成像技术在乳腺癌检测中的应用,但将其与先进的机器学习技术相结合,用于早期诊断和个性化预测的研究仍相对较少。本研究提出了一种新型混合机器学习系统(HMLS),该系统结合了深度自动编码器技术,用于乳腺癌患者的自动早期检测和预后分层。通过利用热成像数据的时间动态,该方法可提供比静态单帧方法更全面的分析。数据处理包括拆分数据集进行训练和测试。选择一个主要的红外图像,并应用矩阵因式分解来捕捉温度随时间的变化。整合凸因子分析和钟形曲线成员函数嵌入,以实现降维和特征提取。自动编码器深度神经网络进一步降低了维度。HMLS 模型开发包括特征选择和通过交叉验证优化生存预测算法。使用准确率和 F-measure 指标评估模型性能。整合临床数据的 HMLS 准确率达到 81.6%,超过了仅使用凸 NMF 的 77.6%。最佳分类器在测试数据上达到了 83.2% 的准确率。这项研究证明了热成像和 HMLS 在准确早期检测和个性化预测乳腺癌方面的有效性。所提出的框架有望加强对患者的护理,并有可能降低死亡率。
{"title":"Predicting the Early Detection of Breast Cancer Using Hybrid Machine Learning Systems and Thermographic Imaging","authors":"Mohammad Mehdi Hosseini,&nbsp;Zahra Mosahebeh,&nbsp;Somenath Chakraborty,&nbsp;Abdorreza Alavi Gharahbagh","doi":"10.1002/ima.23211","DOIUrl":"https://doi.org/10.1002/ima.23211","url":null,"abstract":"<div>\u0000 \u0000 <p>Breast cancer is a leading cause of mortality among women, emphasizing the critical need for precise early detection and prognosis. However, conventional methods often struggle to differentiate precancerous lesions or tailor treatments effectively. Thermal imaging, capturing subtle temperature variations, presents a promising avenue for non-invasive cancer detection. While some studies explore thermography for breast cancer detection, integrating it with advanced machine learning for early diagnosis and personalized prediction remains relatively unexplored. This study proposes a novel hybrid machine learning system (HMLS) incorporating deep autoencoder techniques for automated early detection and prognostic stratification of breast cancer patients. By exploiting the temporal dynamics of thermographic data, this approach offers a more comprehensive analysis than static single-frame approaches. Data processing involves splitting the dataset for training and testing. A predominant infrared image was selected, and matrix factorization was applied to capture temperature changes over time. Integration of convex factor analysis and bell-curve membership function embedding for dimensionality reduction and feature extraction. The autoencoder deep neural network further reduces dimensionality. HMLS model development included feature selection and optimization of survival prediction algorithms through cross-validation. Model performance was assessed using accuracy and F-measure metrics. HMLS, integrating clinical data, achieved 81.6% accuracy, surpassing 77.6% using only convex-NMF. The best classifier attained 83.2% accuracy on test data. This study demonstrates the effectiveness of thermographic imaging and HMLS for accurate early detection and personalized prediction of breast cancer. The proposed framework holds promise for enhancing patient care and potentially reducing mortality rates.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 6","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142664749","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
VMC-UNet: A Vision Mamba-CNN U-Net for Tumor Segmentation in Breast Ultrasound Image VMC-UNet:用于乳腺超声图像肿瘤分割的视觉曼巴-CNN U-Net
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-11-14 DOI: 10.1002/ima.23222
Dongyue Wang, Weiyu Zhao, Kaixuan Cui, Yi Zhu

Breast cancer remains one of the most significant health threats to women, making precise segmentation of target tumors critical for early clinical intervention and postoperative monitoring. While numerous convolutional neural networks (CNNs) and vision transformers have been developed to segment breast tumors from ultrasound images, both architectures encounter difficulties in effectively modeling long-range dependencies, which are essential for accurate segmentation. Drawing inspiration from the Mamba architecture, we introduce the Vision Mamba-CNN U-Net (VMC-UNet) for breast tumor segmentation. This innovative hybrid framework merges the long-range dependency modeling capabilities of Mamba with the detailed local representation power of CNNs. A key feature of our approach is the implementation of a residual connection method within the U-Net architecture, utilizing the visual state space (VSS) module to extract long-range dependency features from convolutional feature maps effectively. Additionally, to better integrate texture and structural features, we have designed a bilinear multi-scale attention module (BMSA), which significantly enhances the network's ability to capture and utilize intricate feature details across multiple scales. Extensive experiments conducted on three public datasets demonstrate that the proposed VMC-UNet surpasses other state-of-the-art methods in breast tumor segmentation, achieving Dice coefficients of 81.52% for BUSI, 88.00% for BUS, and 88.96% for STU. The source code is accessible at https://github.com/windywindyw/VMC-UNet.

乳腺癌仍然是对妇女健康威胁最大的疾病之一,因此目标肿瘤的精确分割对于早期临床干预和术后监测至关重要。虽然已经开发了许多卷积神经网络(CNN)和视觉变换器来分割超声波图像中的乳腺肿瘤,但这两种架构在有效建模长程依赖性方面都遇到了困难,而长程依赖性对精确分割至关重要。我们从 Mamba 架构中汲取灵感,推出了用于乳腺肿瘤分割的视觉 Mamba-CNN U-Net (VMC-UNet)。这一创新的混合框架融合了 Mamba 的远距离依赖建模能力和 CNN 的详细局部表示能力。我们的方法的一个主要特点是在 U-Net 架构中实施了残差连接方法,利用视觉状态空间(VSS)模块从卷积特征图中有效提取长距离依赖特征。此外,为了更好地整合纹理和结构特征,我们还设计了双线性多尺度注意力模块(BMSA),大大增强了网络捕捉和利用多尺度复杂特征细节的能力。在三个公开数据集上进行的广泛实验表明,所提出的 VMC-UNet 在乳腺肿瘤分割方面超越了其他最先进的方法,其 BUSI、BUS 和 STU 的 Dice 系数分别达到了 81.52%、88.00% 和 88.96%。源代码可从 https://github.com/windywindyw/VMC-UNet 获取。
{"title":"VMC-UNet: A Vision Mamba-CNN U-Net for Tumor Segmentation in Breast Ultrasound Image","authors":"Dongyue Wang,&nbsp;Weiyu Zhao,&nbsp;Kaixuan Cui,&nbsp;Yi Zhu","doi":"10.1002/ima.23222","DOIUrl":"https://doi.org/10.1002/ima.23222","url":null,"abstract":"<div>\u0000 \u0000 <p>Breast cancer remains one of the most significant health threats to women, making precise segmentation of target tumors critical for early clinical intervention and postoperative monitoring. While numerous convolutional neural networks (CNNs) and vision transformers have been developed to segment breast tumors from ultrasound images, both architectures encounter difficulties in effectively modeling long-range dependencies, which are essential for accurate segmentation. Drawing inspiration from the Mamba architecture, we introduce the Vision Mamba-CNN U-Net (VMC-UNet) for breast tumor segmentation. This innovative hybrid framework merges the long-range dependency modeling capabilities of Mamba with the detailed local representation power of CNNs. A key feature of our approach is the implementation of a residual connection method within the U-Net architecture, utilizing the visual state space (VSS) module to extract long-range dependency features from convolutional feature maps effectively. Additionally, to better integrate texture and structural features, we have designed a bilinear multi-scale attention module (BMSA), which significantly enhances the network's ability to capture and utilize intricate feature details across multiple scales. Extensive experiments conducted on three public datasets demonstrate that the proposed VMC-UNet surpasses other state-of-the-art methods in breast tumor segmentation, achieving Dice coefficients of 81.52% for BUSI, 88.00% for BUS, and 88.96% for STU. The source code is accessible at https://github.com/windywindyw/VMC-UNet.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 6","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-11-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142641714","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Suppression of the Tissue Component With the Total Least-Squares Algorithm to Improve Second Harmonic Imaging of Ultrasound Contrast Agents 用最小二乘总算法抑制组织成分,改善超声造影剂的二次谐波成像效果
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-11-14 DOI: 10.1002/ima.23218
Jingying Zhu, Yufeng Zhang, Bingbing He, Zhiyao Li, Li Xiong, Xun Lang

The second harmonic (SH) of ultrasound contrast agents (UCAs) is widely used in contrast-enhanced ultrasound imaging; however, is affected by the nonlinearity of surrounding tissue. Suppression of the tissue component based on the total least-squares (STLS) algorithm is proposed to enhance the SH imaging of UCAs. The image blocks of pulse-inversion-based SH images before and after UCA injections are set as the reference and input of the total least-squares model, respectively. The optimal coefficients of the model are obtained by minimizing the Frobenius norm of perturbations in the input and output signals. After processing all image blocks, the complete SH image of UCAs is obtained by subtracting the optimal output of the model (i.e., the estimated tissue SH image) from the SH image after UCA injection. Simulation and in vivo experiments confirm that the STLS approach offers clearer capillaries. For in vivo experiments, the STLS-based contrast-to-tissue ratios and contrasts increase by 26.90% and 56.27%, as well as 26.99% and 56.43%, respectively, compared with those based on bubble-echo deconvolution and pulse inversion bubble-wavelet imaging methods. The STLS approach enhances the SH imaging of UCAs by effectively suppressing more tissue SH components, having the potential to provide more accurate diagnostic information for clinical applications.

超声造影剂(UCAs)的二次谐波(SH)被广泛应用于造影剂增强超声成像,但它会受到周围组织非线性的影响。为了增强 UCA 的二次谐波成像,提出了基于总最小二乘(STLS)算法的组织成分抑制方法。UCA 注射前后基于脉冲反转的 SH 图像块分别被设定为总最小二乘模型的参考和输入。通过最小化输入和输出信号中扰动的 Frobenius 准则,获得模型的最佳系数。处理完所有图像块后,从 UCA 注射后的 SH 图像中减去模型的最佳输出(即估计的组织 SH 图像),就得到了完整的 UCA SH 图像。模拟和活体实验证实,STLS 方法能提供更清晰的毛细血管。在活体实验中,与基于气泡回波解卷积和脉冲反转气泡小波成像方法的对比度和对比度相比,基于 STLS 的对比度和对比度分别提高了 26.90% 和 56.27%,以及 26.99% 和 56.43%。STLS 方法能有效抑制更多的组织 SH 成分,从而增强 UCA 的 SH 成像,有望为临床应用提供更准确的诊断信息。
{"title":"Suppression of the Tissue Component With the Total Least-Squares Algorithm to Improve Second Harmonic Imaging of Ultrasound Contrast Agents","authors":"Jingying Zhu,&nbsp;Yufeng Zhang,&nbsp;Bingbing He,&nbsp;Zhiyao Li,&nbsp;Li Xiong,&nbsp;Xun Lang","doi":"10.1002/ima.23218","DOIUrl":"https://doi.org/10.1002/ima.23218","url":null,"abstract":"<div>\u0000 \u0000 <p>The second harmonic (SH) of ultrasound contrast agents (UCAs) is widely used in contrast-enhanced ultrasound imaging; however, is affected by the nonlinearity of surrounding tissue. Suppression of the tissue component based on the total least-squares (STLS) algorithm is proposed to enhance the SH imaging of UCAs. The image blocks of pulse-inversion-based SH images before and after UCA injections are set as the reference and input of the total least-squares model, respectively. The optimal coefficients of the model are obtained by minimizing the Frobenius norm of perturbations in the input and output signals. After processing all image blocks, the complete SH image of UCAs is obtained by subtracting the optimal output of the model (i.e., the estimated tissue SH image) from the SH image after UCA injection. Simulation and in vivo experiments confirm that the STLS approach offers clearer capillaries. For in vivo experiments, the STLS-based contrast-to-tissue ratios and contrasts increase by 26.90% and 56.27%, as well as 26.99% and 56.43%, respectively, compared with those based on bubble-echo deconvolution and pulse inversion bubble-wavelet imaging methods. The STLS approach enhances the SH imaging of UCAs by effectively suppressing more tissue SH components, having the potential to provide more accurate diagnostic information for clinical applications.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 6","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-11-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142641715","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Segmentation and Classification of Breast Masses From the Whole Mammography Images Using Transfer Learning and BI-RADS Characteristics 利用迁移学习和 BI-RADS 特征从整个乳腺 X 射线照相图像中对乳腺肿块进行分割和分类
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-11-13 DOI: 10.1002/ima.23216
Hayette Oudjer, Assia Cherfa, Yazid Cherfa, Noureddine Belkhamsa

Breast cancer is the most prevalent cancer among women worldwide, highlighting the critical need for its accurate detection and early diagnosis. In this context, the segmentation of breast masses (the most common symptom of breast cancer) plays a crucial role in analyzing mammographic images. In addition, in image processing, the analysis of mammographic images is very common, but certain combinations of mathematical tools have never been exploited. We propose a computer-aided diagnosis (CAD) system designed with different and new algorithm combinations for the segmentation and classification of breast masses based on the Breast Imaging-Reporting and Data System (BI-RADS) lexicon. The image is initially divided into superpixels using the simple linear iterative clustering (SLIC) algorithm. Fine-tuning of ResNet50, EfficientNetB2, MobileNetV2, and InceptionV3 models is employed to extract features from superpixels. The classification of each superpixel as background or breast mass is performed by feeding the extracted features into a support vector machine (SVM) classifier, resulting in an accurate primary segmentation for breast masses, refined by the GrabCut algorithm with automated initialization. Finally, we extract contour, texture, and shape parameters from the segmented regions for the classification of masses into BI-Rads 2, 3, 4, and 5 using the gradient boost (GB) classifier while also examining the impact of the surrounding tissue. The proposed method was evaluated on the INBreast database, achieving a Dice score of 87.65% and a sensitivity of 87.96% for segmentation. For classification, we obtained a sensitivity of 88.66%, a precision of 90.51%, and an area under the curve (AUC) of 97.8%. The CAD system demonstrates high accuracy in both the segmentation and classification of breast masses, providing a reliable tool for aiding breast cancer diagnosis using the BI-Rads lexicon. The study also showed that the surrounding tissue has an impact on classification, leading to the importance of choosing the right size of ROIs.

乳腺癌是全球妇女中发病率最高的癌症,因此准确检测和早期诊断乳腺癌至关重要。在这种情况下,乳腺肿块(乳腺癌最常见的症状)的分割在乳腺 X 射线图像分析中起着至关重要的作用。此外,在图像处理中,乳腺 X 射线图像的分析非常常见,但某些数学工具的组合却从未被利用过。我们提出了一种计算机辅助诊断(CAD)系统,该系统根据乳腺成像-报告和数据系统(BI-RADS)词典设计了不同的新算法组合,用于乳腺肿块的分割和分类。首先使用简单线性迭代聚类(SLIC)算法将图像划分为超像素。然后对 ResNet50、EfficientNetB2、MobileNetV2 和 InceptionV3 模型进行微调,从超像素中提取特征。通过将提取的特征输入支持向量机(SVM)分类器,将每个超像素分类为背景或乳房肿块,最后通过自动初始化的 GrabCut 算法对乳房肿块进行精确的初级分割。最后,我们从分割区域提取轮廓、纹理和形状参数,使用梯度提升(GB)分类器将肿块分类为 BI-Rads2、3、4 和 5,同时还检查了周围组织的影响。我们在 INBreast 数据库中对所提出的方法进行了评估,结果显示该方法的 Dice 得分为 87.65%,分割灵敏度为 87.96%。在分类方面,我们获得了 88.66% 的灵敏度、90.51% 的精确度和 97.8% 的曲线下面积 (AUC)。CAD 系统在乳腺肿块的分割和分类方面都表现出很高的准确性,为使用 BI-Rads 词典辅助乳腺癌诊断提供了可靠的工具。研究还表明,周围组织对分类也有影响,因此选择正确的 ROI 大小非常重要。
{"title":"Segmentation and Classification of Breast Masses From the Whole Mammography Images Using Transfer Learning and BI-RADS Characteristics","authors":"Hayette Oudjer,&nbsp;Assia Cherfa,&nbsp;Yazid Cherfa,&nbsp;Noureddine Belkhamsa","doi":"10.1002/ima.23216","DOIUrl":"https://doi.org/10.1002/ima.23216","url":null,"abstract":"<div>\u0000 \u0000 <p>Breast cancer is the most prevalent cancer among women worldwide, highlighting the critical need for its accurate detection and early diagnosis. In this context, the segmentation of breast masses (the most common symptom of breast cancer) plays a crucial role in analyzing mammographic images. In addition, in image processing, the analysis of mammographic images is very common, but certain combinations of mathematical tools have never been exploited. We propose a computer-aided diagnosis (CAD) system designed with different and new algorithm combinations for the segmentation and classification of breast masses based on the Breast Imaging-Reporting and Data System (BI-RADS) lexicon. The image is initially divided into superpixels using the simple linear iterative clustering (SLIC) algorithm. Fine-tuning of ResNet50, EfficientNetB2, MobileNetV2, and InceptionV3 models is employed to extract features from superpixels. The classification of each superpixel as background or breast mass is performed by feeding the extracted features into a support vector machine (SVM) classifier, resulting in an accurate primary segmentation for breast masses, refined by the GrabCut algorithm with automated initialization. Finally, we extract contour, texture, and shape parameters from the segmented regions for the classification of masses into BI-Rads 2, 3, 4, and 5 using the gradient boost (GB) classifier while also examining the impact of the surrounding tissue. The proposed method was evaluated on the INBreast database, achieving a Dice score of 87.65% and a sensitivity of 87.96% for segmentation. For classification, we obtained a sensitivity of 88.66%, a precision of 90.51%, and an area under the curve (AUC) of 97.8%. The CAD system demonstrates high accuracy in both the segmentation and classification of breast masses, providing a reliable tool for aiding breast cancer diagnosis using the BI-Rads lexicon. The study also showed that the surrounding tissue has an impact on classification, leading to the importance of choosing the right size of ROIs.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 6","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142642168","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dual-Low Technology in Coronary and Abdominal CT Angiography: A Comparative Study of Deep Learning Image Reconstruction and Adaptive Statistic Iterative Reconstruction-Veo 冠状动脉和腹部 CT 血管造影中的双低技术:深度学习图像重建与自适应统计迭代重建的比较研究-Veo
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-11-13 DOI: 10.1002/ima.23217
Zhanao Meng, Qing Xiang, Jian Cao, Yahao Guo, Sisi Deng, Tao Luo, Yue Zhang, Ke Zhang, Xuan Zhu, Kun Ma, Xiaohong Wang, Jie Qin

To investigate the application advantages of dual-low technology (low radiation dose and low contrast agent dose) in deep learning image reconstruction (DLIR) compared to the adaptive statistical iterative reconstruction-Veo (ASIR-V) standard protocol when combing coronary computed tomography angiography (CCTA) and abdominal computed tomography angiography (ACTA). Sixty patients who underwent CCTA and ACTA were recruited. Thirty patients with low body mass index (BMI) (< 24 kg/m2, Group A, standard protocol) were reconstructed using 60% ASIR-V, and 30 patients with high BMI (> 24 kg/m2, Group B, dual-low protocol) were reconstructed using DLIR at high strength (DLIR-H). The effective dose and contrast agent dose were recorded. The CT values, standard deviations, signal-to-noise ratio (SNR), and contrast-to-noise ratio (CNR) were measured. The subjective evaluation criteria were scored by two radiologists using a blind Likert 5-point scale. The general data, objective evaluation, and subjective scores between both groups were compared using corresponding test methods. The consistency of objective and subjective evaluations between the two radiologists were analyzed using Kappa tests. Group B showed a remarkable 44.6% reduction in mean effective dose (p < 0.01) and a 20.3% decrease in contrast agent dose compared to Group A (p < 0.01). The DLIR-H demonstrated the smallest standard deviations and highest SNR and CNR values (p < 0.01). The subjective score of DLIR-H was the highest (p < 0.01), and there was good consistency between the two radiologists in the subjective scoring of CCTA and ACTA image quality (κ = 0.751 ~ 0.919, p < 0.01). In combined CCTA and ACTA protocols, DLIR can significantly reduce the effective dose and contrast agent dose compared to ASIR-V while maintaining good image quality.

研究在冠状动脉计算机断层扫描(CCTA)和腹部计算机断层扫描(ACTA)联合检查时,深度学习图像重建(DLIR)中的双低技术(低辐射剂量和低造影剂剂量)与自适应统计迭代重建-Veo(ASIR-V)标准方案相比的应用优势。研究人员招募了 60 名接受过 CCTA 和 ACTA 检查的患者。30 名低体重指数(BMI)患者(24 kg/m2,A 组,标准方案)使用 60% ASIR-V 重建,30 名高体重指数患者(24 kg/m2,B 组,双低方案)使用高强度 DLIR(DLIR-H)重建。记录了有效剂量和造影剂剂量。测量了 CT 值、标准偏差、信噪比(SNR)和对比度-噪声比(CNR)。主观评价标准由两名放射科医生采用李克特 5 点盲法评分。采用相应的测试方法对两组患者的一般数据、客观评价和主观评分进行比较。两位放射科医生的客观和主观评价的一致性采用 Kappa 检验进行分析。与 A 组相比,B 组的平均有效剂量明显减少了 44.6%(p < 0.01),造影剂剂量减少了 20.3%(p < 0.01)。DLIR-H 组的标准偏差最小,信噪比(SNR)和有线信噪比(CNR)最高(p < 0.01)。DLIR-H 的主观评分最高(p <0.01),两位放射医师对 CCTA 和 ACTA 图像质量的主观评分具有良好的一致性(κ = 0.751 ~ 0.919,p <0.01)。在联合 CCTA 和 ACTA 方案中,与 ASIR-V 相比,DLIR 可显著降低有效剂量和造影剂剂量,同时保持良好的图像质量。
{"title":"Dual-Low Technology in Coronary and Abdominal CT Angiography: A Comparative Study of Deep Learning Image Reconstruction and Adaptive Statistic Iterative Reconstruction-Veo","authors":"Zhanao Meng,&nbsp;Qing Xiang,&nbsp;Jian Cao,&nbsp;Yahao Guo,&nbsp;Sisi Deng,&nbsp;Tao Luo,&nbsp;Yue Zhang,&nbsp;Ke Zhang,&nbsp;Xuan Zhu,&nbsp;Kun Ma,&nbsp;Xiaohong Wang,&nbsp;Jie Qin","doi":"10.1002/ima.23217","DOIUrl":"https://doi.org/10.1002/ima.23217","url":null,"abstract":"<div>\u0000 \u0000 <p>To investigate the application advantages of dual-low technology (low radiation dose and low contrast agent dose) in deep learning image reconstruction (DLIR) compared to the adaptive statistical iterative reconstruction-Veo (ASIR-V) standard protocol when combing coronary computed tomography angiography (CCTA) and abdominal computed tomography angiography (ACTA). Sixty patients who underwent CCTA and ACTA were recruited. Thirty patients with low body mass index (BMI) (&lt; 24 kg/m<sup>2</sup>, Group A, standard protocol) were reconstructed using 60% ASIR-V, and 30 patients with high BMI (&gt; 24 kg/m<sup>2</sup>, Group B, dual-low protocol) were reconstructed using DLIR at high strength (DLIR-H). The effective dose and contrast agent dose were recorded. The CT values, standard deviations, signal-to-noise ratio (SNR), and contrast-to-noise ratio (CNR) were measured. The subjective evaluation criteria were scored by two radiologists using a blind Likert 5-point scale. The general data, objective evaluation, and subjective scores between both groups were compared using corresponding test methods. The consistency of objective and subjective evaluations between the two radiologists were analyzed using Kappa tests. Group B showed a remarkable 44.6% reduction in mean effective dose (<i>p</i> &lt; 0.01) and a 20.3% decrease in contrast agent dose compared to Group A (<i>p</i> &lt; 0.01). The DLIR-H demonstrated the smallest standard deviations and highest SNR and CNR values (<i>p</i> &lt; 0.01). The subjective score of DLIR-H was the highest (<i>p</i> &lt; 0.01), and there was good consistency between the two radiologists in the subjective scoring of CCTA and ACTA image quality (κ = 0.751 ~ 0.919, <i>p</i> &lt; 0.01). In combined CCTA and ACTA protocols, DLIR can significantly reduce the effective dose and contrast agent dose compared to ASIR-V while maintaining good image quality.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 6","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142642169","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-Deep Learning Approach With Transfer Learning for 7-Stages Diabetic Retinopathy Classification 利用迁移学习的多深度学习法进行 7 级糖尿病视网膜病变分类
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-11-11 DOI: 10.1002/ima.23213
Abdul Qadir Khan, Guangmin Sun, Majdi Khalid, Majed Farrash, Anas Bilal

Proposed novel investigation focused on leveraging an innovative diabetic retinopathy (DR) dataset comprising seven severity stages, an approach not previously examined. By capitalizing on this unique resource, this study′s findings set a new benchmark for DR classification, highlighting the transformative potential of incorporating advanced data into AI models. This study developed a Vgg16 transfer learning model and gauged its performance against established algorithms including Vgg-19, AlexNet, and SqueezeNet. Remarkably, our results achieved accuracy rates of 96.95, 96.75, 96.09, and 92.96, respectively, emphasizing the contribution of our work. We strongly emphasized comprehensive severity rating, yielding perfect and impressive F1-scores of 1.00 for “mild NPDR” and 97.00 for “no DR signs.” The Vgg16-TL model consistently outperformed other models across all severity levels, reinforcing the value of our discoveries. Our deep learning training process, carefully selecting a learning rate of 1e-05, allowed continuous refinements in training and validation accuracy. Beyond metrics, our investigation underscores the vital clinical importance of precise DR classification for preventing vision loss. This study conclusively establishes deep learning as a powerful transformative tool for developing effective DR algorithms with the potential to improve patient outcomes and advance ophthalmology standards.

拟议的新调查侧重于利用创新的糖尿病视网膜病变(DR)数据集,该数据集包括七个严重程度阶段,这是一种以前从未研究过的方法。通过利用这一独特的资源,本研究的发现为糖尿病视网膜病变分类设定了新的基准,凸显了将先进数据纳入人工智能模型的变革潜力。本研究开发了一个 Vgg16 转移学习模型,并将其性能与 Vgg-19、AlexNet 和 SqueezeNet 等成熟算法进行了比较。值得注意的是,我们的结果分别达到了 96.95、96.75、96.09 和 92.96 的准确率,这突出了我们工作的贡献。我们非常重视综合严重程度评级,"轻度 NPDR "和 "无 DR 征兆 "的 F1 分数分别达到了 1.00 和 97.00,令人印象深刻。在所有严重程度等级中,Vgg16-TL 模型的表现始终优于其他模型,这加强了我们发现的价值。我们的深度学习训练过程精心选择了 1e-05 的学习率,从而不断提高了训练和验证的准确性。除了指标之外,我们的研究还强调了精确 DR 分类对预防视力丧失的重要临床意义。这项研究最终证明,深度学习是开发有效 DR 算法的强大变革工具,具有改善患者预后和提高眼科标准的潜力。
{"title":"Multi-Deep Learning Approach With Transfer Learning for 7-Stages Diabetic Retinopathy Classification","authors":"Abdul Qadir Khan,&nbsp;Guangmin Sun,&nbsp;Majdi Khalid,&nbsp;Majed Farrash,&nbsp;Anas Bilal","doi":"10.1002/ima.23213","DOIUrl":"https://doi.org/10.1002/ima.23213","url":null,"abstract":"<div>\u0000 \u0000 <p>Proposed novel investigation focused on leveraging an innovative diabetic retinopathy (DR) dataset comprising seven severity stages, an approach not previously examined. By capitalizing on this unique resource, this study′s findings set a new benchmark for DR classification, highlighting the transformative potential of incorporating advanced data into AI models. This study developed a Vgg16 transfer learning model and gauged its performance against established algorithms including Vgg-19, AlexNet, and SqueezeNet. Remarkably, our results achieved accuracy rates of 96.95, 96.75, 96.09, and 92.96, respectively, emphasizing the contribution of our work. We strongly emphasized comprehensive severity rating, yielding perfect and impressive F1-scores of 1.00 for “mild NPDR” and 97.00 for “no DR signs.” The Vgg16-TL model consistently outperformed other models across all severity levels, reinforcing the value of our discoveries. Our deep learning training process, carefully selecting a learning rate of 1e-05, allowed continuous refinements in training and validation accuracy. Beyond metrics, our investigation underscores the vital clinical importance of precise DR classification for preventing vision loss. This study conclusively establishes deep learning as a powerful transformative tool for developing effective DR algorithms with the potential to improve patient outcomes and advance ophthalmology standards.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 6","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142641976","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
YOLOv7-XAI: Multi-Class Skin Lesion Diagnosis Using Explainable AI With Fair Decision Making YOLOv7-XAI:利用可解释的人工智能进行多类皮肤病变诊断并做出公平决策
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-11-11 DOI: 10.1002/ima.23214
Nirmala Veeramani, Premaladha Jayaraman

Skin cancer, a prevalent and potentially life-threatening condition, demands accurate and timely detection for effective intervention. It is an uncontrolled growth of abnormal cells in the human body. Studies are underway to determine if a skin lesion is benign (non-cancerous) or malignant (cancerous), but the biggest challenge for a doctor is determining the type of skin cancer. As a result, determining the type of tumour is crucial for the right course of treatment. In this study, we introduce a groundbreaking approach to multi-class skin cancer detection by harnessing the power of Explainable Artificial Intelligence (XAI) in conjunction with a customised You Only Look Once (YOLOv7) architecture. Our research focuses on enhancing the YOLOv7 framework to accurately discern 8 different skin cancer classes, including melanoma, basal cell carcinoma, and squamous cell carcinoma. The YOLOv7 model is the robust backbone, enriched with features tailored for precise multi-class classification. Concurrently, integrating XAI elements, Local Interpretable Modal-agnostic Explanations (LIME) and Shapley Additive Explanations (SHAP) ensures transparent decision-making processes, enabling healthcare professionals to interpret and trust the model's predictions. This innovative synergy between YOLOv7 and XAI heralds a new era in interpretability, resulting in high-performance skin cancer diagnostics. The obtained results are 96.8%, 94.2%, 95.6%, and 95.8%, evaluated with popular quantitative metrics such as accuracy, precision, recall, and F1 score, respectively.

皮肤癌是一种普遍存在并可能危及生命的疾病,需要准确及时的检测以进行有效干预。皮肤癌是人体内异常细胞不受控制的生长。目前正在进行研究,以确定皮肤病变是良性(非癌症)还是恶性(癌症),但医生面临的最大挑战是确定皮肤癌的类型。因此,确定肿瘤类型对于正确治疗至关重要。在本研究中,我们利用可解释人工智能(XAI)的强大功能,结合定制的 "只看一眼"(YOLOv7)架构,推出了一种开创性的多类皮肤癌检测方法。我们的研究重点是增强 YOLOv7 框架,以准确分辨 8 种不同的皮肤癌类别,包括黑色素瘤、基底细胞癌和鳞状细胞癌。YOLOv7 模型是稳健的骨干,富含为精确多类分类量身定制的特征。同时,整合 XAI 元素、本地可解释模态诊断解释(LIME)和夏普利加法解释(SHAP)可确保决策过程透明,使医疗保健专业人员能够解释和信任模型的预测。YOLOv7 和 XAI 之间的这种创新协同作用预示着可解释性的新时代即将到来,从而带来高性能的皮肤癌诊断。根据准确率、精确度、召回率和 F1 分数等常用量化指标进行评估,结果分别为 96.8%、94.2%、95.6% 和 95.8%。
{"title":"YOLOv7-XAI: Multi-Class Skin Lesion Diagnosis Using Explainable AI With Fair Decision Making","authors":"Nirmala Veeramani,&nbsp;Premaladha Jayaraman","doi":"10.1002/ima.23214","DOIUrl":"https://doi.org/10.1002/ima.23214","url":null,"abstract":"<div>\u0000 \u0000 <p>Skin cancer, a prevalent and potentially life-threatening condition, demands accurate and timely detection for effective intervention. It is an uncontrolled growth of abnormal cells in the human body. Studies are underway to determine if a skin lesion is benign (non-cancerous) or malignant (cancerous), but the biggest challenge for a doctor is determining the type of skin cancer. As a result, determining the type of tumour is crucial for the right course of treatment. In this study, we introduce a groundbreaking approach to multi-class skin cancer detection by harnessing the power of Explainable Artificial Intelligence (XAI) in conjunction with a customised You Only Look Once (YOLOv7) architecture. Our research focuses on enhancing the YOLOv7 framework to accurately discern 8 different skin cancer classes, including melanoma, basal cell carcinoma, and squamous cell carcinoma. The YOLOv7 model is the robust backbone, enriched with features tailored for precise multi-class classification. Concurrently, integrating XAI elements, Local Interpretable Modal-agnostic Explanations (LIME) and Shapley Additive Explanations (SHAP) ensures transparent decision-making processes, enabling healthcare professionals to interpret and trust the model's predictions. This innovative synergy between YOLOv7 and XAI heralds a new era in interpretability, resulting in high-performance skin cancer diagnostics. The obtained results are 96.8%, 94.2%, 95.6%, and 95.8%, evaluated with popular quantitative metrics such as accuracy, precision, recall, and F1 score, respectively.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 6","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142641975","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Risk Stratification Study of Ultrasound Images of Thyroid Nodules Based on Improved DETR 基于改进型 DETR 的甲状腺结节超声图像风险分层研究
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-11-11 DOI: 10.1002/ima.23219
Zhang Le, Yue Liang, Xiaokang Hu, Taorong Qiu, Pan Xu

The Chinese Thyroid Imaging Reporting and Data System (C-TIRADS) standard is based on the Chinese current medical context. However, at present, there is a lack of C-TIRADS-based automatic computer-aided diagnosis system for thyroid nodule ultrasound images, and the existing algorithms for detecting and recognizing thyroid nodules are basically for the dichotomous classification of benign and malignant. We used the DETR (detection transformer) model as a baseline model and carried out model enhancements to address the shortcomings of unsatisfactory classification accuracy and difficulty in detecting small thyroid nodules in the DETR model. First, to investigate the method of acquiring multi-scale features of thyroid nodule ultrasound images, we choose TResNet-L as the feature extraction network and introduce multi-scale feature information and group convolution, thereby enhancing the model's multi-label classification accuracy. Second, a parallel decoder architecture for multi-label thyroid nodule ultrasound image classification is designed to enhance the learning of correlation between pathological feature class labels, aiming to improve the multi-label classification accuracy of the detection model. Third, the loss function of the detection model is improved. We propose a linear combination of Smooth L1-Loss and CIoU Loss as the model's bounding box loss function and asymmetric loss as the model's multi-label classification loss function, aiming to further improve the detection model's detection accuracy for small thyroid nodules. The experiment results show that the improved DETR model achieves an AP of 92.4% and 81.6% with IoU thresholds of 0.5 and 0.75, respectively.

中国甲状腺影像报告和数据系统(C-TIRADS)标准是基于中国当前的医疗背景制定的。然而,目前还缺乏基于 C-TIRADS 的甲状腺结节超声图像计算机辅助自动诊断系统,现有的甲状腺结节检测和识别算法基本上是良性和恶性的二分法。我们以 DETR(检测转换器)模型为基线模型,针对 DETR 模型中分类精度不理想和难以检测甲状腺小结节的缺点进行了模型增强。首先,为了研究甲状腺结节超声图像多尺度特征的获取方法,我们选择了 TResNet-L 作为特征提取网络,并引入了多尺度特征信息和群卷积,从而提高了模型的多标签分类精度。其次,设计了用于甲状腺结节超声图像多标签分类的并行解码器架构,加强病理特征类标签之间的相关性学习,旨在提高检测模型的多标签分类精度。第三,改进检测模型的损失函数。我们提出了平滑 L1 损失和 CIoU 损失的线性组合作为模型的边界框损失函数,非对称损失作为模型的多标签分类损失函数,旨在进一步提高检测模型对甲状腺小结节的检测精度。实验结果表明,改进后的DETR模型在IoU阈值为0.5和0.75时,AP分别达到92.4%和81.6%。
{"title":"A Risk Stratification Study of Ultrasound Images of Thyroid Nodules Based on Improved DETR","authors":"Zhang Le,&nbsp;Yue Liang,&nbsp;Xiaokang Hu,&nbsp;Taorong Qiu,&nbsp;Pan Xu","doi":"10.1002/ima.23219","DOIUrl":"https://doi.org/10.1002/ima.23219","url":null,"abstract":"<div>\u0000 \u0000 <p>The Chinese Thyroid Imaging Reporting and Data System (C-TIRADS) standard is based on the Chinese current medical context. However, at present, there is a lack of C-TIRADS-based automatic computer-aided diagnosis system for thyroid nodule ultrasound images, and the existing algorithms for detecting and recognizing thyroid nodules are basically for the dichotomous classification of benign and malignant. We used the DETR (detection transformer) model as a baseline model and carried out model enhancements to address the shortcomings of unsatisfactory classification accuracy and difficulty in detecting small thyroid nodules in the DETR model. First, to investigate the method of acquiring multi-scale features of thyroid nodule ultrasound images, we choose TResNet-L as the feature extraction network and introduce multi-scale feature information and group convolution, thereby enhancing the model's multi-label classification accuracy. Second, a parallel decoder architecture for multi-label thyroid nodule ultrasound image classification is designed to enhance the learning of correlation between pathological feature class labels, aiming to improve the multi-label classification accuracy of the detection model. Third, the loss function of the detection model is improved. We propose a linear combination of Smooth L1-Loss and CIoU Loss as the model's bounding box loss function and asymmetric loss as the model's multi-label classification loss function, aiming to further improve the detection model's detection accuracy for small thyroid nodules. The experiment results show that the improved DETR model achieves an AP of 92.4% and 81.6% with IoU thresholds of 0.5 and 0.75, respectively.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 6","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142641944","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Learning and Handcrafted Features for Thyroid Nodule Classification 深度学习和人工特征用于甲状腺结节分类
IF 3 4区 计算机科学 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2024-11-08 DOI: 10.1002/ima.23215
Ayoub Abderrazak Maarouf, Hacini meriem, Fella Hachouf

In this research, we present a refined image-based computer-aided diagnosis (CAD) system for thyroid cancer detection using ultrasound imagery. This system integrates a specialized convolutional neural network (CNN) architecture designed to address the unique aspects of thyroid image datasets. Additionally, it incorporates a novel statistical model that utilizes a two-dimensional random coefficient autoregressive (2D-RCA) method to precisely analyze the textural characteristics of thyroid images, thereby capturing essential texture-related information. The classification framework relies on a composite feature vector that combines deep learning features from the CNN and handcrafted features from the 2D-RCA model, processed through a support vector machine (SVM) algorithm. Our evaluation methodology is structured in three phases: initial assessment of the 2D-RCA features, analysis of the CNN-derived features, and a final evaluation of their combined effect on classification performance. Comparative analyses with well-known networks such as VGG16, VGG19, ResNet50, and AlexNet highlight the superior performance of our approach. The outcomes indicate a significant enhancement in diagnostic accuracy, achieving a classification accuracy of 97.2%, a sensitivity of 84.42%, and a specificity of 95.23%. These results not only demonstrate a notable advancement in the classification of thyroid nodules but also establish a new standard in the efficiency of CAD systems.

在这项研究中,我们提出了一种基于图像的计算机辅助诊断(CAD)系统,利用超声图像检测甲状腺癌。该系统集成了专门的卷积神经网络(CNN)架构,旨在解决甲状腺图像数据集的独特问题。此外,它还采用了一种新型统计模型,利用二维随机系数自回归(2D-RCA)方法精确分析甲状腺图像的纹理特征,从而捕捉到与纹理相关的重要信息。分类框架依赖于一个复合特征向量,该向量结合了 CNN 的深度学习特征和 2D-RCA 模型的手工特征,并通过支持向量机 (SVM) 算法进行处理。我们的评估方法分为三个阶段:初步评估 2D-RCA 特征、分析 CNN 衍生特征,以及最终评估它们对分类性能的综合影响。与 VGG16、VGG19、ResNet50 和 AlexNet 等知名网络的对比分析凸显了我们方法的卓越性能。结果表明,我们的方法显著提高了诊断准确性,分类准确率达到 97.2%,灵敏度达到 84.42%,特异性达到 95.23%。这些结果不仅证明了我们在甲状腺结节分类方面的显著进步,还为 CAD 系统的效率建立了新的标准。
{"title":"Deep Learning and Handcrafted Features for Thyroid Nodule Classification","authors":"Ayoub Abderrazak Maarouf,&nbsp;Hacini meriem,&nbsp;Fella Hachouf","doi":"10.1002/ima.23215","DOIUrl":"https://doi.org/10.1002/ima.23215","url":null,"abstract":"<div>\u0000 \u0000 <p>In this research, we present a refined image-based computer-aided diagnosis (CAD) system for thyroid cancer detection using ultrasound imagery. This system integrates a specialized convolutional neural network (CNN) architecture designed to address the unique aspects of thyroid image datasets. Additionally, it incorporates a novel statistical model that utilizes a two-dimensional random coefficient autoregressive (2D-RCA) method to precisely analyze the textural characteristics of thyroid images, thereby capturing essential texture-related information. The classification framework relies on a composite feature vector that combines deep learning features from the CNN and handcrafted features from the 2D-RCA model, processed through a support vector machine (SVM) algorithm. Our evaluation methodology is structured in three phases: initial assessment of the 2D-RCA features, analysis of the CNN-derived features, and a final evaluation of their combined effect on classification performance. Comparative analyses with well-known networks such as VGG16, VGG19, ResNet50, and AlexNet highlight the superior performance of our approach. The outcomes indicate a significant enhancement in diagnostic accuracy, achieving a classification accuracy of 97.2%, a sensitivity of 84.42%, and a specificity of 95.23%. These results not only demonstrate a notable advancement in the classification of thyroid nodules but also establish a new standard in the efficiency of CAD systems.</p>\u0000 </div>","PeriodicalId":14027,"journal":{"name":"International Journal of Imaging Systems and Technology","volume":"34 6","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142641632","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International Journal of Imaging Systems and Technology
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1