首页 > 最新文献

Journal of Digital Imaging最新文献

英文 中文
Convolutional Neural Networks for Segmentation of Pleural Mesothelioma: Analysis of Probability Map Thresholds (CALGB 30901, Alliance) 用于胸膜间皮瘤分割的卷积神经网络:概率图阈值分析(CALGB 30901,联盟)
IF 4.4 2区 工程技术 Q2 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2024-09-12 DOI: 10.1007/s10278-024-01092-z
Mena Shenouda, Eyjólfur Gudmundsson, Feng Li, Christopher M. Straus, Hedy L. Kindler, Arkadiusz Z. Dudek, Thomas Stinchcombe, Xiaofei Wang, Adam Starkey, Samuel G. Armato III

The purpose of this study was to evaluate the impact of probability map threshold on pleural mesothelioma (PM) tumor delineations generated using a convolutional neural network (CNN). One hundred eighty-six CT scans from 48 PM patients were segmented by a VGG16/U-Net CNN. A radiologist modified the contours generated at a 0.5 probability threshold. Percent difference of tumor volume and overlap using the Dice Similarity Coefficient (DSC) were compared between the reference standard provided by the radiologist and CNN outputs for thresholds ranging from 0.001 to 0.9. CNN-derived contours consistently yielded smaller tumor volumes than radiologist contours. Reducing the probability threshold from 0.5 to 0.01 decreased the absolute percent volume difference, on average, from 42.93% to 26.60%. Median and mean DSC ranged from 0.57 to 0.59, with a peak at a threshold of 0.2; no distinct threshold was found for percent volume difference. The CNN exhibited deficiencies with specific disease presentations, such as severe pleural effusion or disease in the pleural fissure. No single output threshold in the CNN probability maps was optimal for both tumor volume and DSC. This study emphasized the importance of considering both figures of merit when evaluating deep learning-based tumor segmentations across probability thresholds. This work underscores the need to simultaneously assess tumor volume and spatial overlap when evaluating CNN performance. While automated segmentations may yield comparable tumor volumes to that of the reference standard, the spatial region delineated by the CNN at a specific threshold is equally important.

本研究旨在评估概率图阈值对使用卷积神经网络(CNN)生成的胸膜间皮瘤(PM)肿瘤划分的影响。使用 VGG16/U-Net CNN 对 48 名胸膜间皮瘤患者的 186 张 CT 扫描图像进行了分割。放射科医生以 0.5 的概率阈值修改生成的轮廓。在 0.001 至 0.9 的阈值范围内,比较了放射科医生提供的参考标准与 CNN 输出结果之间的肿瘤体积百分比差异和使用 Dice 相似性系数 (DSC) 的重叠度。CNN 导出的轮廓得出的肿瘤体积始终小于放射科医生的轮廓。概率阈值从 0.5 降至 0.01 后,肿瘤体积绝对百分比差异平均从 42.93% 降至 26.60%。DSC的中位数和平均值从0.57到0.59不等,在阈值为0.2时达到峰值;在体积百分比差异方面没有发现明显的阈值。CNN 在特定疾病(如严重胸腔积液或胸膜裂孔中的疾病)时表现出缺陷。对于肿瘤体积和 DSC 而言,CNN 概率图中没有一个输出阈值是最佳的。这项研究强调,在评估基于深度学习的跨概率阈值肿瘤分割时,必须同时考虑这两项数据的优劣。这项工作强调了在评估 CNN 性能时同时评估肿瘤体积和空间重叠的必要性。虽然自动分割可能会产生与参考标准相当的肿瘤体积,但 CNN 在特定阈值下划定的空间区域同样重要。
{"title":"Convolutional Neural Networks for Segmentation of Pleural Mesothelioma: Analysis of Probability Map Thresholds (CALGB 30901, Alliance)","authors":"Mena Shenouda, Eyjólfur Gudmundsson, Feng Li, Christopher M. Straus, Hedy L. Kindler, Arkadiusz Z. Dudek, Thomas Stinchcombe, Xiaofei Wang, Adam Starkey, Samuel G. Armato III","doi":"10.1007/s10278-024-01092-z","DOIUrl":"https://doi.org/10.1007/s10278-024-01092-z","url":null,"abstract":"<p>The purpose of this study was to evaluate the impact of probability map threshold on pleural mesothelioma (PM) tumor delineations generated using a convolutional neural network (CNN). One hundred eighty-six CT scans from 48 PM patients were segmented by a VGG16/U-Net CNN. A radiologist modified the contours generated at a 0.5 probability threshold. Percent difference of tumor volume and overlap using the Dice Similarity Coefficient (DSC) were compared between the reference standard provided by the radiologist and CNN outputs for thresholds ranging from 0.001 to 0.9. CNN-derived contours consistently yielded smaller tumor volumes than radiologist contours. Reducing the probability threshold from 0.5 to 0.01 decreased the absolute percent volume difference, on average, from 42.93% to 26.60%. Median and mean DSC ranged from 0.57 to 0.59, with a peak at a threshold of 0.2; no distinct threshold was found for percent volume difference. The CNN exhibited deficiencies with specific disease presentations, such as severe pleural effusion or disease in the pleural fissure. No single output threshold in the CNN probability maps was optimal for both tumor volume and DSC. This study emphasized the importance of considering both figures of merit when evaluating deep learning-based tumor segmentations across probability thresholds. This work underscores the need to simultaneously assess tumor volume and spatial overlap when evaluating CNN performance. While automated segmentations may yield comparable tumor volumes to that of the reference standard, the spatial region delineated by the CNN at a specific threshold is equally important.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"3 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142226943","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automated Three-Dimensional Imaging and Pfirrmann Classification of Intervertebral Disc Using a Graphical Neural Network in Sagittal Magnetic Resonance Imaging of the Lumbar Spine 在腰椎矢状位磁共振成像中使用图形神经网络自动进行椎间盘三维成像和普菲尔曼分类
IF 4.4 2区 工程技术 Q2 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2024-09-12 DOI: 10.1007/s10278-024-01251-2
David Baur, Richard Bieck, Johann Berger, Patrick Schöfer, Tim Stelzner, Juliane Neumann, Thomas Neumuth, Christoph-E. Heyde, Anna Voelker

This study aimed to develop a graph neural network (GNN) for automated three-dimensional (3D) magnetic resonance imaging (MRI) visualization and Pfirrmann grading of intervertebral discs (IVDs), and benchmark it against manual classifications. Lumbar IVD MRI data from 300 patients were retrospectively analyzed. Two clinicians assessed the manual segmentation and grading for inter-rater reliability using Cohen's kappa. The IVDs were then processed and classified using an automated convolutional neural network (CNN)–GNN pipeline, and their performance was evaluated using F1 scores. Manual Pfirrmann grading exhibited moderate agreement (κ = 0.455–0.565) among the clinicians, with higher exact match frequencies at lower lumbar levels. Single-grade discrepancies were prevalent except at L5/S1. Automated segmentation of IVDs using a pretrained U-Net model achieved an F1 score of 0.85, with a precision and recall of 0.83 and 0.88, respectively. Following 3D reconstruction of the automatically segmented IVD into a 3D point-cloud representation of the target intervertebral disc, the GNN model demonstrated moderate performance in Pfirrmann classification. The highest precision (0.81) and F1 score (0.71) were observed at L2/3, whereas the overall metrics indicated moderate performance (precision: 0.46, recall: 0.47, and F1 score: 0.46), with variability across spinal levels. The integration of CNN and GNN offers a new perspective for automating IVD analysis in MRI. Although the current performance highlights the need for further refinement, the moderate accuracy of the model, combined with its 3D visualization capabilities, establishes a promising foundation for more advanced grading systems.

本研究旨在开发一种图神经网络(GNN),用于自动三维(3D)磁共振成像(MRI)可视化和椎间盘(IVD)的普菲尔曼分级,并将其与人工分级进行比较。对 300 名患者的腰椎间盘磁共振成像数据进行了回顾性分析。两名临床医生使用科恩卡帕评估了人工分割和分级的评分者间可靠性。然后使用自动卷积神经网络(CNN)-GNN 管道对 IVD 进行处理和分级,并使用 F1 分数评估其性能。临床医生之间的手动 Pfirrmann 分级显示出中等程度的一致性(κ = 0.455-0.565),腰椎低位的精确匹配频率较高。除 L5/S1 外,普遍存在单级差异。使用预训练的 U-Net 模型自动分割 IVD 的 F1 得分为 0.85,精确度和召回率分别为 0.83 和 0.88。将自动分割的 IVD 三维重建为目标椎间盘的三维点云表示后,GNN 模型在 Pfirrmann 分类中表现出中等水平。最高精度(0.81)和 F1 得分(0.71)出现在 L2/3,而总体指标显示性能适中(精度:0.46,召回率:0.47,F1 得分:0.46),不同脊柱水平之间存在差异。CNN 和 GNN 的整合为磁共振成像中的 IVD 自动分析提供了一个新的视角。虽然目前的性能还需要进一步改进,但该模型的中等准确度及其三维可视化功能为更先进的分级系统奠定了良好的基础。
{"title":"Automated Three-Dimensional Imaging and Pfirrmann Classification of Intervertebral Disc Using a Graphical Neural Network in Sagittal Magnetic Resonance Imaging of the Lumbar Spine","authors":"David Baur, Richard Bieck, Johann Berger, Patrick Schöfer, Tim Stelzner, Juliane Neumann, Thomas Neumuth, Christoph-E. Heyde, Anna Voelker","doi":"10.1007/s10278-024-01251-2","DOIUrl":"https://doi.org/10.1007/s10278-024-01251-2","url":null,"abstract":"<p>This study aimed to develop a graph neural network (GNN) for automated three-dimensional (3D) magnetic resonance imaging (MRI) visualization and Pfirrmann grading of intervertebral discs (IVDs), and benchmark it against manual classifications. Lumbar IVD MRI data from 300 patients were retrospectively analyzed. Two clinicians assessed the manual segmentation and grading for inter-rater reliability using Cohen's kappa. The IVDs were then processed and classified using an automated convolutional neural network (CNN)–GNN pipeline, and their performance was evaluated using F1 scores. Manual Pfirrmann grading exhibited moderate agreement (κ = 0.455–0.565) among the clinicians, with higher exact match frequencies at lower lumbar levels. Single-grade discrepancies were prevalent except at L5/S1. Automated segmentation of IVDs using a pretrained U-Net model achieved an F1 score of 0.85, with a precision and recall of 0.83 and 0.88, respectively. Following 3D reconstruction of the automatically segmented IVD into a 3D point-cloud representation of the target intervertebral disc, the GNN model demonstrated moderate performance in Pfirrmann classification. The highest precision (0.81) and F1 score (0.71) were observed at L2/3, whereas the overall metrics indicated moderate performance (precision: 0.46, recall: 0.47, and F1 score: 0.46), with variability across spinal levels. The integration of CNN and GNN offers a new perspective for automating IVD analysis in MRI. Although the current performance highlights the need for further refinement, the moderate accuracy of the model, combined with its 3D visualization capabilities, establishes a promising foundation for more advanced grading systems.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"7 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142216218","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Screening Patient Misidentification Errors Using a Deep Learning Model of Chest Radiography: A Seven Reader Study 使用胸部 X 射线照相深度学习模型筛查患者误识别错误:七名读者研究
IF 4.4 2区 工程技术 Q2 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2024-09-11 DOI: 10.1007/s10278-024-01245-0
Kiduk Kim, Kyungjin Cho, Yujeong Eo, Jeeyoung Kim, Jihye Yun, Yura Ahn, Joon Beom Seo, Gil-Sun Hong, Namkug Kim

We aimed to evaluate the ability of deep learning (DL) models to identify patients from a paired chest radiograph (CXR) and compare their performance with that of human experts. In this retrospective study, patient identification DL models were developed using 240,004 CXRs. The models were validated using multiple datasets, namely, internal validation, CheXpert, and Chest ImaGenome (CIG), which include different populations. Model performance was analyzed in terms of disease change status. The performance of the models to identify patients from paired CXRs was compared with three junior radiology residents (group I), two senior radiology residents (group II), and two board-certified expert radiologists (group III). For the reader study, 240 patients (age, 56.617 ± 13.690 years, 113 females, 160 same pairs) were evaluated. A one-sided non-inferiority test was performed with a one-sided margin of 0.05. SimChest, our similarity-based DL model, demonstrated the best patient identification performance across multiple datasets, regardless of disease change status (internal validation [area under the receiver operating characteristic curve range: 0.992–0.999], CheXpert [0.933–0.948], and CIG [0.949–0.951]). The radiologists identified patients from the paired CXRs with a mean accuracy of 0.900 (95% confidence interval: 0.852–0.948), with performance increasing with experience (mean accuracy:group I [0.874], group II [0.904], group III [0.935], and SimChest [0.904]). SimChest achieved non-inferior performance compared to the radiologists (P for non-inferiority: 0.015). The findings of this diagnostic study indicate that DL models can screen for patient misidentification using a pair of CXRs non-inferiorly to human experts.

我们旨在评估深度学习(DL)模型从成对胸片(CXR)中识别患者的能力,并将其与人类专家的表现进行比较。在这项回顾性研究中,使用 240,004 张 CXR 开发了患者识别 DL 模型。模型通过多个数据集(即内部验证、CheXpert 和 Chest ImaGenome (CIG))进行了验证,这些数据集包括不同的人群。根据疾病变化状况对模型性能进行了分析。与三位放射科初级住院医师(第一组)、两位放射科高级住院医师(第二组)和两位经委员会认证的放射科专家(第三组)比较了模型从配对的 CXR 图像中识别患者的性能。读者研究共评估了 240 名患者(年龄为 56.617 ± 13.690 岁,113 名女性,160 对相同患者)。进行了单侧非劣效性检验,单侧差值为 0.05。无论疾病变化状况如何,我们基于相似性的 DL 模型 SimChest 在多个数据集中都表现出最佳的患者识别性能(内部验证[接收器操作特征曲线下面积范围:0.992-0.999]、CheXpert [0.933-0.948]和 CIG [0.949-0.951])。放射科医生从配对 CXR 图像中识别患者的平均准确率为 0.900(95% 置信区间:0.852-0.948),准确率随经验的增加而提高(平均准确率:第一组[0.874],第二组[0.904],第三组[0.935],SimChest [0.904])。与放射科医生相比,SimChest 的表现并不逊色(不逊色的 P 值:0.015)。这项诊断研究的结果表明,DL 模型可以使用一对 CXR 对患者进行错误识别,其效果不亚于人类专家。
{"title":"Screening Patient Misidentification Errors Using a Deep Learning Model of Chest Radiography: A Seven Reader Study","authors":"Kiduk Kim, Kyungjin Cho, Yujeong Eo, Jeeyoung Kim, Jihye Yun, Yura Ahn, Joon Beom Seo, Gil-Sun Hong, Namkug Kim","doi":"10.1007/s10278-024-01245-0","DOIUrl":"https://doi.org/10.1007/s10278-024-01245-0","url":null,"abstract":"<p>We aimed to evaluate the ability of deep learning (DL) models to identify patients from a paired chest radiograph (CXR) and compare their performance with that of human experts. In this retrospective study, patient identification DL models were developed using 240,004 CXRs. The models were validated using multiple datasets, namely, internal validation, CheXpert, and Chest ImaGenome (CIG), which include different populations. Model performance was analyzed in terms of disease change status. The performance of the models to identify patients from paired CXRs was compared with three junior radiology residents (group I), two senior radiology residents (group II), and two board-certified expert radiologists (group III). For the reader study, 240 patients (age, 56.617 ± 13.690 years, 113 females, 160 same pairs) were evaluated. A one-sided non-inferiority test was performed with a one-sided margin of 0.05. SimChest, our similarity-based DL model, demonstrated the best patient identification performance across multiple datasets, regardless of disease change status (internal validation [area under the receiver operating characteristic curve range: 0.992–0.999], CheXpert [0.933–0.948], and CIG [0.949–0.951]). The radiologists identified patients from the paired CXRs with a mean accuracy of 0.900 (95% confidence interval: 0.852–0.948), with performance increasing with experience (mean accuracy:group I [0.874], group II [0.904], group III [0.935], and SimChest [0.904]). SimChest achieved non-inferior performance compared to the radiologists (<i>P</i> for non-inferiority: 0.015). The findings of this diagnostic study indicate that DL models can screen for patient misidentification using a pair of CXRs non-inferiorly to human experts.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"6 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142216219","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Novel Network for Low-Dose CT Denoising Based on Dual-Branch Structure and Multi-Scale Residual Attention 基于双分支结构和多尺度残留注意力的低剂量 CT 去噪新型网络
IF 4.4 2区 工程技术 Q2 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2024-09-11 DOI: 10.1007/s10278-024-01254-z
Ju Zhang, Lieli Ye, Weiwei Gong, Mingyang Chen, Guangyu Liu, Yun Cheng

Deep learning-based denoising of low-dose medical CT images has received great attention both from academic researchers and physicians in recent years, and has shown important application value in clinical practice. In this work, a novel two-branch and multi-scale residual attention-based network for low-dose CT image denoising is proposed. It adopts a two-branch framework structure, to extract and fuse image features at shallow and deep levels respectively, to recover image texture and structure information as much as possible. We propose the adaptive dynamic convolution block (ADCB) in the local information extraction layer. It can effectively extract the detailed information of low-dose CT denoising and enables the network to better capture the local details and texture features of the image, thereby improving the denoising effect and image quality. Multi-scale edge enhancement attention block (MEAB) is proposed in the global information extraction layer, to perform feature fusion through dilated convolution and a multi-dimensional attention mechanism. A multi-scale residual convolution block (MRCB) is proposed to integrate feature information and improve the robustness and generalization of the network. To demonstrate the effectiveness of our method, extensive comparison experiments are conducted and the performances evaluated on two publicly available datasets. Our model achieves 29.3004 PSNR, 0.8659 SSIM, and 14.0284 RMSE on the AAPM-Mayo dataset. It is evaluated by adding four different noise levels σ = 15, 30, 45, and 60 on the Qin_LUNG_CT dataset and achieves the best results. Ablation studies show that the proposed ADCB, MEAB, and MRCB modules improve the denoising performances significantly. The source code is available at https://github.com/Ye111-cmd/LDMANet.

近年来,基于深度学习的低剂量医学 CT 图像去噪受到了学术研究人员和医生的极大关注,并在临床实践中显示出了重要的应用价值。本研究提出了一种新颖的基于双分支和多尺度残差注意力的低剂量 CT 图像去噪网络。它采用双分支框架结构,分别从浅层和深层提取和融合图像特征,尽可能地恢复图像纹理和结构信息。我们在局部信息提取层中提出了自适应动态卷积块(ADCB)。它能有效提取低剂量 CT 去噪的细节信息,使网络更好地捕捉图像的局部细节和纹理特征,从而提高去噪效果和图像质量。在全局信息提取层提出了多尺度边缘增强注意块(MEAB),通过扩张卷积和多维注意机制进行特征融合。我们还提出了多尺度残差卷积块(MRCB)来整合特征信息,提高网络的鲁棒性和泛化能力。为了证明我们方法的有效性,我们在两个公开数据集上进行了广泛的对比实验和性能评估。我们的模型在 AAPM-Mayo 数据集上实现了 29.3004 PSNR、0.8659 SSIM 和 14.0284 RMSE。在 Qin_LUNG_CT 数据集上,通过添加四种不同的噪声水平 σ = 15、30、45 和 60 进行评估,结果最佳。消融研究表明,所提出的 ADCB、MEAB 和 MRCB 模块显著提高了去噪性能。源代码见 https://github.com/Ye111-cmd/LDMANet。
{"title":"A Novel Network for Low-Dose CT Denoising Based on Dual-Branch Structure and Multi-Scale Residual Attention","authors":"Ju Zhang, Lieli Ye, Weiwei Gong, Mingyang Chen, Guangyu Liu, Yun Cheng","doi":"10.1007/s10278-024-01254-z","DOIUrl":"https://doi.org/10.1007/s10278-024-01254-z","url":null,"abstract":"<p>Deep learning-based denoising of low-dose medical CT images has received great attention both from academic researchers and physicians in recent years, and has shown important application value in clinical practice. In this work, a novel two-branch and multi-scale residual attention-based network for low-dose CT image denoising is proposed. It adopts a two-branch framework structure, to extract and fuse image features at shallow and deep levels respectively, to recover image texture and structure information as much as possible. We propose the adaptive dynamic convolution block (ADCB) in the local information extraction layer. It can effectively extract the detailed information of low-dose CT denoising and enables the network to better capture the local details and texture features of the image, thereby improving the denoising effect and image quality. Multi-scale edge enhancement attention block (MEAB) is proposed in the global information extraction layer, to perform feature fusion through dilated convolution and a multi-dimensional attention mechanism. A multi-scale residual convolution block (MRCB) is proposed to integrate feature information and improve the robustness and generalization of the network. To demonstrate the effectiveness of our method, extensive comparison experiments are conducted and the performances evaluated on two publicly available datasets. Our model achieves 29.3004 PSNR, 0.8659 SSIM, and 14.0284 RMSE on the AAPM-Mayo dataset. It is evaluated by adding four different noise levels σ = 15, 30, 45, and 60 on the Qin_LUNG_CT dataset and achieves the best results. Ablation studies show that the proposed ADCB, MEAB, and MRCB modules improve the denoising performances significantly. The source code is available at https://github.com/Ye111-cmd/LDMANet.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"2 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142216220","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sex-Specific Imaging Biomarkers for Parkinson’s Disease Diagnosis: A Machine Learning Analysis 用于帕金森病诊断的性别特异性成像生物标记物:机器学习分析
IF 4.4 2区 工程技术 Q2 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2024-09-10 DOI: 10.1007/s10278-024-01235-2
Yifeng Yang, Liangyun Hu, Yang Chen, Weidong Gu, Yuanzhong Xie, Shengdong Nie

This study aimed to identify sex-specific imaging biomarkers for Parkinson’s disease (PD) based on multiple MRI morphological features by using machine learning methods. Participants were categorized into female and male subgroups, and various structural morphological features were extracted. An ensemble Lasso (EnLasso) method was employed to identify a stable optimal feature subset for each sex-based subgroup. Eight typical classifiers were adopted to construct classification models for PD and HC, respectively, to validate whether models specific to sex subgroups could bolster the precision of PD identification. Finally, statistical analysis and correlation tests were carried out on significant brain region features to identify potential sex-specific imaging biomarkers. The best model (MLP) based on the female subgroup and male subgroup achieved average classification accuracy of 92.83% and 92.11%, respectively, which were better than that of the model based on the overall samples (86.88%) and the overall model incorporating gender factor (87.52%). In addition, the most discriminative feature of PD among males was the lh 6r (FD), but among females, it was the lh PreS (GI). The findings indicate that the sex-specific PD diagnosis model yields a significantly higher classification performance compared to previous models that included all participants. Additionally, the male subgroup exhibited a greater number of brain region changes than the female subgroup, suggesting sex-specific differences in PD risk markers. This study underscore the importance of stratifying data by sex and offer insights into sex-specific variations in PD phenotypes, which could aid in the development of precise and personalized diagnostic approaches in the early stages of the disease.

本研究旨在利用机器学习方法,根据多种核磁共振成像形态学特征识别帕金森病(PD)的性别特异性成像生物标志物。研究人员将参与者分为女性和男性亚组,并提取了各种结构形态学特征。采用集合拉索(EnLasso)方法为每个基于性别的亚组确定一个稳定的最佳特征子集。采用八种典型的分类器分别构建了 PD 和 HC 的分类模型,以验证性别亚组的特定模型是否能提高 PD 识别的精确度。最后,对重要的脑区特征进行了统计分析和相关性测试,以确定潜在的性别特异性成像生物标志物。基于女性亚组和男性亚组的最佳模型(MLP)的平均分类准确率分别为92.83%和92.11%,优于基于总体样本的模型(86.88%)和包含性别因素的总体模型(87.52%)。此外,男性 PD 中最具鉴别力的特征是 lh 6r (FD),而女性则是 lh PreS (GI)。研究结果表明,与之前包含所有参与者的模型相比,针对不同性别的帕金森病诊断模型的分类性能明显更高。此外,男性亚组比女性亚组表现出更多的脑区变化,这表明帕金森病风险标志物存在性别差异。这项研究强调了按性别对数据进行分层的重要性,并提供了对帕金森病表型的性别特异性差异的见解,这有助于在疾病的早期阶段开发精确的个性化诊断方法。
{"title":"Sex-Specific Imaging Biomarkers for Parkinson’s Disease Diagnosis: A Machine Learning Analysis","authors":"Yifeng Yang, Liangyun Hu, Yang Chen, Weidong Gu, Yuanzhong Xie, Shengdong Nie","doi":"10.1007/s10278-024-01235-2","DOIUrl":"https://doi.org/10.1007/s10278-024-01235-2","url":null,"abstract":"<p>This study aimed to identify sex-specific imaging biomarkers for Parkinson’s disease (PD) based on multiple MRI morphological features by using machine learning methods. Participants were categorized into female and male subgroups, and various structural morphological features were extracted. An ensemble Lasso (EnLasso) method was employed to identify a stable optimal feature subset for each sex-based subgroup. Eight typical classifiers were adopted to construct classification models for PD and HC, respectively, to validate whether models specific to sex subgroups could bolster the precision of PD identification. Finally, statistical analysis and correlation tests were carried out on significant brain region features to identify potential sex-specific imaging biomarkers. The best model (MLP) based on the female subgroup and male subgroup achieved average classification accuracy of 92.83% and 92.11%, respectively, which were better than that of the model based on the overall samples (86.88%) and the overall model incorporating gender factor (87.52%). In addition, the most discriminative feature of PD among males was the lh 6r (FD), but among females, it was the lh PreS (GI). The findings indicate that the sex-specific PD diagnosis model yields a significantly higher classification performance compared to previous models that included all participants. Additionally, the male subgroup exhibited a greater number of brain region changes than the female subgroup, suggesting sex-specific differences in PD risk markers. This study underscore the importance of stratifying data by sex and offer insights into sex-specific variations in PD phenotypes, which could aid in the development of precise and personalized diagnostic approaches in the early stages of the disease.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"283 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142216221","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing Nasopharyngeal Carcinoma Survival Prediction: Integrating Pre- and Post-Treatment MRI Radiomics with Clinical Data 加强鼻咽癌生存预测:将治疗前后的磁共振成像放射组学与临床数据相结合
IF 4.4 2区 工程技术 Q2 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2024-04-30 DOI: 10.1007/s10278-024-01109-7
Luong Huu Dang, Shih-Han Hung, Nhi Thao Ngoc Le, Wei-Kai Chuang, Jeng-You Wu, Ting-Chieh Huang, Nguyen Quoc Khanh Le

Recurrences are frequent in nasopharyngeal carcinoma (NPC) despite high remission rates with treatment, leading to considerable morbidity. This study aimed to develop a prediction model for NPC survival by harnessing both pre- and post-treatment magnetic resonance imaging (MRI) radiomics in conjunction with clinical data, focusing on 3-year progression-free survival (PFS) as the primary outcome. Our comprehensive approach involved retrospective clinical and MRI data collection of 276 eligible NPC patients from three independent hospitals (180 in the training cohort, 46 in the validation cohort, and 50 in the external cohort) who underwent MRI scans twice, once within 2 months prior to treatment and once within 10 months after treatment. From the contrast-enhanced T1-weighted images before and after treatment, 3404 radiomics features were extracted. These features were not only derived from the primary lesion but also from the adjacent lymph nodes surrounding the tumor. We conducted appropriate feature selection pipelines, followed by Cox proportional hazards models for survival analysis. Model evaluation was performed using receiver operating characteristic (ROC) analysis, the Kaplan–Meier method, and nomogram construction. Our study unveiled several crucial predictors of NPC survival, notably highlighting the synergistic combination of pre- and post-treatment data in both clinical and radiomics assessments. Our prediction model demonstrated robust performance, with an accuracy of AUCs of 0.66 (95% CI: 0.536–0.779) in the training cohort, 0.717 (95% CI: 0.536–0.883) in the testing cohort, and 0.827 (95% CI: 0.684–0.948) in validation cohort in prognosticating patient outcomes. Our study presented a novel and effective prediction model for NPC survival, leveraging both pre- and post-treatment clinical data in conjunction with MRI features. Its constructed nomogram provides potentially significant implications for NPC research, offering clinicians a valuable tool for individualized treatment planning and patient counseling.

尽管鼻咽癌(NPC)的治疗缓解率很高,但复发却很频繁,导致相当高的发病率。本研究旨在结合临床数据,利用治疗前后的磁共振成像(MRI)放射组学建立鼻咽癌生存预测模型,并将 3 年无进展生存期(PFS)作为主要结果。我们的综合方法包括对三家独立医院的 276 名符合条件的鼻咽癌患者进行回顾性临床和 MRI 数据收集(其中 180 人属于训练队列,46 人属于验证队列,50 人属于外部队列),这些患者接受了两次 MRI 扫描,一次在治疗前 2 个月内,一次在治疗后 10 个月内。从治疗前后的对比增强 T1 加权图像中提取了 3404 个放射组学特征。这些特征不仅来自原发病灶,还来自肿瘤周围的邻近淋巴结。我们进行了适当的特征选择流水线,然后使用 Cox 比例危险模型进行生存分析。模型评估采用接收器操作特征(ROC)分析、Kaplan-Meier 法和提名图构建法。我们的研究揭示了鼻咽癌存活率的几个关键预测因素,特别强调了临床和放射组学评估中治疗前后数据的协同组合。我们的预测模型表现出强劲的性能,在预测患者预后方面,训练队列的AUC为0.66(95% CI:0.536-0.779),测试队列的AUC为0.717(95% CI:0.536-0.883),验证队列的AUC为0.827(95% CI:0.684-0.948)。我们的研究利用治疗前后的临床数据和磁共振成像特征,提出了一种新颖有效的鼻咽癌生存预测模型。其构建的提名图为鼻咽癌研究提供了潜在的重要意义,为临床医生提供了个体化治疗计划和患者咨询的宝贵工具。
{"title":"Enhancing Nasopharyngeal Carcinoma Survival Prediction: Integrating Pre- and Post-Treatment MRI Radiomics with Clinical Data","authors":"Luong Huu Dang, Shih-Han Hung, Nhi Thao Ngoc Le, Wei-Kai Chuang, Jeng-You Wu, Ting-Chieh Huang, Nguyen Quoc Khanh Le","doi":"10.1007/s10278-024-01109-7","DOIUrl":"https://doi.org/10.1007/s10278-024-01109-7","url":null,"abstract":"<p>Recurrences are frequent in nasopharyngeal carcinoma (NPC) despite high remission rates with treatment, leading to considerable morbidity. This study aimed to develop a prediction model for NPC survival by harnessing both pre- and post-treatment magnetic resonance imaging (MRI) radiomics in conjunction with clinical data, focusing on 3-year progression-free survival (PFS) as the primary outcome. Our comprehensive approach involved retrospective clinical and MRI data collection of 276 eligible NPC patients from three independent hospitals (180 in the training cohort, 46 in the validation cohort, and 50 in the external cohort) who underwent MRI scans twice, once within 2 months prior to treatment and once within 10 months after treatment. From the contrast-enhanced T1-weighted images before and after treatment, 3404 radiomics features were extracted. These features were not only derived from the primary lesion but also from the adjacent lymph nodes surrounding the tumor. We conducted appropriate feature selection pipelines, followed by Cox proportional hazards models for survival analysis. Model evaluation was performed using receiver operating characteristic (ROC) analysis, the Kaplan–Meier method, and nomogram construction. Our study unveiled several crucial predictors of NPC survival, notably highlighting the synergistic combination of pre- and post-treatment data in both clinical and radiomics assessments. Our prediction model demonstrated robust performance, with an accuracy of AUCs of 0.66 (95% CI: 0.536–0.779) in the training cohort, 0.717 (95% CI: 0.536–0.883) in the testing cohort, and 0.827 (95% CI: 0.684–0.948) in validation cohort in prognosticating patient outcomes. Our study presented a novel and effective prediction model for NPC survival, leveraging both pre- and post-treatment clinical data in conjunction with MRI features. Its constructed nomogram provides potentially significant implications for NPC research, offering clinicians a valuable tool for individualized treatment planning and patient counseling.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"26 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-04-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140840340","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Development of a Secure Web-Based Medical Imaging Analysis Platform: The AWESOMME Project 开发基于网络的安全医学影像分析平台:AWESOMME 项目
IF 4.4 2区 工程技术 Q2 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2024-04-30 DOI: 10.1007/s10278-024-01110-0
Tiphaine Diot-Dejonghe, Benjamin Leporq, Amine Bouhamama, Helene Ratiney, Frank Pilleul, Olivier Beuf, Frederic Cervenansky

Precision medicine research benefits from machine learning in the creation of robust models adapted to the processing of patient data. This applies both to pathology identification in images, i.e., annotation or segmentation, and to computer-aided diagnostic for classification or prediction. It comes with the strong need to exploit and visualize large volumes of images and associated medical data. The work carried out in this paper follows on from a main case study piloted in a cancer center. It proposes an analysis pipeline for patients with osteosarcoma through segmentation, feature extraction and application of a deep learning model to predict response to treatment. The main aim of the AWESOMME project is to leverage this work and implement the pipeline on an easy-to-access, secure web platform. The proposed WEB application is based on a three-component architecture: a data server, a heavy computation and authentication server and a medical imaging web-framework with a user interface. These existing components have been enhanced to meet the needs of security and traceability for the continuous production of expert data. It innovates by covering all steps of medical imaging processing (visualization and segmentation, feature extraction and aided diagnostic) and enables the test and use of machine learning models. The infrastructure is operational, deployed in internal production and is currently being installed in the hospital environment. The extension of the case study and user feedback enabled us to fine-tune functionalities and proved that AWESOMME is a modular solution capable to analyze medical data and share research algorithms with in-house clinicians.

精准医疗研究得益于机器学习,它能创建适合患者数据处理的强大模型。这既适用于图像中的病理识别,即注释或分割,也适用于分类或预测的计算机辅助诊断。这就强烈要求对大量图像和相关医疗数据进行开发和可视化。本文的研究工作是在癌症中心进行的一项主要案例研究的基础上开展的。它通过分割、特征提取和应用深度学习模型预测治疗反应,为骨肉瘤患者提出了一个分析管道。AWESOMME 项目的主要目的是利用这项工作,在一个易于访问的安全网络平台上实施该管道。拟议的 WEB 应用程序基于一个由三部分组成的架构:数据服务器、重计算和身份验证服务器以及带有用户界面的医学影像网络框架。对这些现有组件进行了改进,以满足持续生产专家数据的安全性和可追溯性需求。它的创新之处在于涵盖了医学影像处理的所有步骤(可视化和分割、特征提取和辅助诊断),并能测试和使用机器学习模型。该基础设施已投入运行,在内部生产中部署,目前正在医院环境中安装。案例研究的扩展和用户反馈使我们能够对功能进行微调,并证明 AWESOMME 是一个模块化解决方案,能够分析医疗数据并与内部临床医生共享研究算法。
{"title":"Development of a Secure Web-Based Medical Imaging Analysis Platform: The AWESOMME Project","authors":"Tiphaine Diot-Dejonghe, Benjamin Leporq, Amine Bouhamama, Helene Ratiney, Frank Pilleul, Olivier Beuf, Frederic Cervenansky","doi":"10.1007/s10278-024-01110-0","DOIUrl":"https://doi.org/10.1007/s10278-024-01110-0","url":null,"abstract":"<p>Precision medicine research benefits from machine learning in the creation of robust models adapted to the processing of patient data. This applies both to pathology identification in images, i.e., annotation or segmentation, and to computer-aided diagnostic for classification or prediction. It comes with the strong need to exploit and visualize large volumes of images and associated medical data. The work carried out in this paper follows on from a main case study piloted in a cancer center. It proposes an analysis pipeline for patients with osteosarcoma through segmentation, feature extraction and application of a deep learning model to predict response to treatment. The main aim of the AWESOMME project is to leverage this work and implement the pipeline on an easy-to-access, secure web platform. The proposed WEB application is based on a three-component architecture: a data server, a heavy computation and authentication server and a medical imaging web-framework with a user interface. These existing components have been enhanced to meet the needs of security and traceability for the continuous production of expert data. It innovates by covering all steps of medical imaging processing (visualization and segmentation, feature extraction and aided diagnostic) and enables the test and use of machine learning models. The infrastructure is operational, deployed in internal production and is currently being installed in the hospital environment. The extension of the case study and user feedback enabled us to fine-tune functionalities and proved that AWESOMME is a modular solution capable to analyze medical data and share research algorithms with in-house clinicians.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"30 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-04-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140840107","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automatic Skeleton Segmentation in CT Images Based on U-Net 基于 U-Net 的 CT 图像骨骼自动分割技术
IF 4.4 2区 工程技术 Q2 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2024-04-30 DOI: 10.1007/s10278-024-01127-5
Eva Milara, Adolfo Gómez-Grande, Pilar Sarandeses, Alexander P. Seiffert, Enrique J. Gómez, Patricia Sánchez-González

Bone metastasis, emerging oncological therapies, and osteoporosis represent some of the distinct clinical contexts which can result in morphological alterations in bone structure. The visual assessment of these changes through anatomical images is considered suboptimal, emphasizing the importance of precise skeletal segmentation as a valuable aid for its evaluation. In the present study, a neural network model for automatic skeleton segmentation from bidimensional computerized tomography (CT) slices is proposed. A total of 77 CT images and their semimanual skeleton segmentation from two acquisition protocols (whole-body and femur-to-head) are used to form a training group and a testing group. Preprocessing of the images includes four main steps: stretcher removal, thresholding, image clipping, and normalization (with two different techniques: interpatient and intrapatient). Subsequently, five different sets are created and arranged in a randomized order for the training phase. A neural network model based on U-Net architecture is implemented with different values of the number of channels in each feature map and number of epochs. The model with the best performance obtains a Jaccard index (IoU) of 0.959 and a Dice index of 0.979. The resultant model demonstrates the potential of deep learning applied in medical images and proving its utility in bone segmentation.

骨转移、新出现的肿瘤疗法和骨质疏松症是可能导致骨结构形态改变的一些不同临床情况。通过解剖图像对这些变化进行视觉评估被认为是不理想的,这就强调了精确骨骼分割作为评估的重要辅助工具的重要性。本研究提出了一种用于从二维计算机断层扫描(CT)切片自动分割骨骼的神经网络模型。本研究采用两种采集方案(全身和股骨头)共 77 张 CT 图像及其半手动骨骼分割,组成一个训练组和一个测试组。图像预处理包括四个主要步骤:移除担架、阈值处理、图像剪切和归一化(采用两种不同的技术:患者间和患者内)。随后,创建五个不同的集合,并按随机顺序排列,用于训练阶段。基于 U-Net 架构的神经网络模型在每个特征图中采用不同的通道数和历时数。性能最佳的模型获得了 0.959 的 Jaccard 指数(IoU)和 0.979 的 Dice 指数。结果模型展示了深度学习应用于医学图像的潜力,并证明了其在骨骼分割中的实用性。
{"title":"Automatic Skeleton Segmentation in CT Images Based on U-Net","authors":"Eva Milara, Adolfo Gómez-Grande, Pilar Sarandeses, Alexander P. Seiffert, Enrique J. Gómez, Patricia Sánchez-González","doi":"10.1007/s10278-024-01127-5","DOIUrl":"https://doi.org/10.1007/s10278-024-01127-5","url":null,"abstract":"<p>Bone metastasis, emerging oncological therapies, and osteoporosis represent some of the distinct clinical contexts which can result in morphological alterations in bone structure. The visual assessment of these changes through anatomical images is considered suboptimal, emphasizing the importance of precise skeletal segmentation as a valuable aid for its evaluation. In the present study, a neural network model for automatic skeleton segmentation from bidimensional computerized tomography (CT) slices is proposed. A total of 77 CT images and their semimanual skeleton segmentation from two acquisition protocols (whole-body and femur-to-head) are used to form a training group and a testing group. Preprocessing of the images includes four main steps: stretcher removal, thresholding, image clipping, and normalization (with two different techniques: interpatient and intrapatient). Subsequently, five different sets are created and arranged in a randomized order for the training phase. A neural network model based on U-Net architecture is implemented with different values of the number of channels in each feature map and number of epochs. The model with the best performance obtains a Jaccard index (IoU) of 0.959 and a Dice index of 0.979. The resultant model demonstrates the potential of deep learning applied in medical images and proving its utility in bone segmentation.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"13 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-04-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140840354","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Letter to the Editor Regarding Article “Prior to Initiation of Chemotherapy, Can We Predict Breast Tumor Response? Deep Learning Convolutional Neural Networks Approach Using a Breast MRI Tumor Dataset” 致编辑的信,内容涉及文章 "化疗开始前,我们能预测乳腺肿瘤的反应吗?使用乳腺 MRI 肿瘤数据集的深度学习卷积神经网络方法"
IF 4.4 2区 工程技术 Q2 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2024-04-30 DOI: 10.1007/s10278-024-01129-3
Joren Brunekreef

The cited article reports on a convolutional neural network trained to predict response to neoadjuvant chemotherapy from pre-treatment breast MRI scans. The proposed algorithm attains impressive performance on the test dataset with a mean Area Under the Receiver-Operating Characteristic curve of 0.98 and a mean accuracy of 88%. In this letter, I raise concerns that the reported results can be explained by inadvertent data leakage between training and test datasets. More precisely, I conjecture that the random split of the full dataset in training and test sets did not occur on a patient level, but rather on the level of 2D MRI slices. This allows the neural network to “memorize” a patient’s anatomy and their treatment outcome, as opposed to discovering useful features for treatment response prediction. To provide evidence for these claims, I present results of similar experiments I conducted on a public breast MRI dataset, where I demonstrate that the suspected data leakage mechanism closely reproduces the results reported on in the cited work.

引用的文章报告了一种经过训练的卷积神经网络,用于根据治疗前的乳腺核磁共振扫描结果预测新辅助化疗的反应。所提出的算法在测试数据集上取得了令人印象深刻的性能,其接收者操作特性曲线下的平均面积为 0.98,平均准确率为 88%。在这封信中,我担心所报告的结果可能是由于训练数据集和测试数据集之间的无意数据泄露造成的。更确切地说,我推测训练集和测试集中完整数据集的随机拆分不是发生在患者层面,而是发生在二维 MRI 切片层面。这使得神经网络能够 "记忆 "患者的解剖结构和治疗结果,而不是发现有用的治疗反应预测特征。为了给这些说法提供证据,我介绍了我在一个公开的乳腺 MRI 数据集上进行的类似实验的结果,我在实验中证明了疑似数据泄漏机制与所引用工作中报告的结果密切重现。
{"title":"Letter to the Editor Regarding Article “Prior to Initiation of Chemotherapy, Can We Predict Breast Tumor Response? Deep Learning Convolutional Neural Networks Approach Using a Breast MRI Tumor Dataset”","authors":"Joren Brunekreef","doi":"10.1007/s10278-024-01129-3","DOIUrl":"https://doi.org/10.1007/s10278-024-01129-3","url":null,"abstract":"<p>The cited article reports on a convolutional neural network trained to predict response to neoadjuvant chemotherapy from pre-treatment breast MRI scans. The proposed algorithm attains impressive performance on the test dataset with a mean Area Under the Receiver-Operating Characteristic curve of 0.98 and a mean accuracy of 88%. In this letter, I raise concerns that the reported results can be explained by inadvertent data leakage between training and test datasets. More precisely, I conjecture that the random split of the full dataset in training and test sets did not occur on a patient level, but rather on the level of 2D MRI slices. This allows the neural network to “memorize” a patient’s anatomy and their treatment outcome, as opposed to discovering useful features for treatment response prediction. To provide evidence for these claims, I present results of similar experiments I conducted on a public breast MRI dataset, where I demonstrate that the suspected data leakage mechanism closely reproduces the results reported on in the cited work.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"23 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-04-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140840363","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
UViT-Seg: An Efficient ViT and U-Net-Based Framework for Accurate Colorectal Polyp Segmentation in Colonoscopy and WCE Images UViT-Seg:基于 ViT 和 U-Net 的高效框架,用于在结肠镜和 WCE 图像中准确分割结直肠息肉
IF 4.4 2区 工程技术 Q2 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2024-04-26 DOI: 10.1007/s10278-024-01124-8
Yassine Oukdach, Anass Garbaz, Zakaria Kerkaou, Mohamed El Ansari, Lahcen Koutti, Ahmed Fouad El Ouafdi, Mouna Salihoun

Colorectal cancer (CRC) stands out as one of the most prevalent global cancers. The accurate localization of colorectal polyps in endoscopy images is pivotal for timely detection and removal, contributing significantly to CRC prevention. The manual analysis of images generated by gastrointestinal screening technologies poses a tedious task for doctors. Therefore, computer vision-assisted cancer detection could serve as an efficient tool for polyp segmentation. Numerous efforts have been dedicated to automating polyp localization, with the majority of studies relying on convolutional neural networks (CNNs) to learn features from polyp images. Despite their success in polyp segmentation tasks, CNNs exhibit significant limitations in precisely determining polyp location and shape due to their sole reliance on learning local features from images. While gastrointestinal images manifest significant variation in their features, encompassing both high- and low-level ones, a framework that combines the ability to learn both features of polyps is desired. This paper introduces UViT-Seg, a framework designed for polyp segmentation in gastrointestinal images. Operating on an encoder-decoder architecture, UViT-Seg employs two distinct feature extraction methods. A vision transformer in the encoder section captures long-range semantic information, while a CNN module, integrating squeeze-excitation and dual attention mechanisms, captures low-level features, focusing on critical image regions. Experimental evaluations conducted on five public datasets, including CVC clinic, ColonDB, Kvasir-SEG, ETIS LaribDB, and Kvasir Capsule-SEG, demonstrate UViT-Seg’s effectiveness in polyp localization. To confirm its generalization performance, the model is tested on datasets not used in training. Benchmarking against common segmentation methods and state-of-the-art polyp segmentation approaches, the proposed model yields promising results. For instance, it achieves a mean Dice coefficient of 0.915 and a mean intersection over union of 0.902 on the CVC Colon dataset. Furthermore, UViT-Seg has the advantage of being efficient, requiring fewer computational resources for both training and testing. This feature positions it as an optimal choice for real-world deployment scenarios.

结肠直肠癌(CRC)是全球发病率最高的癌症之一。内窥镜图像中结直肠息肉的准确定位对于及时发现和切除息肉至关重要,对预防结直肠癌大有裨益。人工分析胃肠道筛查技术生成的图像对医生来说是一项繁琐的任务。因此,计算机视觉辅助癌症检测可作为息肉分割的有效工具。目前已有许多研究致力于实现息肉定位的自动化,其中大多数研究依赖卷积神经网络(CNN)从息肉图像中学习特征。尽管卷积神经网络在息肉分割任务中取得了成功,但由于其完全依赖于从图像中学习局部特征,因此在精确确定息肉位置和形状方面表现出很大的局限性。胃肠道图像的特征变化很大,既有高层次特征,也有低层次特征,因此需要一个能同时学习息肉两种特征的框架。本文介绍的 UViT-Seg 是一个专为胃肠道图像中的息肉分割而设计的框架。UViT-Seg 采用编码器-解码器架构,采用两种不同的特征提取方法。编码器部分的视觉转换器可捕捉远距离语义信息,而集成了挤压激发和双重关注机制的 CNN 模块可捕捉低层次特征,重点关注关键图像区域。在五个公共数据集(包括 CVC clinic、ColonDB、Kvasir-SEG、ETIS LaribDB 和 Kvasir Capsule-SEG)上进行的实验评估证明了 UViT-Seg 在息肉定位方面的有效性。为了证实其通用性能,该模型在未用于训练的数据集上进行了测试。以常见的分割方法和最先进的息肉分割方法为基准,所提出的模型取得了令人满意的结果。例如,在 CVC 结肠数据集上,该模型的平均骰子系数(Dice coefficient)为 0.915,平均交集大于联合系数(intersection over union)为 0.902。此外,UViT-Seg 还具有高效的优势,训练和测试所需的计算资源都较少。这一特点使其成为实际部署场景的最佳选择。
{"title":"UViT-Seg: An Efficient ViT and U-Net-Based Framework for Accurate Colorectal Polyp Segmentation in Colonoscopy and WCE Images","authors":"Yassine Oukdach, Anass Garbaz, Zakaria Kerkaou, Mohamed El Ansari, Lahcen Koutti, Ahmed Fouad El Ouafdi, Mouna Salihoun","doi":"10.1007/s10278-024-01124-8","DOIUrl":"https://doi.org/10.1007/s10278-024-01124-8","url":null,"abstract":"<p>Colorectal cancer (CRC) stands out as one of the most prevalent global cancers. The accurate localization of colorectal polyps in endoscopy images is pivotal for timely detection and removal, contributing significantly to CRC prevention. The manual analysis of images generated by gastrointestinal screening technologies poses a tedious task for doctors. Therefore, computer vision-assisted cancer detection could serve as an efficient tool for polyp segmentation. Numerous efforts have been dedicated to automating polyp localization, with the majority of studies relying on convolutional neural networks (CNNs) to learn features from polyp images. Despite their success in polyp segmentation tasks, CNNs exhibit significant limitations in precisely determining polyp location and shape due to their sole reliance on learning local features from images. While gastrointestinal images manifest significant variation in their features, encompassing both high- and low-level ones, a framework that combines the ability to learn both features of polyps is desired. This paper introduces UViT-Seg, a framework designed for polyp segmentation in gastrointestinal images. Operating on an encoder-decoder architecture, UViT-Seg employs two distinct feature extraction methods. A vision transformer in the encoder section captures long-range semantic information, while a CNN module, integrating squeeze-excitation and dual attention mechanisms, captures low-level features, focusing on critical image regions. Experimental evaluations conducted on five public datasets, including CVC clinic, ColonDB, Kvasir-SEG, ETIS LaribDB, and Kvasir Capsule-SEG, demonstrate UViT-Seg’s effectiveness in polyp localization. To confirm its generalization performance, the model is tested on datasets not used in training. Benchmarking against common segmentation methods and state-of-the-art polyp segmentation approaches, the proposed model yields promising results. For instance, it achieves a mean Dice coefficient of 0.915 and a mean intersection over union of 0.902 on the CVC Colon dataset. Furthermore, UViT-Seg has the advantage of being efficient, requiring fewer computational resources for both training and testing. This feature positions it as an optimal choice for real-world deployment scenarios.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"43 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-04-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140799442","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Digital Imaging
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1