首页 > 最新文献

Journal of Medical Imaging最新文献

英文 中文
Optimizing mammography interpretation education: leveraging deep learning for cohort-specific error detection to enhance radiologist training. 优化乳腺 X 射线摄影解读教育:利用深度学习进行队列特定错误检测,以加强放射医师培训。
IF 1.9 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2024-09-01 Epub Date: 2024-10-03 DOI: 10.1117/1.JMI.11.5.055502
Xuetong Tao, Warren M Reed, Tong Li, Patrick C Brennan, Ziba Gandomkar

Purpose: Accurate interpretation of mammograms presents challenges. Tailoring mammography training to reader profiles holds the promise of an effective strategy to reduce these errors. This proof-of-concept study investigated the feasibility of employing convolutional neural networks (CNNs) with transfer learning to categorize regions associated with false-positive (FP) errors within screening mammograms into categories of "low" or "high" likelihood of being a false-positive detection for radiologists sharing similar geographic characteristics.

Approach: Mammography test sets assessed by two geographically distant cohorts of radiologists (cohorts A and B) were collected. FP patches within these mammograms were segmented and categorized as "difficult" or "easy" based on the number of readers committing FP errors. Patches outside 1.5 times the interquartile range above the upper quartile were labeled as difficult, whereas the remaining patches were labeled as easy. Using transfer learning, a patch-wise CNN model for binary patch classification was developed utilizing ResNet as the feature extractor, with modified fully connected layers for the target task. Model performance was assessed using 10-fold cross-validation.

Results: Compared with other architectures, the transferred ResNet-50 achieved the highest performance, obtaining receiver operating characteristics area under the curve values of 0.933 ( ± 0.012 ) and 0.975 ( ± 0.011 ) on the validation sets for cohorts A and B, respectively.

Conclusions: The findings highlight the feasibility of employing CNN-based transfer learning to predict the difficulty levels of local FP patches in screening mammograms for specific radiologist cohort with similar geographic characteristics.

目的:准确判读乳房 X 光照片是一项挑战。根据读者特征定制乳腺 X 光检查培训有望成为减少这些错误的有效策略。这项概念验证研究调查了利用卷积神经网络(CNN)和迁移学习将乳房X光筛查中与假阳性(FP)错误相关的区域分为 "低 "或 "高 "假阳性检测可能性类别的可行性:方法:收集两组地理位置相距较远的放射科医生(A 组和 B 组)评估的乳腺 X 光检查测试集。根据出现 FP 错误的读者人数,对这些乳房 X 光片中的 FP 补丁进行分割并分为 "难 "和 "易 "两类。超出上四分位数 1.5 倍四分位数间范围的片段标记为 "困难",而其余片段标记为 "容易"。利用迁移学习,我们开发了一个用于二元补丁分类的补丁全连接 CNN 模型,使用 ResNet 作为特征提取器,并针对目标任务修改了全连接层。模型性能通过 10 倍交叉验证进行评估:结果:与其他架构相比,转用的 ResNet-50 性能最高,在群组 A 和群组 B 的验证集上分别获得了 0.933 ( ± 0.012 ) 和 0.975 ( ± 0.011 ) 的接收器工作特性曲线下面积值:研究结果凸显了采用基于 CNN 的迁移学习来预测具有相似地理特征的特定放射科医师队列在乳房 X 光筛查中局部 FP 补丁的难度水平的可行性。
{"title":"Optimizing mammography interpretation education: leveraging deep learning for cohort-specific error detection to enhance radiologist training.","authors":"Xuetong Tao, Warren M Reed, Tong Li, Patrick C Brennan, Ziba Gandomkar","doi":"10.1117/1.JMI.11.5.055502","DOIUrl":"10.1117/1.JMI.11.5.055502","url":null,"abstract":"<p><strong>Purpose: </strong>Accurate interpretation of mammograms presents challenges. Tailoring mammography training to reader profiles holds the promise of an effective strategy to reduce these errors. This proof-of-concept study investigated the feasibility of employing convolutional neural networks (CNNs) with transfer learning to categorize regions associated with false-positive (FP) errors within screening mammograms into categories of \"low\" or \"high\" likelihood of being a false-positive detection for radiologists sharing similar geographic characteristics.</p><p><strong>Approach: </strong>Mammography test sets assessed by two geographically distant cohorts of radiologists (cohorts A and B) were collected. FP patches within these mammograms were segmented and categorized as \"difficult\" or \"easy\" based on the number of readers committing FP errors. Patches outside 1.5 times the interquartile range above the upper quartile were labeled as difficult, whereas the remaining patches were labeled as easy. Using transfer learning, a patch-wise CNN model for binary patch classification was developed utilizing ResNet as the feature extractor, with modified fully connected layers for the target task. Model performance was assessed using 10-fold cross-validation.</p><p><strong>Results: </strong>Compared with other architectures, the transferred ResNet-50 achieved the highest performance, obtaining receiver operating characteristics area under the curve values of 0.933 ( <math><mrow><mo>±</mo> <mn>0.012</mn></mrow> </math> ) and 0.975 ( <math><mrow><mo>±</mo> <mn>0.011</mn></mrow> </math> ) on the validation sets for cohorts A and B, respectively.</p><p><strong>Conclusions: </strong>The findings highlight the feasibility of employing CNN-based transfer learning to predict the difficulty levels of local FP patches in screening mammograms for specific radiologist cohort with similar geographic characteristics.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"11 5","pages":"055502"},"PeriodicalIF":1.9,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11447382/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142382053","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Predicting peritumoral glioblastoma infiltration and subsequent recurrence using deep-learning-based analysis of multi-parametric magnetic resonance imaging. 利用基于深度学习的多参数磁共振成像分析预测瘤周胶质母细胞瘤浸润和后续复发。
IF 1.9 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2024-09-01 Epub Date: 2024-08-30 DOI: 10.1117/1.JMI.11.5.054001
Sunwoo Kwak, Hamed Akbari, Jose A Garcia, Suyash Mohan, Yehuda Dicker, Chiharu Sako, Yuji Matsumoto, MacLean P Nasrallah, Mahmoud Shalaby, Donald M O'Rourke, Russel T Shinohara, Fang Liu, Chaitra Badve, Jill S Barnholtz-Sloan, Andrew E Sloan, Matthew Lee, Rajan Jain, Santiago Cepeda, Arnab Chakravarti, Joshua D Palmer, Adam P Dicker, Gaurav Shukla, Adam E Flanders, Wenyin Shi, Graeme F Woodworth, Christos Davatzikos

Purpose: Glioblastoma (GBM) is the most common and aggressive primary adult brain tumor. The standard treatment approach is surgical resection to target the enhancing tumor mass, followed by adjuvant chemoradiotherapy. However, malignant cells often extend beyond the enhancing tumor boundaries and infiltrate the peritumoral edema. Traditional supervised machine learning techniques hold potential in predicting tumor infiltration extent but are hindered by the extensive resources needed to generate expertly delineated regions of interest (ROIs) for training models on tissue most and least likely to be infiltrated.

Approach: We developed a method combining expert knowledge and training-based data augmentation to automatically generate numerous training examples, enhancing the accuracy of our model for predicting tumor infiltration through predictive maps. Such maps can be used for targeted supra-total surgical resection and other therapies that might benefit from intensive yet well-targeted treatment of infiltrated tissue. We apply our method to preoperative multi-parametric magnetic resonance imaging (mpMRI) scans from a subset of 229 patients of a multi-institutional consortium (Radiomics Signatures for Precision Diagnostics) and test the model on subsequent scans with pathology-proven recurrence.

Results: Leave-one-site-out cross-validation was used to train and evaluate the tumor infiltration prediction model using initial pre-surgical scans, comparing the generated prediction maps with follow-up mpMRI scans confirming recurrence through post-resection tissue analysis. Performance was measured by voxel-wised odds ratios (ORs) across six institutions: University of Pennsylvania (OR: 9.97), Ohio State University (OR: 14.03), Case Western Reserve University (OR: 8.13), New York University (OR: 16.43), Thomas Jefferson University (OR: 8.22), and Rio Hortega (OR: 19.48).

Conclusions: The proposed model demonstrates that mpMRI analysis using deep learning can predict infiltration in the peri-tumoral brain region for GBM patients without needing to train a model using expert ROI drawings. Results for each institution demonstrate the model's generalizability and reproducibility.

目的:胶质母细胞瘤(GBM)是最常见的侵袭性原发性成人脑肿瘤。标准的治疗方法是针对增大的肿瘤块进行手术切除,然后进行辅助化放疗。然而,恶性细胞往往会超出增强的肿瘤边界,浸润瘤周水肿。传统的有监督机器学习技术在预测肿瘤浸润范围方面具有潜力,但由于需要大量资源来生成专家划定的感兴趣区(ROI),以便对最有可能和最不可能浸润的组织进行模型训练,因此受到了阻碍:我们开发了一种方法,将专家知识与基于训练的数据增强相结合,自动生成大量训练示例,通过预测图提高模型预测肿瘤浸润的准确性。这种图谱可用于有针对性的超全切手术和其他疗法,对浸润组织进行密集而有针对性的治疗可能会使患者受益。我们将我们的方法应用于术前多参数磁共振成像(mpMRI)扫描,该扫描来自一个多机构联盟(精准诊断放射组学特征)的 229 位患者的子集,并在随后的病理证实复发扫描中对模型进行了测试:使用手术前的初始扫描结果对肿瘤浸润预测模型进行训练和评估,并将生成的预测图与通过切除术后组织分析确认复发的后续 mpMRI 扫描结果进行比较。该模型的性能由六个机构的体素化几率(ORs)来衡量:宾夕法尼亚大学(OR:9.97)、俄亥俄州立大学(OR:14.03)、凯斯西储大学(OR:8.13)、纽约大学(OR:16.43)、托马斯杰斐逊大学(OR:8.22)和里奥霍尔特加大学(OR:19.48):所提出的模型表明,使用深度学习进行 mpMRI 分析可以预测 GBM 患者肿瘤周围脑区的浸润情况,而无需使用专家的 ROI 图纸来训练模型。每个机构的结果都证明了该模型的通用性和可重复性。
{"title":"Predicting peritumoral glioblastoma infiltration and subsequent recurrence using deep-learning-based analysis of multi-parametric magnetic resonance imaging.","authors":"Sunwoo Kwak, Hamed Akbari, Jose A Garcia, Suyash Mohan, Yehuda Dicker, Chiharu Sako, Yuji Matsumoto, MacLean P Nasrallah, Mahmoud Shalaby, Donald M O'Rourke, Russel T Shinohara, Fang Liu, Chaitra Badve, Jill S Barnholtz-Sloan, Andrew E Sloan, Matthew Lee, Rajan Jain, Santiago Cepeda, Arnab Chakravarti, Joshua D Palmer, Adam P Dicker, Gaurav Shukla, Adam E Flanders, Wenyin Shi, Graeme F Woodworth, Christos Davatzikos","doi":"10.1117/1.JMI.11.5.054001","DOIUrl":"10.1117/1.JMI.11.5.054001","url":null,"abstract":"<p><strong>Purpose: </strong>Glioblastoma (GBM) is the most common and aggressive primary adult brain tumor. The standard treatment approach is surgical resection to target the enhancing tumor mass, followed by adjuvant chemoradiotherapy. However, malignant cells often extend beyond the enhancing tumor boundaries and infiltrate the peritumoral edema. Traditional supervised machine learning techniques hold potential in predicting tumor infiltration extent but are hindered by the extensive resources needed to generate expertly delineated regions of interest (ROIs) for training models on tissue most and least likely to be infiltrated.</p><p><strong>Approach: </strong>We developed a method combining expert knowledge and training-based data augmentation to automatically generate numerous training examples, enhancing the accuracy of our model for predicting tumor infiltration through predictive maps. Such maps can be used for targeted supra-total surgical resection and other therapies that might benefit from intensive yet well-targeted treatment of infiltrated tissue. We apply our method to preoperative multi-parametric magnetic resonance imaging (mpMRI) scans from a subset of 229 patients of a multi-institutional consortium (Radiomics Signatures for Precision Diagnostics) and test the model on subsequent scans with pathology-proven recurrence.</p><p><strong>Results: </strong>Leave-one-site-out cross-validation was used to train and evaluate the tumor infiltration prediction model using initial pre-surgical scans, comparing the generated prediction maps with follow-up mpMRI scans confirming recurrence through post-resection tissue analysis. Performance was measured by voxel-wised odds ratios (ORs) across six institutions: University of Pennsylvania (OR: 9.97), Ohio State University (OR: 14.03), Case Western Reserve University (OR: 8.13), New York University (OR: 16.43), Thomas Jefferson University (OR: 8.22), and Rio Hortega (OR: 19.48).</p><p><strong>Conclusions: </strong>The proposed model demonstrates that mpMRI analysis using deep learning can predict infiltration in the peri-tumoral brain region for GBM patients without needing to train a model using expert ROI drawings. Results for each institution demonstrate the model's generalizability and reproducibility.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"11 5","pages":"054001"},"PeriodicalIF":1.9,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11363410/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142113462","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep learning architecture for scatter estimation in cone-beam computed tomography head imaging with varying field-of-measurement settings. 用于锥束计算机断层扫描头部成像中散射估计的深度学习架构,具有不同的测量场设置。
IF 1.9 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2024-09-01 Epub Date: 2024-10-15 DOI: 10.1117/1.JMI.11.5.053501
Harshit Agrawal, Ari Hietanen, Simo Särkkä

Purpose: X-ray scatter causes considerable degradation in the cone-beam computed tomography (CBCT) image quality. To estimate the scatter, deep learning-based methods have been demonstrated to be effective. Modern CBCT systems can scan a wide range of field-of-measurement (FOM) sizes. Variations in the size of FOM can cause a major shift in the scatter-to-primary ratio in CBCT. However, the scatter estimation performance of deep learning networks has not been extensively evaluated under varying FOMs. Therefore, we train the state-of-the-art scatter estimation neural networks for varying FOMs and develop a method to utilize FOM size information to improve performance.

Approach: We used FOM size information as additional features by converting it into two channels and then concatenating it to the encoder of the networks. We compared our approach for a U-Net, Spline-Net, and DSE-Net, by training them with and without the FOM information. We utilized a Monte Carlo-simulated dataset to train the networks on 18 FOM sizes and test on 30 unseen FOM sizes. In addition, we evaluated the models on the water phantoms and real clinical CBCT scans.

Results: The simulation study demonstrates that our method reduced average mean-absolute-percentage-error for U-Net by 38%, Spline-Net by 40%, and DSE-net by 33% for the scatter estimation in the 2D projection domain. Furthermore, the root-mean-square error on the 3D reconstructed volumes was improved for U-Net by 43%, Spline-Net by 30%, and DSE-Net by 23%. Furthermore, our method improved contrast and image quality on real datasets such as water phantom and clinical data.

Conclusion: Providing additional information about FOM size improves the robustness of the neural networks for scatter estimation. Our approach is not limited to utilizing only FOM size information; more variables such as tube voltage, scanning geometry, and patient size can be added to improve the robustness of a single network.

目的:X 射线散射会大大降低锥束计算机断层扫描(CBCT)图像的质量。要估计散射,基于深度学习的方法已被证明是有效的。现代 CBCT 系统可以扫描各种尺寸的测量场(FOM)。FOM 大小的变化会导致 CBCT 中的散射比发生重大变化。然而,深度学习网络的散点估计性能尚未在不同的 FOMs 下得到广泛评估。因此,我们对最先进的散点估计神经网络进行了针对不同FOM的训练,并开发了一种利用FOM大小信息提高性能的方法:方法:我们利用 FOM 大小信息作为附加特征,将其转换为两个通道,然后将其连接到网络编码器中。我们对 U-Net、Spline-Net 和 DSE-Net 的方法进行了比较,分别使用和不使用 FOM 信息对它们进行了训练。我们利用蒙特卡洛模拟数据集在 18 种 FOM 大小上训练网络,并在 30 种未见 FOM 大小上进行测试。此外,我们还在水模型和真实临床 CBCT 扫描上对模型进行了评估:模拟研究表明,在二维投影域的散点估计中,我们的方法将 U-Net 的平均均值-绝对值-百分比误差降低了 38%,将 Spline-Net 的平均均值-绝对值-百分比误差降低了 40%,将 DSE-net 的平均均值-绝对值-百分比误差降低了 33%。此外,在三维重建体积的均方根误差方面,U-Net 提高了 43%,Spline-Net 提高了 30%,DSE-Net 提高了 23%。此外,我们的方法还提高了真实数据集(如水模型和临床数据)的对比度和图像质量:结论:提供有关 FOM 大小的额外信息可提高神经网络对散射估计的鲁棒性。我们的方法不仅限于利用 FOM 大小信息,还可以添加更多变量,如管电压、扫描几何形状和患者体型,以提高单个网络的鲁棒性。
{"title":"Deep learning architecture for scatter estimation in cone-beam computed tomography head imaging with varying field-of-measurement settings.","authors":"Harshit Agrawal, Ari Hietanen, Simo Särkkä","doi":"10.1117/1.JMI.11.5.053501","DOIUrl":"https://doi.org/10.1117/1.JMI.11.5.053501","url":null,"abstract":"<p><strong>Purpose: </strong>X-ray scatter causes considerable degradation in the cone-beam computed tomography (CBCT) image quality. To estimate the scatter, deep learning-based methods have been demonstrated to be effective. Modern CBCT systems can scan a wide range of field-of-measurement (FOM) sizes. Variations in the size of FOM can cause a major shift in the scatter-to-primary ratio in CBCT. However, the scatter estimation performance of deep learning networks has not been extensively evaluated under varying FOMs. Therefore, we train the state-of-the-art scatter estimation neural networks for varying FOMs and develop a method to utilize FOM size information to improve performance.</p><p><strong>Approach: </strong>We used FOM size information as additional features by converting it into two channels and then concatenating it to the encoder of the networks. We compared our approach for a U-Net, Spline-Net, and DSE-Net, by training them with and without the FOM information. We utilized a Monte Carlo-simulated dataset to train the networks on 18 FOM sizes and test on 30 unseen FOM sizes. In addition, we evaluated the models on the water phantoms and real clinical CBCT scans.</p><p><strong>Results: </strong>The simulation study demonstrates that our method reduced average mean-absolute-percentage-error for U-Net by 38%, Spline-Net by 40%, and DSE-net by 33% for the scatter estimation in the 2D projection domain. Furthermore, the root-mean-square error on the 3D reconstructed volumes was improved for U-Net by 43%, Spline-Net by 30%, and DSE-Net by 23%. Furthermore, our method improved contrast and image quality on real datasets such as water phantom and clinical data.</p><p><strong>Conclusion: </strong>Providing additional information about FOM size improves the robustness of the neural networks for scatter estimation. Our approach is not limited to utilizing only FOM size information; more variables such as tube voltage, scanning geometry, and patient size can be added to improve the robustness of a single network.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"11 5","pages":"053501"},"PeriodicalIF":1.9,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11477364/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142477765","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Demystifying the effect of receptive field size in U-Net models for medical image segmentation. 揭示用于医学图像分割的 U-Net 模型中感受野大小的影响。
IF 1.9 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2024-09-01 Epub Date: 2024-10-29 DOI: 10.1117/1.JMI.11.5.054004
Vincent Loos, Rohit Pardasani, Navchetan Awasthi

Purpose: Medical image segmentation is a critical task in healthcare applications, and U-Nets have demonstrated promising results in this domain. We delve into the understudied aspect of receptive field (RF) size and its impact on the U-Net and attention U-Net architectures used for medical imaging segmentation.

Approach: We explore several critical elements including the relationship among RF size, characteristics of the region of interest, and model performance, as well as the balance between RF size and computational costs for U-Net and attention U-Net methods for different datasets. We also propose a mathematical notation for representing the theoretical receptive field (TRF) of a given layer in a network and propose two new metrics, namely, the effective receptive field (ERF) rate and the object rate, to quantify the fraction of significantly contributing pixels within the ERF against the TRF area and assessing the relative size of the segmentation object compared with the TRF size, respectively.

Results: The results demonstrate that there exists an optimal TRF size that successfully strikes a balance between capturing a wider global context and maintaining computational efficiency, thereby optimizing model performance. Interestingly, a distinct correlation is observed between the data complexity and the required TRF size; segmentation based solely on contrast achieved peak performance even with smaller TRF sizes, whereas more complex segmentation tasks necessitated larger TRFs. Attention U-Net models consistently outperformed their U-Net counterparts, highlighting the value of attention mechanisms regardless of TRF size.

Conclusions: These insights present an invaluable resource for developing more efficient U-Net-based architectures for medical imaging and pave the way for future exploration of other segmentation architectures. A tool is also developed, which calculates the TRF for a U-Net (and attention U-Net) model and also suggests an appropriate TRF size for a given model and dataset.

目的:医学图像分割是医疗保健应用中的一项关键任务,U-Net 在这一领域取得了令人鼓舞的成果。我们深入研究了未被充分研究的感受野(RF)大小及其对用于医学影像分割的 U-Net 和注意力 U-Net 架构的影响:方法:我们探讨了几个关键因素,包括感受野大小、感兴趣区域特征和模型性能之间的关系,以及针对不同数据集的 U-Net 和注意力 U-Net 方法在感受野大小和计算成本之间的平衡。我们还提出了一种数学符号,用于表示网络中给定层的理论感受野(TRF),并提出了两个新指标,即有效感受野(ERF)率和对象率,分别用于量化ERF内对TRF区域有显著贡献的像素的比例,以及评估分割对象与TRF大小相比的相对大小:结果表明,存在一个最佳的 TRF 大小,它能成功地在捕捉更广泛的全局背景和保持计算效率之间取得平衡,从而优化模型性能。有趣的是,数据复杂度与所需的 TRF 大小之间存在明显的相关性;即使 TRF 大小较小,仅基于对比度的分割也能达到峰值性能,而更复杂的分割任务则需要更大的 TRF。注意力 U-Net 模型的表现始终优于其 U-Net 模型,这凸显了注意力机制的价值,无论 TRF 大小如何:这些见解为开发更高效的基于 U-Net 的医学成像架构提供了宝贵的资源,并为未来探索其他分割架构铺平了道路。此外,还开发了一种工具,用于计算 U-Net(和注意力 U-Net)模型的 TRF,并为给定模型和数据集建议合适的 TRF 大小。
{"title":"Demystifying the effect of receptive field size in U-Net models for medical image segmentation.","authors":"Vincent Loos, Rohit Pardasani, Navchetan Awasthi","doi":"10.1117/1.JMI.11.5.054004","DOIUrl":"10.1117/1.JMI.11.5.054004","url":null,"abstract":"<p><strong>Purpose: </strong>Medical image segmentation is a critical task in healthcare applications, and U-Nets have demonstrated promising results in this domain. We delve into the understudied aspect of receptive field (RF) size and its impact on the U-Net and attention U-Net architectures used for medical imaging segmentation.</p><p><strong>Approach: </strong>We explore several critical elements including the relationship among RF size, characteristics of the region of interest, and model performance, as well as the balance between RF size and computational costs for U-Net and attention U-Net methods for different datasets. We also propose a mathematical notation for representing the theoretical receptive field (TRF) of a given layer in a network and propose two new metrics, namely, the effective receptive field (ERF) rate and the object rate, to quantify the fraction of significantly contributing pixels within the ERF against the TRF area and assessing the relative size of the segmentation object compared with the TRF size, respectively.</p><p><strong>Results: </strong>The results demonstrate that there exists an optimal TRF size that successfully strikes a balance between capturing a wider global context and maintaining computational efficiency, thereby optimizing model performance. Interestingly, a distinct correlation is observed between the data complexity and the required TRF size; segmentation based solely on contrast achieved peak performance even with smaller TRF sizes, whereas more complex segmentation tasks necessitated larger TRFs. Attention U-Net models consistently outperformed their U-Net counterparts, highlighting the value of attention mechanisms regardless of TRF size.</p><p><strong>Conclusions: </strong>These insights present an invaluable resource for developing more efficient U-Net-based architectures for medical imaging and pave the way for future exploration of other segmentation architectures. A tool is also developed, which calculates the TRF for a U-Net (and attention U-Net) model and also suggests an appropriate TRF size for a given model and dataset.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"11 5","pages":"054004"},"PeriodicalIF":1.9,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11520766/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142548314","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HarmonyTM: multi-center data harmonization applied to distributed learning for Parkinson's disease classification. HarmonyTM:多中心数据协调应用于帕金森病分类的分布式学习。
IF 1.9 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2024-09-01 Epub Date: 2024-09-20 DOI: 10.1117/1.JMI.11.5.054502
Raissa Souza, Emma A M Stanley, Vedant Gulve, Jasmine Moore, Chris Kang, Richard Camicioli, Oury Monchi, Zahinoor Ismail, Matthias Wilms, Nils D Forkert

Purpose: Distributed learning is widely used to comply with data-sharing regulations and access diverse datasets for training machine learning (ML) models. The traveling model (TM) is a distributed learning approach that sequentially trains with data from one center at a time, which is especially advantageous when dealing with limited local datasets. However, a critical concern emerges when centers utilize different scanners for data acquisition, which could potentially lead models to exploit these differences as shortcuts. Although data harmonization can mitigate this issue, current methods typically rely on large or paired datasets, which can be impractical to obtain in distributed setups.

Approach: We introduced HarmonyTM, a data harmonization method tailored for the TM. HarmonyTM effectively mitigates bias in the model's feature representation while retaining crucial disease-related information, all without requiring extensive datasets. Specifically, we employed adversarial training to "unlearn" bias from the features used in the model for classifying Parkinson's disease (PD). We evaluated HarmonyTM using multi-center three-dimensional (3D) neuroimaging datasets from 83 centers using 23 different scanners.

Results: Our results show that HarmonyTM improved PD classification accuracy from 72% to 76% and reduced (unwanted) scanner classification accuracy from 53% to 30% in the TM setup.

Conclusion: HarmonyTM is a method tailored for harmonizing 3D neuroimaging data within the TM approach, aiming to minimize shortcut learning in distributed setups. This prevents the disease classifier from leveraging scanner-specific details to classify patients with or without PD-a key aspect for deploying ML models for clinical applications.

目的:分布式学习被广泛应用于遵守数据共享法规和访问各种数据集以训练机器学习(ML)模型。巡回模型 (TM) 是一种分布式学习方法,每次使用一个中心的数据进行顺序训练,这在处理有限的本地数据集时尤其有利。然而,当各中心使用不同的扫描仪采集数据时,就会出现一个重要的问题,这可能会导致模型利用这些差异作为捷径。虽然数据协调可以缓解这一问题,但目前的方法通常依赖于大型或成对的数据集,而在分布式设置中获取这些数据集可能并不现实:我们引入了 HarmonyTM,这是一种专为 TM 量身定制的数据协调方法。HarmonyTM 能有效减少模型特征表示中的偏差,同时保留关键的疾病相关信息,而这一切都不需要大量的数据集。具体来说,我们采用对抗训练来 "消除 "用于帕金森病(PD)分类模型的特征中的偏差。我们使用来自 83 个中心、使用 23 种不同扫描仪的多中心三维(3D)神经成像数据集对 HarmonyTM 进行了评估:结果表明,在 TM 设置中,HarmonyTM 将 PD 分类准确率从 72% 提高到 76%,将(不需要的)扫描仪分类准确率从 53% 降低到 30%:HarmonyTM 是一种在 TM 方法中协调三维神经成像数据的定制方法,旨在最大限度地减少分布式设置中的捷径学习。这可以防止疾病分类器利用扫描仪的特定细节来对患有或不患有帕金森病的患者进行分类--这是在临床应用中部署 ML 模型的关键环节。
{"title":"HarmonyTM: multi-center data harmonization applied to distributed learning for Parkinson's disease classification.","authors":"Raissa Souza, Emma A M Stanley, Vedant Gulve, Jasmine Moore, Chris Kang, Richard Camicioli, Oury Monchi, Zahinoor Ismail, Matthias Wilms, Nils D Forkert","doi":"10.1117/1.JMI.11.5.054502","DOIUrl":"10.1117/1.JMI.11.5.054502","url":null,"abstract":"<p><strong>Purpose: </strong>Distributed learning is widely used to comply with data-sharing regulations and access diverse datasets for training machine learning (ML) models. The traveling model (TM) is a distributed learning approach that sequentially trains with data from one center at a time, which is especially advantageous when dealing with limited local datasets. However, a critical concern emerges when centers utilize different scanners for data acquisition, which could potentially lead models to exploit these differences as shortcuts. Although data harmonization can mitigate this issue, current methods typically rely on large or paired datasets, which can be impractical to obtain in distributed setups.</p><p><strong>Approach: </strong>We introduced HarmonyTM, a data harmonization method tailored for the TM. HarmonyTM effectively mitigates bias in the model's feature representation while retaining crucial disease-related information, all without requiring extensive datasets. Specifically, we employed adversarial training to \"unlearn\" bias from the features used in the model for classifying Parkinson's disease (PD). We evaluated HarmonyTM using multi-center three-dimensional (3D) neuroimaging datasets from 83 centers using 23 different scanners.</p><p><strong>Results: </strong>Our results show that HarmonyTM improved PD classification accuracy from 72% to 76% and reduced (unwanted) scanner classification accuracy from 53% to 30% in the TM setup.</p><p><strong>Conclusion: </strong>HarmonyTM is a method tailored for harmonizing 3D neuroimaging data within the TM approach, aiming to minimize shortcut learning in distributed setups. This prevents the disease classifier from leveraging scanner-specific details to classify patients with or without PD-a key aspect for deploying ML models for clinical applications.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"11 5","pages":"054502"},"PeriodicalIF":1.9,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11413651/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142298698","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Expanding generalized contrast-to-noise ratio into a clinically relevant measure of lesion detectability by considering size and spatial resolution. 通过考虑大小和空间分辨率,将广义对比度与噪声比扩展为一种与临床相关的病变可探测性测量方法。
IF 1.9 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2024-09-01 Epub Date: 2024-10-23 DOI: 10.1117/1.JMI.11.5.057001
Siegfried Schlunk, Brett Byram

Purpose: Early image quality metrics were often designed with clinicians in mind, and ideal metrics would correlate with the subjective opinion of practitioners. Over time, adaptive beamformers and other post-processing methods have become more common, and these newer methods often violate assumptions of earlier image quality metrics, invalidating the meaning of those metrics. The result is that beamformers may "manipulate" metrics without producing more clinical information.

Approach: In this work, Smith et al.'s signal-to-noise ratio (SNR) metric for lesion detectability is considered, and a more robust version, here called generalized SNR (gSNR), is proposed that uses generalized contrast-to-noise ratio (gCNR) as a core. It is analytically shown that for Rayleigh distributed data, gCNR is a function of Smith et al.'s C ψ (and therefore can be used as a substitution). More robust methods for estimating the resolution cell size are considered. Simulated lesions are included to verify the equations and demonstrate behavior, and it is shown to apply equally well to in vivo data.

Results: gSNR is shown to be equivalent to SNR for delay-and-sum (DAS) beamformed data, as intended. However, it is shown to be more robust against transformations and report lesion detectability more accurately for non-Rayleigh distributed data. In the simulation included, the SNR of DAS was 4.4 ± 0.8 , and minimum variance (MV) was 6.4 ± 1.9 , but the gSNR of DAS was 4.5 ± 0.9 , and MV was 3.0 ± 0.9 , which agrees with the subjective assessment of the image. Likewise, the DAS 2 transformation (which is clinically identical to DAS) had an incorrect SNR of 9.4 ± 1.0 and a correct gSNR of 4.4 ± 0.9 . Similar results are shown in vivo.

Conclusions: Using gCNR as a component to estimate gSNR creates a robust measure of lesion detectability. Like SNR, gSNR can be compared with the Rose criterion and may better correlate with clinical assessments of image quality for modern beamformers.

目的:早期的图像质量指标通常是以临床医生为中心设计的,理想的指标与医生的主观意见相关。随着时间的推移,自适应波束成形器和其他后处理方法变得越来越普遍,而这些新方法往往违反了早期图像质量指标的假设,使这些指标的意义失效。其结果是,波束成形器可能会 "操纵 "指标,而不会产生更多临床信息:在这项工作中,考虑了 Smith 等人的信噪比(SNR)病变可探测性指标,并提出了一个更稳健的版本,这里称为广义信噪比(gSNR),它以广义对比度-噪声比(gCNR)为核心。分析表明,对于瑞利分布数据,gCNR 是 Smith 等人的 C ψ 的函数(因此可用作替代)。我们还考虑了估算分辨单元大小的更稳健的方法。结果表明,gSNR 与延迟和(DAS)波束成形数据的信噪比相当。然而,对于非瑞利分布式数据,gSNR 对变换具有更强的鲁棒性,并能更准确地报告病变可探测性。在模拟中,DAS 的 SNR 为 4.4 ± 0.8,最小方差 (MV) 为 6.4 ± 1.9,但 DAS 的 gSNR 为 4.5 ± 0.9,MV 为 3.0 ± 0.9,这与图像的主观评估一致。同样,DAS 2 转换(临床上与 DAS 相同)的不正确 SNR 为 9.4 ± 1.0,而正确的 gSNR 为 4.4 ± 0.9。类似的结果在体内也有显示:结论:使用 gCNR 作为估算 gSNR 的一个组成部分,可以稳健地衡量病变的可探测性。与信噪比一样,gSNR 也可以与罗斯标准进行比较,并能更好地与临床评估现代波束成形器的图像质量相关联。
{"title":"Expanding generalized contrast-to-noise ratio into a clinically relevant measure of lesion detectability by considering size and spatial resolution.","authors":"Siegfried Schlunk, Brett Byram","doi":"10.1117/1.JMI.11.5.057001","DOIUrl":"https://doi.org/10.1117/1.JMI.11.5.057001","url":null,"abstract":"<p><strong>Purpose: </strong>Early image quality metrics were often designed with clinicians in mind, and ideal metrics would correlate with the subjective opinion of practitioners. Over time, adaptive beamformers and other post-processing methods have become more common, and these newer methods often violate assumptions of earlier image quality metrics, invalidating the meaning of those metrics. The result is that beamformers may \"manipulate\" metrics without producing more clinical information.</p><p><strong>Approach: </strong>In this work, Smith et al.'s signal-to-noise ratio (SNR) metric for lesion detectability is considered, and a more robust version, here called generalized SNR (gSNR), is proposed that uses generalized contrast-to-noise ratio (gCNR) as a core. It is analytically shown that for Rayleigh distributed data, gCNR is a function of Smith et al.'s <math> <mrow><msub><mi>C</mi> <mi>ψ</mi></msub> </mrow> </math> (and therefore can be used as a substitution). More robust methods for estimating the resolution cell size are considered. Simulated lesions are included to verify the equations and demonstrate behavior, and it is shown to apply equally well to <i>in vivo</i> data.</p><p><strong>Results: </strong>gSNR is shown to be equivalent to SNR for delay-and-sum (DAS) beamformed data, as intended. However, it is shown to be more robust against transformations and report lesion detectability more accurately for non-Rayleigh distributed data. In the simulation included, the SNR of DAS was <math><mrow><mn>4.4</mn> <mo>±</mo> <mn>0.8</mn></mrow> </math> , and minimum variance (MV) was <math><mrow><mn>6.4</mn> <mo>±</mo> <mn>1.9</mn></mrow> </math> , but the gSNR of DAS was <math><mrow><mn>4.5</mn> <mo>±</mo> <mn>0.9</mn></mrow> </math> , and MV was <math><mrow><mn>3.0</mn> <mo>±</mo> <mn>0.9</mn></mrow> </math> , which agrees with the subjective assessment of the image. Likewise, the <math> <mrow><msup><mi>DAS</mi> <mn>2</mn></msup> </mrow> </math> transformation (which is clinically identical to DAS) had an incorrect SNR of <math><mrow><mn>9.4</mn> <mo>±</mo> <mn>1.0</mn></mrow> </math> and a correct gSNR of <math><mrow><mn>4.4</mn> <mo>±</mo> <mn>0.9</mn></mrow> </math> . Similar results are shown <i>in vivo</i>.</p><p><strong>Conclusions: </strong>Using gCNR as a component to estimate gSNR creates a robust measure of lesion detectability. Like SNR, gSNR can be compared with the Rose criterion and may better correlate with clinical assessments of image quality for modern beamformers.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"11 5","pages":"057001"},"PeriodicalIF":1.9,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11498315/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142510423","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automated echocardiography view classification and quality assessment with recognition of unknown views. 自动超声心动图视图分类和质量评估,可识别未知视图。
IF 1.9 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2024-09-01 Epub Date: 2024-08-30 DOI: 10.1117/1.JMI.11.5.054002
Gino E Jansen, Bob D de Vos, Mitchel A Molenaar, Mark J Schuuring, Berto J Bouma, Ivana Išgum

Purpose: Interpreting echocardiographic exams requires substantial manual interaction as videos lack scan-plane information and have inconsistent image quality, ranging from clinically relevant to unrecognizable. Thus, a manual prerequisite step for analysis is to select the appropriate views that showcase both the target anatomy and optimal image quality. To automate this selection process, we present a method for automatic classification of routine views, recognition of unknown views, and quality assessment of detected views.

Approach: We train a neural network for view classification and employ the logit activations from the neural network for unknown view recognition. Subsequently, we train a linear regression algorithm that uses feature embeddings from the neural network to predict view quality scores. We evaluate the method on a clinical test set of 2466 echocardiography videos with expert-annotated view labels and a subset of 438 videos with expert-rated view quality scores. A second observer annotated a subset of 894 videos, including all quality-rated videos.

Results: The proposed method achieved an accuracy of 84.9 % ± 0.67 for the joint objective of routine view classification and unknown view recognition, whereas a second observer reached an accuracy of 87.6%. For view quality assessment, the method achieved a Spearman's rank correlation coefficient of 0.71, whereas a second observer reached a correlation coefficient of 0.62.

Conclusion: The proposed method approaches expert-level performance, enabling fully automatic selection of the most appropriate views for manual or automatic downstream analysis.

目的:解读超声心动图检查需要大量的人工互动,因为视频缺乏扫描平面信息,图像质量也不一致,有的与临床相关,有的则无法识别。因此,人工分析的先决条件是选择适当的视图,既能显示目标解剖结构,又能获得最佳图像质量。为了实现这一选择过程的自动化,我们提出了一种方法,用于对常规视图进行自动分类、识别未知视图以及对检测到的视图进行质量评估:方法:我们训练一个神经网络来进行视图分类,并利用神经网络的对数激活来识别未知视图。随后,我们训练了一种线性回归算法,利用神经网络的特征嵌入来预测视图质量得分。我们在一个临床测试集上对该方法进行了评估,该测试集包含 2466 个超声心动图视频,其中有专家标注的视图标签和 438 个专家评定视图质量分数的视频子集。第二名观察者对 894 个视频子集(包括所有质量评分视频)进行了注释:结果:在常规视图分类和未知视图识别的共同目标上,所提出的方法达到了 84.9 % ± 0.67 的准确率,而第二位观察者的准确率则达到了 87.6%。在视图质量评估方面,该方法的斯皮尔曼等级相关系数为 0.71,而第二观察者的相关系数为 0.62:所提出的方法性能接近专家水平,能够全自动选择最合适的视图,用于人工或自动下游分析。
{"title":"Automated echocardiography view classification and quality assessment with recognition of unknown views.","authors":"Gino E Jansen, Bob D de Vos, Mitchel A Molenaar, Mark J Schuuring, Berto J Bouma, Ivana Išgum","doi":"10.1117/1.JMI.11.5.054002","DOIUrl":"10.1117/1.JMI.11.5.054002","url":null,"abstract":"<p><strong>Purpose: </strong>Interpreting echocardiographic exams requires substantial manual interaction as videos lack scan-plane information and have inconsistent image quality, ranging from clinically relevant to unrecognizable. Thus, a manual prerequisite step for analysis is to select the appropriate views that showcase both the target anatomy and optimal image quality. To automate this selection process, we present a method for automatic classification of routine views, recognition of unknown views, and quality assessment of detected views.</p><p><strong>Approach: </strong>We train a neural network for view classification and employ the logit activations from the neural network for unknown view recognition. Subsequently, we train a linear regression algorithm that uses feature embeddings from the neural network to predict view quality scores. We evaluate the method on a clinical test set of 2466 echocardiography videos with expert-annotated view labels and a subset of 438 videos with expert-rated view quality scores. A second observer annotated a subset of 894 videos, including all quality-rated videos.</p><p><strong>Results: </strong>The proposed method achieved an accuracy of <math><mrow><mn>84.9</mn> <mo>%</mo> <mo>±</mo> <mn>0.67</mn></mrow> </math> for the joint objective of routine view classification and unknown view recognition, whereas a second observer reached an accuracy of 87.6%. For view quality assessment, the method achieved a Spearman's rank correlation coefficient of 0.71, whereas a second observer reached a correlation coefficient of 0.62.</p><p><strong>Conclusion: </strong>The proposed method approaches expert-level performance, enabling fully automatic selection of the most appropriate views for manual or automatic downstream analysis.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"11 5","pages":"054002"},"PeriodicalIF":1.9,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11364256/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142113461","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep network and multi-atlas segmentation fusion for delineation of thigh muscle groups in three-dimensional water-fat separated MRI. 融合深度网络和多图谱分割技术,在三维水脂分离磁共振成像中划分大腿肌肉群。
IF 1.9 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2024-09-01 Epub Date: 2024-09-03 DOI: 10.1117/1.JMI.11.5.054003
Nagasoujanya V Annasamudram, Azubuike M Okorie, Richard G Spencer, Rita R Kalyani, Qi Yang, Bennett A Landman, Luigi Ferrucci, Sokratis Makrogiannis

Purpose: Segmentation is essential for tissue quantification and characterization in studies of aging and age-related and metabolic diseases and the development of imaging biomarkers. We propose a multi-method and multi-atlas methodology for automated segmentation of functional muscle groups in three-dimensional (3D) thigh magnetic resonance images. These groups lie anatomically adjacent to each other, rendering their manual delineation a challenging and time-consuming task.

Approach: We introduce a framework for automated segmentation of the four main functional muscle groups of the thigh, gracilis, hamstring, quadriceps femoris, and sartorius, using chemical shift encoded water-fat magnetic resonance imaging (CSE-MRI). We propose fusing anatomical mappings from multiple deformable models with 3D deep learning model-based segmentation. This approach leverages the generalizability of multi-atlas segmentation (MAS) and accuracy of deep networks, hence enabling accurate assessment of volume and fat content of muscle groups.

Results: For segmentation performance evaluation, we calculated the Dice similarity coefficient (DSC) and Hausdorff distance 95th percentile (HD-95). We evaluated the proposed framework, its variants, and baseline methods on 15 healthy subjects by threefold cross-validation and tested on four patients. Fusion of multiple atlases, deformable registration models, and deep learning segmentation produced the top performance with an average DSC of 0.859 and HD-95 of 8.34 over all muscles.

Conclusions: Fusion of multiple anatomical mappings from multiple MAS techniques enriches the template set and improves the segmentation accuracy. Additional fusion with deep network decisions applied to the subject space offers complementary information. The proposed approach can produce accurate segmentation of individual muscle groups in 3D thigh MRI scans.

目的:在研究衰老、年龄相关疾病和代谢性疾病以及开发成像生物标记物时,分割对于组织量化和特征描述至关重要。我们提出了一种多方法和多图谱方法,用于自动分割三维(3D)大腿磁共振图像中的功能性肌肉群。这些肌群在解剖学上彼此相邻,因此人工划分这些肌群是一项具有挑战性且耗时的任务:方法:我们采用化学位移编码水脂磁共振成像(CSE-MRI)技术,为自动分割大腿的四个主要功能肌群(腓肠肌、腘绳肌、股四头肌和滑肌)提供了一个框架。我们建议将多个可变形模型的解剖映射与基于三维深度学习模型的分割相结合。这种方法充分利用了多图谱分割(MAS)的通用性和深度网络的准确性,从而能够准确评估肌肉群的体积和脂肪含量:为了评估分割性能,我们计算了戴斯相似系数(DSC)和豪斯多夫距离第 95 百分位数(HD-95)。通过三重交叉验证,我们在 15 名健康受试者身上评估了所提出的框架、其变体和基线方法,并在 4 名患者身上进行了测试。融合多种地图集、可变形配准模型和深度学习分割法产生了最佳性能,所有肌肉的平均 DSC 为 0.859,HD-95 为 8.34:结论:融合多种 MAS 技术的多种解剖映射可丰富模板集,提高分割准确性。与应用于主体空间的深度网络决策的额外融合提供了补充信息。所提出的方法可以在三维大腿磁共振成像扫描中准确分割出单个肌肉群。
{"title":"Deep network and multi-atlas segmentation fusion for delineation of thigh muscle groups in three-dimensional water-fat separated MRI.","authors":"Nagasoujanya V Annasamudram, Azubuike M Okorie, Richard G Spencer, Rita R Kalyani, Qi Yang, Bennett A Landman, Luigi Ferrucci, Sokratis Makrogiannis","doi":"10.1117/1.JMI.11.5.054003","DOIUrl":"10.1117/1.JMI.11.5.054003","url":null,"abstract":"<p><strong>Purpose: </strong>Segmentation is essential for tissue quantification and characterization in studies of aging and age-related and metabolic diseases and the development of imaging biomarkers. We propose a multi-method and multi-atlas methodology for automated segmentation of functional muscle groups in three-dimensional (3D) thigh magnetic resonance images. These groups lie anatomically adjacent to each other, rendering their manual delineation a challenging and time-consuming task.</p><p><strong>Approach: </strong>We introduce a framework for automated segmentation of the four main functional muscle groups of the thigh, gracilis, hamstring, quadriceps femoris, and sartorius, using chemical shift encoded water-fat magnetic resonance imaging (CSE-MRI). We propose fusing anatomical mappings from multiple deformable models with 3D deep learning model-based segmentation. This approach leverages the generalizability of multi-atlas segmentation (MAS) and accuracy of deep networks, hence enabling accurate assessment of volume and fat content of muscle groups.</p><p><strong>Results: </strong>For segmentation performance evaluation, we calculated the Dice similarity coefficient (DSC) and Hausdorff distance 95th percentile (HD-95). We evaluated the proposed framework, its variants, and baseline methods on 15 healthy subjects by threefold cross-validation and tested on four patients. Fusion of multiple atlases, deformable registration models, and deep learning segmentation produced the top performance with an average DSC of 0.859 and HD-95 of 8.34 over all muscles.</p><p><strong>Conclusions: </strong>Fusion of multiple anatomical mappings from multiple MAS techniques enriches the template set and improves the segmentation accuracy. Additional fusion with deep network decisions applied to the subject space offers complementary information. The proposed approach can produce accurate segmentation of individual muscle groups in 3D thigh MRI scans.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"11 5","pages":"054003"},"PeriodicalIF":1.9,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11369361/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142134214","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Radiomics and quantitative multi-parametric MRI for predicting uterine fibroid growth. 用于预测子宫肌瘤生长的放射组学和定量多参数磁共振成像。
IF 1.9 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2024-09-01 Epub Date: 2024-09-12 DOI: 10.1117/1.JMI.11.5.054501
Karen Drukker, Milica Medved, Carla B Harmath, Maryellen L Giger, Obianuju S Madueke-Laveaux

Significance: Uterine fibroids (UFs) can pose a serious health risk to women. UFs are benign tumors that vary in clinical presentation from asymptomatic to causing debilitating symptoms. UF management is limited by our inability to predict UF growth rate and future morbidity.

Aim: We aim to develop a predictive model to identify UFs with increased growth rates and possible resultant morbidity.

Approach: We retrospectively analyzed 44 expertly outlined UFs from 20 patients who underwent two multi-parametric MR imaging exams as part of a prospective study over an average of 16 months. We identified 44 initial features by extracting quantitative magnetic resonance imaging (MRI) features plus morphological and textural radiomics features from DCE, T2, and apparent diffusion coefficient sequences. Principal component analysis reduced dimensionality, with the smallest number of components explaining over 97.5% of the variance selected. Employing a leave-one-fibroid-out scheme, a linear discriminant analysis classifier utilized these components to output a growth risk score.

Results: The classifier incorporated the first three principal components and achieved an area under the receiver operating characteristic curve of 0.80 (95% confidence interval [0.69; 0.91]), effectively distinguishing UFs growing faster than the median growth rate of 0.93    cm 3 / year / fibroid from slower-growing ones within the cohort. Time-to-event analysis, dividing the cohort based on the median growth risk score, yielded a hazard ratio of 0.33 [0.15; 0.76], demonstrating potential clinical utility.

Conclusion: We developed a promising predictive model utilizing quantitative MRI features and principal component analysis to identify UFs with increased growth rates. Furthermore, the model's discrimination ability supports its potential clinical utility in developing tailored patient and fibroid-specific management once validated on a larger cohort.

意义重大:子宫肌瘤(UFs)会对女性健康造成严重危害。子宫肌瘤是一种良性肿瘤,其临床表现各不相同,有的无症状,有的会导致衰弱症状。我们无法预测 UF 的生长率和未来的发病率,这限制了 UF 的治疗。目的:我们旨在开发一种预测模型,以识别生长率增高并可能导致发病率增高的 UF:我们回顾性分析了 20 名患者的 44 个专家概述 UF,这些患者在平均 16 个月的时间内接受了两次多参数 MR 成像检查,这是前瞻性研究的一部分。我们从 DCE、T2 和表观扩散系数序列中提取了定量磁共振成像(MRI)特征以及形态和纹理放射组学特征,从而确定了 44 个初始特征。主成分分析降低了维度,所选最小数量的成分可解释97.5%以上的方差。线性判别分析分类器采用 "留一剔除 "方案,利用这些成分输出生长风险评分:分类器包含前三个主成分,接收者操作特征曲线下面积为 0.80(95% 置信区间 [0.69; 0.91]),能有效区分生长速度快于中位数 0.93 厘米 3 / 年/肌瘤的 UF 和队列中生长速度较慢的 UF。根据中位生长风险评分对队列进行时间到事件分析,得出的危险比为 0.33 [0.15; 0.76],显示了潜在的临床实用性:我们利用磁共振成像的定量特征和主成分分析建立了一个很有前景的预测模型,用于识别生长率增高的 UFs。此外,该模型的辨别能力支持其在更大范围内验证后,在制定针对患者和肌瘤的定制化管理方面的潜在临床实用性。
{"title":"Radiomics and quantitative multi-parametric MRI for predicting uterine fibroid growth.","authors":"Karen Drukker, Milica Medved, Carla B Harmath, Maryellen L Giger, Obianuju S Madueke-Laveaux","doi":"10.1117/1.JMI.11.5.054501","DOIUrl":"https://doi.org/10.1117/1.JMI.11.5.054501","url":null,"abstract":"<p><strong>Significance: </strong>Uterine fibroids (UFs) can pose a serious health risk to women. UFs are benign tumors that vary in clinical presentation from asymptomatic to causing debilitating symptoms. UF management is limited by our inability to predict UF growth rate and future morbidity.</p><p><strong>Aim: </strong>We aim to develop a predictive model to identify UFs with increased growth rates and possible resultant morbidity.</p><p><strong>Approach: </strong>We retrospectively analyzed 44 expertly outlined UFs from 20 patients who underwent two multi-parametric MR imaging exams as part of a prospective study over an average of 16 months. We identified 44 initial features by extracting quantitative magnetic resonance imaging (MRI) features plus morphological and textural radiomics features from DCE, T2, and apparent diffusion coefficient sequences. Principal component analysis reduced dimensionality, with the smallest number of components explaining over 97.5% of the variance selected. Employing a leave-one-fibroid-out scheme, a linear discriminant analysis classifier utilized these components to output a growth risk score.</p><p><strong>Results: </strong>The classifier incorporated the first three principal components and achieved an area under the receiver operating characteristic curve of 0.80 (95% confidence interval [0.69; 0.91]), effectively distinguishing UFs growing faster than the median growth rate of <math><mrow><mn>0.93</mn> <mtext>  </mtext> <msup><mrow><mi>cm</mi></mrow> <mrow><mn>3</mn></mrow> </msup> <mo>/</mo> <mi>year</mi> <mo>/</mo> <mi>fibroid</mi></mrow> </math> from slower-growing ones within the cohort. Time-to-event analysis, dividing the cohort based on the median growth risk score, yielded a hazard ratio of 0.33 [0.15; 0.76], demonstrating potential clinical utility.</p><p><strong>Conclusion: </strong>We developed a promising predictive model utilizing quantitative MRI features and principal component analysis to identify UFs with increased growth rates. Furthermore, the model's discrimination ability supports its potential clinical utility in developing tailored patient and fibroid-specific management once validated on a larger cohort.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"11 5","pages":"054501"},"PeriodicalIF":1.9,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11391479/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142298699","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Photon-counting computed tomography versus energy-integrating computed tomography for detection of small liver lesions: comparison using a virtual framework imaging. 光子计数计算机断层扫描与能量积分计算机断层扫描在检测肝脏小病变方面的比较:利用虚拟框架成像进行比较。
IF 1.9 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2024-09-01 Epub Date: 2024-10-17 DOI: 10.1117/1.JMI.11.5.053502
Nicholas Felice, Benjamin Wildman-Tobriner, William Paul Segars, Mustafa R Bashir, Daniele Marin, Ehsan Samei, Ehsan Abadi

Purpose: Photon-counting computed tomography (PCCT) has the potential to provide superior image quality to energy-integrating CT (EICT). We objectively compare PCCT to EICT for liver lesion detection.

Approach: Fifty anthropomorphic, computational phantoms with inserted liver lesions were generated. Contrast-enhanced scans of each phantom were simulated at the portal venous phase. The acquisitions were done using DukeSim, a validated CT simulation platform. Scans were simulated at two dose levels ( CTDI vol 1.5 to 6.0 mGy) modeling PCCT (NAEOTOM Alpha, Siemens, Erlangen, Germany) and EICT (SOMATOM Flash, Siemens). Images were reconstructed with varying levels of kernel sharpness (soft, medium, sharp). To provide a quantitative estimate of image quality, the modulation transfer function (MTF), frequency at 50% of the MTF ( f 50 ), noise magnitude, contrast-to-noise ratio (CNR, per lesion), and detectability index ( d ' , per lesion) were measured.

Results: Across all studied conditions, the best detection performance, measured by d ' , was for PCCT images with the highest dose level and softest kernel. With soft kernel reconstruction, PCCT demonstrated improved lesion CNR and d ' compared with EICT, with a mean increase in CNR of 35.0% ( p < 0.001 ) and 21% ( p < 0.001 ) and a mean increase in d ' of 41.0% ( p < 0.001 ) and 23.3% ( p = 0.007 ) for the 1.5 and 6.0 mGy acquisitions, respectively. The improvements were greatest for larger phantoms, low-contrast lesions, and low-dose scans.

Conclusions: PCCT demonstrated objective improvement in liver lesion detection and image quality metrics compared with EICT. These advances may lead to earlier and more accurate liver lesion detection, thus improving patient care.

目的:光子计数计算机断层扫描(PCCT)有望提供优于能量整合 CT(EICT)的图像质量。我们对 PCCT 和 EICT 在肝脏病变检测方面进行了客观比较:方法:生成 50 个拟人化的计算模型,并插入肝脏病变。在门静脉相位模拟每个模型的对比增强扫描。采集使用的是经过验证的 CT 模拟平台 DukeSim。模拟扫描采用两种剂量水平(CTDI vol 1.5 至 6.0 mGy),分别以 PCCT(NAEOTOM Alpha,西门子,德国埃尔兰根)和 EICT(SOMATOM Flash,西门子)为模型。图像以不同的内核锐利度(柔和、中等、锐利)进行重建。为了对图像质量进行定量评估,测量了调制传递函数(MTF)、MTF 50%时的频率(f 50)、噪声大小、对比度与噪声比(CNR,每个病变)和可探测性指数(d ' ,每个病变):在所有研究条件下,剂量水平最高、内核最软的 PCCT 图像的检测性能最佳(以 d ' 为衡量标准)。与 EICT 相比,采用软核重建的 PCCT 提高了病变的 CNR 和 d',1.5 和 6.0 mGy 采集的 CNR 平均分别提高了 35.0% ( p 0.001 ) 和 21% ( p 0.001 ) ,d'平均分别提高了 41.0% ( p 0.001 ) 和 23.3% ( p = 0.007)。较大的模型、低对比度病变和低剂量扫描的改善幅度最大:结论:与 EICT 相比,PCCT 在肝脏病变检测和图像质量指标方面都有客观的改善。这些进步可能会使肝脏病变检测更早、更准确,从而改善患者护理。
{"title":"Photon-counting computed tomography versus energy-integrating computed tomography for detection of small liver lesions: comparison using a virtual framework imaging.","authors":"Nicholas Felice, Benjamin Wildman-Tobriner, William Paul Segars, Mustafa R Bashir, Daniele Marin, Ehsan Samei, Ehsan Abadi","doi":"10.1117/1.JMI.11.5.053502","DOIUrl":"10.1117/1.JMI.11.5.053502","url":null,"abstract":"<p><strong>Purpose: </strong>Photon-counting computed tomography (PCCT) has the potential to provide superior image quality to energy-integrating CT (EICT). We objectively compare PCCT to EICT for liver lesion detection.</p><p><strong>Approach: </strong>Fifty anthropomorphic, computational phantoms with inserted liver lesions were generated. Contrast-enhanced scans of each phantom were simulated at the portal venous phase. The acquisitions were done using DukeSim, a validated CT simulation platform. Scans were simulated at two dose levels ( <math> <mrow> <msub><mrow><mi>CTDI</mi></mrow> <mrow><mi>vol</mi></mrow> </msub> </mrow> </math> 1.5 to 6.0 mGy) modeling PCCT (NAEOTOM Alpha, Siemens, Erlangen, Germany) and EICT (SOMATOM Flash, Siemens). Images were reconstructed with varying levels of kernel sharpness (soft, medium, sharp). To provide a quantitative estimate of image quality, the modulation transfer function (MTF), frequency at 50% of the MTF ( <math> <mrow><msub><mi>f</mi> <mn>50</mn></msub> </mrow> </math> ), noise magnitude, contrast-to-noise ratio (CNR, per lesion), and detectability index ( <math> <mrow> <msup><mrow><mi>d</mi></mrow> <mrow><mo>'</mo></mrow> </msup> </mrow> </math> , per lesion) were measured.</p><p><strong>Results: </strong>Across all studied conditions, the best detection performance, measured by <math> <mrow> <msup><mrow><mi>d</mi></mrow> <mrow><mo>'</mo></mrow> </msup> </mrow> </math> , was for PCCT images with the highest dose level and softest kernel. With soft kernel reconstruction, PCCT demonstrated improved lesion CNR and <math> <mrow> <msup><mrow><mi>d</mi></mrow> <mrow><mo>'</mo></mrow> </msup> </mrow> </math> compared with EICT, with a mean increase in CNR of 35.0% ( <math><mrow><mi>p</mi> <mo><</mo> <mn>0.001</mn></mrow> </math> ) and 21% ( <math><mrow><mi>p</mi> <mo><</mo> <mn>0.001</mn></mrow> </math> ) and a mean increase in <math> <mrow> <msup><mrow><mi>d</mi></mrow> <mrow><mo>'</mo></mrow> </msup> </mrow> </math> of 41.0% ( <math><mrow><mi>p</mi> <mo><</mo> <mn>0.001</mn></mrow> </math> ) and 23.3% ( <math><mrow><mi>p</mi> <mo>=</mo> <mn>0.007</mn></mrow> </math> ) for the 1.5 and 6.0 mGy acquisitions, respectively. The improvements were greatest for larger phantoms, low-contrast lesions, and low-dose scans.</p><p><strong>Conclusions: </strong>PCCT demonstrated objective improvement in liver lesion detection and image quality metrics compared with EICT. These advances may lead to earlier and more accurate liver lesion detection, thus improving patient care.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"11 5","pages":"053502"},"PeriodicalIF":1.9,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11486217/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142477766","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Medical Imaging
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1