首页 > 最新文献

Computerized Medical Imaging and Graphics最新文献

英文 中文
PatchCL-AE: Anomaly detection for medical images using patch-wise contrastive learning-based auto-encoder PatchCL-AE:利用基于对比学习的自动编码器对医学图像进行异常检测
IF 5.7 2区 医学 Q1 Medicine Pub Date : 2024-03-06 DOI: 10.1016/j.compmedimag.2024.102366
Shuai Lu , Weihang Zhang , Jia Guo , Hanruo Liu , Huiqi Li , Ningli Wang

Anomaly detection is an important yet challenging task in medical image analysis. Most anomaly detection methods are based on reconstruction, but the performance of reconstruction-based methods is limited due to over-reliance on pixel-level losses. To address the limitation, we propose a patch-wise contrastive learning-based auto-encoder for medical anomaly detection. The key contribution is the patch-wise contrastive learning loss that provides supervision on local semantics to enforce semantic consistency between corresponding input–output patches. Contrastive learning pulls corresponding patch pairs closer while pushing non-corresponding ones apart between input and output, enabling the model to learn local normal features better and improve discriminability on anomalous regions. Additionally, we design an anomaly score based on local semantic discrepancies to pinpoint abnormalities by comparing feature difference rather than pixel variations. Extensive experiments on three public datasets (i.e., brain MRI, retinal OCT, and chest X-ray) achieve state-of-the-art performance, with our method achieving over 99% AUC on retinal and brain images. Both the contrastive patch-wise supervision and patch-discrepancy score provide targeted advancements to overcome the weaknesses in existing approaches.

异常检测是医学图像分析中一项重要而又具有挑战性的任务。大多数异常检测方法都基于重构,但由于过度依赖像素级损失,基于重构的方法性能有限。为了解决这个问题,我们提出了一种用于医学异常检测的基于补丁对比学习的自动编码器。该方法的主要贡献在于补丁对比学习损失,它提供了对局部语义的监督,以加强相应输入输出补丁之间的语义一致性。对比学习将相应的补丁对拉近,同时将输入和输出之间不对应的补丁对拉开,从而使模型能够更好地学习局部正常特征,并提高对异常区域的判别能力。此外,我们还设计了基于局部语义差异的异常评分,通过比较特征差异而不是像素变化来精确定位异常。我们在三个公共数据集(即脑磁共振成像、视网膜 OCT 和胸部 X 光)上进行了广泛的实验,取得了最先进的性能,我们的方法在视网膜和脑图像上的 AUC 超过 99%。对比性斑块监督和斑块差异得分都为克服现有方法的弱点提供了有针对性的进步。
{"title":"PatchCL-AE: Anomaly detection for medical images using patch-wise contrastive learning-based auto-encoder","authors":"Shuai Lu ,&nbsp;Weihang Zhang ,&nbsp;Jia Guo ,&nbsp;Hanruo Liu ,&nbsp;Huiqi Li ,&nbsp;Ningli Wang","doi":"10.1016/j.compmedimag.2024.102366","DOIUrl":"https://doi.org/10.1016/j.compmedimag.2024.102366","url":null,"abstract":"<div><p>Anomaly detection is an important yet challenging task in medical image analysis. Most anomaly detection methods are based on reconstruction, but the performance of reconstruction-based methods is limited due to over-reliance on pixel-level losses. To address the limitation, we propose a patch-wise contrastive learning-based auto-encoder for medical anomaly detection. The key contribution is the patch-wise contrastive learning loss that provides supervision on local semantics to enforce semantic consistency between corresponding input–output patches. Contrastive learning pulls corresponding patch pairs closer while pushing non-corresponding ones apart between input and output, enabling the model to learn local normal features better and improve discriminability on anomalous regions. Additionally, we design an anomaly score based on local semantic discrepancies to pinpoint abnormalities by comparing feature difference rather than pixel variations. Extensive experiments on three public datasets (<em>i.e.</em>, brain MRI, retinal OCT, and chest X-ray) achieve state-of-the-art performance, with our method achieving over 99% AUC on retinal and brain images. Both the contrastive patch-wise supervision and patch-discrepancy score provide targeted advancements to overcome the weaknesses in existing approaches.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":null,"pages":null},"PeriodicalIF":5.7,"publicationDate":"2024-03-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140095986","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-task global optimization-based method for vascular landmark detection 基于多任务全局优化的血管地标检测方法
IF 5.7 2区 医学 Q1 Medicine Pub Date : 2024-03-01 DOI: 10.1016/j.compmedimag.2024.102364
Zimeng Tan , Jianjiang Feng , Wangsheng Lu , Yin Yin , Guangming Yang , Jie Zhou

Vascular landmark detection plays an important role in medical analysis and clinical treatment. However, due to the complex topology and similar local appearance around landmarks, the popular heatmap regression based methods always suffer from the landmark confusion problem. Vascular landmarks are connected by vascular segments and have special spatial correlations, which can be utilized for performance improvement. In this paper, we propose a multi-task global optimization-based framework for accurate and automatic vascular landmark detection. A multi-task deep learning network is exploited to accomplish landmark heatmap regression, vascular semantic segmentation, and orientation field regression simultaneously. The two auxiliary objectives are highly correlated with the heatmap regression task and help the network incorporate the structural prior knowledge. During inference, instead of performing a max-voting strategy, we propose a global optimization-based post-processing method for final landmark decision. The spatial relationships between neighboring landmarks are utilized explicitly to tackle the landmark confusion problem. We evaluated our method on a cerebral MRA dataset with 564 volumes, a cerebral CTA dataset with 510 volumes, and an aorta CTA dataset with 50 volumes. The experiments demonstrate that the proposed method is effective for vascular landmark localization and achieves state-of-the-art performance.

血管地标检测在医学分析和临床治疗中发挥着重要作用。然而,由于地标周围复杂的拓扑结构和相似的局部外观,目前流行的基于热图回归的方法总是存在地标混淆问题。血管地标由血管节段连接,具有特殊的空间相关性,可以利用这些相关性来提高性能。本文提出了一种基于多任务全局优化的框架,用于准确、自动地检测血管地标。利用多任务深度学习网络同时完成地标热图回归、血管语义分割和方位场回归。这两个辅助目标与热图回归任务高度相关,有助于网络纳入结构先验知识。在推理过程中,我们提出了一种基于全局优化的后处理方法来进行最终的地标决策,而不是执行最大投票策略。我们明确利用相邻地标之间的空间关系来解决地标混淆问题。我们在一个包含 564 个体量的脑 MRA 数据集、一个包含 510 个体量的脑 CTA 数据集和一个包含 50 个体量的主动脉 CTA 数据集上评估了我们的方法。实验证明,所提出的方法对血管地标定位非常有效,并达到了最先进的性能。
{"title":"Multi-task global optimization-based method for vascular landmark detection","authors":"Zimeng Tan ,&nbsp;Jianjiang Feng ,&nbsp;Wangsheng Lu ,&nbsp;Yin Yin ,&nbsp;Guangming Yang ,&nbsp;Jie Zhou","doi":"10.1016/j.compmedimag.2024.102364","DOIUrl":"https://doi.org/10.1016/j.compmedimag.2024.102364","url":null,"abstract":"<div><p>Vascular landmark detection plays an important role in medical analysis and clinical treatment. However, due to the complex topology and similar local appearance around landmarks, the popular heatmap regression based methods always suffer from the landmark confusion problem. Vascular landmarks are connected by vascular segments and have special spatial correlations, which can be utilized for performance improvement. In this paper, we propose a multi-task global optimization-based framework for accurate and automatic vascular landmark detection. A multi-task deep learning network is exploited to accomplish landmark heatmap regression, vascular semantic segmentation, and orientation field regression simultaneously. The two auxiliary objectives are highly correlated with the heatmap regression task and help the network incorporate the structural prior knowledge. During inference, instead of performing a max-voting strategy, we propose a global optimization-based post-processing method for final landmark decision. The spatial relationships between neighboring landmarks are utilized explicitly to tackle the landmark confusion problem. We evaluated our method on a cerebral MRA dataset with 564 volumes, a cerebral CTA dataset with 510 volumes, and an aorta CTA dataset with 50 volumes. The experiments demonstrate that the proposed method is effective for vascular landmark localization and achieves state-of-the-art performance.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":null,"pages":null},"PeriodicalIF":5.7,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140016089","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Detection of abdominopelvic lymph nodes in multi-parametric MRI 在多参数磁共振成像中检测腹盆腔淋巴结
IF 5.7 2区 医学 Q1 Medicine Pub Date : 2024-03-01 DOI: 10.1016/j.compmedimag.2024.102363
Tejas Sudharshan Mathai , Thomas C. Shen , Daniel C. Elton , Sungwon Lee , Zhiyong Lu , Ronald M. Summers

Reliable localization of lymph nodes (LNs) in multi-parametric MRI (mpMRI) studies plays a major role in the assessment of lymphadenopathy and staging of metastatic disease. Radiologists routinely measure the nodal size in order to distinguish benign from malignant nodes, which require subsequent cancer staging. However, identification of lymph nodes is a cumbersome task due to their myriad appearances in mpMRI studies. Multiple sequences are acquired in mpMRI studies, including T2 fat suppressed (T2FS) and diffusion weighted imaging (DWI) sequences among others; consequently, the sizing of LNs is rendered challenging due to the variety of signal intensities in these sequences. Furthermore, radiologists can miss potentially metastatic LNs during a busy clinical day. To lighten these imaging and workflow challenges, we propose a computer-aided detection (CAD) pipeline to detect both benign and malignant LNs in the body for their subsequent measurement. We employed the recently proposed Dynamic Head (DyHead) neural network to detect LNs in mpMRI studies that were acquired using a variety of scanners and exam protocols. The T2FS and DWI series were co-registered, and a selective augmentation technique called Intra-Label LISA (ILL) was used to blend the two volumes with the interpolation factor drawn from a Beta distribution. In this way, ILL diversified the samples that the model encountered during the training phase, while the requirement for both sequences to be present at test time was nullified. Our results showed a mean average precision (mAP) of 53.5% and a sensitivity of 78% with ILL at 4 FP/vol. This corresponded to an improvement of 10% in mAP and 12% in sensitivity at 4FP (p ¡ 0.05) respectively over current LN detection approaches evaluated on the same dataset. We also established the out-of-distribution robustness of the DyHead model by training it on data acquired by a Siemens Aera scanner and testing it on data from the Siemens Verio, Siemens Biograph mMR, and Philips Achieva scanners. Our pilot work represents an important first step towards automated detection, segmentation, and classification of lymph nodes in mpMRI.

多参数磁共振成像(mpMRI)研究中淋巴结(LN)的可靠定位在淋巴结病的评估和转移性疾病的分期中起着重要作用。放射科医生通常会测量淋巴结的大小,以区分良性和恶性淋巴结,这需要随后进行癌症分期。然而,由于淋巴结在 mpMRI 研究中的表现多种多样,淋巴结的识别是一项繁琐的工作。在 mpMRI 研究中会采集多个序列,包括 T2 脂肪抑制(T2FS)和弥散加权成像(DWI)序列等;因此,由于这些序列中信号强度的多样性,确定淋巴结的大小就变得非常具有挑战性。此外,放射科医生可能会在繁忙的临床工作中错过潜在的转移性 LN。为了减轻这些成像和工作流程方面的挑战,我们提出了一种计算机辅助检测(CAD)管道,用于检测体内良性和恶性 LN,以便进行后续测量。我们采用了最近提出的动态头部(DyHead)神经网络来检测使用各种扫描仪和检查方案获得的 mpMRI 研究中的 LN。对 T2FS 和 DWI 序列进行共同注册,并使用一种称为标签内 LISA(ILL)的选择性增强技术,利用从 Beta 分布中提取的插值因子混合两个体量。通过这种方式,ILL 使模型在训练阶段遇到的样本多样化,同时取消了测试时两个序列同时存在的要求。结果表明,在 4 FP/vol 的条件下,ILL 的平均精确度(mAP)为 53.5%,灵敏度为 78%。与目前在相同数据集上评估的 LN 检测方法相比,4FP 的 mAP 和灵敏度分别提高了≥10% 和≥12%(p ¡ 0.05)。我们还在西门子Aera扫描仪获取的数据上对DyHead模型进行了训练,并在西门子Verio、西门子Biograph mMR和飞利浦Achieva扫描仪的数据上进行了测试,从而确定了DyHead模型在分布外的鲁棒性。我们的试验性工作代表着向 mpMRI 中淋巴结的自动检测、分割和分类迈出了重要的第一步。
{"title":"Detection of abdominopelvic lymph nodes in multi-parametric MRI","authors":"Tejas Sudharshan Mathai ,&nbsp;Thomas C. Shen ,&nbsp;Daniel C. Elton ,&nbsp;Sungwon Lee ,&nbsp;Zhiyong Lu ,&nbsp;Ronald M. Summers","doi":"10.1016/j.compmedimag.2024.102363","DOIUrl":"https://doi.org/10.1016/j.compmedimag.2024.102363","url":null,"abstract":"<div><p>Reliable localization of lymph nodes (LNs) in multi-parametric MRI (mpMRI) studies plays a major role in the assessment of lymphadenopathy and staging of metastatic disease. Radiologists routinely measure the nodal size in order to distinguish benign from malignant nodes, which require subsequent cancer staging. However, identification of lymph nodes is a cumbersome task due to their myriad appearances in mpMRI studies. Multiple sequences are acquired in mpMRI studies, including T2 fat suppressed (T2FS) and diffusion weighted imaging (DWI) sequences among others; consequently, the sizing of LNs is rendered challenging due to the variety of signal intensities in these sequences. Furthermore, radiologists can miss potentially metastatic LNs during a busy clinical day. To lighten these imaging and workflow challenges, we propose a computer-aided detection (CAD) pipeline to detect both benign and malignant LNs in the body for their subsequent measurement. We employed the recently proposed Dynamic Head (DyHead) neural network to detect LNs in mpMRI studies that were acquired using a variety of scanners and exam protocols. The T2FS and DWI series were co-registered, and a selective augmentation technique called Intra-Label LISA (ILL) was used to blend the two volumes with the interpolation factor drawn from a Beta distribution. In this way, ILL diversified the samples that the model encountered during the training phase, while the requirement for both sequences to be present at test time was nullified. Our results showed a mean average precision (mAP) of 53.5% and a sensitivity of <span><math><mo>∼</mo></math></span>78% with ILL at 4 FP/vol. This corresponded to an improvement of <span><math><mo>≥</mo></math></span>10% in mAP and <span><math><mo>≥</mo></math></span>12% in sensitivity at 4FP (<em>p</em> ¡ 0.05) respectively over current LN detection approaches evaluated on the same dataset. We also established the out-of-distribution robustness of the DyHead model by training it on data acquired by a Siemens Aera scanner and testing it on data from the Siemens Verio, Siemens Biograph mMR, and Philips Achieva scanners. Our pilot work represents an important first step towards automated detection, segmentation, and classification of lymph nodes in mpMRI.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":null,"pages":null},"PeriodicalIF":5.7,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140030312","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automatic artery/vein classification methods for retinal blood vessel: A review 视网膜血管的自动动脉/静脉分类方法:综述
IF 5.7 2区 医学 Q1 Medicine Pub Date : 2024-02-16 DOI: 10.1016/j.compmedimag.2024.102355
Qihan Chen , Jianqing Peng , Shen Zhao , Wanquan Liu

Automatic retinal arteriovenous classification can assist ophthalmologists in disease early diagnosis. Deep learning-based methods and topological graph-based methods have become the main solutions for retinal arteriovenous classification in recent years. This paper reviews the automatic retinal arteriovenous classification methods from 2003 to 2022. Firstly, we compare different methods and provide comparison tables of the summary results. Secondly, we complete the classification of the public arteriovenous classification datasets and provide the annotation development tables of different datasets. Finally, we sort out the challenges of evaluation methods and provide a comprehensive evaluation system. Quantitative and qualitative analysis shows the changes in research hotspots over time, Quantitative and qualitative analyses reveal the evolution of research hotspots over time, highlighting the significance of exploring the integration of deep learning with topological information in future research.

视网膜动静脉自动分类可以帮助眼科医生进行疾病的早期诊断。近年来,基于深度学习的方法和基于拓扑图的方法已成为视网膜动静脉分类的主要解决方案。本文回顾了 2003 年至 2022 年的视网膜动静脉自动分类方法。首先,我们比较了不同的方法,并提供了汇总结果对比表。其次,我们完成了公共动静脉分类数据集的分类,并提供了不同数据集的注释开发表。最后,我们梳理了评价方法所面临的挑战,并提供了一个全面的评价体系。定量和定性分析显示了研究热点随时间的变化,定量和定性分析揭示了研究热点随时间的演变,凸显了在未来研究中探索深度学习与拓扑信息整合的意义。
{"title":"Automatic artery/vein classification methods for retinal blood vessel: A review","authors":"Qihan Chen ,&nbsp;Jianqing Peng ,&nbsp;Shen Zhao ,&nbsp;Wanquan Liu","doi":"10.1016/j.compmedimag.2024.102355","DOIUrl":"https://doi.org/10.1016/j.compmedimag.2024.102355","url":null,"abstract":"<div><p>Automatic retinal arteriovenous classification can assist ophthalmologists in disease early diagnosis. Deep learning-based methods and topological graph-based methods have become the main solutions for retinal arteriovenous classification in recent years. This paper reviews the automatic retinal arteriovenous classification methods from 2003 to 2022. Firstly, we compare different methods and provide comparison tables of the summary results. Secondly, we complete the classification of the public arteriovenous classification datasets and provide the annotation development tables of different datasets. Finally, we sort out the challenges of evaluation methods and provide a comprehensive evaluation system. Quantitative and qualitative analysis shows the changes in research hotspots over time, Quantitative and qualitative analyses reveal the evolution of research hotspots over time, highlighting the significance of exploring the integration of deep learning with topological information in future research.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":null,"pages":null},"PeriodicalIF":5.7,"publicationDate":"2024-02-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139907922","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SCANED: Siamese collateral assessment network for evaluation of collaterals from ischemic damage SCANED:用于评估缺血性损伤侧支的暹罗侧支评估网络
IF 5.7 2区 医学 Q1 Medicine Pub Date : 2024-02-15 DOI: 10.1016/j.compmedimag.2024.102346
Mumu Aktar , Yiming Xiao , Ali K.Z. Tehrani , Donatella Tampieri , Hassan Rivaz , Marta Kersten-Oertel

This study conducts collateral evaluation from ischemic damage using a deep learning-based Siamese network, addressing the challenges associated with a small and imbalanced dataset. The collateral network provides an alternative oxygen and nutrient supply pathway in ischemic stroke cases, influencing treatment decisions. Research in this area focuses on automated collateral assessment using deep learning (DL) methods to expedite decision-making processes and enhance accuracy. Our study employed a 3D ResNet-based Siamese network, referred to as SCANED, to classify collaterals as good/intermediate or poor. Utilizing non-contrast computed tomography (NCCT) images, the network automates collateral identification and assessment by analyzing tissue degeneration around the ischemic site. Relevant features from the left/right hemispheres were extracted, and Euclidean Distance (ED) was employed for similarity measurement. Finally, dichotomized classification of good/intermediate or poor collateral is performed by SCANED using an optimal threshold derived from ROC analysis. SCANED provides a sensitivity of 0.88, a specificity of 0.63, and a weighted F1 score of 0.86 in the dichotomized classification.

这项研究利用基于深度学习的连体网络对缺血性损伤的侧支进行评估,解决了小型不平衡数据集带来的挑战。侧支网络为缺血性中风病例提供了另一条氧气和营养供应途径,从而影响治疗决策。该领域的研究重点是使用深度学习(DL)方法进行自动侧支评估,以加快决策过程并提高准确性。我们的研究采用了基于三维 ResNet 的连体网络(称为 SCANED),将侧支分为良好/中等或较差。该网络利用非对比计算机断层扫描(NCCT)图像,通过分析缺血部位周围的组织变性,自动进行侧支识别和评估。提取左/右半球的相关特征,并采用欧氏距离(ED)进行相似性测量。最后,SCANED 利用 ROC 分析得出的最佳阈值对良好/中等或不良侧支进行二分分类。在二分法分类中,SCANED 的灵敏度为 0.88,特异度为 0.63,加权 F1 得分为 0.86。
{"title":"SCANED: Siamese collateral assessment network for evaluation of collaterals from ischemic damage","authors":"Mumu Aktar ,&nbsp;Yiming Xiao ,&nbsp;Ali K.Z. Tehrani ,&nbsp;Donatella Tampieri ,&nbsp;Hassan Rivaz ,&nbsp;Marta Kersten-Oertel","doi":"10.1016/j.compmedimag.2024.102346","DOIUrl":"https://doi.org/10.1016/j.compmedimag.2024.102346","url":null,"abstract":"<div><p>This study conducts collateral evaluation from ischemic damage using a deep learning-based Siamese network, addressing the challenges associated with a small and imbalanced dataset. The collateral network provides an alternative oxygen and nutrient supply pathway in ischemic stroke cases, influencing treatment decisions. Research in this area focuses on automated collateral assessment using deep learning (DL) methods to expedite decision-making processes and enhance accuracy. Our study employed a 3D ResNet-based Siamese network, referred to as SCANED, to classify collaterals as good/intermediate or poor. Utilizing non-contrast computed tomography (NCCT) images, the network automates collateral identification and assessment by analyzing tissue degeneration around the ischemic site. Relevant features from the left/right hemispheres were extracted, and Euclidean Distance (ED) was employed for similarity measurement. Finally, dichotomized classification of good/intermediate or poor collateral is performed by SCANED using an optimal threshold derived from ROC analysis. SCANED provides a sensitivity of 0.88, a specificity of 0.63, and a weighted F1 score of 0.86 in the dichotomized classification.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":null,"pages":null},"PeriodicalIF":5.7,"publicationDate":"2024-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0895611124000235/pdfft?md5=6b0fbf08d111cd87a2314d489eeb3849&pid=1-s2.0-S0895611124000235-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139738796","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SC-GAN: Structure-completion generative adversarial network for synthetic CT generation from MR images with truncated anatomy SC-GAN:用于从截断解剖结构的 MR 图像生成合成 CT 的结构补全生成对抗网络
IF 5.7 2区 医学 Q1 Medicine Pub Date : 2024-02-10 DOI: 10.1016/j.compmedimag.2024.102353
Xinru Chen , Yao Zhao , Laurence E. Court , He Wang , Tinsu Pan , Jack Phan , Xin Wang , Yao Ding , Jinzhong Yang

Creating synthetic CT (sCT) from magnetic resonance (MR) images enables MR-based treatment planning in radiation therapy. However, the MR images used for MR-guided adaptive planning are often truncated in the boundary regions due to the limited field of view and the need for sequence optimization. Consequently, the sCT generated from these truncated MR images lacks complete anatomic information, leading to dose calculation error for MR-based adaptive planning. We propose a novel structure-completion generative adversarial network (SC-GAN) to generate sCT with full anatomic details from the truncated MR images. To enable anatomy compensation, we expand input channels of the CT generator by including a body mask and introduce a truncation loss between sCT and real CT. The body mask for each patient was automatically created from the simulation CT scans and transformed to daily MR images by rigid registration as another input for our SC-GAN in addition to the MR images. The truncation loss was constructed by implementing either an auto-segmentor or an edge detector to penalize the difference in body outlines between sCT and real CT. The experimental results show that our SC-GAN achieved much improved accuracy of sCT generation in both truncated and untruncated regions compared to the original cycleGAN and conditional GAN methods.

根据磁共振(MR)图像创建合成 CT(sCT)可实现基于 MR 的放射治疗计划。然而,由于视野有限和序列优化的需要,用于磁共振引导自适应规划的磁共振图像通常会在边界区域被截断。因此,由这些截断的磁共振图像生成的 sCT 缺乏完整的解剖信息,导致基于磁共振的自适应计划的剂量计算错误。我们提出了一种新颖的结构补全生成对抗网络(SC-GAN),可从截断的磁共振图像中生成具有完整解剖细节的 sCT。为了实现解剖补偿,我们通过加入人体遮罩来扩展 CT 生成器的输入通道,并在 sCT 和真实 CT 之间引入截断损失。每个患者的身体掩膜都是根据模拟 CT 扫描自动创建的,并通过刚性配准转换为日常 MR 图像,作为除 MR 图像外 SC-GAN 的另一个输入。截断损失是通过实施自动分割器或边缘检测器来构建的,以惩罚 sCT 和真实 CT 之间身体轮廓的差异。实验结果表明,与原始 cycleGAN 和条件 GAN 方法相比,我们的 SC-GAN 在截断和未截断区域生成 sCT 的准确性都有很大提高。
{"title":"SC-GAN: Structure-completion generative adversarial network for synthetic CT generation from MR images with truncated anatomy","authors":"Xinru Chen ,&nbsp;Yao Zhao ,&nbsp;Laurence E. Court ,&nbsp;He Wang ,&nbsp;Tinsu Pan ,&nbsp;Jack Phan ,&nbsp;Xin Wang ,&nbsp;Yao Ding ,&nbsp;Jinzhong Yang","doi":"10.1016/j.compmedimag.2024.102353","DOIUrl":"https://doi.org/10.1016/j.compmedimag.2024.102353","url":null,"abstract":"<div><p>Creating synthetic CT (sCT) from magnetic resonance (MR) images enables MR-based treatment planning in radiation therapy. However, the MR images used for MR-guided adaptive planning are often truncated in the boundary regions due to the limited field of view and the need for sequence optimization. Consequently, the sCT generated from these truncated MR images lacks complete anatomic information, leading to dose calculation error for MR-based adaptive planning. We propose a novel structure-completion generative adversarial network (SC-GAN) to generate sCT with full anatomic details from the truncated MR images. To enable anatomy compensation, we expand input channels of the CT generator by including a body mask and introduce a truncation loss between sCT and real CT. The body mask for each patient was automatically created from the simulation CT scans and transformed to daily MR images by rigid registration as another input for our SC-GAN in addition to the MR images. The truncation loss was constructed by implementing either an auto-segmentor or an edge detector to penalize the difference in body outlines between sCT and real CT. The experimental results show that our SC-GAN achieved much improved accuracy of sCT generation in both truncated and untruncated regions compared to the original cycleGAN and conditional GAN methods.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":null,"pages":null},"PeriodicalIF":5.7,"publicationDate":"2024-02-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139915280","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improving abdominal image segmentation with overcomplete shape priors 利用过度完整形状先验改进腹部图像分割
IF 5.7 2区 医学 Q1 Medicine Pub Date : 2024-02-09 DOI: 10.1016/j.compmedimag.2024.102356
Amine Sadikine , Bogdan Badic , Jean-Pierre Tasu , Vincent Noblet , Pascal Ballet , Dimitris Visvikis , Pierre-Henri Conze

The extraction of abdominal structures using deep learning has recently experienced a widespread interest in medical image analysis. Automatic abdominal organ and vessel segmentation is highly desirable to guide clinicians in computer-assisted diagnosis, therapy, or surgical planning. Despite a good ability to extract large organs, the capacity of U-Net inspired architectures to automatically delineate smaller structures remains a major issue, especially given the increase in receptive field size as we go deeper into the network. To deal with various abdominal structure sizes while exploiting efficient geometric constraints, we present a novel approach that integrates into deep segmentation shape priors from a semi-overcomplete convolutional auto-encoder (S-OCAE) embedding. Compared to standard convolutional auto-encoders (CAE), it exploits an over-complete branch that projects data onto higher dimensions to better characterize anatomical structures with a small spatial extent. Experiments on abdominal organs and vessel delineation performed on various publicly available datasets highlight the effectiveness of our method compared to state-of-the-art, including U-Net trained without and with shape priors from a traditional CAE. Exploiting a semi-overcomplete convolutional auto-encoder embedding as shape priors improves the ability of deep segmentation models to provide realistic and accurate abdominal structure contours.

利用深度学习提取腹部结构最近在医学图像分析领域受到广泛关注。自动腹部器官和血管分割非常适合指导临床医生进行计算机辅助诊断、治疗或手术规划。尽管 U-Net 架构能够很好地提取大型器官,但其自动划分较小结构的能力仍是一个主要问题,特别是考虑到随着网络的深入,感受野的大小也会增加。为了在利用高效几何约束的同时处理各种腹部结构尺寸,我们提出了一种新方法,将来自半超完全卷积自动编码器(S-OCAE)嵌入的形状先验整合到深度分割中。与标准卷积自动编码器(CAE)相比,它利用了一个超完全分支,将数据投射到更高的维度上,从而更好地描述空间范围较小的解剖结构。在各种公开数据集上进行的腹部器官和血管划分实验表明,与最先进的方法(包括不使用和使用传统 CAE 的形状先验训练的 U-Net)相比,我们的方法非常有效。利用半不完全卷积自动编码器嵌入作为形状先验,提高了深度分割模型提供真实准确的腹部结构轮廓的能力。
{"title":"Improving abdominal image segmentation with overcomplete shape priors","authors":"Amine Sadikine ,&nbsp;Bogdan Badic ,&nbsp;Jean-Pierre Tasu ,&nbsp;Vincent Noblet ,&nbsp;Pascal Ballet ,&nbsp;Dimitris Visvikis ,&nbsp;Pierre-Henri Conze","doi":"10.1016/j.compmedimag.2024.102356","DOIUrl":"https://doi.org/10.1016/j.compmedimag.2024.102356","url":null,"abstract":"<div><p>The extraction of abdominal structures using deep learning has recently experienced a widespread interest in medical image analysis. Automatic abdominal organ and vessel segmentation is highly desirable to guide clinicians in computer-assisted diagnosis, therapy, or surgical planning. Despite a good ability to extract large organs, the capacity of U-Net inspired architectures to automatically delineate smaller structures remains a major issue, especially given the increase in receptive field size as we go deeper into the network. To deal with various abdominal structure sizes while exploiting efficient geometric constraints, we present a novel approach that integrates into deep segmentation shape priors from a semi-overcomplete convolutional auto-encoder (S-OCAE) embedding. Compared to standard convolutional auto-encoders (CAE), it exploits an over-complete branch that projects data onto higher dimensions to better characterize anatomical structures with a small spatial extent. Experiments on abdominal organs and vessel delineation performed on various publicly available datasets highlight the effectiveness of our method compared to state-of-the-art, including U-Net trained without and with shape priors from a traditional CAE. Exploiting a semi-overcomplete convolutional auto-encoder embedding as shape priors improves the ability of deep segmentation models to provide realistic and accurate abdominal structure contours.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":null,"pages":null},"PeriodicalIF":5.7,"publicationDate":"2024-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0895611124000338/pdfft?md5=e83e8e8de56fb1f5b580a68d5fed492b&pid=1-s2.0-S0895611124000338-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139714471","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A transformer-based pyramid network for coronary calcified plaque segmentation in intravascular optical coherence tomography images 基于变压器的金字塔网络,用于血管内光学相干断层扫描图像中冠状动脉钙化斑块的分割。
IF 5.7 2区 医学 Q1 Medicine Pub Date : 2024-02-09 DOI: 10.1016/j.compmedimag.2024.102347
Yiqing Liu , Farhad R. Nezami , Elazer R. Edelman

Characterizing coronary calcified plaque (CCP) provides essential insight into diagnosis and treatment of atherosclerosis. Intravascular optical coherence tomography (OCT) offers significant advantages for detecting CCP and even automated segmentation with recent advances in deep learning techniques. Most of current methods have achieved promising results by adopting existing convolution neural networks (CNNs) in computer vision domain. However, their performance can be detrimentally affected by unseen plaque patterns and artifacts due to inherent limitation of CNNs in contextual reasoning. To overcome this obstacle, we proposed a Transformer-based pyramid network called AFS-TPNet for robust, end-to-end segmentation of CCP from OCT images. Its encoder is built upon CSWin Transformer architecture, allowing for better perceptual understanding of calcified arteries at a higher semantic level. Specifically, an augmented feature split (AFS) module and residual convolutional position encoding (RCPE) mechanism are designed to effectively enhance the capability of Transformer in capturing both fine-grained features and global contexts. Extensive experiments showed that AFS-TPNet trained using Lovasz Loss achieved superior performance in segmentation CCP under various contexts, surpassing prior state-of-the-art CNN and Transformer architectures by more than 6.58% intersection over union (IoU) score. The application of this promising method to extract CCP features is expected to enhance clinical intervention and translational research using OCT.

冠状动脉钙化斑块(CCP)的特征为动脉粥样硬化的诊断和治疗提供了重要依据。血管内光学相干断层扫描(OCT)在检测冠状动脉钙化斑块(CCP)方面具有显著优势,随着深度学习技术的最新进展,甚至可以实现自动分割。目前的大多数方法都采用了计算机视觉领域现有的卷积神经网络(CNN),取得了可喜的成果。然而,由于卷积神经网络在上下文推理方面的固有局限性,它们的性能可能会受到未见斑块模式和伪影的不利影响。为了克服这一障碍,我们提出了一种名为 AFS-TPNet 的基于变换器的金字塔网络,用于从 OCT 图像中对 CCP 进行稳健的端到端分割。它的编码器建立在 CSWin Transformer 架构之上,可以在更高的语义层面上更好地感知钙化动脉。具体来说,设计了增强特征分割(AFS)模块和残差卷积位置编码(RCPE)机制,以有效增强 Transformer 在捕捉细粒度特征和全局上下文方面的能力。广泛的实验表明,使用 Lovasz Loss 训练的 AFS-TPNet 在各种上下文条件下分割 CCP 时都取得了优异的性能,超过了之前最先进的 CNN 和 Transformer 架构 6.58% 以上的 intersection over union (IoU) 分数。应用这种前景广阔的方法来提取 CCP 特征,有望利用 OCT 加强临床干预和转化研究。
{"title":"A transformer-based pyramid network for coronary calcified plaque segmentation in intravascular optical coherence tomography images","authors":"Yiqing Liu ,&nbsp;Farhad R. Nezami ,&nbsp;Elazer R. Edelman","doi":"10.1016/j.compmedimag.2024.102347","DOIUrl":"10.1016/j.compmedimag.2024.102347","url":null,"abstract":"<div><p>Characterizing coronary calcified plaque (CCP) provides essential insight into diagnosis and treatment of atherosclerosis. Intravascular optical coherence tomography (OCT) offers significant advantages for detecting CCP and even automated segmentation with recent advances in deep learning techniques. Most of current methods have achieved promising results by adopting existing convolution neural networks (CNNs) in computer vision domain. However, their performance can be detrimentally affected by unseen plaque patterns and artifacts due to inherent limitation of CNNs in contextual reasoning. To overcome this obstacle, we proposed a Transformer-based pyramid network called AFS-TPNet for robust, end-to-end segmentation of CCP from OCT images. Its encoder is built upon CSWin Transformer architecture, allowing for better perceptual understanding of calcified arteries at a higher semantic level. Specifically, an augmented feature split (AFS) module and residual convolutional position encoding (RCPE) mechanism are designed to effectively enhance the capability of Transformer in capturing both fine-grained features and global contexts. Extensive experiments showed that AFS-TPNet trained using Lovasz Loss achieved superior performance in segmentation CCP under various contexts, surpassing prior state-of-the-art CNN and Transformer architectures by more than 6.58% intersection over union (IoU) score. The application of this promising method to extract CCP features is expected to enhance clinical intervention and translational research using OCT.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":null,"pages":null},"PeriodicalIF":5.7,"publicationDate":"2024-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139718035","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Knowledge distillation on individual vertebrae segmentation exploiting 3D U-Net 利用 3D U-Net 对单个椎骨分割进行知识提炼
IF 5.7 2区 医学 Q1 Medicine Pub Date : 2024-02-08 DOI: 10.1016/j.compmedimag.2024.102350
Luís Serrador , Francesca Pia Villani , Sara Moccia , Cristina P. Santos

Recent advances in medical imaging have highlighted the critical development of algorithms for individual vertebral segmentation on computed tomography (CT) scans. Essential for diagnostic accuracy and treatment planning in orthopaedics, neurosurgery and oncology, these algorithms face challenges in clinical implementation, including integration into healthcare systems. Consequently, our focus lies in exploring the application of knowledge distillation (KD) methods to train shallower networks capable of efficiently segmenting vertebrae in CT scans. This approach aims to reduce segmentation time, enhance suitability for emergency cases, and optimize computational and memory resource efficiency. Building upon prior research in the field, a two-step segmentation approach was employed. Firstly, the spine’s location was determined by predicting a heatmap, indicating the probability of each voxel belonging to the spine. Subsequently, an iterative segmentation of vertebrae was performed from the top to the bottom of the CT volume over the located spine, using a memory instance to record the already segmented vertebrae. KD methods were implemented by training a teacher network with performance similar to that found in the literature, and this knowledge was distilled to a shallower network (student). Two KD methods were applied: (1) using the soft outputs of both networks and (2) matching logits. Two publicly available datasets, comprising 319 CT scans from 300 patients and a total of 611 cervical, 2387 thoracic, and 1507 lumbar vertebrae, were used. To ensure dataset balance and robustness, effective data augmentation methods were applied, including cleaning the memory instance to replicate the first vertebra segmentation. The teacher network achieved an average Dice similarity coefficient (DSC) of 88.22% and a Hausdorff distance (HD) of 7.71 mm, showcasing performance similar to other approaches in the literature. Through knowledge distillation from the teacher network, the student network’s performance improved, with an average DSC increasing from 75.78% to 84.70% and an HD decreasing from 15.17 mm to 8.08 mm. Compared to other methods, our teacher network exhibited up to 99.09% fewer parameters, 90.02% faster inference time, 88.46% shorter total segmentation time, and 89.36% less associated carbon (CO2) emission rate. Regarding our student network, it featured 75.00% fewer parameters than our teacher, resulting in a 36.15% reduction in inference time, a 33.33% decrease in total segmentation time, and a 42.96% reduction in CO2 emissions. This study marks the first exploration of applying KD to the problem of individual vertebrae segmentation in CT, demonstrating the feasibility of achieving comparable performance to existing methods using smaller neural networks.

医学成像技术的最新进展突显了计算机断层扫描(CT)上单个脊椎分割算法的重要发展。这些算法对骨科、神经外科和肿瘤科的诊断准确性和治疗规划至关重要,但在临床应用中却面临着挑战,包括与医疗保健系统的整合。因此,我们的重点在于探索知识蒸馏(KD)方法的应用,以训练能够有效分割 CT 扫描中椎骨的较浅网络。这种方法旨在缩短分割时间,提高紧急情况下的适用性,并优化计算和内存资源效率。在该领域先前研究的基础上,我们采用了两步分割法。首先,通过预测热图确定脊柱的位置,热图显示了每个体素属于脊柱的概率。随后,使用记忆实例记录已分割的椎体,在定位的脊柱上从 CT 容积的顶部到底部对椎体进行迭代分割。KD 方法是通过训练一个性能与文献中相似的教师网络来实现的,然后将这些知识提炼到一个较浅的网络(学生)中。应用了两种 KD 方法:(1) 使用两个网络的软输出;(2) 匹配对数。使用了两个公开可用的数据集,其中包括来自 300 名患者的 319 个 CT 扫描以及共计 611 个颈椎、2387 个胸椎和 1507 个腰椎。为确保数据集的平衡性和鲁棒性,采用了有效的数据增强方法,包括清理内存实例以复制第一个椎体分割。教师网络的平均骰子相似系数(DSC)为 88.22%,豪斯多夫距离(HD)为 7.71 mm,与文献中的其他方法表现相似。通过从教师网络中提炼知识,学生网络的性能得到提高,平均 DSC 从 75.78% 提高到 84.70%,HD 从 15.17 mm 下降到 8.08 mm。与其他方法相比,我们的教师网络的参数数量减少了 99.09%,推理时间缩短了 90.02%,总分割时间缩短了 88.46%,相关碳排放率降低了 89.36%。学生网络的参数比教师网络少 75.00%,推理时间缩短 36.15%,总分割时间缩短 33.33%,二氧化碳排放量减少 42.96%。这项研究标志着将 KD 应用于 CT 中单个椎骨分割问题的首次探索,证明了使用较小的神经网络实现与现有方法相当的性能的可行性。
{"title":"Knowledge distillation on individual vertebrae segmentation exploiting 3D U-Net","authors":"Luís Serrador ,&nbsp;Francesca Pia Villani ,&nbsp;Sara Moccia ,&nbsp;Cristina P. Santos","doi":"10.1016/j.compmedimag.2024.102350","DOIUrl":"https://doi.org/10.1016/j.compmedimag.2024.102350","url":null,"abstract":"<div><p>Recent advances in medical imaging have highlighted the critical development of algorithms for individual vertebral segmentation on computed tomography (CT) scans. Essential for diagnostic accuracy and treatment planning in orthopaedics, neurosurgery and oncology, these algorithms face challenges in clinical implementation, including integration into healthcare systems. Consequently, our focus lies in exploring the application of knowledge distillation (KD) methods to train shallower networks capable of efficiently segmenting vertebrae in CT scans. This approach aims to reduce segmentation time, enhance suitability for emergency cases, and optimize computational and memory resource efficiency. Building upon prior research in the field, a two-step segmentation approach was employed. Firstly, the spine’s location was determined by predicting a heatmap, indicating the probability of each voxel belonging to the spine. Subsequently, an iterative segmentation of vertebrae was performed from the top to the bottom of the CT volume over the located spine, using a memory instance to record the already segmented vertebrae. KD methods were implemented by training a teacher network with performance similar to that found in the literature, and this knowledge was distilled to a shallower network (student). Two KD methods were applied: (1) using the soft outputs of both networks and (2) matching logits. Two publicly available datasets, comprising 319 CT scans from 300 patients and a total of 611 cervical, 2387 thoracic, and 1507 lumbar vertebrae, were used. To ensure dataset balance and robustness, effective data augmentation methods were applied, including cleaning the memory instance to replicate the first vertebra segmentation. The teacher network achieved an average Dice similarity coefficient (DSC) of 88.22% and a Hausdorff distance (HD) of 7.71 mm, showcasing performance similar to other approaches in the literature. Through knowledge distillation from the teacher network, the student network’s performance improved, with an average DSC increasing from 75.78% to 84.70% and an HD decreasing from 15.17 mm to 8.08 mm. Compared to other methods, our teacher network exhibited up to 99.09% fewer parameters, 90.02% faster inference time, 88.46% shorter total segmentation time, and 89.36% less associated carbon (CO<sub>2</sub>) emission rate. Regarding our student network, it featured 75.00% fewer parameters than our teacher, resulting in a 36.15% reduction in inference time, a 33.33% decrease in total segmentation time, and a 42.96% reduction in CO<sub>2</sub> emissions. This study marks the first exploration of applying KD to the problem of individual vertebrae segmentation in CT, demonstrating the feasibility of achieving comparable performance to existing methods using smaller neural networks.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":null,"pages":null},"PeriodicalIF":5.7,"publicationDate":"2024-02-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0895611124000272/pdfft?md5=5527b04bad0cd774436ca9f2fd764d59&pid=1-s2.0-S0895611124000272-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139714472","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dynamic recurrent inference machines for accelerated MRI-guided radiotherapy of the liver 用于加速核磁共振成像引导的肝脏放射治疗的动态循环推理机
IF 5.7 2区 医学 Q1 Medicine Pub Date : 2024-02-08 DOI: 10.1016/j.compmedimag.2024.102348
Kai Lønning , Matthan W.A. Caan , Marlies E. Nowee , Jan-Jakob Sonke

Recurrent inference machines (RIM), a deep learning model that learns an iterative scheme for reconstructing sparsely sampled MRI, has been shown able to perform well on accelerated 2D and 3D MRI scans, learn from small datasets and generalize well to unseen types of data. Here we propose the dynamic recurrent inference machine (DRIM) for reconstructing sparsely sampled 4D MRI by exploiting correlations between respiratory states. The DRIM was applied to a 4D protocol for MR-guided radiotherapy of liver lesions based on repetitive interleaved coronal 2D multi-slice T2-weighted acquisitions. We demonstrate with an ablation study that the DRIM outperforms the RIM, increasing the SSIM score from about 0.89 to 0.95. The DRIM allowed for an approximately 2.7 times faster scan time than the current clinical protocol with only a slight loss in image sharpness. Correlations between slice locations can also be used, but were found to be of less importance, as were a majority of tested variations in network architecture, as long as the respiratory states are processed by the network. Through cross-validation, the DRIM is also shown to be robust in terms of training data. We further demonstrate a good performance across a large range of subsampling factors, and conclude through an evaluation by a radiation oncologist that reconstructed images of the liver contour and inner structures are of a clinically acceptable standard at acceleration factors 10x and 8x, respectively. Finally, we show that binning the data with respect to respiratory states prior to reconstruction comes at a slight cost to reconstruction quality, but at greater speed of the overall protocol.

递归推理机(RIM)是一种深度学习模型,用于学习重建稀疏采样核磁共振成像的迭代方案,已被证明能够在加速的二维和三维核磁共振成像扫描中表现出色,能够从小型数据集中学习,并能很好地泛化到未见过的数据类型。在此,我们提出了动态循环推理机(DRIM),利用呼吸状态之间的相关性重建稀疏采样的 4D MRI。DRIM 被应用于基于重复交错冠状二维多切片 T2 加权采集的肝脏病变磁共振引导放疗 4D 方案。我们通过一项消融研究证明,DRIM 优于 RIM,将 SSIM 分数从 0.89 提高到 0.95。DRIM 使扫描时间比目前的临床方案快了约 2.7 倍,而图像清晰度仅略有下降。切片位置之间的相关性也可以使用,但发现其重要性较低,就像大多数测试过的网络结构变化一样,只要呼吸状态由网络处理即可。通过交叉验证,DRIM 在训练数据方面也显示出很强的鲁棒性。我们进一步证明了 DRIM 在多种子采样因子下的良好性能,并通过一位放射肿瘤专家的评估得出结论:在加速因子分别为 10 倍和 8 倍的情况下,重建的肝脏轮廓和内部结构图像达到了临床可接受的标准。最后,我们证明了在重建之前根据呼吸状态对数据进行分档会对重建质量造成轻微影响,但却能提高整个方案的速度。
{"title":"Dynamic recurrent inference machines for accelerated MRI-guided radiotherapy of the liver","authors":"Kai Lønning ,&nbsp;Matthan W.A. Caan ,&nbsp;Marlies E. Nowee ,&nbsp;Jan-Jakob Sonke","doi":"10.1016/j.compmedimag.2024.102348","DOIUrl":"https://doi.org/10.1016/j.compmedimag.2024.102348","url":null,"abstract":"<div><p>Recurrent inference machines (RIM), a deep learning model that learns an iterative scheme for reconstructing sparsely sampled MRI, has been shown able to perform well on accelerated 2D and 3D MRI scans, learn from small datasets and generalize well to unseen types of data. Here we propose the dynamic recurrent inference machine (DRIM) for reconstructing sparsely sampled 4D MRI by exploiting correlations between respiratory states. The DRIM was applied to a 4D protocol for MR-guided radiotherapy of liver lesions based on repetitive interleaved coronal 2D multi-slice <span><math><msub><mrow><mi>T</mi></mrow><mrow><mn>2</mn></mrow></msub></math></span>-weighted acquisitions. We demonstrate with an ablation study that the DRIM outperforms the RIM, increasing the SSIM score from about 0.89 to 0.95. The DRIM allowed for an approximately 2.7 times faster scan time than the current clinical protocol with only a slight loss in image sharpness. Correlations between slice locations can also be used, but were found to be of less importance, as were a majority of tested variations in network architecture, as long as the respiratory states are processed by the network. Through cross-validation, the DRIM is also shown to be robust in terms of training data. We further demonstrate a good performance across a large range of subsampling factors, and conclude through an evaluation by a radiation oncologist that reconstructed images of the liver contour and inner structures are of a clinically acceptable standard at acceleration factors 10x and 8x, respectively. Finally, we show that binning the data with respect to respiratory states prior to reconstruction comes at a slight cost to reconstruction quality, but at greater speed of the overall protocol.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":null,"pages":null},"PeriodicalIF":5.7,"publicationDate":"2024-02-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139748249","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Computerized Medical Imaging and Graphics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1