首页 > 最新文献

Medical image analysis最新文献

英文 中文
ZygoPlanner: A three-stage graphics-based framework for optimal preoperative planning of zygomatic implant placement. ZygoPlanner:一个基于三个阶段图形的框架,用于颧骨植入的最佳术前规划。
IF 10.7 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-11-28 DOI: 10.1016/j.media.2024.103401
Haitao Li, Xingqi Fan, Baoxin Tao, Wenying Wang, Yiqun Wu, Xiaojun Chen

Zygomatic implant surgery is an essential treatment option of oral rehabilitation for patients with severe maxillary defect, and preoperative planning is an important approach to enhance the surgical outcomes. However, the current planning still heavily relies on manual interventions, which is labor-intensive, experience-dependent, and poorly reproducible. Therefore, we propose ZygoPlanner, a pioneering efficient preoperative planning framework for zygomatic implantation, which may be the first solution that seamlessly involves the positioning of zygomatic bones, the generation of alternative paths, and the computation of optimal implantation paths. To efficiently achieve robust planning, we developed a graphics-based interpretable method for zygomatic bone positioning leveraging the shape prior knowledge. Meanwhile, a surface-faithful point cloud filling algorithm that works for concave geometries was proposed to populate dense points within the zygomatic bones, facilitating generation of alternative paths. Finally, we innovatively realized a graphical representation of the medical bone-to-implant contact to obtain the optimal results under multiple constraints. Clinical experiments confirmed the superiority of our framework across different scenarios. The source code is available at https://github.com/Haitao-Lee/auto_zygomatic_implantation.

颧骨种植手术是上颌严重缺损患者口腔康复的重要治疗选择,而术前规划是提高手术效果的重要方法。然而,目前的规划仍严重依赖人工干预,劳动强度大、经验依赖性强、可重复性差。因此,我们提出了 ZygoPlanner,一个开创性的高效颧骨植入术前规划框架,它可能是第一个能无缝整合颧骨定位、替代路径生成和最佳植入路径计算的解决方案。为了有效实现稳健规划,我们开发了一种基于图形的可解释方法,利用形状先验知识进行颧骨定位。同时,我们还提出了一种适用于凹面几何形状的表面忠实点云填充算法,用于填充颧骨内的密集点,从而促进替代路径的生成。最后,我们创新性地实现了医学骨与种植体接触的图形表示,从而在多重约束条件下获得最佳结果。临床实验证实了我们的框架在不同情况下的优越性。源代码见 https://github.com/Haitao-Lee/auto_zygomatic_implantation。
{"title":"ZygoPlanner: A three-stage graphics-based framework for optimal preoperative planning of zygomatic implant placement.","authors":"Haitao Li, Xingqi Fan, Baoxin Tao, Wenying Wang, Yiqun Wu, Xiaojun Chen","doi":"10.1016/j.media.2024.103401","DOIUrl":"https://doi.org/10.1016/j.media.2024.103401","url":null,"abstract":"<p><p>Zygomatic implant surgery is an essential treatment option of oral rehabilitation for patients with severe maxillary defect, and preoperative planning is an important approach to enhance the surgical outcomes. However, the current planning still heavily relies on manual interventions, which is labor-intensive, experience-dependent, and poorly reproducible. Therefore, we propose ZygoPlanner, a pioneering efficient preoperative planning framework for zygomatic implantation, which may be the first solution that seamlessly involves the positioning of zygomatic bones, the generation of alternative paths, and the computation of optimal implantation paths. To efficiently achieve robust planning, we developed a graphics-based interpretable method for zygomatic bone positioning leveraging the shape prior knowledge. Meanwhile, a surface-faithful point cloud filling algorithm that works for concave geometries was proposed to populate dense points within the zygomatic bones, facilitating generation of alternative paths. Finally, we innovatively realized a graphical representation of the medical bone-to-implant contact to obtain the optimal results under multiple constraints. Clinical experiments confirmed the superiority of our framework across different scenarios. The source code is available at https://github.com/Haitao-Lee/auto_zygomatic_implantation.</p>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"101 ","pages":"103401"},"PeriodicalIF":10.7,"publicationDate":"2024-11-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142818562","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Prediction of the upright articulated spine shape in the operating room using conditioned neural kernel fields 应用条件神经核场预测手术室直立关节脊柱形态
IF 10.7 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-11-27 DOI: 10.1016/j.media.2024.103400
Sylvain Thibeault , Marjolaine Roy-Beaudry , Stefan Parent , Samuel Kadoury
Anterior vertebral tethering (AVT) is a non-invasive spine surgery technique, treating severe spine deformations and preserving lower back mobility. However, patient positioning and surgical strategies greatly influences postoperative results. Predicting the upright geometry from pediatric spines is needed to optimize patient positioning in the operating room (OR) and improve surgical outcomes, but remains a complex task due to immature bone properties. We propose a framework used in the OR predicting the upright spine geometry at the first visit following surgery in idiopathic scoliosis patients. The approach first creates a 3D model of the spine while the patient is on the operating table. For this, multiview Transformers that combine images from different viewpoints are used to generate the intraoperative pose. The postoperative upright shape is then predicted on-the-fly using implicit neural fields, which are trained from geometries at different time points and conditioned with surgical parameters. A Signed Distance Function for shape constellations is used to handle the variability in spine appearance, capturing a disentangled latent domain of the articulation vectors, with separate encoding vectors representing both articulation and shape parameters. A regularization criterion based on a pre-trained group-wise trajectory of spine transformations generates complete spine models. A training set of 652 patients with 3D models was used to train the model, tested on a distinct cohort of 83 surgical patients. The framework based on neural kernels predicted upright 3D geometries with a mean 3D error of 1.3±0.5mm in landmarks points, and IoU of 95.9% in vertebral shapes when compared to actual postop models, falling within the acceptable margins of error below 2 mm.
前路椎体系扎术(AVT)是一种非侵入性脊柱手术技术,用于治疗严重的脊柱变形和保持下背部的活动能力。然而,患者体位和手术策略对术后结果有很大影响。预测儿童脊柱的直立几何形状是优化患者在手术室(OR)的定位和改善手术效果所必需的,但由于骨骼特性不成熟,这仍然是一项复杂的任务。我们提出了一个框架,用于OR预测直立脊柱几何形状在第一次访问后,特发性脊柱侧凸患者的手术。该方法首先在患者躺在手术台上时创建脊柱的3D模型。为此,结合不同视点图像的多视图变形器用于生成术中姿势。然后使用隐式神经场实时预测术后直立形状,隐式神经场从不同时间点的几何形状中训练,并以手术参数为条件。形状星座的签名距离函数用于处理脊柱外观的可变性,捕获关节向量的解纠缠潜在域,使用单独的编码向量表示关节和形状参数。基于预先训练的脊柱转换的组明智轨迹的正则化准则生成完整的脊柱模型。使用652名具有3D模型的患者的训练集来训练模型,并在83名外科患者的不同队列中进行测试。基于神经核的框架预测直立三维几何形状,在地标点上的平均3D误差为1.3±0.5mm,与实际的后支架模型相比,椎体形状的IoU为95.9%,在可接受的误差范围内低于2mm。
{"title":"Prediction of the upright articulated spine shape in the operating room using conditioned neural kernel fields","authors":"Sylvain Thibeault ,&nbsp;Marjolaine Roy-Beaudry ,&nbsp;Stefan Parent ,&nbsp;Samuel Kadoury","doi":"10.1016/j.media.2024.103400","DOIUrl":"10.1016/j.media.2024.103400","url":null,"abstract":"<div><div>Anterior vertebral tethering (AVT) is a non-invasive spine surgery technique, treating severe spine deformations and preserving lower back mobility. However, patient positioning and surgical strategies greatly influences postoperative results. Predicting the upright geometry from pediatric spines is needed to optimize patient positioning in the operating room (OR) and improve surgical outcomes, but remains a complex task due to immature bone properties. We propose a framework used in the OR predicting the upright spine geometry at the first visit following surgery in idiopathic scoliosis patients. The approach first creates a 3D model of the spine while the patient is on the operating table. For this, multiview Transformers that combine images from different viewpoints are used to generate the intraoperative pose. The postoperative upright shape is then predicted on-the-fly using implicit neural fields, which are trained from geometries at different time points and conditioned with surgical parameters. A Signed Distance Function for shape constellations is used to handle the variability in spine appearance, capturing a disentangled latent domain of the articulation vectors, with separate encoding vectors representing both articulation and shape parameters. A regularization criterion based on a pre-trained group-wise trajectory of spine transformations generates complete spine models. A training set of 652 patients with 3D models was used to train the model, tested on a distinct cohort of 83 surgical patients. The framework based on neural kernels predicted upright 3D geometries with a mean 3D error of <span><math><mrow><mn>1</mn><mo>.</mo><mn>3</mn><mo>±</mo><mn>0</mn><mo>.</mo><mn>5</mn><mspace></mspace><mi>mm</mi></mrow></math></span> in landmarks points, and IoU of 95.9% in vertebral shapes when compared to actual postop models, falling within the acceptable margins of error below 2 mm.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"100 ","pages":"Article 103400"},"PeriodicalIF":10.7,"publicationDate":"2024-11-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142756929","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-scale region selection network in deep features for full-field mammogram classification 基于深度特征的多尺度区域选择网络进行乳房x线照片全视野分类
IF 10.7 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-11-26 DOI: 10.1016/j.media.2024.103399
Luhao Sun , Bowen Han , Wenzong Jiang , Weifeng Liu , Baodi Liu , Dapeng Tao , Zhiyong Yu , Chao Li
Early diagnosis and treatment of breast cancer can effectively reduce mortality. Since mammogram is one of the most commonly used methods in the early diagnosis of breast cancer, the classification of mammogram images is an important work of computer-aided diagnosis (CAD) systems. With the development of deep learning in CAD, deep convolutional neural networks have been shown to have the ability to complete the classification of breast cancer tumor patches with high quality, which makes most previous CNN-based full-field mammography classification methods rely on region of interest (ROI) or segmentation annotation to enable the model to locate and focus on small tumor regions. However, the dependence on ROI greatly limits the development of CAD, because obtaining a large number of reliable ROI annotations is expensive and difficult. Some full-field mammography image classification algorithms use multi-stage training or multi-feature extractors to get rid of the dependence on ROI, which increases the computational amount of the model and feature redundancy. In order to reduce the cost of model training and make full use of the feature extraction capability of CNN, we propose a deep multi-scale region selection network (MRSN) in deep features for end-to-end training to classify full-field mammography without ROI or segmentation annotation. Inspired by the idea of multi-example learning and the patch classifier, MRSN filters the feature information and saves only the feature information of the tumor region to make the performance of the full-field image classifier closer to the patch classifier. MRSN first scores different regions under different dimensions to obtain the location information of tumor regions. Then, a few high-scoring regions are selected by location information as feature representations of the entire image, allowing the model to focus on the tumor region. Experiments on two public datasets and one private dataset prove that the proposed MRSN achieves the most advanced performance.
乳腺癌的早期诊断和治疗可以有效降低死亡率。由于乳房x光检查是早期诊断乳腺癌最常用的方法之一,因此乳房x光检查图像的分类是计算机辅助诊断(CAD)系统的一项重要工作。随着CAD中深度学习的发展,深度卷积神经网络已被证明具有高质量完成乳腺癌肿瘤斑块分类的能力,这使得以往大多数基于cnn的全视场乳房x线摄影分类方法依赖于感兴趣区域(ROI)或分割标注,使模型能够定位和关注小肿瘤区域。然而,对ROI的依赖极大地限制了CAD的发展,因为获得大量可靠的ROI注释既昂贵又困难。一些全视场乳腺摄影图像分类算法采用多阶段训练或多特征提取器来摆脱对ROI的依赖,这增加了模型的计算量和特征冗余度。为了降低模型训练成本,充分利用CNN的特征提取能力,我们提出了一种基于深度特征的深度多尺度区域选择网络(MRSN)进行端到端训练,在不需要ROI和分割标注的情况下对全场乳房x线照片进行分类。​MRSN首先对不同维度下的不同区域进行评分,得到肿瘤区域的位置信息。然后,根据位置信息选择几个高分区域作为整幅图像的特征表示,使模型专注于肿瘤区域。在两个公共数据集和一个私有数据集上的实验证明,所提出的MRSN达到了最先进的性能。
{"title":"Multi-scale region selection network in deep features for full-field mammogram classification","authors":"Luhao Sun ,&nbsp;Bowen Han ,&nbsp;Wenzong Jiang ,&nbsp;Weifeng Liu ,&nbsp;Baodi Liu ,&nbsp;Dapeng Tao ,&nbsp;Zhiyong Yu ,&nbsp;Chao Li","doi":"10.1016/j.media.2024.103399","DOIUrl":"10.1016/j.media.2024.103399","url":null,"abstract":"<div><div>Early diagnosis and treatment of breast cancer can effectively reduce mortality. Since mammogram is one of the most commonly used methods in the early diagnosis of breast cancer, the classification of mammogram images is an important work of computer-aided diagnosis (CAD) systems. With the development of deep learning in CAD, deep convolutional neural networks have been shown to have the ability to complete the classification of breast cancer tumor patches with high quality, which makes most previous CNN-based full-field mammography classification methods rely on region of interest (ROI) or segmentation annotation to enable the model to locate and focus on small tumor regions. However, the dependence on ROI greatly limits the development of CAD, because obtaining a large number of reliable ROI annotations is expensive and difficult. Some full-field mammography image classification algorithms use multi-stage training or multi-feature extractors to get rid of the dependence on ROI, which increases the computational amount of the model and feature redundancy. In order to reduce the cost of model training and make full use of the feature extraction capability of CNN, we propose a deep multi-scale region selection network (MRSN) in deep features for end-to-end training to classify full-field mammography without ROI or segmentation annotation. Inspired by the idea of multi-example learning and the patch classifier, MRSN filters the feature information and saves only the feature information of the tumor region to make the performance of the full-field image classifier closer to the patch classifier. MRSN first scores different regions under different dimensions to obtain the location information of tumor regions. Then, a few high-scoring regions are selected by location information as feature representations of the entire image, allowing the model to focus on the tumor region. Experiments on two public datasets and one private dataset prove that the proposed MRSN achieves the most advanced performance.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"100 ","pages":"Article 103399"},"PeriodicalIF":10.7,"publicationDate":"2024-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142744060","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Incorporating spatial information in deep learning parameter estimation with application to the intravoxel incoherent motion model in diffusion-weighted MRI. 融合空间信息的深度学习参数估计及其在弥散加权MRI体内非相干运动模型中的应用。
IF 10.7 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-11-26 DOI: 10.1016/j.media.2024.103414
Misha P T Kaandorp, Frank Zijlstra, Davood Karimi, Ali Gholipour, Peter T While

In medical image analysis, the utilization of biophysical models for signal analysis offers valuable insights into the underlying tissue types and microstructural processes. In diffusion-weighted magnetic resonance imaging (DWI), a major challenge lies in accurately estimating model parameters from the acquired data due to the inherently low signal-to-noise ratio (SNR) of the signal measurements and the complexity of solving the ill-posed inverse problem. Conventional model fitting approaches treat individual voxels as independent. However, the tissue microenvironment is typically homogeneous in a local environment, where neighboring voxels may contain correlated information. To harness the potential benefits of exploiting correlations among signals in adjacent voxels, this study introduces a novel approach to deep learning parameter estimation that effectively incorporates relevant spatial information. This is achieved by training neural networks on patches of synthetic data encompassing plausible combinations of direct correlations between neighboring voxels. We evaluated the approach on the intravoxel incoherent motion (IVIM) model in DWI. We explored the potential of several deep learning architectures to incorporate spatial information using self-supervised and supervised learning. We assessed performance quantitatively using novel fractal-noise-based synthetic data, which provide ground truths possessing spatial correlations. Additionally, we present results of the approach applied to in vivo DWI data consisting of twelve repetitions from a healthy volunteer. We demonstrate that supervised training on larger patch sizes using attention models leads to substantial performance improvements over both conventional voxelwise model fitting and convolution-based approaches.

在医学图像分析中,利用生物物理模型进行信号分析为潜在的组织类型和微观结构过程提供了有价值的见解。在扩散加权磁共振成像(DWI)中,由于信号测量固有的低信噪比(SNR)和求解病态逆问题的复杂性,如何从采集的数据中准确估计模型参数是一个主要的挑战。传统的模型拟合方法将单个体素视为独立的。然而,组织微环境在局部环境中通常是同质的,相邻体素可能包含相关信息。为了利用相邻体素中信号之间的相关性的潜在好处,本研究引入了一种新的深度学习参数估计方法,该方法有效地结合了相关的空间信息。这是通过在包含相邻体素之间直接关联的合理组合的合成数据块上训练神经网络来实现的。我们在DWI的体内非相干运动(IVIM)模型上评估了该方法。我们探索了几种深度学习架构的潜力,利用自监督学习和监督学习来整合空间信息。我们使用新的基于分形噪声的合成数据定量评估性能,这些数据提供了具有空间相关性的地面事实。此外,我们还介绍了将该方法应用于健康志愿者体内DWI数据的结果,该数据由12次重复组成。我们证明,与传统的体素模型拟合和基于卷积的方法相比,使用注意力模型在更大的补丁尺寸上进行监督训练可以显著提高性能。
{"title":"Incorporating spatial information in deep learning parameter estimation with application to the intravoxel incoherent motion model in diffusion-weighted MRI.","authors":"Misha P T Kaandorp, Frank Zijlstra, Davood Karimi, Ali Gholipour, Peter T While","doi":"10.1016/j.media.2024.103414","DOIUrl":"https://doi.org/10.1016/j.media.2024.103414","url":null,"abstract":"<p><p>In medical image analysis, the utilization of biophysical models for signal analysis offers valuable insights into the underlying tissue types and microstructural processes. In diffusion-weighted magnetic resonance imaging (DWI), a major challenge lies in accurately estimating model parameters from the acquired data due to the inherently low signal-to-noise ratio (SNR) of the signal measurements and the complexity of solving the ill-posed inverse problem. Conventional model fitting approaches treat individual voxels as independent. However, the tissue microenvironment is typically homogeneous in a local environment, where neighboring voxels may contain correlated information. To harness the potential benefits of exploiting correlations among signals in adjacent voxels, this study introduces a novel approach to deep learning parameter estimation that effectively incorporates relevant spatial information. This is achieved by training neural networks on patches of synthetic data encompassing plausible combinations of direct correlations between neighboring voxels. We evaluated the approach on the intravoxel incoherent motion (IVIM) model in DWI. We explored the potential of several deep learning architectures to incorporate spatial information using self-supervised and supervised learning. We assessed performance quantitatively using novel fractal-noise-based synthetic data, which provide ground truths possessing spatial correlations. Additionally, we present results of the approach applied to in vivo DWI data consisting of twelve repetitions from a healthy volunteer. We demonstrate that supervised training on larger patch sizes using attention models leads to substantial performance improvements over both conventional voxelwise model fitting and convolution-based approaches.</p>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"101 ","pages":"103414"},"PeriodicalIF":10.7,"publicationDate":"2024-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142910005","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
NCCT-to-CECT synthesis with contrast-enhanced knowledge and anatomical perception for multi-organ segmentation in non-contrast CT images NCCT-to-CECT合成与对比度增强知识和解剖感知在非对比CT图像中进行多器官分割
IF 10.7 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-11-26 DOI: 10.1016/j.media.2024.103397
Liming Zhong , Ruolin Xiao , Hai Shu , Kaiyi Zheng , Xinming Li , Yuankui Wu , Jianhua Ma , Qianjin Feng , Wei Yang
Contrast-enhanced computed tomography (CECT) is constantly used for delineating organs-at-risk (OARs) in radiation therapy planning. The delineated OARs are needed to transfer from CECT to non-contrast CT (NCCT) for dose calculation. Yet, the use of iodinated contrast agents (CA) in CECT and the dose calculation errors caused by the spatial misalignment between NCCT and CECT images pose risks of adverse side effects. A promising solution is synthesizing CECT images from NCCT scans, which can improve the visibility of organs and abnormalities for more effective multi-organ segmentation in NCCT images. However, existing methods neglect the difference between tissues induced by CA and lack the ability to synthesize the details of organ edges and blood vessels. To address these issues, we propose a contrast-enhanced knowledge and anatomical perception network (CKAP-Net) for NCCT-to-CECT synthesis. CKAP-Net leverages a contrast-enhanced knowledge learning network to capture both similarities and dissimilarities in domain characteristics attributable to CA. Specifically, a CA-based perceptual loss function is introduced to enhance the synthesis of CA details. Furthermore, we design a multi-scale anatomical perception transformer that utilizes multi-scale anatomical information from NCCT images, enabling the precise synthesis of tissue details. Our CKAP-Net is evaluated on a multi-center abdominal NCCT-CECT dataset, a head an neck NCCT-CECT dataset, and an NCMRI-CEMRI dataset. It achieves a MAE of 25.96 ± 2.64, a SSIM of 0.855 ± 0.017, and a PSNR of 32.60 ± 0.02 for CECT synthesis, and a DSC of 81.21 ± 4.44 for segmentation on the internal dataset. Extensive experiments demonstrate that CKAP-Net outperforms state-of-the-art CA synthesis methods and has better generalizability across different datasets.
对比增强计算机断层扫描(CECT)在放射治疗计划中经常用于描绘危险器官(OARs)。所描绘的桨需要从CECT转移到非对比CT (NCCT)进行剂量计算。然而,在CECT中使用碘造影剂(CA)以及NCCT与CECT图像空间错位导致的剂量计算误差存在不良副作用的风险。一种很有前景的解决方案是从NCCT扫描中合成CECT图像,这可以提高器官和异常的可见性,从而更有效地在NCCT图像中进行多器官分割。然而,现有的方法忽略了CA诱导的组织之间的差异,缺乏对器官边缘和血管细节的合成能力。为了解决这些问题,我们提出了一个对比增强的知识和解剖感知网络(CKAP-Net)用于ncct到cect的合成。CKAP-Net利用对比度增强的知识学习网络来捕获可归因于CA的领域特征的相似性和差异性。具体而言,引入了基于CA的感知损失函数来增强CA细节的合成。此外,我们设计了一个多尺度解剖感知转换器,利用来自NCCT图像的多尺度解剖信息,实现组织细节的精确合成。我们的CKAP-Net在多中心腹部NCCT-CECT数据集、头部和颈部NCCT-CECT数据集和ncmi - cemri数据集上进行了评估。在内部数据集上,CECT合成的MAE为25.96±2.64,SSIM为0.855±0.017,PSNR为32.60±0.02,DSC为81.21±4.44。大量的实验表明,CKAP-Net优于最先进的CA合成方法,并且在不同的数据集上具有更好的泛化性。
{"title":"NCCT-to-CECT synthesis with contrast-enhanced knowledge and anatomical perception for multi-organ segmentation in non-contrast CT images","authors":"Liming Zhong ,&nbsp;Ruolin Xiao ,&nbsp;Hai Shu ,&nbsp;Kaiyi Zheng ,&nbsp;Xinming Li ,&nbsp;Yuankui Wu ,&nbsp;Jianhua Ma ,&nbsp;Qianjin Feng ,&nbsp;Wei Yang","doi":"10.1016/j.media.2024.103397","DOIUrl":"10.1016/j.media.2024.103397","url":null,"abstract":"<div><div>Contrast-enhanced computed tomography (CECT) is constantly used for delineating organs-at-risk (OARs) in radiation therapy planning. The delineated OARs are needed to transfer from CECT to non-contrast CT (NCCT) for dose calculation. Yet, the use of iodinated contrast agents (CA) in CECT and the dose calculation errors caused by the spatial misalignment between NCCT and CECT images pose risks of adverse side effects. A promising solution is synthesizing CECT images from NCCT scans, which can improve the visibility of organs and abnormalities for more effective multi-organ segmentation in NCCT images. However, existing methods neglect the difference between tissues induced by CA and lack the ability to synthesize the details of organ edges and blood vessels. To address these issues, we propose a contrast-enhanced knowledge and anatomical perception network (CKAP-Net) for NCCT-to-CECT synthesis. CKAP-Net leverages a contrast-enhanced knowledge learning network to capture both similarities and dissimilarities in domain characteristics attributable to CA. Specifically, a CA-based perceptual loss function is introduced to enhance the synthesis of CA details. Furthermore, we design a multi-scale anatomical perception transformer that utilizes multi-scale anatomical information from NCCT images, enabling the precise synthesis of tissue details. Our CKAP-Net is evaluated on a multi-center abdominal NCCT-CECT dataset, a head an neck NCCT-CECT dataset, and an NCMRI-CEMRI dataset. It achieves a MAE of 25.96 ± 2.64, a SSIM of 0.855 ± 0.017, and a PSNR of 32.60 ± 0.02 for CECT synthesis, and a DSC of 81.21 ± 4.44 for segmentation on the internal dataset. Extensive experiments demonstrate that CKAP-Net outperforms state-of-the-art CA synthesis methods and has better generalizability across different datasets.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"100 ","pages":"Article 103397"},"PeriodicalIF":10.7,"publicationDate":"2024-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142744163","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multidimensional Directionality-Enhanced Segmentation via large vision model. 基于大视觉模型的多维方向性增强分割。
IF 10.7 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-11-25 DOI: 10.1016/j.media.2024.103395
Xingru Huang, Changpeng Yue, Yihao Guo, Jian Huang, Zhengyao Jiang, Mingkuan Wang, Zhaoyang Xu, Guangyuan Zhang, Jin Liu, Tianyun Zhang, Zhiwen Zheng, Xiaoshuai Zhang, Hong He, Shaowei Jiang, Yaoqi Sun

Optical Coherence Tomography (OCT) facilitates a comprehensive examination of macular edema and associated lesions. Manual delineation of retinal fluid is labor-intensive and error-prone, necessitating an automated diagnostic and therapeutic planning mechanism. Conventional supervised learning models are hindered by dataset limitations, while Transformer-based large vision models exhibit challenges in medical image segmentation, particularly in detecting small, subtle lesions in OCT images. This paper introduces the Multidimensional Directionality-Enhanced Retinal Fluid Segmentation framework (MD-DERFS), which reduces the limitations inherent in conventional supervised models by adapting a transformer-based large vision model for macular edema segmentation. The proposed MD-DERFS introduces a Multi-Dimensional Feature Re-Encoder Unit (MFU) to augment the model's proficiency in recognizing specific textures and pathological features through directional prior extraction and an Edema Texture Mapping Unit (ETMU), a Cross-scale Directional Insight Network (CDIN) furnishes a holistic perspective spanning local to global details, mitigating the large vision model's deficiencies in capturing localized feature information. Additionally, the framework is augmented by a Harmonic Minutiae Segmentation Equilibrium loss (LHMSE) that can address the challenges of data imbalance and annotation scarcity in macular edema datasets. Empirical validation on the MacuScan-8k dataset shows that MD-DERFS surpasses existing segmentation methodologies, demonstrating its efficacy in adapting large vision models for boundary-sensitive medical imaging tasks. The code is publicly available at https://github.com/IMOP-lab/MD-DERFS-Pytorch.git.

光学相干断层扫描(OCT)有助于黄斑水肿和相关病变的全面检查。手动划定视网膜液是劳动密集型和容易出错,需要一个自动化的诊断和治疗计划机制。传统的监督学习模型受到数据集限制的阻碍,而基于transformer的大视觉模型在医学图像分割方面表现出挑战,特别是在检测OCT图像中的小而微妙的病变方面。本文介绍了多维方向性增强视网膜液体分割框架(MD-DERFS),该框架采用基于变压器的大视觉模型对黄斑水肿进行分割,减少了传统监督模型固有的局限性。提出的MD-DERFS引入了多维特征重新编码器单元(MFU),通过定向先验提取增强模型识别特定纹理和病理特征的熟练程度;引入了水肿纹理映射单元(ETMU);引入了跨尺度定向洞察网络(CDIN),提供了从局部到全局细节的整体视角,减轻了大视觉模型在捕获局部特征信息方面的不足。此外,该框架还增加了调和细节分割平衡损失(LHMSE),可以解决黄斑水肿数据集中数据不平衡和注释稀缺性的挑战。对MacuScan-8k数据集的实证验证表明,MD-DERFS超越了现有的分割方法,证明了其在适应大视觉模型进行边界敏感医学成像任务方面的有效性。该代码可在https://github.com/IMOP-lab/MD-DERFS-Pytorch.git上公开获得。
{"title":"Multidimensional Directionality-Enhanced Segmentation via large vision model.","authors":"Xingru Huang, Changpeng Yue, Yihao Guo, Jian Huang, Zhengyao Jiang, Mingkuan Wang, Zhaoyang Xu, Guangyuan Zhang, Jin Liu, Tianyun Zhang, Zhiwen Zheng, Xiaoshuai Zhang, Hong He, Shaowei Jiang, Yaoqi Sun","doi":"10.1016/j.media.2024.103395","DOIUrl":"https://doi.org/10.1016/j.media.2024.103395","url":null,"abstract":"<p><p>Optical Coherence Tomography (OCT) facilitates a comprehensive examination of macular edema and associated lesions. Manual delineation of retinal fluid is labor-intensive and error-prone, necessitating an automated diagnostic and therapeutic planning mechanism. Conventional supervised learning models are hindered by dataset limitations, while Transformer-based large vision models exhibit challenges in medical image segmentation, particularly in detecting small, subtle lesions in OCT images. This paper introduces the Multidimensional Directionality-Enhanced Retinal Fluid Segmentation framework (MD-DERFS), which reduces the limitations inherent in conventional supervised models by adapting a transformer-based large vision model for macular edema segmentation. The proposed MD-DERFS introduces a Multi-Dimensional Feature Re-Encoder Unit (MFU) to augment the model's proficiency in recognizing specific textures and pathological features through directional prior extraction and an Edema Texture Mapping Unit (ETMU), a Cross-scale Directional Insight Network (CDIN) furnishes a holistic perspective spanning local to global details, mitigating the large vision model's deficiencies in capturing localized feature information. Additionally, the framework is augmented by a Harmonic Minutiae Segmentation Equilibrium loss (L<sub>HMSE</sub>) that can address the challenges of data imbalance and annotation scarcity in macular edema datasets. Empirical validation on the MacuScan-8k dataset shows that MD-DERFS surpasses existing segmentation methodologies, demonstrating its efficacy in adapting large vision models for boundary-sensitive medical imaging tasks. The code is publicly available at https://github.com/IMOP-lab/MD-DERFS-Pytorch.git.</p>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"101 ","pages":"103395"},"PeriodicalIF":10.7,"publicationDate":"2024-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142791576","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CLMS: Bridging domain gaps in medical imaging segmentation with source-free continual learning for robust knowledge transfer and adaptation CLMS:通过无源持续学习弥合医学成像分割领域的差距,实现强大的知识转移和适应
IF 10.7 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-11-24 DOI: 10.1016/j.media.2024.103404
Weilu Li , Yun Zhang , Hao Zhou, Wenhan Yang, Zhi Xie, Yao He
Deep learning shows promise for medical image segmentation but suffers performance declines when applied to diverse healthcare sites due to data discrepancies among the different sites. Translating deep learning models to new clinical environments is challenging, especially when the original source data used for training is unavailable due to privacy restrictions. Source-free domain adaptation (SFDA) aims to adapt models to new unlabeled target domains without requiring access to the original source data. However, existing SFDA methods face challenges such as error propagation, misalignment of visual and structural features, and inability to preserve source knowledge. This paper introduces Continual Learning Multi-Scale domain adaptation (CLMS), an end-to-end SFDA framework integrating multi-scale reconstruction, continual learning, and style alignment to bridge domain gaps across medical sites using only unlabeled target data or publicly available data. Compared to the current state-of-the-art methods, CLMS consistently and significantly achieved top performance for different tasks, including prostate MRI segmentation (improved Dice of 10.87 %), colonoscopy polyp segmentation (improved Dice of 17.73 %), and plus disease classification from retinal images (improved AUC of 11.19 %). Crucially, CLMS preserved source knowledge for all the tasks, avoiding catastrophic forgetting. CLMS demonstrates a promising solution for translating deep learning models to new clinical imaging domains towards safe, reliable deployment across diverse healthcare settings.
深度学习显示了医学图像分割的前景,但由于不同站点之间的数据差异,当应用于不同的医疗保健站点时,性能会下降。将深度学习模型转化为新的临床环境是具有挑战性的,特别是当用于训练的原始源数据由于隐私限制而不可用时。无源域自适应(SFDA)旨在使模型适应新的未标记的目标域,而不需要访问原始源数据。然而,现有的SFDA方法面临着诸如误差传播、视觉和结构特征不一致以及无法保存源知识等挑战。本文介绍了持续学习多尺度域适应(CLMS),这是一个端到端的SFDA框架,集成了多尺度重建、持续学习和风格对齐,仅使用未标记的目标数据或公开可用的数据来弥合医疗站点之间的域差距。与目前最先进的方法相比,CLMS在不同任务上一致且显著地取得了最佳性能,包括前列腺MRI分割(提高了10.87%的Dice),结肠镜息肉分割(提高了17.73%的Dice),以及视网膜图像的疾病分类(提高了11.19%的AUC)。关键是,CLMS保留了所有任务的源知识,避免了灾难性遗忘。CLMS展示了一种很有前途的解决方案,可将深度学习模型转换为新的临床成像领域,从而在各种医疗保健环境中安全、可靠地部署。
{"title":"CLMS: Bridging domain gaps in medical imaging segmentation with source-free continual learning for robust knowledge transfer and adaptation","authors":"Weilu Li ,&nbsp;Yun Zhang ,&nbsp;Hao Zhou,&nbsp;Wenhan Yang,&nbsp;Zhi Xie,&nbsp;Yao He","doi":"10.1016/j.media.2024.103404","DOIUrl":"10.1016/j.media.2024.103404","url":null,"abstract":"<div><div>Deep learning shows promise for medical image segmentation but suffers performance declines when applied to diverse healthcare sites due to data discrepancies among the different sites. Translating deep learning models to new clinical environments is challenging, especially when the original source data used for training is unavailable due to privacy restrictions. Source-free domain adaptation (SFDA) aims to adapt models to new unlabeled target domains without requiring access to the original source data. However, existing SFDA methods face challenges such as error propagation, misalignment of visual and structural features, and inability to preserve source knowledge. This paper introduces Continual Learning Multi-Scale domain adaptation (CLMS), an end-to-end SFDA framework integrating multi-scale reconstruction, continual learning, and style alignment to bridge domain gaps across medical sites using only unlabeled target data or publicly available data. Compared to the current state-of-the-art methods, CLMS consistently and significantly achieved top performance for different tasks, including prostate MRI segmentation (improved Dice of 10.87 %), colonoscopy polyp segmentation (improved Dice of 17.73 %), and plus disease classification from retinal images (improved AUC of 11.19 %). Crucially, CLMS preserved source knowledge for all the tasks, avoiding catastrophic forgetting. CLMS demonstrates a promising solution for translating deep learning models to new clinical imaging domains towards safe, reliable deployment across diverse healthcare settings.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"100 ","pages":"Article 103404"},"PeriodicalIF":10.7,"publicationDate":"2024-11-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142756236","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Highly accelerated MRI via implicit neural representation guided posterior sampling of diffusion models 通过隐式神经表征引导的扩散模型后向采样实现高速磁共振成像
IF 10.7 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-11-23 DOI: 10.1016/j.media.2024.103398
Jiayue Chu , Chenhe Du , Xiyue Lin , Xiaoqun Zhang , Lihui Wang , Yuyao Zhang , Hongjiang Wei
Reconstructing high-fidelity magnetic resonance (MR) images from under-sampled k-space is a commonly used strategy to reduce scan time. The posterior sampling of diffusion models based on the real measurement data holds significant promise of improved reconstruction accuracy. However, traditional posterior sampling methods often lack effective data consistency guidance, leading to inaccurate and unstable reconstructions. Implicit neural representation (INR) has emerged as a powerful paradigm for solving inverse problems by modeling a signal’s attributes as a continuous function of spatial coordinates. In this study, we present a novel posterior sampler for diffusion models using INR, named DiffINR. The INR-based component incorporates both the diffusion prior distribution and the MRI physical model to ensure high data fidelity. DiffINR demonstrates superior performance on in-distribution datasets with remarkable accuracy, even under high acceleration factors (up to R = 12 in single-channel reconstruction). Furthermore, DiffINR exhibits excellent generalizability across various tissue contrasts and anatomical structures with low uncertainty. Overall, DiffINR significantly improves MRI reconstruction in terms of accuracy, generalizability and stability, paving the way for further accelerating MRI acquisition. Notably, our proposed framework can be a generalizable framework to solve inverse problems in other medical imaging tasks.
从采样不足的 k 空间重建高保真磁共振(MR)图像是缩短扫描时间的常用策略。根据真实测量数据对扩散模型进行后向采样,有望显著提高重建精度。然而,传统的后向采样方法往往缺乏有效的数据一致性指导,导致重建不准确、不稳定。隐式神经表示(INR)通过将信号属性建模为空间坐标的连续函数,已成为解决逆问题的强大范例。在这项研究中,我们为使用 INR 的扩散模型提出了一种新型后验采样器,名为 DiffINR。基于 INR 的组件结合了扩散先验分布和核磁共振物理模型,以确保高数据保真度。DiffINR 在分布内数据集上表现出卓越的性能,即使在高加速因子(单通道重建中高达 R = 12)条件下也能保持出色的准确性。此外,DiffINR 在各种组织对比度和解剖结构上都表现出卓越的通用性,不确定性很低。总之,DiffINR 在准确性、通用性和稳定性方面大大提高了磁共振成像重建的效率,为进一步加速磁共振成像采集铺平了道路。值得注意的是,我们提出的框架可用于解决其他医学成像任务中的逆问题。
{"title":"Highly accelerated MRI via implicit neural representation guided posterior sampling of diffusion models","authors":"Jiayue Chu ,&nbsp;Chenhe Du ,&nbsp;Xiyue Lin ,&nbsp;Xiaoqun Zhang ,&nbsp;Lihui Wang ,&nbsp;Yuyao Zhang ,&nbsp;Hongjiang Wei","doi":"10.1016/j.media.2024.103398","DOIUrl":"10.1016/j.media.2024.103398","url":null,"abstract":"<div><div>Reconstructing high-fidelity magnetic resonance (MR) images from under-sampled k-space is a commonly used strategy to reduce scan time. The posterior sampling of diffusion models based on the real measurement data holds significant promise of improved reconstruction accuracy. However, traditional posterior sampling methods often lack effective data consistency guidance, leading to inaccurate and unstable reconstructions. Implicit neural representation (INR) has emerged as a powerful paradigm for solving inverse problems by modeling a signal’s attributes as a continuous function of spatial coordinates. In this study, we present a novel posterior sampler for diffusion models using INR, named DiffINR. The INR-based component incorporates both the diffusion prior distribution and the MRI physical model to ensure high data fidelity. DiffINR demonstrates superior performance on in-distribution datasets with remarkable accuracy, even under high acceleration factors (up to R <span><math><mo>=</mo></math></span> 12 in single-channel reconstruction). Furthermore, DiffINR exhibits excellent generalizability across various tissue contrasts and anatomical structures with low uncertainty. Overall, DiffINR significantly improves MRI reconstruction in terms of accuracy, generalizability and stability, paving the way for further accelerating MRI acquisition. Notably, our proposed framework can be a generalizable framework to solve inverse problems in other medical imaging tasks.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"100 ","pages":"Article 103398"},"PeriodicalIF":10.7,"publicationDate":"2024-11-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142721145","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Noise-aware dynamic image denoising and positron range correction for Rubidium-82 cardiac PET imaging via self-supervision 通过自我监督实现铷-82 心脏正电子发射计算机断层成像的噪声感知动态图像去噪和正电子射程校正。
IF 10.7 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-11-20 DOI: 10.1016/j.media.2024.103391
Huidong Xie , Liang Guo , Alexandre Velo , Zhao Liu , Qiong Liu , Xueqi Guo , Bo Zhou , Xiongchao Chen , Yu-Jung Tsai , Tianshun Miao , Menghua Xia , Yi-Hwa Liu , Ian S. Armstrong , Ge Wang , Richard E. Carson , Albert J. Sinusas , Chi Liu
<div><div>Rubidium-82 (<span><math><mrow><msup><mrow></mrow><mrow><mn>82</mn></mrow></msup><mtext>Rb</mtext></mrow></math></span>) is a radioactive isotope widely used for cardiac PET imaging. Despite numerous benefits of <span><math><mrow><msup><mrow></mrow><mrow><mn>82</mn></mrow></msup><mtext>Rb</mtext></mrow></math></span>, there are several factors that limits its image quality and quantitative accuracy. First, the short half-life of <span><math><mrow><msup><mrow></mrow><mrow><mn>82</mn></mrow></msup><mtext>Rb</mtext></mrow></math></span> results in noisy dynamic frames. Low signal-to-noise ratio would result in inaccurate and biased image quantification. Noisy dynamic frames also lead to highly noisy parametric images. The noise levels also vary substantially in different dynamic frames due to radiotracer decay and short half-life. Existing denoising methods are not applicable for this task due to the lack of paired training inputs/labels and inability to generalize across varying noise levels. Second, <span><math><mrow><msup><mrow></mrow><mrow><mn>82</mn></mrow></msup><mtext>Rb</mtext></mrow></math></span> emits high-energy positrons. Compared with other tracers such as <span><math><mrow><msup><mrow></mrow><mrow><mn>18</mn></mrow></msup><mtext>F</mtext></mrow></math></span>, <span><math><mrow><msup><mrow></mrow><mrow><mn>82</mn></mrow></msup><mtext>Rb</mtext></mrow></math></span> travels a longer distance before annihilation, which negatively affect image spatial resolution. Here, the goal of this study is to propose a self-supervised method for simultaneous (1) noise-aware dynamic image denoising and (2) positron range correction for <span><math><mrow><msup><mrow></mrow><mrow><mn>82</mn></mrow></msup><mtext>Rb</mtext></mrow></math></span> cardiac PET imaging. Tested on a series of PET scans from a cohort of normal volunteers, the proposed method produced images with superior visual quality. To demonstrate the improvement in image quantification, we compared image-derived input functions (IDIFs) with arterial input functions (AIFs) from continuous arterial blood samples. The IDIF derived from the proposed method led to lower AUC differences, decreasing from 11.09<span><math><mtext>%</mtext></math></span> to 7.58<span><math><mtext>%</mtext></math></span> on average, compared to the original dynamic frames. The proposed method also improved the quantification of myocardium blood flow (MBF), as validated against <span><math><mrow><msup><mrow></mrow><mrow><mn>15</mn></mrow></msup><mtext>O-water</mtext></mrow></math></span> scans, with mean MBF differences decreased from 0.43 to 0.09, compared to the original dynamic frames. We also conducted a generalizability experiment on 37 patient scans obtained from a different country using a different scanner. The presented method enhanced defect contrast and resulted in lower regional MBF in areas with perfusion defects. Lastly, comparison with other related methods is included to show the effectivenes
铷-82(82Rb)是一种广泛用于心脏 PET 成像的放射性同位素。尽管 82Rb 有很多优点,但有几个因素限制了它的成像质量和定量准确性。首先,82Rb 的半衰期短,导致动态帧噪声大。低信噪比会导致图像定量不准确和有偏差。动态帧噪声大也会导致参数图像噪声大。由于放射性示踪剂衰变和半衰期短,不同动态帧的噪声水平也有很大差异。由于缺乏成对的训练输入/标签,现有的去噪方法无法在不同的噪声水平下通用,因此不适用于这项任务。其次,82Rb 发射高能正电子。与 18F 等其他示踪剂相比,82Rb 在湮灭前的飞行距离更长,这会对图像的空间分辨率产生负面影响。本研究的目的是提出一种自监督方法,用于同时对 82Rb 心脏 PET 成像进行(1)噪声感知动态图像去噪和(2)正电子射程校正。通过对一系列正常志愿者的正电子发射计算机断层扫描进行测试,所提出的方法生成的图像具有卓越的视觉质量。为了证明图像量化的改进,我们将图像衍生输入函数(IDIF)与来自连续动脉血样本的动脉输入函数(AIF)进行了比较。与原始动态帧相比,由建议方法得出的 IDIF 降低了 AUC 差异,平均从 11.09% 降至 7.58%。经 15O 水扫描验证,提出的方法还改善了心肌血流(MBF)的量化,与原始动态帧相比,MBF 平均差异从 0.43 降至 0.09。我们还对来自不同国家、使用不同扫描仪扫描的 37 名患者进行了通用性实验。所提出的方法增强了缺损对比度,并降低了灌注缺损区域的区域 MBF。最后,我们还与其他相关方法进行了比较,以显示所提方法的有效性。
{"title":"Noise-aware dynamic image denoising and positron range correction for Rubidium-82 cardiac PET imaging via self-supervision","authors":"Huidong Xie ,&nbsp;Liang Guo ,&nbsp;Alexandre Velo ,&nbsp;Zhao Liu ,&nbsp;Qiong Liu ,&nbsp;Xueqi Guo ,&nbsp;Bo Zhou ,&nbsp;Xiongchao Chen ,&nbsp;Yu-Jung Tsai ,&nbsp;Tianshun Miao ,&nbsp;Menghua Xia ,&nbsp;Yi-Hwa Liu ,&nbsp;Ian S. Armstrong ,&nbsp;Ge Wang ,&nbsp;Richard E. Carson ,&nbsp;Albert J. Sinusas ,&nbsp;Chi Liu","doi":"10.1016/j.media.2024.103391","DOIUrl":"10.1016/j.media.2024.103391","url":null,"abstract":"&lt;div&gt;&lt;div&gt;Rubidium-82 (&lt;span&gt;&lt;math&gt;&lt;mrow&gt;&lt;msup&gt;&lt;mrow&gt;&lt;/mrow&gt;&lt;mrow&gt;&lt;mn&gt;82&lt;/mn&gt;&lt;/mrow&gt;&lt;/msup&gt;&lt;mtext&gt;Rb&lt;/mtext&gt;&lt;/mrow&gt;&lt;/math&gt;&lt;/span&gt;) is a radioactive isotope widely used for cardiac PET imaging. Despite numerous benefits of &lt;span&gt;&lt;math&gt;&lt;mrow&gt;&lt;msup&gt;&lt;mrow&gt;&lt;/mrow&gt;&lt;mrow&gt;&lt;mn&gt;82&lt;/mn&gt;&lt;/mrow&gt;&lt;/msup&gt;&lt;mtext&gt;Rb&lt;/mtext&gt;&lt;/mrow&gt;&lt;/math&gt;&lt;/span&gt;, there are several factors that limits its image quality and quantitative accuracy. First, the short half-life of &lt;span&gt;&lt;math&gt;&lt;mrow&gt;&lt;msup&gt;&lt;mrow&gt;&lt;/mrow&gt;&lt;mrow&gt;&lt;mn&gt;82&lt;/mn&gt;&lt;/mrow&gt;&lt;/msup&gt;&lt;mtext&gt;Rb&lt;/mtext&gt;&lt;/mrow&gt;&lt;/math&gt;&lt;/span&gt; results in noisy dynamic frames. Low signal-to-noise ratio would result in inaccurate and biased image quantification. Noisy dynamic frames also lead to highly noisy parametric images. The noise levels also vary substantially in different dynamic frames due to radiotracer decay and short half-life. Existing denoising methods are not applicable for this task due to the lack of paired training inputs/labels and inability to generalize across varying noise levels. Second, &lt;span&gt;&lt;math&gt;&lt;mrow&gt;&lt;msup&gt;&lt;mrow&gt;&lt;/mrow&gt;&lt;mrow&gt;&lt;mn&gt;82&lt;/mn&gt;&lt;/mrow&gt;&lt;/msup&gt;&lt;mtext&gt;Rb&lt;/mtext&gt;&lt;/mrow&gt;&lt;/math&gt;&lt;/span&gt; emits high-energy positrons. Compared with other tracers such as &lt;span&gt;&lt;math&gt;&lt;mrow&gt;&lt;msup&gt;&lt;mrow&gt;&lt;/mrow&gt;&lt;mrow&gt;&lt;mn&gt;18&lt;/mn&gt;&lt;/mrow&gt;&lt;/msup&gt;&lt;mtext&gt;F&lt;/mtext&gt;&lt;/mrow&gt;&lt;/math&gt;&lt;/span&gt;, &lt;span&gt;&lt;math&gt;&lt;mrow&gt;&lt;msup&gt;&lt;mrow&gt;&lt;/mrow&gt;&lt;mrow&gt;&lt;mn&gt;82&lt;/mn&gt;&lt;/mrow&gt;&lt;/msup&gt;&lt;mtext&gt;Rb&lt;/mtext&gt;&lt;/mrow&gt;&lt;/math&gt;&lt;/span&gt; travels a longer distance before annihilation, which negatively affect image spatial resolution. Here, the goal of this study is to propose a self-supervised method for simultaneous (1) noise-aware dynamic image denoising and (2) positron range correction for &lt;span&gt;&lt;math&gt;&lt;mrow&gt;&lt;msup&gt;&lt;mrow&gt;&lt;/mrow&gt;&lt;mrow&gt;&lt;mn&gt;82&lt;/mn&gt;&lt;/mrow&gt;&lt;/msup&gt;&lt;mtext&gt;Rb&lt;/mtext&gt;&lt;/mrow&gt;&lt;/math&gt;&lt;/span&gt; cardiac PET imaging. Tested on a series of PET scans from a cohort of normal volunteers, the proposed method produced images with superior visual quality. To demonstrate the improvement in image quantification, we compared image-derived input functions (IDIFs) with arterial input functions (AIFs) from continuous arterial blood samples. The IDIF derived from the proposed method led to lower AUC differences, decreasing from 11.09&lt;span&gt;&lt;math&gt;&lt;mtext&gt;%&lt;/mtext&gt;&lt;/math&gt;&lt;/span&gt; to 7.58&lt;span&gt;&lt;math&gt;&lt;mtext&gt;%&lt;/mtext&gt;&lt;/math&gt;&lt;/span&gt; on average, compared to the original dynamic frames. The proposed method also improved the quantification of myocardium blood flow (MBF), as validated against &lt;span&gt;&lt;math&gt;&lt;mrow&gt;&lt;msup&gt;&lt;mrow&gt;&lt;/mrow&gt;&lt;mrow&gt;&lt;mn&gt;15&lt;/mn&gt;&lt;/mrow&gt;&lt;/msup&gt;&lt;mtext&gt;O-water&lt;/mtext&gt;&lt;/mrow&gt;&lt;/math&gt;&lt;/span&gt; scans, with mean MBF differences decreased from 0.43 to 0.09, compared to the original dynamic frames. We also conducted a generalizability experiment on 37 patient scans obtained from a different country using a different scanner. The presented method enhanced defect contrast and resulted in lower regional MBF in areas with perfusion defects. Lastly, comparison with other related methods is included to show the effectivenes","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"100 ","pages":"Article 103391"},"PeriodicalIF":10.7,"publicationDate":"2024-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142695596","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DDKG: A Dual Domain Knowledge Guidance strategy for localization and diagnosis of non-displaced femoral neck fractures DDKG:用于非脱位股骨颈骨折定位和诊断的双领域知识指导策略
IF 10.7 1区 医学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-11-19 DOI: 10.1016/j.media.2024.103393
Jing Yang , Lianxin Wang , Chen Lin , Jiacheng Wang , Liansheng Wang
X-ray is the primary tool for diagnosing fractures, crucial for determining their type, location, and severity. However, non-displaced femoral neck fractures (ND-FNF) can pose challenges in identification due to subtle cracks and complex anatomical structures. Most deep learning-based methods for diagnosing ND-FNF rely on cropped images, necessitating manual annotation of the hip location, which increases annotation costs. To address this challenge, we propose Dual Domain Knowledge Guidance (DDKG), which harnesses spatial and semantic domain knowledge to guide the model in acquiring robust representations of ND-FNF across the whole X-ray image. Specifically, DDKG comprises two key modules: the Spatial Aware Module (SAM) and the Semantic Coordination Module (SCM). SAM employs limited positional supervision to guide the model in focusing on the hip joint region and reducing background interference. SCM integrates information from radiological reports, utilizes prior knowledge from large language models to extract critical information related to ND-FNF, and guides the model to learn relevant visual representations. During inference, the model only requires the whole X-ray image for accurate diagnosis without additional information. The model was validated on datasets from four different centers, showing consistent accuracy and robustness. Codes and models are available at https://github.com/Yjing07/DDKG.
X 射线是诊断骨折的主要工具,对于确定骨折的类型、位置和严重程度至关重要。然而,由于存在细微的裂缝和复杂的解剖结构,非移位股骨颈骨折(ND-FNF)的识别面临挑战。大多数基于深度学习的 ND-FNF 诊断方法都依赖于裁剪图像,需要人工标注髋关节位置,这增加了标注成本。为了应对这一挑战,我们提出了双领域知识指导(DDKG),利用空间和语义领域知识指导模型在整个 X 光图像中获取 ND-FNF 的稳健表示。具体来说,DDKG 包括两个关键模块:空间感知模块(SAM)和语义协调模块(SCM)。空间感知模块采用有限的位置监督来引导模型聚焦于髋关节区域并减少背景干扰。SCM 整合了放射报告中的信息,利用大型语言模型中的先验知识提取与 ND-FNF 相关的关键信息,并引导模型学习相关的视觉表征。在推理过程中,该模型只需要整个 X 光图像就能准确诊断,无需额外信息。该模型在四个不同中心的数据集上进行了验证,显示出一致的准确性和鲁棒性。代码和模型可在 https://github.com/Yjing07/DDKG 上获取。
{"title":"DDKG: A Dual Domain Knowledge Guidance strategy for localization and diagnosis of non-displaced femoral neck fractures","authors":"Jing Yang ,&nbsp;Lianxin Wang ,&nbsp;Chen Lin ,&nbsp;Jiacheng Wang ,&nbsp;Liansheng Wang","doi":"10.1016/j.media.2024.103393","DOIUrl":"10.1016/j.media.2024.103393","url":null,"abstract":"<div><div>X-ray is the primary tool for diagnosing fractures, crucial for determining their type, location, and severity. However, non-displaced femoral neck fractures (ND-FNF) can pose challenges in identification due to subtle cracks and complex anatomical structures. Most deep learning-based methods for diagnosing ND-FNF rely on cropped images, necessitating manual annotation of the hip location, which increases annotation costs. To address this challenge, we propose Dual Domain Knowledge Guidance (DDKG), which harnesses spatial and semantic domain knowledge to guide the model in acquiring robust representations of ND-FNF across the whole X-ray image. Specifically, DDKG comprises two key modules: the Spatial Aware Module (SAM) and the Semantic Coordination Module (SCM). SAM employs limited positional supervision to guide the model in focusing on the hip joint region and reducing background interference. SCM integrates information from radiological reports, utilizes prior knowledge from large language models to extract critical information related to ND-FNF, and guides the model to learn relevant visual representations. During inference, the model only requires the whole X-ray image for accurate diagnosis without additional information. The model was validated on datasets from four different centers, showing consistent accuracy and robustness. Codes and models are available at <span><span>https://github.com/Yjing07/DDKG</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"100 ","pages":"Article 103393"},"PeriodicalIF":10.7,"publicationDate":"2024-11-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142696321","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Medical image analysis
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1