首页 > 最新文献

Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention最新文献

英文 中文
A Conditional Flow Variational Autoencoder for Controllable Synthesis of Virtual Populations of Anatomy 一种用于解剖学虚拟种群可控合成的条件流变分自编码器
Haoran Dou, N. Ravikumar, Alejandro F Frangi
The generation of virtual populations (VPs) of anatomy is essential for conducting in silico trials of medical devices. Typically, the generated VP should capture sufficient variability while remaining plausible and should reflect the specific characteristics and demographics of the patients observed in real populations. In several applications, it is desirable to synthesise virtual populations in a textit{controlled} manner, where relevant covariates are used to conditionally synthesise virtual populations that fit a specific target population/characteristics. We propose to equip a conditional variational autoencoder (cVAE) with normalising flows to boost the flexibility and complexity of the approximate posterior learnt, leading to enhanced flexibility for controllable synthesis of VPs of anatomical structures. We demonstrate the performance of our conditional flow VAE using a data set of cardiac left ventricles acquired from 2360 patients, with associated demographic information and clinical measurements (used as covariates/conditional information). The results obtained indicate the superiority of the proposed method for conditional synthesis of virtual populations of cardiac left ventricles relative to a cVAE. Conditional synthesis performance was evaluated in terms of generalisation and specificity errors and in terms of the ability to preserve clinically relevant biomarkers in synthesised VPs, that is, the left ventricular blood pool and myocardial volume, relative to the real observed population.
生成解剖学的虚拟种群(VPs)对于进行医疗设备的计算机试验至关重要。通常,生成的VP应在保持可信的同时捕获足够的变异性,并应反映在实际人群中观察到的患者的具体特征和人口统计学特征。在一些应用中,希望以textit{受控}的方式合成虚拟种群,其中使用相关协变量有条件地合成适合特定目标种群/特征的虚拟种群。我们建议为条件变分自编码器(cVAE)配备归一化流,以提高近似后验学习的灵活性和复杂性,从而提高解剖结构VPs可控合成的灵活性。我们使用从2360名患者中获得的左心室数据集,以及相关的人口统计信息和临床测量(用作协变量/条件信息)来证明条件血流VAE的性能。所获得的结果表明,相对于cVAE,所提出的方法具有条件合成左心室虚拟种群的优越性。条件合成性能的评估依据是泛化和特异性误差,以及在合成的VPs中保留临床相关生物标志物的能力,即相对于实际观察人群的左室血池和心肌体积。
{"title":"A Conditional Flow Variational Autoencoder for Controllable Synthesis of Virtual Populations of Anatomy","authors":"Haoran Dou, N. Ravikumar, Alejandro F Frangi","doi":"10.48550/arXiv.2306.14680","DOIUrl":"https://doi.org/10.48550/arXiv.2306.14680","url":null,"abstract":"The generation of virtual populations (VPs) of anatomy is essential for conducting in silico trials of medical devices. Typically, the generated VP should capture sufficient variability while remaining plausible and should reflect the specific characteristics and demographics of the patients observed in real populations. In several applications, it is desirable to synthesise virtual populations in a textit{controlled} manner, where relevant covariates are used to conditionally synthesise virtual populations that fit a specific target population/characteristics. We propose to equip a conditional variational autoencoder (cVAE) with normalising flows to boost the flexibility and complexity of the approximate posterior learnt, leading to enhanced flexibility for controllable synthesis of VPs of anatomical structures. We demonstrate the performance of our conditional flow VAE using a data set of cardiac left ventricles acquired from 2360 patients, with associated demographic information and clinical measurements (used as covariates/conditional information). The results obtained indicate the superiority of the proposed method for conditional synthesis of virtual populations of cardiac left ventricles relative to a cVAE. Conditional synthesis performance was evaluated in terms of generalisation and specificity errors and in terms of the ability to preserve clinically relevant biomarkers in synthesised VPs, that is, the left ventricular blood pool and myocardial volume, relative to the real observed population.","PeriodicalId":18289,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"62 8 1","pages":"143-152"},"PeriodicalIF":0.0,"publicationDate":"2023-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80813124","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
AME-CAM: Attentive Multiple-Exit CAM for Weakly Supervised Segmentation on MRI Brain Tumor AME-CAM:弱监督分割MRI脑肿瘤的细心多出口CAM
Yu-Jen Chen, Xinrong Hu, Yi Shi, Tsung-Yi Ho
Magnetic resonance imaging (MRI) is commonly used for brain tumor segmentation, which is critical for patient evaluation and treatment planning. To reduce the labor and expertise required for labeling, weakly-supervised semantic segmentation (WSSS) methods with class activation mapping (CAM) have been proposed. However, existing CAM methods suffer from low resolution due to strided convolution and pooling layers, resulting in inaccurate predictions. In this study, we propose a novel CAM method, Attentive Multiple-Exit CAM (AME-CAM), that extracts activation maps from multiple resolutions to hierarchically aggregate and improve prediction accuracy. We evaluate our method on the BraTS 2021 dataset and show that it outperforms state-of-the-art methods.
磁共振成像(MRI)是常用的脑肿瘤分割技术,对患者评估和治疗计划至关重要。为了减少标注所需的人力和专业知识,提出了基于类激活映射(CAM)的弱监督语义分割方法。然而,现有的CAM方法由于存在跨行卷积和池化层,导致分辨率较低,导致预测不准确。在本研究中,我们提出了一种新的CAM方法——细心多出口CAM (AME-CAM),该方法从多个分辨率中提取激活图,分层聚合,提高预测精度。我们在BraTS 2021数据集上评估了我们的方法,并表明它优于最先进的方法。
{"title":"AME-CAM: Attentive Multiple-Exit CAM for Weakly Supervised Segmentation on MRI Brain Tumor","authors":"Yu-Jen Chen, Xinrong Hu, Yi Shi, Tsung-Yi Ho","doi":"10.48550/arXiv.2306.14505","DOIUrl":"https://doi.org/10.48550/arXiv.2306.14505","url":null,"abstract":"Magnetic resonance imaging (MRI) is commonly used for brain tumor segmentation, which is critical for patient evaluation and treatment planning. To reduce the labor and expertise required for labeling, weakly-supervised semantic segmentation (WSSS) methods with class activation mapping (CAM) have been proposed. However, existing CAM methods suffer from low resolution due to strided convolution and pooling layers, resulting in inaccurate predictions. In this study, we propose a novel CAM method, Attentive Multiple-Exit CAM (AME-CAM), that extracts activation maps from multiple resolutions to hierarchically aggregate and improve prediction accuracy. We evaluate our method on the BraTS 2021 dataset and show that it outperforms state-of-the-art methods.","PeriodicalId":18289,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"8 1","pages":"173-182"},"PeriodicalIF":0.0,"publicationDate":"2023-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91025864","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
FeSViBS: Federated Split Learning of Vision Transformer with Block Sampling FeSViBS:基于块采样的视觉变压器的联邦分割学习
Faris Almalik, Naif Alkhunaizi, Ibrahim Almakky, K. Nandakumar
Data scarcity is a significant obstacle hindering the learning of powerful machine learning models in critical healthcare applications. Data-sharing mechanisms among multiple entities (e.g., hospitals) can accelerate model training and yield more accurate predictions. Recently, approaches such as Federated Learning (FL) and Split Learning (SL) have facilitated collaboration without the need to exchange private data. In this work, we propose a framework for medical imaging classification tasks called Federated Split learning of Vision transformer with Block Sampling (FeSViBS). The FeSViBS framework builds upon the existing federated split vision transformer and introduces a block sampling module, which leverages intermediate features extracted by the Vision Transformer (ViT) at the server. This is achieved by sampling features (patch tokens) from an intermediate transformer block and distilling their information content into a pseudo class token before passing them back to the client. These pseudo class tokens serve as an effective feature augmentation strategy and enhances the generalizability of the learned model. We demonstrate the utility of our proposed method compared to other SL and FL approaches on three publicly available medical imaging datasets: HAM1000, BloodMNIST, and Fed-ISIC2019, under both IID and non-IID settings. Code: https://github.com/faresmalik/FeSViBS
数据稀缺是阻碍在关键医疗保健应用中学习强大机器学习模型的一个重要障碍。多个实体(如医院)之间的数据共享机制可以加速模型训练并产生更准确的预测。最近,联邦学习(FL)和分裂学习(SL)等方法促进了协作,而无需交换私有数据。在这项工作中,我们提出了一个医学成像分类任务的框架,称为联邦分割学习视觉变压器与块采样(FeSViBS)。FeSViBS框架建立在现有的联邦分割视觉转换器的基础上,并引入了一个块采样模块,该模块利用了服务器上视觉转换器(ViT)提取的中间特征。这是通过从中间变压器块采样特征(补丁令牌)并在将其传递回客户端之前将其信息内容提取到伪类令牌中来实现的。这些伪类标记作为一种有效的特征增强策略,增强了学习模型的可泛化性。在IID和非IID设置下,我们展示了与其他SL和FL方法相比,我们提出的方法在三个公开可用的医学成像数据集(HAM1000、BloodMNIST和Fed-ISIC2019)上的实用性。代码:https://github.com/faresmalik/FeSViBS
{"title":"FeSViBS: Federated Split Learning of Vision Transformer with Block Sampling","authors":"Faris Almalik, Naif Alkhunaizi, Ibrahim Almakky, K. Nandakumar","doi":"10.48550/arXiv.2306.14638","DOIUrl":"https://doi.org/10.48550/arXiv.2306.14638","url":null,"abstract":"Data scarcity is a significant obstacle hindering the learning of powerful machine learning models in critical healthcare applications. Data-sharing mechanisms among multiple entities (e.g., hospitals) can accelerate model training and yield more accurate predictions. Recently, approaches such as Federated Learning (FL) and Split Learning (SL) have facilitated collaboration without the need to exchange private data. In this work, we propose a framework for medical imaging classification tasks called Federated Split learning of Vision transformer with Block Sampling (FeSViBS). The FeSViBS framework builds upon the existing federated split vision transformer and introduces a block sampling module, which leverages intermediate features extracted by the Vision Transformer (ViT) at the server. This is achieved by sampling features (patch tokens) from an intermediate transformer block and distilling their information content into a pseudo class token before passing them back to the client. These pseudo class tokens serve as an effective feature augmentation strategy and enhances the generalizability of the learned model. We demonstrate the utility of our proposed method compared to other SL and FL approaches on three publicly available medical imaging datasets: HAM1000, BloodMNIST, and Fed-ISIC2019, under both IID and non-IID settings. Code: https://github.com/faresmalik/FeSViBS","PeriodicalId":18289,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"64 1","pages":"350-360"},"PeriodicalIF":0.0,"publicationDate":"2023-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87463940","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TCEIP: Text Condition Embedded Regression Network for Dental Implant Position Prediction 牙种植体位置预测的文本条件嵌入回归网络
Xinquan Yang, Jinheng Xie, Xuguang Li, Xuechen Li, X. Li, Linlin Shen, Yongqiang Deng
When deep neural network has been proposed to assist the dentist in designing the location of dental implant, most of them are targeting simple cases where only one missing tooth is available. As a result, literature works do not work well when there are multiple missing teeth and easily generate false predictions when the teeth are sparsely distributed. In this paper, we are trying to integrate a weak supervision text, the target region, to the implant position regression network, to address above issues. We propose a text condition embedded implant position regression network (TCEIP), to embed the text condition into the encoder-decoder framework for improvement of the regression performance. A cross-modal interaction that consists of cross-modal attention (CMA) and knowledge alignment module (KAM) is proposed to facilitate the interaction between features of images and texts. The CMA module performs a cross-attention between the image feature and the text condition, and the KAM mitigates the knowledge gap between the image feature and the image encoder of the CLIP. Extensive experiments on a dental implant dataset through five-fold cross-validation demonstrated that the proposed TCEIP achieves superior performance than existing methods.
当深度神经网络被提出来帮助牙医设计种植体的位置时,大多数都是针对只有一颗缺牙的简单病例。因此,当有多个缺失牙齿时,文学作品不能很好地工作,当牙齿稀疏分布时,文学作品容易产生错误的预测。在本文中,我们试图将弱监督文本,目标区域,整合到植入位置回归网络中,以解决上述问题。我们提出了一种文本条件嵌入植入位置回归网络(TCEIP),将文本条件嵌入到编码器-解码器框架中,以提高回归性能。为了促进图像和文本特征之间的交互,提出了一种由跨模态注意(CMA)和知识对齐模块(KAM)组成的跨模态交互方法。CMA模块在图像特征和文本条件之间进行交叉关注,KAM模块减轻了图像特征和CLIP图像编码器之间的知识差距。在牙科种植体数据集上进行的大量实验通过五倍交叉验证表明,所提出的TCEIP比现有方法具有更好的性能。
{"title":"TCEIP: Text Condition Embedded Regression Network for Dental Implant Position Prediction","authors":"Xinquan Yang, Jinheng Xie, Xuguang Li, Xuechen Li, X. Li, Linlin Shen, Yongqiang Deng","doi":"10.48550/arXiv.2306.14406","DOIUrl":"https://doi.org/10.48550/arXiv.2306.14406","url":null,"abstract":"When deep neural network has been proposed to assist the dentist in designing the location of dental implant, most of them are targeting simple cases where only one missing tooth is available. As a result, literature works do not work well when there are multiple missing teeth and easily generate false predictions when the teeth are sparsely distributed. In this paper, we are trying to integrate a weak supervision text, the target region, to the implant position regression network, to address above issues. We propose a text condition embedded implant position regression network (TCEIP), to embed the text condition into the encoder-decoder framework for improvement of the regression performance. A cross-modal interaction that consists of cross-modal attention (CMA) and knowledge alignment module (KAM) is proposed to facilitate the interaction between features of images and texts. The CMA module performs a cross-attention between the image feature and the text condition, and the KAM mitigates the knowledge gap between the image feature and the image encoder of the CLIP. Extensive experiments on a dental implant dataset through five-fold cross-validation demonstrated that the proposed TCEIP achieves superior performance than existing methods.","PeriodicalId":18289,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"112 1","pages":"317-326"},"PeriodicalIF":0.0,"publicationDate":"2023-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90390230","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A denoised Mean Teacher for domain adaptive point cloud registration 域自适应点云配准的去噪均值教师算法
Alexander Bigalke, M. Heinrich
Point cloud-based medical registration promises increased computational efficiency, robustness to intensity shifts, and anonymity preservation but is limited by the inefficacy of unsupervised learning with similarity metrics. Supervised training on synthetic deformations is an alternative but, in turn, suffers from the domain gap to the real domain. In this work, we aim to tackle this gap through domain adaptation. Self-training with the Mean Teacher is an established approach to this problem but is impaired by the inherent noise of the pseudo labels from the teacher. As a remedy, we present a denoised teacher-student paradigm for point cloud registration, comprising two complementary denoising strategies. First, we propose to filter pseudo labels based on the Chamfer distances of teacher and student registrations, thus preventing detrimental supervision by the teacher. Second, we make the teacher dynamically synthesize novel training pairs with noise-free labels by warping its moving inputs with the predicted deformations. Evaluation is performed for inhale-to-exhale registration of lung vessel trees on the public PVT dataset under two domain shifts. Our method surpasses the baseline Mean Teacher by 13.5/62.8%, consistently outperforms diverse competitors, and sets a new state-of-the-art accuracy (TRE=2.31mm). Code is available at https://github.com/multimodallearning/denoised_mt_pcd_reg.
基于点云的医疗注册有望提高计算效率,对强度变化的鲁棒性和匿名性保持,但受无监督学习与相似度量的无效限制。对合成变形的监督训练是另一种选择,但反过来又受到与真实域的域差距的影响。在这项工作中,我们的目标是通过领域适应来解决这一差距。与“刻薄的老师”一起进行自我训练是解决这个问题的一种既定方法,但受到来自老师的伪标签的固有噪声的影响。作为补救措施,我们提出了一个去噪的师生模式点云配准,包括两个互补的去噪策略。首先,我们建议根据教师和学生注册的Chamfer距离来过滤伪标签,从而防止教师的有害监督。其次,我们通过用预测的变形来扭曲其运动输入,使教师动态地合成具有无噪声标签的新训练对。在两个域移位下,对公共PVT数据集上的肺血管树的吸气到呼气注册进行了评估。我们的方法比基线Mean Teacher高出13.5/62.8%,始终优于各种竞争对手,并设定了新的最先进的精度(TRE=2.31mm)。代码可从https://github.com/multimodallearning/denoised_mt_pcd_reg获得。
{"title":"A denoised Mean Teacher for domain adaptive point cloud registration","authors":"Alexander Bigalke, M. Heinrich","doi":"10.48550/arXiv.2306.14749","DOIUrl":"https://doi.org/10.48550/arXiv.2306.14749","url":null,"abstract":"Point cloud-based medical registration promises increased computational efficiency, robustness to intensity shifts, and anonymity preservation but is limited by the inefficacy of unsupervised learning with similarity metrics. Supervised training on synthetic deformations is an alternative but, in turn, suffers from the domain gap to the real domain. In this work, we aim to tackle this gap through domain adaptation. Self-training with the Mean Teacher is an established approach to this problem but is impaired by the inherent noise of the pseudo labels from the teacher. As a remedy, we present a denoised teacher-student paradigm for point cloud registration, comprising two complementary denoising strategies. First, we propose to filter pseudo labels based on the Chamfer distances of teacher and student registrations, thus preventing detrimental supervision by the teacher. Second, we make the teacher dynamically synthesize novel training pairs with noise-free labels by warping its moving inputs with the predicted deformations. Evaluation is performed for inhale-to-exhale registration of lung vessel trees on the public PVT dataset under two domain shifts. Our method surpasses the baseline Mean Teacher by 13.5/62.8%, consistently outperforms diverse competitors, and sets a new state-of-the-art accuracy (TRE=2.31mm). Code is available at https://github.com/multimodallearning/denoised_mt_pcd_reg.","PeriodicalId":18289,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"129 1","pages":"666-676"},"PeriodicalIF":0.0,"publicationDate":"2023-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82925660","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Histopathology Image Classification using Deep Manifold Contrastive Learning 基于深度流形对比学习的组织病理学图像分类
J. Tan, Won-Ki Jeong
Contrastive learning has gained popularity due to its robustness with good feature representation performance. However, cosine distance, the commonly used similarity metric in contrastive learning, is not well suited to represent the distance between two data points, especially on a nonlinear feature manifold. Inspired by manifold learning, we propose a novel extension of contrastive learning that leverages geodesic distance between features as a similarity metric for histopathology whole slide image classification. To reduce the computational overhead in manifold learning, we propose geodesic-distance-based feature clustering for efficient contrastive loss evaluation using prototypes without time-consuming pairwise feature similarity comparison. The efficacy of the proposed method is evaluated on two real-world histopathology image datasets. Results demonstrate that our method outperforms state-of-the-art cosine-distance-based contrastive learning methods.
对比学习以其鲁棒性和良好的特征表示性能而受到广泛的关注。然而,余弦距离是对比学习中常用的相似度度量,并不适合表示两个数据点之间的距离,特别是在非线性特征流形上。受流形学习的启发,我们提出了一种新的对比学习扩展,利用特征之间的测地线距离作为组织病理学全幻灯片图像分类的相似性度量。为了减少流形学习的计算开销,我们提出了基于测地线距离的特征聚类,使用原型进行有效的对比损失评估,而不需要耗时的两两特征相似性比较。在两个真实世界的组织病理学图像数据集上评估了所提出方法的有效性。结果表明,我们的方法优于最先进的基于余弦距离的对比学习方法。
{"title":"Histopathology Image Classification using Deep Manifold Contrastive Learning","authors":"J. Tan, Won-Ki Jeong","doi":"10.48550/arXiv.2306.14459","DOIUrl":"https://doi.org/10.48550/arXiv.2306.14459","url":null,"abstract":"Contrastive learning has gained popularity due to its robustness with good feature representation performance. However, cosine distance, the commonly used similarity metric in contrastive learning, is not well suited to represent the distance between two data points, especially on a nonlinear feature manifold. Inspired by manifold learning, we propose a novel extension of contrastive learning that leverages geodesic distance between features as a similarity metric for histopathology whole slide image classification. To reduce the computational overhead in manifold learning, we propose geodesic-distance-based feature clustering for efficient contrastive loss evaluation using prototypes without time-consuming pairwise feature similarity comparison. The efficacy of the proposed method is evaluated on two real-world histopathology image datasets. Results demonstrate that our method outperforms state-of-the-art cosine-distance-based contrastive learning methods.","PeriodicalId":18289,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"30 1","pages":"683-692"},"PeriodicalIF":0.0,"publicationDate":"2023-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76653705","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CDiffMR: Can We Replace the Gaussian Noise with K-Space Undersampling for Fast MRI? CDiffMR:能否用k空间欠采样代替高斯噪声用于快速MRI?
Jiahao Huang, Angelica I. Avilés-Rivero, C. Schönlieb, Guang Yang
Deep learning has shown the capability to substantially accelerate MRI reconstruction while acquiring fewer measurements. Recently, diffusion models have gained burgeoning interests as a novel group of deep learning-based generative methods. These methods seek to sample data points that belong to a target distribution from a Gaussian distribution, which has been successfully extended to MRI reconstruction. In this work, we proposed a Cold Diffusion-based MRI reconstruction method called CDiffMR. Different from conventional diffusion models, the degradation operation of our CDiffMR is based on textit{k}-space undersampling instead of adding Gaussian noise, and the restoration network is trained to harness a de-aliaseing function. We also design starting point and data consistency conditioning strategies to guide and accelerate the reverse process. More intriguingly, the pre-trained CDiffMR model can be reused for reconstruction tasks with different undersampling rates. We demonstrated, through extensive numerical and visual experiments, that the proposed CDiffMR can achieve comparable or even superior reconstruction results than state-of-the-art models. Compared to the diffusion model-based counterpart, CDiffMR reaches readily competing results using only $1.6 sim 3.4%$ for inference time. The code is publicly available at https://github.com/ayanglab/CDiffMR.
深度学习已经显示出大大加快MRI重建的能力,同时获得更少的测量值。近年来,扩散模型作为一种新的基于深度学习的生成方法得到了广泛的关注。这些方法寻求从高斯分布中采样属于目标分布的数据点,这已经成功地扩展到MRI重建中。在这项工作中,我们提出了一种基于冷扩散的MRI重建方法,称为CDiffMR。与传统的扩散模型不同,CDiffMR的退化操作是基于textit{k}空间欠采样而不是添加高斯噪声,并且恢复网络被训练成利用去混叠函数。我们还设计了起点和数据一致性调节策略来指导和加速反向过程。更有趣的是,预训练的CDiffMR模型可以重复用于不同欠采样率的重建任务。通过广泛的数值和视觉实验,我们证明了所提出的CDiffMR可以达到与最先进的模型相当甚至更好的重建结果。与基于扩散模型的对应模型相比,CDiffMR仅使用$1.6 sim 3.4%$进行推理时间就可以获得容易竞争的结果。该代码可在https://github.com/ayanglab/CDiffMR上公开获得。
{"title":"CDiffMR: Can We Replace the Gaussian Noise with K-Space Undersampling for Fast MRI?","authors":"Jiahao Huang, Angelica I. Avilés-Rivero, C. Schönlieb, Guang Yang","doi":"10.48550/arXiv.2306.14350","DOIUrl":"https://doi.org/10.48550/arXiv.2306.14350","url":null,"abstract":"Deep learning has shown the capability to substantially accelerate MRI reconstruction while acquiring fewer measurements. Recently, diffusion models have gained burgeoning interests as a novel group of deep learning-based generative methods. These methods seek to sample data points that belong to a target distribution from a Gaussian distribution, which has been successfully extended to MRI reconstruction. In this work, we proposed a Cold Diffusion-based MRI reconstruction method called CDiffMR. Different from conventional diffusion models, the degradation operation of our CDiffMR is based on textit{k}-space undersampling instead of adding Gaussian noise, and the restoration network is trained to harness a de-aliaseing function. We also design starting point and data consistency conditioning strategies to guide and accelerate the reverse process. More intriguingly, the pre-trained CDiffMR model can be reused for reconstruction tasks with different undersampling rates. We demonstrated, through extensive numerical and visual experiments, that the proposed CDiffMR can achieve comparable or even superior reconstruction results than state-of-the-art models. Compared to the diffusion model-based counterpart, CDiffMR reaches readily competing results using only $1.6 sim 3.4%$ for inference time. The code is publicly available at https://github.com/ayanglab/CDiffMR.","PeriodicalId":18289,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"10 1","pages":"3-12"},"PeriodicalIF":0.0,"publicationDate":"2023-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79174098","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Regular SE(3) Group Convolutions for Volumetric Medical Image Analysis 正则SE(3)群卷积用于体积医学图像分析
T. Kuipers, E. Bekkers
Regular group convolutional neural networks (G-CNNs) have been shown to increase model performance and improve equivariance to different geometrical symmetries. This work addresses the problem of SE(3), i.e., roto-translation equivariance, on volumetric data. Volumetric image data is prevalent in many medical settings. Motivated by the recent work on separable group convolutions, we devise a SE(3) group convolution kernel separated into a continuous SO(3) (rotation) kernel and a spatial kernel. We approximate equivariance to the continuous setting by sampling uniform SO(3) grids. Our continuous SO(3) kernel is parameterized via RBF interpolation on similarly uniform grids. We demonstrate the advantages of our approach in volumetric medical image analysis. Our SE(3) equivariant models consistently outperform CNNs and regular discrete G-CNNs on challenging medical classification tasks and show significantly improved generalization capabilities. Our approach achieves up to a 16.5% gain in accuracy over regular CNNs.
正则群卷积神经网络(g - cnn)已被证明可以提高模型性能并改善对不同几何对称的等方差。这项工作解决了体积数据上的SE(3)问题,即旋转平移等方差。体积图像数据在许多医疗环境中很普遍。受最近关于可分离群卷积研究的启发,我们设计了一个SE(3)群卷积核,它分为连续SO(3)(旋转)核和空间核。我们通过采样均匀的SO(3)网格来近似连续设置的等方差。我们的连续SO(3)核通过RBF插值在相似的均匀网格上参数化。我们证明了我们的方法在体积医学图像分析中的优势。我们的SE(3)等变模型在具有挑战性的医学分类任务上始终优于cnn和常规离散g - cnn,并显示出显著提高的泛化能力。我们的方法比常规cnn的准确率提高了16.5%。
{"title":"Regular SE(3) Group Convolutions for Volumetric Medical Image Analysis","authors":"T. Kuipers, E. Bekkers","doi":"10.48550/arXiv.2306.13960","DOIUrl":"https://doi.org/10.48550/arXiv.2306.13960","url":null,"abstract":"Regular group convolutional neural networks (G-CNNs) have been shown to increase model performance and improve equivariance to different geometrical symmetries. This work addresses the problem of SE(3), i.e., roto-translation equivariance, on volumetric data. Volumetric image data is prevalent in many medical settings. Motivated by the recent work on separable group convolutions, we devise a SE(3) group convolution kernel separated into a continuous SO(3) (rotation) kernel and a spatial kernel. We approximate equivariance to the continuous setting by sampling uniform SO(3) grids. Our continuous SO(3) kernel is parameterized via RBF interpolation on similarly uniform grids. We demonstrate the advantages of our approach in volumetric medical image analysis. Our SE(3) equivariant models consistently outperform CNNs and regular discrete G-CNNs on challenging medical classification tasks and show significantly improved generalization capabilities. Our approach achieves up to a 16.5% gain in accuracy over regular CNNs.","PeriodicalId":18289,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"33 1","pages":"252-261"},"PeriodicalIF":0.0,"publicationDate":"2023-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79090356","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
DiffuseIR: Diffusion Models For Isotropic Reconstruction of 3D Microscopic Images 扩散ir:用于三维显微图像各向同性重建的扩散模型
Mingjie Pan, Yulu Gan, Fangxu Zhou, Jiaming Liu, Aimin Wang, Shanghang Zhang, Dawei Li
Three-dimensional microscopy is often limited by anisotropic spatial resolution, resulting in lower axial resolution than lateral resolution. Current State-of-The-Art (SoTA) isotropic reconstruction methods utilizing deep neural networks can achieve impressive super-resolution performance in fixed imaging settings. However, their generality in practical use is limited by degraded performance caused by artifacts and blurring when facing unseen anisotropic factors. To address these issues, we propose DiffuseIR, an unsupervised method for isotropic reconstruction based on diffusion models. First, we pre-train a diffusion model to learn the structural distribution of biological tissue from lateral microscopic images, resulting in generating naturally high-resolution images. Then we use low-axial-resolution microscopy images to condition the generation process of the diffusion model and generate high-axial-resolution reconstruction results. Since the diffusion model learns the universal structural distribution of biological tissues, which is independent of the axial resolution, DiffuseIR can reconstruct authentic images with unseen low-axial resolutions into a high-axial resolution without requiring re-training. The proposed DiffuseIR achieves SoTA performance in experiments on EM data and can even compete with supervised methods.
三维显微镜通常受到各向异性空间分辨率的限制,导致轴向分辨率低于横向分辨率。目前最先进的(SoTA)各向同性重建方法利用深度神经网络可以在固定成像设置中实现令人印象深刻的超分辨率性能。然而,它们在实际使用中的通用性受到在面对看不见的各向异性因素时由伪影和模糊引起的性能下降的限制。为了解决这些问题,我们提出了DiffuseIR,一种基于扩散模型的无监督各向同性重建方法。首先,我们预训练扩散模型,从横向显微图像中学习生物组织的结构分布,从而生成自然的高分辨率图像。然后利用低轴向分辨率的显微图像来调节扩散模型的生成过程,生成高轴向分辨率的重建结果。由于扩散模型学习了生物组织的普遍结构分布,与轴向分辨率无关,因此DiffuseIR可以将未见过的低轴向分辨率的真实图像重建为高轴向分辨率的图像,而无需重新训练。本文提出的扩散红外方法在EM数据实验中达到了SoTA的性能,甚至可以与监督方法相媲美。
{"title":"DiffuseIR: Diffusion Models For Isotropic Reconstruction of 3D Microscopic Images","authors":"Mingjie Pan, Yulu Gan, Fangxu Zhou, Jiaming Liu, Aimin Wang, Shanghang Zhang, Dawei Li","doi":"10.48550/arXiv.2306.12109","DOIUrl":"https://doi.org/10.48550/arXiv.2306.12109","url":null,"abstract":"Three-dimensional microscopy is often limited by anisotropic spatial resolution, resulting in lower axial resolution than lateral resolution. Current State-of-The-Art (SoTA) isotropic reconstruction methods utilizing deep neural networks can achieve impressive super-resolution performance in fixed imaging settings. However, their generality in practical use is limited by degraded performance caused by artifacts and blurring when facing unseen anisotropic factors. To address these issues, we propose DiffuseIR, an unsupervised method for isotropic reconstruction based on diffusion models. First, we pre-train a diffusion model to learn the structural distribution of biological tissue from lateral microscopic images, resulting in generating naturally high-resolution images. Then we use low-axial-resolution microscopy images to condition the generation process of the diffusion model and generate high-axial-resolution reconstruction results. Since the diffusion model learns the universal structural distribution of biological tissues, which is independent of the axial resolution, DiffuseIR can reconstruct authentic images with unseen low-axial resolutions into a high-axial resolution without requiring re-training. The proposed DiffuseIR achieves SoTA performance in experiments on EM data and can even compete with supervised methods.","PeriodicalId":18289,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"14 1","pages":"323-332"},"PeriodicalIF":0.0,"publicationDate":"2023-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89778535","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Reliable and Interpretable Framework of Multi-view Learning for Liver Fibrosis Staging 肝纤维化分期的可靠和可解释的多视角学习框架
Zheyao Gao, Yuanye Liu, Fuping Wu, N. Shi, Yuxin Shi, X. Zhuang
Staging of liver fibrosis is important in the diagnosis and treatment planning of patients suffering from liver diseases. Current deep learning-based methods using abdominal magnetic resonance imaging (MRI) usually take a sub-region of the liver as an input, which nevertheless could miss critical information. To explore richer representations, we formulate this task as a multi-view learning problem and employ multiple sub-regions of the liver. Previously, features or predictions are usually combined in an implicit manner, and uncertainty-aware methods have been proposed. However, these methods could be challenged to capture cross-view representations, which can be important in the accurate prediction of staging. Therefore, we propose a reliable multi-view learning method with interpretable combination rules, which can model global representations to improve the accuracy of predictions. Specifically, the proposed method estimates uncertainties based on subjective logic to improve reliability, and an explicit combination rule is applied based on Dempster-Shafer's evidence theory with good power of interpretability. Moreover, a data-efficient transformer is introduced to capture representations in the global view. Results evaluated on enhanced MRI data show that our method delivers superior performance over existing multi-view learning methods.
肝纤维化分期在肝病患者的诊断和治疗计划中具有重要意义。目前使用腹部磁共振成像(MRI)的基于深度学习的方法通常将肝脏的一个亚区域作为输入,然而这可能会遗漏关键信息。为了探索更丰富的表征,我们将此任务制定为一个多视图学习问题,并使用肝脏的多个子区域。以前,特征或预测通常以隐式方式组合,并提出了不确定性感知方法。然而,这些方法在捕获交叉视图表示方面可能会受到挑战,这对于准确预测分期很重要。因此,我们提出了一种可靠的多视图学习方法,该方法具有可解释的组合规则,可以对全局表示进行建模,以提高预测的准确性。具体而言,该方法基于主观逻辑估计不确定性,提高了可靠性,并基于Dempster-Shafer证据理论采用显式组合规则,具有良好的可解释性。此外,还引入了一个数据高效的转换器来捕获全局视图中的表示。对增强MRI数据的评估结果表明,我们的方法比现有的多视图学习方法具有更好的性能。
{"title":"A Reliable and Interpretable Framework of Multi-view Learning for Liver Fibrosis Staging","authors":"Zheyao Gao, Yuanye Liu, Fuping Wu, N. Shi, Yuxin Shi, X. Zhuang","doi":"10.48550/arXiv.2306.12054","DOIUrl":"https://doi.org/10.48550/arXiv.2306.12054","url":null,"abstract":"Staging of liver fibrosis is important in the diagnosis and treatment planning of patients suffering from liver diseases. Current deep learning-based methods using abdominal magnetic resonance imaging (MRI) usually take a sub-region of the liver as an input, which nevertheless could miss critical information. To explore richer representations, we formulate this task as a multi-view learning problem and employ multiple sub-regions of the liver. Previously, features or predictions are usually combined in an implicit manner, and uncertainty-aware methods have been proposed. However, these methods could be challenged to capture cross-view representations, which can be important in the accurate prediction of staging. Therefore, we propose a reliable multi-view learning method with interpretable combination rules, which can model global representations to improve the accuracy of predictions. Specifically, the proposed method estimates uncertainties based on subjective logic to improve reliability, and an explicit combination rule is applied based on Dempster-Shafer's evidence theory with good power of interpretability. Moreover, a data-efficient transformer is introduced to capture representations in the global view. Results evaluated on enhanced MRI data show that our method delivers superior performance over existing multi-view learning methods.","PeriodicalId":18289,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"16 1","pages":"178-188"},"PeriodicalIF":0.0,"publicationDate":"2023-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77486057","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1