首页 > 最新文献

Information processing in medical imaging : proceedings of the ... conference最新文献

英文 中文
Don't PANIC: Prototypical Additive Neural Network for Interpretable Classification of Alzheimer's Disease 不要惊慌:用于阿尔茨海默病可解释分类的原型加法神经网络
Pub Date : 2023-03-13 DOI: 10.48550/arXiv.2303.07125
Thomas Wolf, Sebastian Pölsterl, C. Wachinger
Alzheimer's disease (AD) has a complex and multifactorial etiology, which requires integrating information about neuroanatomy, genetics, and cerebrospinal fluid biomarkers for accurate diagnosis. Hence, recent deep learning approaches combined image and tabular information to improve diagnostic performance. However, the black-box nature of such neural networks is still a barrier for clinical applications, in which understanding the decision of a heterogeneous model is integral. We propose PANIC, a prototypical additive neural network for interpretable AD classification that integrates 3D image and tabular data. It is interpretable by design and, thus, avoids the need for post-hoc explanations that try to approximate the decision of a network. Our results demonstrate that PANIC achieves state-of-the-art performance in AD classification, while directly providing local and global explanations. Finally, we show that PANIC extracts biologically meaningful signatures of AD, and satisfies a set of desirable desiderata for trustworthy machine learning. Our implementation is available at https://github.com/ai-med/PANIC .
阿尔茨海默病(AD)具有复杂的多因素病因,需要整合神经解剖学、遗传学和脑脊液生物标志物的信息才能准确诊断。因此,最近的深度学习方法结合了图像和表格信息来提高诊断性能。然而,这种神经网络的黑箱性质仍然是临床应用的障碍,在临床应用中,理解异构模型的决策是不可或缺的。我们提出了PANIC,一个典型的用于可解释AD分类的加性神经网络,它集成了3D图像和表格数据。它可以通过设计来解释,因此,避免了试图近似网络决策的事后解释的需要。我们的研究结果表明,PANIC在AD分类中达到了最先进的性能,同时直接提供了局部和全局解释。最后,我们证明了PANIC提取了AD的生物学意义签名,并满足了一组值得信赖的机器学习的理想需求。我们的实现可以在https://github.com/ai-med/PANIC上获得。
{"title":"Don't PANIC: Prototypical Additive Neural Network for Interpretable Classification of Alzheimer's Disease","authors":"Thomas Wolf, Sebastian Pölsterl, C. Wachinger","doi":"10.48550/arXiv.2303.07125","DOIUrl":"https://doi.org/10.48550/arXiv.2303.07125","url":null,"abstract":"Alzheimer's disease (AD) has a complex and multifactorial etiology, which requires integrating information about neuroanatomy, genetics, and cerebrospinal fluid biomarkers for accurate diagnosis. Hence, recent deep learning approaches combined image and tabular information to improve diagnostic performance. However, the black-box nature of such neural networks is still a barrier for clinical applications, in which understanding the decision of a heterogeneous model is integral. We propose PANIC, a prototypical additive neural network for interpretable AD classification that integrates 3D image and tabular data. It is interpretable by design and, thus, avoids the need for post-hoc explanations that try to approximate the decision of a network. Our results demonstrate that PANIC achieves state-of-the-art performance in AD classification, while directly providing local and global explanations. Finally, we show that PANIC extracts biologically meaningful signatures of AD, and satisfies a set of desirable desiderata for trustworthy machine learning. Our implementation is available at https://github.com/ai-med/PANIC .","PeriodicalId":73379,"journal":{"name":"Information processing in medical imaging : proceedings of the ... conference","volume":"28 1","pages":"82-94"},"PeriodicalIF":0.0,"publicationDate":"2023-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76432496","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Token Sparsification for Faster Medical Image Segmentation 用于快速医学图像分割的标记稀疏化
Pub Date : 2023-03-11 DOI: 10.48550/arXiv.2303.06522
Lei Zhou, Huidong Liu, Joseph Bae, Junjun He, D. Samaras, P. Prasanna
Can we use sparse tokens for dense prediction, e.g., segmentation? Although token sparsification has been applied to Vision Transformers (ViT) to accelerate classification, it is still unknown how to perform segmentation from sparse tokens. To this end, we reformulate segmentation as a sparse encoding ->token completion ->dense decoding (SCD) pipeline. We first empirically show that naively applying existing approaches from classification token pruning and masked image modeling (MIM) leads to failure and inefficient training caused by inappropriate sampling algorithms and the low quality of the restored dense features. In this paper, we propose Soft-topK Token Pruning (STP) and Multi-layer Token Assembly (MTA) to address these problems. In sparse encoding, STP predicts token importance scores with a lightweight sub-network and samples the topK tokens. The intractable topK gradients are approximated through a continuous perturbed score distribution. In token completion, MTA restores a full token sequence by assembling both sparse output tokens and pruned multi-layer intermediate ones. The last dense decoding stage is compatible with existing segmentation decoders, e.g., UNETR. Experiments show SCD pipelines equipped with STP and MTA are much faster than baselines without token pruning in both training (up to 120% higher throughput and inference up to 60.6% higher throughput) while maintaining segmentation quality.
我们可以使用稀疏标记进行密集预测,例如分割吗?尽管标记稀疏化已被应用于视觉变形器(Vision transformer, ViT)来加速分类,但如何从稀疏标记中进行分割仍然是一个未知的问题。为此,我们将分割重新制定为稀疏编码->令牌补全->密集解码(SCD)管道。我们首先通过经验证明,天真地应用分类令牌修剪和掩蔽图像建模(MIM)等现有方法会导致采样算法不合适以及恢复的密集特征质量低,从而导致训练失败和低效。在本文中,我们提出软顶令牌修剪(STP)和多层令牌组装(MTA)来解决这些问题。在稀疏编码中,STP使用轻量级子网络预测令牌重要性分数,并对topK令牌进行采样。顽固性topK梯度通过连续扰动分数分布近似。在令牌补全中,MTA通过组合稀疏的输出令牌和经过修剪的多层中间令牌来恢复完整的令牌序列。最后一个密集解码阶段与现有的分割解码器兼容,例如UNETR。实验表明,在保持分割质量的同时,配备STP和MTA的SCD管道在训练(高达120%的吞吐量和高达60.6%的吞吐量)上都比没有令牌修剪的基线快得多。
{"title":"Token Sparsification for Faster Medical Image Segmentation","authors":"Lei Zhou, Huidong Liu, Joseph Bae, Junjun He, D. Samaras, P. Prasanna","doi":"10.48550/arXiv.2303.06522","DOIUrl":"https://doi.org/10.48550/arXiv.2303.06522","url":null,"abstract":"Can we use sparse tokens for dense prediction, e.g., segmentation? Although token sparsification has been applied to Vision Transformers (ViT) to accelerate classification, it is still unknown how to perform segmentation from sparse tokens. To this end, we reformulate segmentation as a sparse encoding ->token completion ->dense decoding (SCD) pipeline. We first empirically show that naively applying existing approaches from classification token pruning and masked image modeling (MIM) leads to failure and inefficient training caused by inappropriate sampling algorithms and the low quality of the restored dense features. In this paper, we propose Soft-topK Token Pruning (STP) and Multi-layer Token Assembly (MTA) to address these problems. In sparse encoding, STP predicts token importance scores with a lightweight sub-network and samples the topK tokens. The intractable topK gradients are approximated through a continuous perturbed score distribution. In token completion, MTA restores a full token sequence by assembling both sparse output tokens and pruned multi-layer intermediate ones. The last dense decoding stage is compatible with existing segmentation decoders, e.g., UNETR. Experiments show SCD pipelines equipped with STP and MTA are much faster than baselines without token pruning in both training (up to 120% higher throughput and inference up to 60.6% higher throughput) while maintaining segmentation quality.","PeriodicalId":73379,"journal":{"name":"Information processing in medical imaging : proceedings of the ... conference","volume":"19 1","pages":"743-754"},"PeriodicalIF":0.0,"publicationDate":"2023-03-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84304658","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Resolving quantitative MRI model degeneracy with machine learning via training data distribution design 通过训练数据分布设计,用机器学习解决定量MRI模型退化问题
Pub Date : 2023-03-09 DOI: 10.48550/arXiv.2303.05464
Michele Guerreri, Sean C. Epstein, H. Azadbakht, Hui Zhang
Quantitative MRI (qMRI) aims to map tissue properties non-invasively via models that relate these unknown quantities to measured MRI signals. Estimating these unknowns, which has traditionally required model fitting - an often iterative procedure, can now be done with one-shot machine learning (ML) approaches. Such parameter estimation may be complicated by intrinsic qMRI signal model degeneracy: different combinations of tissue properties produce the same signal. Despite their many advantages, it remains unclear whether ML approaches can resolve this issue. Growing empirical evidence appears to suggest ML approaches remain susceptible to model degeneracy. Here we demonstrate under the right circumstances ML can address this issue. Inspired by recent works on the impact of training data distributions on ML-based parameter estimation, we propose to resolve model degeneracy by designing training data distributions. We put forward a classification of model degeneracies and identify one particular kind of degeneracies amenable to the proposed attack. The strategy is demonstrated successfully using the Revised NODDI model with standard multi-shell diffusion MRI data as an exemplar. Our results illustrate the importance of training set design which has the potential to allow accurate estimation of tissue properties with ML.
定量MRI (qMRI)旨在通过将这些未知量与测量的MRI信号相关联的模型,非侵入性地绘制组织特性。估计这些未知数,传统上需要模型拟合,这通常是一个迭代的过程,现在可以用一次性机器学习(ML)方法来完成。这种参数估计可能会因qMRI信号模型的内在退化而变得复杂:组织特性的不同组合产生相同的信号。尽管ML方法有很多优点,但它是否能解决这个问题还不清楚。越来越多的经验证据似乎表明ML方法仍然容易受到模型退化的影响。这里我们演示了在适当的情况下ML可以解决这个问题。受最近研究训练数据分布对基于ml的参数估计影响的工作的启发,我们提出通过设计训练数据分布来解决模型退化问题。我们提出了一种模型退化的分类方法,并确定了一种适合这种攻击的模型退化。以标准多壳扩散MRI数据为例,对修正的NODDI模型进行了验证。我们的结果说明了训练集设计的重要性,它有可能允许用ML准确估计组织特性。
{"title":"Resolving quantitative MRI model degeneracy with machine learning via training data distribution design","authors":"Michele Guerreri, Sean C. Epstein, H. Azadbakht, Hui Zhang","doi":"10.48550/arXiv.2303.05464","DOIUrl":"https://doi.org/10.48550/arXiv.2303.05464","url":null,"abstract":"Quantitative MRI (qMRI) aims to map tissue properties non-invasively via models that relate these unknown quantities to measured MRI signals. Estimating these unknowns, which has traditionally required model fitting - an often iterative procedure, can now be done with one-shot machine learning (ML) approaches. Such parameter estimation may be complicated by intrinsic qMRI signal model degeneracy: different combinations of tissue properties produce the same signal. Despite their many advantages, it remains unclear whether ML approaches can resolve this issue. Growing empirical evidence appears to suggest ML approaches remain susceptible to model degeneracy. Here we demonstrate under the right circumstances ML can address this issue. Inspired by recent works on the impact of training data distributions on ML-based parameter estimation, we propose to resolve model degeneracy by designing training data distributions. We put forward a classification of model degeneracies and identify one particular kind of degeneracies amenable to the proposed attack. The strategy is demonstrated successfully using the Revised NODDI model with standard multi-shell diffusion MRI data as an exemplar. Our results illustrate the importance of training set design which has the potential to allow accurate estimation of tissue properties with ML.","PeriodicalId":73379,"journal":{"name":"Information processing in medical imaging : proceedings of the ... conference","volume":"3 1","pages":"3-14"},"PeriodicalIF":0.0,"publicationDate":"2023-03-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86424637","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MetaMorph: Learning Metamorphic Image Transformation With Appearance Changes 变形:学习变形图像的外观变化
Pub Date : 2023-03-08 DOI: 10.48550/arXiv.2303.04849
Jian Wang, Jiarui Xing, J.T. Druzgal, W. Wells, Miaomiao Zhang
This paper presents a novel predictive model, MetaMorph, for metamorphic registration of images with appearance changes (i.e., caused by brain tumors). In contrast to previous learning-based registration methods that have little or no control over appearance-changes, our model introduces a new regularization that can effectively suppress the negative effects of appearance changing areas. In particular, we develop a piecewise regularization on the tangent space of diffeomorphic transformations (also known as initial velocity fields) via learned segmentation maps of abnormal regions. The geometric transformation and appearance changes are treated as joint tasks that are mutually beneficial. Our model MetaMorph is more robust and accurate when searching for an optimal registration solution under the guidance of segmentation, which in turn improves the segmentation performance by providing appropriately augmented training labels. We validate MetaMorph on real 3D human brain tumor magnetic resonance imaging (MRI) scans. Experimental results show that our model outperforms the state-of-the-art learning-based registration models. The proposed MetaMorph has great potential in various image-guided clinical interventions, e.g., real-time image-guided navigation systems for tumor removal surgery.
本文提出了一种新的预测模型MetaMorph,用于具有外观变化(即由脑肿瘤引起)的图像的变质配准。与之前基于学习的注册方法很少或根本不控制外观变化相比,我们的模型引入了一种新的正则化方法,可以有效地抑制外观变化区域的负面影响。特别是,我们通过学习异常区域的分割映射,在微分同构变换的切空间(也称为初始速度场)上开发了分段正则化。几何变换和外观变化被视为互惠互利的联合任务。我们的模型MetaMorph在分割指导下寻找最优配准解时更加鲁棒和准确,从而通过提供适当的增强训练标签来提高分割性能。我们在真实的三维人类脑肿瘤磁共振成像(MRI)扫描上验证了MetaMorph。实验结果表明,我们的模型优于目前最先进的基于学习的配准模型。所提出的MetaMorph在各种图像引导的临床干预中具有很大的潜力,例如用于肿瘤切除手术的实时图像引导导航系统。
{"title":"MetaMorph: Learning Metamorphic Image Transformation With Appearance Changes","authors":"Jian Wang, Jiarui Xing, J.T. Druzgal, W. Wells, Miaomiao Zhang","doi":"10.48550/arXiv.2303.04849","DOIUrl":"https://doi.org/10.48550/arXiv.2303.04849","url":null,"abstract":"This paper presents a novel predictive model, MetaMorph, for metamorphic registration of images with appearance changes (i.e., caused by brain tumors). In contrast to previous learning-based registration methods that have little or no control over appearance-changes, our model introduces a new regularization that can effectively suppress the negative effects of appearance changing areas. In particular, we develop a piecewise regularization on the tangent space of diffeomorphic transformations (also known as initial velocity fields) via learned segmentation maps of abnormal regions. The geometric transformation and appearance changes are treated as joint tasks that are mutually beneficial. Our model MetaMorph is more robust and accurate when searching for an optimal registration solution under the guidance of segmentation, which in turn improves the segmentation performance by providing appropriately augmented training labels. We validate MetaMorph on real 3D human brain tumor magnetic resonance imaging (MRI) scans. Experimental results show that our model outperforms the state-of-the-art learning-based registration models. The proposed MetaMorph has great potential in various image-guided clinical interventions, e.g., real-time image-guided navigation systems for tumor removal surgery.","PeriodicalId":73379,"journal":{"name":"Information processing in medical imaging : proceedings of the ... conference","volume":"67 1","pages":"576-587"},"PeriodicalIF":0.0,"publicationDate":"2023-03-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83589864","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Improved Segmentation of Deep Sulci in Cortical Gray Matter Using a Deep Learning Framework Incorporating Laplace's Equation 基于拉普拉斯方程的深度学习框架改进皮层灰质深沟分割
Pub Date : 2023-03-01 DOI: 10.48550/arXiv.2303.00795
S. Ravikumar, Ranjit Itttyerah, Sydney A. Lim, L. Xie, Sandhitsu R. Das, Pulkit Khandelwal, L. Wisse, M. Bedard, John L. Robinson, Terry K. Schuck, M. Grossman, J. Trojanowski, Eddie B. Lee, M. Tisdall, K. Prabhakaran, J. Detre, D. Irwin, Winifred Trotman, G. Mizsei, Emilio Artacho-P'erula, Maria Mercedes Iniguez de Onzono Martin, Maria del Mar Arroyo Jim'enez, M. Muñoz, Francisco Javier Molina Romero, M. Rabal, Sandra Cebada-S'anchez, J. Gonz'alez, C. Rosa-Prieto, Marta Córcoles Parada, D. Wolk, R. Insausti, Paul Yushkevich
When developing tools for automated cortical segmentation, the ability to produce topologically correct segmentations is important in order to compute geometrically valid morphometry measures. In practice, accurate cortical segmentation is challenged by image artifacts and the highly convoluted anatomy of the cortex itself. To address this, we propose a novel deep learning-based cortical segmentation method in which prior knowledge about the geometry of the cortex is incorporated into the network during the training process. We design a loss function which uses the theory of Laplace's equation applied to the cortex to locally penalize unresolved boundaries between tightly folded sulci. Using an ex vivo MRI dataset of human medial temporal lobe specimens, we demonstrate that our approach outperforms baseline segmentation networks, both quantitatively and qualitatively.
在开发自动皮层分割工具时,为了计算几何上有效的形态测量,生成拓扑正确分割的能力是很重要的。在实践中,准确的皮层分割受到图像伪影和皮层本身高度复杂的解剖结构的挑战。为了解决这个问题,我们提出了一种新的基于深度学习的皮层分割方法,该方法在训练过程中将皮层几何形状的先验知识整合到网络中。我们设计了一个损失函数,该函数使用拉普拉斯方程理论应用于皮层,以局部惩罚紧密折叠沟之间未解决的边界。使用人类内侧颞叶标本的离体MRI数据集,我们证明了我们的方法在定量和定性上都优于基线分割网络。
{"title":"Improved Segmentation of Deep Sulci in Cortical Gray Matter Using a Deep Learning Framework Incorporating Laplace's Equation","authors":"S. Ravikumar, Ranjit Itttyerah, Sydney A. Lim, L. Xie, Sandhitsu R. Das, Pulkit Khandelwal, L. Wisse, M. Bedard, John L. Robinson, Terry K. Schuck, M. Grossman, J. Trojanowski, Eddie B. Lee, M. Tisdall, K. Prabhakaran, J. Detre, D. Irwin, Winifred Trotman, G. Mizsei, Emilio Artacho-P'erula, Maria Mercedes Iniguez de Onzono Martin, Maria del Mar Arroyo Jim'enez, M. Muñoz, Francisco Javier Molina Romero, M. Rabal, Sandra Cebada-S'anchez, J. Gonz'alez, C. Rosa-Prieto, Marta Córcoles Parada, D. Wolk, R. Insausti, Paul Yushkevich","doi":"10.48550/arXiv.2303.00795","DOIUrl":"https://doi.org/10.48550/arXiv.2303.00795","url":null,"abstract":"When developing tools for automated cortical segmentation, the ability to produce topologically correct segmentations is important in order to compute geometrically valid morphometry measures. In practice, accurate cortical segmentation is challenged by image artifacts and the highly convoluted anatomy of the cortex itself. To address this, we propose a novel deep learning-based cortical segmentation method in which prior knowledge about the geometry of the cortex is incorporated into the network during the training process. We design a loss function which uses the theory of Laplace's equation applied to the cortex to locally penalize unresolved boundaries between tightly folded sulci. Using an ex vivo MRI dataset of human medial temporal lobe specimens, we demonstrate that our approach outperforms baseline segmentation networks, both quantitatively and qualitatively.","PeriodicalId":73379,"journal":{"name":"Information processing in medical imaging : proceedings of the ... conference","volume":"146 1","pages":"692-704"},"PeriodicalIF":0.0,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77647104","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
X-TRA: Improving Chest X-ray Tasks with Cross-Modal Retrieval Augmentation X-TRA:改进胸部x线任务与跨模态检索增强
Pub Date : 2023-02-22 DOI: 10.48550/arXiv.2302.11352
Tom van Sonsbeek, M. Worring
An important component of human analysis of medical images and their context is the ability to relate newly seen things to related instances in our memory. In this paper we mimic this ability by using multi-modal retrieval augmentation and apply it to several tasks in chest X-ray analysis. By retrieving similar images and/or radiology reports we expand and regularize the case at hand with additional knowledge, while maintaining factual knowledge consistency. The method consists of two components. First, vision and language modalities are aligned using a pre-trained CLIP model. To enforce that the retrieval focus will be on detailed disease-related content instead of global visual appearance it is fine-tuned using disease class information. Subsequently, we construct a non-parametric retrieval index, which reaches state-of-the-art retrieval levels. We use this index in our downstream tasks to augment image representations through multi-head attention for disease classification and report retrieval. We show that retrieval augmentation gives considerable improvements on these tasks. Our downstream report retrieval even shows to be competitive with dedicated report generation methods, paving the path for this method in medical imaging.
人类分析医学图像及其背景的一个重要组成部分是将新看到的事物与我们记忆中的相关实例联系起来的能力。在本文中,我们通过使用多模态检索增强来模拟这种能力,并将其应用于胸部x射线分析中的几个任务。通过检索相似的图像和/或放射学报告,我们用额外的知识扩展和规范手边的病例,同时保持事实知识的一致性。该方法由两个部分组成。首先,使用预训练的CLIP模型对齐视觉和语言模式。为了确保检索重点将放在与疾病相关的详细内容上,而不是全局视觉外观上,它使用疾病类别信息进行了微调。随后,我们构建了一个非参数检索索引,该索引达到了最先进的检索水平。我们在下游任务中使用该索引,通过多头关注来增强图像表示,用于疾病分类和报告检索。我们表明,检索增强在这些任务上提供了相当大的改进。我们的下游报告检索甚至显示出与专用报告生成方法的竞争力,为该方法在医学成像中铺平了道路。
{"title":"X-TRA: Improving Chest X-ray Tasks with Cross-Modal Retrieval Augmentation","authors":"Tom van Sonsbeek, M. Worring","doi":"10.48550/arXiv.2302.11352","DOIUrl":"https://doi.org/10.48550/arXiv.2302.11352","url":null,"abstract":"An important component of human analysis of medical images and their context is the ability to relate newly seen things to related instances in our memory. In this paper we mimic this ability by using multi-modal retrieval augmentation and apply it to several tasks in chest X-ray analysis. By retrieving similar images and/or radiology reports we expand and regularize the case at hand with additional knowledge, while maintaining factual knowledge consistency. The method consists of two components. First, vision and language modalities are aligned using a pre-trained CLIP model. To enforce that the retrieval focus will be on detailed disease-related content instead of global visual appearance it is fine-tuned using disease class information. Subsequently, we construct a non-parametric retrieval index, which reaches state-of-the-art retrieval levels. We use this index in our downstream tasks to augment image representations through multi-head attention for disease classification and report retrieval. We show that retrieval augmentation gives considerable improvements on these tasks. Our downstream report retrieval even shows to be competitive with dedicated report generation methods, paving the path for this method in medical imaging.","PeriodicalId":73379,"journal":{"name":"Information processing in medical imaging : proceedings of the ... conference","volume":"21 1","pages":"471-482"},"PeriodicalIF":0.0,"publicationDate":"2023-02-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84151625","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Non-rigid Medical Image Registration using Physics-informed Neural Networks 使用物理信息神经网络的非刚性医学图像配准
Pub Date : 2023-02-20 DOI: 10.48550/arXiv.2302.10343
Z. Min, Zachary Michael Cieman Baum, Shaheer U. Saeed, M. Emberton, D. Barratt, Z. Taylor, Yipeng Hu
Biomechanical modelling of soft tissue provides a non-data-driven method for constraining medical image registration, such that the estimated spatial transformation is considered biophysically plausible. This has not only been adopted in real-world clinical applications, such as the MR-to-ultrasound registration for prostate intervention of interest in this work, but also provides an explainable means of understanding the organ motion and spatial correspondence establishment. This work instantiates the recently-proposed physics-informed neural networks (PINNs) to a 3D linear elastic model for modelling prostate motion commonly encountered during transrectal ultrasound guided procedures. To overcome a widely-recognised challenge in generalising PINNs to different subjects, we propose to use PointNet as the nodal-permutation-invariant feature extractor, together with a registration algorithm that aligns point sets and simultaneously takes into account the PINN-imposed biomechanics. The proposed method has been both developed and validated in both patient-specific and multi-patient manner.
软组织的生物力学建模提供了一种非数据驱动的方法来约束医学图像配准,这样估计的空间变换被认为是生物物理上可信的。这不仅在现实世界的临床应用中被采用,例如本研究中感兴趣的前列腺介入的MR-to-ultrasound registration,而且还提供了一种理解器官运动和空间对应建立的可解释的方法。这项工作将最近提出的物理信息神经网络(pinn)实例化为三维线性弹性模型,用于模拟经直肠超声引导过程中常见的前列腺运动。为了克服将pinn推广到不同主题的广泛公认的挑战,我们建议使用PointNet作为节点置换不变特征提取器,以及对齐点集并同时考虑pinn施加的生物力学的配准算法。所提出的方法已在患者特异性和多患者方式中开发和验证。
{"title":"Non-rigid Medical Image Registration using Physics-informed Neural Networks","authors":"Z. Min, Zachary Michael Cieman Baum, Shaheer U. Saeed, M. Emberton, D. Barratt, Z. Taylor, Yipeng Hu","doi":"10.48550/arXiv.2302.10343","DOIUrl":"https://doi.org/10.48550/arXiv.2302.10343","url":null,"abstract":"Biomechanical modelling of soft tissue provides a non-data-driven method for constraining medical image registration, such that the estimated spatial transformation is considered biophysically plausible. This has not only been adopted in real-world clinical applications, such as the MR-to-ultrasound registration for prostate intervention of interest in this work, but also provides an explainable means of understanding the organ motion and spatial correspondence establishment. This work instantiates the recently-proposed physics-informed neural networks (PINNs) to a 3D linear elastic model for modelling prostate motion commonly encountered during transrectal ultrasound guided procedures. To overcome a widely-recognised challenge in generalising PINNs to different subjects, we propose to use PointNet as the nodal-permutation-invariant feature extractor, together with a registration algorithm that aligns point sets and simultaneously takes into account the PINN-imposed biomechanics. The proposed method has been both developed and validated in both patient-specific and multi-patient manner.","PeriodicalId":73379,"journal":{"name":"Information processing in medical imaging : proceedings of the ... conference","volume":"23 1","pages":"601-613"},"PeriodicalIF":0.0,"publicationDate":"2023-02-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87288866","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Heterogeneous Graph Convolutional Neural Network via Hodge-Laplacian for Brain Functional Data 基于Hodge-Laplacian的脑功能数据异构图卷积神经网络
Pub Date : 2023-02-18 DOI: 10.48550/arXiv.2302.09323
Jinghan Huang, M. Chung, A. Qiu
This study proposes a novel heterogeneous graph convolutional neural network (HGCNN) to handle complex brain fMRI data at regional and across-region levels. We introduce a generic formulation of spectral filters on heterogeneous graphs by introducing the $k-th$ Hodge-Laplacian (HL) operator. In particular, we propose Laguerre polynomial approximations of HL spectral filters and prove that their spatial localization on graphs is related to the polynomial order. Furthermore, based on the bijection property of boundary operators on simplex graphs, we introduce a generic topological graph pooling (TGPool) method that can be used at any dimensional simplices. This study designs HL-node, HL-edge, and HL-HGCNN neural networks to learn signal representation at a graph node, edge levels, and both, respectively. Our experiments employ fMRI from the Adolescent Brain Cognitive Development (ABCD; n=7693) to predict general intelligence. Our results demonstrate the advantage of the HL-edge network over the HL-node network when functional brain connectivity is considered as features. The HL-HGCNN outperforms the state-of-the-art graph neural networks (GNNs) approaches, such as GAT, BrainGNN, dGCN, BrainNetCNN, and Hypergraph NN. The functional connectivity features learned from the HL-HGCNN are meaningful in interpreting neural circuits related to general intelligence.
本研究提出了一种新的异构图卷积神经网络(HGCNN)来处理区域和跨区域水平的复杂脑功能磁共振数据。通过引入第k阶霍奇-拉普拉斯算子,给出了异质图上光谱滤波器的一般公式。特别地,我们提出了HL光谱滤波器的Laguerre多项式近似,并证明了它们在图上的空间定位与多项式阶数有关。在此基础上,基于单纯形图边界算子的双射性质,提出了一种适用于任意维度单纯图的通用拓扑图池化方法(TGPool)。本研究设计了HL-node、HL-edge和HL-HGCNN神经网络,分别在图节点、边水平和两者上学习信号表示。我们的实验使用了青少年大脑认知发展(ABCD;N =7693)来预测一般智力。我们的研究结果表明,当脑功能连接被视为特征时,HL-edge网络优于HL-node网络。HL-HGCNN优于GAT、BrainGNN、dGCN、BrainNetCNN和Hypergraph NN等最先进的图神经网络(gnn)方法。从HL-HGCNN中学习到的功能连接特征对于解释与一般智能相关的神经回路具有重要意义。
{"title":"Heterogeneous Graph Convolutional Neural Network via Hodge-Laplacian for Brain Functional Data","authors":"Jinghan Huang, M. Chung, A. Qiu","doi":"10.48550/arXiv.2302.09323","DOIUrl":"https://doi.org/10.48550/arXiv.2302.09323","url":null,"abstract":"This study proposes a novel heterogeneous graph convolutional neural network (HGCNN) to handle complex brain fMRI data at regional and across-region levels. We introduce a generic formulation of spectral filters on heterogeneous graphs by introducing the $k-th$ Hodge-Laplacian (HL) operator. In particular, we propose Laguerre polynomial approximations of HL spectral filters and prove that their spatial localization on graphs is related to the polynomial order. Furthermore, based on the bijection property of boundary operators on simplex graphs, we introduce a generic topological graph pooling (TGPool) method that can be used at any dimensional simplices. This study designs HL-node, HL-edge, and HL-HGCNN neural networks to learn signal representation at a graph node, edge levels, and both, respectively. Our experiments employ fMRI from the Adolescent Brain Cognitive Development (ABCD; n=7693) to predict general intelligence. Our results demonstrate the advantage of the HL-edge network over the HL-node network when functional brain connectivity is considered as features. The HL-HGCNN outperforms the state-of-the-art graph neural networks (GNNs) approaches, such as GAT, BrainGNN, dGCN, BrainNetCNN, and Hypergraph NN. The functional connectivity features learned from the HL-HGCNN are meaningful in interpreting neural circuits related to general intelligence.","PeriodicalId":73379,"journal":{"name":"Information processing in medical imaging : proceedings of the ... conference","volume":"65 1","pages":"278-290"},"PeriodicalIF":0.0,"publicationDate":"2023-02-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80907025","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Fast-MC-PET: A Novel Deep Learning-aided Motion Correction and Reconstruction Framework for Accelerated PET Fast-MC-PET:一种新的加速PET的深度学习辅助运动校正和重建框架
Pub Date : 2023-02-14 DOI: 10.48550/arXiv.2302.07135
Bo Zhou, Yu-Jung Tsai, Jiazhen Zhang, Xueqi Guo, Huidong Xie, Xiongchao Chen, T. Miao, Yihuan Lu, J. Duncan, Chi Liu
Patient motion during PET is inevitable. Its long acquisition time not only increases the motion and the associated artifacts but also the patient's discomfort, thus PET acceleration is desirable. However, accelerating PET acquisition will result in reconstructed images with low SNR, and the image quality will still be degraded by motion-induced artifacts. Most of the previous PET motion correction methods are motion type specific that require motion modeling, thus may fail when multiple types of motion present together. Also, those methods are customized for standard long acquisition and could not be directly applied to accelerated PET. To this end, modeling-free universal motion correction reconstruction for accelerated PET is still highly under-explored. In this work, we propose a novel deep learning-aided motion correction and reconstruction framework for accelerated PET, called Fast-MC-PET. Our framework consists of a universal motion correction (UMC) and a short-to-long acquisition reconstruction (SL-Reon) module. The UMC enables modeling-free motion correction by estimating quasi-continuous motion from ultra-short frame reconstructions and using this information for motion-compensated reconstruction. Then, the SL-Recon converts the accelerated UMC image with low counts to a high-quality image with high counts for our final reconstruction output. Our experimental results on human studies show that our Fast-MC-PET can enable 7-fold acceleration and use only 2 minutes acquisition to generate high-quality reconstruction images that outperform/match previous motion correction reconstruction methods using standard 15 minutes long acquisition data.
患者在PET期间的运动是不可避免的。它的长采集时间不仅增加了运动和相关的伪影,而且增加了患者的不适感,因此PET加速是可取的。然而,加速PET采集将导致重建图像信噪比较低,并且图像质量仍然会因运动引起的伪影而下降。以前的PET运动校正方法大多是特定于运动类型的,需要运动建模,因此当多种类型的运动同时存在时可能会失败。此外,这些方法是为标准长采集定制的,不能直接应用于加速PET。为此,加速PET的无建模通用运动校正重建还有待进一步探索。在这项工作中,我们提出了一种新的用于加速PET的深度学习辅助运动校正和重建框架,称为Fast-MC-PET。我们的框架由一个通用运动校正(UMC)和一个短到长采集重建(SL-Reon)模块组成。UMC通过从超短帧重建中估计准连续运动,并将此信息用于运动补偿重建,从而实现无建模运动校正。然后,SL-Recon将具有低计数的加速UMC图像转换为具有高计数的高质量图像,用于我们的最终重建输出。我们在人体研究上的实验结果表明,我们的Fast-MC-PET可以实现7倍的加速,只需2分钟的采集就可以生成高质量的重建图像,优于/匹配以前使用标准15分钟采集数据的运动校正重建方法。
{"title":"Fast-MC-PET: A Novel Deep Learning-aided Motion Correction and Reconstruction Framework for Accelerated PET","authors":"Bo Zhou, Yu-Jung Tsai, Jiazhen Zhang, Xueqi Guo, Huidong Xie, Xiongchao Chen, T. Miao, Yihuan Lu, J. Duncan, Chi Liu","doi":"10.48550/arXiv.2302.07135","DOIUrl":"https://doi.org/10.48550/arXiv.2302.07135","url":null,"abstract":"Patient motion during PET is inevitable. Its long acquisition time not only increases the motion and the associated artifacts but also the patient's discomfort, thus PET acceleration is desirable. However, accelerating PET acquisition will result in reconstructed images with low SNR, and the image quality will still be degraded by motion-induced artifacts. Most of the previous PET motion correction methods are motion type specific that require motion modeling, thus may fail when multiple types of motion present together. Also, those methods are customized for standard long acquisition and could not be directly applied to accelerated PET. To this end, modeling-free universal motion correction reconstruction for accelerated PET is still highly under-explored. In this work, we propose a novel deep learning-aided motion correction and reconstruction framework for accelerated PET, called Fast-MC-PET. Our framework consists of a universal motion correction (UMC) and a short-to-long acquisition reconstruction (SL-Reon) module. The UMC enables modeling-free motion correction by estimating quasi-continuous motion from ultra-short frame reconstructions and using this information for motion-compensated reconstruction. Then, the SL-Recon converts the accelerated UMC image with low counts to a high-quality image with high counts for our final reconstruction output. Our experimental results on human studies show that our Fast-MC-PET can enable 7-fold acceleration and use only 2 minutes acquisition to generate high-quality reconstruction images that outperform/match previous motion correction reconstruction methods using standard 15 minutes long acquisition data.","PeriodicalId":73379,"journal":{"name":"Information processing in medical imaging : proceedings of the ... conference","volume":"89 1","pages":"523-535"},"PeriodicalIF":0.0,"publicationDate":"2023-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85905357","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Transient Hemodynamics Prediction Using an Efficient Octree-Based Deep Learning Model 基于高效八叉树深度学习模型的瞬态血流动力学预测
Pub Date : 2023-02-13 DOI: 10.48550/arXiv.2302.06557
Noah Maul, Katharina Zinn, Fabian Wagner, Mareike Thies, M. Rohleder, Laura Pfaff, M. Kowarschik, A. Birkhold, Andreas K. Maier
Patient-specific hemodynamics assessment could support diagnosis and treatment of neurovascular diseases. Currently, conventional medical imaging modalities are not able to accurately acquire high-resolution hemodynamic information that would be required to assess complex neurovascular pathologies. Therefore, computational fluid dynamics (CFD) simulations can be applied to tomographic reconstructions to obtain clinically relevant information. However, three-dimensional (3D) CFD simulations require enormous computational resources and simulation-related expert knowledge that are usually not available in clinical environments. Recently, deep-learning-based methods have been proposed as CFD surrogates to improve computational efficiency. Nevertheless, the prediction of high-resolution transient CFD simulations for complex vascular geometries poses a challenge to conventional deep learning models. In this work, we present an architecture that is tailored to predict high-resolution (spatial and temporal) velocity fields for complex synthetic vascular geometries. For this, an octree-based spatial discretization is combined with an implicit neural function representation to efficiently handle the prediction of the 3D velocity field for each time step. The presented method is evaluated for the task of cerebral hemodynamics prediction before and during the injection of contrast agent in the internal carotid artery (ICA). Compared to CFD simulations, the velocity field can be estimated with a mean absolute error of 0.024 m/s, whereas the run time reduces from several hours on a high-performance cluster to a few seconds on a consumer graphical processing unit.
患者特异性血流动力学评估可以支持神经血管疾病的诊断和治疗。目前,传统的医学成像方式不能准确地获得高分辨率的血流动力学信息,这将需要评估复杂的神经血管病变。因此,计算流体动力学(CFD)模拟可以应用于层析重建,以获得临床相关信息。然而,三维(3D) CFD模拟需要大量的计算资源和与模拟相关的专家知识,这些通常在临床环境中无法获得。近年来,人们提出了基于深度学习的CFD替代方法来提高计算效率。然而,复杂血管几何形状的高分辨率瞬态CFD模拟预测对传统的深度学习模型提出了挑战。在这项工作中,我们提出了一个专门用于预测复杂合成血管几何形状的高分辨率(空间和时间)速度场的架构。为此,将基于八叉树的空间离散化与隐式神经函数表示相结合,有效地处理每个时间步长的三维速度场预测。评价了该方法在颈内动脉注射造影剂前和注射过程中的脑血流动力学预测任务。与CFD模拟相比,速度场的估计平均绝对误差为0.024 m/s,而运行时间从高性能集群上的几个小时减少到消费者图形处理单元上的几秒钟。
{"title":"Transient Hemodynamics Prediction Using an Efficient Octree-Based Deep Learning Model","authors":"Noah Maul, Katharina Zinn, Fabian Wagner, Mareike Thies, M. Rohleder, Laura Pfaff, M. Kowarschik, A. Birkhold, Andreas K. Maier","doi":"10.48550/arXiv.2302.06557","DOIUrl":"https://doi.org/10.48550/arXiv.2302.06557","url":null,"abstract":"Patient-specific hemodynamics assessment could support diagnosis and treatment of neurovascular diseases. Currently, conventional medical imaging modalities are not able to accurately acquire high-resolution hemodynamic information that would be required to assess complex neurovascular pathologies. Therefore, computational fluid dynamics (CFD) simulations can be applied to tomographic reconstructions to obtain clinically relevant information. However, three-dimensional (3D) CFD simulations require enormous computational resources and simulation-related expert knowledge that are usually not available in clinical environments. Recently, deep-learning-based methods have been proposed as CFD surrogates to improve computational efficiency. Nevertheless, the prediction of high-resolution transient CFD simulations for complex vascular geometries poses a challenge to conventional deep learning models. In this work, we present an architecture that is tailored to predict high-resolution (spatial and temporal) velocity fields for complex synthetic vascular geometries. For this, an octree-based spatial discretization is combined with an implicit neural function representation to efficiently handle the prediction of the 3D velocity field for each time step. The presented method is evaluated for the task of cerebral hemodynamics prediction before and during the injection of contrast agent in the internal carotid artery (ICA). Compared to CFD simulations, the velocity field can be estimated with a mean absolute error of 0.024 m/s, whereas the run time reduces from several hours on a high-performance cluster to a few seconds on a consumer graphical processing unit.","PeriodicalId":73379,"journal":{"name":"Information processing in medical imaging : proceedings of the ... conference","volume":"3 1","pages":"183-194"},"PeriodicalIF":0.0,"publicationDate":"2023-02-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90378726","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Information processing in medical imaging : proceedings of the ... conference
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1