首页 > 最新文献

Proceedings. IEEE International Symposium on Biomedical Imaging最新文献

英文 中文
JOINT SEGMENTATION OF MULTIPLE SCLEROSIS LESIONS AND BRAIN ANATOMY IN MRI SCANS OF ANY CONTRAST AND RESOLUTION WITH CNNs. 用 CNN 对任何对比度和分辨率的 MRI 扫描中的多发性溃疡和大脑解剖进行联合分段。
Pub Date : 2021-04-01 Epub Date: 2021-05-25 DOI: 10.1109/isbi48211.2021.9434127
Benjamin Billot, Stefano Cerri, Koen Van Leemput, Adrian V Dalca, Juan Eugenio Iglesias

We present the first deep learning method to segment Multiple Sclerosis lesions and brain structures from MRI scans of any (possibly multimodal) contrast and resolution. Our method only requires segmentations to be trained (no images), as it leverages the generative model of Bayesian segmentation to generate synthetic scans with simulated lesions, which are then used to train a CNN. Our method can be retrained to segment at any resolution by adjusting the amount of synthesised partial volume. By construction, the synthetic scans are perfectly aligned with their labels, which enables training with noisy labels obtained with automatic methods. The training data are generated on the fly, and aggressive augmentation (including artefacts) is applied for improved generalisation. We demonstrate our method on two public datasets, comparing it with a state-of-the-art Bayesian approach implemented in FreeSurfer, and dataset specific CNNs trained on real data. The code is available at https://github.com/BBillot/SynthSeg.

我们提出了第一种深度学习方法,用于从任何(可能是多模态)对比度和分辨率的 MRI 扫描中分割多发性硬化症病灶和大脑结构。我们的方法只需要对分割进行训练(不需要图像),因为它利用贝叶斯分割的生成模型生成带有模拟病灶的合成扫描,然后用于训练 CNN。通过调整合成部分体积的数量,我们的方法可以在任何分辨率下重新训练分割。根据构造,合成扫描图像与其标签完全对齐,因此可以使用自动方法获得的噪声标签进行训练。训练数据是即时生成的,为了提高泛化能力,我们采用了积极的增强方法(包括伪影)。我们在两个公共数据集上演示了我们的方法,并将其与在 FreeSurfer 中实现的最先进的贝叶斯方法以及在真实数据上训练的特定数据集 CNN 进行了比较。代码可在 https://github.com/BBillot/SynthSeg 上获取。
{"title":"JOINT SEGMENTATION OF MULTIPLE SCLEROSIS LESIONS AND BRAIN ANATOMY IN MRI SCANS OF ANY CONTRAST AND RESOLUTION WITH CNNs.","authors":"Benjamin Billot, Stefano Cerri, Koen Van Leemput, Adrian V Dalca, Juan Eugenio Iglesias","doi":"10.1109/isbi48211.2021.9434127","DOIUrl":"10.1109/isbi48211.2021.9434127","url":null,"abstract":"<p><p>We present the first deep learning method to segment Multiple Sclerosis lesions and brain structures from MRI scans of any (possibly multimodal) contrast and resolution. Our method only requires segmentations to be trained (no images), as it leverages the generative model of Bayesian segmentation to generate synthetic scans with simulated lesions, which are then used to train a CNN. Our method can be retrained to segment at any resolution by adjusting the amount of synthesised partial volume. By construction, the synthetic scans are perfectly aligned with their labels, which enables training with noisy labels obtained with automatic methods. The training data are generated on the fly, and aggressive augmentation (including artefacts) is applied for improved generalisation. We demonstrate our method on two public datasets, comparing it with a state-of-the-art Bayesian approach implemented in FreeSurfer, and dataset specific CNNs trained on real data. The code is available at https://github.com/BBillot/SynthSeg.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":"2021 ","pages":"1971-1974"},"PeriodicalIF":0.0,"publicationDate":"2021-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8340983/pdf/nihms-1727379.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"39291017","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DEEP GENERATIVE STORM MODEL FOR DYNAMIC IMAGING. 用于动态成像的深度风暴生成模型
Pub Date : 2021-04-01 Epub Date: 2021-03-25 DOI: 10.1109/isbi48211.2021.9433839
Qing Zou, Abdul Haseeb Ahmed, Prashant Nagpal, Stanley Kruger, Mathews Jacob

We introduce a novel generative smoothness regularization on manifolds (SToRM) model for the recovery of dynamic image data from highly undersampled measurements. The proposed generative framework represents the image time series as a smooth non-linear function of low-dimensional latent vectors that capture the cardiac and respiratory phases. The non-linear function is represented using a deep convolutional neural network (CNN). Unlike the popular CNN approaches that require extensive fully-sampled training data that is not available in this setting, the parameters of the CNN generator as well as the latent vectors are jointly estimated from the undersampled measurements using stochastic gradient descent. We penalize the norm of the gradient of the generator to encourage the learning of a smooth surface/manifold, while temporal gradients of the latent vectors are penalized to encourage the time series to be smooth. The main benefits of the proposed scheme are (a) the quite significant reduction in memory demand compared to the analysis based SToRM model, and (b) the spatial regularization brought in by the CNN model. We also introduce efficient progressive approaches to minimize the computational complexity of the algorithm.

我们介绍了一种新颖的流形平滑正则化(STORM)生成模型,用于从高度采样不足的测量数据中恢复动态图像数据。所提出的生成框架将图像时间序列表示为捕捉心脏和呼吸阶段的低维潜在向量的平滑非线性函数。非线性函数使用深度卷积神经网络(CNN)表示。流行的卷积神经网络方法需要大量完全采样的训练数据,而在这种情况下无法获得这些数据,与之不同的是,卷积神经网络生成器的参数以及潜向量是通过随机梯度下降法,从采样不足的测量数据中共同估算出来的。我们对生成器梯度的常模进行惩罚,以鼓励学习平滑的曲面/manifold,同时对潜在向量的时间梯度进行惩罚,以鼓励时间序列变得平滑。与基于分析的 SToRM 模型相比,拟议方案的主要优势在于:(a) 显著减少了内存需求;(b) CNN 模型带来的空间正则化。我们还引入了高效的渐进方法,以最大限度地降低算法的计算复杂度。
{"title":"DEEP GENERATIVE STORM MODEL FOR DYNAMIC IMAGING.","authors":"Qing Zou, Abdul Haseeb Ahmed, Prashant Nagpal, Stanley Kruger, Mathews Jacob","doi":"10.1109/isbi48211.2021.9433839","DOIUrl":"10.1109/isbi48211.2021.9433839","url":null,"abstract":"<p><p>We introduce a novel generative smoothness regularization on manifolds (SToRM) model for the recovery of dynamic image data from highly undersampled measurements. The proposed generative framework represents the image time series as a smooth non-linear function of low-dimensional latent vectors that capture the cardiac and respiratory phases. The non-linear function is represented using a deep convolutional neural network (CNN). Unlike the popular CNN approaches that require extensive fully-sampled training data that is not available in this setting, the parameters of the CNN generator as well as the latent vectors are jointly estimated from the undersampled measurements using stochastic gradient descent. We penalize the norm of the gradient of the generator to encourage the learning of a smooth surface/manifold, while temporal gradients of the latent vectors are penalized to encourage the time series to be smooth. The main benefits of the proposed scheme are (a) the quite significant reduction in memory demand compared to the analysis based SToRM model, and (b) the spatial regularization brought in by the CNN model. We also introduce efficient progressive approaches to minimize the computational complexity of the algorithm.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":"2021 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2021-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8320670/pdf/nihms-1668003.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"39267150","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CALIBRATIONLESS MRI RECONSTRUCTION WITH A PLUG-IN DENOISER. 带插件去噪器的无校准mri重建。
Pub Date : 2021-04-01 Epub Date: 2021-05-25 DOI: 10.1109/isbi48211.2021.9433815
Shen Zhao, Lee C Potter, Rizwan Ahmad

Magnetic Resonance Imaging (MRI) is a noninvasive imaging technique that provides excellent soft-tissue contrast without using ionizing radiation. MRI's clinical application may be limited by long data acquisition time; therefore, MR image reconstruction from highly under-sampled k-space data has been an active research area. Calibrationless MRI not only enables a higher acceleration rate but also increases flexibility for sampling pattern design. To leverage non-linear machine learning priors, we pair our High-dimensional Fast Convolutional Framework (HICU) [1] with a plug-in denoiser and demonstrate its feasibility using 2D brain data.

磁共振成像(MRI)是一种无创成像技术,无需电离辐射即可提供出色的软组织对比。MRI的临床应用可能受到数据采集时间长的限制;因此,从高度欠采样的k空间数据中重建MR图像一直是一个活跃的研究领域。无校准MRI不仅可以实现更高的加速速率,而且可以增加采样模式设计的灵活性。为了利用非线性机器学习先验,我们将高维快速卷积框架(HICU)[1]与插件去噪器配对,并使用2D大脑数据证明其可行性。
{"title":"CALIBRATIONLESS MRI RECONSTRUCTION WITH A PLUG-IN DENOISER.","authors":"Shen Zhao,&nbsp;Lee C Potter,&nbsp;Rizwan Ahmad","doi":"10.1109/isbi48211.2021.9433815","DOIUrl":"https://doi.org/10.1109/isbi48211.2021.9433815","url":null,"abstract":"<p><p>Magnetic Resonance Imaging (MRI) is a noninvasive imaging technique that provides excellent soft-tissue contrast without using ionizing radiation. MRI's clinical application may be limited by long data acquisition time; therefore, MR image reconstruction from highly under-sampled k-space data has been an active research area. Calibrationless MRI not only enables a higher acceleration rate but also increases flexibility for sampling pattern design. To leverage non-linear machine learning priors, we pair our High-dimensional Fast Convolutional Framework (HICU) [1] with a plug-in denoiser and demonstrate its feasibility using 2D brain data.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":"2021 ","pages":"1846-1849"},"PeriodicalIF":0.0,"publicationDate":"2021-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/isbi48211.2021.9433815","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"39834658","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
STATISTICAL COMPARISONS OF CHROMOSOMAL SHAPE POPULATIONS. 染色体形状群体的统计比较。
Pub Date : 2021-04-01 Epub Date: 2021-05-25 DOI: 10.1109/isbi48211.2021.9433812
Carlos J Soto, Peiyao A Zhao, Kyle N Klein, David M Gilbert, Anuj Srivastava

This paper develops statistical tools for testing differences in shapes of chromosomes resulting from certain gene knockouts (KO), specifically RIF1 gene KO (RKO) and the cohesin subunit RAD21 gene KO (CKO). It utilizes a two-sample test for comparing shapes of KO chromosomes with wild type (WT) at two levels: (1) Coarse shape analysis, where one compares shapes of full or large parts of chromosomes, and (2) Fine shape analysis, where chromosomes are first segmented into (TAD-based) pieces and then the corresponding pieces are compared across populations. The shape comparisons - coarse and fine - are based on an elastic shape metric for comparing shapes of 3D curves. The experiments show that the KO populations, RKO and CKO, have statistically significant differences from WT at both coarse and fine levels. Furthermore, this framework highlights local regions where these differences are most prominent.

本文开发了统计工具,用于测试某些基因敲除(KO)导致的染色体形状差异,特别是 RIF1 基因敲除(RKO)和粘合素亚基 RAD21 基因敲除(CKO)。它利用双样本检验在两个层面上比较 KO 染色体与野生型(WT)的形状:(1) 粗形状分析,即比较染色体全部或大部分的形状;(2) 细形状分析,即首先将染色体分割成(基于 TAD 的)片段,然后比较不同群体的相应片段。粗略和精细的形状比较基于一种用于比较三维曲线形状的弹性形状指标。实验表明,KO 群体(RKO 和 CKO)与 WT 群体在粗略和精细程度上都有显著的统计学差异。此外,该框架还突出了这些差异最显著的局部区域。
{"title":"STATISTICAL COMPARISONS OF CHROMOSOMAL SHAPE POPULATIONS.","authors":"Carlos J Soto, Peiyao A Zhao, Kyle N Klein, David M Gilbert, Anuj Srivastava","doi":"10.1109/isbi48211.2021.9433812","DOIUrl":"10.1109/isbi48211.2021.9433812","url":null,"abstract":"<p><p>This paper develops statistical tools for testing differences in shapes of chromosomes resulting from certain gene knockouts (KO), specifically RIF1 gene KO (RKO) and the cohesin subunit RAD21 gene KO (CKO). It utilizes a <i>two-sample test</i> for comparing shapes of KO chromosomes with wild type (WT) at two levels: (1) <i>Coarse shape analysis</i>, where one compares shapes of full or large parts of chromosomes, and (2) <i>Fine shape analysis</i>, where chromosomes are first segmented into (TAD-based) pieces and then the corresponding pieces are compared across populations. The shape comparisons - coarse and fine - are based on an elastic shape metric for comparing shapes of 3D curves. The experiments show that the KO populations, RKO and CKO, have statistically significant differences from WT at both coarse and fine levels. Furthermore, this framework highlights local regions where these differences are most prominent.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":"2021 ","pages":"788-791"},"PeriodicalIF":0.0,"publicationDate":"2021-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8840943/pdf/nihms-1776570.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"39924997","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
VISUALIZING MISSING SURFACES IN COLONOSCOPY VIDEOS USING SHARED LATENT SPACE REPRESENTATIONS. 使用共享的潜在空间表示可视化结肠镜检查视频中缺失的表面。
Pub Date : 2021-04-01 Epub Date: 2021-05-25 DOI: 10.1109/isbi48211.2021.9433982
Shawn Mathew, Saad Nadeem, Arie Kaufman

Optical colonoscopy (OC), the most prevalent colon cancer screening tool, has a high miss rate due to a number of factors, including the geometry of the colon (haustral fold and sharp bends occlusions), endoscopist inexperience or fatigue, endoscope field of view. We present a framework to visualize the missed regions per-frame during OC, and provides a workable clinical solution. Specifically, we make use of 3D reconstructed virtual colonoscopy (VC) data and the insight that VC and OC share the same underlying geometry but differ in color, texture and specular reflections, embedded in the OC. A lossy unpaired image-to-image translation model is introduced with enforced shared latent space for OC and VC. This shared space captures the geometric information while deferring the color, texture, and specular information creation to additional Gaussian noise input. The latter can be utilized to generate one-to-many mappings from VC to OC and OC to OC. The code, data and trained models will be released via our Computational Endoscopy Platform at https://github.com/nadeemlab/CEP.

光学结肠镜检查(OC)是最流行的结肠癌癌症筛查工具,由于多种因素,包括结肠的几何形状(吸器折叠和急弯闭塞)、内镜医生缺乏经验或疲劳、内窥镜视野,其漏诊率很高。我们提出了一个框架来可视化OC期间每帧的遗漏区域,并提供了一个可行的临床解决方案。具体来说,我们利用了3D重建的虚拟结肠镜检查(VC)数据,以及VC和OC共享相同的底层几何结构,但在颜色、纹理和镜面反射方面有所不同,嵌入在OC中。引入了一种有损的不成对图像到图像的转换模型,该模型具有针对OC和VC的强制共享潜在空间。该共享空间捕获几何信息,同时将颜色、纹理和镜面信息的创建推迟到额外的高斯噪声输入。后者可用于生成从VC到OC和从OC到OC的一对多映射。代码、数据和经过训练的模型将通过我们的计算内窥镜平台在https://github.com/nadeemlab/CEP.
{"title":"VISUALIZING MISSING SURFACES IN COLONOSCOPY VIDEOS USING SHARED LATENT SPACE REPRESENTATIONS.","authors":"Shawn Mathew,&nbsp;Saad Nadeem,&nbsp;Arie Kaufman","doi":"10.1109/isbi48211.2021.9433982","DOIUrl":"10.1109/isbi48211.2021.9433982","url":null,"abstract":"<p><p>Optical colonoscopy (OC), the most prevalent colon cancer screening tool, has a high miss rate due to a number of factors, including the geometry of the colon (haustral fold and sharp bends occlusions), endoscopist inexperience or fatigue, endoscope field of view. We present a framework to visualize the missed regions per-frame during OC, and provides a workable clinical solution. Specifically, we make use of 3D reconstructed virtual colonoscopy (VC) data and the insight that VC and OC share the same underlying geometry but differ in color, texture and specular reflections, embedded in the OC. A lossy unpaired image-to-image translation model is introduced with enforced shared latent space for OC and VC. This shared space captures the geometric information while deferring the color, texture, and specular information creation to additional Gaussian noise input. The latter can be utilized to generate one-to-many mappings from VC to OC and OC to OC. The code, data and trained models will be released via our Computational Endoscopy Platform at https://github.com/nadeemlab/CEP.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":"2021 ","pages":"329-333"},"PeriodicalIF":0.0,"publicationDate":"2021-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/isbi48211.2021.9433982","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"39513772","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Predicting Progression from Mild Cognitive Impairment to Alzheimer's Disease using MRI-based Cortical Features and a Two-State Markov Model. 使用基于mri的皮质特征和双状态马尔可夫模型预测从轻度认知障碍到阿尔茨海默病的进展。
Pub Date : 2021-04-01 Epub Date: 2021-05-25 DOI: 10.1109/isbi48211.2021.9434143
Eleonora Ficiarà, Valentino Crespi, Shruti Prashant Gadewar, Sophia I Thomopoulos, Joshua Boyd, Paul M Thompson, Neda Jahanshad, Fabrizio Pizzagalli

Magnetic resonance imaging (MRI) has a potential for early diagnosis of individuals at risk for developing Alzheimer's disease (AD). Cognitive performance in healthy elderly people and in those with mild cognitive impairment (MCI) has been associated with measures of cortical gyrification [1] and thickness (CT) [2], yet the extent to which sulcal measures can help to predict AD conversion above and beyond CT measures is not known. Here, we analyzed 721 participants with MCI from phases 1 and 2 of the Alzheimer's Disease Neuroimaging Initiative, applying a two-state Markov model to study the conversion from MCI to AD condition. Our preliminary results suggest that MRI-based cortical features, including sulcal morphometry, may help to predict conversion from MCI to AD.

磁共振成像(MRI)具有早期诊断阿尔茨海默病(AD)风险个体的潜力。健康老年人和轻度认知障碍(MCI)患者的认知表现与皮质旋回[1]和厚度(CT)[2]相关,但脑沟测量在多大程度上可以帮助预测AD的转换,而不是CT测量。在这里,我们分析了721名来自阿尔茨海默病神经成像计划第一和第二阶段的MCI参与者,应用双状态马尔可夫模型研究从MCI到AD的转换。我们的初步结果表明,基于mri的皮层特征,包括脑沟形态测定,可能有助于预测从轻度认知损伤到AD的转化。
{"title":"Predicting Progression from Mild Cognitive Impairment to Alzheimer's Disease using MRI-based Cortical Features and a Two-State Markov Model.","authors":"Eleonora Ficiarà,&nbsp;Valentino Crespi,&nbsp;Shruti Prashant Gadewar,&nbsp;Sophia I Thomopoulos,&nbsp;Joshua Boyd,&nbsp;Paul M Thompson,&nbsp;Neda Jahanshad,&nbsp;Fabrizio Pizzagalli","doi":"10.1109/isbi48211.2021.9434143","DOIUrl":"https://doi.org/10.1109/isbi48211.2021.9434143","url":null,"abstract":"<p><p>Magnetic resonance imaging (MRI) has a potential for early diagnosis of individuals at risk for developing Alzheimer's disease (AD). Cognitive performance in healthy elderly people and in those with mild cognitive impairment (MCI) has been associated with measures of cortical gyrification [1] and thickness (CT) [2], yet the extent to which sulcal measures can help to predict AD conversion above and beyond CT measures is not known. Here, we analyzed 721 participants with MCI from phases 1 and 2 of the Alzheimer's Disease Neuroimaging Initiative, applying a two-state Markov model to study the conversion from MCI to AD condition. Our preliminary results suggest that MRI-based cortical features, including sulcal morphometry, may help to predict conversion from MCI to AD.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":" ","pages":"1145-1149"},"PeriodicalIF":0.0,"publicationDate":"2021-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/isbi48211.2021.9434143","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"40317904","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
LEARNING TO SYNTHESIZE CORTICAL MORPHOLOGICAL CHANGES USING GRAPH CONDITIONAL VARIATIONAL AUTOENCODER. 学习用图条件变分自编码器合成皮层形态变化。
Pub Date : 2021-04-01 Epub Date: 2021-05-25 DOI: 10.1109/isbi48211.2021.9433837
Yaqiong Chai, Mengting Liu, Ben A Duffy, Hosung Kim

Changes in brain morphology, such as cortical thinning are of great value for understanding the trajectory of brain aging and various neurodegenerative diseases. In this work, we employed a generative neural network variational autoencoder (VAE) that is conditional on age and is able to generate cortical thickness maps at various ages given an input cortical thickness map. To take into account the mesh topology in the model, we proposed a loss function based on weighted adjacency to integrate the surface topography defined as edge connections with the cortical thickness mapped as vertices. Compared to traditional conditional VAE that did not use the surface topological information, our method better predicted "future" cortical thickness maps, especially when the age gap became wider. Our model has the potential to predict the distinctive temporospatial pattern of individual cortical morphology in relation to aging and neurodegenerative diseases.

脑形态学的变化,如皮层变薄,对理解脑老化和各种神经退行性疾病的轨迹具有重要价值。在这项工作中,我们采用了一种以年龄为条件的生成式神经网络变分自编码器(VAE),它能够在给定输入皮层厚度图的情况下生成不同年龄的皮层厚度图。考虑到模型中的网格拓扑结构,我们提出了一个基于加权邻接的损失函数,将定义为边缘连接的表面形貌与映射为顶点的皮质厚度进行整合。与不使用表面拓扑信息的传统条件VAE相比,我们的方法更好地预测了“未来”皮层厚度图,特别是当年龄差距变得更大时。我们的模型有潜力预测与衰老和神经退行性疾病相关的个体皮层形态的独特时空模式。
{"title":"LEARNING TO SYNTHESIZE CORTICAL MORPHOLOGICAL CHANGES USING GRAPH CONDITIONAL VARIATIONAL AUTOENCODER.","authors":"Yaqiong Chai,&nbsp;Mengting Liu,&nbsp;Ben A Duffy,&nbsp;Hosung Kim","doi":"10.1109/isbi48211.2021.9433837","DOIUrl":"https://doi.org/10.1109/isbi48211.2021.9433837","url":null,"abstract":"<p><p>Changes in brain morphology, such as cortical thinning are of great value for understanding the trajectory of brain aging and various neurodegenerative diseases. In this work, we employed a generative neural network variational autoencoder (VAE) that is conditional on age and is able to generate cortical thickness maps at various ages given an input cortical thickness map. To take into account the mesh topology in the model, we proposed a loss function based on weighted adjacency to integrate the surface topography defined as edge connections with the cortical thickness mapped as vertices. Compared to traditional conditional VAE that did not use the surface topological information, our method better predicted \"future\" cortical thickness maps, especially when the age gap became wider. Our model has the potential to predict the distinctive temporospatial pattern of individual cortical morphology in relation to aging and neurodegenerative diseases.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":" ","pages":"1495-1499"},"PeriodicalIF":0.0,"publicationDate":"2021-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/isbi48211.2021.9433837","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"40322297","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Discovering Salient Anatomical Landmarks by Predicting Human Gaze. 通过预测人类凝视发现显著的解剖标志。
Pub Date : 2020-04-03 DOI: 10.1109/ISBI45749.2020.9098505
R Droste, P Chatelain, L Drukker, H Sharma, A T Papageorghiou, J A Noble

Anatomical landmarks are a crucial prerequisite for many medical imaging tasks. Usually, the set of landmarks for a given task is predefined by experts. The landmark locations for a given image are then annotated manually or via machine learning methods trained on manual annotations. In this paper, in contrast, we present a method to automatically discover and localize anatomical landmarks in medical images. Specifically, we consider landmarks that attract the visual attention of humans, which we term visually salient landmarks. We illustrate the method for fetal neurosonographic images. First, full-length clinical fetal ultrasound scans are recorded with live sonographer gaze-tracking. Next, a convolutional neural network (CNN) is trained to predict the gaze point distribution (saliency map) of the sonographers on scan video frames. The CNN is then used to predict saliency maps of unseen fetal neurosonographic images, and the landmarks are extracted as the local maxima of these saliency maps. Finally, the landmarks are matched across images by clustering the landmark CNN features. We show that the discovered landmarks can be used within affine image registration, with average landmark alignment errors between 4.1% and 10.9% of the fetal head long axis length.

解剖标志是许多医学成像任务的关键先决条件。通常,给定任务的标志集是由专家预定义的。然后对给定图像的地标位置进行手动注释,或者通过对手动注释进行训练的机器学习方法进行注释。在本文中,我们提出了一种自动发现和定位医学图像中解剖标志的方法。具体来说,我们考虑吸引人类视觉注意力的地标,我们称之为视觉显著地标。我们说明胎儿神经超声图像的方法。首先,全程临床胎儿超声扫描记录与现场超声仪的目光跟踪。接下来,训练卷积神经网络(CNN)来预测超声医师在扫描视频帧上的注视点分布(显著性图)。然后使用CNN预测未见的胎儿神经超声图像的显著性图,并提取地标作为这些显著性图的局部最大值。最后,通过对地标CNN特征进行聚类,实现图像间地标的匹配。结果表明,发现的标记可以用于仿射图像配准,平均标记对齐误差在胎儿头部长轴长度的4.1% ~ 10.9%之间。
{"title":"Discovering Salient Anatomical Landmarks by Predicting Human Gaze.","authors":"R Droste,&nbsp;P Chatelain,&nbsp;L Drukker,&nbsp;H Sharma,&nbsp;A T Papageorghiou,&nbsp;J A Noble","doi":"10.1109/ISBI45749.2020.9098505","DOIUrl":"https://doi.org/10.1109/ISBI45749.2020.9098505","url":null,"abstract":"<p><p>Anatomical landmarks are a crucial prerequisite for many medical imaging tasks. Usually, the set of landmarks for a given task is predefined by experts. The landmark locations for a given image are then annotated manually or via machine learning methods trained on manual annotations. In this paper, in contrast, we present a method to automatically discover and localize anatomical landmarks in medical images. Specifically, we consider landmarks that attract the visual attention of humans, which we term <i>visually salient landmarks</i>. We illustrate the method for fetal neurosonographic images. First, full-length clinical fetal ultrasound scans are recorded with live sonographer gaze-tracking. Next, a convolutional neural network (CNN) is trained to predict the gaze point distribution (saliency map) of the sonographers on scan video frames. The CNN is then used to predict saliency maps of unseen fetal neurosonographic images, and the landmarks are extracted as the local maxima of these saliency maps. Finally, the landmarks are matched across images by clustering the landmark CNN features. We show that the discovered landmarks can be used within affine image registration, with average landmark alignment errors between 4.1% and 10.9% of the fetal head long axis length.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":" ","pages":"1711-1714"},"PeriodicalIF":0.0,"publicationDate":"2020-04-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/ISBI45749.2020.9098505","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"38006116","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Self-Supervised Representation Learning for Ultrasound Video. 超声视频的自监督表示学习。
Pub Date : 2020-04-03 DOI: 10.1109/ISBI45749.2020.9098666
Jianbo Jiao, Richard Droste, Lior Drukker, Aris T Papageorghiou, J Alison Noble

Recent advances in deep learning have achieved promising performance for medical image analysis, while in most cases ground-truth annotations from human experts are necessary to train the deep model. In practice, such annotations are expensive to collect and can be scarce for medical imaging applications. Therefore, there is significant interest in learning representations from unlabelled raw data. In this paper, we propose a self-supervised learning approach to learn meaningful and transferable representations from medical imaging video without any type of human annotation. We assume that in order to learn such a representation, the model should identify anatomical structures from the unlabelled data. Therefore we force the model to address anatomy-aware tasks with free supervision from the data itself. Specifically, the model is designed to correct the order of a reshuffled video clip and at the same time predict the geometric transformation applied to the video clip. Experiments on fetal ultrasound video show that the proposed approach can effectively learn meaningful and strong representations, which transfer well to downstream tasks like standard plane detection and saliency prediction.

深度学习的最新进展已经在医学图像分析中取得了很好的表现,而在大多数情况下,来自人类专家的真值注释是训练深度模型所必需的。在实践中,这种注释的收集成本很高,而且对于医学成像应用程序来说可能很少。因此,人们对从未标记的原始数据中学习表示非常感兴趣。在本文中,我们提出了一种自监督学习方法,从医学成像视频中学习有意义和可转移的表征,而无需任何类型的人类注释。我们假设为了学习这种表示,模型应该从未标记的数据中识别解剖结构。因此,我们迫使模型在数据本身的自由监督下解决解剖学感知任务。具体来说,该模型旨在纠正重新洗牌的视频片段的顺序,同时预测应用于视频片段的几何变换。胎儿超声视频实验表明,该方法可以有效地学习有意义的强表征,并能很好地转移到标准平面检测和显著性预测等下游任务中。
{"title":"Self-Supervised Representation Learning for Ultrasound Video.","authors":"Jianbo Jiao,&nbsp;Richard Droste,&nbsp;Lior Drukker,&nbsp;Aris T Papageorghiou,&nbsp;J Alison Noble","doi":"10.1109/ISBI45749.2020.9098666","DOIUrl":"https://doi.org/10.1109/ISBI45749.2020.9098666","url":null,"abstract":"<p><p>Recent advances in deep learning have achieved promising performance for medical image analysis, while in most cases ground-truth annotations from human experts are necessary to train the deep model. In practice, such annotations are expensive to collect and can be scarce for medical imaging applications. Therefore, there is significant interest in learning representations from unlabelled raw data. In this paper, we propose a self-supervised learning approach to learn meaningful and transferable representations from medical imaging video without any type of human annotation. We assume that in order to learn such a representation, the model should identify anatomical structures from the unlabelled data. Therefore we force the model to address anatomy-aware tasks with free supervision from the data itself. Specifically, the model is designed to correct the order of a reshuffled video clip and at the same time predict the geometric transformation applied to the video clip. Experiments on fetal ultrasound video show that the proposed approach can effectively learn meaningful and strong representations, which transfer well to downstream tasks like standard plane detection and saliency prediction.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":" ","pages":"1847-1850"},"PeriodicalIF":0.0,"publicationDate":"2020-04-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/ISBI45749.2020.9098666","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"38006117","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 41
Model-Based Deep Learning for Reconstruction of Joint k-q Under-sampled High Resolution Diffusion MRI. 基于模型的深度学习重建关节k-q欠采样高分辨率扩散MRI。
Pub Date : 2020-04-01 Epub Date: 2020-05-22 DOI: 10.1109/isbi45749.2020.9098593
Merry P Mani, Hemant K Aggarwal, Sanjay Ghosh, Mathews Jacob

We propose a model-based deep learning architecture for the reconstruction of highly accelerated diffusion magnetic resonance imaging (MRI) that enables high resolution imaging. The proposed reconstruction jointly recovers all the diffusion weighted images in a single step from a joint k-q under-sampled acquisition in a parallel MRI setting. We propose the novel use of a pre-trained denoiser as a regularizer in a model-based reconstruction for the recovery of highly under-sampled data. Specifically, we designed the denoiser based on a general diffusion MRI tissue microstructure model for multi-compartmental modeling. By using a wide range of biologically plausible parameter values for the multi-compartmental microstructure model, we simulated diffusion signal that spans the entire microstructure parameter space. A neural network was trained in an unsupervised manner using an autoencoder to learn the diffusion MRI signal subspace. We employed the autoencoder in a model-based reconstruction and show that the autoencoder provides a strong denoising prior to recover the q-space signal. We show reconstruction results on a simulated brain dataset that shows high acceleration capabilities of the proposed method.

我们提出了一种基于模型的深度学习架构,用于高加速扩散磁共振成像(MRI)的重建,从而实现高分辨率成像。本文提出的重建方法在一个步骤中从一个联合k-q欠采样采集中联合恢复所有扩散加权图像。我们提出在基于模型的重建中使用预训练的去噪器作为正则器,用于恢复高度欠采样数据。具体来说,我们设计了基于一般扩散MRI组织微观结构模型的去噪器,用于多室建模。通过使用广泛的生物学上合理的多室微结构模型参数值,我们模拟了跨越整个微结构参数空间的扩散信号。利用自编码器以无监督的方式训练神经网络学习扩散核磁共振信号子空间。我们在基于模型的重建中使用了自编码器,并表明自编码器在恢复q空间信号之前提供了强去噪。我们展示了模拟大脑数据集的重建结果,显示了所提出方法的高加速能力。
{"title":"Model-Based Deep Learning for Reconstruction of Joint k-q Under-sampled High Resolution Diffusion MRI.","authors":"Merry P Mani,&nbsp;Hemant K Aggarwal,&nbsp;Sanjay Ghosh,&nbsp;Mathews Jacob","doi":"10.1109/isbi45749.2020.9098593","DOIUrl":"https://doi.org/10.1109/isbi45749.2020.9098593","url":null,"abstract":"<p><p>We propose a model-based deep learning architecture for the reconstruction of highly accelerated diffusion magnetic resonance imaging (MRI) that enables high resolution imaging. The proposed reconstruction jointly recovers all the diffusion weighted images in a single step from a joint k-q under-sampled acquisition in a parallel MRI setting. We propose the novel use of a pre-trained denoiser as a regularizer in a model-based reconstruction for the recovery of highly under-sampled data. Specifically, we designed the denoiser based on a general diffusion MRI tissue microstructure model for multi-compartmental modeling. By using a wide range of biologically plausible parameter values for the multi-compartmental microstructure model, we simulated diffusion signal that spans the entire microstructure parameter space. A neural network was trained in an unsupervised manner using an autoencoder to learn the diffusion MRI signal subspace. We employed the autoencoder in a model-based reconstruction and show that the autoencoder provides a strong denoising prior to recover the q-space signal. We show reconstruction results on a simulated brain dataset that shows high acceleration capabilities of the proposed method.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":" ","pages":"913-916"},"PeriodicalIF":0.0,"publicationDate":"2020-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/isbi45749.2020.9098593","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"25360296","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
期刊
Proceedings. IEEE International Symposium on Biomedical Imaging
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1