首页 > 最新文献

Proceedings. IEEE International Symposium on Biomedical Imaging最新文献

英文 中文
Brain Hemisphere Dissimilarity, a Self-Supervised Learning Approach for alpha-synucleinopathies prediction with FDG PET. 用 FDG PET 预测α-突触核蛋白病的自监督学习方法--脑半球相似性。
Pub Date : 2023-04-01 Epub Date: 2023-09-01 DOI: 10.1109/isbi53787.2023.10230560
S Tripathi, P Mattioli, C Liguori, A Chiaravalloti, D Arnaldi, L Giancardo

Idiopathic Rem sleep Behavior Disorder (iRBD) is a significant biomarker for the development of alpha-synucleinopathies, such as Parkinson's disease (PD) or Dementia with Lewy bodies (DLB). Methods to identify patterns in iRBD patients can help in the prediction of the future conversion to these diseases during the long prodromal phase when symptoms are non-specific. These methods are essential for disease management and clinical trial recruitment. Brain PET scans with 18F-FDG PET radiotracers have recently shown promise, however, the scarcity of longitudinal data and PD/DLB conversion information makes the use of representation learning approaches such as deep convolutional networks not feasible if trained in a supervised manner. In this work, we propose a self-supervised learning strategy to learn features by comparing the brain hemispheres of iRBD non-convertor subjects, which allows for pre-training a convolutional network on a small data regimen. We introduce a loss function called hemisphere dissimilarity loss (HDL), which extends the Barlow Twins loss, that promotes the creation of invariant and non-redundant features for brain hemispheres of the same subject, and the opposite for hemispheres of different subjects. This loss enables the pre-training of a network without any information about the disease, which is then used to generate full brain feature vectors that are fine-tuned to two downstream tasks: follow-up conversion, and the type of conversion (PD or DLB) using baseline 18F-FDG PET. In our results, we find that the HDL outperforms the variational autoencoder with different forms of inputs.

特发性睡眠行为障碍(iRBD)是帕金森病(PD)或路易体痴呆(DLB)等α-突触核蛋白病发展的重要生物标志物。在症状无特异性的漫长前驱期,识别 iRBD 患者模式的方法有助于预测这些疾病的未来转归。这些方法对于疾病管理和临床试验招募至关重要。最近,使用 18F-FDG PET 放射性同位素进行的脑 PET 扫描显示了前景,然而,由于纵向数据和 PD/DLB 转换信息的稀缺,使用深度卷积网络等表征学习方法进行监督训练并不可行。在这项工作中,我们提出了一种自监督学习策略,通过比较 iRBD 非转换者受试者的大脑半球来学习特征,这样就可以在小数据方案上对卷积网络进行预训练。我们引入了一种称为半球不相似性损失(HDL)的损失函数,它扩展了巴洛双胞胎损失(Barlow Twins loss),可促进为同一受试者的大脑半球创建不变且非冗余的特征,而为不同受试者的大脑半球创建相反的特征。通过这种损失,可以在没有任何疾病信息的情况下对网络进行预训练,然后利用预训练生成完整的大脑特征向量,并根据两个下游任务对其进行微调:随访转换和利用基线 18F-FDG PET 确定转换类型(PD 或 DLB)。在我们的研究结果中,我们发现 HDL 在不同形式的输入下都优于变异自动编码器。
{"title":"Brain Hemisphere Dissimilarity, a Self-Supervised Learning Approach for alpha-synucleinopathies prediction with FDG PET.","authors":"S Tripathi, P Mattioli, C Liguori, A Chiaravalloti, D Arnaldi, L Giancardo","doi":"10.1109/isbi53787.2023.10230560","DOIUrl":"10.1109/isbi53787.2023.10230560","url":null,"abstract":"<p><p>Idiopathic Rem sleep Behavior Disorder (iRBD) is a significant biomarker for the development of alpha-synucleinopathies, such as Parkinson's disease (PD) or Dementia with Lewy bodies (DLB). Methods to identify patterns in iRBD patients can help in the prediction of the future conversion to these diseases during the long prodromal phase when symptoms are non-specific. These methods are essential for disease management and clinical trial recruitment. Brain PET scans with 18F-FDG PET radiotracers have recently shown promise, however, the scarcity of longitudinal data and PD/DLB conversion information makes the use of representation learning approaches such as deep convolutional networks not feasible if trained in a supervised manner. In this work, we propose a self-supervised learning strategy to learn features by comparing the brain hemispheres of iRBD non-convertor subjects, which allows for pre-training a convolutional network on a small data regimen. We introduce a loss function called hemisphere dissimilarity loss (HDL), which extends the Barlow Twins loss, that promotes the creation of invariant and non-redundant features for brain hemispheres of the same subject, and the opposite for hemispheres of different subjects. This loss enables the pre-training of a network without any information about the disease, which is then used to generate full brain feature vectors that are fine-tuned to two downstream tasks: follow-up conversion, and the type of conversion (PD or DLB) using baseline 18F-FDG PET. In our results, we find that the HDL outperforms the variational autoencoder with different forms of inputs.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":"2023 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10496490/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10264588","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Intermediate Deformable Image Registration via Windowed Cross-Correlation. 中间变形图像配准通过窗口交叉相关。
Pub Date : 2023-04-01 DOI: 10.1109/isbi53787.2023.10230715
Iman Aganj, Bruce Fischl

In population and longitudinal imaging studies that employ deformable image registration, more accurate results can be achieved by initializing deformable registration with the results of affine registration where global misalignments have been considerably reduced. Such affine registration, however, is limited to linear transformations and it cannot account for large nonlinear anatomical variations, such as those between pre- and post-operative images or across different subject anatomies. In this work, we introduce a new intermediate deformable image registration (IDIR) technique that recovers large deformations via windowed cross-correlation, and provide an efficient implementation based on the fast Fourier transform. We evaluate our method on 2D X-ray and 3D magnetic resonance images, demonstrating its ability to align substantial nonlinear anatomical variations within a few iterations.

在采用可变形图像配准的总体和纵向成像研究中,通过将可变形配准与仿射配准的结果初始化,可以获得更准确的结果,其中全局不对准已经大大减少。然而,这种仿射配准仅限于线性变换,不能解释大的非线性解剖变化,例如术前和术后图像之间或不同主体解剖结构之间的变化。在这项工作中,我们引入了一种新的中间可变形图像配准(IDIR)技术,该技术通过加窗互相关恢复大变形,并提供了一种基于快速傅里叶变换的有效实现。我们在2D x射线和3D磁共振图像上评估了我们的方法,证明了它在几次迭代内对齐大量非线性解剖变化的能力。
{"title":"Intermediate Deformable Image Registration via Windowed Cross-Correlation.","authors":"Iman Aganj,&nbsp;Bruce Fischl","doi":"10.1109/isbi53787.2023.10230715","DOIUrl":"https://doi.org/10.1109/isbi53787.2023.10230715","url":null,"abstract":"<p><p>In population and longitudinal imaging studies that employ deformable image registration, more accurate results can be achieved by initializing deformable registration with the results of affine registration where global misalignments have been considerably reduced. Such affine registration, however, is limited to linear transformations and it cannot account for large nonlinear anatomical variations, such as those between pre- and post-operative images or across different subject anatomies. In this work, we introduce a new <i>intermediate</i> deformable image registration (IDIR) technique that recovers large deformations via windowed cross-correlation, and provide an efficient implementation based on the fast Fourier transform. We evaluate our method on 2D X-ray and 3D magnetic resonance images, demonstrating its ability to align substantial nonlinear anatomical variations within a few iterations.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":"2023 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10485808/pdf/nihms-1872292.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10241374","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Foveal avascular zone segmentation using deep learning-driven image-level optimization and fundus photographs. 利用深度学习驱动的图像级优化和眼底照片进行眼窝无血管区分割。
Pub Date : 2023-04-01 Epub Date: 2023-09-01 DOI: 10.1109/isbi53787.2023.10230410
I Coronado, S Pachade, H Dawoodally, S Salazar Marioni, J Yan, R Abdelkhaleq, M Bahrainian, A Jagolino-Cole, R Channa, S A Sheth, L Giancardo

The foveal avascular zone (FAZ) is a retinal area devoid of capillaries and associated with multiple retinal pathologies and visual acuity. Optical Coherence Tomography Angiography (OCT-A) is a very effective means of visualizing retinal vascular and avascular areas, but its use remains limited to research settings due to its complex optics limiting availability. On the other hand, fundus photography is widely available and often adopted in population studies. In this work, we test the feasibility of estimating the FAZ from fundus photos using three different approaches. The first two approaches rely on pixel-level and image-level FAZ information to segment FAZ pixels and regress FAZ area, respectively. The third is a training mask-free pipeline combining saliency maps with an active contours approach to segment FAZ pixels while being trained on image-level measures of the FAZ areas. This enables training FAZ segmentation methods without manual alignment of fundus and OCT-A images, a time-consuming process, which limits the dataset that can be used for training. Segmentation methods trained on pixel-level labels and image-level labels had good agreement with masks from a human grader (respectively DICE of 0.45 and 0.4). Results indicate the feasibility of using fundus images as a proxy to estimate the FAZ when angiography data is not available.

眼窝无血管区(FAZ)是一个没有毛细血管的视网膜区域,与多种视网膜病变和视敏度有关。光学相干断层扫描血管造影术(OCT-A)是观察视网膜血管和无血管区的一种非常有效的方法,但由于其光学结构复杂,可用性受到限制,因此仍仅限于研究环境中使用。另一方面,眼底照相技术应用广泛,经常被用于人群研究。在这项工作中,我们使用三种不同的方法测试了从眼底照片估算 FAZ 的可行性。前两种方法分别依靠像素级和图像级 FAZ 信息来分割 FAZ 像素和回归 FAZ 面积。第三种是无训练掩码管道,结合显著性地图和主动轮廓方法来分割 FAZ 像素,同时根据 FAZ 区域的图像级测量进行训练。这样就可以训练 FAZ 分割方法,而无需手动对准眼底和 OCT-A 图像(这是一个耗时的过程,会限制可用于训练的数据集)。根据像素级标签和图像级标签训练的分割方法与人类分级者的掩膜具有良好的一致性(DICE 分别为 0.45 和 0.4)。结果表明,在没有血管造影数据的情况下,使用眼底图像作为代理来估计 FAZ 是可行的。
{"title":"Foveal avascular zone segmentation using deep learning-driven image-level optimization and fundus photographs.","authors":"I Coronado, S Pachade, H Dawoodally, S Salazar Marioni, J Yan, R Abdelkhaleq, M Bahrainian, A Jagolino-Cole, R Channa, S A Sheth, L Giancardo","doi":"10.1109/isbi53787.2023.10230410","DOIUrl":"10.1109/isbi53787.2023.10230410","url":null,"abstract":"<p><p>The foveal avascular zone (FAZ) is a retinal area devoid of capillaries and associated with multiple retinal pathologies and visual acuity. Optical Coherence Tomography Angiography (OCT-A) is a very effective means of visualizing retinal vascular and avascular areas, but its use remains limited to research settings due to its complex optics limiting availability. On the other hand, fundus photography is widely available and often adopted in population studies. In this work, we test the feasibility of estimating the FAZ from fundus photos using three different approaches. The first two approaches rely on pixel-level and image-level FAZ information to segment FAZ pixels and regress FAZ area, respectively. The third is a training mask-free pipeline combining saliency maps with an active contours approach to segment FAZ pixels while being trained on image-level measures of the FAZ areas. This enables training FAZ segmentation methods without manual alignment of fundus and OCT-A images, a time-consuming process, which limits the dataset that can be used for training. Segmentation methods trained on pixel-level labels and image-level labels had good agreement with masks from a human grader (respectively DICE of 0.45 and 0.4). Results indicate the feasibility of using fundus images as a proxy to estimate the FAZ when angiography data is not available.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":"2023 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10498664/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10264596","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
OPTIMAL TRANSPORT GUIDED UNSUPERVISED LEARNING FOR ENHANCING LOW-QUALITY RETINAL IMAGES. 用于增强低质量视网膜图像的最优运输引导无监督学习。
Pub Date : 2023-04-01 Epub Date: 2023-09-01 DOI: 10.1109/isbi53787.2023.10230719
Wenhui Zhu, Peijie Qiu, Mohammad Farazi, Keshav Nandakumar, Oana M Dumitrascu, Yalin Wang

Real-world non-mydriatic retinal fundus photography is prone to artifacts, imperfections, and low-quality when certain ocular or systemic co-morbidities exist. Artifacts may result in inaccuracy or ambiguity in clinical diagnoses. In this paper, we proposed a simple but effective end-to-end framework for enhancing poor-quality retinal fundus images. Leveraging the optimal transport theory, we proposed an unpaired image-to-image translation scheme for transporting low-quality images to their high-quality counterparts. We theoretically proved that a Generative Adversarial Networks (GAN) model with a generator and discriminator is sufficient for this task. Furthermore, to mitigate the inconsistency of information between the low-quality images and their enhancements, an information consistency mechanism was proposed to maximally maintain structural consistency (optical discs, blood vessels, lesions) between the source and enhanced domains. Extensive experiments were conducted on the EyeQ dataset to demonstrate the superiority of our proposed method perceptually and quantitatively.

当存在某些眼部或全身并发症时,真实世界的非散瞳视网膜眼底摄影容易出现伪影、缺陷和低质量。伪影可能导致临床诊断不准确或不明确。在本文中,我们提出了一种简单但有效的端到端框架来增强低质量的视网膜眼底图像。利用最优传输理论,我们提出了一种不成对的图像到图像转换方案,用于将低质量图像传输到高质量图像。我们从理论上证明了一个具有生成器和鉴别器的生成对抗网络(GAN)模型足以完成这项任务。此外,为了缓解低质量图像及其增强之间的信息不一致,提出了一种信息一致性机制,以最大限度地保持源域和增强域之间的结构一致性(光盘、血管、病变)。在EyeQ数据集上进行了大量实验,以从感知和定量上证明我们提出的方法的优越性。
{"title":"OPTIMAL TRANSPORT GUIDED UNSUPERVISED LEARNING FOR ENHANCING LOW-QUALITY RETINAL IMAGES.","authors":"Wenhui Zhu,&nbsp;Peijie Qiu,&nbsp;Mohammad Farazi,&nbsp;Keshav Nandakumar,&nbsp;Oana M Dumitrascu,&nbsp;Yalin Wang","doi":"10.1109/isbi53787.2023.10230719","DOIUrl":"https://doi.org/10.1109/isbi53787.2023.10230719","url":null,"abstract":"<p><p>Real-world non-mydriatic retinal fundus photography is prone to artifacts, imperfections, and low-quality when certain ocular or systemic co-morbidities exist. Artifacts may result in inaccuracy or ambiguity in clinical diagnoses. In this paper, we proposed a simple but effective end-to-end framework for enhancing poor-quality retinal fundus images. Leveraging the optimal transport theory, we proposed an unpaired image-to-image translation scheme for transporting low-quality images to their high-quality counterparts. We theoretically proved that a Generative Adversarial Networks (GAN) model with a generator and discriminator is sufficient for this task. Furthermore, to mitigate the inconsistency of information between the low-quality images and their enhancements, an information consistency mechanism was proposed to maximally maintain structural consistency (optical discs, blood vessels, lesions) between the source and enhanced domains. Extensive experiments were conducted on the EyeQ dataset to demonstrate the superiority of our proposed method perceptually and quantitatively.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":"2023 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10513403/pdf/nihms-1880846.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41179829","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Predicting Alzheimer's Disease and Quantifying Asymmetric Degeneration of the Hippocampus Using Deep Learning of Magnetic Resonance Imaging Data. 利用磁共振成像数据的深度学习预测阿尔茨海默病和量化海马不对称变性。
Pub Date : 2023-04-01 Epub Date: 2023-09-01 DOI: 10.1109/isbi53787.2023.10230830
Xi Liu, Hongming Li, Yong Fan

In order to quantify lateral asymmetric degeneration of the hippocampus for early predicting Alzheimer's disease (AD), we develop a deep learning (DL) model to learn informative features from the hippocampal magnetic resonance imaging (MRI) data for predicting AD conversion in a time-to-event prediction modeling framework. The DL model is trained on unilateral hippocampal data with an autoencoder based regularizer, facilitating quantification of lateral asymmetry in the hippocampal prediction power of AD conversion and identification of the optimal strategy to integrate the bilateral hippocampal MRI data for predicting AD. Experimental results on MRI scans of 1307 subjects (817 for training and 490 for validation) have demonstrated that the left hippocampus can better predict AD than the right hippocampus, and an integration of the bilateral hippocampal data with the instance based DL method improved AD prediction, compared with alternative predictive modeling strategies.

为了量化海马侧不对称变性以早期预测阿尔茨海默病(AD),我们开发了一个深度学习(DL)模型,从海马磁共振成像(MRI)数据中学习信息特征,以在时间-事件预测模型框架中预测AD转换。DL模型使用基于自动编码器的正则化子在单侧海马数据上进行训练,有助于量化AD转换的海马预测能力的横向不对称性,并确定整合双侧海马MRI数据预测AD的最佳策略。1307名受试者(817名用于训练,490名用于验证)的MRI扫描实验结果表明,与其他预测建模策略相比,左侧海马体比右侧海马体能够更好地预测AD,并且将双侧海马体数据与基于实例的DL方法相结合改进了AD预测。
{"title":"Predicting Alzheimer's Disease and Quantifying Asymmetric Degeneration of the Hippocampus Using Deep Learning of Magnetic Resonance Imaging Data.","authors":"Xi Liu, Hongming Li, Yong Fan","doi":"10.1109/isbi53787.2023.10230830","DOIUrl":"10.1109/isbi53787.2023.10230830","url":null,"abstract":"<p><p>In order to quantify lateral asymmetric degeneration of the hippocampus for early predicting Alzheimer's disease (AD), we develop a deep learning (DL) model to learn informative features from the hippocampal magnetic resonance imaging (MRI) data for predicting AD conversion in a time-to-event prediction modeling framework. The DL model is trained on unilateral hippocampal data with an autoencoder based regularizer, facilitating quantification of lateral asymmetry in the hippocampal prediction power of AD conversion and identification of the optimal strategy to integrate the bilateral hippocampal MRI data for predicting AD. Experimental results on MRI scans of 1307 subjects (817 for training and 490 for validation) have demonstrated that the left hippocampus can better predict AD than the right hippocampus, and an integration of the bilateral hippocampal data with the instance based DL method improved AD prediction, compared with alternative predictive modeling strategies.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":"2023 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10544795/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41170627","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SELF-SUPERVISED LEARNING WITH RADIOLOGY REPORTS, A COMPARATIVE ANALYSIS OF STRATEGIES FOR LARGE VESSEL OCCLUSION AND BRAIN CTA IMAGES. 利用放射学报告进行自我监督学习,对大血管闭塞和脑 cta 图像的策略进行比较分析。
Pub Date : 2023-04-01 Epub Date: 2023-09-01 DOI: 10.1109/isbi53787.2023.10230623
S Pachade, S Datta, Y Dong, S Salazar-Marioni, R Abdelkhaleq, A Niktabe, K Roberts, S A Sheth, L Giancardo

Scarcity of labels for medical images is a significant barrier for training representation learning approaches based on deep neural networks. This limitation is also present when using imaging data collected during routine clinical care stored in picture archiving communication systems (PACS), as these data rarely have attached the high-quality labels required for medical image computing tasks. However, medical images extracted from PACS are commonly coupled with descriptive radiology reports that contain significant information and could be leveraged to pre-train imaging models, which could serve as starting points for further task-specific fine-tuning. In this work, we perform a head-to-head comparison of three different self-supervised strategies to pre-train the same imaging model on 3D brain computed tomography angiogram (CTA) images, with large vessel occlusion (LVO) detection as the downstream task. These strategies evaluate two natural language processing (NLP) approaches, one to extract 100 explicit radiology concepts (Rad-SpatialNet) and the other to create general-purpose radiology reports embeddings (DistilBERT). In addition, we experiment with learning radiology concepts directly or by using a recent self-supervised learning approach (CLIP) that learns by ranking the distance between language and image vector embeddings. The LVO detection task was selected because it requires 3D imaging data, is clinically important, and requires the algorithm to learn outputs not explicitly stated in the radiology report. Pre-training was performed on an unlabeled dataset containing 1,542 3D CTA - reports pairs. The downstream task was tested on a labeled dataset of 402 subjects for LVO. We find that the pre-training performed with CLIP-based strategies improve the performance of the imaging model to detect LVO compared to a model trained only on the labeled data. The best performance was achieved by pre-training using the explicit radiology concepts and CLIP strategy.

医学图像标签的匮乏是训练基于深度神经网络的表示学习方法的一大障碍。在使用存储在图片存档通信系统(PACS)中的常规临床护理过程中收集的成像数据时也存在这种限制,因为这些数据很少附带医学图像计算任务所需的高质量标签。不过,从 PACS 中提取的医学图像通常与包含重要信息的描述性放射学报告结合在一起,可用于预训练成像模型,这些模型可作为进一步针对特定任务进行微调的起点。在这项工作中,我们对三种不同的自监督策略进行了正面比较,以在三维脑计算机断层扫描血管造影(CTA)图像上预训练相同的成像模型,并将大血管闭塞(LVO)检测作为下游任务。这些策略评估了两种自然语言处理 (NLP) 方法,一种用于提取 100 个明确的放射学概念(Rad-SpatialNet),另一种用于创建通用放射学报告嵌入(DistilBERT)。此外,我们还尝试了直接学习放射学概念或使用最新的自我监督学习方法(CLIP)学习放射学概念,该方法通过对语言和图像向量嵌入之间的距离进行排序来学习。之所以选择 LVO 检测任务,是因为它需要三维成像数据,在临床上非常重要,而且要求算法学习放射学报告中没有明确说明的输出结果。预训练在一个包含 1,542 对 3D CTA - 报告的无标记数据集上进行。下游任务在一个包含 402 名 LVO 受试者的标注数据集上进行了测试。我们发现,与仅在标记数据上训练的模型相比,使用基于 CLIP 的策略进行的预训练提高了成像模型检测 LVO 的性能。使用明确的放射学概念和 CLIP 策略进行预训练可获得最佳性能。
{"title":"SELF-SUPERVISED LEARNING WITH RADIOLOGY REPORTS, A COMPARATIVE ANALYSIS OF STRATEGIES FOR LARGE VESSEL OCCLUSION AND BRAIN CTA IMAGES.","authors":"S Pachade, S Datta, Y Dong, S Salazar-Marioni, R Abdelkhaleq, A Niktabe, K Roberts, S A Sheth, L Giancardo","doi":"10.1109/isbi53787.2023.10230623","DOIUrl":"10.1109/isbi53787.2023.10230623","url":null,"abstract":"<p><p>Scarcity of labels for medical images is a significant barrier for training representation learning approaches based on deep neural networks. This limitation is also present when using imaging data collected during routine clinical care stored in picture archiving communication systems (PACS), as these data rarely have attached the high-quality labels required for medical image computing tasks. However, medical images extracted from PACS are commonly coupled with descriptive radiology reports that contain significant information and could be leveraged to pre-train imaging models, which could serve as starting points for further task-specific fine-tuning. In this work, we perform a head-to-head comparison of three different self-supervised strategies to pre-train the same imaging model on 3D brain computed tomography angiogram (CTA) images, with large vessel occlusion (LVO) detection as the downstream task. These strategies evaluate two natural language processing (NLP) approaches, one to extract 100 explicit radiology concepts (Rad-SpatialNet) and the other to create general-purpose radiology reports embeddings (DistilBERT). In addition, we experiment with learning radiology concepts directly or by using a recent self-supervised learning approach (CLIP) that learns by ranking the distance between language and image vector embeddings. The LVO detection task was selected because it requires 3D imaging data, is clinically important, and requires the algorithm to learn outputs not explicitly stated in the radiology report. Pre-training was performed on an unlabeled dataset containing 1,542 3D CTA - reports pairs. The downstream task was tested on a labeled dataset of 402 subjects for LVO. We find that the pre-training performed with CLIP-based strategies improve the performance of the imaging model to detect LVO compared to a model trained only on the labeled data. The best performance was achieved by pre-training using the explicit radiology concepts and CLIP strategy.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":"2023 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10498780/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10306669","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SurfNN: Joint Reconstruction of Multiple Cortical Surfaces from Magnetic Resonance Images. SurfNN:从磁共振图像中联合重建多个皮层表面。
Pub Date : 2023-04-01 Epub Date: 2023-09-01 DOI: 10.1109/isbi53787.2023.10230488
Hao Zheng, Hongming Li, Yong Fan

To achieve fast, robust, and accurate reconstruction of the human cortical surfaces from 3D magnetic resonance images (MRIs), we develop a novel deep learning-based framework, referred to as SurfNN, to reconstruct simultaneously both inner (between white matter and gray matter) and outer (pial) surfaces from MRIs. Different from existing deep learning-based cortical surface reconstruction methods that either reconstruct the cortical surfaces separately or neglect the interdependence between the inner and outer surfaces, SurfNN reconstructs both the inner and outer cortical surfaces jointly by training a single network to predict a midthickness surface that lies at the center of the inner and outer cortical surfaces. The input of SurfNN consists of a 3D MRI and an initialization of the midthickness surface that is represented both implicitly as a 3D distance map and explicitly as a triangular mesh with spherical topology, and its output includes both the inner and outer cortical surfaces, as well as the midthickness surface. The method has been evaluated on a large-scale MRI dataset and demonstrated competitive cortical surface reconstruction performance.

为了从3D磁共振图像(MRI)中快速、稳健和准确地重建人类皮层表面,我们开发了一种新的基于深度学习的框架,称为SurfNN,以从MRI中同时重建内部(白质和灰质之间)和外部(pial)表面。与现有的基于深度学习的皮层表面重建方法不同,该方法要么单独重建皮层表面,SurfNN通过训练单个网络来预测位于皮层内外表面中心的中厚表面,从而联合重建皮层内外表面。SurfNN的输入包括3D MRI和中厚表面的初始化,该表面隐式表示为3D距离图,显式表示为具有球形拓扑结构的三角形网格,其输出包括皮层内表面和皮层外表面以及中厚表面。该方法已在大规模MRI数据集上进行了评估,并证明了具有竞争力的皮层表面重建性能。
{"title":"<i>Surf</i>NN: Joint Reconstruction of Multiple Cortical Surfaces from Magnetic Resonance Images.","authors":"Hao Zheng, Hongming Li, Yong Fan","doi":"10.1109/isbi53787.2023.10230488","DOIUrl":"10.1109/isbi53787.2023.10230488","url":null,"abstract":"<p><p>To achieve fast, robust, and accurate reconstruction of the human cortical surfaces from 3D magnetic resonance images (MRIs), we develop a novel deep learning-based framework, referred to as <i>Surf</i>NN, to reconstruct simultaneously both inner (between white matter and gray matter) and outer (pial) surfaces from MRIs. Different from existing deep learning-based cortical surface reconstruction methods that either reconstruct the cortical surfaces separately or neglect the interdependence between the inner and outer surfaces, <i>Surf</i>NN reconstructs both the inner and outer cortical surfaces jointly by training a single network to predict a midthickness surface that lies at the center of the inner and outer cortical surfaces. The input of <i>Surf</i>NN consists of a 3D MRI and an initialization of the midthickness surface that is represented both implicitly as a 3D distance map and explicitly as a triangular mesh with spherical topology, and its output includes both the inner and outer cortical surfaces, as well as the midthickness surface. The method has been evaluated on a large-scale MRI dataset and demonstrated competitive cortical surface reconstruction performance.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":"2023 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10544794/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"41176109","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
End-to-end First Trimester Fetal Ultrasound Video Automated CRL and NT Segmentation. 端到端第一孕期胎儿超声视频自动 CRL 和 NT 分段。
Pub Date : 2022-04-28 DOI: 10.1109/ISBI52829.2022.9761400
Robail Yasrab, Zeyu Fu, Lior Drukker, Lok Hin Lee, He Zhao, Aris T Papageorghiou, Alison J Noble

This study presents a novel approach to automatic detection and segmentation of the Crown Rump Length (CRL) and Nuchal Translucency (NT), two essential measurements in the first trimester US scan. The proposed method automatically localises a standard plane within a video clip as defined by the UK Fetal Abnormality Screening Programme. A Nested Hourglass (NHG) based network performs semantic pixel-wise segmentation to extract NT and CRL structures. Our results show that the NHG network is faster (19.52% < GFlops than FCN32) and offers high pixel agreement (mean-IoU=80.74) with expert manual annotations.

本研究提出了一种自动检测和分割胎儿臀长(CRL)和颈部透明层(NT)的新方法,这两项指标是孕期前三个月 US 扫描的基本测量指标。根据英国胎儿畸形筛查计划的规定,该方法可自动定位视频片段中的标准平面。基于嵌套沙漏(NHG)的网络执行语义像素分割,以提取 NT 和 CRL 结构。我们的结果表明,NHG 网络的速度更快(比 FCN32 快 19.52%),与专家手动注释的像素一致性更高(平均 IoU=80.74)。
{"title":"End-to-end First Trimester Fetal Ultrasound Video Automated CRL and NT Segmentation.","authors":"Robail Yasrab, Zeyu Fu, Lior Drukker, Lok Hin Lee, He Zhao, Aris T Papageorghiou, Alison J Noble","doi":"10.1109/ISBI52829.2022.9761400","DOIUrl":"10.1109/ISBI52829.2022.9761400","url":null,"abstract":"<p><p>This study presents a novel approach to automatic detection and segmentation of the Crown Rump Length (CRL) and Nuchal Translucency (NT), two essential measurements in the first trimester US scan. The proposed method automatically localises a standard plane within a video clip as defined by the UK Fetal Abnormality Screening Programme. A Nested Hourglass (NHG) based network performs semantic pixel-wise segmentation to extract NT and CRL structures. Our results show that the NHG network is faster (19.52% < GFlops than FCN32) and offers high pixel agreement (mean-IoU=80.74) with expert manual annotations.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":"2022 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2022-04-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7614066/pdf/EMS159390.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9359298","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
First Trimester video Saliency Prediction using CLSTMU-NET with Stochastic Augmentation. 基于随机增强的CLSTMU-NET的早期妊娠视频显著性预测。
Pub Date : 2022-04-26 DOI: 10.1109/ISBI52829.2022.9761585
Elizaveta Savochkina, Lok Hin Lee, He Zhao, Lior Drukker, Aris T Papageorghiou, J Alison Noble

In this paper we develop a multi-modal video analysis algorithm to predict where a sonographer should look next. Our approach uses video and expert knowledge, defined by gaze tracking data, which is acquired during routine first-trimester fetal ultrasound scanning. Specifically, we propose a spatio-temporal convolutional LSTMU-Net neural network (cLSTMU-Net) for video saliency prediction with stochastic augmentation. The architecture design consists of a U-Net based encoder-decoder network and a cLSTM to take into account temporal information. We compare the performance of the cLSTMU-Net alongside spatial-only architectures for the task of predicting gaze in first trimester ultrasound videos. Our study dataset consists of 115 clinically acquired first trimester US videos and a total of 45, 666 video frames. We adopt a Random Augmentation strategy (RA) from a stochastic augmentation policy search to improve model performance and reduce over-fitting. The proposed cLSTMU-Net using a video clip of 6 frames outperforms the baseline approach on all saliency metrics: KLD, SIM, NSS and CC (2.08, 0.28, 4.53 and 0.42 versus 2.16, 0.27, 4.34 and 0.39).

在本文中,我们开发了一种多模态视频分析算法来预测超声医师下一步应该看哪里。我们的方法使用视频和专家知识,由凝视跟踪数据定义,这些数据是在常规妊娠早期胎儿超声扫描中获得的。具体来说,我们提出了一种时空卷积LSTMU-Net神经网络(cLSTMU-Net),用于随机增强视频显著性预测。架构设计包括基于U-Net的编码器-解码器网络和考虑时序信息的cLSTM。我们比较了cLSTMU-Net和纯空间架构在预测妊娠早期超声视频凝视任务中的性能。我们的研究数据集包括115个临床获得的妊娠早期美国视频和总共45,666个视频帧。我们采用随机增强策略搜索中的随机增强策略(RA)来提高模型性能并减少过拟合。使用6帧视频剪辑的拟议cLSTMU-Net在所有显着性指标上优于基线方法:KLD, SIM, NSS和CC(2.08, 0.28, 4.53和0.42相对于2.16,0.27,4.34和0.39)。
{"title":"First Trimester video Saliency Prediction using CLSTMU-NET with Stochastic Augmentation.","authors":"Elizaveta Savochkina, Lok Hin Lee, He Zhao, Lior Drukker, Aris T Papageorghiou, J Alison Noble","doi":"10.1109/ISBI52829.2022.9761585","DOIUrl":"10.1109/ISBI52829.2022.9761585","url":null,"abstract":"<p><p>In this paper we develop a multi-modal video analysis algorithm to predict where a sonographer should look next. Our approach uses video and expert knowledge, defined by gaze tracking data, which is acquired during routine first-trimester fetal ultrasound scanning. Specifically, we propose a spatio-temporal convolutional LSTMU-Net neural network (cLSTMU-Net) for video saliency prediction with stochastic augmentation. The architecture design consists of a U-Net based encoder-decoder network and a cLSTM to take into account temporal information. We compare the performance of the cLSTMU-Net alongside spatial-only architectures for the task of predicting gaze in first trimester ultrasound videos. Our study dataset consists of 115 clinically acquired first trimester US videos and a total of 45, 666 video frames. We adopt a Random Augmentation strategy (RA) from a stochastic augmentation policy search to improve model performance and reduce over-fitting. The proposed cLSTMU-Net using a video clip of 6 frames outperforms the baseline approach on all saliency metrics: KLD, SIM, NSS and CC (2.08, 0.28, 4.53 and 0.42 versus 2.16, 0.27, 4.34 and 0.39).</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":"2022 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2022-04-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7614063/pdf/EMS159391.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9350355","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SELF-SEMANTIC CONTOUR ADAPTATION FOR CROSS MODALITY BRAIN TUMOR SEGMENTATION. 自语义轮廓自适应跨模态脑肿瘤分割。
Pub Date : 2022-03-01 Epub Date: 2022-04-26 DOI: 10.1109/isbi52829.2022.9761629
Xiaofeng Liu, Fangxu Xing, Georges El Fakhri, Jonghye Woo

Unsupervised domain adaptation (UDA) between two significantly disparate domains to learn high-level semantic alignment is a crucial yet challenging task. To this end, in this work, we propose exploiting low-level edge information to facilitate the adaptation as a precursor task, which has a small cross-domain gap, compared with semantic segmentation. The precise contour then provides spatial information to guide the semantic adaptation. More specifically, we propose a multi-task framework to learn a contouring adaptation network along with a semantic segmentation adaptation network, which takes both magnetic resonance imaging (MRI) slice and its initial edge map as input. These two networks are jointly trained with source domain labels, and the feature and edge map level adversarial learning is carried out for cross-domain alignment. In addition, self-entropy minimization is incorporated to further enhance segmentation performance. We evaluated our framework on the BraTS2018 database for cross-modality segmentation of brain tumors, showing the validity and superiority of our approach, compared with competing methods.

在两个明显不同的领域之间进行无监督域自适应(UDA)以学习高级语义对齐是一项至关重要但具有挑战性的任务。为此,在本工作中,我们提出利用底层边缘信息来促进自适应作为前置任务,与语义分割相比,该任务具有较小的跨域差距。然后,精确的轮廓提供空间信息来指导语义适应。更具体地说,我们提出了一个多任务框架来学习轮廓自适应网络和语义分割自适应网络,该网络以磁共振成像(MRI)切片及其初始边缘图为输入。利用源域标签对这两个网络进行联合训练,并进行特征和边缘映射级对抗学习进行跨域对齐。此外,引入自熵最小化来进一步提高分割性能。我们在BraTS2018数据库上评估了我们的框架用于脑肿瘤的跨模态分割,与竞争对手的方法相比,显示了我们的方法的有效性和优越性。
{"title":"SELF-SEMANTIC CONTOUR ADAPTATION FOR CROSS MODALITY BRAIN TUMOR SEGMENTATION.","authors":"Xiaofeng Liu,&nbsp;Fangxu Xing,&nbsp;Georges El Fakhri,&nbsp;Jonghye Woo","doi":"10.1109/isbi52829.2022.9761629","DOIUrl":"https://doi.org/10.1109/isbi52829.2022.9761629","url":null,"abstract":"<p><p>Unsupervised domain adaptation (UDA) between two significantly disparate domains to learn high-level semantic alignment is a crucial yet challenging task. To this end, in this work, we propose exploiting low-level edge information to facilitate the adaptation as a precursor task, which has a small cross-domain gap, compared with semantic segmentation. The precise contour then provides spatial information to guide the semantic adaptation. More specifically, we propose a multi-task framework to learn a contouring adaptation network along with a semantic segmentation adaptation network, which takes both magnetic resonance imaging (MRI) slice and its initial edge map as input. These two networks are jointly trained with source domain labels, and the feature and edge map level adversarial learning is carried out for cross-domain alignment. In addition, self-entropy minimization is incorporated to further enhance segmentation performance. We evaluated our framework on the BraTS2018 database for cross-modality segmentation of brain tumors, showing the validity and superiority of our approach, compared with competing methods.</p>","PeriodicalId":74566,"journal":{"name":"Proceedings. IEEE International Symposium on Biomedical Imaging","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2022-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9387767/pdf/nihms-1779009.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"40431593","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
期刊
Proceedings. IEEE International Symposium on Biomedical Imaging
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1