首页 > 最新文献

Machine learning in medical imaging. MLMI (Workshop)最新文献

英文 中文
Probabilistic 3D Correspondence Prediction from Sparse Unsegmented Images. 从稀疏未分割图像进行概率三维对应预测
Pub Date : 2025-01-01 Epub Date: 2024-10-23 DOI: 10.1007/978-3-031-73290-4_12
Krithika Iyer, Shireen Y Elhabian

The study of physiology demonstrates that the form (shape) of anatomical structures dictates their functions, and analyzing the form of anatomies plays a crucial role in clinical research. Statistical shape modeling (SSM) is a widely used tool for quantitative analysis of forms of anatomies, aiding in characterizing and identifying differences within a population of subjects. Despite its utility, the conventional SSM construction pipeline is often complex and time-consuming. Additionally, reliance on linearity assumptions further limits the model from capturing clinically relevant variations. Recent advancements in deep learning solutions enable the direct inference of SSM from unsegmented medical images, streamlining the process and improving accessibility. However, the new methods of SSM from images do not adequately account for situations where the imaging data quality is poor or where only sparse information is available. Moreover, quantifying aleatoric uncertainty, which represents inherent data variability, is crucial in deploying deep learning for clinical tasks to ensure reliable model predictions and robust decision-making, especially in challenging imaging conditions. Therefore, we propose SPI-CorrNet, a unified model that predicts 3D correspondences from sparse imaging data. It leverages a teacher network to regularize feature learning and quantifies data-dependent aleatoric uncertainty by adapting the network to predict intrinsic input variances. Experiments on the LGE MRI left atrium dataset and Abdomen CT-1K liver datasets demonstrate that our technique enhances the accuracy and robustness of sparse image-driven SSM.

生理学研究表明,解剖结构的形态(形状)决定了其功能,分析解剖结构的形态在临床研究中起着至关重要的作用。统计形状建模(SSM)是一种广泛使用的工具,用于对解剖结构的形状进行定量分析,帮助描述和识别受试者群体中的差异。尽管它很有用,但传统的 SSM 构建流程往往复杂且耗时。此外,对线性假设的依赖进一步限制了模型捕捉临床相关变异的能力。深度学习解决方案的最新进展使得从未分类的医学图像中直接推断 SSM 成为可能,从而简化了流程并提高了可及性。然而,从图像推断 SSM 的新方法并不能充分考虑成像数据质量较差或仅有稀疏信息的情况。此外,量化代表固有数据变异性的不确定性(aleatoric uncertainty)对于将深度学习应用于临床任务以确保可靠的模型预测和稳健的决策至关重要,尤其是在具有挑战性的成像条件下。因此,我们提出了 SPI-CorrNet 模型,这是一个能从稀疏成像数据中预测三维对应关系的统一模型。它利用教师网络来正则化特征学习,并通过调整网络来预测内在输入方差,从而量化与数据相关的不确定性。在 LGE MRI 左心房数据集和腹部 CT-1K 肝脏数据集上的实验表明,我们的技术提高了稀疏图像驱动 SSM 的准确性和鲁棒性。
{"title":"Probabilistic 3D Correspondence Prediction from Sparse Unsegmented Images.","authors":"Krithika Iyer, Shireen Y Elhabian","doi":"10.1007/978-3-031-73290-4_12","DOIUrl":"10.1007/978-3-031-73290-4_12","url":null,"abstract":"<p><p>The study of physiology demonstrates that the form (shape) of anatomical structures dictates their functions, and analyzing the form of anatomies plays a crucial role in clinical research. Statistical shape modeling (SSM) is a widely used tool for quantitative analysis of forms of anatomies, aiding in characterizing and identifying differences within a population of subjects. Despite its utility, the conventional SSM construction pipeline is often complex and time-consuming. Additionally, reliance on linearity assumptions further limits the model from capturing clinically relevant variations. Recent advancements in deep learning solutions enable the direct inference of SSM from unsegmented medical images, streamlining the process and improving accessibility. However, the new methods of SSM from images do not adequately account for situations where the imaging data quality is poor or where only sparse information is available. Moreover, quantifying aleatoric uncertainty, which represents inherent data variability, is crucial in deploying deep learning for clinical tasks to ensure reliable model predictions and robust decision-making, especially in challenging imaging conditions. Therefore, we propose SPI-CorrNet, a unified model that predicts 3D correspondences from sparse imaging data. It leverages a teacher network to regularize feature learning and quantifies data-dependent aleatoric uncertainty by adapting the network to predict intrinsic input variances. Experiments on the LGE MRI left atrium dataset and Abdomen CT-1K liver datasets demonstrate that our technique enhances the accuracy and robustness of sparse image-driven SSM.</p>","PeriodicalId":74092,"journal":{"name":"Machine learning in medical imaging. MLMI (Workshop)","volume":"15242 ","pages":"117-127"},"PeriodicalIF":0.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11568407/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142649865","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Robust Unsupervised Super-Resolution of Infant MRI via Dual-Modal Deep Image Prior. 通过双模态深度图像先验实现婴儿磁共振成像的稳健无监督超分辨率
Pub Date : 2024-01-01 Epub Date: 2023-10-15 DOI: 10.1007/978-3-031-45673-2_5
Cheng Che Tsai, Xiaoyang Chen, Sahar Ahmad, Pew-Thian Yap

Magnetic resonance imaging (MRI) is commonly used for studying infant brain development. However, due to the lengthy image acquisition time and limited subject compliance, high-quality infant MRI can be challenging. Without imposing additional burden on image acquisition, image super-resolution (SR) can be used to enhance image quality post-acquisition. Most SR techniques are supervised and trained on multiple aligned low-resolution (LR) and high-resolution (HR) image pairs, which in practice are not usually available. Unlike supervised approaches, Deep Image Prior (DIP) can be employed for unsupervised single-image SR, utilizing solely the input LR image for de novo optimization to produce an HR image. However, determining when to stop early in DIP training is non-trivial and presents a challenge to fully automating the SR process. To address this issue, we constrain the low-frequency k-space of the SR image to be similar to that of the LR image. We further improve performance by designing a dual-modal framework that leverages shared anatomical information between T1-weighted and T2-weighted images. We evaluated our model, dual-modal DIP (dmDIP), on infant MRI data acquired from birth to one year of age, demonstrating that enhanced image quality can be obtained with substantially reduced sensitivity to early stopping.

磁共振成像(MRI)通常用于研究婴儿的大脑发育。然而,由于图像采集时间长、受试者服从性有限,高质量的婴儿磁共振成像具有挑战性。在不增加图像采集负担的情况下,图像超分辨率(SR)可用于提高采集后的图像质量。大多数超分辨率技术都是有监督的,并在多个对齐的低分辨率(LR)和高分辨率(HR)图像对上进行训练,但实际上通常无法获得这些图像对。与有监督的方法不同,深度图像优先(DIP)可用于无监督的单图像 SR,仅利用输入的低分辨率图像进行全新优化,生成高分辨率图像。然而,确定何时停止 DIP 训练并非易事,这对 SR 过程的完全自动化提出了挑战。为了解决这个问题,我们限制 SR 图像的低频 k 空间与 LR 图像的低频 k 空间相似。我们通过设计双模态框架,利用 T1 加权和 T2 加权图像之间共享的解剖信息,进一步提高了性能。我们在从出生到一岁的婴儿磁共振成像数据上评估了我们的模型--双模态 DIP(dmDIP),结果表明,在大幅降低对早期停跳敏感性的同时,还能获得更高的图像质量。
{"title":"Robust Unsupervised Super-Resolution of Infant MRI via Dual-Modal Deep Image Prior.","authors":"Cheng Che Tsai, Xiaoyang Chen, Sahar Ahmad, Pew-Thian Yap","doi":"10.1007/978-3-031-45673-2_5","DOIUrl":"10.1007/978-3-031-45673-2_5","url":null,"abstract":"<p><p>Magnetic resonance imaging (MRI) is commonly used for studying infant brain development. However, due to the lengthy image acquisition time and limited subject compliance, high-quality infant MRI can be challenging. Without imposing additional burden on image acquisition, image super-resolution (SR) can be used to enhance image quality post-acquisition. Most SR techniques are supervised and trained on multiple aligned low-resolution (LR) and high-resolution (HR) image pairs, which in practice are not usually available. Unlike supervised approaches, Deep Image Prior (DIP) can be employed for unsupervised single-image SR, utilizing solely the input LR image for <i>de novo</i> optimization to produce an HR image. However, determining when to stop early in DIP training is non-trivial and presents a challenge to fully automating the SR process. To address this issue, we constrain the low-frequency k-space of the SR image to be similar to that of the LR image. We further improve performance by designing a dual-modal framework that leverages shared anatomical information between T1-weighted and T2-weighted images. We evaluated our model, dual-modal DIP (dmDIP), on infant MRI data acquired from birth to one year of age, demonstrating that enhanced image quality can be obtained with substantially reduced sensitivity to early stopping.</p>","PeriodicalId":74092,"journal":{"name":"Machine learning in medical imaging. MLMI (Workshop)","volume":"14348 ","pages":"42-51"},"PeriodicalIF":0.0,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11323077/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141989667","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MoViT: Memorizing Vision Transformers for Medical Image Analysis. MoViT:为医学图像分析记忆视觉变换器。
Pub Date : 2024-01-01 Epub Date: 2023-10-15 DOI: 10.1007/978-3-031-45676-3_21
Yiqing Shen, Pengfei Guo, Jingpu Wu, Qianqi Huang, Nhat Le, Jinyuan Zhou, Shanshan Jiang, Mathias Unberath

The synergy of long-range dependencies from transformers and local representations of image content from convolutional neural networks (CNNs) has led to advanced architectures and increased performance for various medical image analysis tasks due to their complementary benefits. However, compared with CNNs, transformers require considerably more training data, due to a larger number of parameters and an absence of inductive bias. The need for increasingly large datasets continues to be problematic, particularly in the context of medical imaging, where both annotation efforts and data protection result in limited data availability. In this work, inspired by the human decision-making process of correlating new "evidence" with previously memorized "experience", we propose a Memorizing Vision Transformer (MoViT) to alleviate the need for large-scale datasets to successfully train and deploy transformer-based architectures. MoViT leverages an external memory structure to cache history attention snapshots during the training stage. To prevent overfitting, we incorporate an innovative memory update scheme, attention temporal moving average, to update the stored external memories with the historical moving average. For inference speedup, we design a prototypical attention learning method to distill the external memory into smaller representative subsets. We evaluate our method on a public histology image dataset and an in-house MRI dataset, demonstrating that MoViT applied to varied medical image analysis tasks, can outperform vanilla transformer models across varied data regimes, especially in cases where only a small amount of annotated data is available. More importantly, MoViT can reach a competitive performance of ViT with only 3.0% of the training data. In conclusion, MoViT provides a simple plug-in for transformer architectures which may contribute to reducing the training data needed to achieve acceptable models for a broad range of medical image analysis tasks.

变压器的长程依赖性和卷积神经网络(CNN)对图像内容的局部表征的协同作用,为各种医学图像分析任务提供了先进的架构和更高的性能,因为它们具有互补优势。然而,与卷积神经网络相比,变换器需要更多的训练数据,这是因为变换器需要更多的参数,而且不存在归纳偏差。对越来越大的数据集的需求仍然是个问题,特别是在医学成像领域,注释工作和数据保护都导致数据可用性有限。在这项工作中,受将新 "证据 "与先前记忆的 "经验 "关联起来的人类决策过程的启发,我们提出了记忆视觉转换器(MoViT),以缓解对大规模数据集的需求,从而成功地训练和部署基于转换器的架构。MoViT 利用外部内存结构,在训练阶段缓存历史注意力快照。为防止过度拟合,我们采用了一种创新的内存更新方案--注意力时空移动平均法,用历史移动平均值更新存储的外部内存。为了加快推理速度,我们设计了一种原型注意力学习方法,将外部记忆提炼为更小的代表性子集。我们在一个公共组织学图像数据集和一个内部核磁共振成像数据集上对我们的方法进行了评估,结果表明,将 MoViT 应用于各种医学图像分析任务时,它在各种数据环境下的表现都优于香草变换器模型,尤其是在只有少量注释数据的情况下。更重要的是,MoViT 只需 3.0% 的训练数据就能达到 ViT 的竞争性能。总之,MoViT 为变换器架构提供了一个简单的插件,它可以帮助减少训练数据,从而为广泛的医学图像分析任务建立可接受的模型。
{"title":"MoViT: Memorizing Vision Transformers for Medical Image Analysis.","authors":"Yiqing Shen, Pengfei Guo, Jingpu Wu, Qianqi Huang, Nhat Le, Jinyuan Zhou, Shanshan Jiang, Mathias Unberath","doi":"10.1007/978-3-031-45676-3_21","DOIUrl":"https://doi.org/10.1007/978-3-031-45676-3_21","url":null,"abstract":"<p><p>The synergy of long-range dependencies from transformers and local representations of image content from convolutional neural networks (CNNs) has led to advanced architectures and increased performance for various medical image analysis tasks due to their complementary benefits. However, compared with CNNs, transformers require considerably more training data, due to a larger number of parameters and an absence of inductive bias. The need for increasingly large datasets continues to be problematic, particularly in the context of medical imaging, where both annotation efforts and data protection result in limited data availability. In this work, inspired by the human decision-making process of correlating new \"evidence\" with previously memorized \"experience\", we propose a Memorizing Vision Transformer (MoViT) to alleviate the need for large-scale datasets to successfully train and deploy transformer-based architectures. MoViT leverages an external memory structure to cache history attention snapshots during the training stage. To prevent overfitting, we incorporate an innovative memory update scheme, attention temporal moving average, to update the stored external memories with the historical moving average. For inference speedup, we design a prototypical attention learning method to distill the external memory into smaller representative subsets. We evaluate our method on a public histology image dataset and an in-house MRI dataset, demonstrating that MoViT applied to varied medical image analysis tasks, can outperform vanilla transformer models across varied data regimes, especially in cases where only a small amount of annotated data is available. More importantly, MoViT can reach a competitive performance of ViT with only 3.0% of the training data. In conclusion, MoViT provides a simple plug-in for transformer architectures which may contribute to reducing the training data needed to achieve acceptable models for a broad range of medical image analysis tasks.</p>","PeriodicalId":74092,"journal":{"name":"Machine learning in medical imaging. MLMI (Workshop)","volume":"14349 ","pages":"205-213"},"PeriodicalIF":0.0,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11008051/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140863811","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Class-Balanced Deep Learning with Adaptive Vector Scaling Loss for Dementia Stage Detection. 利用自适应向量缩放损失的类平衡深度学习进行痴呆症阶段检测。
Pub Date : 2024-01-01 Epub Date: 2023-10-15 DOI: 10.1007/978-3-031-45676-3_15
Boning Tong, Zhuoping Zhou, Davoud Ataee Tarzanagh, Bojian Hou, Andrew J Saykin, Jason Moore, Marylyn Ritchie, Li Shen

Alzheimer's disease (AD) leads to irreversible cognitive decline, with Mild Cognitive Impairment (MCI) as its prodromal stage. Early detection of AD and related dementia is crucial for timely treatment and slowing disease progression. However, classifying cognitive normal (CN), MCI, and AD subjects using machine learning models faces class imbalance, necessitating the use of balanced accuracy as a suitable metric. To enhance model performance and balanced accuracy, we introduce a novel method called VS-Opt-Net. This approach incorporates the recently developed vector-scaling (VS) loss into a machine learning pipeline named STREAMLINE. Moreover, it employs Bayesian optimization for hyperparameter learning of both the model and loss function. VS-Opt-Net not only amplifies the contribution of minority examples in proportion to the imbalance level but also addresses the challenge of generalization in training deep networks. In our empirical study, we use MRI-based brain regional measurements as features to conduct the CN vs MCI and AD vs MCI binary classifications. We compare the balanced accuracy of our model with other machine learning models and deep neural network loss functions that also employ class-balanced strategies. Our findings demonstrate that after hyperparameter optimization, the deep neural network using the VS loss function substantially improves balanced accuracy. It also surpasses other models in performance on the AD dataset. Moreover, our feature importance analysis highlights VS-Opt-Net's ability to elucidate biomarker differences across dementia stages.

阿尔茨海默病(AD)会导致不可逆的认知能力下降,轻度认知障碍(MCI)是其前驱阶段。早期发现阿尔茨海默病和相关痴呆症对于及时治疗和延缓疾病进展至关重要。然而,使用机器学习模型对认知正常(CN)、MCI 和 AD 受试者进行分类面临着类别不平衡的问题,因此有必要使用平衡准确性作为合适的衡量标准。为了提高模型性能和平衡准确性,我们引入了一种名为 VS-Opt-Net 的新方法。这种方法将最近开发的向量缩放(VS)损失纳入名为 STREAMLINE 的机器学习管道中。此外,它还采用贝叶斯优化方法对模型和损失函数进行超参数学习。VS-Opt-Net 不仅能根据不平衡程度放大少数实例的贡献,还能解决深度网络训练中的泛化难题。在实证研究中,我们使用基于 MRI 的大脑区域测量作为特征,进行 CN vs MCI 和 AD vs MCI 的二元分类。我们将模型的平衡准确性与其他同样采用类平衡策略的机器学习模型和深度神经网络损失函数进行了比较。我们的研究结果表明,经过超参数优化后,使用 VS 损失函数的深度神经网络大大提高了均衡准确率。它在 AD 数据集上的表现也超过了其他模型。此外,我们的特征重要性分析突出了 VS-Opt-Net 在阐明不同痴呆症阶段的生物标记物差异方面的能力。
{"title":"Class-Balanced Deep Learning with Adaptive Vector Scaling Loss for Dementia Stage Detection.","authors":"Boning Tong, Zhuoping Zhou, Davoud Ataee Tarzanagh, Bojian Hou, Andrew J Saykin, Jason Moore, Marylyn Ritchie, Li Shen","doi":"10.1007/978-3-031-45676-3_15","DOIUrl":"10.1007/978-3-031-45676-3_15","url":null,"abstract":"<p><p>Alzheimer's disease (AD) leads to irreversible cognitive decline, with Mild Cognitive Impairment (MCI) as its prodromal stage. Early detection of AD and related dementia is crucial for timely treatment and slowing disease progression. However, classifying cognitive normal (CN), MCI, and AD subjects using machine learning models faces class imbalance, necessitating the use of balanced accuracy as a suitable metric. To enhance model performance and balanced accuracy, we introduce a novel method called VS-Opt-Net. This approach incorporates the recently developed vector-scaling (VS) loss into a machine learning pipeline named STREAMLINE. Moreover, it employs Bayesian optimization for hyperparameter learning of both the model and loss function. VS-Opt-Net not only amplifies the contribution of minority examples in proportion to the imbalance level but also addresses the challenge of generalization in training deep networks. In our empirical study, we use MRI-based brain regional measurements as features to conduct the CN vs MCI and AD vs MCI binary classifications. We compare the balanced accuracy of our model with other machine learning models and deep neural network loss functions that also employ class-balanced strategies. Our findings demonstrate that after hyperparameter optimization, the deep neural network using the VS loss function substantially improves balanced accuracy. It also surpasses other models in performance on the AD dataset. Moreover, our feature importance analysis highlights VS-Opt-Net's ability to elucidate biomarker differences across dementia stages.</p>","PeriodicalId":74092,"journal":{"name":"Machine learning in medical imaging. MLMI (Workshop)","volume":"14349 ","pages":"144-154"},"PeriodicalIF":0.0,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10924683/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140095306","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Bayesian Quantization for Supervised Neuroimage Search. 用于监督神经图像搜索的深度贝叶斯量化。
Pub Date : 2023-10-01 Epub Date: 2023-10-15 DOI: 10.1007/978-3-031-45676-3_40
Erkun Yang, Cheng Deng, Mingxia Liu

Neuroimage retrieval plays a crucial role in providing physicians with access to previous similar cases, which is essential for case-based reasoning and evidence-based medicine. Due to low computation and storage costs, hashing-based search techniques have been widely adopted for establishing image retrieval systems. However, these methods often suffer from nonnegligible quantization loss, which can degrade the overall search performance. To address this issue, this paper presents a compact coding solution namely Deep Bayesian Quantization (DBQ), which focuses on deep compact quantization that can estimate continuous neuroimage representations and achieve superior performance over existing hashing solutions. Specifically, DBQ seamlessly combines the deep representation learning and the representation compact quantization within a novel Bayesian learning framework, where a proxy embedding-based likelihood function is developed to alleviate the sampling issue for traditional similarity supervision. Additionally, a Gaussian prior is employed to reduce the quantization losses. By utilizing pre-computed lookup tables, the proposed DBQ can enable efficient and effective similarity search. Extensive experiments conducted on 2, 008 structural MRI scans from three benchmark neuroimage datasets demonstrate that our method outperforms previous state-of-the-arts.

神经图像检索在为医生提供以往类似病例方面发挥着至关重要的作用,这对于基于病例的推理和循证医学至关重要。由于计算和存储成本低,基于散列的搜索技术已被广泛用于建立图像检索系统。然而,这些方法往往存在不可忽略的量化损失,这会降低整体搜索性能。为了解决这个问题,本文提出了一种紧凑型编码解决方案,即深度贝叶斯量化(DBQ),它侧重于深度紧凑量化,可以估计连续的神经图像表征,并实现优于现有散列解决方案的性能。具体来说,DBQ 在一个新颖的贝叶斯学习框架内无缝结合了深度表征学习和表征紧凑量化,其中开发了一个基于代理嵌入的似然函数,以减轻传统相似性监督的采样问题。此外,还采用了高斯先验来减少量化损失。通过利用预先计算的查找表,所提出的 DBQ 可以实现高效和有效的相似性搜索。在三个基准神经图像数据集的 2,008 个结构性 MRI 扫描上进行的广泛实验表明,我们的方法优于之前的先进方法。
{"title":"Deep Bayesian Quantization for Supervised Neuroimage Search.","authors":"Erkun Yang, Cheng Deng, Mingxia Liu","doi":"10.1007/978-3-031-45676-3_40","DOIUrl":"10.1007/978-3-031-45676-3_40","url":null,"abstract":"<p><p>Neuroimage retrieval plays a crucial role in providing physicians with access to previous similar cases, which is essential for case-based reasoning and evidence-based medicine. Due to low computation and storage costs, hashing-based search techniques have been widely adopted for establishing image retrieval systems. However, these methods often suffer from nonnegligible quantization loss, which can degrade the overall search performance. To address this issue, this paper presents a compact coding solution namely <i>Deep Bayesian Quantization</i> (DBQ), which focuses on deep compact quantization that can estimate continuous neuroimage representations and achieve superior performance over existing hashing solutions. Specifically, DBQ seamlessly combines the deep representation learning and the representation compact quantization within a novel Bayesian learning framework, where a proxy embedding-based likelihood function is developed to alleviate the sampling issue for traditional similarity supervision. Additionally, a Gaussian prior is employed to reduce the quantization losses. By utilizing pre-computed lookup tables, the proposed DBQ can enable efficient and effective similarity search. Extensive experiments conducted on 2, 008 structural MRI scans from three benchmark neuroimage datasets demonstrate that our method outperforms previous state-of-the-arts.</p>","PeriodicalId":74092,"journal":{"name":"Machine learning in medical imaging. MLMI (Workshop)","volume":"14349 ","pages":"396-406"},"PeriodicalIF":0.0,"publicationDate":"2023-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10883338/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139934532","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IA-GCN: Interpretable Attention based Graph Convolutional Network for Disease Prediction. IACN:用于疾病预测的可解释的基于注意力的图卷积网络。
Pub Date : 2023-10-01 Epub Date: 2023-10-15 DOI: 10.1007/978-3-031-45673-2_38
Anees Kazi, Soroush Farghadani, Iman Aganj, Nassir Navab

Interpretability in Graph Convolutional Networks (GCNs) has been explored to some extent in general in computer vision; yet, in the medical domain, it requires further examination. Most of the interpretability approaches for GCNs, especially in the medical domain, focus on interpreting the output of the model in a post-hoc fashion. In this paper, we propose an interpretable attention module (IAM) that explains the relevance of the input features to the classification task on a GNN Model. The model uses these interpretations to improve its performance. In a clinical scenario, such a model can assist the clinical experts in better decision-making for diagnosis and treatment planning. The main novelty lies in the IAM, which directly operates on input features. IAM learns the attention for each feature based on the unique interpretability-specific losses. We show the application of our model on two publicly available datasets, Tadpole and the UK Biobank (UKBB). For Tadpole we choose the task of disease classification, and for UKBB, age, and sex prediction. The proposed model achieves an increase in an average accuracy of 3.2% for Tadpole and 1.6% for UKBB sex and 2% for the UKBB age prediction task compared to the state-of-the-art. Further, we show exhaustive validation and clinical interpretation of our results.

在计算机视觉中,图卷积网络(GCN)的可解释性已经得到了一定程度的探索;然而,在医学领域,它还需要进一步的检查。GCN的大多数可解释性方法,特别是在医学领域,都侧重于以事后的方式解释模型的输出。在本文中,我们提出了一个可解释注意力模块(IAM),用于解释GNN模型上输入特征与分类任务的相关性。该模型使用这些解释来提高其性能。在临床场景中,这样的模型可以帮助临床专家更好地进行诊断和治疗计划的决策。主要的新颖之处在于IAM,它直接对输入特性进行操作。IAM根据独特的可解释性特定损失来学习每个特征的注意力。我们在Tadpole和英国生物库(UKBB)这两个公开可用的数据集上展示了我们的模型的应用。对于蝌蚪,我们选择疾病分类的任务,对于UKBB,选择年龄和性别预测的任务。与现有技术相比,所提出的模型实现了蝌蚪3.2%、UKBB性别1.6%和UKBB年龄预测任务2%的平均准确率的提高。此外,我们对我们的结果进行了详尽的验证和临床解释。
{"title":"IA-GCN: Interpretable Attention based Graph Convolutional Network for Disease Prediction.","authors":"Anees Kazi,&nbsp;Soroush Farghadani,&nbsp;Iman Aganj,&nbsp;Nassir Navab","doi":"10.1007/978-3-031-45673-2_38","DOIUrl":"https://doi.org/10.1007/978-3-031-45673-2_38","url":null,"abstract":"<p><p>Interpretability in Graph Convolutional Networks (GCNs) has been explored to some extent in general in computer vision; yet, in the medical domain, it requires further examination. Most of the interpretability approaches for GCNs, especially in the medical domain, focus on interpreting the output of the model in a <i>post</i>-<i>hoc</i> fashion. In this paper, we propose an interpretable attention module (IAM) that explains the relevance of the input features to the classification task on a GNN Model. The model uses these interpretations to improve its performance. In a clinical scenario, such a model can assist the clinical experts in better decision-making for diagnosis and treatment planning. The main novelty lies in the IAM, which directly operates on input features. IAM learns the attention for each feature based on the unique interpretability-specific losses. We show the application of our model on two publicly available datasets, Tadpole and the UK Biobank (UKBB). For Tadpole we choose the task of disease classification, and for UKBB, age, and sex prediction. The proposed model achieves an increase in an average accuracy of 3.2% for Tadpole and 1.6% for UKBB sex and 2% for the UKBB age prediction task compared to the state-of-the-art. Further, we show exhaustive validation and clinical interpretation of our results.</p>","PeriodicalId":74092,"journal":{"name":"Machine learning in medical imaging. MLMI (Workshop)","volume":"14348 ","pages":"382-392"},"PeriodicalIF":0.0,"publicationDate":"2023-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10583839/pdf/nihms-1932968.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49685875","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Radiomics Boosts Deep Learning Model for IPMN Classification. 放射组学提升了用于 IPMN 分类的深度学习模型。
Pub Date : 2023-10-01 Epub Date: 2023-10-15 DOI: 10.1007/978-3-031-45676-3_14
Lanhong Yao, Zheyuan Zhang, Ugur Demir, Elif Keles, Camila Vendrami, Emil Agarunov, Candice Bolan, Ivo Schoots, Marc Bruno, Rajesh Keswani, Frank Miller, Tamas Gonda, Cemal Yazici, Temel Tirkes, Michael Wallace, Concetto Spampinato, Ulas Bagci

Intraductal Papillary Mucinous Neoplasm (IPMN) cysts are pre-malignant pancreas lesions, and they can progress into pancreatic cancer. Therefore, detecting and stratifying their risk level is of ultimate importance for effective treatment planning and disease control. However, this is a highly challenging task because of the diverse and irregular shape, texture, and size of the IPMN cysts as well as the pancreas. In this study, we propose a novel computer-aided diagnosis pipeline for IPMN risk classification from multi-contrast MRI scans. Our proposed analysis framework includes an efficient volumetric self-adapting segmentation strategy for pancreas delineation, followed by a newly designed deep learning-based classification scheme with a radiomics-based predictive approach. We test our proposed decision-fusion model in multi-center data sets of 246 multi-contrast MRI scans and obtain superior performance to the state of the art (SOTA) in this field. Our ablation studies demonstrate the significance of both radiomics and deep learning modules for achieving the new SOTA performance compared to international guidelines and published studies (81.9% vs 61.3% in accuracy). Our findings have important implications for clinical decision-making. In a series of rigorous experiments on multi-center data sets (246 MRI scans from five centers), we achieved unprecedented performance (81.9% accuracy). The code is available upon publication.

导管内乳头状黏液瘤(IPMN)囊肿是胰腺恶性肿瘤的前期病变,可发展为胰腺癌。因此,检测其风险水平并对其进行分层对于制定有效的治疗计划和控制疾病至关重要。然而,由于 IPMN 囊肿和胰腺的形状、质地和大小各不相同且不规则,因此这是一项极具挑战性的任务。在本研究中,我们提出了一种新型计算机辅助诊断管道,用于从多重对比 MRI 扫描中进行 IPMN 风险分类。我们提出的分析框架包括一种高效的胰腺体积自适应分割策略,以及一种新设计的基于深度学习的分类方案和一种基于放射组学的预测方法。我们在 246 个多对比度 MRI 扫描的多中心数据集中测试了我们提出的决策融合模型,结果表明其性能优于该领域的最新技术(SOTA)。我们的消融研究表明,与国际指南和已发表的研究相比,放射组学和深度学习模块对于实现新的 SOTA 性能具有重要意义(准确率为 81.9% 对 61.3%)。我们的研究结果对临床决策具有重要意义。在对多中心数据集(来自五个中心的 246 例核磁共振扫描)进行的一系列严格实验中,我们取得了前所未有的性能(准确率为 81.9%)。代码发布后即可获得。
{"title":"Radiomics Boosts Deep Learning Model for IPMN Classification.","authors":"Lanhong Yao, Zheyuan Zhang, Ugur Demir, Elif Keles, Camila Vendrami, Emil Agarunov, Candice Bolan, Ivo Schoots, Marc Bruno, Rajesh Keswani, Frank Miller, Tamas Gonda, Cemal Yazici, Temel Tirkes, Michael Wallace, Concetto Spampinato, Ulas Bagci","doi":"10.1007/978-3-031-45676-3_14","DOIUrl":"10.1007/978-3-031-45676-3_14","url":null,"abstract":"<p><p>Intraductal Papillary Mucinous Neoplasm (IPMN) cysts are pre-malignant pancreas lesions, and they can progress into pancreatic cancer. Therefore, detecting and stratifying their risk level is of ultimate importance for effective treatment planning and disease control. However, this is a highly challenging task because of the diverse and irregular shape, texture, and size of the IPMN cysts as well as the pancreas. In this study, we propose a novel computer-aided diagnosis pipeline for IPMN risk classification from multi-contrast MRI scans. Our proposed analysis framework includes an efficient volumetric self-adapting segmentation strategy for pancreas delineation, followed by a newly designed deep learning-based classification scheme with a radiomics-based predictive approach. We test our proposed decision-fusion model in multi-center data sets of 246 multi-contrast MRI scans and obtain superior performance to the state of the art (SOTA) in this field. Our ablation studies demonstrate the significance of both radiomics and deep learning modules for achieving the new SOTA performance compared to international guidelines and published studies (81.9% vs 61.3% in accuracy). Our findings have important implications for clinical decision-making. In a series of rigorous experiments on multi-center data sets (246 MRI scans from five centers), we achieved unprecedented performance (81.9% accuracy). The code is available upon publication.</p>","PeriodicalId":74092,"journal":{"name":"Machine learning in medical imaging. MLMI (Workshop)","volume":"14349 ","pages":"134-143"},"PeriodicalIF":0.0,"publicationDate":"2023-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10810260/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139563708","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Structural MRI Harmonization via Disentangled Latent Energy-Based Style Translation. 通过基于潜能的风格翻译进行结构磁共振成像协调。
Pub Date : 2023-10-01 Epub Date: 2023-10-15 DOI: 10.1007/978-3-031-45673-2_1
Mengqi Wu, Lintao Zhang, Pew-Thian Yap, Weili Lin, Hongtu Zhu, Mingxia Liu

Multi-site brain magnetic resonance imaging (MRI) has been widely used in clinical and research domains, but usually is sensitive to non-biological variations caused by site effects (e.g., field strengths and scanning protocols). Several retrospective data harmonization methods have shown promising results in removing these non-biological variations at feature or whole-image level. Most existing image-level harmonization methods are implemented through generative adversarial networks, which are generally computationally expensive and generalize poorly on independent data. To this end, this paper proposes a disentangled latent energy-based style translation (DLEST) framework for image-level structural MRI harmonization. Specifically, DLEST disentangles site-invariant image generation and site-specific style translation via a latent autoencoder and an energy-based model. The autoencoder learns to encode images into low-dimensional latent space, and generates faithful images from latent codes. The energy-based model is placed in between the encoding and generation steps, facilitating style translation from a source domain to a target domain implicitly. This allows highly generalizable image generation and efficient style translation through the latent space. We train our model on 4,092 T1-weighted MRIs in 3 tasks: histogram comparison, acquisition site classification, and brain tissue segmentation. Qualitative and quantitative results demonstrate the superiority of our approach, which generally outperforms several state-of-the-art methods.

多部位脑磁共振成像(MRI)已广泛应用于临床和研究领域,但通常对部位效应(如场强和扫描协议)引起的非生物变异很敏感。有几种回顾性数据协调方法在特征或整个图像层面消除这些非生物变异方面取得了可喜的成果。现有的大多数图像级协调方法都是通过生成式对抗网络实现的,这种网络通常计算成本高,对独立数据的泛化能力差。为此,本文提出了一种基于潜能的风格转换(DLEST)框架,用于图像级结构磁共振成像协调。具体来说,DLEST 通过一个潜在自动编码器和一个基于能量的模型,将部位不变的图像生成和特定部位的风格转换分离开来。自动编码器学习将图像编码到低维潜在空间,并从潜在代码生成忠实图像。基于能量的模型被置于编码和生成步骤之间,促进了从源域到目标域的隐式风格转换。这使得图像生成具有高度的通用性,并能通过潜空间进行高效的风格转换。我们在 4,092 张 T1 加权核磁共振图像上对模型进行了 3 项任务的训练:直方图比较、采集部位分类和脑组织分割。定性和定量结果都证明了我们方法的优越性,总体上优于几种最先进的方法。
{"title":"Structural MRI Harmonization via Disentangled Latent Energy-Based Style Translation.","authors":"Mengqi Wu, Lintao Zhang, Pew-Thian Yap, Weili Lin, Hongtu Zhu, Mingxia Liu","doi":"10.1007/978-3-031-45673-2_1","DOIUrl":"10.1007/978-3-031-45673-2_1","url":null,"abstract":"<p><p>Multi-site brain magnetic resonance imaging (MRI) has been widely used in clinical and research domains, but usually is sensitive to non-biological variations caused by site effects (<i>e.g.</i>, field strengths and scanning protocols). Several retrospective data harmonization methods have shown promising results in removing these non-biological variations at feature or whole-image level. Most existing image-level harmonization methods are implemented through generative adversarial networks, which are generally computationally expensive and generalize poorly on independent data. To this end, this paper proposes a disentangled latent energy-based style translation (DLEST) framework for image-level structural MRI harmonization. Specifically, DLEST disentangles <i>site-invariant image generation</i> and <i>site-specific style translation</i> via a latent autoencoder and an energy-based model. The autoencoder learns to encode images into low-dimensional latent space, and generates faithful images from latent codes. The energy-based model is placed in between the encoding and generation steps, facilitating style translation from a source domain to a target domain implicitly. This allows <i>highly generalizable image generation and efficient style translation</i> through the latent space. We train our model on 4,092 T1-weighted MRIs in 3 tasks: histogram comparison, acquisition site classification, and brain tissue segmentation. Qualitative and quantitative results demonstrate the superiority of our approach, which generally outperforms several state-of-the-art methods.</p>","PeriodicalId":74092,"journal":{"name":"Machine learning in medical imaging. MLMI (Workshop)","volume":"14348 ","pages":"1-11"},"PeriodicalIF":0.0,"publicationDate":"2023-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10883146/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139934531","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Understanding Clinical Progression of Late-Life Depression to Alzheimer's Disease Over 5 Years with Structural MRI. 通过结构MRI了解老年抑郁症到阿尔茨海默病的临床进展。
Pub Date : 2022-09-01 DOI: 10.1007/978-3-031-21014-3_27
Lintao Zhang, Minhui Yu, Lihong Wang, David C Steffens, Rong Wu, Guy G Potter, Mingxia Liu

Previous studies have shown that late-life depression (LLD) may be a precursor of neurodegenerative diseases and may increase the risk of dementia. At present, the pathological relationship between LLD and dementia, in particularly Alzheimer's disease (AD) is unclear. Structural MRI (sMRI) can provide objective biomarkers for the computer-aided diagnosis of LLD and AD, providing a promising solution to understand the clinical progression of brain disorders. But few studies have focused on sMRI-based predictive analysis of clinical progression from LLD to AD. In this paper, we develop a deep learning method to predict the clinical progression of LLD to AD up to 5 years after baseline time using T1-weighted structural MRIs. We also analyze several important factors that limit the diagnostic performance of learning-based methods, including data imbalance, small-sample-size, and multi-site data heterogeneity, by leveraging a relatively large-scale database to aid model training. Experimental results on 308 subjects with sMRIs acquired from 2 imaging sites and the publicly available ADNI database demonstrate the potential of deep learning in predicting the clinical progression of LLD to AD. To the best of our knowledge, this is among the first attempts to explore the complex pathophysiological relationship between LLD and AD based on structural MRI using a deep learning method.

先前的研究表明,晚年抑郁(LLD)可能是神经退行性疾病的前兆,并可能增加患痴呆的风险。目前,LLD与痴呆,特别是阿尔茨海默病(AD)的病理关系尚不清楚。结构磁共振成像(sMRI)可以为LLD和AD的计算机辅助诊断提供客观的生物标志物,为了解脑部疾病的临床进展提供了有希望的解决方案。但很少有研究关注基于smri的从LLD到AD临床进展的预测分析。在本文中,我们开发了一种深度学习方法,使用t1加权结构mri预测LLD在基线时间后长达5年的临床进展。我们还分析了限制基于学习的方法诊断性能的几个重要因素,包括数据不平衡、小样本量和多站点数据异质性,通过利用相对大规模的数据库来辅助模型训练。来自2个影像站点和公开可用的ADNI数据库的308名sMRIs受试者的实验结果表明,深度学习在预测LLD到AD的临床进展方面具有潜力。据我们所知,这是首次尝试使用深度学习方法基于结构MRI探索LLD和AD之间复杂的病理生理关系。
{"title":"Understanding Clinical Progression of Late-Life Depression to Alzheimer's Disease Over 5 Years with Structural MRI.","authors":"Lintao Zhang,&nbsp;Minhui Yu,&nbsp;Lihong Wang,&nbsp;David C Steffens,&nbsp;Rong Wu,&nbsp;Guy G Potter,&nbsp;Mingxia Liu","doi":"10.1007/978-3-031-21014-3_27","DOIUrl":"https://doi.org/10.1007/978-3-031-21014-3_27","url":null,"abstract":"<p><p>Previous studies have shown that late-life depression (LLD) may be a precursor of neurodegenerative diseases and may increase the risk of dementia. At present, the pathological relationship between LLD and dementia, in particularly Alzheimer's disease (AD) is unclear. Structural MRI (sMRI) can provide objective biomarkers for the computer-aided diagnosis of LLD and AD, providing a promising solution to understand the clinical progression of brain disorders. But few studies have focused on sMRI-based predictive analysis of clinical progression from LLD to AD. In this paper, we develop a deep learning method to predict the clinical progression of LLD to AD up to 5 years after baseline time using T1-weighted structural MRIs. We also analyze several important factors that limit the diagnostic performance of learning-based methods, including data imbalance, small-sample-size, and multi-site data heterogeneity, by leveraging a relatively large-scale database to aid model training. Experimental results on 308 subjects with sMRIs acquired from 2 imaging sites and the publicly available ADNI database demonstrate the potential of deep learning in predicting the clinical progression of LLD to AD. To the best of our knowledge, this is among the first attempts to explore the complex pathophysiological relationship between LLD and AD based on structural MRI using a deep learning method.</p>","PeriodicalId":74092,"journal":{"name":"Machine learning in medical imaging. MLMI (Workshop)","volume":"13583 ","pages":"259-268"},"PeriodicalIF":0.0,"publicationDate":"2022-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9805302/pdf/nihms-1859375.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9838060","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Multi-scale Multi-structure Siamese Network (MMSNet) for Primary Open-Angle Glaucoma Prediction. 用于原发性开角型青光眼预测的多尺度多结构连体网络 (MMSNet)。
Pub Date : 2022-09-01 Epub Date: 2022-12-16 DOI: 10.1007/978-3-031-21014-3_45
Mingquan Lin, Lei Liu, Mae Gorden, Michael Kass, Sarah Van Tassel, Fei Wang, Yifan Peng

Primary open-angle glaucoma (POAG) is one of the leading causes of irreversible blindness in the United States and worldwide. POAG prediction before onset plays an important role in early treatment. Although deep learning methods have been proposed to predict POAG, these methods mainly focus on current status prediction. In addition, all these methods used a single image as input. On the other hand, glaucoma specialists determine a glaucomatous eye by comparing the follow-up optic nerve image with the baseline along with supplementary clinical data. To simulate this process, we proposed a Multi-scale Multi-structure Siamese Network (MMSNet) to predict future POAG event from fundus photographs. The MMSNet consists of two side-outputs for deep supervision and 2D blocks to utilize two-dimensional features to assist classification. The MMSNet network was trained and evaluated on a large dataset: 37,339 fundus photographs from 1,636 Ocular Hypertension Treatment Study (OHTS) participants. Extensive experiments show that MMSNet outperforms the state-of-the-art on two "POAG prediction before onset" tasks. Our AUC are 0.9312 and 0.9507, which are 0.2204 and 0.1490 higher than the state-of-the-art, respectively. In addition, an ablation study is performed to check the contribution of different components. These results highlight the potential of deep learning to assist and enhance the prediction of future POAG event. The proposed network will be publicly available on https://github.com/bionlplab/MMSNet.

原发性开角型青光眼(POAG)是美国乃至全世界造成不可逆转性失明的主要原因之一。在发病前预测 POAG 对早期治疗起着重要作用。虽然已有人提出了预测 POAG 的深度学习方法,但这些方法主要侧重于现状预测。此外,所有这些方法都使用单一图像作为输入。另一方面,青光眼专家通过比较随访视神经图像和基线图像以及补充临床数据来确定是否患有青光眼。为了模拟这一过程,我们提出了一种多尺度多结构连体网络(MMSNet),用于根据眼底照片预测未来的 POAG 事件。MMSNet 由两个用于深度监督的侧输出和二维块组成,利用二维特征来辅助分类。MMSNet 网络在一个大型数据集上进行了训练和评估:37,339 张眼底照片,这些照片来自 1,636 名眼压过高治疗研究(OHTS)参与者。广泛的实验表明,MMSNet 在两项 "发病前预测 POAG "任务中的表现优于最先进的网络。我们的 AUC 分别为 0.9312 和 0.9507,分别比先进水平高出 0.2204 和 0.1490。此外,我们还进行了一项消融研究,以检查不同成分的贡献。这些结果凸显了深度学习在辅助和增强未来 POAG 事件预测方面的潜力。拟议的网络将在 https://github.com/bionlplab/MMSNet 上公开发布。
{"title":"Multi-scale Multi-structure Siamese Network (MMSNet) for Primary Open-Angle Glaucoma Prediction.","authors":"Mingquan Lin, Lei Liu, Mae Gorden, Michael Kass, Sarah Van Tassel, Fei Wang, Yifan Peng","doi":"10.1007/978-3-031-21014-3_45","DOIUrl":"10.1007/978-3-031-21014-3_45","url":null,"abstract":"<p><p>Primary open-angle glaucoma (POAG) is one of the leading causes of irreversible blindness in the United States and worldwide. POAG prediction before onset plays an important role in early treatment. Although deep learning methods have been proposed to predict POAG, these methods mainly focus on current status prediction. In addition, all these methods used a single image as input. On the other hand, glaucoma specialists determine a glaucomatous eye by comparing the follow-up optic nerve image with the baseline along with supplementary clinical data. To simulate this process, we proposed a Multi-scale Multi-structure Siamese Network (MMSNet) to predict future POAG event from fundus photographs. The MMSNet consists of two side-outputs for deep supervision and 2D blocks to utilize two-dimensional features to assist classification. The MMSNet network was trained and evaluated on a large dataset: 37,339 fundus photographs from 1,636 Ocular Hypertension Treatment Study (OHTS) participants. Extensive experiments show that MMSNet outperforms the state-of-the-art on two \"POAG prediction before onset\" tasks. Our AUC are 0.9312 and 0.9507, which are 0.2204 and 0.1490 higher than the state-of-the-art, respectively. In addition, an ablation study is performed to check the contribution of different components. These results highlight the potential of deep learning to assist and enhance the prediction of future POAG event. The proposed network will be publicly available on https://github.com/bionlplab/MMSNet.</p>","PeriodicalId":74092,"journal":{"name":"Machine learning in medical imaging. MLMI (Workshop)","volume":"13583 ","pages":"436-445"},"PeriodicalIF":0.0,"publicationDate":"2022-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9844668/pdf/nihms-1864372.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10604661","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Machine learning in medical imaging. MLMI (Workshop)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1