首页 > 最新文献

Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention最新文献

英文 中文
LOTUS: Learning to Optimize Task-based US representations LOTUS:学习优化基于任务的US表示
Yordanka Velikova, Mohammad Farid Azampour, Walter Simson, Vanessa Gonzalez Duque, N. Navab
Anatomical segmentation of organs in ultrasound images is essential to many clinical applications, particularly for diagnosis and monitoring. Existing deep neural networks require a large amount of labeled data for training in order to achieve clinically acceptable performance. Yet, in ultrasound, due to characteristic properties such as speckle and clutter, it is challenging to obtain accurate segmentation boundaries, and precise pixel-wise labeling of images is highly dependent on the expertise of physicians. In contrast, CT scans have higher resolution and improved contrast, easing organ identification. In this paper, we propose a novel approach for learning to optimize task-based ultra-sound image representations. Given annotated CT segmentation maps as a simulation medium, we model acoustic propagation through tissue via ray-casting to generate ultrasound training data. Our ultrasound simulator is fully differentiable and learns to optimize the parameters for generating physics-based ultrasound images guided by the downstream segmentation task. In addition, we train an image adaptation network between real and simulated images to achieve simultaneous image synthesis and automatic segmentation on US images in an end-to-end training setting. The proposed method is evaluated on aorta and vessel segmentation tasks and shows promising quantitative results. Furthermore, we also conduct qualitative results of optimized image representations on other organs.
超声图像中器官的解剖分割在许多临床应用中是必不可少的,特别是在诊断和监测方面。为了达到临床可接受的性能,现有的深度神经网络需要大量的标记数据进行训练。然而,在超声中,由于散斑和杂波等特征特性,很难获得准确的分割边界,并且图像的精确像素标记高度依赖于医生的专业知识。相比之下,CT扫描具有更高的分辨率和更好的对比度,易于器官识别。在本文中,我们提出了一种新的学习方法来优化基于任务的超声波图像表示。给定带注释的CT分割图作为模拟介质,我们通过射线投射来模拟声波在组织中的传播,以生成超声训练数据。我们的超声模拟器是完全可微分的,并学习优化参数,以生成基于物理的超声图像,由下游分割任务引导。此外,我们在真实图像和模拟图像之间训练了一个图像自适应网络,在端到端训练环境中实现了对美国图像的同步图像合成和自动分割。该方法对主动脉和血管分割任务进行了评估,并显示出有希望的定量结果。此外,我们还在其他器官上进行了优化图像表示的定性结果。
{"title":"LOTUS: Learning to Optimize Task-based US representations","authors":"Yordanka Velikova, Mohammad Farid Azampour, Walter Simson, Vanessa Gonzalez Duque, N. Navab","doi":"10.48550/arXiv.2307.16021","DOIUrl":"https://doi.org/10.48550/arXiv.2307.16021","url":null,"abstract":"Anatomical segmentation of organs in ultrasound images is essential to many clinical applications, particularly for diagnosis and monitoring. Existing deep neural networks require a large amount of labeled data for training in order to achieve clinically acceptable performance. Yet, in ultrasound, due to characteristic properties such as speckle and clutter, it is challenging to obtain accurate segmentation boundaries, and precise pixel-wise labeling of images is highly dependent on the expertise of physicians. In contrast, CT scans have higher resolution and improved contrast, easing organ identification. In this paper, we propose a novel approach for learning to optimize task-based ultra-sound image representations. Given annotated CT segmentation maps as a simulation medium, we model acoustic propagation through tissue via ray-casting to generate ultrasound training data. Our ultrasound simulator is fully differentiable and learns to optimize the parameters for generating physics-based ultrasound images guided by the downstream segmentation task. In addition, we train an image adaptation network between real and simulated images to achieve simultaneous image synthesis and automatic segmentation on US images in an end-to-end training setting. The proposed method is evaluated on aorta and vessel segmentation tasks and shows promising quantitative results. Furthermore, we also conduct qualitative results of optimized image representations on other organs.","PeriodicalId":18289,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"54 1","pages":"435-445"},"PeriodicalIF":0.0,"publicationDate":"2023-07-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77935210","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Robust vertebra identification using simultaneous node and edge predicting Graph Neural Networks 基于节点和边缘同时预测的图神经网络的鲁棒椎体识别
Vincent Bürgin, R. Prevost, Marijn F. Stollenga
Automatic vertebra localization and identification in CT scans is important for numerous clinical applications. Much progress has been made on this topic, but it mostly targets positional localization of vertebrae, ignoring their orientation. Additionally, most methods employ heuristics in their pipeline that can be sensitive in real clinical images which tend to contain abnormalities. We introduce a simple pipeline that employs a standard prediction with a U-Net, followed by a single graph neural network to associate and classify vertebrae with full orientation. To test our method, we introduce a new vertebra dataset that also contains pedicle detections that are associated with vertebra bodies, creating a more challenging landmark prediction, association and classification task. Our method is able to accurately associate the correct body and pedicle landmarks, ignore false positives and classify vertebrae in a simple, fully trainable pipeline avoiding application-specific heuristics. We show our method outperforms traditional approaches such as Hungarian Matching and Hidden Markov Models. We also show competitive performance on the standard VerSe challenge body identification task.
在CT扫描中自动定位和识别椎体在许多临床应用中是重要的。在这方面已经取得了很大的进展,但它主要针对椎骨的位置定位,而忽略了它们的方向。此外,大多数方法在他们的管道中使用启发式,这在真实的临床图像中是敏感的,往往包含异常。我们引入了一个简单的管道,该管道使用U-Net进行标准预测,然后使用单图神经网络进行全方向椎骨的关联和分类。为了测试我们的方法,我们引入了一个新的椎体数据集,该数据集还包含与椎体相关的椎弓根检测,从而创建了更具挑战性的地标预测、关联和分类任务。我们的方法能够准确地将正确的身体和椎弓根标志联系起来,忽略假阳性,并以简单,完全可训练的管道对椎骨进行分类,避免了特定应用的启发式。我们证明我们的方法优于传统的方法,如匈牙利匹配和隐马尔可夫模型。我们还在标准的VerSe挑战身体识别任务中显示了竞争表现。
{"title":"Robust vertebra identification using simultaneous node and edge predicting Graph Neural Networks","authors":"Vincent Bürgin, R. Prevost, Marijn F. Stollenga","doi":"10.48550/arXiv.2308.02509","DOIUrl":"https://doi.org/10.48550/arXiv.2308.02509","url":null,"abstract":"Automatic vertebra localization and identification in CT scans is important for numerous clinical applications. Much progress has been made on this topic, but it mostly targets positional localization of vertebrae, ignoring their orientation. Additionally, most methods employ heuristics in their pipeline that can be sensitive in real clinical images which tend to contain abnormalities. We introduce a simple pipeline that employs a standard prediction with a U-Net, followed by a single graph neural network to associate and classify vertebrae with full orientation. To test our method, we introduce a new vertebra dataset that also contains pedicle detections that are associated with vertebra bodies, creating a more challenging landmark prediction, association and classification task. Our method is able to accurately associate the correct body and pedicle landmarks, ignore false positives and classify vertebrae in a simple, fully trainable pipeline avoiding application-specific heuristics. We show our method outperforms traditional approaches such as Hungarian Matching and Hidden Markov Models. We also show competitive performance on the standard VerSe challenge body identification task.","PeriodicalId":18289,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"43 1","pages":"483-493"},"PeriodicalIF":0.0,"publicationDate":"2023-07-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80245364","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
vox2vec: A Framework for Self-supervised Contrastive Learning of Voxel-level Representations in Medical Images vox2vec:医学图像体素级表示的自监督对比学习框架
M. Goncharov, Vera Soboleva, Anvar Kurmukov, M. Pisov, M. Belyaev
This paper introduces vox2vec - a contrastive method for self-supervised learning (SSL) of voxel-level representations. vox2vec representations are modeled by a Feature Pyramid Network (FPN): a voxel representation is a concatenation of the corresponding feature vectors from different pyramid levels. The FPN is pre-trained to produce similar representations for the same voxel in different augmented contexts and distinctive representations for different voxels. This results in unified multi-scale representations that capture both global semantics (e.g., body part) and local semantics (e.g., different small organs or healthy versus tumor tissue). We use vox2vec to pre-train a FPN on more than 6500 publicly available computed tomography images. We evaluate the pre-trained representations by attaching simple heads on top of them and training the resulting models for 22 segmentation tasks. We show that vox2vec outperforms existing medical imaging SSL techniques in three evaluation setups: linear and non-linear probing and end-to-end fine-tuning. Moreover, a non-linear head trained on top of the frozen vox2vec representations achieves competitive performance with the FPN trained from scratch while having 50 times fewer trainable parameters. The code is available at https://github.com/mishgon/vox2vec .
本文介绍了vox2vec——一种体素级表示的自监督学习(SSL)对比方法。vox2vec表示由特征金字塔网络(FPN)建模:体素表示是来自不同金字塔级别的相应特征向量的连接。FPN经过预训练,可以在不同的增强环境中为相同的体素产生相似的表示,并为不同的体素产生不同的表示。这导致了统一的多尺度表示,可以捕获全局语义(例如,身体部位)和局部语义(例如,不同的小器官或健康与肿瘤组织)。我们使用vox2vec在6500多张公开可用的计算机断层扫描图像上预训练FPN。我们通过在预训练的表示上附加简单的头部来评估它们,并为22个分割任务训练生成的模型。我们表明,vox2vec在三个评估设置中优于现有的医学成像SSL技术:线性和非线性探测以及端到端微调。此外,在冻结的vox2vec表示之上训练的非线性头部与从头开始训练的FPN相比具有竞争性性能,同时可训练参数减少了50倍。代码可在https://github.com/mishgon/vox2vec上获得。
{"title":"vox2vec: A Framework for Self-supervised Contrastive Learning of Voxel-level Representations in Medical Images","authors":"M. Goncharov, Vera Soboleva, Anvar Kurmukov, M. Pisov, M. Belyaev","doi":"10.48550/arXiv.2307.14725","DOIUrl":"https://doi.org/10.48550/arXiv.2307.14725","url":null,"abstract":"This paper introduces vox2vec - a contrastive method for self-supervised learning (SSL) of voxel-level representations. vox2vec representations are modeled by a Feature Pyramid Network (FPN): a voxel representation is a concatenation of the corresponding feature vectors from different pyramid levels. The FPN is pre-trained to produce similar representations for the same voxel in different augmented contexts and distinctive representations for different voxels. This results in unified multi-scale representations that capture both global semantics (e.g., body part) and local semantics (e.g., different small organs or healthy versus tumor tissue). We use vox2vec to pre-train a FPN on more than 6500 publicly available computed tomography images. We evaluate the pre-trained representations by attaching simple heads on top of them and training the resulting models for 22 segmentation tasks. We show that vox2vec outperforms existing medical imaging SSL techniques in three evaluation setups: linear and non-linear probing and end-to-end fine-tuning. Moreover, a non-linear head trained on top of the frozen vox2vec representations achieves competitive performance with the FPN trained from scratch while having 50 times fewer trainable parameters. The code is available at https://github.com/mishgon/vox2vec .","PeriodicalId":18289,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"87 1","pages":"605-614"},"PeriodicalIF":0.0,"publicationDate":"2023-07-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86843413","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Centroid-aware feature recalibration for cancer grading in pathology images 质心感知特征重新校准用于病理图像的癌症分级
Jaeung Lee, Keunho Byeon, J. T. Kwak
Cancer grading is an essential task in pathology. The recent developments of artificial neural networks in computational pathology have shown that these methods hold great potential for improving the accuracy and quality of cancer diagnosis. However, the issues with the robustness and reliability of such methods have not been fully resolved yet. Herein, we propose a centroid-aware feature recalibration network that can conduct cancer grading in an accurate and robust manner. The proposed network maps an input pathology image into an embedding space and adjusts it by using centroids embedding vectors of different cancer grades via attention mechanism. Equipped with the recalibrated embedding vector, the proposed network classifiers the input pathology image into a pertinent class label, i.e., cancer grade. We evaluate the proposed network using colorectal cancer datasets that were collected under different environments. The experimental results confirm that the proposed network is able to conduct cancer grading in pathology images with high accuracy regardless of the environmental changes in the datasets.
肿瘤分级是病理学中的一项重要任务。人工神经网络在计算病理学中的最新发展表明,这些方法在提高癌症诊断的准确性和质量方面具有巨大的潜力。然而,这些方法的鲁棒性和可靠性问题尚未得到充分解决。在此,我们提出了一个质心感知特征再校准网络,可以准确和稳健地进行癌症分级。该网络将输入的病理图像映射到嵌入空间中,并通过注意机制使用不同癌症等级的质心嵌入向量对其进行调整。利用重新校准的嵌入向量,该网络将输入的病理图像分类为相关的类别标签,即癌症等级。我们使用在不同环境下收集的结直肠癌数据集来评估所提出的网络。实验结果证实,无论数据集的环境变化如何,所提出的网络都能够以较高的准确率对病理图像进行癌症分级。
{"title":"Centroid-aware feature recalibration for cancer grading in pathology images","authors":"Jaeung Lee, Keunho Byeon, J. T. Kwak","doi":"10.48550/arXiv.2307.13947","DOIUrl":"https://doi.org/10.48550/arXiv.2307.13947","url":null,"abstract":"Cancer grading is an essential task in pathology. The recent developments of artificial neural networks in computational pathology have shown that these methods hold great potential for improving the accuracy and quality of cancer diagnosis. However, the issues with the robustness and reliability of such methods have not been fully resolved yet. Herein, we propose a centroid-aware feature recalibration network that can conduct cancer grading in an accurate and robust manner. The proposed network maps an input pathology image into an embedding space and adjusts it by using centroids embedding vectors of different cancer grades via attention mechanism. Equipped with the recalibrated embedding vector, the proposed network classifiers the input pathology image into a pertinent class label, i.e., cancer grade. We evaluate the proposed network using colorectal cancer datasets that were collected under different environments. The experimental results confirm that the proposed network is able to conduct cancer grading in pathology images with high accuracy regardless of the environmental changes in the datasets.","PeriodicalId":18289,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"11 1","pages":"212-221"},"PeriodicalIF":0.0,"publicationDate":"2023-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90959424","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards multi-modal anatomical landmark detection for ultrasound-guided brain tumor resection with contrastive learning 超声引导下对比学习脑肿瘤切除术的多模态解剖地标检测
Soorena Salari, Amir Rasoulian, H. Rivaz, Yiming Xiao
Homologous anatomical landmarks between medical scans are instrumental in quantitative assessment of image registration quality in various clinical applications, such as MRI-ultrasound registration for tissue shift correction in ultrasound-guided brain tumor resection. While manually identified landmark pairs between MRI and ultrasound (US) have greatly facilitated the validation of different registration algorithms for the task, the procedure requires significant expertise, labor, and time, and can be prone to inter- and intra-rater inconsistency. So far, many traditional and machine learning approaches have been presented for anatomical landmark detection, but they primarily focus on mono-modal applications. Unfortunately, despite the clinical needs, inter-modal/contrast landmark detection has very rarely been attempted. Therefore, we propose a novel contrastive learning framework to detect corresponding landmarks between MRI and intra-operative US scans in neurosurgery. Specifically, two convolutional neural networks were trained jointly to encode image features in MRI and US scans to help match the US image patch that contain the corresponding landmarks in the MRI. We developed and validated the technique using the public RESECT database. With a mean landmark detection accuracy of 5.88+-4.79 mm against 18.78+-4.77 mm with SIFT features, the proposed method offers promising results for MRI-US landmark detection in neurosurgical applications for the first time.
在各种临床应用中,医学扫描之间的同源解剖标志有助于定量评估图像配准质量,例如超声引导下脑肿瘤切除术中组织移位校正的mri超声配准。虽然手动识别MRI和超声(US)之间的地标对极大地促进了不同配准算法的验证,但该过程需要大量的专业知识、人力和时间,并且容易出现内部和内部的不一致。到目前为止,已经提出了许多用于解剖地标检测的传统方法和机器学习方法,但它们主要集中在单模态应用上。不幸的是,尽管有临床需要,但很少有人尝试检测多模态/造影剂地标。因此,我们提出了一种新的对比学习框架来检测神经外科MRI和术中US扫描之间的相应标志。具体来说,两个卷积神经网络被联合训练来编码MRI和US扫描中的图像特征,以帮助匹配包含MRI中相应地标的US图像补丁。我们使用公共RESECT数据库开发并验证了该技术。该方法的平均地标检测精度为5.88+-4.79 mm,而SIFT特征为18.78+-4.77 mm,首次为神经外科应用的MRI-US地标检测提供了有希望的结果。
{"title":"Towards multi-modal anatomical landmark detection for ultrasound-guided brain tumor resection with contrastive learning","authors":"Soorena Salari, Amir Rasoulian, H. Rivaz, Yiming Xiao","doi":"10.48550/arXiv.2307.14523","DOIUrl":"https://doi.org/10.48550/arXiv.2307.14523","url":null,"abstract":"Homologous anatomical landmarks between medical scans are instrumental in quantitative assessment of image registration quality in various clinical applications, such as MRI-ultrasound registration for tissue shift correction in ultrasound-guided brain tumor resection. While manually identified landmark pairs between MRI and ultrasound (US) have greatly facilitated the validation of different registration algorithms for the task, the procedure requires significant expertise, labor, and time, and can be prone to inter- and intra-rater inconsistency. So far, many traditional and machine learning approaches have been presented for anatomical landmark detection, but they primarily focus on mono-modal applications. Unfortunately, despite the clinical needs, inter-modal/contrast landmark detection has very rarely been attempted. Therefore, we propose a novel contrastive learning framework to detect corresponding landmarks between MRI and intra-operative US scans in neurosurgery. Specifically, two convolutional neural networks were trained jointly to encode image features in MRI and US scans to help match the US image patch that contain the corresponding landmarks in the MRI. We developed and validated the technique using the public RESECT database. With a mean landmark detection accuracy of 5.88+-4.79 mm against 18.78+-4.77 mm with SIFT features, the proposed method offers promising results for MRI-US landmark detection in neurosurgical applications for the first time.","PeriodicalId":18289,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"24 1","pages":"668-678"},"PeriodicalIF":0.0,"publicationDate":"2023-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81069447","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ProtoASNet: Dynamic Prototypes for Inherently Interpretable and Uncertainty-Aware Aortic Stenosis Classification in Echocardiography 超声心动图中固有可解释和不确定性感知主动脉狭窄分类的动态原型
H. Vaseli, A. Gu, S. Neda, Ahmadi Amiri, M. Tsang, A. Fung, Nima Kondori, Armin Saadat, P. Abolmaesumi, T. Tsang
Aortic stenosis (AS) is a common heart valve disease that requires accurate and timely diagnosis for appropriate treatment. Most current automatic AS severity detection methods rely on black-box models with a low level of trustworthiness, which hinders clinical adoption. To address this issue, we propose ProtoASNet, a prototypical network that directly detects AS from B-mode echocardiography videos, while making interpretable predictions based on the similarity between the input and learned spatio-temporal prototypes. This approach provides supporting evidence that is clinically relevant, as the prototypes typically highlight markers such as calcification and restricted movement of aortic valve leaflets. Moreover, ProtoASNet utilizes abstention loss to estimate aleatoric uncertainty by defining a set of prototypes that capture ambiguity and insufficient information in the observed data. This provides a reliable system that can detect and explain when it may fail. We evaluate ProtoASNet on a private dataset and the publicly available TMED-2 dataset, where it outperforms existing state-of-the-art methods with an accuracy of 80.0% and 79.7%, respectively. Furthermore, ProtoASNet provides interpretability and an uncertainty measure for each prediction, which can improve transparency and facilitate the interactive usage of deep networks to aid clinical decision-making. Our source code is available at: https://github.com/hooman007/ProtoASNet.
主动脉瓣狭窄(AS)是一种常见的心脏瓣膜疾病,需要准确及时的诊断和适当的治疗。目前大多数AS严重程度自动检测方法依赖于黑盒模型,可信度较低,阻碍了临床应用。为了解决这个问题,我们提出了ProtoASNet,这是一个原型网络,可以直接从b模式超声心动图视频中检测AS,同时根据输入和学习的时空原型之间的相似性做出可解释的预测。这种方法提供了临床相关的支持性证据,因为原型通常突出标记,如钙化和主动脉瓣小叶运动受限。此外,ProtoASNet通过定义一组捕获观察数据中的模糊和信息不足的原型,利用弃权损失来估计任意不确定性。这提供了一个可靠的系统,可以检测和解释何时可能会失败。我们在私有数据集和公开可用的TMED-2数据集上对ProtoASNet进行了评估,其中它的准确率分别为80.0%和79.7%,优于现有的最先进的方法。此外,ProtoASNet为每个预测提供了可解释性和不确定性度量,这可以提高透明度并促进深度网络的交互式使用,以帮助临床决策。我们的源代码可从https://github.com/hooman007/ProtoASNet获得。
{"title":"ProtoASNet: Dynamic Prototypes for Inherently Interpretable and Uncertainty-Aware Aortic Stenosis Classification in Echocardiography","authors":"H. Vaseli, A. Gu, S. Neda, Ahmadi Amiri, M. Tsang, A. Fung, Nima Kondori, Armin Saadat, P. Abolmaesumi, T. Tsang","doi":"10.48550/arXiv.2307.14433","DOIUrl":"https://doi.org/10.48550/arXiv.2307.14433","url":null,"abstract":"Aortic stenosis (AS) is a common heart valve disease that requires accurate and timely diagnosis for appropriate treatment. Most current automatic AS severity detection methods rely on black-box models with a low level of trustworthiness, which hinders clinical adoption. To address this issue, we propose ProtoASNet, a prototypical network that directly detects AS from B-mode echocardiography videos, while making interpretable predictions based on the similarity between the input and learned spatio-temporal prototypes. This approach provides supporting evidence that is clinically relevant, as the prototypes typically highlight markers such as calcification and restricted movement of aortic valve leaflets. Moreover, ProtoASNet utilizes abstention loss to estimate aleatoric uncertainty by defining a set of prototypes that capture ambiguity and insufficient information in the observed data. This provides a reliable system that can detect and explain when it may fail. We evaluate ProtoASNet on a private dataset and the publicly available TMED-2 dataset, where it outperforms existing state-of-the-art methods with an accuracy of 80.0% and 79.7%, respectively. Furthermore, ProtoASNet provides interpretability and an uncertainty measure for each prediction, which can improve transparency and facilitate the interactive usage of deep networks to aid clinical decision-making. Our source code is available at: https://github.com/hooman007/ProtoASNet.","PeriodicalId":18289,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"23 1","pages":"368-378"},"PeriodicalIF":0.0,"publicationDate":"2023-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82308171","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Client-Level Differential Privacy via Adaptive Intermediary in Federated Medical Imaging 联邦医学成像中基于自适应中介的客户级差异隐私
Meirui Jiang, Yuan Zhong, Anjie Le, Xiaoxiao Li, Qianming Dou
Despite recent progress in enhancing the privacy of federated learning (FL) via differential privacy (DP), the trade-off of DP between privacy protection and performance is still underexplored for real-world medical scenario. In this paper, we propose to optimize the trade-off under the context of client-level DP, which focuses on privacy during communications. However, FL for medical imaging involves typically much fewer participants (hospitals) than other domains (e.g., mobile devices), thus ensuring clients be differentially private is much more challenging. To tackle this problem, we propose an adaptive intermediary strategy to improve performance without harming privacy. Specifically, we theoretically find splitting clients into sub-clients, which serve as intermediaries between hospitals and the server, can mitigate the noises introduced by DP without harming privacy. Our proposed approach is empirically evaluated on both classification and segmentation tasks using two public datasets, and its effectiveness is demonstrated with significant performance improvements and comprehensive analytical studies. Code is available at: https://github.com/med-air/Client-DP-FL.
尽管最近在通过差分隐私(DP)增强联邦学习(FL)的隐私方面取得了进展,但在现实医疗场景中,差分隐私保护与性能之间的权衡仍然没有得到充分的探讨。在本文中,我们提出在客户端级DP环境下优化权衡,该环境关注通信过程中的隐私。然而,医学成像的FL涉及的参与者(医院)通常比其他领域(例如,移动设备)少得多,因此确保客户与众不同的隐私更具挑战性。为了解决这个问题,我们提出了一种自适应中介策略,在不损害隐私的情况下提高性能。具体来说,我们在理论上发现将客户端划分为子客户端,这些子客户端充当医院和服务器之间的中介,可以在不损害隐私的情况下减轻DP引入的噪声。我们提出的方法使用两个公共数据集对分类和分割任务进行了实证评估,并通过显着的性能改进和全面的分析研究证明了其有效性。代码可从https://github.com/med-air/Client-DP-FL获得。
{"title":"Client-Level Differential Privacy via Adaptive Intermediary in Federated Medical Imaging","authors":"Meirui Jiang, Yuan Zhong, Anjie Le, Xiaoxiao Li, Qianming Dou","doi":"10.48550/arXiv.2307.12542","DOIUrl":"https://doi.org/10.48550/arXiv.2307.12542","url":null,"abstract":"Despite recent progress in enhancing the privacy of federated learning (FL) via differential privacy (DP), the trade-off of DP between privacy protection and performance is still underexplored for real-world medical scenario. In this paper, we propose to optimize the trade-off under the context of client-level DP, which focuses on privacy during communications. However, FL for medical imaging involves typically much fewer participants (hospitals) than other domains (e.g., mobile devices), thus ensuring clients be differentially private is much more challenging. To tackle this problem, we propose an adaptive intermediary strategy to improve performance without harming privacy. Specifically, we theoretically find splitting clients into sub-clients, which serve as intermediaries between hospitals and the server, can mitigate the noises introduced by DP without harming privacy. Our proposed approach is empirically evaluated on both classification and segmentation tasks using two public datasets, and its effectiveness is demonstrated with significant performance improvements and comprehensive analytical studies. Code is available at: https://github.com/med-air/Client-DP-FL.","PeriodicalId":18289,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"61 1","pages":"500-510"},"PeriodicalIF":0.0,"publicationDate":"2023-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79492490","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dense Transformer based Enhanced Coding Network for Unsupervised Metal Artifact Reduction 基于密集变压器的增强编码网络无监督金属伪影减少
Wangduo Xie, Matthew B. Blaschko
CT images corrupted by metal artifacts have serious negative effects on clinical diagnosis. Considering the difficulty of collecting paired data with ground truth in clinical settings, unsupervised methods for metal artifact reduction are of high interest. However, it is difficult for previous unsupervised methods to retain structural information from CT images while handling the non-local characteristics of metal artifacts. To address these challenges, we proposed a novel Dense Transformer based Enhanced Coding Network (DTEC-Net) for unsupervised metal artifact reduction. Specifically, we introduce a Hierarchical Disentangling Encoder, supported by the high-order dense process, and transformer to obtain densely encoded sequences with long-range correspondence. Then, we present a second-order disentanglement method to improve the dense sequence's decoding process. Extensive experiments and model discussions illustrate DTEC-Net's effectiveness, which outperforms the previous state-of-the-art methods on a benchmark dataset, and greatly reduces metal artifacts while restoring richer texture details.
金属伪影对CT图像的破坏对临床诊断有严重的负面影响。考虑到在临床环境中收集具有基础真理的配对数据的困难,金属伪影减少的无监督方法受到高度关注。然而,以往的无监督方法在处理金属伪影的非局部特征时,难以保留CT图像中的结构信息。为了解决这些挑战,我们提出了一种新的基于密集变压器的增强编码网络(DTEC-Net),用于无监督金属伪影的减少。具体来说,我们引入了一种分层解纠缠编码器,由高阶密集过程和变压器支持,以获得具有远程对应的密集编码序列。在此基础上,提出了一种二阶解纠缠方法来改善密集序列的解码过程。大量的实验和模型讨论说明了DTEC-Net的有效性,它在基准数据集上优于以前最先进的方法,并且在恢复更丰富的纹理细节的同时大大减少了金属伪像。
{"title":"Dense Transformer based Enhanced Coding Network for Unsupervised Metal Artifact Reduction","authors":"Wangduo Xie, Matthew B. Blaschko","doi":"10.48550/arXiv.2307.12717","DOIUrl":"https://doi.org/10.48550/arXiv.2307.12717","url":null,"abstract":"CT images corrupted by metal artifacts have serious negative effects on clinical diagnosis. Considering the difficulty of collecting paired data with ground truth in clinical settings, unsupervised methods for metal artifact reduction are of high interest. However, it is difficult for previous unsupervised methods to retain structural information from CT images while handling the non-local characteristics of metal artifacts. To address these challenges, we proposed a novel Dense Transformer based Enhanced Coding Network (DTEC-Net) for unsupervised metal artifact reduction. Specifically, we introduce a Hierarchical Disentangling Encoder, supported by the high-order dense process, and transformer to obtain densely encoded sequences with long-range correspondence. Then, we present a second-order disentanglement method to improve the dense sequence's decoding process. Extensive experiments and model discussions illustrate DTEC-Net's effectiveness, which outperforms the previous state-of-the-art methods on a benchmark dataset, and greatly reduces metal artifacts while restoring richer texture details.","PeriodicalId":18289,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"151 1","pages":"77-86"},"PeriodicalIF":0.0,"publicationDate":"2023-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86680007","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AMAE: Adaptation of Pre-Trained Masked Autoencoder for Dual-Distribution Anomaly Detection in Chest X-Rays 基于预训练掩码自编码器的胸部x线双分布异常检测
B. Bozorgtabar, D. Mahapatra, J. Thiran
Unsupervised anomaly detection in medical images such as chest radiographs is stepping into the spotlight as it mitigates the scarcity of the labor-intensive and costly expert annotation of anomaly data. However, nearly all existing methods are formulated as a one-class classification trained only on representations from the normal class and discard a potentially significant portion of the unlabeled data. This paper focuses on a more practical setting, dual distribution anomaly detection for chest X-rays, using the entire training data, including both normal and unlabeled images. Inspired by a modern self-supervised vision transformer model trained using partial image inputs to reconstruct missing image regions -- we propose AMAE, a two-stage algorithm for adaptation of the pre-trained masked autoencoder (MAE). Starting from MAE initialization, AMAE first creates synthetic anomalies from only normal training images and trains a lightweight classifier on frozen transformer features. Subsequently, we propose an adaptation strategy to leverage unlabeled images containing anomalies. The adaptation scheme is accomplished by assigning pseudo-labels to unlabeled images and using two separate MAE based modules to model the normative and anomalous distributions of pseudo-labeled images. The effectiveness of the proposed adaptation strategy is evaluated with different anomaly ratios in an unlabeled training set. AMAE leads to consistent performance gains over competing self-supervised and dual distribution anomaly detection methods, setting the new state-of-the-art on three public chest X-ray benchmarks: RSNA, NIH-CXR, and VinDr-CXR.
医学图像(如胸部x线片)中的无监督异常检测正逐渐成为人们关注的焦点,因为它减轻了异常数据的人工密集型和昂贵的专家注释的稀缺性。然而,几乎所有现有的方法都被表述为一个单类分类,只训练来自正常类的表示,并丢弃了未标记数据的潜在重要部分。本文的重点是一个更实际的设置,双分布异常检测胸部x射线,使用整个训练数据,包括正常和未标记的图像。受使用部分图像输入训练的现代自监督视觉变压器模型的启发,我们提出了AMAE,一种用于自适应预训练的掩码自编码器(MAE)的两阶段算法。从MAE初始化开始,AMAE首先仅从正常训练图像中创建合成异常,并在冻结变压器特征上训练轻量级分类器。随后,我们提出了一种适应策略,以利用包含异常的未标记图像。该自适应方案通过为未标记的图像分配伪标签,并使用两个独立的基于MAE的模块对伪标签图像的规范分布和异常分布进行建模来实现。在一个未标记的训练集中,用不同的异常比率来评估所提出的自适应策略的有效性。AMAE的性能优于竞争对手的自我监督和双分布异常检测方法,在三个公共胸部x射线基准:RSNA、NIH-CXR和VinDr-CXR上设定了最新的技术水平。
{"title":"AMAE: Adaptation of Pre-Trained Masked Autoencoder for Dual-Distribution Anomaly Detection in Chest X-Rays","authors":"B. Bozorgtabar, D. Mahapatra, J. Thiran","doi":"10.48550/arXiv.2307.12721","DOIUrl":"https://doi.org/10.48550/arXiv.2307.12721","url":null,"abstract":"Unsupervised anomaly detection in medical images such as chest radiographs is stepping into the spotlight as it mitigates the scarcity of the labor-intensive and costly expert annotation of anomaly data. However, nearly all existing methods are formulated as a one-class classification trained only on representations from the normal class and discard a potentially significant portion of the unlabeled data. This paper focuses on a more practical setting, dual distribution anomaly detection for chest X-rays, using the entire training data, including both normal and unlabeled images. Inspired by a modern self-supervised vision transformer model trained using partial image inputs to reconstruct missing image regions -- we propose AMAE, a two-stage algorithm for adaptation of the pre-trained masked autoencoder (MAE). Starting from MAE initialization, AMAE first creates synthetic anomalies from only normal training images and trains a lightweight classifier on frozen transformer features. Subsequently, we propose an adaptation strategy to leverage unlabeled images containing anomalies. The adaptation scheme is accomplished by assigning pseudo-labels to unlabeled images and using two separate MAE based modules to model the normative and anomalous distributions of pseudo-labeled images. The effectiveness of the proposed adaptation strategy is evaluated with different anomaly ratios in an unlabeled training set. AMAE leads to consistent performance gains over competing self-supervised and dual distribution anomaly detection methods, setting the new state-of-the-art on three public chest X-ray benchmarks: RSNA, NIH-CXR, and VinDr-CXR.","PeriodicalId":18289,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"109 ","pages":"195-205"},"PeriodicalIF":0.0,"publicationDate":"2023-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91461891","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Right for the Wrong Reason: Can Interpretable ML Techniques Detect Spurious Correlations? 正确的错误原因:可解释的ML技术能检测到虚假的相关性吗?
Susu Sun, Lisa M. Koch, Christian F. Baumgartner
While deep neural network models offer unmatched classification performance, they are prone to learning spurious correlations in the data. Such dependencies on confounding information can be difficult to detect using performance metrics if the test data comes from the same distribution as the training data. Interpretable ML methods such as post-hoc explanations or inherently interpretable classifiers promise to identify faulty model reasoning. However, there is mixed evidence whether many of these techniques are actually able to do so. In this paper, we propose a rigorous evaluation strategy to assess an explanation technique's ability to correctly identify spurious correlations. Using this strategy, we evaluate five post-hoc explanation techniques and one inherently interpretable method for their ability to detect three types of artificially added confounders in a chest x-ray diagnosis task. We find that the post-hoc technique SHAP, as well as the inherently interpretable Attri-Net provide the best performance and can be used to reliably identify faulty model behavior.
虽然深度神经网络模型提供了无与伦比的分类性能,但它们容易学习数据中的虚假相关性。如果测试数据来自与训练数据相同的分布,那么使用性能度量很难检测到这种对混杂信息的依赖。可解释的ML方法,如事后解释或固有可解释分类器,承诺识别错误的模型推理。然而,这些技术是否真的能做到这一点,证据不一。在本文中,我们提出了一个严格的评估策略来评估解释技术正确识别虚假相关的能力。使用该策略,我们评估了五种事后解释技术和一种内在可解释方法,以检测胸部x线诊断任务中人为添加的三种类型的混杂因素的能力。我们发现,事后技术SHAP,以及固有的可解释的Attri-Net提供了最好的性能,可以用来可靠地识别错误的模型行为。
{"title":"Right for the Wrong Reason: Can Interpretable ML Techniques Detect Spurious Correlations?","authors":"Susu Sun, Lisa M. Koch, Christian F. Baumgartner","doi":"10.48550/arXiv.2307.12344","DOIUrl":"https://doi.org/10.48550/arXiv.2307.12344","url":null,"abstract":"While deep neural network models offer unmatched classification performance, they are prone to learning spurious correlations in the data. Such dependencies on confounding information can be difficult to detect using performance metrics if the test data comes from the same distribution as the training data. Interpretable ML methods such as post-hoc explanations or inherently interpretable classifiers promise to identify faulty model reasoning. However, there is mixed evidence whether many of these techniques are actually able to do so. In this paper, we propose a rigorous evaluation strategy to assess an explanation technique's ability to correctly identify spurious correlations. Using this strategy, we evaluate five post-hoc explanation techniques and one inherently interpretable method for their ability to detect three types of artificially added confounders in a chest x-ray diagnosis task. We find that the post-hoc technique SHAP, as well as the inherently interpretable Attri-Net provide the best performance and can be used to reliably identify faulty model behavior.","PeriodicalId":18289,"journal":{"name":"Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention","volume":"72 1","pages":"425-434"},"PeriodicalIF":0.0,"publicationDate":"2023-07-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79126303","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
Medical image computing and computer-assisted intervention : MICCAI ... International Conference on Medical Image Computing and Computer-Assisted Intervention
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1