首页 > 最新文献

Computerized Medical Imaging and Graphics最新文献

英文 中文
Main challenges on the curation of large scale datasets for pancreas segmentation using deep learning in multi-phase CT scans: Focus on cardinality, manual refinement, and annotation quality 利用深度学习在多期 CT 扫描中进行胰腺分割的大型数据集整理工作面临的主要挑战:重点关注万有引力、人工完善和注释质量
IF 5.4 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-09-13 DOI: 10.1016/j.compmedimag.2024.102434
Matteo Cavicchioli , Andrea Moglia , Ludovica Pierelli , Giacomo Pugliese , Pietro Cerveri

Accurate segmentation of the pancreas in computed tomography (CT) holds paramount importance in diagnostics, surgical planning, and interventions. Recent studies have proposed supervised deep-learning models for segmentation, but their efficacy relies on the quality and quantity of the training data. Most of such works employed small-scale public datasets, without proving the efficacy of generalization to external datasets. This study explored the optimization of pancreas segmentation accuracy by pinpointing the ideal dataset size, understanding resource implications, examining manual refinement impact, and assessing the influence of anatomical subregions. We present the AIMS-1300 dataset encompassing 1,300 CT scans. Its manual annotation by medical experts required 938 h. A 2.5D UNet was implemented to assess the impact of training sample size on segmentation accuracy by partitioning the original AIMS-1300 dataset into 11 smaller subsets of progressively increasing numerosity. The findings revealed that training sets exceeding 440 CTs did not lead to better segmentation performance. In contrast, nnU-Net and UNet with Attention Gate reached a plateau for 585 CTs. Tests on generalization on the publicly available AMOS-CT dataset confirmed this outcome. As the size of the partition of the AIMS-1300 training set increases, the number of error slices decreases, reaching a minimum with 730 and 440 CTs, for AIMS-1300 and AMOS-CT datasets, respectively. Segmentation metrics on the AIMS-1300 and AMOS-CT datasets improved more on the head than the body and tail of the pancreas as the dataset size increased. By carefully considering the task and the characteristics of the available data, researchers can develop deep learning models without sacrificing performance even with limited data. This could accelerate developing and deploying artificial intelligence tools for pancreas surgery and other surgical data science applications.

计算机断层扫描(CT)中胰腺的精确分割在诊断、手术规划和干预中至关重要。最近的研究提出了用于分割的有监督深度学习模型,但其有效性取决于训练数据的质量和数量。大多数此类研究都采用了小规模的公共数据集,但并未证明其对外部数据集的泛化效果。本研究通过确定理想的数据集大小、了解资源影响、检查人工细化的影响以及评估解剖亚区的影响,来探索胰腺分割准确性的优化。我们展示的 AIMS-1300 数据集包含 1,300 张 CT 扫描图像。通过将原始 AIMS-1300 数据集划分为 11 个数量逐渐增加的较小子集,我们采用了 2.5D UNet 来评估训练样本数量对分割准确性的影响。结果显示,训练集超过 440 个 CT 并不能带来更好的分割性能。相比之下,nnU-Net 和 UNet with Attention Gate 在 585 CTs 时达到了最高点。在公开的 AMOS-CT 数据集上进行的泛化测试证实了这一结果。随着 AIMS-1300 训练集分区大小的增加,错误片段的数量也在减少,AIMS-1300 和 AMOS-CT 数据集分别在 730 和 440 CT 时达到最小值。随着数据集大小的增加,AIMS-1300 和 AMOS-CT 数据集上胰腺头部的分割指标比胰腺体部和尾部的改进更大。通过仔细考虑任务和可用数据的特征,研究人员可以开发出深度学习模型,即使数据有限,也不会牺牲性能。这将加速开发和部署用于胰腺手术和其他手术数据科学应用的人工智能工具。
{"title":"Main challenges on the curation of large scale datasets for pancreas segmentation using deep learning in multi-phase CT scans: Focus on cardinality, manual refinement, and annotation quality","authors":"Matteo Cavicchioli ,&nbsp;Andrea Moglia ,&nbsp;Ludovica Pierelli ,&nbsp;Giacomo Pugliese ,&nbsp;Pietro Cerveri","doi":"10.1016/j.compmedimag.2024.102434","DOIUrl":"10.1016/j.compmedimag.2024.102434","url":null,"abstract":"<div><p>Accurate segmentation of the pancreas in computed tomography (CT) holds paramount importance in diagnostics, surgical planning, and interventions. Recent studies have proposed supervised deep-learning models for segmentation, but their efficacy relies on the quality and quantity of the training data. Most of such works employed small-scale public datasets, without proving the efficacy of generalization to external datasets. This study explored the optimization of pancreas segmentation accuracy by pinpointing the ideal dataset size, understanding resource implications, examining manual refinement impact, and assessing the influence of anatomical subregions. We present the AIMS-1300 dataset encompassing 1,300 CT scans. Its manual annotation by medical experts required 938 h. A 2.5D UNet was implemented to assess the impact of training sample size on segmentation accuracy by partitioning the original AIMS-1300 dataset into 11 smaller subsets of progressively increasing numerosity. The findings revealed that training sets exceeding 440 CTs did not lead to better segmentation performance. In contrast, nnU-Net and UNet with Attention Gate reached a plateau for 585 CTs. Tests on generalization on the publicly available AMOS-CT dataset confirmed this outcome. As the size of the partition of the AIMS-1300 training set increases, the number of error slices decreases, reaching a minimum with 730 and 440 CTs, for AIMS-1300 and AMOS-CT datasets, respectively. Segmentation metrics on the AIMS-1300 and AMOS-CT datasets improved more on the head than the body and tail of the pancreas as the dataset size increased. By carefully considering the task and the characteristics of the available data, researchers can develop deep learning models without sacrificing performance even with limited data. This could accelerate developing and deploying artificial intelligence tools for pancreas surgery and other surgical data science applications.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"117 ","pages":"Article 102434"},"PeriodicalIF":5.4,"publicationDate":"2024-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0895611124001113/pdfft?md5=45c57efca48ce73af49c585837730bf7&pid=1-s2.0-S0895611124001113-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142233870","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards explainable oral cancer recognition: Screening on imperfect images via Informed Deep Learning and Case-Based Reasoning 实现可解释的口腔癌识别:通过知情深度学习和基于案例的推理对不完美图像进行筛查
IF 5.4 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-09-11 DOI: 10.1016/j.compmedimag.2024.102433
Marco Parola , Federico A. Galatolo , Gaetano La Mantia , Mario G.C.A. Cimino , Giuseppina Campisi , Olga Di Fede

Oral squamous cell carcinoma recognition presents a challenge due to late diagnosis and costly data acquisition. A cost-efficient, computerized screening system is crucial for early disease detection, minimizing the need for expert intervention and expensive analysis. Besides, transparency is essential to align these systems with critical sector applications. Explainable Artificial Intelligence (XAI) provides techniques for understanding models. However, current XAI is mostly data-driven and focused on addressing developers’ requirements of improving models rather than clinical users’ demands for expressing relevant insights. Among different XAI strategies, we propose a solution composed of Case-Based Reasoning paradigm to provide visual output explanations and Informed Deep Learning (IDL) to integrate medical knowledge within the system. A key aspect of our solution lies in its capability to handle data imperfections, including labeling inaccuracies and artifacts, thanks to an ensemble architecture on top of the deep learning (DL) workflow. We conducted several experimental benchmarks on a dataset collected in collaboration with medical centers. Our findings reveal that employing the IDL approach yields an accuracy of 85%, surpassing the 77% accuracy achieved by DL alone. Furthermore, we measured the human-centered explainability of the two approaches and IDL generates explanations more congruent with the clinical user demands.

口腔鳞状细胞癌的识别因诊断较晚和数据采集成本高昂而面临挑战。具有成本效益的计算机化筛查系统对于早期疾病检测至关重要,可最大限度地减少对专家干预和昂贵分析的需求。此外,要使这些系统与关键领域的应用保持一致,透明度也至关重要。可解释人工智能(XAI)提供了理解模型的技术。然而,目前的 XAI 大多是数据驱动的,侧重于满足开发人员改进模型的要求,而不是满足临床用户表达相关见解的需求。在不同的 XAI 策略中,我们提出了一种由基于案例的推理范式和知情深度学习(IDL)组成的解决方案,前者用于提供可视化的输出解释,后者用于在系统中整合医学知识。我们的解决方案的一个关键方面在于它能够处理数据不完善的问题,包括标记不准确和伪造,这要归功于深度学习(DL)工作流程之上的集合架构。我们在与医疗中心合作收集的数据集上进行了多项实验基准测试。我们的研究结果表明,采用 IDL 方法可获得 85% 的准确率,超过了单独使用 DL 所获得的 77% 的准确率。此外,我们还测量了这两种方法以人为本的可解释性,IDL 生成的解释更符合临床用户的需求。
{"title":"Towards explainable oral cancer recognition: Screening on imperfect images via Informed Deep Learning and Case-Based Reasoning","authors":"Marco Parola ,&nbsp;Federico A. Galatolo ,&nbsp;Gaetano La Mantia ,&nbsp;Mario G.C.A. Cimino ,&nbsp;Giuseppina Campisi ,&nbsp;Olga Di Fede","doi":"10.1016/j.compmedimag.2024.102433","DOIUrl":"10.1016/j.compmedimag.2024.102433","url":null,"abstract":"<div><p>Oral squamous cell carcinoma recognition presents a challenge due to late diagnosis and costly data acquisition. A cost-efficient, computerized screening system is crucial for early disease detection, minimizing the need for expert intervention and expensive analysis. Besides, transparency is essential to align these systems with critical sector applications. Explainable Artificial Intelligence (XAI) provides techniques for understanding models. However, current XAI is mostly data-driven and focused on addressing developers’ requirements of improving models rather than clinical users’ demands for expressing relevant insights. Among different XAI strategies, we propose a solution composed of Case-Based Reasoning paradigm to provide visual output explanations and Informed Deep Learning (IDL) to integrate medical knowledge within the system. A key aspect of our solution lies in its capability to handle data imperfections, including labeling inaccuracies and artifacts, thanks to an ensemble architecture on top of the deep learning (DL) workflow. We conducted several experimental benchmarks on a dataset collected in collaboration with medical centers. Our findings reveal that employing the IDL approach yields an accuracy of 85%, surpassing the 77% accuracy achieved by DL alone. Furthermore, we measured the human-centered explainability of the two approaches and IDL generates explanations more congruent with the clinical user demands.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"117 ","pages":"Article 102433"},"PeriodicalIF":5.4,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0895611124001101/pdfft?md5=96dc2c2aa07fe4189f5653f58184b9b0&pid=1-s2.0-S0895611124001101-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142228887","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hematoma expansion prediction in intracerebral hemorrhage patients by using synthesized CT images in an end-to-end deep learning framework 在端到端深度学习框架中使用合成 CT 图像预测脑内出血患者血肿扩大情况
IF 5.4 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-09-05 DOI: 10.1016/j.compmedimag.2024.102430
Cansu Yalcin , Valeriia Abramova , Mikel Terceño , Arnau Oliver , Yolanda Silva , Xavier Lladó

Spontaneous intracerebral hemorrhage (ICH) is a type of stroke less prevalent than ischemic stroke but associated with high mortality rates. Hematoma expansion (HE) is an increase in the bleeding that affects 30%–38% of hemorrhagic stroke patients. It is observed within 24 h of onset and associated with patient worsening. Clinically it is relevant to detect the patients that will develop HE from their initial computed tomography (CT) scans which could improve patient management and treatment decisions. However, this is a significant challenge due to the predictive nature of the task and its low prevalence, which hinders the availability of large datasets with the required longitudinal information. In this work, we present an end-to-end deep learning framework capable of predicting which cases will exhibit HE using only the initial basal image. We introduce a deep learning framework based on the 2D EfficientNet B0 model to predict the occurrence of HE using initial non-contrasted CT scans and their corresponding lesion annotation as priors. We used an in-house acquired dataset of 122 ICH patients, including 35 HE cases, containing longitudinal CT scans with manual lesion annotations in both basal and follow-up (obtained within 24 h after the basal scan). Experiments were conducted using a 5-fold cross-validation strategy. We addressed the limited data problem by incorporating synthetic images into the training process. To the best of our knowledge, our approach is novel in the field of HE prediction, being the first to use image synthesis to enhance results. We studied different scenarios such as training only with the original scans, using standard image augmentation techniques, and using synthetic image generation. The best performance was achieved by adding five generated versions of each image, along with standard data augmentation, during the training process. This significantly improved (p=0.0003) the performance obtained with our baseline model using directly the original CT scans from an Accuracy of 0.56 to 0.84, F1-Score of 0.53 to 0.82, Sensitivity of 0.51 to 0.77, and Specificity of 0.60 to 0.91, respectively. The proposed approach shows promising results in predicting HE, especially with the inclusion of synthetically generated images. The obtained results highlight the significance of this research direction, which has the potential to improve the clinical management of patients with hemorrhagic stroke. The code is available at: https://github.com/NIC-VICOROB/HE-prediction-SynthCT.

自发性脑内出血(ICH)是一种发病率低于缺血性中风但死亡率很高的中风类型。血肿扩大(HE)是指出血量增加,影响 30%-38% 的出血性中风患者。在发病后 24 小时内即可观察到,并与患者病情恶化有关。在临床上,从最初的计算机断层扫描(CT)中检测出会出现 HE 的患者具有重要意义,可改善患者管理和治疗决策。然而,由于这项任务具有预测性,而且发病率较低,这阻碍了具有所需纵向信息的大型数据集的可用性,因此这是一项重大挑战。在这项工作中,我们提出了一种端到端的深度学习框架,能够仅利用初始基底图像预测哪些病例会表现出 HE。我们引入了一个基于二维 EfficientNet B0 模型的深度学习框架,利用初始非对比 CT 扫描及其相应的病变注释作为先验,预测 HE 的发生。我们使用了内部获得的 122 例 ICH 患者数据集,其中包括 35 例 HE 病例,该数据集包含纵向 CT 扫描,并在基底扫描和随访(基底扫描后 24 小时内获得)中进行了人工病灶注释。实验采用了 5 倍交叉验证策略。我们在训练过程中加入了合成图像,从而解决了数据有限的问题。据我们所知,我们的方法是 HE 预测领域的新方法,也是第一个使用图像合成来提高结果的方法。我们研究了不同的情况,如仅使用原始扫描图像进行训练、使用标准图像增强技术以及使用合成图像生成技术。在训练过程中,在标准数据增强的同时,每张图像添加五个生成版本,从而达到最佳效果。这大大提高了(p=0.0003)直接使用原始 CT 扫描图像的基线模型的性能,准确率从 0.56 提高到 0.84,F1 分数从 0.53 提高到 0.82,灵敏度从 0.51 提高到 0.77,特异性从 0.60 提高到 0.91。所提出的方法在预测 HE 方面取得了可喜的成果,尤其是在包含合成图像的情况下。所取得的结果凸显了这一研究方向的重要意义,有望改善出血性中风患者的临床管理。代码见:https://github.com/NIC-VICOROB/HE-prediction-SynthCT。
{"title":"Hematoma expansion prediction in intracerebral hemorrhage patients by using synthesized CT images in an end-to-end deep learning framework","authors":"Cansu Yalcin ,&nbsp;Valeriia Abramova ,&nbsp;Mikel Terceño ,&nbsp;Arnau Oliver ,&nbsp;Yolanda Silva ,&nbsp;Xavier Lladó","doi":"10.1016/j.compmedimag.2024.102430","DOIUrl":"10.1016/j.compmedimag.2024.102430","url":null,"abstract":"<div><p>Spontaneous intracerebral hemorrhage (ICH) is a type of stroke less prevalent than ischemic stroke but associated with high mortality rates. Hematoma expansion (HE) is an increase in the bleeding that affects 30%–38% of hemorrhagic stroke patients. It is observed within 24 h of onset and associated with patient worsening. Clinically it is relevant to detect the patients that will develop HE from their initial computed tomography (CT) scans which could improve patient management and treatment decisions. However, this is a significant challenge due to the predictive nature of the task and its low prevalence, which hinders the availability of large datasets with the required longitudinal information. In this work, we present an end-to-end deep learning framework capable of predicting which cases will exhibit HE using only the initial basal image. We introduce a deep learning framework based on the 2D EfficientNet B0 model to predict the occurrence of HE using initial non-contrasted CT scans and their corresponding lesion annotation as priors. We used an in-house acquired dataset of 122 ICH patients, including 35 HE cases, containing longitudinal CT scans with manual lesion annotations in both basal and follow-up (obtained within 24 h after the basal scan). Experiments were conducted using a 5-fold cross-validation strategy. We addressed the limited data problem by incorporating synthetic images into the training process. To the best of our knowledge, our approach is novel in the field of HE prediction, being the first to use image synthesis to enhance results. We studied different scenarios such as training only with the original scans, using standard image augmentation techniques, and using synthetic image generation. The best performance was achieved by adding five generated versions of each image, along with standard data augmentation, during the training process. This significantly improved (<span><math><mrow><mi>p</mi><mo>=</mo><mn>0</mn><mo>.</mo><mn>0003</mn></mrow></math></span>) the performance obtained with our baseline model using directly the original CT scans from an Accuracy of 0.56 to 0.84, F1-Score of 0.53 to 0.82, Sensitivity of 0.51 to 0.77, and Specificity of 0.60 to 0.91, respectively. The proposed approach shows promising results in predicting HE, especially with the inclusion of synthetically generated images. The obtained results highlight the significance of this research direction, which has the potential to improve the clinical management of patients with hemorrhagic stroke. The code is available at: <span><span>https://github.com/NIC-VICOROB/HE-prediction-SynthCT</span><svg><path></path></svg></span>.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"117 ","pages":"Article 102430"},"PeriodicalIF":5.4,"publicationDate":"2024-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0895611124001071/pdfft?md5=c274f51b14553bc438e5f61e76ba2a00&pid=1-s2.0-S0895611124001071-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142161734","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CycleSGAN: A cycle-consistent and semantics-preserving generative adversarial network for unpaired MR-to-CT image synthesis CycleSGAN:用于无配对 MR-CT 图像合成的周期一致性和语义保留生成对抗网络。
IF 5.4 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-09-04 DOI: 10.1016/j.compmedimag.2024.102431
Runze Wang , Alexander F. Heimann , Moritz Tannast , Guoyan Zheng

CycleGAN has been leveraged to synthesize a CT image from an available MR image after trained on unpaired data. Due to the lack of direct constraints between the synthetic and the input images, CycleGAN cannot guarantee structural consistency and often generates inaccurate mappings that shift the anatomy, which is highly undesirable for downstream clinical applications such as MRI-guided radiotherapy treatment planning and PET/MRI attenuation correction. In this paper, we propose a cycle-consistent and semantics-preserving generative adversarial network, referred as CycleSGAN, for unpaired MR-to-CT image synthesis. Our design features a novel and generic way to incorporate semantic information into CycleGAN. This is done by designing a pair of three-player games within the CycleGAN framework where each three-player game consists of one generator and two discriminators to formulate two distinct types of adversarial learning: appearance adversarial learning and structure adversarial learning. These two types of adversarial learning are alternately trained to ensure both realistic image synthesis and semantic structure preservation. Results on unpaired hip MR-to-CT image synthesis show that our method produces better synthetic CT images in both accuracy and visual quality as compared to other state-of-the-art (SOTA) unpaired MR-to-CT image synthesis methods.

在对非配对数据进行训练后,CycleGAN 被用于根据可用的 MR 图像合成 CT 图像。由于合成图像与输入图像之间缺乏直接约束,CycleGAN 无法保证结构的一致性,经常会生成不准确的映射,从而使解剖结构发生偏移,这对于下游临床应用(如 MRI 引导的放射治疗规划和 PET/MRI 衰减校正)来说是非常不可取的。在本文中,我们提出了一种循环一致性和语义保护生成对抗网络(称为 CycleSGAN),用于非配对 MR-CT 图像合成。我们的设计采用了一种新颖而通用的方法,将语义信息纳入 CycleGAN。具体做法是在 CycleGAN 框架内设计一对三人博弈,每个三人博弈由一个生成器和两个判别器组成,从而形成两种不同类型的对抗学习:外观对抗学习和结构对抗学习。这两类对抗学习交替进行训练,以确保既能合成真实图像,又能保留语义结构。非配对髋关节 MR 到 CT 图像合成的结果表明,与其他最先进的(SOTA)非配对 MR 到 CT 图像合成方法相比,我们的方法在准确性和视觉质量方面都能生成更好的合成 CT 图像。
{"title":"CycleSGAN: A cycle-consistent and semantics-preserving generative adversarial network for unpaired MR-to-CT image synthesis","authors":"Runze Wang ,&nbsp;Alexander F. Heimann ,&nbsp;Moritz Tannast ,&nbsp;Guoyan Zheng","doi":"10.1016/j.compmedimag.2024.102431","DOIUrl":"10.1016/j.compmedimag.2024.102431","url":null,"abstract":"<div><p>CycleGAN has been leveraged to synthesize a CT image from an available MR image after trained on unpaired data. Due to the lack of direct constraints between the synthetic and the input images, CycleGAN cannot guarantee structural consistency and often generates inaccurate mappings that shift the anatomy, which is highly undesirable for downstream clinical applications such as MRI-guided radiotherapy treatment planning and PET/MRI attenuation correction. In this paper, we propose a cycle-consistent and semantics-preserving generative adversarial network, referred as CycleSGAN, for unpaired MR-to-CT image synthesis. Our design features a novel and generic way to incorporate semantic information into CycleGAN. This is done by designing a pair of three-player games within the CycleGAN framework where each three-player game consists of one generator and two discriminators to formulate two distinct types of adversarial learning: appearance adversarial learning and structure adversarial learning. These two types of adversarial learning are alternately trained to ensure both realistic image synthesis and semantic structure preservation. Results on unpaired hip MR-to-CT image synthesis show that our method produces better synthetic CT images in both accuracy and visual quality as compared to other state-of-the-art (SOTA) unpaired MR-to-CT image synthesis methods.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"117 ","pages":"Article 102431"},"PeriodicalIF":5.4,"publicationDate":"2024-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142146765","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A lung biopsy path planning algorithm based on the double spherical constraint Pareto and indicators’ importance-correlation degree 基于双球约束帕累托和指标重要性相关度的肺活检路径规划算法
IF 5.4 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-08-31 DOI: 10.1016/j.compmedimag.2024.102426
Hui Yang , Yu Zhang , Yuhang Gong , Jing Zhang , Ling He , Jianquan Zhong , Ling Tang

Lung cancer has the highest mortality rate among cancers. The commonly used clinical method for diagnosing lung cancer is the CT-guided percutaneous transthoracic lung biopsy (CT-PTLB), but this method requires a high level of clinical experience from doctors. In this work, an automatic path planning method for CT-PTLB is proposed to provide doctors with auxiliary advice on puncture paths. The proposed method comprises three steps: preprocessing, initial path selection, and path evaluation. During preprocessing, the chest organs required for subsequent path planning are segmented. During the initial path selection, a target point selection method for selecting biopsy samples according to biopsy sampling requirements is proposed, which includes a down-sampling algorithm suitable for different nodule shapes. Entry points are selected according to the selected target points and clinical constraints. During the path evaluation, the clinical needs of lung biopsy surgery are first quantified as path evaluation indicators and then divided according to their evaluation perspective into risk and execution indicators. Then, considering the impact of the correlation between indicators, a path scoring system based on the double spherical constraint Pareto and the importance-correlation degree of the indicators is proposed to evaluate the comprehensive performance of the planned paths. The proposed method is retrospectively tested on 6 CT images and prospectively tested on 25 CT images. The experimental results indicate that the method proposed in this work can be used to plan feasible puncture paths for different cases and can serve as an auxiliary tool for lung biopsy surgery.

肺癌是死亡率最高的癌症。临床上常用的肺癌诊断方法是CT引导下经皮经胸肺穿刺活检术(CT-PTLB),但这种方法对医生的临床经验要求很高。本研究提出了一种 CT-PTLB 自动路径规划方法,为医生提供穿刺路径的辅助建议。该方法包括三个步骤:预处理、初始路径选择和路径评估。在预处理过程中,对后续路径规划所需的胸部器官进行分割。在初始路径选择过程中,提出了根据活检取样要求选择活检样本的目标点选择方法,其中包括适合不同结节形状的向下取样算法。根据选定的目标点和临床限制条件选择入口点。在路径评价过程中,首先将肺活检手术的临床需求量化为路径评价指标,然后根据其评价角度分为风险指标和执行指标。然后,考虑到指标间相关性的影响,提出了基于双球面约束帕累托和指标重要性-相关度的路径评分体系,对规划路径的综合性能进行评价。提出的方法在 6 幅 CT 图像上进行了回顾性测试,并在 25 幅 CT 图像上进行了前瞻性测试。实验结果表明,本文提出的方法可用于规划不同病例的可行穿刺路径,可作为肺活检手术的辅助工具。
{"title":"A lung biopsy path planning algorithm based on the double spherical constraint Pareto and indicators’ importance-correlation degree","authors":"Hui Yang ,&nbsp;Yu Zhang ,&nbsp;Yuhang Gong ,&nbsp;Jing Zhang ,&nbsp;Ling He ,&nbsp;Jianquan Zhong ,&nbsp;Ling Tang","doi":"10.1016/j.compmedimag.2024.102426","DOIUrl":"10.1016/j.compmedimag.2024.102426","url":null,"abstract":"<div><p>Lung cancer has the highest mortality rate among cancers. The commonly used clinical method for diagnosing lung cancer is the CT-guided percutaneous transthoracic lung biopsy (CT-PTLB), but this method requires a high level of clinical experience from doctors. In this work, an automatic path planning method for CT-PTLB is proposed to provide doctors with auxiliary advice on puncture paths. The proposed method comprises three steps: preprocessing, initial path selection, and path evaluation. During preprocessing, the chest organs required for subsequent path planning are segmented. During the initial path selection, a target point selection method for selecting biopsy samples according to biopsy sampling requirements is proposed, which includes a down-sampling algorithm suitable for different nodule shapes. Entry points are selected according to the selected target points and clinical constraints. During the path evaluation, the clinical needs of lung biopsy surgery are first quantified as path evaluation indicators and then divided according to their evaluation perspective into risk and execution indicators. Then, considering the impact of the correlation between indicators, a path scoring system based on the double spherical constraint Pareto and the importance-correlation degree of the indicators is proposed to evaluate the comprehensive performance of the planned paths. The proposed method is retrospectively tested on 6 CT images and prospectively tested on 25 CT images. The experimental results indicate that the method proposed in this work can be used to plan feasible puncture paths for different cases and can serve as an auxiliary tool for lung biopsy surgery.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"117 ","pages":"Article 102426"},"PeriodicalIF":5.4,"publicationDate":"2024-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142243550","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Economical hybrid novelty detection leveraging global aleatoric semantic uncertainty for enhanced MRI-based ACL tear diagnosis 经济型混合新颖性检测利用全局历时语义不确定性,增强基于核磁共振成像的前交叉韧带撕裂诊断。
IF 5.4 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-08-29 DOI: 10.1016/j.compmedimag.2024.102424
Athanasios Siouras , Serafeim Moustakidis , George Chalatsis , Tuan Aqeel Bohoran , Michael Hantes , Marianna Vlychou , Sotiris Tasoulis , Archontis Giannakidis , Dimitrios Tsaopoulos

This study presents an innovative hybrid deep learning (DL) framework that reformulates the sagittal MRI-based anterior cruciate ligament (ACL) tear classification task as a novelty detection problem to tackle class imbalance. We introduce a highly discriminative novelty score, which leverages the aleatoric semantic uncertainty as this is modeled in the class scores outputted by the YOLOv5-nano object detection (OD) model. To account for tissue continuity, we propose using the global scores (probability vector) when the model is applied to the entire sagittal sequence. The second module of the proposed pipeline constitutes the MINIROCKET timeseries classification model for determining whether a knee has an ACL tear. To better evaluate the generalization capabilities of our approach, we also carry out cross-database testing involving two public databases (KneeMRI and MRNet) and a validation-only database from University General Hospital of Larissa, Greece. Our method consistently outperformed (p-value<0.05) the state-of-the-art (SOTA) approaches on the KneeMRI dataset and achieved better accuracy and sensitivity on the MRNet dataset. It also generalized remarkably good, especially when the model had been trained on KneeMRI. The presented framework generated at least 2.1 times less carbon emissions and consumed at least 2.6 times less energy, when compared with SOTA. The integration of aleatoric semantic uncertainty-based scores into a novelty detection framework, when combined with the use of lightweight OD and timeseries classification models, have the potential to revolutionize the MRI-based injury detection by setting a new precedent in diagnostic precision, speed and environmental sustainability. Our resource-efficient framework offers potential for widespread application.

本研究提出了一种创新的混合深度学习(DL)框架,该框架将基于矢状磁共振成像的前十字韧带(ACL)撕裂分类任务重新表述为新颖性检测问题,以解决类不平衡问题。我们引入了一种高区分度的新颖性评分,它利用了 YOLOv5-nano 物体检测(OD)模型输出的类评分中的不确定性语义建模。为了考虑组织的连续性,我们建议在将模型应用于整个矢状序列时使用全局分数(概率向量)。拟议流水线的第二个模块是 MINIROCKET 时间序列分类模型,用于确定膝关节是否有前交叉韧带撕裂。为了更好地评估我们方法的泛化能力,我们还进行了跨数据库测试,涉及两个公共数据库(KneeMRI 和 MRNet)和一个来自希腊拉里萨大学综合医院的验证数据库。我们的方法始终优于其他方法(p-value
{"title":"Economical hybrid novelty detection leveraging global aleatoric semantic uncertainty for enhanced MRI-based ACL tear diagnosis","authors":"Athanasios Siouras ,&nbsp;Serafeim Moustakidis ,&nbsp;George Chalatsis ,&nbsp;Tuan Aqeel Bohoran ,&nbsp;Michael Hantes ,&nbsp;Marianna Vlychou ,&nbsp;Sotiris Tasoulis ,&nbsp;Archontis Giannakidis ,&nbsp;Dimitrios Tsaopoulos","doi":"10.1016/j.compmedimag.2024.102424","DOIUrl":"10.1016/j.compmedimag.2024.102424","url":null,"abstract":"<div><p>This study presents an innovative hybrid deep learning (DL) framework that reformulates the sagittal MRI-based anterior cruciate ligament (ACL) tear classification task as a novelty detection problem to tackle class imbalance. We introduce a highly discriminative novelty score, which leverages the aleatoric semantic uncertainty as this is modeled in the class scores outputted by the YOLOv5-nano object detection (OD) model. To account for tissue continuity, we propose using the global scores (probability vector) when the model is applied to the entire sagittal sequence. The second module of the proposed pipeline constitutes the MINIROCKET timeseries classification model for determining whether a knee has an ACL tear. To better evaluate the generalization capabilities of our approach, we also carry out cross-database testing involving two public databases (KneeMRI and MRNet) and a validation-only database from University General Hospital of Larissa, Greece. Our method consistently outperformed (p-value&lt;0.05) the state-of-the-art (SOTA) approaches on the KneeMRI dataset and achieved better accuracy and sensitivity on the MRNet dataset. It also generalized remarkably good, especially when the model had been trained on KneeMRI. The presented framework generated at least 2.1 times less carbon emissions and consumed at least 2.6 times less energy, when compared with SOTA. The integration of aleatoric semantic uncertainty-based scores into a novelty detection framework, when combined with the use of lightweight OD and timeseries classification models, have the potential to revolutionize the MRI-based injury detection by setting a new precedent in diagnostic precision, speed and environmental sustainability. Our resource-efficient framework offers potential for widespread application.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"117 ","pages":"Article 102424"},"PeriodicalIF":5.4,"publicationDate":"2024-08-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0895611124001010/pdfft?md5=08030d7bf6b1d50cc1742401e8b8a65d&pid=1-s2.0-S0895611124001010-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142143465","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Unsupervised adversarial neural network for enhancing vasculature in photoacoustic tomography images using optical coherence tomography angiography 利用光学相干断层血管成像技术增强光声断层图像中血管的无监督对抗神经网络
IF 5.4 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-08-28 DOI: 10.1016/j.compmedimag.2024.102425
Yutian Zhong , Zhenyang Liu , Xiaoming Zhang , Zhaoyong Liang , Wufan Chen , Cuixia Dai , Li Qi

Photoacoustic tomography (PAT) is a powerful imaging modality for visualizing tissue physiology and exogenous contrast agents. However, PAT faces challenges in visualizing deep-seated vascular structures due to light scattering, absorption, and reduced signal intensity with depth. Optical coherence tomography angiography (OCTA) offers high-contrast visualization of vasculature networks, yet its imaging depth is limited to a millimeter scale. Herein, we propose OCPA-Net, a novel unsupervised deep learning method that utilizes the rich vascular feature of OCTA to enhance PAT images. Trained on unpaired OCTA and PAT images, OCPA-Net incorporates a vessel-aware attention module to enhance deep-seated vessel details captured from OCTA. It leverages a domain-adversarial loss function to enforce structural consistency and a novel identity invariant loss to mitigate excessive image content generation. We validate the structural fidelity of OCPA-Net on simulation experiments, and then demonstrate its vascular enhancement performance on in vivo imaging experiments of tumor-bearing mice and contrast-enhanced pregnant mice. The results show the promise of our method for comprehensive vessel-related image analysis in preclinical research applications.

光声断层成像(PAT)是一种强大的成像模式,可用于观察组织生理学和外源性造影剂。然而,由于光散射、吸收和信号强度随深度降低等原因,PAT 在观察深层血管结构方面面临挑战。光学相干断层血管成像(OCTA)可提供高对比度的血管网络可视化,但其成像深度仅限于毫米级。在此,我们提出了一种新颖的无监督深度学习方法 OCPA-Net,利用 OCTA 丰富的血管特征来增强 PAT 图像。OCPA-Net 在未配对的 OCTA 和 PAT 图像上进行训练,结合了血管感知注意模块,以增强从 OCTA 捕捉到的深层血管细节。它利用领域对抗损失函数来执行结构一致性,并利用新颖的身份不变损失来减少过多图像内容的生成。我们在模拟实验中验证了 OCPA-Net 的结构保真度,然后在肿瘤小鼠和对比度增强怀孕小鼠的体内成像实验中证明了它的血管增强性能。结果表明,我们的方法有望在临床前研究应用中进行全面的血管相关图像分析。
{"title":"Unsupervised adversarial neural network for enhancing vasculature in photoacoustic tomography images using optical coherence tomography angiography","authors":"Yutian Zhong ,&nbsp;Zhenyang Liu ,&nbsp;Xiaoming Zhang ,&nbsp;Zhaoyong Liang ,&nbsp;Wufan Chen ,&nbsp;Cuixia Dai ,&nbsp;Li Qi","doi":"10.1016/j.compmedimag.2024.102425","DOIUrl":"10.1016/j.compmedimag.2024.102425","url":null,"abstract":"<div><p>Photoacoustic tomography (PAT) is a powerful imaging modality for visualizing tissue physiology and exogenous contrast agents. However, PAT faces challenges in visualizing deep-seated vascular structures due to light scattering, absorption, and reduced signal intensity with depth. Optical coherence tomography angiography (OCTA) offers high-contrast visualization of vasculature networks, yet its imaging depth is limited to a millimeter scale. Herein, we propose OCPA-Net, a novel unsupervised deep learning method that utilizes the rich vascular feature of OCTA to enhance PAT images. Trained on unpaired OCTA and PAT images, OCPA-Net incorporates a vessel-aware attention module to enhance deep-seated vessel details captured from OCTA. It leverages a domain-adversarial loss function to enforce structural consistency and a novel identity invariant loss to mitigate excessive image content generation. We validate the structural fidelity of OCPA-Net on simulation experiments, and then demonstrate its vascular enhancement performance on <em>in vivo</em> imaging experiments of tumor-bearing mice and contrast-enhanced pregnant mice. The results show the promise of our method for comprehensive vessel-related image analysis in preclinical research applications.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"117 ","pages":"Article 102425"},"PeriodicalIF":5.4,"publicationDate":"2024-08-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142099452","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Cell comparative learning: A cervical cytopathology whole slide image classification method using normal and abnormal cells 细胞比较学习:使用正常和异常细胞的宫颈细胞病理学全玻片图像分类方法
IF 5.4 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-08-28 DOI: 10.1016/j.compmedimag.2024.102427
Jian Qin , Yongjun He , Yiqin Liang , Lanlan Kang , Jing Zhao , Bo Ding

Automated cervical cancer screening through computer-assisted diagnosis has shown considerable potential to improve screening accessibility and reduce associated costs and errors. However, classification performance on whole slide images (WSIs) remains suboptimal due to patient-specific variations. To improve the precision of the screening, pathologists not only analyze the characteristics of suspected abnormal cells, but also compare them with normal cells. Motivated by this practice, we propose a novel cervical cell comparative learning method that leverages pathologist knowledge to learn the differences between normal and suspected abnormal cells within the same WSI. Our method employs two pre-trained YOLOX models to detect suspected abnormal and normal cells in a given WSI. A self-supervised model then extracts features for the detected cells. Subsequently, a tailored Transformer encoder fuses the cell features to obtain WSI instance embeddings. Finally, attention-based multi-instance learning is applied to achieve classification. The experimental results show an AUC of 0.9319 for our proposed method. Moreover, the method achieved professional pathologist-level performance, indicating its potential for clinical applications.

通过计算机辅助诊断进行宫颈癌自动筛查在提高筛查的可及性、降低相关成本和减少误差方面具有相当大的潜力。然而,由于患者的个体差异,整张切片图像(WSI)的分类性能仍不理想。为了提高筛查的精确度,病理学家不仅要分析疑似异常细胞的特征,还要将它们与正常细胞进行比较。受这种做法的启发,我们提出了一种新颖的宫颈细胞比较学习方法,利用病理学家的知识来学习同一 WSI 中正常细胞和疑似异常细胞之间的差异。我们的方法采用两个预先训练好的 YOLOX 模型来检测给定 WSI 中的疑似异常细胞和正常细胞。然后,一个自监督模型提取检测到的细胞的特征。随后,量身定制的 Transformer 编码器会融合细胞特征,从而获得 WSI 实例嵌入。最后,应用基于注意力的多实例学习来实现分类。实验结果显示,我们提出的方法的 AUC 为 0.9319。此外,该方法还达到了专业病理学家的水平,这表明它具有临床应用的潜力。
{"title":"Cell comparative learning: A cervical cytopathology whole slide image classification method using normal and abnormal cells","authors":"Jian Qin ,&nbsp;Yongjun He ,&nbsp;Yiqin Liang ,&nbsp;Lanlan Kang ,&nbsp;Jing Zhao ,&nbsp;Bo Ding","doi":"10.1016/j.compmedimag.2024.102427","DOIUrl":"10.1016/j.compmedimag.2024.102427","url":null,"abstract":"<div><p>Automated cervical cancer screening through computer-assisted diagnosis has shown considerable potential to improve screening accessibility and reduce associated costs and errors. However, classification performance on whole slide images (WSIs) remains suboptimal due to patient-specific variations. To improve the precision of the screening, pathologists not only analyze the characteristics of suspected abnormal cells, but also compare them with normal cells. Motivated by this practice, we propose a novel cervical cell comparative learning method that leverages pathologist knowledge to learn the differences between normal and suspected abnormal cells within the same WSI. Our method employs two pre-trained YOLOX models to detect suspected abnormal and normal cells in a given WSI. A self-supervised model then extracts features for the detected cells. Subsequently, a tailored Transformer encoder fuses the cell features to obtain WSI instance embeddings. Finally, attention-based multi-instance learning is applied to achieve classification. The experimental results show an AUC of 0.9319 for our proposed method. Moreover, the method achieved professional pathologist-level performance, indicating its potential for clinical applications.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"117 ","pages":"Article 102427"},"PeriodicalIF":5.4,"publicationDate":"2024-08-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142099356","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Evidence modeling for reliability learning and interpretable decision-making under multi-modality medical image segmentation 多模态医学影像分割下可靠性学习和可解释决策的证据建模。
IF 5.4 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-08-07 DOI: 10.1016/j.compmedimag.2024.102422
Jianfeng Zhao , Shuo Li

Reliability learning and interpretable decision-making are crucial for multi-modality medical image segmentation. Although many works have attempted multi-modality medical image segmentation, they rarely explore how much reliability is provided by each modality for segmentation. Moreover, the existing approach of decision-making such as the softmax function lacks the interpretability for multi-modality fusion. In this study, we proposed a novel approach named contextual discounted evidential network (CDE-Net) for reliability learning and interpretable decision-making under multi-modality medical image segmentation. Specifically, the CDE-Net first models the semantic evidence by uncertainty measurement using the proposed evidential decision-making module. Then, it leverages the contextual discounted fusion layer to learn the reliability provided by each modality. Finally, a multi-level loss function is deployed for the optimization of evidence modeling and reliability learning. Moreover, this study elaborates on the framework interpretability by discussing the consistency between pixel attribution maps and the learned reliability coefficients. Extensive experiments are conducted on both multi-modality brain and liver datasets. The CDE-Net gains high performance with an average Dice score of 0.914 for brain tumor segmentation and 0.913 for liver tumor segmentation, which proves CDE-Net has great potential to facilitate the interpretation of artificial intelligence-based multi-modality medical image fusion.

可靠性学习和可解释的决策对于多模态医学图像分割至关重要。虽然许多研究都尝试过多模态医学影像分割,但很少探讨每种模态为分割提供了多少可靠性。此外,现有的决策方法(如 softmax 函数)缺乏多模态融合的可解释性。在这项研究中,我们提出了一种名为 "上下文折扣证据网络(CDE-Net)"的新方法,用于多模态医学图像分割下的可靠性学习和可解释性决策。具体来说,CDE-Net 首先利用所提出的证据决策模块,通过不确定性测量建立语义证据模型。然后,它利用上下文折扣融合层来学习每种模态提供的可靠性。最后,采用多级损失函数对证据建模和可靠性学习进行优化。此外,本研究还通过讨论像素归因图与学习到的可靠性系数之间的一致性,详细阐述了框架的可解释性。在多模态大脑和肝脏数据集上进行了广泛的实验。CDE-Net 在脑肿瘤分割和肝脏肿瘤分割中分别获得了 0.914 和 0.913 的平均 Dice 分数,这证明 CDE-Net 在促进基于人工智能的多模态医学影像融合的解释方面具有巨大潜力。
{"title":"Evidence modeling for reliability learning and interpretable decision-making under multi-modality medical image segmentation","authors":"Jianfeng Zhao ,&nbsp;Shuo Li","doi":"10.1016/j.compmedimag.2024.102422","DOIUrl":"10.1016/j.compmedimag.2024.102422","url":null,"abstract":"<div><p>Reliability learning and interpretable decision-making are crucial for multi-modality medical image segmentation. Although many works have attempted multi-modality medical image segmentation, they rarely explore how much reliability is provided by each modality for segmentation. Moreover, the existing approach of decision-making such as the <span><math><mrow><mi>s</mi><mi>o</mi><mi>f</mi><mi>t</mi><mi>m</mi><mi>a</mi><mi>x</mi></mrow></math></span> function lacks the interpretability for multi-modality fusion. In this study, we proposed a novel approach named contextual discounted evidential network (CDE-Net) for reliability learning and interpretable decision-making under multi-modality medical image segmentation. Specifically, the CDE-Net first models the semantic evidence by uncertainty measurement using the proposed evidential decision-making module. Then, it leverages the contextual discounted fusion layer to learn the reliability provided by each modality. Finally, a multi-level loss function is deployed for the optimization of evidence modeling and reliability learning. Moreover, this study elaborates on the framework interpretability by discussing the consistency between pixel attribution maps and the learned reliability coefficients. Extensive experiments are conducted on both multi-modality brain and liver datasets. The CDE-Net gains high performance with an average Dice score of 0.914 for brain tumor segmentation and 0.913 for liver tumor segmentation, which proves CDE-Net has great potential to facilitate the interpretation of artificial intelligence-based multi-modality medical image fusion.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"116 ","pages":"Article 102422"},"PeriodicalIF":5.4,"publicationDate":"2024-08-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141908211","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TLF: Triple learning framework for intracranial aneurysms segmentation from unreliable labeled CTA scans TLF:从不可靠的标记 CTA 扫描中分割颅内动脉瘤的三重学习框架
IF 5.4 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-07-26 DOI: 10.1016/j.compmedimag.2024.102421
Lei Chai , Shuangqian Xue , Daodao Tang , Jixin Liu , Ning Sun , Xiujuan Liu

Intracranial aneurysm (IA) is a prevalent disease that poses a significant threat to human health. The use of computed tomography angiography (CTA) as a diagnostic tool for IAs remains time-consuming and challenging. Deep neural networks (DNNs) have made significant advancements in the field of medical image segmentation. Nevertheless, training large-scale DNNs demands substantial quantities of high-quality labeled data, making the annotation of numerous brain CTA scans a challenging endeavor. To address these challenges and effectively develop a robust IAs segmentation model from a large amount of unlabeled training data, we propose a triple learning framework (TLF). The framework primarily consists of three learning paradigms: pseudo-supervised learning, contrastive learning, and confident learning. This paper introduces an enhanced mean teacher model and voxel-selective strategy to conduct pseudo-supervised learning on unreliable labeled training data. Concurrently, we construct the positive and negative training pairs within the high-level semantic feature space to improve the overall learning efficiency of the TLF through contrastive learning. In addition, a multi-scale confident learning is proposed to correct unreliable labels, which enables the acquisition of broader local structural information instead of relying on individual voxels. To evaluate the effectiveness of our method, we conducted extensive experiments on a self-built database of hundreds of cases of brain CTA scans with IAs. Experimental results demonstrate that our method can effectively learn a robust CTA scan-based IAs segmentation model using unreliable labeled data, outperforming state-of-the-art methods in terms of segmentation accuracy. Codes are released at https://github.com/XueShuangqian/TLF.

颅内动脉瘤(IA)是一种对人类健康构成重大威胁的常见疾病。使用计算机断层扫描血管造影术(CTA)作为动脉瘤的诊断工具仍然耗时且具有挑战性。深度神经网络(DNN)在医学图像分割领域取得了重大进展。然而,大规模 DNNs 的训练需要大量高质量的标记数据,因此对大量脑部 CTA 扫描进行标注是一项极具挑战性的工作。为了应对这些挑战,并从大量未标注的训练数据中有效地开发出稳健的 IAs 分割模型,我们提出了一个三重学习框架(TLF)。该框架主要包括三种学习范式:伪监督学习、对比学习和自信学习。本文引入了增强的平均教师模型和体素选择策略,在不可靠的标注训练数据上进行伪监督学习。同时,我们在高级语义特征空间中构建了正负训练对,通过对比学习提高了 TLF 的整体学习效率。此外,我们还提出了一种多尺度自信学习方法来纠正不可靠标签,从而获取更广泛的局部结构信息,而不是依赖单个体素。为了评估我们方法的有效性,我们在一个自建的数据库中进行了大量实验,该数据库包含数百例带有 IAs 的脑 CTA 扫描。实验结果表明,我们的方法能利用不可靠的标记数据有效地学习基于CTA扫描的IAs分割模型,在分割准确率方面优于最先进的方法。代码发布于 https://github.com/XueShuangqian/TLF。
{"title":"TLF: Triple learning framework for intracranial aneurysms segmentation from unreliable labeled CTA scans","authors":"Lei Chai ,&nbsp;Shuangqian Xue ,&nbsp;Daodao Tang ,&nbsp;Jixin Liu ,&nbsp;Ning Sun ,&nbsp;Xiujuan Liu","doi":"10.1016/j.compmedimag.2024.102421","DOIUrl":"10.1016/j.compmedimag.2024.102421","url":null,"abstract":"<div><p>Intracranial aneurysm (IA) is a prevalent disease that poses a significant threat to human health. The use of computed tomography angiography (CTA) as a diagnostic tool for IAs remains time-consuming and challenging. Deep neural networks (DNNs) have made significant advancements in the field of medical image segmentation. Nevertheless, training large-scale DNNs demands substantial quantities of high-quality labeled data, making the annotation of numerous brain CTA scans a challenging endeavor. To address these challenges and effectively develop a robust IAs segmentation model from a large amount of unlabeled training data, we propose a triple learning framework (TLF). The framework primarily consists of three learning paradigms: pseudo-supervised learning, contrastive learning, and confident learning. This paper introduces an enhanced mean teacher model and voxel-selective strategy to conduct pseudo-supervised learning on unreliable labeled training data. Concurrently, we construct the positive and negative training pairs within the high-level semantic feature space to improve the overall learning efficiency of the TLF through contrastive learning. In addition, a multi-scale confident learning is proposed to correct unreliable labels, which enables the acquisition of broader local structural information instead of relying on individual voxels. To evaluate the effectiveness of our method, we conducted extensive experiments on a self-built database of hundreds of cases of brain CTA scans with IAs. Experimental results demonstrate that our method can effectively learn a robust CTA scan-based IAs segmentation model using unreliable labeled data, outperforming state-of-the-art methods in terms of segmentation accuracy. Codes are released at https://github.com/XueShuangqian/TLF.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"116 ","pages":"Article 102421"},"PeriodicalIF":5.4,"publicationDate":"2024-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141842424","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Computerized Medical Imaging and Graphics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1