Pub Date : 2024-09-13DOI: 10.1016/j.compmedimag.2024.102434
Matteo Cavicchioli , Andrea Moglia , Ludovica Pierelli , Giacomo Pugliese , Pietro Cerveri
Accurate segmentation of the pancreas in computed tomography (CT) holds paramount importance in diagnostics, surgical planning, and interventions. Recent studies have proposed supervised deep-learning models for segmentation, but their efficacy relies on the quality and quantity of the training data. Most of such works employed small-scale public datasets, without proving the efficacy of generalization to external datasets. This study explored the optimization of pancreas segmentation accuracy by pinpointing the ideal dataset size, understanding resource implications, examining manual refinement impact, and assessing the influence of anatomical subregions. We present the AIMS-1300 dataset encompassing 1,300 CT scans. Its manual annotation by medical experts required 938 h. A 2.5D UNet was implemented to assess the impact of training sample size on segmentation accuracy by partitioning the original AIMS-1300 dataset into 11 smaller subsets of progressively increasing numerosity. The findings revealed that training sets exceeding 440 CTs did not lead to better segmentation performance. In contrast, nnU-Net and UNet with Attention Gate reached a plateau for 585 CTs. Tests on generalization on the publicly available AMOS-CT dataset confirmed this outcome. As the size of the partition of the AIMS-1300 training set increases, the number of error slices decreases, reaching a minimum with 730 and 440 CTs, for AIMS-1300 and AMOS-CT datasets, respectively. Segmentation metrics on the AIMS-1300 and AMOS-CT datasets improved more on the head than the body and tail of the pancreas as the dataset size increased. By carefully considering the task and the characteristics of the available data, researchers can develop deep learning models without sacrificing performance even with limited data. This could accelerate developing and deploying artificial intelligence tools for pancreas surgery and other surgical data science applications.
{"title":"Main challenges on the curation of large scale datasets for pancreas segmentation using deep learning in multi-phase CT scans: Focus on cardinality, manual refinement, and annotation quality","authors":"Matteo Cavicchioli , Andrea Moglia , Ludovica Pierelli , Giacomo Pugliese , Pietro Cerveri","doi":"10.1016/j.compmedimag.2024.102434","DOIUrl":"10.1016/j.compmedimag.2024.102434","url":null,"abstract":"<div><p>Accurate segmentation of the pancreas in computed tomography (CT) holds paramount importance in diagnostics, surgical planning, and interventions. Recent studies have proposed supervised deep-learning models for segmentation, but their efficacy relies on the quality and quantity of the training data. Most of such works employed small-scale public datasets, without proving the efficacy of generalization to external datasets. This study explored the optimization of pancreas segmentation accuracy by pinpointing the ideal dataset size, understanding resource implications, examining manual refinement impact, and assessing the influence of anatomical subregions. We present the AIMS-1300 dataset encompassing 1,300 CT scans. Its manual annotation by medical experts required 938 h. A 2.5D UNet was implemented to assess the impact of training sample size on segmentation accuracy by partitioning the original AIMS-1300 dataset into 11 smaller subsets of progressively increasing numerosity. The findings revealed that training sets exceeding 440 CTs did not lead to better segmentation performance. In contrast, nnU-Net and UNet with Attention Gate reached a plateau for 585 CTs. Tests on generalization on the publicly available AMOS-CT dataset confirmed this outcome. As the size of the partition of the AIMS-1300 training set increases, the number of error slices decreases, reaching a minimum with 730 and 440 CTs, for AIMS-1300 and AMOS-CT datasets, respectively. Segmentation metrics on the AIMS-1300 and AMOS-CT datasets improved more on the head than the body and tail of the pancreas as the dataset size increased. By carefully considering the task and the characteristics of the available data, researchers can develop deep learning models without sacrificing performance even with limited data. This could accelerate developing and deploying artificial intelligence tools for pancreas surgery and other surgical data science applications.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"117 ","pages":"Article 102434"},"PeriodicalIF":5.4,"publicationDate":"2024-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0895611124001113/pdfft?md5=45c57efca48ce73af49c585837730bf7&pid=1-s2.0-S0895611124001113-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142233870","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-11DOI: 10.1016/j.compmedimag.2024.102433
Marco Parola , Federico A. Galatolo , Gaetano La Mantia , Mario G.C.A. Cimino , Giuseppina Campisi , Olga Di Fede
Oral squamous cell carcinoma recognition presents a challenge due to late diagnosis and costly data acquisition. A cost-efficient, computerized screening system is crucial for early disease detection, minimizing the need for expert intervention and expensive analysis. Besides, transparency is essential to align these systems with critical sector applications. Explainable Artificial Intelligence (XAI) provides techniques for understanding models. However, current XAI is mostly data-driven and focused on addressing developers’ requirements of improving models rather than clinical users’ demands for expressing relevant insights. Among different XAI strategies, we propose a solution composed of Case-Based Reasoning paradigm to provide visual output explanations and Informed Deep Learning (IDL) to integrate medical knowledge within the system. A key aspect of our solution lies in its capability to handle data imperfections, including labeling inaccuracies and artifacts, thanks to an ensemble architecture on top of the deep learning (DL) workflow. We conducted several experimental benchmarks on a dataset collected in collaboration with medical centers. Our findings reveal that employing the IDL approach yields an accuracy of 85%, surpassing the 77% accuracy achieved by DL alone. Furthermore, we measured the human-centered explainability of the two approaches and IDL generates explanations more congruent with the clinical user demands.
{"title":"Towards explainable oral cancer recognition: Screening on imperfect images via Informed Deep Learning and Case-Based Reasoning","authors":"Marco Parola , Federico A. Galatolo , Gaetano La Mantia , Mario G.C.A. Cimino , Giuseppina Campisi , Olga Di Fede","doi":"10.1016/j.compmedimag.2024.102433","DOIUrl":"10.1016/j.compmedimag.2024.102433","url":null,"abstract":"<div><p>Oral squamous cell carcinoma recognition presents a challenge due to late diagnosis and costly data acquisition. A cost-efficient, computerized screening system is crucial for early disease detection, minimizing the need for expert intervention and expensive analysis. Besides, transparency is essential to align these systems with critical sector applications. Explainable Artificial Intelligence (XAI) provides techniques for understanding models. However, current XAI is mostly data-driven and focused on addressing developers’ requirements of improving models rather than clinical users’ demands for expressing relevant insights. Among different XAI strategies, we propose a solution composed of Case-Based Reasoning paradigm to provide visual output explanations and Informed Deep Learning (IDL) to integrate medical knowledge within the system. A key aspect of our solution lies in its capability to handle data imperfections, including labeling inaccuracies and artifacts, thanks to an ensemble architecture on top of the deep learning (DL) workflow. We conducted several experimental benchmarks on a dataset collected in collaboration with medical centers. Our findings reveal that employing the IDL approach yields an accuracy of 85%, surpassing the 77% accuracy achieved by DL alone. Furthermore, we measured the human-centered explainability of the two approaches and IDL generates explanations more congruent with the clinical user demands.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"117 ","pages":"Article 102433"},"PeriodicalIF":5.4,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0895611124001101/pdfft?md5=96dc2c2aa07fe4189f5653f58184b9b0&pid=1-s2.0-S0895611124001101-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142228887","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-05DOI: 10.1016/j.compmedimag.2024.102430
Cansu Yalcin , Valeriia Abramova , Mikel Terceño , Arnau Oliver , Yolanda Silva , Xavier Lladó
Spontaneous intracerebral hemorrhage (ICH) is a type of stroke less prevalent than ischemic stroke but associated with high mortality rates. Hematoma expansion (HE) is an increase in the bleeding that affects 30%–38% of hemorrhagic stroke patients. It is observed within 24 h of onset and associated with patient worsening. Clinically it is relevant to detect the patients that will develop HE from their initial computed tomography (CT) scans which could improve patient management and treatment decisions. However, this is a significant challenge due to the predictive nature of the task and its low prevalence, which hinders the availability of large datasets with the required longitudinal information. In this work, we present an end-to-end deep learning framework capable of predicting which cases will exhibit HE using only the initial basal image. We introduce a deep learning framework based on the 2D EfficientNet B0 model to predict the occurrence of HE using initial non-contrasted CT scans and their corresponding lesion annotation as priors. We used an in-house acquired dataset of 122 ICH patients, including 35 HE cases, containing longitudinal CT scans with manual lesion annotations in both basal and follow-up (obtained within 24 h after the basal scan). Experiments were conducted using a 5-fold cross-validation strategy. We addressed the limited data problem by incorporating synthetic images into the training process. To the best of our knowledge, our approach is novel in the field of HE prediction, being the first to use image synthesis to enhance results. We studied different scenarios such as training only with the original scans, using standard image augmentation techniques, and using synthetic image generation. The best performance was achieved by adding five generated versions of each image, along with standard data augmentation, during the training process. This significantly improved () the performance obtained with our baseline model using directly the original CT scans from an Accuracy of 0.56 to 0.84, F1-Score of 0.53 to 0.82, Sensitivity of 0.51 to 0.77, and Specificity of 0.60 to 0.91, respectively. The proposed approach shows promising results in predicting HE, especially with the inclusion of synthetically generated images. The obtained results highlight the significance of this research direction, which has the potential to improve the clinical management of patients with hemorrhagic stroke. The code is available at: https://github.com/NIC-VICOROB/HE-prediction-SynthCT.
自发性脑内出血(ICH)是一种发病率低于缺血性中风但死亡率很高的中风类型。血肿扩大(HE)是指出血量增加,影响 30%-38% 的出血性中风患者。在发病后 24 小时内即可观察到,并与患者病情恶化有关。在临床上,从最初的计算机断层扫描(CT)中检测出会出现 HE 的患者具有重要意义,可改善患者管理和治疗决策。然而,由于这项任务具有预测性,而且发病率较低,这阻碍了具有所需纵向信息的大型数据集的可用性,因此这是一项重大挑战。在这项工作中,我们提出了一种端到端的深度学习框架,能够仅利用初始基底图像预测哪些病例会表现出 HE。我们引入了一个基于二维 EfficientNet B0 模型的深度学习框架,利用初始非对比 CT 扫描及其相应的病变注释作为先验,预测 HE 的发生。我们使用了内部获得的 122 例 ICH 患者数据集,其中包括 35 例 HE 病例,该数据集包含纵向 CT 扫描,并在基底扫描和随访(基底扫描后 24 小时内获得)中进行了人工病灶注释。实验采用了 5 倍交叉验证策略。我们在训练过程中加入了合成图像,从而解决了数据有限的问题。据我们所知,我们的方法是 HE 预测领域的新方法,也是第一个使用图像合成来提高结果的方法。我们研究了不同的情况,如仅使用原始扫描图像进行训练、使用标准图像增强技术以及使用合成图像生成技术。在训练过程中,在标准数据增强的同时,每张图像添加五个生成版本,从而达到最佳效果。这大大提高了(p=0.0003)直接使用原始 CT 扫描图像的基线模型的性能,准确率从 0.56 提高到 0.84,F1 分数从 0.53 提高到 0.82,灵敏度从 0.51 提高到 0.77,特异性从 0.60 提高到 0.91。所提出的方法在预测 HE 方面取得了可喜的成果,尤其是在包含合成图像的情况下。所取得的结果凸显了这一研究方向的重要意义,有望改善出血性中风患者的临床管理。代码见:https://github.com/NIC-VICOROB/HE-prediction-SynthCT。
{"title":"Hematoma expansion prediction in intracerebral hemorrhage patients by using synthesized CT images in an end-to-end deep learning framework","authors":"Cansu Yalcin , Valeriia Abramova , Mikel Terceño , Arnau Oliver , Yolanda Silva , Xavier Lladó","doi":"10.1016/j.compmedimag.2024.102430","DOIUrl":"10.1016/j.compmedimag.2024.102430","url":null,"abstract":"<div><p>Spontaneous intracerebral hemorrhage (ICH) is a type of stroke less prevalent than ischemic stroke but associated with high mortality rates. Hematoma expansion (HE) is an increase in the bleeding that affects 30%–38% of hemorrhagic stroke patients. It is observed within 24 h of onset and associated with patient worsening. Clinically it is relevant to detect the patients that will develop HE from their initial computed tomography (CT) scans which could improve patient management and treatment decisions. However, this is a significant challenge due to the predictive nature of the task and its low prevalence, which hinders the availability of large datasets with the required longitudinal information. In this work, we present an end-to-end deep learning framework capable of predicting which cases will exhibit HE using only the initial basal image. We introduce a deep learning framework based on the 2D EfficientNet B0 model to predict the occurrence of HE using initial non-contrasted CT scans and their corresponding lesion annotation as priors. We used an in-house acquired dataset of 122 ICH patients, including 35 HE cases, containing longitudinal CT scans with manual lesion annotations in both basal and follow-up (obtained within 24 h after the basal scan). Experiments were conducted using a 5-fold cross-validation strategy. We addressed the limited data problem by incorporating synthetic images into the training process. To the best of our knowledge, our approach is novel in the field of HE prediction, being the first to use image synthesis to enhance results. We studied different scenarios such as training only with the original scans, using standard image augmentation techniques, and using synthetic image generation. The best performance was achieved by adding five generated versions of each image, along with standard data augmentation, during the training process. This significantly improved (<span><math><mrow><mi>p</mi><mo>=</mo><mn>0</mn><mo>.</mo><mn>0003</mn></mrow></math></span>) the performance obtained with our baseline model using directly the original CT scans from an Accuracy of 0.56 to 0.84, F1-Score of 0.53 to 0.82, Sensitivity of 0.51 to 0.77, and Specificity of 0.60 to 0.91, respectively. The proposed approach shows promising results in predicting HE, especially with the inclusion of synthetically generated images. The obtained results highlight the significance of this research direction, which has the potential to improve the clinical management of patients with hemorrhagic stroke. The code is available at: <span><span>https://github.com/NIC-VICOROB/HE-prediction-SynthCT</span><svg><path></path></svg></span>.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"117 ","pages":"Article 102430"},"PeriodicalIF":5.4,"publicationDate":"2024-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0895611124001071/pdfft?md5=c274f51b14553bc438e5f61e76ba2a00&pid=1-s2.0-S0895611124001071-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142161734","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-04DOI: 10.1016/j.compmedimag.2024.102431
Runze Wang , Alexander F. Heimann , Moritz Tannast , Guoyan Zheng
CycleGAN has been leveraged to synthesize a CT image from an available MR image after trained on unpaired data. Due to the lack of direct constraints between the synthetic and the input images, CycleGAN cannot guarantee structural consistency and often generates inaccurate mappings that shift the anatomy, which is highly undesirable for downstream clinical applications such as MRI-guided radiotherapy treatment planning and PET/MRI attenuation correction. In this paper, we propose a cycle-consistent and semantics-preserving generative adversarial network, referred as CycleSGAN, for unpaired MR-to-CT image synthesis. Our design features a novel and generic way to incorporate semantic information into CycleGAN. This is done by designing a pair of three-player games within the CycleGAN framework where each three-player game consists of one generator and two discriminators to formulate two distinct types of adversarial learning: appearance adversarial learning and structure adversarial learning. These two types of adversarial learning are alternately trained to ensure both realistic image synthesis and semantic structure preservation. Results on unpaired hip MR-to-CT image synthesis show that our method produces better synthetic CT images in both accuracy and visual quality as compared to other state-of-the-art (SOTA) unpaired MR-to-CT image synthesis methods.
{"title":"CycleSGAN: A cycle-consistent and semantics-preserving generative adversarial network for unpaired MR-to-CT image synthesis","authors":"Runze Wang , Alexander F. Heimann , Moritz Tannast , Guoyan Zheng","doi":"10.1016/j.compmedimag.2024.102431","DOIUrl":"10.1016/j.compmedimag.2024.102431","url":null,"abstract":"<div><p>CycleGAN has been leveraged to synthesize a CT image from an available MR image after trained on unpaired data. Due to the lack of direct constraints between the synthetic and the input images, CycleGAN cannot guarantee structural consistency and often generates inaccurate mappings that shift the anatomy, which is highly undesirable for downstream clinical applications such as MRI-guided radiotherapy treatment planning and PET/MRI attenuation correction. In this paper, we propose a cycle-consistent and semantics-preserving generative adversarial network, referred as CycleSGAN, for unpaired MR-to-CT image synthesis. Our design features a novel and generic way to incorporate semantic information into CycleGAN. This is done by designing a pair of three-player games within the CycleGAN framework where each three-player game consists of one generator and two discriminators to formulate two distinct types of adversarial learning: appearance adversarial learning and structure adversarial learning. These two types of adversarial learning are alternately trained to ensure both realistic image synthesis and semantic structure preservation. Results on unpaired hip MR-to-CT image synthesis show that our method produces better synthetic CT images in both accuracy and visual quality as compared to other state-of-the-art (SOTA) unpaired MR-to-CT image synthesis methods.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"117 ","pages":"Article 102431"},"PeriodicalIF":5.4,"publicationDate":"2024-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142146765","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-31DOI: 10.1016/j.compmedimag.2024.102426
Hui Yang , Yu Zhang , Yuhang Gong , Jing Zhang , Ling He , Jianquan Zhong , Ling Tang
Lung cancer has the highest mortality rate among cancers. The commonly used clinical method for diagnosing lung cancer is the CT-guided percutaneous transthoracic lung biopsy (CT-PTLB), but this method requires a high level of clinical experience from doctors. In this work, an automatic path planning method for CT-PTLB is proposed to provide doctors with auxiliary advice on puncture paths. The proposed method comprises three steps: preprocessing, initial path selection, and path evaluation. During preprocessing, the chest organs required for subsequent path planning are segmented. During the initial path selection, a target point selection method for selecting biopsy samples according to biopsy sampling requirements is proposed, which includes a down-sampling algorithm suitable for different nodule shapes. Entry points are selected according to the selected target points and clinical constraints. During the path evaluation, the clinical needs of lung biopsy surgery are first quantified as path evaluation indicators and then divided according to their evaluation perspective into risk and execution indicators. Then, considering the impact of the correlation between indicators, a path scoring system based on the double spherical constraint Pareto and the importance-correlation degree of the indicators is proposed to evaluate the comprehensive performance of the planned paths. The proposed method is retrospectively tested on 6 CT images and prospectively tested on 25 CT images. The experimental results indicate that the method proposed in this work can be used to plan feasible puncture paths for different cases and can serve as an auxiliary tool for lung biopsy surgery.
{"title":"A lung biopsy path planning algorithm based on the double spherical constraint Pareto and indicators’ importance-correlation degree","authors":"Hui Yang , Yu Zhang , Yuhang Gong , Jing Zhang , Ling He , Jianquan Zhong , Ling Tang","doi":"10.1016/j.compmedimag.2024.102426","DOIUrl":"10.1016/j.compmedimag.2024.102426","url":null,"abstract":"<div><p>Lung cancer has the highest mortality rate among cancers. The commonly used clinical method for diagnosing lung cancer is the CT-guided percutaneous transthoracic lung biopsy (CT-PTLB), but this method requires a high level of clinical experience from doctors. In this work, an automatic path planning method for CT-PTLB is proposed to provide doctors with auxiliary advice on puncture paths. The proposed method comprises three steps: preprocessing, initial path selection, and path evaluation. During preprocessing, the chest organs required for subsequent path planning are segmented. During the initial path selection, a target point selection method for selecting biopsy samples according to biopsy sampling requirements is proposed, which includes a down-sampling algorithm suitable for different nodule shapes. Entry points are selected according to the selected target points and clinical constraints. During the path evaluation, the clinical needs of lung biopsy surgery are first quantified as path evaluation indicators and then divided according to their evaluation perspective into risk and execution indicators. Then, considering the impact of the correlation between indicators, a path scoring system based on the double spherical constraint Pareto and the importance-correlation degree of the indicators is proposed to evaluate the comprehensive performance of the planned paths. The proposed method is retrospectively tested on 6 CT images and prospectively tested on 25 CT images. The experimental results indicate that the method proposed in this work can be used to plan feasible puncture paths for different cases and can serve as an auxiliary tool for lung biopsy surgery.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"117 ","pages":"Article 102426"},"PeriodicalIF":5.4,"publicationDate":"2024-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142243550","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This study presents an innovative hybrid deep learning (DL) framework that reformulates the sagittal MRI-based anterior cruciate ligament (ACL) tear classification task as a novelty detection problem to tackle class imbalance. We introduce a highly discriminative novelty score, which leverages the aleatoric semantic uncertainty as this is modeled in the class scores outputted by the YOLOv5-nano object detection (OD) model. To account for tissue continuity, we propose using the global scores (probability vector) when the model is applied to the entire sagittal sequence. The second module of the proposed pipeline constitutes the MINIROCKET timeseries classification model for determining whether a knee has an ACL tear. To better evaluate the generalization capabilities of our approach, we also carry out cross-database testing involving two public databases (KneeMRI and MRNet) and a validation-only database from University General Hospital of Larissa, Greece. Our method consistently outperformed (p-value<0.05) the state-of-the-art (SOTA) approaches on the KneeMRI dataset and achieved better accuracy and sensitivity on the MRNet dataset. It also generalized remarkably good, especially when the model had been trained on KneeMRI. The presented framework generated at least 2.1 times less carbon emissions and consumed at least 2.6 times less energy, when compared with SOTA. The integration of aleatoric semantic uncertainty-based scores into a novelty detection framework, when combined with the use of lightweight OD and timeseries classification models, have the potential to revolutionize the MRI-based injury detection by setting a new precedent in diagnostic precision, speed and environmental sustainability. Our resource-efficient framework offers potential for widespread application.
{"title":"Economical hybrid novelty detection leveraging global aleatoric semantic uncertainty for enhanced MRI-based ACL tear diagnosis","authors":"Athanasios Siouras , Serafeim Moustakidis , George Chalatsis , Tuan Aqeel Bohoran , Michael Hantes , Marianna Vlychou , Sotiris Tasoulis , Archontis Giannakidis , Dimitrios Tsaopoulos","doi":"10.1016/j.compmedimag.2024.102424","DOIUrl":"10.1016/j.compmedimag.2024.102424","url":null,"abstract":"<div><p>This study presents an innovative hybrid deep learning (DL) framework that reformulates the sagittal MRI-based anterior cruciate ligament (ACL) tear classification task as a novelty detection problem to tackle class imbalance. We introduce a highly discriminative novelty score, which leverages the aleatoric semantic uncertainty as this is modeled in the class scores outputted by the YOLOv5-nano object detection (OD) model. To account for tissue continuity, we propose using the global scores (probability vector) when the model is applied to the entire sagittal sequence. The second module of the proposed pipeline constitutes the MINIROCKET timeseries classification model for determining whether a knee has an ACL tear. To better evaluate the generalization capabilities of our approach, we also carry out cross-database testing involving two public databases (KneeMRI and MRNet) and a validation-only database from University General Hospital of Larissa, Greece. Our method consistently outperformed (p-value<0.05) the state-of-the-art (SOTA) approaches on the KneeMRI dataset and achieved better accuracy and sensitivity on the MRNet dataset. It also generalized remarkably good, especially when the model had been trained on KneeMRI. The presented framework generated at least 2.1 times less carbon emissions and consumed at least 2.6 times less energy, when compared with SOTA. The integration of aleatoric semantic uncertainty-based scores into a novelty detection framework, when combined with the use of lightweight OD and timeseries classification models, have the potential to revolutionize the MRI-based injury detection by setting a new precedent in diagnostic precision, speed and environmental sustainability. Our resource-efficient framework offers potential for widespread application.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"117 ","pages":"Article 102424"},"PeriodicalIF":5.4,"publicationDate":"2024-08-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0895611124001010/pdfft?md5=08030d7bf6b1d50cc1742401e8b8a65d&pid=1-s2.0-S0895611124001010-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142143465","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-28DOI: 10.1016/j.compmedimag.2024.102425
Yutian Zhong , Zhenyang Liu , Xiaoming Zhang , Zhaoyong Liang , Wufan Chen , Cuixia Dai , Li Qi
Photoacoustic tomography (PAT) is a powerful imaging modality for visualizing tissue physiology and exogenous contrast agents. However, PAT faces challenges in visualizing deep-seated vascular structures due to light scattering, absorption, and reduced signal intensity with depth. Optical coherence tomography angiography (OCTA) offers high-contrast visualization of vasculature networks, yet its imaging depth is limited to a millimeter scale. Herein, we propose OCPA-Net, a novel unsupervised deep learning method that utilizes the rich vascular feature of OCTA to enhance PAT images. Trained on unpaired OCTA and PAT images, OCPA-Net incorporates a vessel-aware attention module to enhance deep-seated vessel details captured from OCTA. It leverages a domain-adversarial loss function to enforce structural consistency and a novel identity invariant loss to mitigate excessive image content generation. We validate the structural fidelity of OCPA-Net on simulation experiments, and then demonstrate its vascular enhancement performance on in vivo imaging experiments of tumor-bearing mice and contrast-enhanced pregnant mice. The results show the promise of our method for comprehensive vessel-related image analysis in preclinical research applications.
光声断层成像(PAT)是一种强大的成像模式,可用于观察组织生理学和外源性造影剂。然而,由于光散射、吸收和信号强度随深度降低等原因,PAT 在观察深层血管结构方面面临挑战。光学相干断层血管成像(OCTA)可提供高对比度的血管网络可视化,但其成像深度仅限于毫米级。在此,我们提出了一种新颖的无监督深度学习方法 OCPA-Net,利用 OCTA 丰富的血管特征来增强 PAT 图像。OCPA-Net 在未配对的 OCTA 和 PAT 图像上进行训练,结合了血管感知注意模块,以增强从 OCTA 捕捉到的深层血管细节。它利用领域对抗损失函数来执行结构一致性,并利用新颖的身份不变损失来减少过多图像内容的生成。我们在模拟实验中验证了 OCPA-Net 的结构保真度,然后在肿瘤小鼠和对比度增强怀孕小鼠的体内成像实验中证明了它的血管增强性能。结果表明,我们的方法有望在临床前研究应用中进行全面的血管相关图像分析。
{"title":"Unsupervised adversarial neural network for enhancing vasculature in photoacoustic tomography images using optical coherence tomography angiography","authors":"Yutian Zhong , Zhenyang Liu , Xiaoming Zhang , Zhaoyong Liang , Wufan Chen , Cuixia Dai , Li Qi","doi":"10.1016/j.compmedimag.2024.102425","DOIUrl":"10.1016/j.compmedimag.2024.102425","url":null,"abstract":"<div><p>Photoacoustic tomography (PAT) is a powerful imaging modality for visualizing tissue physiology and exogenous contrast agents. However, PAT faces challenges in visualizing deep-seated vascular structures due to light scattering, absorption, and reduced signal intensity with depth. Optical coherence tomography angiography (OCTA) offers high-contrast visualization of vasculature networks, yet its imaging depth is limited to a millimeter scale. Herein, we propose OCPA-Net, a novel unsupervised deep learning method that utilizes the rich vascular feature of OCTA to enhance PAT images. Trained on unpaired OCTA and PAT images, OCPA-Net incorporates a vessel-aware attention module to enhance deep-seated vessel details captured from OCTA. It leverages a domain-adversarial loss function to enforce structural consistency and a novel identity invariant loss to mitigate excessive image content generation. We validate the structural fidelity of OCPA-Net on simulation experiments, and then demonstrate its vascular enhancement performance on <em>in vivo</em> imaging experiments of tumor-bearing mice and contrast-enhanced pregnant mice. The results show the promise of our method for comprehensive vessel-related image analysis in preclinical research applications.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"117 ","pages":"Article 102425"},"PeriodicalIF":5.4,"publicationDate":"2024-08-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142099452","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-28DOI: 10.1016/j.compmedimag.2024.102427
Jian Qin , Yongjun He , Yiqin Liang , Lanlan Kang , Jing Zhao , Bo Ding
Automated cervical cancer screening through computer-assisted diagnosis has shown considerable potential to improve screening accessibility and reduce associated costs and errors. However, classification performance on whole slide images (WSIs) remains suboptimal due to patient-specific variations. To improve the precision of the screening, pathologists not only analyze the characteristics of suspected abnormal cells, but also compare them with normal cells. Motivated by this practice, we propose a novel cervical cell comparative learning method that leverages pathologist knowledge to learn the differences between normal and suspected abnormal cells within the same WSI. Our method employs two pre-trained YOLOX models to detect suspected abnormal and normal cells in a given WSI. A self-supervised model then extracts features for the detected cells. Subsequently, a tailored Transformer encoder fuses the cell features to obtain WSI instance embeddings. Finally, attention-based multi-instance learning is applied to achieve classification. The experimental results show an AUC of 0.9319 for our proposed method. Moreover, the method achieved professional pathologist-level performance, indicating its potential for clinical applications.
通过计算机辅助诊断进行宫颈癌自动筛查在提高筛查的可及性、降低相关成本和减少误差方面具有相当大的潜力。然而,由于患者的个体差异,整张切片图像(WSI)的分类性能仍不理想。为了提高筛查的精确度,病理学家不仅要分析疑似异常细胞的特征,还要将它们与正常细胞进行比较。受这种做法的启发,我们提出了一种新颖的宫颈细胞比较学习方法,利用病理学家的知识来学习同一 WSI 中正常细胞和疑似异常细胞之间的差异。我们的方法采用两个预先训练好的 YOLOX 模型来检测给定 WSI 中的疑似异常细胞和正常细胞。然后,一个自监督模型提取检测到的细胞的特征。随后,量身定制的 Transformer 编码器会融合细胞特征,从而获得 WSI 实例嵌入。最后,应用基于注意力的多实例学习来实现分类。实验结果显示,我们提出的方法的 AUC 为 0.9319。此外,该方法还达到了专业病理学家的水平,这表明它具有临床应用的潜力。
{"title":"Cell comparative learning: A cervical cytopathology whole slide image classification method using normal and abnormal cells","authors":"Jian Qin , Yongjun He , Yiqin Liang , Lanlan Kang , Jing Zhao , Bo Ding","doi":"10.1016/j.compmedimag.2024.102427","DOIUrl":"10.1016/j.compmedimag.2024.102427","url":null,"abstract":"<div><p>Automated cervical cancer screening through computer-assisted diagnosis has shown considerable potential to improve screening accessibility and reduce associated costs and errors. However, classification performance on whole slide images (WSIs) remains suboptimal due to patient-specific variations. To improve the precision of the screening, pathologists not only analyze the characteristics of suspected abnormal cells, but also compare them with normal cells. Motivated by this practice, we propose a novel cervical cell comparative learning method that leverages pathologist knowledge to learn the differences between normal and suspected abnormal cells within the same WSI. Our method employs two pre-trained YOLOX models to detect suspected abnormal and normal cells in a given WSI. A self-supervised model then extracts features for the detected cells. Subsequently, a tailored Transformer encoder fuses the cell features to obtain WSI instance embeddings. Finally, attention-based multi-instance learning is applied to achieve classification. The experimental results show an AUC of 0.9319 for our proposed method. Moreover, the method achieved professional pathologist-level performance, indicating its potential for clinical applications.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"117 ","pages":"Article 102427"},"PeriodicalIF":5.4,"publicationDate":"2024-08-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142099356","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-07DOI: 10.1016/j.compmedimag.2024.102422
Jianfeng Zhao , Shuo Li
Reliability learning and interpretable decision-making are crucial for multi-modality medical image segmentation. Although many works have attempted multi-modality medical image segmentation, they rarely explore how much reliability is provided by each modality for segmentation. Moreover, the existing approach of decision-making such as the function lacks the interpretability for multi-modality fusion. In this study, we proposed a novel approach named contextual discounted evidential network (CDE-Net) for reliability learning and interpretable decision-making under multi-modality medical image segmentation. Specifically, the CDE-Net first models the semantic evidence by uncertainty measurement using the proposed evidential decision-making module. Then, it leverages the contextual discounted fusion layer to learn the reliability provided by each modality. Finally, a multi-level loss function is deployed for the optimization of evidence modeling and reliability learning. Moreover, this study elaborates on the framework interpretability by discussing the consistency between pixel attribution maps and the learned reliability coefficients. Extensive experiments are conducted on both multi-modality brain and liver datasets. The CDE-Net gains high performance with an average Dice score of 0.914 for brain tumor segmentation and 0.913 for liver tumor segmentation, which proves CDE-Net has great potential to facilitate the interpretation of artificial intelligence-based multi-modality medical image fusion.
{"title":"Evidence modeling for reliability learning and interpretable decision-making under multi-modality medical image segmentation","authors":"Jianfeng Zhao , Shuo Li","doi":"10.1016/j.compmedimag.2024.102422","DOIUrl":"10.1016/j.compmedimag.2024.102422","url":null,"abstract":"<div><p>Reliability learning and interpretable decision-making are crucial for multi-modality medical image segmentation. Although many works have attempted multi-modality medical image segmentation, they rarely explore how much reliability is provided by each modality for segmentation. Moreover, the existing approach of decision-making such as the <span><math><mrow><mi>s</mi><mi>o</mi><mi>f</mi><mi>t</mi><mi>m</mi><mi>a</mi><mi>x</mi></mrow></math></span> function lacks the interpretability for multi-modality fusion. In this study, we proposed a novel approach named contextual discounted evidential network (CDE-Net) for reliability learning and interpretable decision-making under multi-modality medical image segmentation. Specifically, the CDE-Net first models the semantic evidence by uncertainty measurement using the proposed evidential decision-making module. Then, it leverages the contextual discounted fusion layer to learn the reliability provided by each modality. Finally, a multi-level loss function is deployed for the optimization of evidence modeling and reliability learning. Moreover, this study elaborates on the framework interpretability by discussing the consistency between pixel attribution maps and the learned reliability coefficients. Extensive experiments are conducted on both multi-modality brain and liver datasets. The CDE-Net gains high performance with an average Dice score of 0.914 for brain tumor segmentation and 0.913 for liver tumor segmentation, which proves CDE-Net has great potential to facilitate the interpretation of artificial intelligence-based multi-modality medical image fusion.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"116 ","pages":"Article 102422"},"PeriodicalIF":5.4,"publicationDate":"2024-08-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141908211","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-26DOI: 10.1016/j.compmedimag.2024.102421
Lei Chai , Shuangqian Xue , Daodao Tang , Jixin Liu , Ning Sun , Xiujuan Liu
Intracranial aneurysm (IA) is a prevalent disease that poses a significant threat to human health. The use of computed tomography angiography (CTA) as a diagnostic tool for IAs remains time-consuming and challenging. Deep neural networks (DNNs) have made significant advancements in the field of medical image segmentation. Nevertheless, training large-scale DNNs demands substantial quantities of high-quality labeled data, making the annotation of numerous brain CTA scans a challenging endeavor. To address these challenges and effectively develop a robust IAs segmentation model from a large amount of unlabeled training data, we propose a triple learning framework (TLF). The framework primarily consists of three learning paradigms: pseudo-supervised learning, contrastive learning, and confident learning. This paper introduces an enhanced mean teacher model and voxel-selective strategy to conduct pseudo-supervised learning on unreliable labeled training data. Concurrently, we construct the positive and negative training pairs within the high-level semantic feature space to improve the overall learning efficiency of the TLF through contrastive learning. In addition, a multi-scale confident learning is proposed to correct unreliable labels, which enables the acquisition of broader local structural information instead of relying on individual voxels. To evaluate the effectiveness of our method, we conducted extensive experiments on a self-built database of hundreds of cases of brain CTA scans with IAs. Experimental results demonstrate that our method can effectively learn a robust CTA scan-based IAs segmentation model using unreliable labeled data, outperforming state-of-the-art methods in terms of segmentation accuracy. Codes are released at https://github.com/XueShuangqian/TLF.
{"title":"TLF: Triple learning framework for intracranial aneurysms segmentation from unreliable labeled CTA scans","authors":"Lei Chai , Shuangqian Xue , Daodao Tang , Jixin Liu , Ning Sun , Xiujuan Liu","doi":"10.1016/j.compmedimag.2024.102421","DOIUrl":"10.1016/j.compmedimag.2024.102421","url":null,"abstract":"<div><p>Intracranial aneurysm (IA) is a prevalent disease that poses a significant threat to human health. The use of computed tomography angiography (CTA) as a diagnostic tool for IAs remains time-consuming and challenging. Deep neural networks (DNNs) have made significant advancements in the field of medical image segmentation. Nevertheless, training large-scale DNNs demands substantial quantities of high-quality labeled data, making the annotation of numerous brain CTA scans a challenging endeavor. To address these challenges and effectively develop a robust IAs segmentation model from a large amount of unlabeled training data, we propose a triple learning framework (TLF). The framework primarily consists of three learning paradigms: pseudo-supervised learning, contrastive learning, and confident learning. This paper introduces an enhanced mean teacher model and voxel-selective strategy to conduct pseudo-supervised learning on unreliable labeled training data. Concurrently, we construct the positive and negative training pairs within the high-level semantic feature space to improve the overall learning efficiency of the TLF through contrastive learning. In addition, a multi-scale confident learning is proposed to correct unreliable labels, which enables the acquisition of broader local structural information instead of relying on individual voxels. To evaluate the effectiveness of our method, we conducted extensive experiments on a self-built database of hundreds of cases of brain CTA scans with IAs. Experimental results demonstrate that our method can effectively learn a robust CTA scan-based IAs segmentation model using unreliable labeled data, outperforming state-of-the-art methods in terms of segmentation accuracy. Codes are released at https://github.com/XueShuangqian/TLF.</p></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"116 ","pages":"Article 102421"},"PeriodicalIF":5.4,"publicationDate":"2024-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141842424","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}