Pub Date : 2024-08-01Epub Date: 2024-02-21DOI: 10.1007/s10278-024-01011-2
Dina A Ragab, Salema Fayed, Noha Ghatwary
Worldwide, the COVID-19 epidemic, which started in 2019, has resulted in millions of deaths. The medical research community has widely used computer analysis of medical data during the pandemic, specifically deep learning models. Deploying models on devices with constrained resources is a significant challenge due to the increased storage demands associated with larger deep learning models. Accordingly, in this paper, we propose a novel compression strategy that compresses deep features with a compression ratio of 10 to 90% to accurately classify the COVID-19 and non-COVID-19 computed tomography scans. Additionally, we extensively validated the compression using various available deep learning methods to extract the most suitable features from different models. Finally, the suggested DeepCSFusion model compresses the extracted features and applies fusion to achieve the highest classification accuracy with fewer features. The proposed DeepCSFusion model was validated on the publicly available dataset "SARS-CoV-2 CT" scans composed of 1252 CT. This study demonstrates that the proposed DeepCSFusion reduced the computational time with an overall accuracy of 99.3%. Also, it outperforms state-of-the-art pipelines in terms of various classification measures.
{"title":"DeepCSFusion: Deep Compressive Sensing Fusion for Efficient COVID-19 Classification.","authors":"Dina A Ragab, Salema Fayed, Noha Ghatwary","doi":"10.1007/s10278-024-01011-2","DOIUrl":"10.1007/s10278-024-01011-2","url":null,"abstract":"<p><p>Worldwide, the COVID-19 epidemic, which started in 2019, has resulted in millions of deaths. The medical research community has widely used computer analysis of medical data during the pandemic, specifically deep learning models. Deploying models on devices with constrained resources is a significant challenge due to the increased storage demands associated with larger deep learning models. Accordingly, in this paper, we propose a novel compression strategy that compresses deep features with a compression ratio of 10 to 90% to accurately classify the COVID-19 and non-COVID-19 computed tomography scans. Additionally, we extensively validated the compression using various available deep learning methods to extract the most suitable features from different models. Finally, the suggested DeepCSFusion model compresses the extracted features and applies fusion to achieve the highest classification accuracy with fewer features. The proposed DeepCSFusion model was validated on the publicly available dataset \"SARS-CoV-2 CT\" scans composed of 1252 CT. This study demonstrates that the proposed DeepCSFusion reduced the computational time with an overall accuracy of 99.3%. Also, it outperforms state-of-the-art pipelines in terms of various classification measures.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11300776/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139914401","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Orthodontically induced external root resorption (OIERR) is a common complication of orthodontic treatments. Accurate OIERR grading is crucial for clinical intervention. This study aimed to evaluate six deep convolutional neural networks (CNNs) for performing OIERR grading on tooth slices to construct an automatic grading system for OIERR. A total of 2146 tooth slices of different OIERR grades were collected and preprocessed. Six pre-trained CNNs (EfficientNet-B1, EfficientNet-B2, EfficientNet-B3, EfficientNet-B4, EfficientNet-B5, and MobileNet-V3) were trained and validated on the pre-processed images based on four different cross-validation methods. The performances of the CNNs on a test set were evaluated and compared with those of orthodontists. The gradient-weighted class activation mapping (Grad-CAM) technique was used to explore the area of maximum impact on the model decisions in the tooth slices. The six CNN models performed remarkably well in OIERR grading, with a mean accuracy of 0.92, surpassing that of the orthodontists (mean accuracy of 0.82). EfficientNet-B4 trained with fivefold cross-validation emerged as the final OIERR grading system, with a high accuracy of 0.94. Grad-CAM revealed that the apical region had the greatest effect on the OIERR grading system. The six CNNs demonstrated excellent OIERR grading and outperformed orthodontists. The proposed OIERR grading system holds potential as a reliable diagnostic support for orthodontists in clinical practice.
{"title":"An Automatic Grading System for Orthodontically Induced External Root Resorption Based on Deep Convolutional Neural Network.","authors":"Shuxi Xu, Houli Peng, Lanxin Yang, Wenjie Zhong, Xiang Gao, Jinlin Song","doi":"10.1007/s10278-024-01045-6","DOIUrl":"10.1007/s10278-024-01045-6","url":null,"abstract":"<p><p>Orthodontically induced external root resorption (OIERR) is a common complication of orthodontic treatments. Accurate OIERR grading is crucial for clinical intervention. This study aimed to evaluate six deep convolutional neural networks (CNNs) for performing OIERR grading on tooth slices to construct an automatic grading system for OIERR. A total of 2146 tooth slices of different OIERR grades were collected and preprocessed. Six pre-trained CNNs (EfficientNet-B1, EfficientNet-B2, EfficientNet-B3, EfficientNet-B4, EfficientNet-B5, and MobileNet-V3) were trained and validated on the pre-processed images based on four different cross-validation methods. The performances of the CNNs on a test set were evaluated and compared with those of orthodontists. The gradient-weighted class activation mapping (Grad-CAM) technique was used to explore the area of maximum impact on the model decisions in the tooth slices. The six CNN models performed remarkably well in OIERR grading, with a mean accuracy of 0.92, surpassing that of the orthodontists (mean accuracy of 0.82). EfficientNet-B4 trained with fivefold cross-validation emerged as the final OIERR grading system, with a high accuracy of 0.94. Grad-CAM revealed that the apical region had the greatest effect on the OIERR grading system. The six CNNs demonstrated excellent OIERR grading and outperformed orthodontists. The proposed OIERR grading system holds potential as a reliable diagnostic support for orthodontists in clinical practice.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11300848/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139935079","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The goal of this study was to evaluate the performance of a convolutional neural network (CNN) with preoperative MRI and clinical factors in predicting the treatment response of unresectable hepatocellular carcinoma (HCC) patients receiving hepatic arterial infusion chemotherapy (HAIC). A total of 191 patients with unresectable HCC who underwent HAIC in our hospital between May 2019 and March 2022 were retrospectively recruited. We selected InceptionV4 from three representative CNN models, AlexNet, ResNet, and InceptionV4, according to the cross-entropy loss (CEL). We subsequently developed InceptionV4 to fuse the information from qualified pretreatment MRI data and patient clinical factors. Radiomic information was evaluated based on several constant sequences, including enhanced T1-weighted sequences (with arterial, portal, and delayed phases), T2 FSE sequences, and dual-echo sequences. The performance of InceptionV4 was cross-validated in the training cohort (n = 127) and internally validated in an independent cohort (n = 64), with comparisons against single important clinical factors and radiologists in terms of receiver operating characteristic (ROC) curves. Class activation mapping was used to visualize the InceptionV4 model. The InceptionV4 model achieved an AUC of 0.871 (95% confidence interval [CI] 0.761-0.981) in the cross-validation cohort and an AUC of 0.826 (95% CI 0.682-0.970) in the internal validation cohort; these two models performed better than did the other methods (AUC ranges 0.783-0.873 and 0.708-0.806 for cross- and internal validations, respectively; P < 0.01). The present InceptionV4 model, which integrates radiomic information and clinical factors, helps predict the treatment response of unresectable HCC patients receiving HAIC treatment.
{"title":"Development and Preliminary Validation of a Novel Convolutional Neural Network Model for Predicting Treatment Response in Patients with Unresectable Hepatocellular Carcinoma Receiving Hepatic Arterial Infusion Chemotherapy.","authors":"Bing Quan, Jinghuan Li, Hailin Mi, Miao Li, Wenfeng Liu, Fan Yao, Rongxin Chen, Yan Shan, Pengju Xu, Zhenggang Ren, Xin Yin","doi":"10.1007/s10278-024-01003-2","DOIUrl":"10.1007/s10278-024-01003-2","url":null,"abstract":"<p><p>The goal of this study was to evaluate the performance of a convolutional neural network (CNN) with preoperative MRI and clinical factors in predicting the treatment response of unresectable hepatocellular carcinoma (HCC) patients receiving hepatic arterial infusion chemotherapy (HAIC). A total of 191 patients with unresectable HCC who underwent HAIC in our hospital between May 2019 and March 2022 were retrospectively recruited. We selected InceptionV4 from three representative CNN models, AlexNet, ResNet, and InceptionV4, according to the cross-entropy loss (CEL). We subsequently developed InceptionV4 to fuse the information from qualified pretreatment MRI data and patient clinical factors. Radiomic information was evaluated based on several constant sequences, including enhanced T1-weighted sequences (with arterial, portal, and delayed phases), T2 FSE sequences, and dual-echo sequences. The performance of InceptionV4 was cross-validated in the training cohort (n = 127) and internally validated in an independent cohort (n = 64), with comparisons against single important clinical factors and radiologists in terms of receiver operating characteristic (ROC) curves. Class activation mapping was used to visualize the InceptionV4 model. The InceptionV4 model achieved an AUC of 0.871 (95% confidence interval [CI] 0.761-0.981) in the cross-validation cohort and an AUC of 0.826 (95% CI 0.682-0.970) in the internal validation cohort; these two models performed better than did the other methods (AUC ranges 0.783-0.873 and 0.708-0.806 for cross- and internal validations, respectively; P < 0.01). The present InceptionV4 model, which integrates radiomic information and clinical factors, helps predict the treatment response of unresectable HCC patients receiving HAIC treatment.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11300745/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139935091","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01Epub Date: 2024-03-04DOI: 10.1007/s10278-024-01033-w
H J H Dreesen, C Stroszczynski, M M Lell
Coronary computed tomography angiography (CCTA) is an essential part of the diagnosis of chronic coronary syndrome (CCS) in patients with low-to-intermediate pre-test probability. The minimum technical requirement is 64-row multidetector CT (64-MDCT), which is still frequently used, although it is prone to motion artifacts because of its limited temporal resolution and z-coverage. In this study, we evaluate the potential of a deep-learning-based motion correction algorithm (MCA) to eliminate these motion artifacts. 124 64-MDCT-acquired CCTA examinations with at least minor motion artifacts were included. Images were reconstructed using a conventional reconstruction algorithm (CA) and a MCA. Image quality (IQ), according to a 5-point Likert score, was evaluated per-segment, per-artery, and per-patient and was correlated with potentially disturbing factors (heart rate (HR), intra-cycle HR changes, BMI, age, and sex). Comparison was done by Wilcoxon-Signed-Rank test, and correlation by Spearman's Rho. Per-patient, insufficient IQ decreased by 5.26%, and sufficient IQ increased by 9.66% with MCA. Per-artery, insufficient IQ of the right coronary artery (RCA) decreased by 18.18%, and sufficient IQ increased by 27.27%. Per-segment, insufficient IQ in segments 1 and 2 decreased by 11.51% and 24.78%, respectively, and sufficient IQ increased by 10.62% and 18.58%, respectively. Total artifacts per-artery decreased in the RCA from 3.11 ± 1.65 to 2.26 ± 1.52. HR dependence of RCA IQ decreased to intermediate correlation in images with MCA reconstruction. The applied MCA improves the IQ of 64-MDCT-acquired images and reduces the influence of HR on IQ, increasing 64-MDCT validity in the diagnosis of CCS.
{"title":"Optimizing Coronary Computed Tomography Angiography Using a Novel Deep Learning-Based Algorithm.","authors":"H J H Dreesen, C Stroszczynski, M M Lell","doi":"10.1007/s10278-024-01033-w","DOIUrl":"10.1007/s10278-024-01033-w","url":null,"abstract":"<p><p>Coronary computed tomography angiography (CCTA) is an essential part of the diagnosis of chronic coronary syndrome (CCS) in patients with low-to-intermediate pre-test probability. The minimum technical requirement is 64-row multidetector CT (64-MDCT), which is still frequently used, although it is prone to motion artifacts because of its limited temporal resolution and z-coverage. In this study, we evaluate the potential of a deep-learning-based motion correction algorithm (MCA) to eliminate these motion artifacts. 124 64-MDCT-acquired CCTA examinations with at least minor motion artifacts were included. Images were reconstructed using a conventional reconstruction algorithm (CA) and a MCA. Image quality (IQ), according to a 5-point Likert score, was evaluated per-segment, per-artery, and per-patient and was correlated with potentially disturbing factors (heart rate (HR), intra-cycle HR changes, BMI, age, and sex). Comparison was done by Wilcoxon-Signed-Rank test, and correlation by Spearman's Rho. Per-patient, insufficient IQ decreased by 5.26%, and sufficient IQ increased by 9.66% with MCA. Per-artery, insufficient IQ of the right coronary artery (RCA) decreased by 18.18%, and sufficient IQ increased by 27.27%. Per-segment, insufficient IQ in segments 1 and 2 decreased by 11.51% and 24.78%, respectively, and sufficient IQ increased by 10.62% and 18.58%, respectively. Total artifacts per-artery decreased in the RCA from 3.11 ± 1.65 to 2.26 ± 1.52. HR dependence of RCA IQ decreased to intermediate correlation in images with MCA reconstruction. The applied MCA improves the IQ of 64-MDCT-acquired images and reduces the influence of HR on IQ, increasing 64-MDCT validity in the diagnosis of CCS.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11300758/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140029985","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01Epub Date: 2024-02-26DOI: 10.1007/s10278-024-01041-w
Adriano Barbosa Silva, Alessandro Santana Martins, Thaína Aparecida Azevedo Tosta, Adriano Mota Loyola, Sérgio Vitorino Cardoso, Leandro Alves Neves, Paulo Rogério de Faria, Marcelo Zanchetta do Nascimento
Early diagnosis of potentially malignant disorders, such as oral epithelial dysplasia, is the most reliable way to prevent oral cancer. Computational algorithms have been used as an auxiliary tool to aid specialists in this process. Usually, experiments are performed on private data, making it difficult to reproduce the results. There are several public datasets of histological images, but studies focused on oral dysplasia images use inaccessible datasets. This prevents the improvement of algorithms aimed at this lesion. This study introduces an annotated public dataset of oral epithelial dysplasia tissue images. The dataset includes 456 images acquired from 30 mouse tongues. The images were categorized among the lesion grades, with nuclear structures manually marked by a trained specialist and validated by a pathologist. Also, experiments were carried out in order to illustrate the potential of the proposed dataset in classification and segmentation processes commonly explored in the literature. Convolutional neural network (CNN) models for semantic and instance segmentation were employed on the images, which were pre-processed with stain normalization methods. Then, the segmented and non-segmented images were classified with CNN architectures and machine learning algorithms. The data obtained through these processes is available in the dataset. The segmentation stage showed the F1-score value of 0.83, obtained with the U-Net model using the ResNet-50 as a backbone. At the classification stage, the most expressive result was achieved with the Random Forest method, with an accuracy value of 94.22%. The results show that the segmentation contributed to the classification results, but studies are needed for the improvement of these stages of automated diagnosis. The original, gold standard, normalized, and segmented images are publicly available and may be used for the improvement of clinical applications of CAD methods on oral epithelial dysplasia tissue images.
{"title":"OralEpitheliumDB: A Dataset for Oral Epithelial Dysplasia Image Segmentation and Classification.","authors":"Adriano Barbosa Silva, Alessandro Santana Martins, Thaína Aparecida Azevedo Tosta, Adriano Mota Loyola, Sérgio Vitorino Cardoso, Leandro Alves Neves, Paulo Rogério de Faria, Marcelo Zanchetta do Nascimento","doi":"10.1007/s10278-024-01041-w","DOIUrl":"10.1007/s10278-024-01041-w","url":null,"abstract":"<p><p>Early diagnosis of potentially malignant disorders, such as oral epithelial dysplasia, is the most reliable way to prevent oral cancer. Computational algorithms have been used as an auxiliary tool to aid specialists in this process. Usually, experiments are performed on private data, making it difficult to reproduce the results. There are several public datasets of histological images, but studies focused on oral dysplasia images use inaccessible datasets. This prevents the improvement of algorithms aimed at this lesion. This study introduces an annotated public dataset of oral epithelial dysplasia tissue images. The dataset includes 456 images acquired from 30 mouse tongues. The images were categorized among the lesion grades, with nuclear structures manually marked by a trained specialist and validated by a pathologist. Also, experiments were carried out in order to illustrate the potential of the proposed dataset in classification and segmentation processes commonly explored in the literature. Convolutional neural network (CNN) models for semantic and instance segmentation were employed on the images, which were pre-processed with stain normalization methods. Then, the segmented and non-segmented images were classified with CNN architectures and machine learning algorithms. The data obtained through these processes is available in the dataset. The segmentation stage showed the F1-score value of 0.83, obtained with the U-Net model using the ResNet-50 as a backbone. At the classification stage, the most expressive result was achieved with the Random Forest method, with an accuracy value of 94.22%. The results show that the segmentation contributed to the classification results, but studies are needed for the improvement of these stages of automated diagnosis. The original, gold standard, normalized, and segmented images are publicly available and may be used for the improvement of clinical applications of CAD methods on oral epithelial dysplasia tissue images.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139975360","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01Epub Date: 2024-02-29DOI: 10.1007/s10278-024-01035-8
Selvakanmani S, G Dharani Devi, Rekha V, J Jeyalakshmi
Breast cancer is deadly cancer causing a considerable number of fatalities among women in worldwide. To enhance patient outcomes as well as survival rates, early and accurate detection is crucial. Machine learning techniques, particularly deep learning, have demonstrated impressive success in various image recognition tasks, including breast cancer classification. However, the reliance on large labeled datasets poses challenges in the medical domain due to privacy issues and data silos. This study proposes a novel transfer learning approach integrated into a federated learning framework to solve the limitations of limited labeled data and data privacy in collaborative healthcare settings. For breast cancer classification, the mammography and MRO images were gathered from three different medical centers. Federated learning, an emerging privacy-preserving paradigm, empowers multiple medical institutions to jointly train the global model while maintaining data decentralization. Our proposed methodology capitalizes on the power of pre-trained ResNet, a deep neural network architecture, as a feature extractor. By fine-tuning the higher layers of ResNet using breast cancer datasets from diverse medical centers, we enable the model to learn specialized features relevant to different domains while leveraging the comprehensive image representations acquired from large-scale datasets like ImageNet. To overcome domain shift challenges caused by variations in data distributions across medical centers, we introduce domain adversarial training. The model learns to minimize the domain discrepancy while maximizing classification accuracy, facilitating the acquisition of domain-invariant features. We conducted extensive experiments on diverse breast cancer datasets obtained from multiple medical centers. Comparative analysis was performed to evaluate the proposed approach against traditional standalone training and federated learning without domain adaptation. When compared with traditional models, our proposed model showed a classification accuracy of 98.8% and a computational time of 12.22 s. The results showcase promising enhancements in classification accuracy and model generalization, underscoring the potential of our method in improving breast cancer classification performance while upholding data privacy in a federated healthcare environment.
{"title":"Privacy-Preserving Breast Cancer Classification: A Federated Transfer Learning Approach.","authors":"Selvakanmani S, G Dharani Devi, Rekha V, J Jeyalakshmi","doi":"10.1007/s10278-024-01035-8","DOIUrl":"10.1007/s10278-024-01035-8","url":null,"abstract":"<p><p>Breast cancer is deadly cancer causing a considerable number of fatalities among women in worldwide. To enhance patient outcomes as well as survival rates, early and accurate detection is crucial. Machine learning techniques, particularly deep learning, have demonstrated impressive success in various image recognition tasks, including breast cancer classification. However, the reliance on large labeled datasets poses challenges in the medical domain due to privacy issues and data silos. This study proposes a novel transfer learning approach integrated into a federated learning framework to solve the limitations of limited labeled data and data privacy in collaborative healthcare settings. For breast cancer classification, the mammography and MRO images were gathered from three different medical centers. Federated learning, an emerging privacy-preserving paradigm, empowers multiple medical institutions to jointly train the global model while maintaining data decentralization. Our proposed methodology capitalizes on the power of pre-trained ResNet, a deep neural network architecture, as a feature extractor. By fine-tuning the higher layers of ResNet using breast cancer datasets from diverse medical centers, we enable the model to learn specialized features relevant to different domains while leveraging the comprehensive image representations acquired from large-scale datasets like ImageNet. To overcome domain shift challenges caused by variations in data distributions across medical centers, we introduce domain adversarial training. The model learns to minimize the domain discrepancy while maximizing classification accuracy, facilitating the acquisition of domain-invariant features. We conducted extensive experiments on diverse breast cancer datasets obtained from multiple medical centers. Comparative analysis was performed to evaluate the proposed approach against traditional standalone training and federated learning without domain adaptation. When compared with traditional models, our proposed model showed a classification accuracy of 98.8% and a computational time of 12.22 s. The results showcase promising enhancements in classification accuracy and model generalization, underscoring the potential of our method in improving breast cancer classification performance while upholding data privacy in a federated healthcare environment.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11300768/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139998780","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01Epub Date: 2024-02-27DOI: 10.1007/s10278-024-01055-4
Miguel Molina-Moreno, Iván González-Díaz, Maite Rivera Gorrín, Víctor Burguera Vion, Fernando Díaz-de-María
Ultrasound is a widespread imaging modality, with special application in medical fields such as nephrology. However, automated approaches for ultrasound renal interpretation still pose some challenges: (1) the need for manual supervision by experts at various stages of the system, which prevents its adoption in primary healthcare, and (2) their limited considered taxonomy (e.g., reduced number of pathologies), which makes them unsuitable for training practitioners and providing support to experts. This paper proposes a fully automated computer-aided diagnosis system for ultrasound renal imaging addressing both of these challenges. Our system is based in a multi-task architecture, which is implemented by a three-branched convolutional neural network and is capable of segmenting the kidney and detecting global and local pathologies with no need of human interaction during diagnosis. The integration of different image perspectives at distinct granularities enhanced the proposed diagnosis. We employ a large (1985 images) and demanding ultrasound renal imaging database, publicly released with the system and annotated on the basis of an exhaustive taxonomy of two global and nine local pathologies (including cysts, lithiasis, hydronephrosis, angiomyolipoma), establishing a benchmark for ultrasound renal interpretation. Experiments show that our proposed method outperforms several state-of-the-art methods in both segmentation and diagnosis tasks and leverages the combination of global and local image information to improve the diagnosis. Our results, with a 87.41% of AUC in healthy-pathological diagnosis and 81.90% in multi-pathological diagnosis, support the use of our system as a helpful tool in the healthcare system.
{"title":"URI-CADS: A Fully Automated Computer-Aided Diagnosis System for Ultrasound Renal Imaging.","authors":"Miguel Molina-Moreno, Iván González-Díaz, Maite Rivera Gorrín, Víctor Burguera Vion, Fernando Díaz-de-María","doi":"10.1007/s10278-024-01055-4","DOIUrl":"10.1007/s10278-024-01055-4","url":null,"abstract":"<p><p>Ultrasound is a widespread imaging modality, with special application in medical fields such as nephrology. However, automated approaches for ultrasound renal interpretation still pose some challenges: (1) the need for manual supervision by experts at various stages of the system, which prevents its adoption in primary healthcare, and (2) their limited considered taxonomy (e.g., reduced number of pathologies), which makes them unsuitable for training practitioners and providing support to experts. This paper proposes a fully automated computer-aided diagnosis system for ultrasound renal imaging addressing both of these challenges. Our system is based in a multi-task architecture, which is implemented by a three-branched convolutional neural network and is capable of segmenting the kidney and detecting global and local pathologies with no need of human interaction during diagnosis. The integration of different image perspectives at distinct granularities enhanced the proposed diagnosis. We employ a large (1985 images) and demanding ultrasound renal imaging database, publicly released with the system and annotated on the basis of an exhaustive taxonomy of two global and nine local pathologies (including cysts, lithiasis, hydronephrosis, angiomyolipoma), establishing a benchmark for ultrasound renal interpretation. Experiments show that our proposed method outperforms several state-of-the-art methods in both segmentation and diagnosis tasks and leverages the combination of global and local image information to improve the diagnosis. Our results, with a 87.41% of AUC in healthy-pathological diagnosis and 81.90% in multi-pathological diagnosis, support the use of our system as a helpful tool in the healthcare system.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11300425/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139984980","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01Epub Date: 2024-02-29DOI: 10.1007/s10278-023-00934-6
Jianning Chi, Zhiyi Sun, Shuyu Tian, Huan Wang, Siqi Wang
Low-dose computer tomography (LDCT) has been widely used in medical diagnosis. Various denoising methods have been presented to remove noise in LDCT scans. However, existing methods cannot achieve satisfactory results due to the difficulties in (1) distinguishing the characteristics of structures, textures, and noise confused in the image domain, and (2) representing local details and global semantics in the hierarchical features. In this paper, we propose a novel denoising method consisting of (1) a 2D dual-domain restoration framework to reconstruct noise-free structure and texture signals separately, and (2) a 3D multi-depth reinforcement U-Net model to further recover image details with enhanced hierarchical features. In the 2D dual-domain restoration framework, the convolutional neural networks are adopted in both the image domain where the image structures are well preserved through the spatial continuity, and the sinogram domain where the textures and noise are separately represented by different wavelet coefficients and processed adaptively. In the 3D multi-depth reinforcement U-Net model, the hierarchical features from the 3D U-Net are enhanced by the cross-resolution attention module (CRAM) and dual-branch graph convolution module (DBGCM). The CRAM preserves local details by integrating adjacent low-level features with different resolutions, while the DBGCM enhances global semantics by building graphs for high-level features in intra-feature and inter-feature dimensions. Experimental results on the LUNA16 dataset and 2016 NIH-AAPM-Mayo Clinic LDCT Grand Challenge dataset illustrate the proposed method outperforms the state-of-the-art methods on removing noise from LDCT images with clear structures and textures, proving its potential in clinical practice.
{"title":"A Hybrid Framework of Dual-Domain Signal Restoration and Multi-depth Feature Reinforcement for Low-Dose Lung CT Denoising.","authors":"Jianning Chi, Zhiyi Sun, Shuyu Tian, Huan Wang, Siqi Wang","doi":"10.1007/s10278-023-00934-6","DOIUrl":"10.1007/s10278-023-00934-6","url":null,"abstract":"<p><p>Low-dose computer tomography (LDCT) has been widely used in medical diagnosis. Various denoising methods have been presented to remove noise in LDCT scans. However, existing methods cannot achieve satisfactory results due to the difficulties in (1) distinguishing the characteristics of structures, textures, and noise confused in the image domain, and (2) representing local details and global semantics in the hierarchical features. In this paper, we propose a novel denoising method consisting of (1) a 2D dual-domain restoration framework to reconstruct noise-free structure and texture signals separately, and (2) a 3D multi-depth reinforcement U-Net model to further recover image details with enhanced hierarchical features. In the 2D dual-domain restoration framework, the convolutional neural networks are adopted in both the image domain where the image structures are well preserved through the spatial continuity, and the sinogram domain where the textures and noise are separately represented by different wavelet coefficients and processed adaptively. In the 3D multi-depth reinforcement U-Net model, the hierarchical features from the 3D U-Net are enhanced by the cross-resolution attention module (CRAM) and dual-branch graph convolution module (DBGCM). The CRAM preserves local details by integrating adjacent low-level features with different resolutions, while the DBGCM enhances global semantics by building graphs for high-level features in intra-feature and inter-feature dimensions. Experimental results on the LUNA16 dataset and 2016 NIH-AAPM-Mayo Clinic LDCT Grand Challenge dataset illustrate the proposed method outperforms the state-of-the-art methods on removing noise from LDCT images with clear structures and textures, proving its potential in clinical practice.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11300419/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139998751","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01Epub Date: 2024-02-29DOI: 10.1007/s10278-023-00923-9
Kepei Xu, Meiqi Hua, Ting Mai, Xiaojing Ren, Xiaozheng Fang, Chunjie Wang, Min Ge, Hua Qian, Maosheng Xu, Ruixin Zhang
This study aims to develop an MRI-based radiomics model to assess the likelihood of recurrence in luminal B breast cancer. The study analyzed medical images and clinical data from 244 patients with luminal B breast cancer. Of 244 patients, 35 had experienced recurrence and 209 had not. The patients were randomly divided into the training set (51.5 ± 12.5 years old; n = 171) and the test set (51.7 ± 11.3 years old; n = 73) in a ratio of 7:3. The study employed univariate and multivariate Cox regression along with the least absolute shrinkage and selection operator (LASSO) regression methods to select radiomics features and calculate a risk score. A combined model was constructed by integrating the risk score with the clinical and pathological characteristics. The study identified two radiomics features (GLSZM and GLRLM) from DCE-MRI that were used to calculate a risk score. The AUCs were 0.860 and 0.868 in the training set and 0.816 and 0.714 in the testing set for 3- and 5-year recurrence risk, respectively. The combined model incorporating the risk score, pN, and endocrine therapy showed improved predictive power, with AUCs of 0.857 and 0.912 in the training set and 0.943 and 0.945 in the testing set for 3- and 5-year recurrence risk, respectively. The calibration curve of the combined model showed good consistency between predicted and measured values. Our study developed an MRI-based radiomics model that integrates clinical and radiomics features to assess the likelihood of recurrence in luminal B breast cancer. The model shows promise for improving clinical risk stratification and treatment decision-making.
{"title":"A Multiparametric MRI-based Radiomics Model for Stratifying Postoperative Recurrence in Luminal B Breast Cancer.","authors":"Kepei Xu, Meiqi Hua, Ting Mai, Xiaojing Ren, Xiaozheng Fang, Chunjie Wang, Min Ge, Hua Qian, Maosheng Xu, Ruixin Zhang","doi":"10.1007/s10278-023-00923-9","DOIUrl":"10.1007/s10278-023-00923-9","url":null,"abstract":"<p><p>This study aims to develop an MRI-based radiomics model to assess the likelihood of recurrence in luminal B breast cancer. The study analyzed medical images and clinical data from 244 patients with luminal B breast cancer. Of 244 patients, 35 had experienced recurrence and 209 had not. The patients were randomly divided into the training set (51.5 ± 12.5 years old; n = 171) and the test set (51.7 ± 11.3 years old; n = 73) in a ratio of 7:3. The study employed univariate and multivariate Cox regression along with the least absolute shrinkage and selection operator (LASSO) regression methods to select radiomics features and calculate a risk score. A combined model was constructed by integrating the risk score with the clinical and pathological characteristics. The study identified two radiomics features (GLSZM and GLRLM) from DCE-MRI that were used to calculate a risk score. The AUCs were 0.860 and 0.868 in the training set and 0.816 and 0.714 in the testing set for 3- and 5-year recurrence risk, respectively. The combined model incorporating the risk score, pN, and endocrine therapy showed improved predictive power, with AUCs of 0.857 and 0.912 in the training set and 0.943 and 0.945 in the testing set for 3- and 5-year recurrence risk, respectively. The calibration curve of the combined model showed good consistency between predicted and measured values. Our study developed an MRI-based radiomics model that integrates clinical and radiomics features to assess the likelihood of recurrence in luminal B breast cancer. The model shows promise for improving clinical risk stratification and treatment decision-making.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11300413/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139998752","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-01Epub Date: 2024-02-21DOI: 10.1007/s10278-024-01036-7
Ronghui Tian, Guoxiu Lu, Nannan Zhao, Wei Qian, He Ma, Wei Yang
The purpose of this study was to fuse conventional radiomic and deep features from digital breast tomosynthesis craniocaudal projection (DBT-CC) and ultrasound (US) images to establish a multimodal benign-malignant classification model and evaluate its clinical value. Data were obtained from a total of 487 patients at three centers, each of whom underwent DBT-CC and US examinations. A total of 322 patients from dataset 1 were used to construct the model, while 165 patients from datasets 2 and 3 formed the prospective testing cohort. Two radiologists with 10-20 years of work experience and three sonographers with 12-20 years of work experience semiautomatically segmented the lesions using ITK-SNAP software while considering the surrounding tissue. For the experiments, we extracted conventional radiomic and deep features from tumors from DBT-CCs and US images using PyRadiomics and Inception-v3. Additionally, we extracted conventional radiomic features from four peritumoral layers around the tumors via DBT-CC and US images. Features were fused separately from the intratumoral and peritumoral regions. For the models, we tested the SVM, KNN, decision tree, RF, XGBoost, and LightGBM classifiers. Early fusion and late fusion (ensemble and stacking) strategies were employed for feature fusion. Using the SVM classifier, stacking fusion of deep features and three peritumoral radiomic features from tumors in DBT-CC and US images achieved the optimal performance, with an accuracy and AUC of 0.953 and 0.959 [CI: 0.886-0.996], a sensitivity and specificity of 0.952 [CI: 0.888-0.992] and 0.955 [0.868-0.985], and a precision of 0.976. The experimental results indicate that the fusion model of deep features and peritumoral radiomic features from tumors in DBT-CC and US images shows promise in differentiating benign and malignant breast tumors.
{"title":"Constructing the Optimal Classification Model for Benign and Malignant Breast Tumors Based on Multifeature Analysis from Multimodal Images.","authors":"Ronghui Tian, Guoxiu Lu, Nannan Zhao, Wei Qian, He Ma, Wei Yang","doi":"10.1007/s10278-024-01036-7","DOIUrl":"10.1007/s10278-024-01036-7","url":null,"abstract":"<p><p>The purpose of this study was to fuse conventional radiomic and deep features from digital breast tomosynthesis craniocaudal projection (DBT-CC) and ultrasound (US) images to establish a multimodal benign-malignant classification model and evaluate its clinical value. Data were obtained from a total of 487 patients at three centers, each of whom underwent DBT-CC and US examinations. A total of 322 patients from dataset 1 were used to construct the model, while 165 patients from datasets 2 and 3 formed the prospective testing cohort. Two radiologists with 10-20 years of work experience and three sonographers with 12-20 years of work experience semiautomatically segmented the lesions using ITK-SNAP software while considering the surrounding tissue. For the experiments, we extracted conventional radiomic and deep features from tumors from DBT-CCs and US images using PyRadiomics and Inception-v3. Additionally, we extracted conventional radiomic features from four peritumoral layers around the tumors via DBT-CC and US images. Features were fused separately from the intratumoral and peritumoral regions. For the models, we tested the SVM, KNN, decision tree, RF, XGBoost, and LightGBM classifiers. Early fusion and late fusion (ensemble and stacking) strategies were employed for feature fusion. Using the SVM classifier, stacking fusion of deep features and three peritumoral radiomic features from tumors in DBT-CC and US images achieved the optimal performance, with an accuracy and AUC of 0.953 and 0.959 [CI: 0.886-0.996], a sensitivity and specificity of 0.952 [CI: 0.888-0.992] and 0.955 [0.868-0.985], and a precision of 0.976. The experimental results indicate that the fusion model of deep features and peritumoral radiomic features from tumors in DBT-CC and US images shows promise in differentiating benign and malignant breast tumors.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11300407/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139914400","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}