Pub Date : 2024-01-12DOI: 10.1007/s10278-023-00954-2
Evgin Goceri
Accurate and early detection of precursor adenomatous polyps and their removal at the early stage can significantly decrease the mortality rate and the occurrence of the disease since most colorectal cancer evolve from adenomatous polyps. However, accurate detection and segmentation of the polyps by doctors are difficult mainly these factors: (i) quality of the screening of the polyps with colonoscopy depends on the imaging quality and the experience of the doctors; (ii) visual inspection by doctors is time-consuming, burdensome, and tiring; (iii) prolonged visual inspections can lead to polyps being missed even when the physician is experienced. To overcome these problems, computer-aided methods have been proposed. However, they have some disadvantages or limitations. Therefore, in this work, a new architecture based on residual transformer layers has been designed and used for polyp segmentation. In the proposed segmentation, both high-level semantic features and low-level spatial features have been utilized. Also, a novel hybrid loss function has been proposed. The loss function designed with focal Tversky loss, binary cross-entropy, and Jaccard index reduces image-wise and pixel-wise differences as well as improves regional consistencies. Experimental works have indicated the effectiveness of the proposed approach in terms of dice similarity (0.9048), recall (0.9041), precision (0.9057), and F2 score (0.8993). Comparisons with the state-of-the-art methods have shown its better performance.
{"title":"Polyp Segmentation Using a Hybrid Vision Transformer and a Hybrid Loss Function","authors":"Evgin Goceri","doi":"10.1007/s10278-023-00954-2","DOIUrl":"https://doi.org/10.1007/s10278-023-00954-2","url":null,"abstract":"<p>Accurate and early detection of precursor adenomatous polyps and their removal at the early stage can significantly decrease the mortality rate and the occurrence of the disease since most colorectal cancer evolve from adenomatous polyps. However, accurate detection and segmentation of the polyps by doctors are difficult mainly these factors: (i) quality of the screening of the polyps with colonoscopy depends on the imaging quality and the experience of the doctors; (ii) visual inspection by doctors is time-consuming, burdensome, and tiring; (iii) prolonged visual inspections can lead to polyps being missed even when the physician is experienced. To overcome these problems, computer-aided methods have been proposed. However, they have some disadvantages or limitations. Therefore, in this work, a new architecture based on residual transformer layers has been designed and used for polyp segmentation. In the proposed segmentation, both high-level semantic features and low-level spatial features have been utilized. Also, a novel hybrid loss function has been proposed. The loss function designed with focal Tversky loss, binary cross-entropy, and Jaccard index reduces image-wise and pixel-wise differences as well as improves regional consistencies. Experimental works have indicated the effectiveness of the proposed approach in terms of dice similarity (0.9048), recall (0.9041), precision (0.9057), and F2 score (0.8993). Comparisons with the state-of-the-art methods have shown its better performance.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"28 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139463428","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-12DOI: 10.1007/s10278-023-00953-3
Hamid Alavi, Mehdi Seifi, Mahboubeh Rouhollahei, Mehravar Rafati, Masoud Arabfard
Proximal femur geometry is an important risk factor for diagnosing and predicting hip and femur injuries. Hence, the development of an automated approach for measuring these parameters could help physicians with the early identification of hip and femur ailments. This paper presents a technique that combines the active shape model (ASM) and deep learning methodologies. First, the femur boundary is extracted by a deep learning neural network. Then, the femur’s anatomical landmarks are fitted to the extracted border using the ASM method. Finally, the geometric parameters of the proximal femur, including femur neck axis length (FNAL), femur head diameter (FHD), femur neck width (FNW), shaft width (SW), neck shaft angle (NSA), and alpha angle (AA), are calculated by measuring the distances and angles between the landmarks. The dataset of hip radiographic images consisted of 428 images, with 208 men and 220 women. These images were split into training and testing sets for analysis. The deep learning network and ASM were subsequently trained on the training dataset. In the testing dataset, the automatic measurement of FNAL, FHD, FNW, SW, NSA, and AA parameters resulted in mean errors of 1.19%, 1.46%, 2.28%, 2.43%, 1.95%, and 4.53%, respectively.
{"title":"Development of Local Software for Automatic Measurement of Geometric Parameters in the Proximal Femur Using a Combination of a Deep Learning Approach and an Active Shape Model on X-ray Images","authors":"Hamid Alavi, Mehdi Seifi, Mahboubeh Rouhollahei, Mehravar Rafati, Masoud Arabfard","doi":"10.1007/s10278-023-00953-3","DOIUrl":"https://doi.org/10.1007/s10278-023-00953-3","url":null,"abstract":"<p>Proximal femur geometry is an important risk factor for diagnosing and predicting hip and femur injuries. Hence, the development of an automated approach for measuring these parameters could help physicians with the early identification of hip and femur ailments. This paper presents a technique that combines the active shape model (ASM) and deep learning methodologies. First, the femur boundary is extracted by a deep learning neural network. Then, the femur’s anatomical landmarks are fitted to the extracted border using the ASM method. Finally, the geometric parameters of the proximal femur, including femur neck axis length (FNAL), femur head diameter (FHD), femur neck width (FNW), shaft width (SW), neck shaft angle (NSA), and alpha angle (AA), are calculated by measuring the distances and angles between the landmarks. The dataset of hip radiographic images consisted of 428 images, with 208 men and 220 women. These images were split into training and testing sets for analysis. The deep learning network and ASM were subsequently trained on the training dataset. In the testing dataset, the automatic measurement of FNAL, FHD, FNW, SW, NSA, and AA parameters resulted in mean errors of 1.19%, 1.46%, 2.28%, 2.43%, 1.95%, and 4.53%, respectively.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"30 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139463382","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-12DOI: 10.1007/s10278-023-00900-2
Shamija Sherryl R. M. R., Jaya T.
Kidney tumor segmentation is a difficult task because of the complex spatial and volumetric information present in medical images. Recent advances in deep convolutional neural networks (DCNNs) have improved tumor segmentation accuracy. However, the practical usability of current CNN-based networks is constrained by their high computational complexity. Additionally, these techniques often struggle to make adaptive modifications based on the structure of the tumors, which can lead to blurred edges in segmentation results. A lightweight architecture called the contextual deformable attention and edge-enhanced U-Net (CDA2E-Net) for high-accuracy pixel-level kidney tumor segmentation is proposed to address these challenges. Rather than using complex deep encoders, the approach includes a lightweight depthwise dilated ShuffleNetV2 (LDS-Net) encoder integrated into the CDA2E-Net framework. The proposed method also contains a multiscale attention feature pyramid pooling (MAF2P) module that improves the ability of multiscale features to adapt to various tumor shapes. Finally, an edge-enhanced loss function is introduced to guide the CDA2E-Net to concentrate on tumor edge information. The CDA2E-Net is evaluated on the KiTS19 and KiTS21 datasets, and the results demonstrate its superiority over existing approaches in terms of Hausdorff distance (HD), intersection over union (IoU), and dice coefficient (DSC) metrics.
{"title":"Multi-Scale and Spatial Information Extraction for Kidney Tumor Segmentation: A Contextual Deformable Attention and Edge-Enhanced U-Net","authors":"Shamija Sherryl R. M. R., Jaya T.","doi":"10.1007/s10278-023-00900-2","DOIUrl":"https://doi.org/10.1007/s10278-023-00900-2","url":null,"abstract":"<p>Kidney tumor segmentation is a difficult task because of the complex spatial and volumetric information present in medical images. Recent advances in deep convolutional neural networks (DCNNs) have improved tumor segmentation accuracy. However, the practical usability of current CNN-based networks is constrained by their high computational complexity. Additionally, these techniques often struggle to make adaptive modifications based on the structure of the tumors, which can lead to blurred edges in segmentation results. A lightweight architecture called the contextual deformable attention and edge-enhanced U-Net (CDA2E-Net) for high-accuracy pixel-level kidney tumor segmentation is proposed to address these challenges. Rather than using complex deep encoders, the approach includes a lightweight depthwise dilated ShuffleNetV2 (LDS-Net) encoder integrated into the CDA2E-Net framework. The proposed method also contains a multiscale attention feature pyramid pooling (MAF2P) module that improves the ability of multiscale features to adapt to various tumor shapes. Finally, an edge-enhanced loss function is introduced to guide the CDA2E-Net to concentrate on tumor edge information. The CDA2E-Net is evaluated on the KiTS19 and KiTS21 datasets, and the results demonstrate its superiority over existing approaches in terms of Hausdorff distance (HD), intersection over union (IoU), and dice coefficient (DSC) metrics.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"39 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139463326","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-12DOI: 10.1007/s10278-023-00940-8
Daniel Berthe, Anna Kolb, Abdulrahman Rabi, Thorsten Sellerer, Villseveri Somerkivi, Georg Constantin Feuerriegel, Andreas Philipp Sauter, Felix Meurer, York Hämisch, Tuomas Pantsar, Henrik Lohman, Daniela Pfeiffer, Franz Pfeiffer
Modern photon counting detectors allow the calculation of virtual monoenergetic or material decomposed X-ray images but are not yet used for dental panoramic radiography systems. To assess the diagnostic potential and image quality of photon counting detectors in dental panoramic radiography, ethics approval from the local ethics committee was obtained for this retrospective study. Conventional CT scans of the head and neck region were segmented into bone and soft tissue. The resulting datasets were used to calculate panoramic equivalent thickness bone and soft tissue images by forward projection, using a geometry like that of conventional panoramic radiographic systems. The panoramic equivalent thickness images were utilized to generate synthetic conventional panoramic radiographs and panoramic virtual monoenergetic radiographs at various energies. The conventional, two virtual monoenergetic images at 40 keV and 60 keV, and material-separated bone and soft tissue panoramic equivalent thickness X-ray images simulated from 17 head CTs were evaluated in a reader study involving three experienced radiologists regarding their diagnostic value and image quality. Compared to conventional panoramic radiographs, the material-separated bone panoramic equivalent thickness image exhibits a higher image quality and diagnostic value in assessing the bone structure (left(p<.001right)) and details such as teeth or root canals (left(p<.001right)). Panoramic virtual monoenergetic radiographs do not show a significant advantage over conventional panoramic radiographs. The conducted reader study shows the potential of spectral X-ray imaging for dental panoramic imaging to improve the diagnostic value and image quality.
现代光子计数探测器可以计算虚拟单能或物质分解 X 射线图像,但尚未用于牙科全景放射摄影系统。为了评估光子计数探测器在牙科全景放射摄影中的诊断潜力和图像质量,这项回顾性研究获得了当地伦理委员会的伦理批准。头颈部的常规 CT 扫描分为骨骼和软组织两部分。所得数据集通过正投影法计算出全景等效厚度的骨骼和软组织图像,几何形状与传统的全景放射摄影系统相似。利用全景等效厚度图像生成各种能量下的合成传统全景射线照片和全景虚拟单能射线照片。由三位经验丰富的放射科医生参与的读者研究评估了从 17 个头部 CT 模拟出的传统图像、40 千兆赫和 60 千兆赫的两个虚拟单能量图像以及材料分离的骨和软组织全景等效厚度 X 射线图像的诊断价值和图像质量。与传统的全景X光片相比,材料分离的骨全景等效厚度图像在评估骨结构((left(p<.001right))和牙齿或根管等细节((left(p<.001right))方面显示出更高的图像质量和诊断价值。全景虚拟单能X光片与传统的全景X光片相比并没有显示出明显的优势。这项读者研究显示了光谱 X 射线成像在牙科全景成像中提高诊断价值和图像质量的潜力。
{"title":"Evaluation of Spectral X-Ray Imaging for Panoramic Dental Images Based on a Simulation Framework","authors":"Daniel Berthe, Anna Kolb, Abdulrahman Rabi, Thorsten Sellerer, Villseveri Somerkivi, Georg Constantin Feuerriegel, Andreas Philipp Sauter, Felix Meurer, York Hämisch, Tuomas Pantsar, Henrik Lohman, Daniela Pfeiffer, Franz Pfeiffer","doi":"10.1007/s10278-023-00940-8","DOIUrl":"https://doi.org/10.1007/s10278-023-00940-8","url":null,"abstract":"<p>Modern photon counting detectors allow the calculation of virtual monoenergetic or material decomposed X-ray images but are not yet used for dental panoramic radiography systems. To assess the diagnostic potential and image quality of photon counting detectors in dental panoramic radiography, ethics approval from the local ethics committee was obtained for this retrospective study. Conventional CT scans of the head and neck region were segmented into bone and soft tissue. The resulting datasets were used to calculate panoramic equivalent thickness bone and soft tissue images by forward projection, using a geometry like that of conventional panoramic radiographic systems. The panoramic equivalent thickness images were utilized to generate synthetic conventional panoramic radiographs and panoramic virtual monoenergetic radiographs at various energies. The conventional, two virtual monoenergetic images at 40 keV and 60 keV, and material-separated bone and soft tissue panoramic equivalent thickness X-ray images simulated from 17 head CTs were evaluated in a reader study involving three experienced radiologists regarding their diagnostic value and image quality. Compared to conventional panoramic radiographs, the material-separated bone panoramic equivalent thickness image exhibits a higher image quality and diagnostic value in assessing the bone structure <span>(left(p<.001right))</span> and details such as teeth or root canals <span>(left(p<.001right))</span>. Panoramic virtual monoenergetic radiographs do not show a significant advantage over conventional panoramic radiographs. The conducted reader study shows the potential of spectral X-ray imaging for dental panoramic imaging to improve the diagnostic value and image quality.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"83 2 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139463479","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-12DOI: 10.1007/s10278-023-00941-7
Serkan Savaş
Monkeypox (MPox) is an infectious disease caused by the monkeypox virus, presenting challenges in accurate identification due to its resemblance to other diseases. This study introduces a deep learning-based method to distinguish visually similar diseases, specifically MPox, chickenpox, and measles, addressing the 2022 global MPox outbreak. A two-stage optimization approach was presented in the study. By analyzing pre-trained deep neural networks including 71 models, this study optimizes accuracy through transfer learning, fine-tuning, and ensemble learning techniques. ConvNeXtBase, Large, and XLarge models were identified achieving 97.5% accuracy in the first stage. Afterwards, some selection criteria were followed for the models identified in the first stage for use in ensemble learning technique within the optimization approach. The top-performing ensemble model, EM3 (composed of RegNetX160, ResNetRS101, and ResNet101), attains an AUC of 0.9971 in the second stage. Evaluation on unseen data ensures model robustness and enhances the study’s overall validity and reliability. The design and implementation of the study have been optimized to address the limitations identified in the literature. This approach offers a rapid and highly accurate decision support system for timely MPox diagnosis, reducing human error, manual processes, and enhancing clinic efficiency. It aids in early MPox detection, addresses diverse disease challenges, and informs imaging device software development. The study’s broad implications support global health efforts and showcase artificial intelligence potential in medical informatics for disease identification and diagnosis.
{"title":"Enhancing Disease Classification with Deep Learning: a Two-Stage Optimization Approach for Monkeypox and Similar Skin Lesion Diseases","authors":"Serkan Savaş","doi":"10.1007/s10278-023-00941-7","DOIUrl":"https://doi.org/10.1007/s10278-023-00941-7","url":null,"abstract":"<p>Monkeypox (MPox) is an infectious disease caused by the monkeypox virus, presenting challenges in accurate identification due to its resemblance to other diseases. This study introduces a deep learning-based method to distinguish visually similar diseases, specifically MPox, chickenpox, and measles, addressing the 2022 global MPox outbreak. A two-stage optimization approach was presented in the study. By analyzing pre-trained deep neural networks including 71 models, this study optimizes accuracy through transfer learning, fine-tuning, and ensemble learning techniques. ConvNeXtBase, Large, and XLarge models were identified achieving 97.5% accuracy in the first stage. Afterwards, some selection criteria were followed for the models identified in the first stage for use in ensemble learning technique within the optimization approach. The top-performing ensemble model, EM3 (composed of RegNetX160, ResNetRS101, and ResNet101), attains an AUC of 0.9971 in the second stage. Evaluation on unseen data ensures model robustness and enhances the study’s overall validity and reliability. The design and implementation of the study have been optimized to address the limitations identified in the literature. This approach offers a rapid and highly accurate decision support system for timely MPox diagnosis, reducing human error, manual processes, and enhancing clinic efficiency. It aids in early MPox detection, addresses diverse disease challenges, and informs imaging device software development. The study’s broad implications support global health efforts and showcase artificial intelligence potential in medical informatics for disease identification and diagnosis.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"95 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139463132","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-12DOI: 10.1007/s10278-023-00957-z
Fei Zheng, Ping Yin, Li Yang, Yujian Wang, Wenhan Hao, Qi Hao, Xuzhu Chen, Nan Hong
This paper aims to compare the performance of the classical machine learning (CML) model and the deep learning (DL) model, and to assess the effectiveness of utilizing fusion radiomics from both CML and DL in distinguishing encephalitis from glioma in atypical cases. We analysed the axial FLAIR images of preoperative MRI in 116 patients pathologically confirmed as gliomas and clinically diagnosed with encephalitis. The 3 CML models (logistic regression (LR), support vector machine (SVM) and multi-layer perceptron (MLP)), 3 DL models (DenseNet 121, ResNet 50 and ResNet 18) and a deep learning radiomic (DLR) model were established, respectively. The area under the receiver operating curve (AUC) and sensitivity, specificity, accuracy, negative predictive value (NPV) and positive predictive value (PPV) were calculated for the training and validation sets. In addition, a deep learning radiomic nomogram (DLRN) and a web calculator were designed as a tool to aid clinical decision-making. The best DL model (ResNet50) consistently outperformed the best CML model (LR). The DLR model had the best predictive performance, with AUC, sensitivity, specificity, accuracy, NPV and PPV of 0.879, 0.929, 0.800, 0.875, 0.867 and 0.889 in the validation sets, respectively. Calibration curve of DLR model shows good agreement between prediction and observation, and the decision curve analysis (DCA) indicated that the DLR model had higher overall net benefit than the other two models (ResNet50 and LR). Meanwhile, the DLRN and web calculator can provide dynamic assessments. Machine learning (ML) models have the potential to non-invasively differentiate between encephalitis and glioma in atypical cases. Furthermore, combining DL and CML techniques could enhance the performance of the ML models.
{"title":"MRI-Based Machine Learning Fusion Models to Distinguish Encephalitis and Gliomas","authors":"Fei Zheng, Ping Yin, Li Yang, Yujian Wang, Wenhan Hao, Qi Hao, Xuzhu Chen, Nan Hong","doi":"10.1007/s10278-023-00957-z","DOIUrl":"https://doi.org/10.1007/s10278-023-00957-z","url":null,"abstract":"<p>This paper aims to compare the performance of the classical machine learning (CML) model and the deep learning (DL) model, and to assess the effectiveness of utilizing fusion radiomics from both CML and DL in distinguishing encephalitis from glioma in atypical cases. We analysed the axial FLAIR images of preoperative MRI in 116 patients pathologically confirmed as gliomas and clinically diagnosed with encephalitis. The 3 CML models (logistic regression (LR), support vector machine (SVM) and multi-layer perceptron (MLP)), 3 DL models (DenseNet 121, ResNet 50 and ResNet 18) and a deep learning radiomic (DLR) model were established, respectively. The area under the receiver operating curve (AUC) and sensitivity, specificity, accuracy, negative predictive value (NPV) and positive predictive value (PPV) were calculated for the training and validation sets. In addition, a deep learning radiomic nomogram (DLRN) and a web calculator were designed as a tool to aid clinical decision-making. The best DL model (ResNet50) consistently outperformed the best CML model (LR). The DLR model had the best predictive performance, with AUC, sensitivity, specificity, accuracy, NPV and PPV of 0.879, 0.929, 0.800, 0.875, 0.867 and 0.889 in the validation sets, respectively. Calibration curve of DLR model shows good agreement between prediction and observation, and the decision curve analysis (DCA) indicated that the DLR model had higher overall net benefit than the other two models (ResNet50 and LR). Meanwhile, the DLRN and web calculator can provide dynamic assessments. Machine learning (ML) models have the potential to non-invasively differentiate between encephalitis and glioma in atypical cases. Furthermore, combining DL and CML techniques could enhance the performance of the ML models.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"19 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139463191","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-12DOI: 10.1007/s10278-023-00925-7
Sami Azam, Sidratul Montaha, Mohaimenul Azam Khan Raiaan, A. K. M. Rakibul Haque Rafid, Saddam Hossain Mukta, Mirjam Jonkman
An automated computer-aided approach might aid radiologists in diagnosing breast cancer at a primary stage. This study proposes a novel decision support system to classify breast tumors into benign and malignant based on clinically important features, using ultrasound images. Nine handcrafted features, which align with the clinical markers used by radiologists, are extracted from the region of interest (ROI) of ultrasound images. To validate that these elected clinical markers have a significant impact on predicting the benign and malignant classes, ten machine learning (ML) models are experimented with resulting in test accuracies in the range of 96 to 99%. In addition, four feature selection techniques are explored where two features are eliminated according to the feature ranking score of each feature selection method. The Random Forest classifier is trained with the resultant four feature sets. Results indicate that even when eliminating only two features, the performance of the model is reduced for each feature selection technique. These experiments validate the efficiency and effectiveness of the clinically important features. To develop the decision support system, a probability density function (PDF) graph is generated for each feature in order to find a threshold range to distinguish benign and malignant tumors. Based on the threshold range of particular features, a decision support system is developed in such a way that if at least eight out of nine features are within the threshold range, the image will be denoted as true predicted. With this algorithm, a test accuracy of 99.38% and an F1 Score of 99.05% is achieved, which means that our decision support system outperforms all the previously trained ML models. Moreover, after calculating individual class-based test accuracies, for the benign class, a test accuracy of 99.31% has been attained where only three benign instances are misclassified out of 437 instances, and for the malignant class, a test accuracy of 99.52% has been attained where only one malignant instance is misclassified out of 210 instances. This system is robust, time-effective, and reliable as the radiologists’ criteria are followed and may aid specialists in making a diagnosis.
{"title":"An Automated Decision Support System to Analyze Malignancy Patterns of Breast Masses Employing Medically Relevant Features of Ultrasound Images","authors":"Sami Azam, Sidratul Montaha, Mohaimenul Azam Khan Raiaan, A. K. M. Rakibul Haque Rafid, Saddam Hossain Mukta, Mirjam Jonkman","doi":"10.1007/s10278-023-00925-7","DOIUrl":"https://doi.org/10.1007/s10278-023-00925-7","url":null,"abstract":"<p>An automated computer-aided approach might aid radiologists in diagnosing breast cancer at a primary stage. This study proposes a novel decision support system to classify breast tumors into benign and malignant based on clinically important features, using ultrasound images. Nine handcrafted features, which align with the clinical markers used by radiologists, are extracted from the region of interest (ROI) of ultrasound images. To validate that these elected clinical markers have a significant impact on predicting the benign and malignant classes, ten machine learning (ML) models are experimented with resulting in test accuracies in the range of 96 to 99%. In addition, four feature selection techniques are explored where two features are eliminated according to the feature ranking score of each feature selection method. The Random Forest classifier is trained with the resultant four feature sets. Results indicate that even when eliminating only two features, the performance of the model is reduced for each feature selection technique. These experiments validate the efficiency and effectiveness of the clinically important features. To develop the decision support system, a probability density function (PDF) graph is generated for each feature in order to find a threshold range to distinguish benign and malignant tumors. Based on the threshold range of particular features, a decision support system is developed in such a way that if at least eight out of nine features are within the threshold range, the image will be denoted as true predicted. With this algorithm, a test accuracy of 99.38% and an F1 Score of 99.05% is achieved, which means that our decision support system outperforms all the previously trained ML models. Moreover, after calculating individual class-based test accuracies, for the benign class, a test accuracy of 99.31% has been attained where only three benign instances are misclassified out of 437 instances, and for the malignant class, a test accuracy of 99.52% has been attained where only one malignant instance is misclassified out of 210 instances. This system is robust, time-effective, and reliable as the radiologists’ criteria are followed and may aid specialists in making a diagnosis.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"52 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139463330","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-11DOI: 10.1007/s10278-023-00914-w
Marcello Chang, Joshua J. Reicher, Angad Kalra, Michael Muelly, Yousef Ahmad
We previously validated Fibresolve, a machine learning classifier system that non-invasively predicts idiopathic pulmonary fibrosis (IPF) diagnosis. The system incorporates an automated deep learning algorithm that analyzes chest computed tomography (CT) imaging to assess for features associated with idiopathic pulmonary fibrosis. Here, we assess performance in assessment of patterns beyond those that are characteristic features of usual interstitial pneumonia (UIP) pattern. The machine learning classifier was previously developed and validated using standard training, validation, and test sets, with clinical plus pathologically determined ground truth. The multi-site 295-patient validation dataset was used for focused subgroup analysis in this investigation to evaluate the classifier’s performance range in cases with and without radiologic UIP and probable UIP designations. Radiologic assessment of specific features for UIP including the presence and distribution of reticulation, ground glass, bronchiectasis, and honeycombing was used for assignment of radiologic pattern. Output from the classifier was assessed within various UIP subgroups. The machine learning classifier was able to classify cases not meeting the criteria for UIP or probable UIP as IPF with estimated sensitivity of 56–65% and estimated specificity of 92–94%. Example cases demonstrated non-basilar-predominant as well as ground glass patterns that were indeterminate for UIP by subjective imaging criteria but for which the classifier system was able to correctly identify the case as IPF as confirmed by multidisciplinary discussion generally inclusive of histopathology. The machine learning classifier Fibresolve may be helpful in the diagnosis of IPF in cases without radiological UIP and probable UIP patterns.
{"title":"Analysis of Validation Performance of a Machine Learning Classifier in Interstitial Lung Disease Cases Without Definite or Probable Usual Interstitial Pneumonia Pattern on CT Using Clinical and Pathology-Supported Diagnostic Labels","authors":"Marcello Chang, Joshua J. Reicher, Angad Kalra, Michael Muelly, Yousef Ahmad","doi":"10.1007/s10278-023-00914-w","DOIUrl":"https://doi.org/10.1007/s10278-023-00914-w","url":null,"abstract":"<p>We previously validated Fibresolve, a machine learning classifier system that non-invasively predicts idiopathic pulmonary fibrosis (IPF) diagnosis. The system incorporates an automated deep learning algorithm that analyzes chest computed tomography (CT) imaging to assess for features associated with idiopathic pulmonary fibrosis. Here, we assess performance in assessment of patterns beyond those that are characteristic features of usual interstitial pneumonia (UIP) pattern. The machine learning classifier was previously developed and validated using standard training, validation, and test sets, with clinical plus pathologically determined ground truth. The multi-site 295-patient validation dataset was used for focused subgroup analysis in this investigation to evaluate the classifier’s performance range in cases with and without radiologic UIP and probable UIP designations. Radiologic assessment of specific features for UIP including the presence and distribution of reticulation, ground glass, bronchiectasis, and honeycombing was used for assignment of radiologic pattern. Output from the classifier was assessed within various UIP subgroups. The machine learning classifier was able to classify cases not meeting the criteria for UIP or probable UIP as IPF with estimated sensitivity of 56–65% and estimated specificity of 92–94%. Example cases demonstrated non-basilar-predominant as well as ground glass patterns that were indeterminate for UIP by subjective imaging criteria but for which the classifier system was able to correctly identify the case as IPF as confirmed by multidisciplinary discussion generally inclusive of histopathology. The machine learning classifier Fibresolve may be helpful in the diagnosis of IPF in cases without radiological UIP and probable UIP patterns.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"7 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139463329","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Lung cancer is one of the leading causes of death worldwide and early detection is crucial to reduce the mortality. A reliable computer-aided diagnosis (CAD) system can help facilitate early detection of malignant nodules. Although existing methods provide adequate classification accuracy, there is still room for further improvement. This study is dedicated to investigating a new CAD scheme for predicting the malignant likelihood of lung nodules in computed tomography (CT) images in light of a deep learning strategy. Conceived from the residual learning and selective kernel, we investigated an efficient residual selective kernel (RSK) block to handle the diversity of lung nodules with various shapes and obscure structures. Founded on this RSK block, we established a multiview RSK network (MRSKNet), to which three anatomical planes in the axial, coronal, and sagittal directions were fed. To reinforce the classification efficiency, seven handcrafted texture features with a filter-like computation strategy were explored, among which the homogeneity (HOM) feature maps are combined with the corresponding intensity CT images for concatenation input, leading to an improved network architecture. Evaluated on the public benchmark Lung Image Database Consortium and Image Database Resource Initiative (LIDC-IDRI) challenge database with ten-fold cross validation of binary classification, our experimental results indicated high area under receiver operating characteristic (AUC) and accuracy scores. A better compromise between recall and specificity was struck using the suggested concatenation strategy comparing to many state-of-the-art approaches. The proposed pulmonary nodule classification framework exhibited great efficacy and achieved a higher AUC of 0.9711. The association of handcrafted texture features with deep learning models is promising in advancing the classification performance. The developed pulmonary nodule CAD network architecture is of potential in facilitating the diagnosis of lung cancer for further image processing applications.
{"title":"Pulmonary Nodule Classification Using a Multiview Residual Selective Kernel Network","authors":"Herng-Hua Chang, Cheng-Zhe Wu, Audrey Haihong Gallogly","doi":"10.1007/s10278-023-00928-4","DOIUrl":"https://doi.org/10.1007/s10278-023-00928-4","url":null,"abstract":"<p>Lung cancer is one of the leading causes of death worldwide and early detection is crucial to reduce the mortality. A reliable computer-aided diagnosis (CAD) system can help facilitate early detection of malignant nodules. Although existing methods provide adequate classification accuracy, there is still room for further improvement. This study is dedicated to investigating a new CAD scheme for predicting the malignant likelihood of lung nodules in computed tomography (CT) images in light of a deep learning strategy. Conceived from the residual learning and selective kernel, we investigated an efficient residual selective kernel (RSK) block to handle the diversity of lung nodules with various shapes and obscure structures. Founded on this RSK block, we established a multiview RSK network (MRSKNet), to which three anatomical planes in the axial, coronal, and sagittal directions were fed. To reinforce the classification efficiency, seven handcrafted texture features with a filter-like computation strategy were explored, among which the homogeneity (HOM) feature maps are combined with the corresponding intensity CT images for concatenation input, leading to an improved network architecture. Evaluated on the public benchmark Lung Image Database Consortium and Image Database Resource Initiative (LIDC-IDRI) challenge database with ten-fold cross validation of binary classification, our experimental results indicated high area under receiver operating characteristic (AUC) and accuracy scores. A better compromise between recall and specificity was struck using the suggested concatenation strategy comparing to many state-of-the-art approaches. The proposed pulmonary nodule classification framework exhibited great efficacy and achieved a higher AUC of 0.9711. The association of handcrafted texture features with deep learning models is promising in advancing the classification performance. The developed pulmonary nodule CAD network architecture is of potential in facilitating the diagnosis of lung cancer for further image processing applications.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"4 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139463328","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-11DOI: 10.1007/s10278-023-00920-y
Michael K. Hoy, Vishal Desai, Simukayi Mutasa, Robert C. Hoy, Richard Gorniak, Jeffrey A. Belair
To use a novel deep learning system to localize the hip joints and detect findings of cam-type femoroacetabular impingement (FAI). A retrospective search of hip/pelvis radiographs obtained in patients to evaluate for FAI yielded 3050 total studies. Each hip was classified separately by the original interpreting radiologist in the following manner: 724 hips had severe cam-type FAI morphology, 962 moderate cam-type FAI morphology, 846 mild cam-type FAI morphology, and 518 hips were normal. The anteroposterior (AP) view from each study was anonymized and extracted. After localization of the hip joints by a novel convolutional neural network (CNN) based on the focal loss principle, a second CNN classified the images of the hip as cam positive, or no FAI. Accuracy was 74% for diagnosing normal vs. abnormal cam-type FAI morphology, with aggregate sensitivity and specificity of 0.821 and 0.669, respectively, at the chosen operating point. The aggregate AUC was 0.736. A deep learning system can be applied to detect FAI-related changes on single view pelvic radiographs. Deep learning is useful for quickly identifying and categorizing pathology on imaging, which may aid the interpreting radiologist.
使用新型深度学习系统定位髋关节并检测凸轮型股骨髋臼撞击症(FAI)的发现。对患者为评估 FAI 而获得的髋关节/骨盆 X 光片进行回顾性检索,共获得 3050 项研究结果。每个髋关节都由最初的放射科医生按以下方式进行了分类:724个髋关节有重度凸轮型FAI形态,962个髋关节有中度凸轮型FAI形态,846个髋关节有轻度凸轮型FAI形态,518个髋关节正常。每项研究的前胸(AP)切面都经过匿名处理和提取。基于焦点损失原理的新型卷积神经网络(CNN)对髋关节进行定位后,第二个 CNN 将髋关节图像分为凸轮阳性或无 FAI 两类。在所选操作点上,诊断正常与异常凸轮型 FAI 形态的准确率为 74%,总灵敏度和特异度分别为 0.821 和 0.669。总的 AUC 为 0.736。深度学习系统可用于检测单视角骨盆X光片上与FAI相关的变化。深度学习有助于快速识别成像上的病理变化并对其进行分类,从而为放射科医生提供帮助。
{"title":"Deep Learning–Assisted Identification of Femoroacetabular Impingement (FAI) on Routine Pelvic Radiographs","authors":"Michael K. Hoy, Vishal Desai, Simukayi Mutasa, Robert C. Hoy, Richard Gorniak, Jeffrey A. Belair","doi":"10.1007/s10278-023-00920-y","DOIUrl":"https://doi.org/10.1007/s10278-023-00920-y","url":null,"abstract":"<p>To use a novel deep learning system to localize the hip joints and detect findings of cam-type femoroacetabular impingement (FAI). A retrospective search of hip/pelvis radiographs obtained in patients to evaluate for FAI yielded 3050 total studies. Each hip was classified separately by the original interpreting radiologist in the following manner: 724 hips had severe cam-type FAI morphology, 962 moderate cam-type FAI morphology, 846 mild cam-type FAI morphology, and 518 hips were normal. The anteroposterior (AP) view from each study was anonymized and extracted. After localization of the hip joints by a novel convolutional neural network (CNN) based on the focal loss principle, a second CNN classified the images of the hip as cam positive, or no FAI. Accuracy was 74% for diagnosing normal vs. abnormal cam-type FAI morphology, with aggregate sensitivity and specificity of 0.821 and 0.669, respectively, at the chosen operating point. The aggregate AUC was 0.736. A deep learning system can be applied to detect FAI-related changes on single view pelvic radiographs. Deep learning is useful for quickly identifying and categorizing pathology on imaging, which may aid the interpreting radiologist.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"80 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139463123","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}