Pub Date : 2024-08-19DOI: 10.3390/jimaging10080201
Rumana Islam, Mohammed Tarique
Breast cancer is one of the paramount causes of new cancer cases worldwide annually. It is a malignant neoplasm that develops in the breast cells. The early screening of this disease is essential to prevent its metastasis. A mammogram X-ray image is the most common screening tool practiced currently when this disease is suspected; all the breast lesions identified are not malignant. The invasive fine needle aspiration (FNA) of a breast mass sample is the secondary screening tool to clinically examine cancerous lesions. The visual image analysis of the stained aspirated sample imposes a challenge for the cytologist to identify the malignant cells accurately. The formulation of an artificial intelligence-based objective technique on top of the introspective assessment is essential to avoid misdiagnosis. This paper addresses several artificial intelligence (AI)-based techniques to diagnose breast cancer from the nuclear features of FNA samples. The Wisconsin Breast Cancer dataset (WBCD) from the UCI machine learning repository is applied for this investigation. Significant statistical parameters are measured to evaluate the performance of the proposed techniques. The best detection accuracy of 98.10% is achieved with a two-layer feed-forward neural network (FFNN). Finally, the developed algorithm's performance is compared with some state-of-the-art works in the literature.
乳腺癌是全球每年新增癌症病例的主要原因之一。它是一种发生在乳腺细胞中的恶性肿瘤。对这种疾病进行早期筛查对于防止其转移至关重要。乳房 X 射线造影是目前最常用的筛查工具,当怀疑患有这种疾病时,所有发现的乳房病变都不是恶性的。乳房肿块样本的侵入性细针穿刺术(FNA)是临床上检查癌症病灶的辅助筛查工具。对抽吸出的染色样本进行视觉图像分析是细胞学专家准确识别恶性细胞的一项挑战。在内省评估的基础上,制定一种基于人工智能的客观技术对于避免误诊至关重要。本文探讨了几种基于人工智能(AI)的技术,以从 FNA 样本的核特征诊断乳腺癌。本研究采用了 UCI 机器学习库中的威斯康星乳腺癌数据集(WBCD)。对重要的统计参数进行了测量,以评估所建议技术的性能。双层前馈神经网络(FFNN)的最佳检测准确率为 98.10%。最后,将所开发算法的性能与文献中一些最先进的作品进行了比较。
{"title":"Artificial Intelligence (AI) and Nuclear Features from the Fine Needle Aspirated (FNA) Tissue Samples to Recognize Breast Cancer.","authors":"Rumana Islam, Mohammed Tarique","doi":"10.3390/jimaging10080201","DOIUrl":"10.3390/jimaging10080201","url":null,"abstract":"<p><p>Breast cancer is one of the paramount causes of new cancer cases worldwide annually. It is a malignant neoplasm that develops in the breast cells. The early screening of this disease is essential to prevent its metastasis. A mammogram X-ray image is the most common screening tool practiced currently when this disease is suspected; all the breast lesions identified are not malignant. The invasive fine needle aspiration (FNA) of a breast mass sample is the secondary screening tool to clinically examine cancerous lesions. The visual image analysis of the stained aspirated sample imposes a challenge for the cytologist to identify the malignant cells accurately. The formulation of an artificial intelligence-based objective technique on top of the introspective assessment is essential to avoid misdiagnosis. This paper addresses several artificial intelligence (AI)-based techniques to diagnose breast cancer from the nuclear features of FNA samples. The Wisconsin Breast Cancer dataset (WBCD) from the UCI machine learning repository is applied for this investigation. Significant statistical parameters are measured to evaluate the performance of the proposed techniques. The best detection accuracy of 98.10% is achieved with a two-layer feed-forward neural network (FFNN). Finally, the developed algorithm's performance is compared with some state-of-the-art works in the literature.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"10 8","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11355253/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142082033","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-16DOI: 10.3390/jimaging10080200
Joaquim Carreras
Celiac disease (CD) is a gluten-sensitive immune-mediated enteropathy. This proof-of-concept study used a convolutional neural network (CNN) to classify hematoxylin and eosin (H&E) CD histological images, normal small intestine control, and non-specified duodenal inflammation (7294, 11,642, and 5966 images, respectively). The trained network classified CD with high performance (accuracy 99.7%, precision 99.6%, recall 99.3%, F1-score 99.5%, and specificity 99.8%). Interestingly, when the same network (already trained for the 3 class images), analyzed duodenal adenocarcinoma (3723 images), the new images were classified as duodenal inflammation in 63.65%, small intestine control in 34.73%, and CD in 1.61% of the cases; and when the network was retrained using the 4 histological subtypes, the performance was above 99% for CD and 97% for adenocarcinoma. Finally, the model added 13,043 images of Crohn's disease to include other inflammatory bowel diseases; a comparison between different CNN architectures was performed, and the gradient-weighted class activation mapping (Grad-CAM) technique was used to understand why the deep learning network made its classification decisions. In conclusion, the CNN-based deep neural system classified 5 diagnoses with high performance. Narrow artificial intelligence (AI) is designed to perform tasks that typically require human intelligence, but it operates within limited constraints and is task-specific.
{"title":"Celiac Disease Deep Learning Image Classification Using Convolutional Neural Networks.","authors":"Joaquim Carreras","doi":"10.3390/jimaging10080200","DOIUrl":"10.3390/jimaging10080200","url":null,"abstract":"<p><p>Celiac disease (CD) is a gluten-sensitive immune-mediated enteropathy. This proof-of-concept study used a convolutional neural network (CNN) to classify hematoxylin and eosin (H&E) CD histological images, normal small intestine control, and non-specified duodenal inflammation (7294, 11,642, and 5966 images, respectively). The trained network classified CD with high performance (accuracy 99.7%, precision 99.6%, recall 99.3%, F1-score 99.5%, and specificity 99.8%). Interestingly, when the same network (already trained for the 3 class images), analyzed duodenal adenocarcinoma (3723 images), the new images were classified as duodenal inflammation in 63.65%, small intestine control in 34.73%, and CD in 1.61% of the cases; and when the network was retrained using the 4 histological subtypes, the performance was above 99% for CD and 97% for adenocarcinoma. Finally, the model added 13,043 images of Crohn's disease to include other inflammatory bowel diseases; a comparison between different CNN architectures was performed, and the gradient-weighted class activation mapping (Grad-CAM) technique was used to understand why the deep learning network made its classification decisions. In conclusion, the CNN-based deep neural system classified 5 diagnoses with high performance. Narrow artificial intelligence (AI) is designed to perform tasks that typically require human intelligence, but it operates within limited constraints and is task-specific.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"10 8","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-08-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11355344/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142082071","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-14DOI: 10.3390/jimaging10080198
Tobias Schmähling, Tobias Müller, Jörg Eberhardt, Stefan Elser
In this paper, we present a multi-task model that predicts disparities and confidence levels in deep stereo matching simultaneously. We do this by combining its successful model for each separate task and obtaining a multi-task model that can be trained with a proposed loss function. We show the advantages of this model compared to training and predicting disparity and confidence sequentially. This method enables an improvement of 15% to 30% in the area under the curve (AUC) metric when trained in parallel rather than sequentially. In addition, the effect of weighting the components in the loss function on the stereo and confidence performance is investigated. By improving the confidence estimate, the practicality of stereo estimators for creating distance images is increased.
{"title":"Simultaneous Stereo Matching and Confidence Estimation Network.","authors":"Tobias Schmähling, Tobias Müller, Jörg Eberhardt, Stefan Elser","doi":"10.3390/jimaging10080198","DOIUrl":"10.3390/jimaging10080198","url":null,"abstract":"<p><p>In this paper, we present a multi-task model that predicts disparities and confidence levels in deep stereo matching simultaneously. We do this by combining its successful model for each separate task and obtaining a multi-task model that can be trained with a proposed loss function. We show the advantages of this model compared to training and predicting disparity and confidence sequentially. This method enables an improvement of 15% to 30% in the area under the curve (AUC) metric when trained in parallel rather than sequentially. In addition, the effect of weighting the components in the loss function on the stereo and confidence performance is investigated. By improving the confidence estimate, the practicality of stereo estimators for creating distance images is increased.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"10 8","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11355914/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142082085","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-14DOI: 10.3390/jimaging10080199
Raffaella Marzullo, Alessandro Capestro, Renato Cosimo, Marco Fogante, Alessandro Aprile, Liliana Balardi, Mario Giordano, Gianpiero Gaio, Gabriella Gauderi, Maria Giovanna Russo, Nicolò Schicchi
Congenital absence of the pericardium (CAP) is an unusual condition discovered, in most cases, incidentally but can potentially lead to fatal complications, including severe arrhythmias and sudden death. Recently, the use of modern imaging technologies has increased the diagnosis of CAP, providing important findings for risk stratification. Nevertheless, there is not yet consensus regarding therapeutic decisions, and the management of patients with CAP remains challenging. In this paper, we discuss the pathophysiological implication of CAP, review the current literature and explain the role of multimodality imaging tools for its diagnosis, management and treatment.
先天性心包缺失(CAP)是一种不常见的疾病,大多数情况下是偶然发现的,但有可能导致致命的并发症,包括严重心律失常和猝死。近来,现代成像技术的应用提高了对 CAP 的诊断率,为风险分层提供了重要发现。然而,关于治疗决策尚未达成共识,CAP 患者的管理仍然充满挑战。在本文中,我们将讨论 CAP 的病理生理学含义,回顾目前的文献,并解释多模态成像工具在诊断、管理和治疗中的作用。
{"title":"Congenital Absence of Pericardium: The Swinging Heart.","authors":"Raffaella Marzullo, Alessandro Capestro, Renato Cosimo, Marco Fogante, Alessandro Aprile, Liliana Balardi, Mario Giordano, Gianpiero Gaio, Gabriella Gauderi, Maria Giovanna Russo, Nicolò Schicchi","doi":"10.3390/jimaging10080199","DOIUrl":"10.3390/jimaging10080199","url":null,"abstract":"<p><p>Congenital absence of the pericardium (CAP) is an unusual condition discovered, in most cases, incidentally but can potentially lead to fatal complications, including severe arrhythmias and sudden death. Recently, the use of modern imaging technologies has increased the diagnosis of CAP, providing important findings for risk stratification. Nevertheless, there is not yet consensus regarding therapeutic decisions, and the management of patients with CAP remains challenging. In this paper, we discuss the pathophysiological implication of CAP, review the current literature and explain the role of multimodality imaging tools for its diagnosis, management and treatment.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"10 8","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11355882/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142082072","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Currently, existing deep learning methods exhibit many limitations in multi-target detection, such as low accuracy and high rates of false detection and missed detections. This paper proposes an improved Faster R-CNN algorithm, aiming to enhance the algorithm's capability in detecting multi-scale targets. This algorithm has three improvements based on Faster R-CNN. Firstly, the new algorithm uses the ResNet101 network for feature extraction of the detection image, which achieves stronger feature extraction capabilities. Secondly, the new algorithm integrates Online Hard Example Mining (OHEM), Soft non-maximum suppression (Soft-NMS), and Distance Intersection Over Union (DIOU) modules, which improves the positive and negative sample imbalance and the problem of small targets being easily missed during model training. Finally, the Region Proposal Network (RPN) is simplified to achieve a faster detection speed and a lower miss rate. The multi-scale training (MST) strategy is also used to train the improved Faster R-CNN to achieve a balance between detection accuracy and efficiency. Compared to the other detection models, the improved Faster R-CNN demonstrates significant advantages in terms of mAP@0.5, F1-score, and Log average miss rate (LAMR). The model proposed in this paper provides valuable insights and inspiration for many fields, such as smart agriculture, medical diagnosis, and face recognition.
{"title":"A Multi-Scale Target Detection Method Using an Improved Faster Region Convolutional Neural Network Based on Enhanced Backbone and Optimized Mechanisms.","authors":"Qianyong Chen, Mengshan Li, Zhenghui Lai, Jihong Zhu, Lixin Guan","doi":"10.3390/jimaging10080197","DOIUrl":"10.3390/jimaging10080197","url":null,"abstract":"<p><p>Currently, existing deep learning methods exhibit many limitations in multi-target detection, such as low accuracy and high rates of false detection and missed detections. This paper proposes an improved Faster R-CNN algorithm, aiming to enhance the algorithm's capability in detecting multi-scale targets. This algorithm has three improvements based on Faster R-CNN. Firstly, the new algorithm uses the ResNet101 network for feature extraction of the detection image, which achieves stronger feature extraction capabilities. Secondly, the new algorithm integrates Online Hard Example Mining (OHEM), Soft non-maximum suppression (Soft-NMS), and Distance Intersection Over Union (DIOU) modules, which improves the positive and negative sample imbalance and the problem of small targets being easily missed during model training. Finally, the Region Proposal Network (RPN) is simplified to achieve a faster detection speed and a lower miss rate. The multi-scale training (MST) strategy is also used to train the improved Faster R-CNN to achieve a balance between detection accuracy and efficiency. Compared to the other detection models, the improved Faster R-CNN demonstrates significant advantages in terms of mAP@0.5, F1-score, and Log average miss rate (LAMR). The model proposed in this paper provides valuable insights and inspiration for many fields, such as smart agriculture, medical diagnosis, and face recognition.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"10 8","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11355408/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142082030","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-13DOI: 10.3390/jimaging10080196
Bihi Sabiri, Amal Khtira, Bouchra El Asri, Maryem Rhanoui
In recent years, contrastive learning has been a highly favored method for self-supervised representation learning, which significantly improves the unsupervised training of deep image models. Self-supervised learning is a subset of unsupervised learning in which the learning process is supervised by creating pseudolabels from the data themselves. Using supervised final adjustments after unsupervised pretraining is one way to take the most valuable information from a vast collection of unlabeled data and teach from a small number of labeled instances. This study aims firstly to compare contrastive learning with other traditional learning models; secondly to demonstrate by experimental studies the superiority of contrastive learning during classification; thirdly to fine-tune performance using pretrained models and appropriate hyperparameter selection; and finally to address the challenge of using contrastive learning techniques to produce data representations with semantic meaning that are independent of irrelevant factors like position, lighting, and background. Relying on contrastive techniques, the model efficiently captures meaningful representations by discerning similarities and differences between modified copies of the same image. The proposed strategy, involving unsupervised pretraining followed by supervised fine-tuning, improves the robustness, accuracy, and knowledge extraction of deep image models. The results show that even with a modest 5% of data labeled, the semisupervised model achieves an accuracy of 57.72%. However, the use of supervised learning with a contrastive approach and careful hyperparameter tuning increases accuracy to 85.43%. Further adjustment of the hyperparameters resulted in an excellent accuracy of 88.70%.
{"title":"Investigating Contrastive Pair Learning's Frontiers in Supervised, Semisupervised, and Self-Supervised Learning.","authors":"Bihi Sabiri, Amal Khtira, Bouchra El Asri, Maryem Rhanoui","doi":"10.3390/jimaging10080196","DOIUrl":"10.3390/jimaging10080196","url":null,"abstract":"<p><p>In recent years, contrastive learning has been a highly favored method for self-supervised representation learning, which significantly improves the unsupervised training of deep image models. Self-supervised learning is a subset of unsupervised learning in which the learning process is supervised by creating pseudolabels from the data themselves. Using supervised final adjustments after unsupervised pretraining is one way to take the most valuable information from a vast collection of unlabeled data and teach from a small number of labeled instances. This study aims firstly to compare contrastive learning with other traditional learning models; secondly to demonstrate by experimental studies the superiority of contrastive learning during classification; thirdly to fine-tune performance using pretrained models and appropriate hyperparameter selection; and finally to address the challenge of using contrastive learning techniques to produce data representations with semantic meaning that are independent of irrelevant factors like position, lighting, and background. Relying on contrastive techniques, the model efficiently captures meaningful representations by discerning similarities and differences between modified copies of the same image. The proposed strategy, involving unsupervised pretraining followed by supervised fine-tuning, improves the robustness, accuracy, and knowledge extraction of deep image models. The results show that even with a modest 5% of data labeled, the semisupervised model achieves an accuracy of 57.72%. However, the use of supervised learning with a contrastive approach and careful hyperparameter tuning increases accuracy to 85.43%. Further adjustment of the hyperparameters resulted in an excellent accuracy of 88.70%.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"10 8","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11355693/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142082078","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-10DOI: 10.3390/jimaging10080195
Andrea Loddo, Marco Usai, Cecilia Di Ruberto
Gastric cancer is the fifth most common and fourth deadliest cancer worldwide, with a bleak 5-year survival rate of about 20%. Despite significant research into its pathobiology, prognostic predictability remains insufficient due to pathologists' heavy workloads and the potential for diagnostic errors. Consequently, there is a pressing need for automated and precise histopathological diagnostic tools. This study leverages Machine Learning and Deep Learning techniques to classify histopathological images into healthy and cancerous categories. By utilizing both handcrafted and deep features and shallow learning classifiers on the GasHisSDB dataset, we conduct a comparative analysis to identify the most effective combinations of features and classifiers for differentiating normal from abnormal histopathological images without employing fine-tuning strategies. Our methodology achieves an accuracy of 95% with the SVM classifier, underscoring the effectiveness of feature fusion strategies. Additionally, cross-magnification experiments produced promising results with accuracies close to 80% and 90% when testing the models on unseen testing images with different resolutions.
{"title":"Gastric Cancer Image Classification: A Comparative Analysis and Feature Fusion Strategies.","authors":"Andrea Loddo, Marco Usai, Cecilia Di Ruberto","doi":"10.3390/jimaging10080195","DOIUrl":"10.3390/jimaging10080195","url":null,"abstract":"<p><p>Gastric cancer is the fifth most common and fourth deadliest cancer worldwide, with a bleak 5-year survival rate of about 20%. Despite significant research into its pathobiology, prognostic predictability remains insufficient due to pathologists' heavy workloads and the potential for diagnostic errors. Consequently, there is a pressing need for automated and precise histopathological diagnostic tools. This study leverages Machine Learning and Deep Learning techniques to classify histopathological images into healthy and cancerous categories. By utilizing both handcrafted and deep features and shallow learning classifiers on the GasHisSDB dataset, we conduct a comparative analysis to identify the most effective combinations of features and classifiers for differentiating normal from abnormal histopathological images without employing fine-tuning strategies. Our methodology achieves an accuracy of 95% with the SVM classifier, underscoring the effectiveness of feature fusion strategies. Additionally, cross-magnification experiments produced promising results with accuracies close to 80% and 90% when testing the models on unseen testing images with different resolutions.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"10 8","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-08-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11355805/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142082075","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-09DOI: 10.3390/jimaging10080194
Antonio Malvasi, Lorenzo E Malgieri, Ettore Cicinelli, Antonella Vimercati, Reuven Achiron, Radmila Sparić, Antonio D'Amato, Giorgio Maria Baldini, Miriam Dellino, Giuseppe Trojano, Renata Beck, Tommaso Difonzo, Andrea Tinelli
<p><p>Asynclitism, a misalignment of the fetal head with respect to the plane of passage through the birth canal, represents a significant obstetric challenge. High degrees of asynclitism are associated with labor dystocia, difficult operative delivery, and cesarean delivery. Despite its clinical relevance, the diagnosis of asynclitism and its influence on the outcome of labor remain matters of debate. This study analyzes the role of the degree of asynclitism (AD) in assessing labor progress and predicting labor outcome, focusing on its ability to predict intrapartum cesarean delivery (ICD) versus non-cesarean delivery. The study also aims to assess the performance of the AIDA (Artificial Intelligence Dystocia Algorithm) algorithm in integrating AD with other ultrasound parameters for predicting labor outcome. This retrospective study involved 135 full-term nulliparous patients with singleton fetuses in cephalic presentation undergoing neuraxial analgesia. Data were collected at three Italian hospitals between January 2014 and December 2020. In addition to routine digital vaginal examination, all patients underwent intrapartum ultrasound (IU) during protracted second stage of labor (greater than three hours). Four geometric parameters were measured using standard 3.5 MHz transabdominal ultrasound probes: head-to-symphysis distance (HSD), degree of asynclitism (AD), angle of progression (AoP), and midline angle (MLA). The AIDA algorithm, a machine learning-based decision support system, was used to classify patients into five classes (from 0 to 4) based on the values of the four geometric parameters and to predict labor outcome (ICD or non-ICD). Six machine learning algorithms were used: MLP (multi-layer perceptron), RF (random forest), SVM (support vector machine), XGBoost, LR (logistic regression), and DT (decision tree). Pearson's correlation was used to investigate the relationship between AD and the other parameters. A degree of asynclitism greater than 70 mm was found to be significantly associated with an increased rate of cesarean deliveries. Pearson's correlation analysis showed a weak to very weak correlation between AD and AoP (PC = 0.36, <i>p</i> < 0.001), AD and HSD (PC = 0.18, <i>p</i> < 0.05), and AD and MLA (PC = 0.14). The AIDA algorithm demonstrated high accuracy in predicting labor outcome, particularly for AIDA classes 0 and 4, with 100% agreement with physician-practiced labor outcome in two cases (RF and SVM algorithms) and slightly lower agreement with MLP. For AIDA class 3, the RF algorithm performed best, with an accuracy of 92%. AD, in combination with HSD, MLA, and AoP, plays a significant role in predicting labor dystocia and labor outcome. The AIDA algorithm, based on these four geometric parameters, has proven to be a promising decision support tool for predicting labor outcome and may help reduce the need for unnecessary cesarean deliveries, while improving maternal-fetal outcomes. Future studies with larger cohorts
{"title":"AIDA (Artificial Intelligence Dystocia Algorithm) in Prolonged Dystocic Labor: Focus on Asynclitism Degree.","authors":"Antonio Malvasi, Lorenzo E Malgieri, Ettore Cicinelli, Antonella Vimercati, Reuven Achiron, Radmila Sparić, Antonio D'Amato, Giorgio Maria Baldini, Miriam Dellino, Giuseppe Trojano, Renata Beck, Tommaso Difonzo, Andrea Tinelli","doi":"10.3390/jimaging10080194","DOIUrl":"10.3390/jimaging10080194","url":null,"abstract":"<p><p>Asynclitism, a misalignment of the fetal head with respect to the plane of passage through the birth canal, represents a significant obstetric challenge. High degrees of asynclitism are associated with labor dystocia, difficult operative delivery, and cesarean delivery. Despite its clinical relevance, the diagnosis of asynclitism and its influence on the outcome of labor remain matters of debate. This study analyzes the role of the degree of asynclitism (AD) in assessing labor progress and predicting labor outcome, focusing on its ability to predict intrapartum cesarean delivery (ICD) versus non-cesarean delivery. The study also aims to assess the performance of the AIDA (Artificial Intelligence Dystocia Algorithm) algorithm in integrating AD with other ultrasound parameters for predicting labor outcome. This retrospective study involved 135 full-term nulliparous patients with singleton fetuses in cephalic presentation undergoing neuraxial analgesia. Data were collected at three Italian hospitals between January 2014 and December 2020. In addition to routine digital vaginal examination, all patients underwent intrapartum ultrasound (IU) during protracted second stage of labor (greater than three hours). Four geometric parameters were measured using standard 3.5 MHz transabdominal ultrasound probes: head-to-symphysis distance (HSD), degree of asynclitism (AD), angle of progression (AoP), and midline angle (MLA). The AIDA algorithm, a machine learning-based decision support system, was used to classify patients into five classes (from 0 to 4) based on the values of the four geometric parameters and to predict labor outcome (ICD or non-ICD). Six machine learning algorithms were used: MLP (multi-layer perceptron), RF (random forest), SVM (support vector machine), XGBoost, LR (logistic regression), and DT (decision tree). Pearson's correlation was used to investigate the relationship between AD and the other parameters. A degree of asynclitism greater than 70 mm was found to be significantly associated with an increased rate of cesarean deliveries. Pearson's correlation analysis showed a weak to very weak correlation between AD and AoP (PC = 0.36, <i>p</i> < 0.001), AD and HSD (PC = 0.18, <i>p</i> < 0.05), and AD and MLA (PC = 0.14). The AIDA algorithm demonstrated high accuracy in predicting labor outcome, particularly for AIDA classes 0 and 4, with 100% agreement with physician-practiced labor outcome in two cases (RF and SVM algorithms) and slightly lower agreement with MLP. For AIDA class 3, the RF algorithm performed best, with an accuracy of 92%. AD, in combination with HSD, MLA, and AoP, plays a significant role in predicting labor dystocia and labor outcome. The AIDA algorithm, based on these four geometric parameters, has proven to be a promising decision support tool for predicting labor outcome and may help reduce the need for unnecessary cesarean deliveries, while improving maternal-fetal outcomes. Future studies with larger cohorts ","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"10 8","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-08-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11355300/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142082032","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Railway track defects pose significant safety risks and can lead to accidents, economic losses, and loss of life. Traditional manual inspection methods are either time-consuming, costly, or prone to human error. This paper proposes RailTrack-DaViT, a novel vision transformer-based approach for railway track defect classification. By leveraging the Dual Attention Vision Transformer (DaViT) architecture, RailTrack-DaViT effectively captures both global and local information, enabling accurate defect detection. The model is trained and evaluated on multiple datasets including rail, fastener and fishplate, multi-faults, and ThaiRailTrack. A comprehensive analysis of the model's performance is provided including confusion matrices, training visualizations, and classification metrics. RailTrack-DaViT demonstrates superior performance compared to state-of-the-art CNN-based methods, achieving the highest accuracies: 96.9% on the rail dataset, 98.9% on the fastener and fishplate dataset, and 98.8% on the multi-faults dataset. Moreover, RailTrack-DaViT outperforms baselines on the ThaiRailTrack dataset with 99.2% accuracy, quickly adapts to unseen images, and shows better model stability during fine-tuning. This capability can significantly reduce time consumption when applying the model to novel datasets in practical applications.
{"title":"RailTrack-DaViT: A Vision Transformer-Based Approach for Automated Railway Track Defect Detection.","authors":"Aniwat Phaphuangwittayakul, Napat Harnpornchai, Fangli Ying, Jinming Zhang","doi":"10.3390/jimaging10080192","DOIUrl":"10.3390/jimaging10080192","url":null,"abstract":"<p><p>Railway track defects pose significant safety risks and can lead to accidents, economic losses, and loss of life. Traditional manual inspection methods are either time-consuming, costly, or prone to human error. This paper proposes RailTrack-DaViT, a novel vision transformer-based approach for railway track defect classification. By leveraging the Dual Attention Vision Transformer (DaViT) architecture, RailTrack-DaViT effectively captures both global and local information, enabling accurate defect detection. The model is trained and evaluated on multiple datasets including rail, fastener and fishplate, multi-faults, and ThaiRailTrack. A comprehensive analysis of the model's performance is provided including confusion matrices, training visualizations, and classification metrics. RailTrack-DaViT demonstrates superior performance compared to state-of-the-art CNN-based methods, achieving the highest accuracies: 96.9% on the rail dataset, 98.9% on the fastener and fishplate dataset, and 98.8% on the multi-faults dataset. Moreover, RailTrack-DaViT outperforms baselines on the ThaiRailTrack dataset with 99.2% accuracy, quickly adapts to unseen images, and shows better model stability during fine-tuning. This capability can significantly reduce time consumption when applying the model to novel datasets in practical applications.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"10 8","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-08-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11355430/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142082082","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-07DOI: 10.3390/jimaging10080191
Qi Chang, Danish Ahmad, Jennifer Toth, Rebecca Bascom, William E Higgins
For patients at risk of developing either lung cancer or colorectal cancer, the identification of suspect lesions in endoscopic video is an important procedure. The physician performs an endoscopic exam by navigating an endoscope through the organ of interest, be it the lungs or intestinal tract, and performs a visual inspection of the endoscopic video stream to identify lesions. Unfortunately, this entails a tedious, error-prone search over a lengthy video sequence. We propose a deep learning architecture that enables the real-time detection and segmentation of lesion regions from endoscopic video, with our experiments focused on autofluorescence bronchoscopy (AFB) for the lungs and colonoscopy for the intestinal tract. Our architecture, dubbed ESFPNet, draws on a pretrained Mix Transformer (MiT) encoder and a decoder structure that incorporates a new Efficient Stage-Wise Feature Pyramid (ESFP) to promote accurate lesion segmentation. In comparison to existing deep learning models, the ESFPNet model gave superior lesion segmentation performance for an AFB dataset. It also produced superior segmentation results for three widely used public colonoscopy databases and nearly the best results for two other public colonoscopy databases. In addition, the lightweight ESFPNet architecture requires fewer model parameters and less computation than other competing models, enabling the real-time analysis of input video frames. Overall, these studies point to the combined superior analysis performance and architectural efficiency of the ESFPNet for endoscopic video analysis. Lastly, additional experiments with the public colonoscopy databases demonstrate the learning ability and generalizability of ESFPNet, implying that the model could be effective for region segmentation in other domains.
{"title":"ESFPNet: Efficient Stage-Wise Feature Pyramid on Mix Transformer for Deep Learning-Based Cancer Analysis in Endoscopic Video.","authors":"Qi Chang, Danish Ahmad, Jennifer Toth, Rebecca Bascom, William E Higgins","doi":"10.3390/jimaging10080191","DOIUrl":"10.3390/jimaging10080191","url":null,"abstract":"<p><p>For patients at risk of developing either lung cancer or colorectal cancer, the identification of suspect lesions in endoscopic video is an important procedure. The physician performs an endoscopic exam by navigating an endoscope through the organ of interest, be it the lungs or intestinal tract, and performs a visual inspection of the endoscopic video stream to identify lesions. Unfortunately, this entails a tedious, error-prone search over a lengthy video sequence. We propose a deep learning architecture that enables the real-time detection and segmentation of lesion regions from endoscopic video, with our experiments focused on autofluorescence bronchoscopy (AFB) for the lungs and colonoscopy for the intestinal tract. Our architecture, dubbed ESFPNet, draws on a pretrained Mix Transformer (MiT) encoder and a decoder structure that incorporates a new Efficient Stage-Wise Feature Pyramid (ESFP) to promote accurate lesion segmentation. In comparison to existing deep learning models, the ESFPNet model gave superior lesion segmentation performance for an AFB dataset. It also produced superior segmentation results for three widely used public colonoscopy databases and nearly the best results for two other public colonoscopy databases. In addition, the lightweight ESFPNet architecture requires fewer model parameters and less computation than other competing models, enabling the real-time analysis of input video frames. Overall, these studies point to the combined superior analysis performance and architectural efficiency of the ESFPNet for endoscopic video analysis. Lastly, additional experiments with the public colonoscopy databases demonstrate the learning ability and generalizability of ESFPNet, implying that the model could be effective for region segmentation in other domains.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"10 8","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-08-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11355868/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142082073","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}