首页 > 最新文献

Journal of Imaging最新文献

英文 中文
Artificial Intelligence (AI) and Nuclear Features from the Fine Needle Aspirated (FNA) Tissue Samples to Recognize Breast Cancer. 人工智能(AI)与细针抽吸(FNA)组织样本中的核特征识别乳腺癌。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2024-08-19 DOI: 10.3390/jimaging10080201
Rumana Islam, Mohammed Tarique

Breast cancer is one of the paramount causes of new cancer cases worldwide annually. It is a malignant neoplasm that develops in the breast cells. The early screening of this disease is essential to prevent its metastasis. A mammogram X-ray image is the most common screening tool practiced currently when this disease is suspected; all the breast lesions identified are not malignant. The invasive fine needle aspiration (FNA) of a breast mass sample is the secondary screening tool to clinically examine cancerous lesions. The visual image analysis of the stained aspirated sample imposes a challenge for the cytologist to identify the malignant cells accurately. The formulation of an artificial intelligence-based objective technique on top of the introspective assessment is essential to avoid misdiagnosis. This paper addresses several artificial intelligence (AI)-based techniques to diagnose breast cancer from the nuclear features of FNA samples. The Wisconsin Breast Cancer dataset (WBCD) from the UCI machine learning repository is applied for this investigation. Significant statistical parameters are measured to evaluate the performance of the proposed techniques. The best detection accuracy of 98.10% is achieved with a two-layer feed-forward neural network (FFNN). Finally, the developed algorithm's performance is compared with some state-of-the-art works in the literature.

乳腺癌是全球每年新增癌症病例的主要原因之一。它是一种发生在乳腺细胞中的恶性肿瘤。对这种疾病进行早期筛查对于防止其转移至关重要。乳房 X 射线造影是目前最常用的筛查工具,当怀疑患有这种疾病时,所有发现的乳房病变都不是恶性的。乳房肿块样本的侵入性细针穿刺术(FNA)是临床上检查癌症病灶的辅助筛查工具。对抽吸出的染色样本进行视觉图像分析是细胞学专家准确识别恶性细胞的一项挑战。在内省评估的基础上,制定一种基于人工智能的客观技术对于避免误诊至关重要。本文探讨了几种基于人工智能(AI)的技术,以从 FNA 样本的核特征诊断乳腺癌。本研究采用了 UCI 机器学习库中的威斯康星乳腺癌数据集(WBCD)。对重要的统计参数进行了测量,以评估所建议技术的性能。双层前馈神经网络(FFNN)的最佳检测准确率为 98.10%。最后,将所开发算法的性能与文献中一些最先进的作品进行了比较。
{"title":"Artificial Intelligence (AI) and Nuclear Features from the Fine Needle Aspirated (FNA) Tissue Samples to Recognize Breast Cancer.","authors":"Rumana Islam, Mohammed Tarique","doi":"10.3390/jimaging10080201","DOIUrl":"10.3390/jimaging10080201","url":null,"abstract":"<p><p>Breast cancer is one of the paramount causes of new cancer cases worldwide annually. It is a malignant neoplasm that develops in the breast cells. The early screening of this disease is essential to prevent its metastasis. A mammogram X-ray image is the most common screening tool practiced currently when this disease is suspected; all the breast lesions identified are not malignant. The invasive fine needle aspiration (FNA) of a breast mass sample is the secondary screening tool to clinically examine cancerous lesions. The visual image analysis of the stained aspirated sample imposes a challenge for the cytologist to identify the malignant cells accurately. The formulation of an artificial intelligence-based objective technique on top of the introspective assessment is essential to avoid misdiagnosis. This paper addresses several artificial intelligence (AI)-based techniques to diagnose breast cancer from the nuclear features of FNA samples. The Wisconsin Breast Cancer dataset (WBCD) from the UCI machine learning repository is applied for this investigation. Significant statistical parameters are measured to evaluate the performance of the proposed techniques. The best detection accuracy of 98.10% is achieved with a two-layer feed-forward neural network (FFNN). Finally, the developed algorithm's performance is compared with some state-of-the-art works in the literature.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"10 8","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11355253/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142082033","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Celiac Disease Deep Learning Image Classification Using Convolutional Neural Networks. 利用卷积神经网络进行乳糜泻深度学习图像分类
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2024-08-16 DOI: 10.3390/jimaging10080200
Joaquim Carreras

Celiac disease (CD) is a gluten-sensitive immune-mediated enteropathy. This proof-of-concept study used a convolutional neural network (CNN) to classify hematoxylin and eosin (H&E) CD histological images, normal small intestine control, and non-specified duodenal inflammation (7294, 11,642, and 5966 images, respectively). The trained network classified CD with high performance (accuracy 99.7%, precision 99.6%, recall 99.3%, F1-score 99.5%, and specificity 99.8%). Interestingly, when the same network (already trained for the 3 class images), analyzed duodenal adenocarcinoma (3723 images), the new images were classified as duodenal inflammation in 63.65%, small intestine control in 34.73%, and CD in 1.61% of the cases; and when the network was retrained using the 4 histological subtypes, the performance was above 99% for CD and 97% for adenocarcinoma. Finally, the model added 13,043 images of Crohn's disease to include other inflammatory bowel diseases; a comparison between different CNN architectures was performed, and the gradient-weighted class activation mapping (Grad-CAM) technique was used to understand why the deep learning network made its classification decisions. In conclusion, the CNN-based deep neural system classified 5 diagnoses with high performance. Narrow artificial intelligence (AI) is designed to perform tasks that typically require human intelligence, but it operates within limited constraints and is task-specific.

乳糜泻(CD)是一种麸质敏感性免疫介导的肠病。这项概念验证研究使用卷积神经网络(CNN)对苏木精和伊红(H&E)CD 组织学图像、正常小肠对照和非特定十二指肠炎症(分别为 7294、11642 和 5966 张图像)进行分类。训练有素的网络对 CD 进行了高效分类(准确率 99.7%、精确率 99.6%、召回率 99.3%、F1 分数 99.5%、特异性 99.8%)。有趣的是,当同一网络(已针对 3 类图像进行过训练)分析十二指肠腺癌(3723 幅图像)时,63.65% 的新图像被归类为十二指肠炎症,34.73% 的新图像被归类为小肠控制,1.61% 的新图像被归类为 CD;当使用 4 种组织学亚型对网络进行再训练时,CD 的分类准确率超过 99%,腺癌的分类准确率超过 97%。最后,该模型添加了 13,043 张克罗恩病图像,以包括其他炎症性肠病;对不同的 CNN 架构进行了比较,并使用梯度加权类激活映射(Grad-CAM)技术来了解深度学习网络做出分类决定的原因。总之,基于 CNN 的深度神经系统对 5 种诊断进行了高性能分类。狭义人工智能(AI)旨在执行通常需要人类智能的任务,但它在有限的限制条件下运行,并且针对特定任务。
{"title":"Celiac Disease Deep Learning Image Classification Using Convolutional Neural Networks.","authors":"Joaquim Carreras","doi":"10.3390/jimaging10080200","DOIUrl":"10.3390/jimaging10080200","url":null,"abstract":"<p><p>Celiac disease (CD) is a gluten-sensitive immune-mediated enteropathy. This proof-of-concept study used a convolutional neural network (CNN) to classify hematoxylin and eosin (H&E) CD histological images, normal small intestine control, and non-specified duodenal inflammation (7294, 11,642, and 5966 images, respectively). The trained network classified CD with high performance (accuracy 99.7%, precision 99.6%, recall 99.3%, F1-score 99.5%, and specificity 99.8%). Interestingly, when the same network (already trained for the 3 class images), analyzed duodenal adenocarcinoma (3723 images), the new images were classified as duodenal inflammation in 63.65%, small intestine control in 34.73%, and CD in 1.61% of the cases; and when the network was retrained using the 4 histological subtypes, the performance was above 99% for CD and 97% for adenocarcinoma. Finally, the model added 13,043 images of Crohn's disease to include other inflammatory bowel diseases; a comparison between different CNN architectures was performed, and the gradient-weighted class activation mapping (Grad-CAM) technique was used to understand why the deep learning network made its classification decisions. In conclusion, the CNN-based deep neural system classified 5 diagnoses with high performance. Narrow artificial intelligence (AI) is designed to perform tasks that typically require human intelligence, but it operates within limited constraints and is task-specific.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"10 8","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-08-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11355344/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142082071","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Simultaneous Stereo Matching and Confidence Estimation Network. 同步立体匹配和可信度估计网络。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2024-08-14 DOI: 10.3390/jimaging10080198
Tobias Schmähling, Tobias Müller, Jörg Eberhardt, Stefan Elser

In this paper, we present a multi-task model that predicts disparities and confidence levels in deep stereo matching simultaneously. We do this by combining its successful model for each separate task and obtaining a multi-task model that can be trained with a proposed loss function. We show the advantages of this model compared to training and predicting disparity and confidence sequentially. This method enables an improvement of 15% to 30% in the area under the curve (AUC) metric when trained in parallel rather than sequentially. In addition, the effect of weighting the components in the loss function on the stereo and confidence performance is investigated. By improving the confidence estimate, the practicality of stereo estimators for creating distance images is increased.

在本文中,我们提出了一种多任务模型,可同时预测深度立体匹配中的差距和置信度。为此,我们将其成功的模型与每个独立任务相结合,得到了一个可使用建议的损失函数进行训练的多任务模型。我们展示了该模型与按顺序训练和预测差异和置信度相比的优势。与顺序训练相比,这种方法可将曲线下面积 (AUC) 指标提高 15% 至 30%。此外,还研究了加权损失函数中的成分对立体和置信度性能的影响。通过改进置信度估计,提高了用于创建距离图像的立体估计器的实用性。
{"title":"Simultaneous Stereo Matching and Confidence Estimation Network.","authors":"Tobias Schmähling, Tobias Müller, Jörg Eberhardt, Stefan Elser","doi":"10.3390/jimaging10080198","DOIUrl":"10.3390/jimaging10080198","url":null,"abstract":"<p><p>In this paper, we present a multi-task model that predicts disparities and confidence levels in deep stereo matching simultaneously. We do this by combining its successful model for each separate task and obtaining a multi-task model that can be trained with a proposed loss function. We show the advantages of this model compared to training and predicting disparity and confidence sequentially. This method enables an improvement of 15% to 30% in the area under the curve (AUC) metric when trained in parallel rather than sequentially. In addition, the effect of weighting the components in the loss function on the stereo and confidence performance is investigated. By improving the confidence estimate, the practicality of stereo estimators for creating distance images is increased.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"10 8","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11355914/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142082085","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Congenital Absence of Pericardium: The Swinging Heart. 先天性心包缺失:摇摆的心脏
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2024-08-14 DOI: 10.3390/jimaging10080199
Raffaella Marzullo, Alessandro Capestro, Renato Cosimo, Marco Fogante, Alessandro Aprile, Liliana Balardi, Mario Giordano, Gianpiero Gaio, Gabriella Gauderi, Maria Giovanna Russo, Nicolò Schicchi

Congenital absence of the pericardium (CAP) is an unusual condition discovered, in most cases, incidentally but can potentially lead to fatal complications, including severe arrhythmias and sudden death. Recently, the use of modern imaging technologies has increased the diagnosis of CAP, providing important findings for risk stratification. Nevertheless, there is not yet consensus regarding therapeutic decisions, and the management of patients with CAP remains challenging. In this paper, we discuss the pathophysiological implication of CAP, review the current literature and explain the role of multimodality imaging tools for its diagnosis, management and treatment.

先天性心包缺失(CAP)是一种不常见的疾病,大多数情况下是偶然发现的,但有可能导致致命的并发症,包括严重心律失常和猝死。近来,现代成像技术的应用提高了对 CAP 的诊断率,为风险分层提供了重要发现。然而,关于治疗决策尚未达成共识,CAP 患者的管理仍然充满挑战。在本文中,我们将讨论 CAP 的病理生理学含义,回顾目前的文献,并解释多模态成像工具在诊断、管理和治疗中的作用。
{"title":"Congenital Absence of Pericardium: The Swinging Heart.","authors":"Raffaella Marzullo, Alessandro Capestro, Renato Cosimo, Marco Fogante, Alessandro Aprile, Liliana Balardi, Mario Giordano, Gianpiero Gaio, Gabriella Gauderi, Maria Giovanna Russo, Nicolò Schicchi","doi":"10.3390/jimaging10080199","DOIUrl":"10.3390/jimaging10080199","url":null,"abstract":"<p><p>Congenital absence of the pericardium (CAP) is an unusual condition discovered, in most cases, incidentally but can potentially lead to fatal complications, including severe arrhythmias and sudden death. Recently, the use of modern imaging technologies has increased the diagnosis of CAP, providing important findings for risk stratification. Nevertheless, there is not yet consensus regarding therapeutic decisions, and the management of patients with CAP remains challenging. In this paper, we discuss the pathophysiological implication of CAP, review the current literature and explain the role of multimodality imaging tools for its diagnosis, management and treatment.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"10 8","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11355882/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142082072","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Multi-Scale Target Detection Method Using an Improved Faster Region Convolutional Neural Network Based on Enhanced Backbone and Optimized Mechanisms. 基于增强骨干和优化机制的改进型快速区域卷积神经网络的多尺度目标检测方法。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2024-08-13 DOI: 10.3390/jimaging10080197
Qianyong Chen, Mengshan Li, Zhenghui Lai, Jihong Zhu, Lixin Guan

Currently, existing deep learning methods exhibit many limitations in multi-target detection, such as low accuracy and high rates of false detection and missed detections. This paper proposes an improved Faster R-CNN algorithm, aiming to enhance the algorithm's capability in detecting multi-scale targets. This algorithm has three improvements based on Faster R-CNN. Firstly, the new algorithm uses the ResNet101 network for feature extraction of the detection image, which achieves stronger feature extraction capabilities. Secondly, the new algorithm integrates Online Hard Example Mining (OHEM), Soft non-maximum suppression (Soft-NMS), and Distance Intersection Over Union (DIOU) modules, which improves the positive and negative sample imbalance and the problem of small targets being easily missed during model training. Finally, the Region Proposal Network (RPN) is simplified to achieve a faster detection speed and a lower miss rate. The multi-scale training (MST) strategy is also used to train the improved Faster R-CNN to achieve a balance between detection accuracy and efficiency. Compared to the other detection models, the improved Faster R-CNN demonstrates significant advantages in terms of mAP@0.5, F1-score, and Log average miss rate (LAMR). The model proposed in this paper provides valuable insights and inspiration for many fields, such as smart agriculture, medical diagnosis, and face recognition.

目前,现有的深度学习方法在多目标检测中表现出许多局限性,如准确率低、误检率和漏检率高等。本文提出了一种改进的 Faster R-CNN 算法,旨在增强该算法检测多尺度目标的能力。该算法在 Faster R-CNN 的基础上进行了三方面的改进。首先,新算法使用 ResNet101 网络对检测图像进行特征提取,实现了更强的特征提取能力。其次,新算法集成了在线硬样本挖掘(OHEM)、软非最大抑制(Soft-NMS)和距离交叉联合(DIOU)模块,改善了正负样本不平衡和模型训练过程中容易遗漏小目标的问题。最后,简化了区域建议网络(RPN),以实现更快的检测速度和更低的漏检率。多尺度训练(MST)策略也被用于训练改进后的 Faster R-CNN,以实现检测精度和效率之间的平衡。与其他检测模型相比,改进的 Faster R-CNN 在 mAP@0.5、F1 分数和对数平均漏检率(LAMR)方面都有显著优势。本文提出的模型为智能农业、医疗诊断和人脸识别等多个领域提供了宝贵的见解和启发。
{"title":"A Multi-Scale Target Detection Method Using an Improved Faster Region Convolutional Neural Network Based on Enhanced Backbone and Optimized Mechanisms.","authors":"Qianyong Chen, Mengshan Li, Zhenghui Lai, Jihong Zhu, Lixin Guan","doi":"10.3390/jimaging10080197","DOIUrl":"10.3390/jimaging10080197","url":null,"abstract":"<p><p>Currently, existing deep learning methods exhibit many limitations in multi-target detection, such as low accuracy and high rates of false detection and missed detections. This paper proposes an improved Faster R-CNN algorithm, aiming to enhance the algorithm's capability in detecting multi-scale targets. This algorithm has three improvements based on Faster R-CNN. Firstly, the new algorithm uses the ResNet101 network for feature extraction of the detection image, which achieves stronger feature extraction capabilities. Secondly, the new algorithm integrates Online Hard Example Mining (OHEM), Soft non-maximum suppression (Soft-NMS), and Distance Intersection Over Union (DIOU) modules, which improves the positive and negative sample imbalance and the problem of small targets being easily missed during model training. Finally, the Region Proposal Network (RPN) is simplified to achieve a faster detection speed and a lower miss rate. The multi-scale training (MST) strategy is also used to train the improved Faster R-CNN to achieve a balance between detection accuracy and efficiency. Compared to the other detection models, the improved Faster R-CNN demonstrates significant advantages in terms of mAP@0.5, F1-score, and Log average miss rate (LAMR). The model proposed in this paper provides valuable insights and inspiration for many fields, such as smart agriculture, medical diagnosis, and face recognition.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"10 8","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11355408/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142082030","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Investigating Contrastive Pair Learning's Frontiers in Supervised, Semisupervised, and Self-Supervised Learning. 调查对比配对学习在监督、半监督和自我监督学习中的应用前沿。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2024-08-13 DOI: 10.3390/jimaging10080196
Bihi Sabiri, Amal Khtira, Bouchra El Asri, Maryem Rhanoui

In recent years, contrastive learning has been a highly favored method for self-supervised representation learning, which significantly improves the unsupervised training of deep image models. Self-supervised learning is a subset of unsupervised learning in which the learning process is supervised by creating pseudolabels from the data themselves. Using supervised final adjustments after unsupervised pretraining is one way to take the most valuable information from a vast collection of unlabeled data and teach from a small number of labeled instances. This study aims firstly to compare contrastive learning with other traditional learning models; secondly to demonstrate by experimental studies the superiority of contrastive learning during classification; thirdly to fine-tune performance using pretrained models and appropriate hyperparameter selection; and finally to address the challenge of using contrastive learning techniques to produce data representations with semantic meaning that are independent of irrelevant factors like position, lighting, and background. Relying on contrastive techniques, the model efficiently captures meaningful representations by discerning similarities and differences between modified copies of the same image. The proposed strategy, involving unsupervised pretraining followed by supervised fine-tuning, improves the robustness, accuracy, and knowledge extraction of deep image models. The results show that even with a modest 5% of data labeled, the semisupervised model achieves an accuracy of 57.72%. However, the use of supervised learning with a contrastive approach and careful hyperparameter tuning increases accuracy to 85.43%. Further adjustment of the hyperparameters resulted in an excellent accuracy of 88.70%.

近年来,对比学习是一种备受青睐的自监督表征学习方法,它能显著改善深度图像模型的无监督训练。自监督学习是无监督学习的一个子集,它通过从数据本身创建伪标签来监督学习过程。在无监督预训练后进行有监督的最终调整,是一种从大量无标签数据中获取最有价值信息,并从少量有标签实例中进行教学的方法。本研究的目的首先是将对比学习与其他传统学习模型进行比较;其次是通过实验研究证明对比学习在分类过程中的优越性;第三是利用预训练模型和适当的超参数选择对性能进行微调;最后是解决利用对比学习技术生成具有语义的数据表示所面临的挑战,这种数据表示不受位置、光照和背景等无关因素的影响。依靠对比技术,该模型通过辨别同一图像的修改副本之间的异同,有效地捕捉到有意义的表征。所提出的策略包括无监督预训练和有监督微调,从而提高了深度图像模型的鲁棒性、准确性和知识提取。结果表明,即使只有区区 5%的标注数据,半监督模型也能达到 57.72% 的准确率。然而,使用监督学习的对比方法和仔细的超参数调整可将准确率提高到 85.43%。进一步调整超参数后,准确率达到了 88.70%。
{"title":"Investigating Contrastive Pair Learning's Frontiers in Supervised, Semisupervised, and Self-Supervised Learning.","authors":"Bihi Sabiri, Amal Khtira, Bouchra El Asri, Maryem Rhanoui","doi":"10.3390/jimaging10080196","DOIUrl":"10.3390/jimaging10080196","url":null,"abstract":"<p><p>In recent years, contrastive learning has been a highly favored method for self-supervised representation learning, which significantly improves the unsupervised training of deep image models. Self-supervised learning is a subset of unsupervised learning in which the learning process is supervised by creating pseudolabels from the data themselves. Using supervised final adjustments after unsupervised pretraining is one way to take the most valuable information from a vast collection of unlabeled data and teach from a small number of labeled instances. This study aims firstly to compare contrastive learning with other traditional learning models; secondly to demonstrate by experimental studies the superiority of contrastive learning during classification; thirdly to fine-tune performance using pretrained models and appropriate hyperparameter selection; and finally to address the challenge of using contrastive learning techniques to produce data representations with semantic meaning that are independent of irrelevant factors like position, lighting, and background. Relying on contrastive techniques, the model efficiently captures meaningful representations by discerning similarities and differences between modified copies of the same image. The proposed strategy, involving unsupervised pretraining followed by supervised fine-tuning, improves the robustness, accuracy, and knowledge extraction of deep image models. The results show that even with a modest 5% of data labeled, the semisupervised model achieves an accuracy of 57.72%. However, the use of supervised learning with a contrastive approach and careful hyperparameter tuning increases accuracy to 85.43%. Further adjustment of the hyperparameters resulted in an excellent accuracy of 88.70%.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"10 8","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11355693/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142082078","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Gastric Cancer Image Classification: A Comparative Analysis and Feature Fusion Strategies. 胃癌图像分类:对比分析与特征融合策略
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2024-08-10 DOI: 10.3390/jimaging10080195
Andrea Loddo, Marco Usai, Cecilia Di Ruberto

Gastric cancer is the fifth most common and fourth deadliest cancer worldwide, with a bleak 5-year survival rate of about 20%. Despite significant research into its pathobiology, prognostic predictability remains insufficient due to pathologists' heavy workloads and the potential for diagnostic errors. Consequently, there is a pressing need for automated and precise histopathological diagnostic tools. This study leverages Machine Learning and Deep Learning techniques to classify histopathological images into healthy and cancerous categories. By utilizing both handcrafted and deep features and shallow learning classifiers on the GasHisSDB dataset, we conduct a comparative analysis to identify the most effective combinations of features and classifiers for differentiating normal from abnormal histopathological images without employing fine-tuning strategies. Our methodology achieves an accuracy of 95% with the SVM classifier, underscoring the effectiveness of feature fusion strategies. Additionally, cross-magnification experiments produced promising results with accuracies close to 80% and 90% when testing the models on unseen testing images with different resolutions.

胃癌是全球第五大常见癌症,也是第四大致命癌症,5 年生存率仅为 20%。尽管对胃癌的病理生物学进行了大量研究,但由于病理学家的工作量繁重以及可能出现诊断错误,预后预测能力仍然不足。因此,迫切需要自动化和精确的组织病理学诊断工具。本研究利用机器学习和深度学习技术将组织病理学图像分为健康和癌症两类。通过在 GasHisSDB 数据集上使用手工制作的深度特征和浅层学习分类器,我们进行了比较分析,以确定最有效的特征和分类器组合,从而在不使用微调策略的情况下区分正常和异常组织病理学图像。我们的方法使 SVM 分类器的准确率达到了 95%,突出了特征融合策略的有效性。此外,在不同分辨率的未见测试图像上对模型进行测试时,交叉放大实验也取得了很好的结果,准确率分别接近 80% 和 90%。
{"title":"Gastric Cancer Image Classification: A Comparative Analysis and Feature Fusion Strategies.","authors":"Andrea Loddo, Marco Usai, Cecilia Di Ruberto","doi":"10.3390/jimaging10080195","DOIUrl":"10.3390/jimaging10080195","url":null,"abstract":"<p><p>Gastric cancer is the fifth most common and fourth deadliest cancer worldwide, with a bleak 5-year survival rate of about 20%. Despite significant research into its pathobiology, prognostic predictability remains insufficient due to pathologists' heavy workloads and the potential for diagnostic errors. Consequently, there is a pressing need for automated and precise histopathological diagnostic tools. This study leverages Machine Learning and Deep Learning techniques to classify histopathological images into healthy and cancerous categories. By utilizing both handcrafted and deep features and shallow learning classifiers on the GasHisSDB dataset, we conduct a comparative analysis to identify the most effective combinations of features and classifiers for differentiating normal from abnormal histopathological images without employing fine-tuning strategies. Our methodology achieves an accuracy of 95% with the SVM classifier, underscoring the effectiveness of feature fusion strategies. Additionally, cross-magnification experiments produced promising results with accuracies close to 80% and 90% when testing the models on unseen testing images with different resolutions.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"10 8","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-08-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11355805/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142082075","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AIDA (Artificial Intelligence Dystocia Algorithm) in Prolonged Dystocic Labor: Focus on Asynclitism Degree. AIDA (人工智能难产算法)在难产产程延长中的应用:关注非同步化程度。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2024-08-09 DOI: 10.3390/jimaging10080194
Antonio Malvasi, Lorenzo E Malgieri, Ettore Cicinelli, Antonella Vimercati, Reuven Achiron, Radmila Sparić, Antonio D'Amato, Giorgio Maria Baldini, Miriam Dellino, Giuseppe Trojano, Renata Beck, Tommaso Difonzo, Andrea Tinelli
<p><p>Asynclitism, a misalignment of the fetal head with respect to the plane of passage through the birth canal, represents a significant obstetric challenge. High degrees of asynclitism are associated with labor dystocia, difficult operative delivery, and cesarean delivery. Despite its clinical relevance, the diagnosis of asynclitism and its influence on the outcome of labor remain matters of debate. This study analyzes the role of the degree of asynclitism (AD) in assessing labor progress and predicting labor outcome, focusing on its ability to predict intrapartum cesarean delivery (ICD) versus non-cesarean delivery. The study also aims to assess the performance of the AIDA (Artificial Intelligence Dystocia Algorithm) algorithm in integrating AD with other ultrasound parameters for predicting labor outcome. This retrospective study involved 135 full-term nulliparous patients with singleton fetuses in cephalic presentation undergoing neuraxial analgesia. Data were collected at three Italian hospitals between January 2014 and December 2020. In addition to routine digital vaginal examination, all patients underwent intrapartum ultrasound (IU) during protracted second stage of labor (greater than three hours). Four geometric parameters were measured using standard 3.5 MHz transabdominal ultrasound probes: head-to-symphysis distance (HSD), degree of asynclitism (AD), angle of progression (AoP), and midline angle (MLA). The AIDA algorithm, a machine learning-based decision support system, was used to classify patients into five classes (from 0 to 4) based on the values of the four geometric parameters and to predict labor outcome (ICD or non-ICD). Six machine learning algorithms were used: MLP (multi-layer perceptron), RF (random forest), SVM (support vector machine), XGBoost, LR (logistic regression), and DT (decision tree). Pearson's correlation was used to investigate the relationship between AD and the other parameters. A degree of asynclitism greater than 70 mm was found to be significantly associated with an increased rate of cesarean deliveries. Pearson's correlation analysis showed a weak to very weak correlation between AD and AoP (PC = 0.36, <i>p</i> < 0.001), AD and HSD (PC = 0.18, <i>p</i> < 0.05), and AD and MLA (PC = 0.14). The AIDA algorithm demonstrated high accuracy in predicting labor outcome, particularly for AIDA classes 0 and 4, with 100% agreement with physician-practiced labor outcome in two cases (RF and SVM algorithms) and slightly lower agreement with MLP. For AIDA class 3, the RF algorithm performed best, with an accuracy of 92%. AD, in combination with HSD, MLA, and AoP, plays a significant role in predicting labor dystocia and labor outcome. The AIDA algorithm, based on these four geometric parameters, has proven to be a promising decision support tool for predicting labor outcome and may help reduce the need for unnecessary cesarean deliveries, while improving maternal-fetal outcomes. Future studies with larger cohorts
胎头不齐(Asynclitism)是指胎儿头部与通过产道的平面错位,是产科的一大难题。高度异位与分娩难产、难产手术和剖宫产有关。尽管与临床相关,异位的诊断及其对分娩结果的影响仍存在争议。本研究分析了异位程度(AD)在评估产程进展和预测分娩结局中的作用,重点关注其预测产中剖宫产(ICD)与非剖宫产的能力。该研究还旨在评估 AIDA(人工智能难产算法)算法在将 AD 与其他超声参数相结合以预测分娩结果方面的性能。这项回顾性研究涉及135名接受神经轴镇痛的头位单胎足月无痛分娩患者。数据收集于 2014 年 1 月至 2020 年 12 月期间的三家意大利医院。除常规数字阴道检查外,所有患者均在第二产程延长期间(超过三小时)接受了产程超声检查(IU)。使用标准的 3.5 MHz 经腹超声探头测量了四个几何参数:头到骨骺的距离 (HSD)、不对称度 (AD)、进展角 (AoP) 和中线角 (MLA)。AIDA 算法是一种基于机器学习的决策支持系统,用于根据四个几何参数的值将患者分为五个等级(从 0 到 4),并预测分娩结果(ICD 或非 ICD)。该系统使用了六种机器学习算法:MLP(多层感知器)、RF(随机森林)、SVM(支持向量机)、XGBoost、LR(逻辑回归)和 DT(决策树)。采用皮尔逊相关性来研究 AD 与其他参数之间的关系。结果发现,不对称程度大于 70 毫米与剖宫产率增加有显著相关性。皮尔逊相关分析表明,AD 与 AoP(PC = 0.36,p < 0.001)、AD 与 HSD(PC = 0.18,p < 0.05)、AD 与 MLA(PC = 0.14)之间存在弱到极弱的相关性。AIDA 算法在预测分娩结局方面表现出很高的准确性,尤其是对于 AIDA 0 级和 4 级,在两种情况下(RF 算法和 SVM 算法)与医生实践分娩结局的一致性达到 100%,而与 MLP 的一致性略低。对于 AIDA 分级 3,RF 算法表现最佳,准确率为 92%。AD与HSD、MLA和AoP相结合,在预测分娩难产和分娩结局方面发挥着重要作用。事实证明,基于这四个几何参数的AIDA算法是预测分娩结局的一种很有前途的决策支持工具,可帮助减少不必要的剖宫产,同时改善母胎结局。未来需要对更大的队列进行研究,以进一步验证这些发现,并完善AIDA算法中AD和其他参数的临界值。
{"title":"AIDA (Artificial Intelligence Dystocia Algorithm) in Prolonged Dystocic Labor: Focus on Asynclitism Degree.","authors":"Antonio Malvasi, Lorenzo E Malgieri, Ettore Cicinelli, Antonella Vimercati, Reuven Achiron, Radmila Sparić, Antonio D'Amato, Giorgio Maria Baldini, Miriam Dellino, Giuseppe Trojano, Renata Beck, Tommaso Difonzo, Andrea Tinelli","doi":"10.3390/jimaging10080194","DOIUrl":"10.3390/jimaging10080194","url":null,"abstract":"&lt;p&gt;&lt;p&gt;Asynclitism, a misalignment of the fetal head with respect to the plane of passage through the birth canal, represents a significant obstetric challenge. High degrees of asynclitism are associated with labor dystocia, difficult operative delivery, and cesarean delivery. Despite its clinical relevance, the diagnosis of asynclitism and its influence on the outcome of labor remain matters of debate. This study analyzes the role of the degree of asynclitism (AD) in assessing labor progress and predicting labor outcome, focusing on its ability to predict intrapartum cesarean delivery (ICD) versus non-cesarean delivery. The study also aims to assess the performance of the AIDA (Artificial Intelligence Dystocia Algorithm) algorithm in integrating AD with other ultrasound parameters for predicting labor outcome. This retrospective study involved 135 full-term nulliparous patients with singleton fetuses in cephalic presentation undergoing neuraxial analgesia. Data were collected at three Italian hospitals between January 2014 and December 2020. In addition to routine digital vaginal examination, all patients underwent intrapartum ultrasound (IU) during protracted second stage of labor (greater than three hours). Four geometric parameters were measured using standard 3.5 MHz transabdominal ultrasound probes: head-to-symphysis distance (HSD), degree of asynclitism (AD), angle of progression (AoP), and midline angle (MLA). The AIDA algorithm, a machine learning-based decision support system, was used to classify patients into five classes (from 0 to 4) based on the values of the four geometric parameters and to predict labor outcome (ICD or non-ICD). Six machine learning algorithms were used: MLP (multi-layer perceptron), RF (random forest), SVM (support vector machine), XGBoost, LR (logistic regression), and DT (decision tree). Pearson's correlation was used to investigate the relationship between AD and the other parameters. A degree of asynclitism greater than 70 mm was found to be significantly associated with an increased rate of cesarean deliveries. Pearson's correlation analysis showed a weak to very weak correlation between AD and AoP (PC = 0.36, &lt;i&gt;p&lt;/i&gt; &lt; 0.001), AD and HSD (PC = 0.18, &lt;i&gt;p&lt;/i&gt; &lt; 0.05), and AD and MLA (PC = 0.14). The AIDA algorithm demonstrated high accuracy in predicting labor outcome, particularly for AIDA classes 0 and 4, with 100% agreement with physician-practiced labor outcome in two cases (RF and SVM algorithms) and slightly lower agreement with MLP. For AIDA class 3, the RF algorithm performed best, with an accuracy of 92%. AD, in combination with HSD, MLA, and AoP, plays a significant role in predicting labor dystocia and labor outcome. The AIDA algorithm, based on these four geometric parameters, has proven to be a promising decision support tool for predicting labor outcome and may help reduce the need for unnecessary cesarean deliveries, while improving maternal-fetal outcomes. Future studies with larger cohorts ","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"10 8","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-08-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11355300/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142082032","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
RailTrack-DaViT: A Vision Transformer-Based Approach for Automated Railway Track Defect Detection. RailTrack-DaViT:基于视觉转换器的铁路轨道缺陷自动检测方法。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2024-08-07 DOI: 10.3390/jimaging10080192
Aniwat Phaphuangwittayakul, Napat Harnpornchai, Fangli Ying, Jinming Zhang

Railway track defects pose significant safety risks and can lead to accidents, economic losses, and loss of life. Traditional manual inspection methods are either time-consuming, costly, or prone to human error. This paper proposes RailTrack-DaViT, a novel vision transformer-based approach for railway track defect classification. By leveraging the Dual Attention Vision Transformer (DaViT) architecture, RailTrack-DaViT effectively captures both global and local information, enabling accurate defect detection. The model is trained and evaluated on multiple datasets including rail, fastener and fishplate, multi-faults, and ThaiRailTrack. A comprehensive analysis of the model's performance is provided including confusion matrices, training visualizations, and classification metrics. RailTrack-DaViT demonstrates superior performance compared to state-of-the-art CNN-based methods, achieving the highest accuracies: 96.9% on the rail dataset, 98.9% on the fastener and fishplate dataset, and 98.8% on the multi-faults dataset. Moreover, RailTrack-DaViT outperforms baselines on the ThaiRailTrack dataset with 99.2% accuracy, quickly adapts to unseen images, and shows better model stability during fine-tuning. This capability can significantly reduce time consumption when applying the model to novel datasets in practical applications.

铁路轨道缺陷构成重大安全风险,可能导致事故、经济损失和人员伤亡。传统的人工检测方法要么耗时长、成本高,要么容易出现人为错误。本文提出了一种基于视觉转换器的铁路轨道缺陷分类新方法--RailTrack-DaViT。通过利用双注意力视觉转换器(DaViT)架构,RailTrack-DaViT 可有效捕捉全局和局部信息,从而实现准确的缺陷检测。该模型在多个数据集上进行了训练和评估,包括钢轨、扣件和鱼板、多重故障和 ThaiRailTrack。提供了对模型性能的全面分析,包括混淆矩阵、训练可视化和分类指标。与最先进的基于 CNN 的方法相比,RailTrack-DaViT 表现出了卓越的性能,达到了最高的准确率:在铁路数据集上达到 96.9%,在扣件和鱼板数据集上达到 98.9%,在多重故障数据集上达到 98.8%。此外,RailTrack-DaViT 在 ThaiRailTrack 数据集上的准确率为 99.2%,优于基线方法,能快速适应未见图像,并在微调过程中显示出更好的模型稳定性。在实际应用中,将模型应用于新数据集时,这种能力可以大大减少时间消耗。
{"title":"RailTrack-DaViT: A Vision Transformer-Based Approach for Automated Railway Track Defect Detection.","authors":"Aniwat Phaphuangwittayakul, Napat Harnpornchai, Fangli Ying, Jinming Zhang","doi":"10.3390/jimaging10080192","DOIUrl":"10.3390/jimaging10080192","url":null,"abstract":"<p><p>Railway track defects pose significant safety risks and can lead to accidents, economic losses, and loss of life. Traditional manual inspection methods are either time-consuming, costly, or prone to human error. This paper proposes RailTrack-DaViT, a novel vision transformer-based approach for railway track defect classification. By leveraging the Dual Attention Vision Transformer (DaViT) architecture, RailTrack-DaViT effectively captures both global and local information, enabling accurate defect detection. The model is trained and evaluated on multiple datasets including rail, fastener and fishplate, multi-faults, and ThaiRailTrack. A comprehensive analysis of the model's performance is provided including confusion matrices, training visualizations, and classification metrics. RailTrack-DaViT demonstrates superior performance compared to state-of-the-art CNN-based methods, achieving the highest accuracies: 96.9% on the rail dataset, 98.9% on the fastener and fishplate dataset, and 98.8% on the multi-faults dataset. Moreover, RailTrack-DaViT outperforms baselines on the ThaiRailTrack dataset with 99.2% accuracy, quickly adapts to unseen images, and shows better model stability during fine-tuning. This capability can significantly reduce time consumption when applying the model to novel datasets in practical applications.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"10 8","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-08-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11355430/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142082082","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ESFPNet: Efficient Stage-Wise Feature Pyramid on Mix Transformer for Deep Learning-Based Cancer Analysis in Endoscopic Video. ESFPNet:基于混合变换器的高效阶段性特征金字塔,用于基于深度学习的内窥镜视频癌症分析。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2024-08-07 DOI: 10.3390/jimaging10080191
Qi Chang, Danish Ahmad, Jennifer Toth, Rebecca Bascom, William E Higgins

For patients at risk of developing either lung cancer or colorectal cancer, the identification of suspect lesions in endoscopic video is an important procedure. The physician performs an endoscopic exam by navigating an endoscope through the organ of interest, be it the lungs or intestinal tract, and performs a visual inspection of the endoscopic video stream to identify lesions. Unfortunately, this entails a tedious, error-prone search over a lengthy video sequence. We propose a deep learning architecture that enables the real-time detection and segmentation of lesion regions from endoscopic video, with our experiments focused on autofluorescence bronchoscopy (AFB) for the lungs and colonoscopy for the intestinal tract. Our architecture, dubbed ESFPNet, draws on a pretrained Mix Transformer (MiT) encoder and a decoder structure that incorporates a new Efficient Stage-Wise Feature Pyramid (ESFP) to promote accurate lesion segmentation. In comparison to existing deep learning models, the ESFPNet model gave superior lesion segmentation performance for an AFB dataset. It also produced superior segmentation results for three widely used public colonoscopy databases and nearly the best results for two other public colonoscopy databases. In addition, the lightweight ESFPNet architecture requires fewer model parameters and less computation than other competing models, enabling the real-time analysis of input video frames. Overall, these studies point to the combined superior analysis performance and architectural efficiency of the ESFPNet for endoscopic video analysis. Lastly, additional experiments with the public colonoscopy databases demonstrate the learning ability and generalizability of ESFPNet, implying that the model could be effective for region segmentation in other domains.

对于有患肺癌或大肠癌风险的患者来说,在内窥镜视频中识别可疑病灶是一项重要的程序。医生在进行内窥镜检查时,会将内窥镜穿过相关器官(无论是肺部还是肠道),并对内窥镜视频流进行目视检查,以识别病变。遗憾的是,这需要在冗长的视频序列中进行乏味且容易出错的搜索。我们提出了一种深度学习架构,可以从内窥镜视频中实时检测和分割病变区域,我们的实验重点是肺部的自动荧光支气管镜(AFB)和肠道的结肠镜检查。我们的架构被称为 ESFPNet,它借鉴了预先训练好的混合变换器(MiT)编码器和解码器结构,其中包含了一种新的高效阶段性特征金字塔(ESFP),以促进准确的病变分割。与现有的深度学习模型相比,ESFPNet 模型在 AFB 数据集上的病变分割性能更为出色。它还为三个广泛使用的公共结肠镜检查数据库提供了出色的分割结果,并为另外两个公共结肠镜检查数据库提供了接近最佳的结果。此外,与其他同类模型相比,轻量级 ESFPNet 架构所需的模型参数和计算量更少,因此可以对输入的视频帧进行实时分析。总之,这些研究表明,ESFPNet 在内窥镜视频分析方面具有卓越的分析性能和架构效率。最后,利用公共结肠镜数据库进行的其他实验证明了 ESFPNet 的学习能力和通用性,这意味着该模型可以有效地用于其他领域的区域分割。
{"title":"ESFPNet: Efficient Stage-Wise Feature Pyramid on Mix Transformer for Deep Learning-Based Cancer Analysis in Endoscopic Video.","authors":"Qi Chang, Danish Ahmad, Jennifer Toth, Rebecca Bascom, William E Higgins","doi":"10.3390/jimaging10080191","DOIUrl":"10.3390/jimaging10080191","url":null,"abstract":"<p><p>For patients at risk of developing either lung cancer or colorectal cancer, the identification of suspect lesions in endoscopic video is an important procedure. The physician performs an endoscopic exam by navigating an endoscope through the organ of interest, be it the lungs or intestinal tract, and performs a visual inspection of the endoscopic video stream to identify lesions. Unfortunately, this entails a tedious, error-prone search over a lengthy video sequence. We propose a deep learning architecture that enables the real-time detection and segmentation of lesion regions from endoscopic video, with our experiments focused on autofluorescence bronchoscopy (AFB) for the lungs and colonoscopy for the intestinal tract. Our architecture, dubbed ESFPNet, draws on a pretrained Mix Transformer (MiT) encoder and a decoder structure that incorporates a new Efficient Stage-Wise Feature Pyramid (ESFP) to promote accurate lesion segmentation. In comparison to existing deep learning models, the ESFPNet model gave superior lesion segmentation performance for an AFB dataset. It also produced superior segmentation results for three widely used public colonoscopy databases and nearly the best results for two other public colonoscopy databases. In addition, the lightweight ESFPNet architecture requires fewer model parameters and less computation than other competing models, enabling the real-time analysis of input video frames. Overall, these studies point to the combined superior analysis performance and architectural efficiency of the ESFPNet for endoscopic video analysis. Lastly, additional experiments with the public colonoscopy databases demonstrate the learning ability and generalizability of ESFPNet, implying that the model could be effective for region segmentation in other domains.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"10 8","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-08-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11355868/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142082073","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Imaging
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1