首页 > 最新文献

Journal of Pathology Informatics最新文献

英文 中文
Towards interactive AI-authoring with prototypical few-shot classifiers in histopathology 利用组织病理学中的原型少量分类器实现交互式人工智能创作
Q2 Medicine Pub Date : 2024-06-06 DOI: 10.1016/j.jpi.2024.100388
Petr Kuritcyn , Rosalie Kletzander , Sophia Eisenberg , Thomas Wittenberg , Volker Bruns , Katja Evert , Felix Keil , Paul K. Ziegler , Katrin Bankov , Peter Wild , Markus Eckstein , Arndt Hartmann , Carol I. Geppert , Michaela Benz

A vast multitude of tasks in histopathology could potentially benefit from the support of artificial intelligence (AI). Many examples have been shown in the literature and first commercial products with FDA or CE-IVDR clearance are available. However, two key challenges remain: (1) a scarcity of thoroughly annotated images, respectively the laboriousness of this task, and (2) the creation of robust models that can cope with the data heterogeneity in the field (domain generalization). In this work, we investigate how the combination of prototypical few-shot classification models and data augmentation can address both of these challenges. Based on annotated data sets that include multiple centers, multiple scanners, and two tumor entities, we examine the robustness and the adaptability of few-shot classifiers in multiple scenarios. We demonstrate that data from one scanner and one site are sufficient to train robust few-shot classification models by applying domain-specific data augmentation. The models achieved classification performance of around 90% on a multiscanner and multicenter database, which is on par with the accuracy achieved on the primary single-center single-scanner data. Various convolutional neural network (CNN) architectures can be used for feature extraction in the few-shot model. A comparison of nine state-of-the-art architectures yielded that EfficientNet B0 provides the best trade-off between accuracy and inference time. The classification of prototypical few-shot models directly relies on class prototypes derived from example images of each class. Therefore, we investigated the influence of prototypes originating from images from different scanners and evaluated their performance also on the multiscanner database. Again, our few-shot model showed a stable performance with an average absolute deviation in accuracy compared to the primary prototypes of 1.8% points. Finally, we examined the adaptability to a new tumor entity: classification of tissue sections containing urothelial carcinoma into normal, tumor, and necrotic regions. Only three annotations per subclass (e.g., muscle and adipose tissue are subclasses of normal tissue) were provided to adapt the few-shot model, which obtained an overall accuracy of 93.6%. These results demonstrate that prototypical few-shot classification is an ideal technology for realizing an interactive AI authoring system as it only requires few annotations and can be adapted to new tasks without involving retraining of the underlying feature extraction CNN, which would in turn require a selection of hyper-parameters based on data science expert knowledge. Similarly, it can be regarded as a guided annotation system. To this end, we realized a workflow and user interface that targets non-technical users.

组织病理学中的大量任务都有可能受益于人工智能(AI)的支持。许多实例已见诸于文献,第一批获得美国食品及药物管理局(FDA)或 CE-IVDR 认证的商业产品也已问世。然而,两个关键挑战依然存在:(1) 缺乏全面注释的图像,这也是这项任务的艰巨性所在;(2) 如何创建稳健的模型,以应对该领域的数据异质性(领域泛化)。在这项工作中,我们研究了如何将原型少镜头分类模型与数据增强相结合来应对这两个挑战。基于包含多个中心、多台扫描仪和两个肿瘤实体的注释数据集,我们考察了少次分类器在多种情况下的鲁棒性和适应性。我们证明,通过应用特定领域的数据扩增,来自一台扫描仪和一个部位的数据足以训练出稳健的少次分类模型。这些模型在多扫描仪和多中心数据库中的分类性能达到了约 90%,与主要单中心单扫描仪数据的准确性相当。各种卷积神经网络(CNN)架构均可用于少数镜头模型的特征提取。对九种最先进的架构进行比较后发现,EfficientNet B0 在准确性和推理时间之间实现了最佳平衡。原型 few-shot 模型的分类直接依赖于从每个类别的示例图像中提取的类别原型。因此,我们研究了来自不同扫描仪图像的原型的影响,并在多扫描仪数据库中评估了它们的性能。同样,我们的少拍模型显示出稳定的性能,与主要原型相比,准确度的平均绝对偏差为 1.8%。最后,我们检验了对新肿瘤实体的适应性:将含有尿道癌的组织切片分为正常区、肿瘤区和坏死区。每个子类(例如,肌肉和脂肪组织是正常组织的子类)只提供了三个注释,以适应少拍模型,该模型的总体准确率为 93.6%。这些结果表明,原型少镜头分类法是实现交互式人工智能创作系统的理想技术,因为它只需要很少的注释,就能适应新的任务,而无需重新训练底层特征提取 CNN,这反过来又需要根据数据科学专家的知识来选择超参数。同样,它也可以被视为一种引导式注释系统。为此,我们实现了针对非技术用户的工作流程和用户界面。
{"title":"Towards interactive AI-authoring with prototypical few-shot classifiers in histopathology","authors":"Petr Kuritcyn ,&nbsp;Rosalie Kletzander ,&nbsp;Sophia Eisenberg ,&nbsp;Thomas Wittenberg ,&nbsp;Volker Bruns ,&nbsp;Katja Evert ,&nbsp;Felix Keil ,&nbsp;Paul K. Ziegler ,&nbsp;Katrin Bankov ,&nbsp;Peter Wild ,&nbsp;Markus Eckstein ,&nbsp;Arndt Hartmann ,&nbsp;Carol I. Geppert ,&nbsp;Michaela Benz","doi":"10.1016/j.jpi.2024.100388","DOIUrl":"10.1016/j.jpi.2024.100388","url":null,"abstract":"<div><p>A vast multitude of tasks in histopathology could potentially benefit from the support of artificial intelligence (AI). Many examples have been shown in the literature and first commercial products with FDA or CE-IVDR clearance are available. However, two key challenges remain: (1) a scarcity of thoroughly annotated images, respectively the laboriousness of this task, and (2) the creation of robust models that can cope with the data heterogeneity in the field (domain generalization). In this work, we investigate how the combination of prototypical few-shot classification models and data augmentation can address both of these challenges. Based on annotated data sets that include multiple centers, multiple scanners, and two tumor entities, we examine the robustness and the adaptability of few-shot classifiers in multiple scenarios. We demonstrate that data from one scanner and one site are sufficient to train robust few-shot classification models by applying domain-specific data augmentation. The models achieved classification performance of around 90% on a multiscanner and multicenter database, which is on par with the accuracy achieved on the primary single-center single-scanner data. Various convolutional neural network (CNN) architectures can be used for feature extraction in the few-shot model. A comparison of nine state-of-the-art architectures yielded that EfficientNet B0 provides the best trade-off between accuracy and inference time. The classification of prototypical few-shot models directly relies on class prototypes derived from example images of each class. Therefore, we investigated the influence of prototypes originating from images from different scanners and evaluated their performance also on the multiscanner database. Again, our few-shot model showed a stable performance with an average absolute deviation in accuracy compared to the primary prototypes of 1.8% points. Finally, we examined the adaptability to a new tumor entity: classification of tissue sections containing urothelial carcinoma into normal, tumor, and necrotic regions. Only three annotations per subclass (e.g., muscle and adipose tissue are subclasses of normal tissue) were provided to adapt the few-shot model, which obtained an overall accuracy of 93.6%. These results demonstrate that prototypical few-shot classification is an ideal technology for realizing an interactive AI authoring system as it only requires few annotations and can be adapted to new tasks without involving retraining of the underlying feature extraction CNN, which would in turn require a selection of hyper-parameters based on data science expert knowledge. Similarly, it can be regarded as a guided annotation system. To this end, we realized a workflow and user interface that targets non-technical users.</p></div>","PeriodicalId":37769,"journal":{"name":"Journal of Pathology Informatics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-06-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2153353924000270/pdfft?md5=05adcd36f07ac4f905fe1929289c6160&pid=1-s2.0-S2153353924000270-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141415124","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Masked pre-training of transformers for histology image analysis 用于组织学图像分析的变换器屏蔽预训练
Q2 Medicine Pub Date : 2024-05-31 DOI: 10.1016/j.jpi.2024.100386
Shuai Jiang , Liesbeth Hondelink , Arief A. Suriawinata , Saeed Hassanpour

In digital pathology, whole-slide images (WSIs) are widely used for applications such as cancer diagnosis and prognosis prediction. Vision transformer (ViT) models have recently emerged as a promising method for encoding large regions of WSIs while preserving spatial relationships among patches. However, due to the large number of model parameters and limited labeled data, applying transformer models to WSIs remains challenging. In this study, we propose a pretext task to train the transformer model in a self-supervised manner. Our model, MaskHIT, uses the transformer output to reconstruct masked patches, measured by contrastive loss. We pre-trained MaskHIT model using over 7000 WSIs from TCGA and extensively evaluated its performance in multiple experiments, covering survival prediction, cancer subtype classification, and grade prediction tasks. Our experiments demonstrate that the pre-training procedure enables context-aware understanding of WSIs, facilitates the learning of representative histological features based on patch positions and visual patterns, and is essential for the ViT model to achieve optimal results on WSI-level tasks. The pre-trained MaskHIT surpasses various multiple instance learning approaches by 3% and 2% on survival prediction and cancer subtype classification tasks, and also outperforms recent state-of-the-art transformer-based methods. Finally, a comparison between the attention maps generated by the MaskHIT model with pathologist's annotations indicates that the model can accurately identify clinically relevant histological structures on the whole slide for each task.

在数字病理学中,整幅图像(WSI)被广泛应用于癌症诊断和预后预测等领域。视觉变换器(ViT)模型是最近出现的一种很有前途的方法,它可以对大区域的 WSIs 进行编码,同时保留斑块之间的空间关系。然而,由于模型参数较多且标注数据有限,将变换器模型应用于 WSIs 仍然具有挑战性。在本研究中,我们提出了一个借口任务,以自我监督的方式训练变换器模型。我们的模型 MaskHIT 使用变换器输出来重构被遮蔽的斑块,以对比度损失来衡量。我们使用 TCGA 的 7000 多个 WSI 对 MaskHIT 模型进行了预训练,并在多个实验中对其性能进行了广泛评估,包括生存预测、癌症亚型分类和等级预测任务。我们的实验证明,预训练程序能够实现对 WSI 的上下文感知理解,促进了基于斑块位置和视觉模式的代表性组织学特征的学习,对于 ViT 模型在 WSI 级别任务中取得最佳结果至关重要。在生存预测和癌症亚型分类任务上,预训练的 MaskHIT 比各种多实例学习方法分别高出 3% 和 2%,也优于最近最先进的基于变换器的方法。最后,将 MaskHIT 模型生成的注意图与病理学家的注释进行比较,结果表明该模型能在每项任务中准确识别整张幻灯片上与临床相关的组织结构。
{"title":"Masked pre-training of transformers for histology image analysis","authors":"Shuai Jiang ,&nbsp;Liesbeth Hondelink ,&nbsp;Arief A. Suriawinata ,&nbsp;Saeed Hassanpour","doi":"10.1016/j.jpi.2024.100386","DOIUrl":"https://doi.org/10.1016/j.jpi.2024.100386","url":null,"abstract":"<div><p>In digital pathology, whole-slide images (WSIs) are widely used for applications such as cancer diagnosis and prognosis prediction. Vision transformer (ViT) models have recently emerged as a promising method for encoding large regions of WSIs while preserving spatial relationships among patches. However, due to the large number of model parameters and limited labeled data, applying transformer models to WSIs remains challenging. In this study, we propose a pretext task to train the transformer model in a self-supervised manner. Our model, MaskHIT, uses the transformer output to reconstruct masked patches, measured by contrastive loss. We pre-trained MaskHIT model using over 7000 WSIs from TCGA and extensively evaluated its performance in multiple experiments, covering survival prediction, cancer subtype classification, and grade prediction tasks. Our experiments demonstrate that the pre-training procedure enables context-aware understanding of WSIs, facilitates the learning of representative histological features based on patch positions and visual patterns, and is essential for the ViT model to achieve optimal results on WSI-level tasks. The pre-trained MaskHIT surpasses various multiple instance learning approaches by 3% and 2% on survival prediction and cancer subtype classification tasks, and also outperforms recent state-of-the-art transformer-based methods. Finally, a comparison between the attention maps generated by the MaskHIT model with pathologist's annotations indicates that the model can accurately identify clinically relevant histological structures on the whole slide for each task.</p></div>","PeriodicalId":37769,"journal":{"name":"Journal of Pathology Informatics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2153353924000257/pdfft?md5=3dfddd9f11d8384fd0c39d65dbfab6b4&pid=1-s2.0-S2153353924000257-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141434521","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Smartphone-based machine learning model for real-time assessment of medical kidney biopsy 基于智能手机的机器学习模型用于医学肾活检的实时评估
Q2 Medicine Pub Date : 2024-05-31 DOI: 10.1016/j.jpi.2024.100385
Odianosen J. Eigbire-Molen , Clarissa A. Cassol , Daniel J. Kenan , Johnathan O.H. Napier , Lyle J. Burdine , Shana M. Coley , Shree G. Sharma

Background

Kidney biopsy is the gold-standard for diagnosing medical renal diseases, but the accuracy of the diagnosis greatly depends on the quality of the biopsy specimen, particularly the amount of renal cortex obtained. Inadequate biopsies, characterized by insufficient cortex or predominant medulla, can lead to inconclusive or incorrect diagnoses, and repeat biopsy. Unfortunately, there has been a concerning increase in the rate of inadequate kidney biopsies, and not all medical centers have access to trained professionals who can assess biopsy adequacy in real time. In response to this challenge, we aimed to develop a machine learning model capable of assessing the percentage cortex of each biopsy pass using smartphone images of the kidney biopsy tissue at the time of biopsy.

Methods

747 kidney biopsy cores and corresponding smartphone macro images were collected from five unused deceased donor kidneys. Each core was imaged, formalin-fixed, sectioned, and stained with Periodic acid–Schiff (PAS) to determine cortex percentage. The fresh unfixed core images were captured using the macro camera on an iPhone 13 Pro. Two experienced renal pathologists independently reviewed the PAS-stained sections to determine the cortex percentage. For the purpose of this study, the biopsies with less than 30% cortex were labeled as inadequate, while those with 30% or more cortex were classified as adequate. The dataset was divided into training (n=643), validation (n=30), and test (n=74) sets. Preprocessing steps involved converting High-Efficiency Image Container iPhone format images to JPEG, normalization, and renal tissue segmentation using a U-Net deep learning model. Subsequently, a classification deep learning model was trained on the renal tissue region of interest and corresponding class label.

Results

The deep learning model achieved an accuracy of 85% on the training data. On the independent test dataset, the model exhibited an accuracy of 81%. For inadequate samples in the test dataset, the model showed a sensitivity of 71%, suggesting its capability to identify cases with inadequate cortical representation. The area under the receiver-operating curve (AUC-ROC) on the test dataset was 0.80.

Conclusion

We successfully developed and tested a machine learning model for classifying smartphone images of kidney biopsies as either adequate or inadequate, based on the amount of cortex determined by expert renal pathologists. The model's promising results suggest its potential as a smartphone application to assist real-time assessment of kidney biopsy tissue, particularly in settings with limited access to trained personnel. Further refinements and validations are warranted to optimize the model's performance.

背景肾活检是诊断内科肾脏疾病的金标准,但诊断的准确性在很大程度上取决于活检标本的质量,尤其是获得的肾皮质的数量。活检不充分,表现为皮质不足或髓质占优势,可导致诊断不确定或不正确,并导致重复活检。遗憾的是,肾脏活检不充分的比例一直在上升,而且并非所有医疗中心都有训练有素的专业人员来实时评估活检是否充分。为了应对这一挑战,我们旨在开发一种机器学习模型,该模型能够利用活检时肾脏活检组织的智能手机图像评估每次活检的皮质百分比。每个肾芯都经过成像、福尔马林固定、切片和过硫酸希夫(PAS)染色,以确定皮质百分比。使用 iPhone 13 Pro 的微距摄像头拍摄了新鲜的未固定肾芯图像。两名经验丰富的肾脏病理学家独立审查 PAS 染色切片,以确定皮质百分比。在本研究中,皮质少于 30% 的活检样本被标记为皮质不足,而皮质达到或超过 30% 的活检样本则被归类为皮质充足。数据集分为训练集(样本数=643)、验证集(样本数=30)和测试集(样本数=74)。预处理步骤包括将高效图像容器 iPhone 格式图像转换为 JPEG、归一化,以及使用 U-Net 深度学习模型进行肾组织分割。随后,根据感兴趣的肾组织区域和相应的类标签训练分类深度学习模型。在独立测试数据集上,该模型的准确率为 81%。对于测试数据集中的不足样本,该模型显示出 71% 的灵敏度,表明它有能力识别皮质表征不足的病例。结论我们成功开发并测试了一种机器学习模型,该模型可根据肾脏病理专家确定的皮质数量,将肾脏活检的智能手机图像分类为充分或不充分。该模型取得了令人鼓舞的结果,表明它有潜力作为智能手机应用来协助实时评估肾活检组织,尤其是在训练有素的人员有限的情况下。为了优化该模型的性能,还需要进一步的改进和验证。
{"title":"Smartphone-based machine learning model for real-time assessment of medical kidney biopsy","authors":"Odianosen J. Eigbire-Molen ,&nbsp;Clarissa A. Cassol ,&nbsp;Daniel J. Kenan ,&nbsp;Johnathan O.H. Napier ,&nbsp;Lyle J. Burdine ,&nbsp;Shana M. Coley ,&nbsp;Shree G. Sharma","doi":"10.1016/j.jpi.2024.100385","DOIUrl":"https://doi.org/10.1016/j.jpi.2024.100385","url":null,"abstract":"<div><h3>Background</h3><p>Kidney biopsy is the gold-standard for diagnosing medical renal diseases, but the accuracy of the diagnosis greatly depends on the quality of the biopsy specimen, particularly the amount of renal cortex obtained. Inadequate biopsies, characterized by insufficient cortex or predominant medulla, can lead to inconclusive or incorrect diagnoses, and repeat biopsy. Unfortunately, there has been a concerning increase in the rate of inadequate kidney biopsies, and not all medical centers have access to trained professionals who can assess biopsy adequacy in real time. In response to this challenge, we aimed to develop a machine learning model capable of assessing the percentage cortex of each biopsy pass using smartphone images of the kidney biopsy tissue at the time of biopsy.</p></div><div><h3>Methods</h3><p>747 kidney biopsy cores and corresponding smartphone macro images were collected from five unused deceased donor kidneys. Each core was imaged, formalin-fixed, sectioned, and stained with Periodic acid–Schiff (PAS) to determine cortex percentage. The fresh unfixed core images were captured using the macro camera on an iPhone 13 Pro. Two experienced renal pathologists independently reviewed the PAS-stained sections to determine the cortex percentage. For the purpose of this study, the biopsies with less than 30% cortex were labeled as inadequate, while those with 30% or more cortex were classified as adequate. The dataset was divided into training (<em>n</em>=643), validation (<em>n</em>=30), and test (<em>n</em>=74) sets. Preprocessing steps involved converting High-Efficiency Image Container iPhone format images to JPEG, normalization, and renal tissue segmentation using a U-Net deep learning model. Subsequently, a classification deep learning model was trained on the renal tissue region of interest and corresponding class label.</p></div><div><h3>Results</h3><p>The deep learning model achieved an accuracy of 85% on the training data. On the independent test dataset, the model exhibited an accuracy of 81%. For inadequate samples in the test dataset, the model showed a sensitivity of 71%, suggesting its capability to identify cases with inadequate cortical representation. The area under the receiver-operating curve (AUC-ROC) on the test dataset was 0.80.</p></div><div><h3>Conclusion</h3><p>We successfully developed and tested a machine learning model for classifying smartphone images of kidney biopsies as either adequate or inadequate, based on the amount of cortex determined by expert renal pathologists. The model's promising results suggest its potential as a smartphone application to assist real-time assessment of kidney biopsy tissue, particularly in settings with limited access to trained personnel. Further refinements and validations are warranted to optimize the model's performance.</p></div>","PeriodicalId":37769,"journal":{"name":"Journal of Pathology Informatics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2153353924000245/pdfft?md5=aa0cdf6fbf647b60d197599f7a7fc32d&pid=1-s2.0-S2153353924000245-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141541797","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Joining forces for pathology diagnostics with AI assistance: The EMPAIA initiative 联合起来,利用人工智能辅助病理诊断:EMPAIA 倡议
Q2 Medicine Pub Date : 2024-05-31 DOI: 10.1016/j.jpi.2024.100387
Norman Zerbe , Lars Ole Schwen , Christian Geißler , Katja Wiesemann , Tom Bisson , Peter Boor , Rita Carvalho , Michael Franz , Christoph Jansen , Tim-Rasmus Kiehl , Björn Lindequist , Nora Charlotte Pohlan , Sarah Schmell , Klaus Strohmenger , Falk Zakrzewski , Markus Plass , Michael Takla , Tobias Küster , André Homeyer , Peter Hufnagl

Over the past decade, artificial intelligence (AI) methods in pathology have advanced substantially. However, integration into routine clinical practice has been slow due to numerous challenges, including technical and regulatory hurdles in translating research results into clinical diagnostic products and the lack of standardized interfaces.

The open and vendor-neutral EMPAIA initiative addresses these challenges. Here, we provide an overview of EMPAIA's achievements and lessons learned. EMPAIA integrates various stakeholders of the pathology AI ecosystem, i.e., pathologists, computer scientists, and industry. In close collaboration, we developed technical interoperability standards, recommendations for AI testing and product development, and explainability methods. We implemented the modular and open-source EMPAIA Platform and successfully integrated 14 AI-based image analysis apps from eight different vendors, demonstrating how different apps can use a single standardized interface. We prioritized requirements and evaluated the use of AI in real clinical settings with 14 different pathology laboratories in Europe and Asia. In addition to technical developments, we created a forum for all stakeholders to share information and experiences on digital pathology and AI. Commercial, clinical, and academic stakeholders can now adopt EMPAIA's common open-source interfaces, providing a unique opportunity for large-scale standardization and streamlining of processes.

Further efforts are needed to effectively and broadly establish AI assistance in routine laboratory use. To this end, a sustainable infrastructure, the non-profit association EMPAIA International, has been established to continue standardization and support broad implementation and advocacy for an AI-assisted digital pathology future.

过去十年间,病理学领域的人工智能(AI)方法取得了长足的进步。然而,由于面临诸多挑战,包括将研究成果转化为临床诊断产品的技术和监管障碍,以及缺乏标准化接口等,将人工智能方法融入常规临床实践的进程十分缓慢。在此,我们将概述 EMPAIA 的成就和经验教训。EMPAIA 整合了病理人工智能生态系统的各利益相关方,即病理学家、计算机科学家和业界。通过密切合作,我们制定了技术互操作性标准、人工智能测试和产品开发建议以及可解释性方法。我们实施了模块化和开源的 EMPAIA 平台,并成功集成了来自 8 个不同供应商的 14 个基于人工智能的图像分析应用程序,展示了不同应用程序如何使用单一的标准化界面。我们对需求进行了优先排序,并与欧洲和亚洲的 14 家不同病理实验室一起评估了人工智能在实际临床环境中的应用。除了技术开发,我们还为所有利益相关者创建了一个论坛,以分享数字病理学和人工智能方面的信息和经验。商业、临床和学术利益相关者现在可以采用EMPAIA的通用开源接口,为大规模标准化和简化流程提供了一个独特的机会。为此,我们建立了一个可持续的基础设施,即非营利性协会 EMPAIA 国际,以继续实现标准化并支持广泛实施和宣传人工智能辅助数字病理学的未来。
{"title":"Joining forces for pathology diagnostics with AI assistance: The EMPAIA initiative","authors":"Norman Zerbe ,&nbsp;Lars Ole Schwen ,&nbsp;Christian Geißler ,&nbsp;Katja Wiesemann ,&nbsp;Tom Bisson ,&nbsp;Peter Boor ,&nbsp;Rita Carvalho ,&nbsp;Michael Franz ,&nbsp;Christoph Jansen ,&nbsp;Tim-Rasmus Kiehl ,&nbsp;Björn Lindequist ,&nbsp;Nora Charlotte Pohlan ,&nbsp;Sarah Schmell ,&nbsp;Klaus Strohmenger ,&nbsp;Falk Zakrzewski ,&nbsp;Markus Plass ,&nbsp;Michael Takla ,&nbsp;Tobias Küster ,&nbsp;André Homeyer ,&nbsp;Peter Hufnagl","doi":"10.1016/j.jpi.2024.100387","DOIUrl":"https://doi.org/10.1016/j.jpi.2024.100387","url":null,"abstract":"<div><p>Over the past decade, artificial intelligence (AI) methods in pathology have advanced substantially. However, integration into routine clinical practice has been slow due to numerous challenges, including technical and regulatory hurdles in translating research results into clinical diagnostic products and the lack of standardized interfaces.</p><p>The open and vendor-neutral EMPAIA initiative addresses these challenges. Here, we provide an overview of EMPAIA's achievements and lessons learned. EMPAIA integrates various stakeholders of the pathology AI ecosystem, i.e., pathologists, computer scientists, and industry. In close collaboration, we developed technical interoperability standards, recommendations for AI testing and product development, and explainability methods. We implemented the modular and open-source EMPAIA Platform and successfully integrated 14 AI-based image analysis apps from eight different vendors, demonstrating how different apps can use a single standardized interface. We prioritized requirements and evaluated the use of AI in real clinical settings with 14 different pathology laboratories in Europe and Asia. In addition to technical developments, we created a forum for all stakeholders to share information and experiences on digital pathology and AI. Commercial, clinical, and academic stakeholders can now adopt EMPAIA's common open-source interfaces, providing a unique opportunity for large-scale standardization and streamlining of processes.</p><p>Further efforts are needed to effectively and broadly establish AI assistance in routine laboratory use. To this end, a sustainable infrastructure, the non-profit association EMPAIA International, has been established to continue standardization and support broad implementation and advocacy for an AI-assisted digital pathology future.</p></div>","PeriodicalId":37769,"journal":{"name":"Journal of Pathology Informatics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2153353924000269/pdfft?md5=93cff7c5dd94e55a015f5beb1d21f7eb&pid=1-s2.0-S2153353924000269-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141422757","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A novel Slide-seq based image processing software to identify gene expression at the single cell level 基于 Slide-seq 图像处理软件的新型单细胞基因表达鉴定软件
Q2 Medicine Pub Date : 2024-05-31 DOI: 10.1016/j.jpi.2024.100384
Th.I. Götz , X. Cong , S. Rauber , M. Angeli , E.W. Lang , A. Ramming , C. Schmidkonz

Analysis of gene expression at the single-cell level could help predict the effectiveness of therapies in the field of chronic inflammatory diseases such as arthritis. Here, we demonstrate an adopted approach for processing images from the Slide-seq method. Using a puck, which consists of about 50,000 DNA barcode beads, an RNA sequence of a cell is to be read. The pucks are repeatedly brought into contact with liquids and then recorded with a conventional epifluorescence microscope. The image analysis initially consists of stitching the partial images of a sequence recording, registering images from different sequences, and finally reading out the bases. The new method enables the use of an inexpensive epifluorescence microscope instead of a confocal microscope.

单细胞水平的基因表达分析有助于预测关节炎等慢性炎症性疾病的治疗效果。在这里,我们展示了一种采用 Slide-seq 方法处理图像的方法。使用由大约 50,000 个 DNA 条形码珠组成的小球,可以读取细胞的 RNA 序列。小球反复与液体接触,然后用传统的荧光显微镜进行记录。图像分析最初包括拼接序列记录的部分图像、登记不同序列的图像以及最终读出碱基。这种新方法可以使用廉价的外荧光显微镜代替共聚焦显微镜。
{"title":"A novel Slide-seq based image processing software to identify gene expression at the single cell level","authors":"Th.I. Götz ,&nbsp;X. Cong ,&nbsp;S. Rauber ,&nbsp;M. Angeli ,&nbsp;E.W. Lang ,&nbsp;A. Ramming ,&nbsp;C. Schmidkonz","doi":"10.1016/j.jpi.2024.100384","DOIUrl":"https://doi.org/10.1016/j.jpi.2024.100384","url":null,"abstract":"<div><p>Analysis of gene expression at the single-cell level could help predict the effectiveness of therapies in the field of chronic inflammatory diseases such as arthritis. Here, we demonstrate an adopted approach for processing images from the Slide-seq method. Using a puck, which consists of about 50,000 DNA barcode beads, an RNA sequence of a cell is to be read. The pucks are repeatedly brought into contact with liquids and then recorded with a conventional epifluorescence microscope. The image analysis initially consists of stitching the partial images of a sequence recording, registering images from different sequences, and finally reading out the bases. The new method enables the use of an inexpensive epifluorescence microscope instead of a confocal microscope.</p></div>","PeriodicalId":37769,"journal":{"name":"Journal of Pathology Informatics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2153353924000233/pdfft?md5=4839c565f8920eeea61c5ef01d5bb248&pid=1-s2.0-S2153353924000233-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141444222","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Eye tracking in digital pathology: A comprehensive literature review 数字病理学中的眼动仪:综合文献综述
Q2 Medicine Pub Date : 2024-05-17 DOI: 10.1016/j.jpi.2024.100383
Alana Lopes , Aaron D. Ward , Matthew Cecchini

Eye tracking has been used for decades in attempt to understand the cognitive processes of individuals. From memory access to problem-solving to decision-making, such insight has the potential to improve workflows and the education of students to become experts in relevant fields. Until recently, the traditional use of microscopes in pathology made eye tracking exceptionally difficult. However, the digital revolution of pathology from conventional microscopes to digital whole slide images allows for new research to be conducted and information to be learned with regards to pathologist visual search patterns and learning experiences. This has the promise to make pathology education more efficient and engaging, ultimately creating stronger and more proficient generations of pathologists to come. The goal of this review on eye tracking in pathology is to characterize and compare the visual search patterns of pathologists. The PubMed and Web of Science databases were searched using ‘pathology’ AND ‘eye tracking’ synonyms. A total of 22 relevant full-text articles published up to and including 2023 were identified and included in this review. Thematic analysis was conducted to organize each study into one or more of the 10 themes identified to characterize the visual search patterns of pathologists: (1) effect of experience, (2) fixations, (3) zooming, (4) panning, (5) saccades, (6) pupil diameter, (7) interpretation time, (8) strategies, (9) machine learning, and (10) education. Expert pathologists were found to have higher diagnostic accuracy, fewer fixations, and shorter interpretation times than pathologists with less experience. Further, literature on eye tracking in pathology indicates that there are several visual strategies for diagnostic interpretation of digital pathology images, but no evidence of a superior strategy exists. The educational implications of eye tracking in pathology have also been explored but the effect of teaching novices how to search as an expert remains unclear. In this article, the main challenges and prospects of eye tracking in pathology are briefly discussed along with their implications to the field.

几十年来,人们一直在使用眼动仪试图了解个人的认知过程。从获取记忆到解决问题再到决策,这种洞察力有可能改进工作流程和学生教育,使其成为相关领域的专家。直到最近,病理学中显微镜的传统使用还使得眼球跟踪异常困难。然而,病理学从传统显微镜到全玻片数字图像的数字化革命,使病理学家的视觉搜索模式和学习经验方面的新研究和新信息得以开展。这有望提高病理学教育的效率和吸引力,最终培养出更强大、更精通的新一代病理学家。这篇关于病理学眼动追踪的综述旨在描述和比较病理学家的视觉搜索模式。我们使用 "病理学 "和 "眼动仪 "同义词在 PubMed 和 Web of Science 数据库中进行了搜索。共找到 22 篇截至 2023 年(含 2023 年)发表的相关全文文章,并将其纳入本综述。我们对每篇研究进行了主题分析,将其归纳为 10 个主题中的一个或多个,以描述病理学家的视觉搜索模式:(1) 经验的影响;(2) 固定;(3) 缩放;(4) 平移;(5) 囊视;(6) 瞳孔直径;(7) 解释时间;(8) 策略;(9) 机器学习;(10) 教育。研究发现,与经验较少的病理学家相比,专家级病理学家的诊断准确率更高、定点次数更少、判读时间更短。此外,有关病理学眼动追踪的文献表明,有几种视觉策略可用于数字病理图像的诊断解读,但没有证据表明存在一种更优越的策略。也有人探讨了眼动追踪在病理学中的教育意义,但教导新手如何像专家一样进行搜索的效果仍不明确。本文简要讨论了眼动追踪技术在病理学领域的主要挑战和前景,以及对该领域的影响。
{"title":"Eye tracking in digital pathology: A comprehensive literature review","authors":"Alana Lopes ,&nbsp;Aaron D. Ward ,&nbsp;Matthew Cecchini","doi":"10.1016/j.jpi.2024.100383","DOIUrl":"10.1016/j.jpi.2024.100383","url":null,"abstract":"<div><p>Eye tracking has been used for decades in attempt to understand the cognitive processes of individuals. From memory access to problem-solving to decision-making, such insight has the potential to improve workflows and the education of students to become experts in relevant fields. Until recently, the traditional use of microscopes in pathology made eye tracking exceptionally difficult. However, the digital revolution of pathology from conventional microscopes to digital whole slide images allows for new research to be conducted and information to be learned with regards to pathologist visual search patterns and learning experiences. This has the promise to make pathology education more efficient and engaging, ultimately creating stronger and more proficient generations of pathologists to come. The goal of this review on eye tracking in pathology is to characterize and compare the visual search patterns of pathologists. The PubMed and Web of Science databases were searched using ‘pathology’ AND ‘eye tracking’ synonyms. A total of 22 relevant full-text articles published up to and including 2023 were identified and included in this review. Thematic analysis was conducted to organize each study into one or more of the 10 themes identified to characterize the visual search patterns of pathologists: (1) effect of experience, (2) fixations, (3) zooming, (4) panning, (5) saccades, (6) pupil diameter, (7) interpretation time, (8) strategies, (9) machine learning, and (10) education. Expert pathologists were found to have higher diagnostic accuracy, fewer fixations, and shorter interpretation times than pathologists with less experience. Further, literature on eye tracking in pathology indicates that there are several visual strategies for diagnostic interpretation of digital pathology images, but no evidence of a superior strategy exists. The educational implications of eye tracking in pathology have also been explored but the effect of teaching novices how to search as an expert remains unclear. In this article, the main challenges and prospects of eye tracking in pathology are briefly discussed along with their implications to the field.</p></div>","PeriodicalId":37769,"journal":{"name":"Journal of Pathology Informatics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2153353924000221/pdfft?md5=01458aa9d7539a3f8a155a98d18ad8ba&pid=1-s2.0-S2153353924000221-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141024415","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CDK: A novel high-performance transfer feature technique for early detection of osteoarthritis CDK:用于早期检测骨关节炎的新型高性能转移特征技术
Q2 Medicine Pub Date : 2024-05-08 DOI: 10.1016/j.jpi.2024.100382
Mohammad Shariful Islam , Mohammad Abu Tareq Rony

Knee osteoarthritis (OA) is a prevalent condition causing significant disability, particularly among the elderly, necessitating advancements in diagnostic methodologies to facilitate early detection and treatment. Traditional OA diagnosis, relying on radiography and physical exams, faces limitations in accuracy and objectivity. This underscores the need for more advanced diagnostic methods, such as machine learning (ML) and deep learning (DL), to improve OA detection and classification. This research introduces a novel ensemble learning approach for image data feature extraction which ingeniously combines the strengths of 2 advanced (ML) models with a (DL) method to substantially improve the accuracy of OA detection from radiographic images. This innovative strategy aims to address the limitations of traditional diagnostic tools by leveraging the enhanced sensitivity and specificity of combined ML and DL models. The methodology deployed in this study encompasses the application of 10 ML models to a comprehensive publicly available Kaggle dataset with a total of 3615 samples of knee X-ray images. Through rigorous k-fold cross-validation and meticulous hyperparameter optimization, we also included evaluation metrics like accuracy, receiver operating characteristic, precision, recall, and F1-score to assess our models' performance effectively. The proposed novel CDK (convolutional neural network, decision tree, K-nearest classifier) ensemble approach for feature extraction is designed to synergize the predictive capabilities of individual models, thereby significantly improving the detection accuracy of OA indicators within radiographic images. We applied several ML and DL approaches to the newly created feature set to evaluate performance. The CDK ensemble model outperformed state-of-the-art studies with a high-performance score of 99.72% accuracy. This remarkable achievement underscores the model's exceptional capability in the early detection of OA, highlighting its superiority in comparison to existing methods.

膝关节骨关节炎(OA)是一种导致严重残疾的常见疾病,尤其是在老年人中,因此有必要改进诊断方法,以便及早发现和治疗。传统的 OA 诊断依赖于射线照相和体格检查,在准确性和客观性方面存在局限性。这凸显了对更先进诊断方法的需求,如机器学习(ML)和深度学习(DL),以改善 OA 检测和分类。本研究介绍了一种用于图像数据特征提取的新型集合学习方法,该方法巧妙地结合了两种先进(ML)模型和一种(DL)方法的优势,从而大幅提高了从放射图像中检测 OA 的准确性。这一创新策略旨在利用 ML 和 DL 组合模型增强的灵敏度和特异性,解决传统诊断工具的局限性。本研究采用的方法包括将 10 个 ML 模型应用于一个全面的公开 Kaggle 数据集,该数据集共有 3615 个膝关节 X 光图像样本。通过严格的 k 倍交叉验证和细致的超参数优化,我们还纳入了准确率、接收者操作特征、精确度、召回率和 F1 分数等评价指标,以有效评估模型的性能。我们提出的新型 CDK(卷积神经网络、决策树、K-最近分类器)特征提取集合方法旨在协同各个模型的预测能力,从而显著提高放射影像中 OA 指标的检测准确率。我们对新创建的特征集采用了多种 ML 和 DL 方法来评估性能。CDK 组合模型的准确率高达 99.72%,超过了最先进的研究结果。这一骄人成绩彰显了该模型在早期检测 OA 方面的卓越能力,突出了它与现有方法相比的优越性。
{"title":"CDK: A novel high-performance transfer feature technique for early detection of osteoarthritis","authors":"Mohammad Shariful Islam ,&nbsp;Mohammad Abu Tareq Rony","doi":"10.1016/j.jpi.2024.100382","DOIUrl":"10.1016/j.jpi.2024.100382","url":null,"abstract":"<div><p>Knee osteoarthritis (OA) is a prevalent condition causing significant disability, particularly among the elderly, necessitating advancements in diagnostic methodologies to facilitate early detection and treatment. Traditional OA diagnosis, relying on radiography and physical exams, faces limitations in accuracy and objectivity. This underscores the need for more advanced diagnostic methods, such as machine learning (ML) and deep learning (DL), to improve OA detection and classification. This research introduces a novel ensemble learning approach for image data feature extraction which ingeniously combines the strengths of 2 advanced (ML) models with a (DL) method to substantially improve the accuracy of OA detection from radiographic images. This innovative strategy aims to address the limitations of traditional diagnostic tools by leveraging the enhanced sensitivity and specificity of combined ML and DL models. The methodology deployed in this study encompasses the application of 10 ML models to a comprehensive publicly available Kaggle dataset with a total of 3615 samples of knee X-ray images. Through rigorous k-fold cross-validation and meticulous hyperparameter optimization, we also included evaluation metrics like accuracy, receiver operating characteristic, precision, recall, and F1-score to assess our models' performance effectively. The proposed novel CDK (convolutional neural network, decision tree, K-nearest classifier) ensemble approach for feature extraction is designed to synergize the predictive capabilities of individual models, thereby significantly improving the detection accuracy of OA indicators within radiographic images. We applied several ML and DL approaches to the newly created feature set to evaluate performance. The CDK ensemble model outperformed state-of-the-art studies with a high-performance score of 99.72% accuracy. This remarkable achievement underscores the model's exceptional capability in the early detection of OA, highlighting its superiority in comparison to existing methods.</p></div>","PeriodicalId":37769,"journal":{"name":"Journal of Pathology Informatics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S215335392400021X/pdfft?md5=7758136ea8b6ff5e5d21a4755ae40c6d&pid=1-s2.0-S215335392400021X-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141028032","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A selective CutMix approach improves generalizability of deep learning-based grading and risk assessment of prostate cancer 选择性 CutMix 方法提高了基于深度学习的前列腺癌分级和风险评估的普适性
Q2 Medicine Pub Date : 2024-05-07 DOI: 10.1016/j.jpi.2024.100381
Sushant Patkar , Stephanie Harmon , Isabell Sesterhenn , Rosina Lis , Maria Merino , Denise Young , G. Thomas Brown , Kimberly M. Greenfield , John D. McGeeney , Sally Elsamanoudi , Shyh-Han Tan , Cara Schafer , Jiji Jiang , Gyorgy Petrovics , Albert Dobi , Francisco J. Rentas , Peter A. Pinto , Gregory T. Chesnut , Peter Choyke , Baris Turkbey , Joel T. Moncur

The Gleason score is an important predictor of prognosis in prostate cancer. However, its subjective nature can result in over- or under-grading. Our objective was to train an artificial intelligence (AI)-based algorithm to grade prostate cancer in specimens from patients who underwent radical prostatectomy (RP) and to assess the correlation of AI-estimated proportions of different Gleason patterns with biochemical recurrence-free survival (RFS), metastasis-free survival (MFS), and overall survival (OS). Training and validation of algorithms for cancer detection and grading were completed with three large datasets containing a total of 580 whole-mount prostate slides from 191 RP patients at two centers and 6218 annotated needle biopsy slides from the publicly available Prostate Cancer Grading Assessment dataset. A cancer detection model was trained using MobileNetV3 on 0.5 mm × 0.5 mm cancer areas (tiles) captured at 10× magnification. For cancer grading, a Gleason pattern detector was trained on tiles using a ResNet50 convolutional neural network and a selective CutMix training strategy involving a mixture of real and artificial examples. This strategy resulted in improved model generalizability in the test set compared with three different control experiments when evaluated on both needle biopsy slides and whole-mount prostate slides from different centers. In an additional test cohort of RP patients who were clinically followed over 30 years, quantitative Gleason pattern AI estimates achieved concordance indexes of 0.69, 0.72, and 0.64 for predicting RFS, MFS, and OS times, outperforming the control experiments and International Society of Urological Pathology system (ISUP) grading by pathologists. Finally, unsupervised clustering of test RP patient specimens into low-, medium-, and high-risk groups based on AI-estimated proportions of each Gleason pattern resulted in significantly improved RFS and MFS stratification compared with ISUP grading. In summary, deep learning-based quantitative Gleason scoring using a selective CutMix training strategy may improve prognostication after prostate cancer surgery.

格里森评分是预测前列腺癌预后的重要指标。然而,其主观性可能导致评分过高或过低。我们的目标是训练一种基于人工智能(AI)的算法,对接受根治性前列腺切除术(RP)患者标本中的前列腺癌进行分级,并评估人工智能估计的不同Gleason模式比例与无生化复发生存期(RFS)、无转移生存期(MFS)和总生存期(OS)之间的相关性。利用三个大型数据集完成了癌症检测和分级算法的训练和验证,这三个数据集包含来自两个中心 191 名前列腺癌患者的共 580 张全切前列腺切片,以及来自公开的前列腺癌分级评估数据集的 6218 张带注释的针刺活检切片。使用 MobileNetV3 对以 10 倍放大率捕获的 0.5 mm × 0.5 mm 癌症区域(瓦片)进行了癌症检测模型训练。在癌症分级方面,使用 ResNet50 卷积神经网络和选择性 CutMix 训练策略(包括真实和人工示例的混合)在瓷砖上训练格里森模式检测器。在对来自不同中心的针刺活检切片和全装前列腺切片进行评估时,与三个不同的对照实验相比,这种策略提高了模型在测试集中的通用性。在对临床随访超过 30 年的前列腺癌患者进行的另一个测试组中,定量格里森模式 AI 估计值在预测 RFS、MFS 和 OS 时间方面的一致性指数分别为 0.69、0.72 和 0.64,优于对照实验和国际泌尿病理学会病理学家分级系统(ISUP)。最后,与 ISUP 分级相比,根据人工智能估计的每种 Gleason 模式的比例将测试 RP 患者标本无监督聚类为低、中、高风险组,可显著改善 RFS 和 MFS 分层。总之,使用选择性 CutMix 训练策略进行基于深度学习的定量格里森评分可以改善前列腺癌手术后的预后。
{"title":"A selective CutMix approach improves generalizability of deep learning-based grading and risk assessment of prostate cancer","authors":"Sushant Patkar ,&nbsp;Stephanie Harmon ,&nbsp;Isabell Sesterhenn ,&nbsp;Rosina Lis ,&nbsp;Maria Merino ,&nbsp;Denise Young ,&nbsp;G. Thomas Brown ,&nbsp;Kimberly M. Greenfield ,&nbsp;John D. McGeeney ,&nbsp;Sally Elsamanoudi ,&nbsp;Shyh-Han Tan ,&nbsp;Cara Schafer ,&nbsp;Jiji Jiang ,&nbsp;Gyorgy Petrovics ,&nbsp;Albert Dobi ,&nbsp;Francisco J. Rentas ,&nbsp;Peter A. Pinto ,&nbsp;Gregory T. Chesnut ,&nbsp;Peter Choyke ,&nbsp;Baris Turkbey ,&nbsp;Joel T. Moncur","doi":"10.1016/j.jpi.2024.100381","DOIUrl":"10.1016/j.jpi.2024.100381","url":null,"abstract":"<div><p>The Gleason score is an important predictor of prognosis in prostate cancer. However, its subjective nature can result in over- or under-grading. Our objective was to train an artificial intelligence (AI)-based algorithm to grade prostate cancer in specimens from patients who underwent radical prostatectomy (RP) and to assess the correlation of AI-estimated proportions of different Gleason patterns with biochemical recurrence-free survival (RFS), metastasis-free survival (MFS), and overall survival (OS). Training and validation of algorithms for cancer detection and grading were completed with three large datasets containing a total of 580 whole-mount prostate slides from 191 RP patients at two centers and 6218 annotated needle biopsy slides from the publicly available Prostate Cancer Grading Assessment dataset. A cancer detection model was trained using MobileNetV3 on 0.5 mm × 0.5 mm cancer areas (tiles) captured at 10× magnification. For cancer grading, a Gleason pattern detector was trained on tiles using a ResNet50 convolutional neural network and a selective CutMix training strategy involving a mixture of real and artificial examples. This strategy resulted in improved model generalizability in the test set compared with three different control experiments when evaluated on both needle biopsy slides and whole-mount prostate slides from different centers. In an additional test cohort of RP patients who were clinically followed over 30 years, quantitative Gleason pattern AI estimates achieved concordance indexes of 0.69, 0.72, and 0.64 for predicting RFS, MFS, and OS times, outperforming the control experiments and International Society of Urological Pathology system (ISUP) grading by pathologists. Finally, unsupervised clustering of test RP patient specimens into low-, medium-, and high-risk groups based on AI-estimated proportions of each Gleason pattern resulted in significantly improved RFS and MFS stratification compared with ISUP grading. In summary, deep learning-based quantitative Gleason scoring using a selective CutMix training strategy may improve prognostication after prostate cancer surgery.</p></div>","PeriodicalId":37769,"journal":{"name":"Journal of Pathology Informatics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2153353924000208/pdfft?md5=3e256673a8f52212fb05001cc4e50f5b&pid=1-s2.0-S2153353924000208-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141042139","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Validation and three years of clinical experience in using an artificial intelligence algorithm as a second read system for prostate cancer diagnosis—real-world experience 将人工智能算法作为前列腺癌诊断的第二读取系统的验证和三年临床经验--真实世界的经验
Q2 Medicine Pub Date : 2024-04-30 DOI: 10.1016/j.jpi.2024.100378
Juan Carlos Santa-Rosario, Erik A. Gustafson, Dario E. Sanabria Bellassai, Phillip E. Gustafson, Mariano de Socarraz

Background

Prostate cancer ranks as the most frequently diagnosed cancer in men in the USA, with significant mortality rates. Early detection is pivotal for optimal patient outcomes, providing increased treatment options and potentially less invasive interventions. There remain significant challenges in prostate cancer histopathology, including the potential for missed diagnoses due to pathologist variability and subjective interpretations.

Methods

To address these challenges, this study investigates the ability of artificial intelligence (AI) to enhance diagnostic accuracy. The Galen™ Prostate AI algorithm was validated on a cohort of Puerto Rican men to demonstrate its efficacy in cancer detection and Gleason grading. Subsequently, the AI algorithm was integrated into routine clinical practice during a 3-year period at a CLIA certified precision pathology laboratory.

Results

The Galen™ Prostate AI algorithm showed a 96.7% (95% CI 95.6–97.8) specificity and a 96.6% (95% CI 93.3–98.8) sensitivity for prostate cancer detection and 82.1% specificity (95% CI 73.9–88.5) and 81.1% sensitivity (95% CI 73.7–87.2) for distinction of Gleason Grade Group 1 from Grade Group 2+. The subsequent AI integration into routine clinical use examined prostate cancer diagnoses on >122,000 slides and 9200 cases over 3 years and had an overall AI Impact ™ factor of 1.8%.

Conclusions

The potential of AI to be a powerful, reliable, and effective diagnostic tool for pathologists is highlighted, while the AI Impact™ in a real-world setting demonstrates the ability of AI to standardize prostate cancer diagnosis at a high level of performance across pathologists.

背景在美国,前列腺癌是最常见的男性癌症,死亡率很高。早期检测是获得最佳治疗效果的关键,它提供了更多的治疗选择,并可能减少侵入性干预。前列腺癌组织病理学仍面临重大挑战,包括由于病理学家的变异和主观解释可能导致漏诊。Galen™ 前列腺人工智能算法在一组波多黎各男性中进行了验证,以证明其在癌症检测和格里森分级方面的功效。结果 Galen™ 前列腺 AI 算法在前列腺癌检测方面的特异性为 96.7%(95% CI 95.6-97.8),灵敏度为 96.6%(95% CI 93.3-98.8);在区分格里森 1 级和 2+ 级方面,特异性为 82.1%(95% CI 73.9-88.5),灵敏度为 81.1%(95% CI 73.7-87.2)。结论人工智能作为病理学家强大、可靠和有效的诊断工具的潜力得到了强调,而在真实世界环境中的人工智能影响(AI Impact™)表明,人工智能有能力使病理学家的前列腺癌诊断达到高水平的标准化。
{"title":"Validation and three years of clinical experience in using an artificial intelligence algorithm as a second read system for prostate cancer diagnosis—real-world experience","authors":"Juan Carlos Santa-Rosario,&nbsp;Erik A. Gustafson,&nbsp;Dario E. Sanabria Bellassai,&nbsp;Phillip E. Gustafson,&nbsp;Mariano de Socarraz","doi":"10.1016/j.jpi.2024.100378","DOIUrl":"https://doi.org/10.1016/j.jpi.2024.100378","url":null,"abstract":"<div><h3>Background</h3><p>Prostate cancer ranks as the most frequently diagnosed cancer in men in the USA, with significant mortality rates. Early detection is pivotal for optimal patient outcomes, providing increased treatment options and potentially less invasive interventions. There remain significant challenges in prostate cancer histopathology, including the potential for missed diagnoses due to pathologist variability and subjective interpretations.</p></div><div><h3>Methods</h3><p>To address these challenges, this study investigates the ability of artificial intelligence (AI) to enhance diagnostic accuracy. The Galen™ Prostate AI algorithm was validated on a cohort of Puerto Rican men to demonstrate its efficacy in cancer detection and Gleason grading. Subsequently, the AI algorithm was integrated into routine clinical practice during a 3-year period at a CLIA certified precision pathology laboratory.</p></div><div><h3>Results</h3><p>The Galen™ Prostate AI algorithm showed a 96.7% (95% CI 95.6–97.8) specificity and a 96.6% (95% CI 93.3–98.8) sensitivity for prostate cancer detection and 82.1% specificity (95% CI 73.9–88.5) and 81.1% sensitivity (95% CI 73.7–87.2) for distinction of Gleason Grade Group 1 from Grade Group 2+. The subsequent AI integration into routine clinical use examined prostate cancer diagnoses on &gt;122,000 slides and 9200 cases over 3 years and had an overall AI Impact ™ factor of 1.8%.</p></div><div><h3>Conclusions</h3><p>The potential of AI to be a powerful, reliable, and effective diagnostic tool for pathologists is highlighted, while the AI Impact™ in a real-world setting demonstrates the ability of AI to standardize prostate cancer diagnosis at a high level of performance across pathologists.</p></div>","PeriodicalId":37769,"journal":{"name":"Journal of Pathology Informatics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-04-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2153353924000178/pdfft?md5=9c3ac8cee3a0de7c2dbd11239bf2bccb&pid=1-s2.0-S2153353924000178-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141164206","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Pathologists light level preferences using the microscope—study to guide digital pathology display use 病理学家对显微镜光照度的偏好--指导数字病理显示屏使用的研究
Q2 Medicine Pub Date : 2024-04-29 DOI: 10.1016/j.jpi.2024.100379
Charlotte Jennings , Darren Treanor , David Brettle

Background

Currently, there is a paucity of guidelines relating to displays used for digital pathology making procurement decisions, and optimal display configuration, challenging.

Experience suggests pathologists have personal preferences for brightness when using a conventional microscope which we hypothesized could be used as a predictor for display setup.

Methods

We conducted an online survey across six NHS hospitals, totalling 108 practicing pathologists, to capture brightness adjustment habits on both microscopes and displays.

A convenience subsample of respondents was then invited to take part in a practical task to determine microscope brightness and display luminance preferences in the normal working environment. A novel adaptation for a lightmeter was developed to directly measure the light output from the microscope eyepiece.

Results

The survey (response rate 59% n=64) indicates 81% of respondents adjust the brightness on their microscope. In comparison, only 11% report adjusting their digital display. Display adjustments were more likely to be for visual comfort and ambient light compensation rather than for tissue factors, common for microscope adjustments. Part of this discrepancy relates to lack of knowledge of how to adjust displays and lack of guidance on whether this is safe; But, 66% felt that the ability to adjust the light on the display was important.

Twenty consultants took part in the practical brightness assessment. Light preferences on the microscope showed no correlation with display preferences, except where a pathologist has a markedly brighter microscope light preference. All of the preferences in this cohort were for a display luminance of <500 cd/m2, with 90% preferring 350 cd/m2 or less. There was no correlation between these preferences and the ambient lighting in the room.

Conclusions

We conclude that microscope preferences can only be used to predict display luminance requirements where the microscope is being used at very high brightness levels. A display capable of a brightness of 500 cd/m2 should be suitable for almost all pathologists with 300 cd/m2 suitable for the majority. Although display luminance is not frequently changed by users, the ability to do so was felt to be important by the majority of respondents.

Further work needs to be undertaken to establish the relationship between diagnostic performance, luminance preferences, and ambient lighting levels.

背景目前,与数字病理学所用显示器相关的指南很少,这使得采购决定和最佳显示器配置具有挑战性。经验表明,病理学家在使用传统显微镜时对亮度有个人偏好,我们假设这可以用作显示器设置的预测因素。方法我们在六家英国国家医疗服务系统(NHS)医院开展了一项在线调查,共调查了 108 名执业病理学家,以了解他们对显微镜和显示屏亮度的调节习惯,然后邀请方便的受访者子样本参加一项实际任务,以确定他们在正常工作环境中对显微镜亮度和显示屏亮度的偏好。结果调查(回复率为 59% n=64)显示,81% 的受访者会调整显微镜的亮度。相比之下,只有 11% 的受访者表示调整过数字显示屏。显示屏的调整更可能是为了视觉舒适度和环境光补偿,而不是显微镜调整中常见的组织因素。造成这种差异的部分原因是对如何调节显示屏缺乏了解,以及缺乏关于这样做是否安全的指导;但是,66% 的人认为能够调节显示屏上的光线非常重要。对显微镜光线的偏好与对显示屏的偏好没有相关性,除非病理学家对显微镜光线有明显的偏好。该组研究人员的所有偏好都是显示亮度为 500 cd/m2,其中 90% 的人偏好 350 cd/m2 或更低的亮度。结论我们得出结论,只有在显微镜以极高亮度水平使用的情况下,显微镜偏好才可用于预测对显示亮度的要求。亮度为 500 cd/m2 的显示屏几乎适合所有病理学家,而 300 cd/m2 则适合大多数病理学家。虽然用户并不经常改变显示亮度,但大多数受访者认为改变亮度的能力非常重要。
{"title":"Pathologists light level preferences using the microscope—study to guide digital pathology display use","authors":"Charlotte Jennings ,&nbsp;Darren Treanor ,&nbsp;David Brettle","doi":"10.1016/j.jpi.2024.100379","DOIUrl":"https://doi.org/10.1016/j.jpi.2024.100379","url":null,"abstract":"<div><h3>Background</h3><p>Currently, there is a paucity of guidelines relating to displays used for digital pathology making procurement decisions, and optimal display configuration, challenging.</p><p>Experience suggests pathologists have personal preferences for brightness when using a conventional microscope which we hypothesized could be used as a predictor for display setup.</p></div><div><h3>Methods</h3><p>We conducted an online survey across six NHS hospitals, totalling 108 practicing pathologists, to capture brightness adjustment habits on both microscopes and displays.</p><p>A convenience subsample of respondents was then invited to take part in a practical task to determine microscope brightness and display luminance preferences in the normal working environment. A novel adaptation for a lightmeter was developed to directly measure the light output from the microscope eyepiece.</p></div><div><h3>Results</h3><p>The survey (response rate 59% <em>n</em>=64) indicates 81% of respondents adjust the brightness on their microscope. In comparison, only 11% report adjusting their digital display. Display adjustments were more likely to be for visual comfort and ambient light compensation rather than for tissue factors, common for microscope adjustments. Part of this discrepancy relates to lack of knowledge of how to adjust displays and lack of guidance on whether this is safe; But, 66% felt that the ability to adjust the light on the display was important.</p><p>Twenty consultants took part in the practical brightness assessment. Light preferences on the microscope showed no correlation with display preferences, except where a pathologist has a markedly brighter microscope light preference. All of the preferences in this cohort were for a display luminance of &lt;500 cd/m<sup>2</sup>, with 90% preferring 350 cd/m<sup>2</sup> or less. There was no correlation between these preferences and the ambient lighting in the room.</p></div><div><h3>Conclusions</h3><p>We conclude that microscope preferences can only be used to predict display luminance requirements where the microscope is being used at very high brightness levels. A display capable of a brightness of 500 cd/m<sup>2</sup> should be suitable for almost all pathologists with 300 cd/m<sup>2</sup> suitable for the majority. Although display luminance is not frequently changed by users, the ability to do so was felt to be important by the majority of respondents.</p><p>Further work needs to be undertaken to establish the relationship between diagnostic performance, luminance preferences, and ambient lighting levels.</p></div>","PeriodicalId":37769,"journal":{"name":"Journal of Pathology Informatics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-04-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S215335392400018X/pdfft?md5=0134b221667c45b419ce808d463b9b22&pid=1-s2.0-S215335392400018X-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141164205","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Pathology Informatics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1