Towards interactive AI-authoring with prototypical few-shot classifiers in histopathology

Petr Kuritcyn , Rosalie Kletzander , Sophia Eisenberg , Thomas Wittenberg , Volker Bruns , Katja Evert , Felix Keil , Paul K. Ziegler , Katrin Bankov , Peter Wild , Markus Eckstein , Arndt Hartmann , Carol I. Geppert , Michaela Benz
{"title":"Towards interactive AI-authoring with prototypical few-shot classifiers in histopathology","authors":"Petr Kuritcyn ,&nbsp;Rosalie Kletzander ,&nbsp;Sophia Eisenberg ,&nbsp;Thomas Wittenberg ,&nbsp;Volker Bruns ,&nbsp;Katja Evert ,&nbsp;Felix Keil ,&nbsp;Paul K. Ziegler ,&nbsp;Katrin Bankov ,&nbsp;Peter Wild ,&nbsp;Markus Eckstein ,&nbsp;Arndt Hartmann ,&nbsp;Carol I. Geppert ,&nbsp;Michaela Benz","doi":"10.1016/j.jpi.2024.100388","DOIUrl":null,"url":null,"abstract":"<div><p>A vast multitude of tasks in histopathology could potentially benefit from the support of artificial intelligence (AI). Many examples have been shown in the literature and first commercial products with FDA or CE-IVDR clearance are available. However, two key challenges remain: (1) a scarcity of thoroughly annotated images, respectively the laboriousness of this task, and (2) the creation of robust models that can cope with the data heterogeneity in the field (domain generalization). In this work, we investigate how the combination of prototypical few-shot classification models and data augmentation can address both of these challenges. Based on annotated data sets that include multiple centers, multiple scanners, and two tumor entities, we examine the robustness and the adaptability of few-shot classifiers in multiple scenarios. We demonstrate that data from one scanner and one site are sufficient to train robust few-shot classification models by applying domain-specific data augmentation. The models achieved classification performance of around 90% on a multiscanner and multicenter database, which is on par with the accuracy achieved on the primary single-center single-scanner data. Various convolutional neural network (CNN) architectures can be used for feature extraction in the few-shot model. A comparison of nine state-of-the-art architectures yielded that EfficientNet B0 provides the best trade-off between accuracy and inference time. The classification of prototypical few-shot models directly relies on class prototypes derived from example images of each class. Therefore, we investigated the influence of prototypes originating from images from different scanners and evaluated their performance also on the multiscanner database. Again, our few-shot model showed a stable performance with an average absolute deviation in accuracy compared to the primary prototypes of 1.8% points. Finally, we examined the adaptability to a new tumor entity: classification of tissue sections containing urothelial carcinoma into normal, tumor, and necrotic regions. Only three annotations per subclass (e.g., muscle and adipose tissue are subclasses of normal tissue) were provided to adapt the few-shot model, which obtained an overall accuracy of 93.6%. These results demonstrate that prototypical few-shot classification is an ideal technology for realizing an interactive AI authoring system as it only requires few annotations and can be adapted to new tasks without involving retraining of the underlying feature extraction CNN, which would in turn require a selection of hyper-parameters based on data science expert knowledge. Similarly, it can be regarded as a guided annotation system. To this end, we realized a workflow and user interface that targets non-technical users.</p></div>","PeriodicalId":37769,"journal":{"name":"Journal of Pathology Informatics","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-06-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2153353924000270/pdfft?md5=05adcd36f07ac4f905fe1929289c6160&pid=1-s2.0-S2153353924000270-main.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Pathology Informatics","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2153353924000270","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"Medicine","Score":null,"Total":0}
引用次数: 0

Abstract

A vast multitude of tasks in histopathology could potentially benefit from the support of artificial intelligence (AI). Many examples have been shown in the literature and first commercial products with FDA or CE-IVDR clearance are available. However, two key challenges remain: (1) a scarcity of thoroughly annotated images, respectively the laboriousness of this task, and (2) the creation of robust models that can cope with the data heterogeneity in the field (domain generalization). In this work, we investigate how the combination of prototypical few-shot classification models and data augmentation can address both of these challenges. Based on annotated data sets that include multiple centers, multiple scanners, and two tumor entities, we examine the robustness and the adaptability of few-shot classifiers in multiple scenarios. We demonstrate that data from one scanner and one site are sufficient to train robust few-shot classification models by applying domain-specific data augmentation. The models achieved classification performance of around 90% on a multiscanner and multicenter database, which is on par with the accuracy achieved on the primary single-center single-scanner data. Various convolutional neural network (CNN) architectures can be used for feature extraction in the few-shot model. A comparison of nine state-of-the-art architectures yielded that EfficientNet B0 provides the best trade-off between accuracy and inference time. The classification of prototypical few-shot models directly relies on class prototypes derived from example images of each class. Therefore, we investigated the influence of prototypes originating from images from different scanners and evaluated their performance also on the multiscanner database. Again, our few-shot model showed a stable performance with an average absolute deviation in accuracy compared to the primary prototypes of 1.8% points. Finally, we examined the adaptability to a new tumor entity: classification of tissue sections containing urothelial carcinoma into normal, tumor, and necrotic regions. Only three annotations per subclass (e.g., muscle and adipose tissue are subclasses of normal tissue) were provided to adapt the few-shot model, which obtained an overall accuracy of 93.6%. These results demonstrate that prototypical few-shot classification is an ideal technology for realizing an interactive AI authoring system as it only requires few annotations and can be adapted to new tasks without involving retraining of the underlying feature extraction CNN, which would in turn require a selection of hyper-parameters based on data science expert knowledge. Similarly, it can be regarded as a guided annotation system. To this end, we realized a workflow and user interface that targets non-technical users.

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
利用组织病理学中的原型少量分类器实现交互式人工智能创作
组织病理学中的大量任务都有可能受益于人工智能(AI)的支持。许多实例已见诸于文献,第一批获得美国食品及药物管理局(FDA)或 CE-IVDR 认证的商业产品也已问世。然而,两个关键挑战依然存在:(1) 缺乏全面注释的图像,这也是这项任务的艰巨性所在;(2) 如何创建稳健的模型,以应对该领域的数据异质性(领域泛化)。在这项工作中,我们研究了如何将原型少镜头分类模型与数据增强相结合来应对这两个挑战。基于包含多个中心、多台扫描仪和两个肿瘤实体的注释数据集,我们考察了少次分类器在多种情况下的鲁棒性和适应性。我们证明,通过应用特定领域的数据扩增,来自一台扫描仪和一个部位的数据足以训练出稳健的少次分类模型。这些模型在多扫描仪和多中心数据库中的分类性能达到了约 90%,与主要单中心单扫描仪数据的准确性相当。各种卷积神经网络(CNN)架构均可用于少数镜头模型的特征提取。对九种最先进的架构进行比较后发现,EfficientNet B0 在准确性和推理时间之间实现了最佳平衡。原型 few-shot 模型的分类直接依赖于从每个类别的示例图像中提取的类别原型。因此,我们研究了来自不同扫描仪图像的原型的影响,并在多扫描仪数据库中评估了它们的性能。同样,我们的少拍模型显示出稳定的性能,与主要原型相比,准确度的平均绝对偏差为 1.8%。最后,我们检验了对新肿瘤实体的适应性:将含有尿道癌的组织切片分为正常区、肿瘤区和坏死区。每个子类(例如,肌肉和脂肪组织是正常组织的子类)只提供了三个注释,以适应少拍模型,该模型的总体准确率为 93.6%。这些结果表明,原型少镜头分类法是实现交互式人工智能创作系统的理想技术,因为它只需要很少的注释,就能适应新的任务,而无需重新训练底层特征提取 CNN,这反过来又需要根据数据科学专家的知识来选择超参数。同样,它也可以被视为一种引导式注释系统。为此,我们实现了针对非技术用户的工作流程和用户界面。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Journal of Pathology Informatics
Journal of Pathology Informatics Medicine-Pathology and Forensic Medicine
CiteScore
3.70
自引率
0.00%
发文量
2
审稿时长
18 weeks
期刊介绍: The Journal of Pathology Informatics (JPI) is an open access peer-reviewed journal dedicated to the advancement of pathology informatics. This is the official journal of the Association for Pathology Informatics (API). The journal aims to publish broadly about pathology informatics and freely disseminate all articles worldwide. This journal is of interest to pathologists, informaticians, academics, researchers, health IT specialists, information officers, IT staff, vendors, and anyone with an interest in informatics. We encourage submissions from anyone with an interest in the field of pathology informatics. We publish all types of papers related to pathology informatics including original research articles, technical notes, reviews, viewpoints, commentaries, editorials, symposia, meeting abstracts, book reviews, and correspondence to the editors. All submissions are subject to rigorous peer review by the well-regarded editorial board and by expert referees in appropriate specialties.
期刊最新文献
Digital mapping of resected cancer specimens: The visual pathology report A precise machine learning model: Detecting cervical cancer using feature selection and explainable AI ViCE: An automated and quantitative program to assess intestinal tissue morphology Deep feature batch correction using ComBat for machine learning applications in computational pathology LVI-PathNet: Segmentation-classification pipeline for detection of lymphovascular invasion in whole slide images of lung adenocarcinoma
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1