Disease-Informed Adaptation of Vision-Language Models

Jiajin Zhang;Ge Wang;Mannudeep K. Kalra;Pingkun Yan
{"title":"Disease-Informed Adaptation of Vision-Language Models","authors":"Jiajin Zhang;Ge Wang;Mannudeep K. Kalra;Pingkun Yan","doi":"10.1109/TMI.2024.3484294","DOIUrl":null,"url":null,"abstract":"Expertise scarcity and high cost of data annotation hinder the development of artificial intelligence (AI) foundation models for medical image analysis. Transfer learning provides a way to utilize the off-the-shelf foundation models to address the clinical challenges. However, such models encounter difficulties when adapting to new diseases not presented in their original pre-training datasets. Compounding this challenge is the limited availability of example cases for a new disease, which further leads to the poor performance of the existing transfer learning techniques. This paper proposes a novel method for transfer learning of foundation Vision-Language Models (VLMs) to efficiently adapt them to a new disease with only a few examples. Such an effective adaptation of VLMs hinges on learning the nuanced representation of new disease concepts. By capitalizing on the joint visual-linguistic capabilities of VLMs, we introduce disease-informed contextual prompting in a novel disease prototype learning framework, which enables VLMs to quickly grasp the concept of the new disease, even with limited data. Extensive experiments across multiple pre-trained medical VLMs and multiple tasks showcase the notable enhancements in performance compared to other existing adaptation techniques. The code will be made publicly available at <uri>https://github.com/RPIDIAL/Disease-informed-VLM-Adaptation</uri>.","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"44 10","pages":"3984-3996"},"PeriodicalIF":0.0000,"publicationDate":"2024-10-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on medical imaging","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10723745/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Expertise scarcity and high cost of data annotation hinder the development of artificial intelligence (AI) foundation models for medical image analysis. Transfer learning provides a way to utilize the off-the-shelf foundation models to address the clinical challenges. However, such models encounter difficulties when adapting to new diseases not presented in their original pre-training datasets. Compounding this challenge is the limited availability of example cases for a new disease, which further leads to the poor performance of the existing transfer learning techniques. This paper proposes a novel method for transfer learning of foundation Vision-Language Models (VLMs) to efficiently adapt them to a new disease with only a few examples. Such an effective adaptation of VLMs hinges on learning the nuanced representation of new disease concepts. By capitalizing on the joint visual-linguistic capabilities of VLMs, we introduce disease-informed contextual prompting in a novel disease prototype learning framework, which enables VLMs to quickly grasp the concept of the new disease, even with limited data. Extensive experiments across multiple pre-trained medical VLMs and multiple tasks showcase the notable enhancements in performance compared to other existing adaptation techniques. The code will be made publicly available at https://github.com/RPIDIAL/Disease-informed-VLM-Adaptation.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
根据疾病调整视觉语言模型
医学图像分析人工智能(AI)基础模型的发展受到专业知识稀缺和数据标注成本高的制约。迁移学习提供了一种利用现成的基础模型来解决临床挑战的方法。然而,这些模型在适应原始预训练数据集中没有出现的新疾病时遇到困难。使这一挑战更加复杂的是,一种新疾病的病例可用性有限,这进一步导致现有迁移学习技术的性能不佳。本文提出了一种新的迁移学习方法,使基础视觉语言模型(VLMs)在实例较少的情况下能够有效地适应新的疾病。如此有效地适应VLMs取决于学习新疾病概念的细微表征。通过利用vlm的联合视觉语言能力,我们在一个新的疾病原型学习框架中引入了疾病信息上下文提示,这使得vlm即使在数据有限的情况下也能快速掌握新疾病的概念。在多个预训练的医疗vlm和多个任务上进行的大量实验表明,与其他现有的自适应技术相比,该方法在性能上有显著提高。该代码将在https://github.com/RPIDIAL/Disease-informed-VLM-Adaptation上公开发布。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
PWD: Prior-aware Wavelet Diffusion for Efficient Dental Limited-angle CT Reconstruction. Semantic Augmentation Variational Autoencoder for Unsupervised Anomaly Detection in Retinal OCT Images. Biomechanics-informed Non-rigid Medical Image Registration with Elasticity Theories. Universal Scale Transformer for Histology Image Segmentation. MambaMatch: A Novel Model for Semi-Supervised Spinal Cortical and Cancellous Bone Segmentation.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1