ClickSAM: Fine-tuning Segment Anything Model using click prompts for ultrasound image segmentation.

Aimee Guo, Grace Fei, Hemanth Pasupuleti, Jing Wang
{"title":"ClickSAM: Fine-tuning Segment Anything Model using click prompts for ultrasound image segmentation.","authors":"Aimee Guo, Grace Fei, Hemanth Pasupuleti, Jing Wang","doi":"10.1117/12.3005879","DOIUrl":null,"url":null,"abstract":"<p><p>The newly released Segment Anything Model (SAM) is a popular tool used in image processing due to its superior segmentation accuracy, variety of input prompts, training capabilities, and efficient model design. However, its current model is trained on a diverse dataset not tailored to medical images, particularly ultrasound images. Ultrasound images tend to have a lot of noise, making it difficult to segment out important structures. In this project, we developed <i>ClickSAM</i>, which fine-tunes the Segment Anything Model using click prompts for ultrasound images. ClickSAM has two stages of training: the first stage is trained on single-click prompts centered in the ground-truth contours, and the second stage focuses on improving the model performance through additional positive and negative click prompts. By comparing the first stage's predictions to the ground-truth masks, true positive, false positive, and false negative segments are calculated. Positive clicks are generated using the true positive and false negative segments, and negative clicks are generated using the false positive segments. The Centroidal Voronoi Tessellation algorithm is then employed to collect positive and negative click prompts in each segment that are used to enhance the model performance during the second stage of training. With click-train methods, ClickSAM exhibits superior performance compared to other existing models for ultrasound image segmentation.</p>","PeriodicalId":74505,"journal":{"name":"Proceedings of SPIE--the International Society for Optical Engineering","volume":"12932 ","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11141328/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of SPIE--the International Society for Optical Engineering","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1117/12.3005879","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2024/4/1 0:00:00","PubModel":"Epub","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

The newly released Segment Anything Model (SAM) is a popular tool used in image processing due to its superior segmentation accuracy, variety of input prompts, training capabilities, and efficient model design. However, its current model is trained on a diverse dataset not tailored to medical images, particularly ultrasound images. Ultrasound images tend to have a lot of noise, making it difficult to segment out important structures. In this project, we developed ClickSAM, which fine-tunes the Segment Anything Model using click prompts for ultrasound images. ClickSAM has two stages of training: the first stage is trained on single-click prompts centered in the ground-truth contours, and the second stage focuses on improving the model performance through additional positive and negative click prompts. By comparing the first stage's predictions to the ground-truth masks, true positive, false positive, and false negative segments are calculated. Positive clicks are generated using the true positive and false negative segments, and negative clicks are generated using the false positive segments. The Centroidal Voronoi Tessellation algorithm is then employed to collect positive and negative click prompts in each segment that are used to enhance the model performance during the second stage of training. With click-train methods, ClickSAM exhibits superior performance compared to other existing models for ultrasound image segmentation.

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
ClickSAM:利用点击提示微调超声波图像分割模型。
新发布的 "任意分割模型"(Segment Anything Model,SAM)因其卓越的分割准确性、多种输入提示、训练能力和高效的模型设计而成为图像处理领域的常用工具。然而,它目前的模型是在一个多样化的数据集上训练的,并不适合医学图像,尤其是超声波图像。超声波图像往往有很多噪声,因此很难分割出重要的结构。在这个项目中,我们开发了 ClickSAM,利用超声波图像的点击提示对 "任意分割模型 "进行微调。ClickSAM 的训练分为两个阶段:第一阶段以地面实况轮廓为中心,通过单击提示进行训练;第二阶段则通过额外的正负单击提示来提高模型性能。通过将第一阶段的预测与地面实况掩码进行比较,计算出真阳性、假阳性和假阴性段。使用真阳性和假阴性片段生成阳性点击,使用假阳性片段生成阴性点击。然后采用中心 Voronoi Tessellation 算法收集每个片段中的正负点击提示,用于在第二阶段训练中提高模型性能。通过点击训练方法,ClickSAM 与其他现有的超声图像分割模型相比,表现出更优越的性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
CiteScore
0.50
自引率
0.00%
发文量
0
期刊最新文献
Automated multi-lesion annotation in chest X-rays: annotating over 450,000 images from public datasets using the AI-based Smart Imagery Framing and Truthing (SIFT) system. High-Fidelity 3D Reconstruction for Accurate Anatomical Measurements in Endoscopic Sinus Surgery. Optimizing parylene and photoconductor thickness in indirect conversion amorphous selenium detectors. Intra- and inter-scanner CT variability and their impact on diagnostic tasks. Quantitative Accuracy of CT Protocols for Cross-sectional and Longitudinal Assessment of COPD: A Virtual Imaging Study.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1