Haobo Chen, Yehua Cai, Changyan Wang, Lin Chen, Bo Zhang, Hong Han, Yuqing Guo, Hong Ding, Qi Zhang
{"title":"Multi-Organ Foundation Model for Universal Ultrasound Image Segmentation with Task Prompt and Anatomical Prior.","authors":"Haobo Chen, Yehua Cai, Changyan Wang, Lin Chen, Bo Zhang, Hong Han, Yuqing Guo, Hong Ding, Qi Zhang","doi":"10.1109/TMI.2024.3472672","DOIUrl":null,"url":null,"abstract":"<p><p>Semantic segmentation of ultrasound (US) images with deep learning has played a crucial role in computer-aided disease screening, diagnosis and prognosis. However, due to the scarcity of US images and small field of view, resulting segmentation models are tailored for a specific single organ and may lack robustness, overlooking correlations among anatomical structures of multiple organs. To address these challenges, we propose the Multi-Organ FOundation (MOFO) model for universal US image segmentation. The MOFO is optimized jointly from multiple organs across various anatomical regions to overcome the data scarcity and explore correlations between multiple organs. The MOFO extracts organ-invariant representations from US images. Simultaneously, the task prompt is employed to refine organ-specific representations for segmentation predictions. Moreover, the anatomical prior is incorporated to enhance the consistency of the anatomical structures. A multi-organ US database, comprising 7039 images from 10 organs across various regions of the human body, has been established to evaluate our model. Results demonstrate that the MOFO outperforms single-organ methods in terms of the Dice coefficient, 95% Hausdorff distance and average symmetric surface distance with statistically sufficient margins. Our experiments in multi-organ universal segmentation for US images serve as a pioneering exploration of improving segmentation performance by leveraging semantic and anatomical relationships within US images of multiple organs.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on medical imaging","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/TMI.2024.3472672","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Semantic segmentation of ultrasound (US) images with deep learning has played a crucial role in computer-aided disease screening, diagnosis and prognosis. However, due to the scarcity of US images and small field of view, resulting segmentation models are tailored for a specific single organ and may lack robustness, overlooking correlations among anatomical structures of multiple organs. To address these challenges, we propose the Multi-Organ FOundation (MOFO) model for universal US image segmentation. The MOFO is optimized jointly from multiple organs across various anatomical regions to overcome the data scarcity and explore correlations between multiple organs. The MOFO extracts organ-invariant representations from US images. Simultaneously, the task prompt is employed to refine organ-specific representations for segmentation predictions. Moreover, the anatomical prior is incorporated to enhance the consistency of the anatomical structures. A multi-organ US database, comprising 7039 images from 10 organs across various regions of the human body, has been established to evaluate our model. Results demonstrate that the MOFO outperforms single-organ methods in terms of the Dice coefficient, 95% Hausdorff distance and average symmetric surface distance with statistically sufficient margins. Our experiments in multi-organ universal segmentation for US images serve as a pioneering exploration of improving segmentation performance by leveraging semantic and anatomical relationships within US images of multiple organs.
利用深度学习对超声波(US)图像进行语义分割在计算机辅助疾病筛查、诊断和预后方面发挥了至关重要的作用。然而,由于 US 图像的稀缺性和小视场,由此产生的分割模型都是为特定的单一器官量身定制的,可能缺乏鲁棒性,忽略了多个器官解剖结构之间的相关性。为了应对这些挑战,我们提出了用于通用 US 图像分割的多器官基金化(MOFO)模型。MOFO 从不同解剖区域的多个器官中联合优化,以克服数据稀缺性并探索多个器官之间的相关性。MOFO 可从 US 图像中提取与器官无关的表征。同时,利用任务提示来完善特定器官的表征,以进行分割预测。此外,还纳入了解剖先验,以增强解剖结构的一致性。为了评估我们的模型,我们建立了一个多器官 US 数据库,其中包括来自人体不同区域 10 个器官的 7039 幅图像。结果表明,MOFO 在 Dice 系数、95% Hausdorff 距离和平均对称面距离方面均优于单器官方法,且在统计学上有足够的优势。我们的 US 图像多器官通用分割实验是利用 US 图像中多个器官的语义和解剖关系提高分割性能的开创性探索。