AVP-AP: Self-Supervised Automatic View Positioning in 3D Cardiac CT via Atlas Prompting

Xiaolin Fan;Yan Wang;Yingying Zhang;Mingkun Bao;Bosen Jia;Dong Lu;Yifan Gu;Jian Cheng;Haogang Zhu
{"title":"AVP-AP: Self-Supervised Automatic View Positioning in 3D Cardiac CT via Atlas Prompting","authors":"Xiaolin Fan;Yan Wang;Yingying Zhang;Mingkun Bao;Bosen Jia;Dong Lu;Yifan Gu;Jian Cheng;Haogang Zhu","doi":"10.1109/TMI.2025.3554785","DOIUrl":null,"url":null,"abstract":"Automatic view positioning is crucial for cardiac computed tomography (CT) examinations, including disease diagnosis and surgical planning. However, it is highly challenging due to individual variability and large 3D search space. Existing work needs labor-intensive and time-consuming manual annotations to train view-specific models, which are limited to predicting only a fixed set of planes. However, in real clinical scenarios, the challenge of positioning semantic 2D slices with any orientation into varying coordinate space in arbitrary 3D volume remains unsolved. We thus introduce a novel framework, AVP-AP, the first to use Atlas Prompting for self-supervised Automatic View Positioning in the 3D CT volume. Specifically, this paper first proposes an atlas prompting method, which generates a 3D canonical atlas and trains a network to map slices into their corresponding positions in the atlas space via a self-supervised manner. Then, guided by atlas prompts corresponding to the given query images in a reference CT, we identify the coarse positions of slices in the target CT volume using rigid transformation between the 3D atlas and target CT volume, effectively reducing the search space. Finally, we refine the coarse positions by maximizing the similarity between the predicted slices and the query images in the feature space of a given foundation model. Our framework is flexible and efficient compared to other methods, outperforming other methods by 19.8% average structural similarity (SSIM) in arbitrary view positioning and achieving 9% SSIM in two-chamber view compared to four radiologists. Meanwhile, experiments on a public dataset validate our framework’s generalizability.","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"44 7","pages":"2921-2932"},"PeriodicalIF":0.0000,"publicationDate":"2025-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on medical imaging","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10942463/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Automatic view positioning is crucial for cardiac computed tomography (CT) examinations, including disease diagnosis and surgical planning. However, it is highly challenging due to individual variability and large 3D search space. Existing work needs labor-intensive and time-consuming manual annotations to train view-specific models, which are limited to predicting only a fixed set of planes. However, in real clinical scenarios, the challenge of positioning semantic 2D slices with any orientation into varying coordinate space in arbitrary 3D volume remains unsolved. We thus introduce a novel framework, AVP-AP, the first to use Atlas Prompting for self-supervised Automatic View Positioning in the 3D CT volume. Specifically, this paper first proposes an atlas prompting method, which generates a 3D canonical atlas and trains a network to map slices into their corresponding positions in the atlas space via a self-supervised manner. Then, guided by atlas prompts corresponding to the given query images in a reference CT, we identify the coarse positions of slices in the target CT volume using rigid transformation between the 3D atlas and target CT volume, effectively reducing the search space. Finally, we refine the coarse positions by maximizing the similarity between the predicted slices and the query images in the feature space of a given foundation model. Our framework is flexible and efficient compared to other methods, outperforming other methods by 19.8% average structural similarity (SSIM) in arbitrary view positioning and achieving 9% SSIM in two-chamber view compared to four radiologists. Meanwhile, experiments on a public dataset validate our framework’s generalizability.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
AVP-AP:基于Atlas提示的三维心脏CT自监督自动视图定位
自动视角定位对于心脏CT检查至关重要,包括疾病诊断和手术计划。然而,由于个体的可变性和大的3D搜索空间,这是极具挑战性的。现有的工作需要耗费大量劳动和时间的手工注释来训练特定于视图的模型,这些模型仅限于预测一组固定的平面。然而,在实际的临床场景中,将任意方向的语义二维切片定位到任意三维体的不同坐标空间的挑战仍然没有得到解决。因此,我们引入了一个新的框架,AVP-AP,这是第一个在3D CT体中使用Atlas提示进行自我监督自动视图定位的框架。具体而言,本文首先提出了一种地图集提示方法,该方法生成三维规范地图集,并训练网络以自监督的方式将切片映射到地图集空间中的相应位置。然后,在参考CT中给定查询图像对应的图集提示符的指导下,利用三维图集与目标CT体积之间的刚性变换,识别出目标CT体积中切片的粗糙位置,有效地缩小了搜索空间。最后,我们通过在给定基础模型的特征空间中最大化预测切片与查询图像之间的相似性来细化粗糙位置。与其他方法相比,我们的框架灵活高效,在任意视图定位中比其他方法高出19.8%的平均结构相似性(SSIM),在双腔视图中与四位放射科医生相比达到9%的SSIM。同时,在公共数据集上的实验验证了我们的框架的泛化性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Disentangled Multi-modal Learning of Histology and Transcriptomics for Cancer Characterization. Tomographic Foundation Model-FORCE: Flow-Oriented Reconstruction Conditioning Engine. Decouple, Reorganize, and Fuse: A Multimodal Framework for Cancer Survival Prediction. MARVEL: Motion-Aware Reconstruction Via Embedded Learning of Motion Prior for Time-Resolved Cardiac CT. QuPaS: SAM-based Semi-supervised Histopathological Image Segmentation with Quantum Force Field Finetuning and Adversarial Estimation.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1