首页 > 最新文献

IEEE transactions on medical imaging最新文献

英文 中文
Few-Shot Contrastive Learning for Cross-Task Stroke Prognosis Prediction with Multimodal Data. 基于多模态数据的跨任务脑卒中预后预测的短时对比学习。
Pub Date : 2026-02-06 DOI: 10.1109/TMI.2026.3661971
Haoran Peng, Yuxiang Dai, Rencheng Zheng, Mingming Wang, Chengyan Wang, Yu Luo, He Wang

Predicting stroke outcome remains challenging due to inherent heterogeneity, misalignment of multimodal clinical data, and the availability of well-annotated longitudinal datasets. Current methodologies often lack robustness and generalizability across these tasks. We propose a few-shot contrastive learning framework that integrates brain MRI images and structured clinical records for cross-task prognosis prediction, addressing both morphological and functional outcomes. Our method combines Model-Agnostic Meta-Learning (MAML) with a two-step contrastive learning strategy including self-awareness learning that captures task-specific features and domain learning that facilitates cross-dataset generalization. To handle inconsistencies in tabular data, a Misalignment Separation technique was adopted. The framework jointly trains a domain encoder on multimodal inputs, capturing shared and task-specific prior knowledge to enhance predictive robustness. Evaluations on 309 patients for morphological outcome and 341 patients for functional outcome, as well as on external validation datasets, demonstrated that our approach outperformed SimCLR and conventional supervised methods, and could effectively integrate cross-task datasets. This framework highlights the potential of multimodal few-shot learning for robust stroke prognosis prediction for small-sample datasets.

由于固有的异质性、多模式临床数据的不一致以及有良好注释的纵向数据集的可用性,预测卒中结局仍然具有挑战性。当前的方法在这些任务中往往缺乏健壮性和通用性。我们提出了一个整合脑MRI图像和结构化临床记录的几次对比学习框架,用于跨任务预后预测,解决形态和功能结果。我们的方法将模型不可知元学习(MAML)与两步对比学习策略相结合,包括捕获任务特定特征的自我意识学习和促进跨数据集泛化的领域学习。为了处理表格数据的不一致性,采用了Misalignment Separation技术。该框架在多模态输入上联合训练域编码器,捕获共享和任务特定的先验知识,以增强预测鲁棒性。对309例形态学结果和341例功能结果以及外部验证数据集的评估表明,我们的方法优于SimCLR和传统的监督方法,并且可以有效地整合跨任务数据集。该框架强调了多模态少镜头学习对小样本数据集的鲁棒脑卒中预后预测的潜力。
{"title":"Few-Shot Contrastive Learning for Cross-Task Stroke Prognosis Prediction with Multimodal Data.","authors":"Haoran Peng, Yuxiang Dai, Rencheng Zheng, Mingming Wang, Chengyan Wang, Yu Luo, He Wang","doi":"10.1109/TMI.2026.3661971","DOIUrl":"https://doi.org/10.1109/TMI.2026.3661971","url":null,"abstract":"<p><p>Predicting stroke outcome remains challenging due to inherent heterogeneity, misalignment of multimodal clinical data, and the availability of well-annotated longitudinal datasets. Current methodologies often lack robustness and generalizability across these tasks. We propose a few-shot contrastive learning framework that integrates brain MRI images and structured clinical records for cross-task prognosis prediction, addressing both morphological and functional outcomes. Our method combines Model-Agnostic Meta-Learning (MAML) with a two-step contrastive learning strategy including self-awareness learning that captures task-specific features and domain learning that facilitates cross-dataset generalization. To handle inconsistencies in tabular data, a Misalignment Separation technique was adopted. The framework jointly trains a domain encoder on multimodal inputs, capturing shared and task-specific prior knowledge to enhance predictive robustness. Evaluations on 309 patients for morphological outcome and 341 patients for functional outcome, as well as on external validation datasets, demonstrated that our approach outperformed SimCLR and conventional supervised methods, and could effectively integrate cross-task datasets. This framework highlights the potential of multimodal few-shot learning for robust stroke prognosis prediction for small-sample datasets.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146133805","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AnatoDiff: Synthesizing Anatomically Truthful Radiographs With Limited Training Images. AnatoDiff:用有限的训练图像合成解剖学真实的x线照片。
Pub Date : 2026-02-06 DOI: 10.1109/TMI.2026.3661433
Ka-Wai Yung, Jayaram Sivaraj, Lodovico di Giura, Simon Eaton, Paolo De Coppi, Danail Stoyanov, Stavros Loukogeorgakis, Evangelos B Mazomenos

Rapid advancements in diffusion models have enabled synthesis of realistic and anonymized imagery in radiography. However, due to their complexity, these models typically require large training volumes, often exceeding 10,000 images. Pre-training on natural images can partly mitigate this issue, but often fails to generate anatomically accurate shapes due to the significant domain gap. This prohibits applications in specialized medical conditions with limited data. We propose AnatoDiff, a diffusion model synthesizing high-quality X-Ray images with accurate anatomical shapes using only 500 to 1,000 training samples. AnatoDiff incorporates a Shape Prototype Module and Anatomical Fidelity loss, allowing for smaller training volumes through targeted supervision. We extensively validate AnatoDiff across three open-source datasets from distinct anatomical regions: Neonatal Abdomen (1,000 images); Adult Chest (500 images); and Humerus (500 images). Results demonstrate significant benefits, with an average improvement of 14.9% in Fréchet Inception Distance, 9.7% in Improved Precision, and 2.3% in Improved Recall compared to state-of-the-art (SOTA) few-shot and data-limited natural image synthesis methods. Unlike other models, AnatoDiff consistently generates anatomically correct images with accurate shapes. Additionally, a ResNet-50 classifier trained on AnatoDiff-generated images shows a 2.1% to 5.3% increase in F1-score, compared to being trained on SOTA diffusion images, across 500 to 10,000 samples. A survey with 10 medical professionals reveals that images generated by AnatoDiff are challenging to distinguish from real ones, with a Matthews correlation coefficient of 0.277 and Fleiss' Kappa of 0.126, highlighting the effectiveness of AnatoDiff in generating high-quality, anatomically accurate radiographs. Our code is available at https://github.com/KawaiYung/AnatoDiff.

扩散模型的快速发展使放射照相中真实和匿名图像的合成成为可能。然而,由于它们的复杂性,这些模型通常需要大量的训练量,通常超过10,000张图像。自然图像的预训练可以部分缓解这一问题,但由于显着的域间隙,通常无法生成解剖学上准确的形状。这禁止在数据有限的特殊医疗条件下应用。我们提出AnatoDiff,一个仅使用500到1000个训练样本就能合成具有精确解剖形状的高质量x射线图像的扩散模型。AnatoDiff结合了形状原型模块和解剖保真度损失,允许通过有针对性的监督更小的训练量。我们在三个来自不同解剖区域的开源数据集上广泛验证AnatoDiff:新生儿腹部(1000张图像);成人胸部(500张);和肱骨(500张图像)。结果显示了显著的优势,与最先进的(SOTA)少量拍摄和数据有限的自然图像合成方法相比,fr起始距离平均提高14.9%,精度提高9.7%,召回率提高2.3%。与其他模型不同,AnatoDiff始终生成具有准确形状的解剖学正确图像。此外,在anatodiff生成的图像上训练的ResNet-50分类器与在SOTA扩散图像上训练的500至10,000个样本相比,f1分数提高了2.1%至5.3%。一项针对10名医学专业人员的调查显示,AnatoDiff生成的图像很难与真实图像区分开来,Matthews相关系数为0.277,Fleiss Kappa为0.126,突出了AnatoDiff在生成高质量、解剖准确的x线照片方面的有效性。我们的代码可在https://github.com/KawaiYung/AnatoDiff上获得。
{"title":"AnatoDiff: Synthesizing Anatomically Truthful Radiographs With Limited Training Images.","authors":"Ka-Wai Yung, Jayaram Sivaraj, Lodovico di Giura, Simon Eaton, Paolo De Coppi, Danail Stoyanov, Stavros Loukogeorgakis, Evangelos B Mazomenos","doi":"10.1109/TMI.2026.3661433","DOIUrl":"https://doi.org/10.1109/TMI.2026.3661433","url":null,"abstract":"<p><p>Rapid advancements in diffusion models have enabled synthesis of realistic and anonymized imagery in radiography. However, due to their complexity, these models typically require large training volumes, often exceeding 10,000 images. Pre-training on natural images can partly mitigate this issue, but often fails to generate anatomically accurate shapes due to the significant domain gap. This prohibits applications in specialized medical conditions with limited data. We propose AnatoDiff, a diffusion model synthesizing high-quality X-Ray images with accurate anatomical shapes using only 500 to 1,000 training samples. AnatoDiff incorporates a Shape Prototype Module and Anatomical Fidelity loss, allowing for smaller training volumes through targeted supervision. We extensively validate AnatoDiff across three open-source datasets from distinct anatomical regions: Neonatal Abdomen (1,000 images); Adult Chest (500 images); and Humerus (500 images). Results demonstrate significant benefits, with an average improvement of 14.9% in Fréchet Inception Distance, 9.7% in Improved Precision, and 2.3% in Improved Recall compared to state-of-the-art (SOTA) few-shot and data-limited natural image synthesis methods. Unlike other models, AnatoDiff consistently generates anatomically correct images with accurate shapes. Additionally, a ResNet-50 classifier trained on AnatoDiff-generated images shows a 2.1% to 5.3% increase in F1-score, compared to being trained on SOTA diffusion images, across 500 to 10,000 samples. A survey with 10 medical professionals reveals that images generated by AnatoDiff are challenging to distinguish from real ones, with a Matthews correlation coefficient of 0.277 and Fleiss' Kappa of 0.126, highlighting the effectiveness of AnatoDiff in generating high-quality, anatomically accurate radiographs. Our code is available at https://github.com/KawaiYung/AnatoDiff.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146133845","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Few-Shot Pulmonary Vessel Segmentation based on Tubular-Aware Prompt-Tuning. 基于小管感知提示调谐的肺血管分割。
Pub Date : 2026-02-06 DOI: 10.1109/TMI.2026.3662001
Zijian Gao, Lai Jiang, Yichen Guo, Sukun Tian, Yuchun Sun, Mai Xu, Liyuan Tao

Segmentation of the pulmonary vessel from computed tomography (CT) images plays a crucial role in the diagnosis and treatment of various lung diseases. Although deep learning-based approaches have shown remarkable progress in recent years, their performance is often hindered by the lack of high-quality annotated datasets, in which the complex anatomy and morphology of pulmonary vessels make manual annotation challenging, time-consuming, and prone to errors. To address this, we propose PV25, the first dataset that features finely paired annotations of both pulmonary vessels and airways. Moreover, we propose TPNet, a novel tubular-aware prompt-tuning framework for pulmonary vessel segmentation under few-shot training with limited annotations. Specifically, based on an advanced and frozen segmentation backbone, TPNet proposes tunable encoding and decoding networks that learn tubular structures as transfer learning priors, bridging the gap between the source and target pulmonary vessel domains. Specifically, TPNet is built in an encoder-decoder manner, including the fixed segmentation backbone, tunable encoding and decoding networks. In encoding stage, the Morphology-Driven Region Growing (MDRG) module is developed to leverage the tubular connectivity of vessels to guide the network in capturing fine-grained features of pulmonary vessels. In decoding stage, the Cross-Correlation Guidance (CCG) module is introduced to integrate multi-scale correlations between airway and vessel structures in a coarse-to-fine manner. Extensive experiments conducted on multiple datasets demonstrate that TPNet achieves state-of-the-art performance in pulmonary vessel segmentation under limited training data. Besides, TPNet shows strong performance in related tasks such as airway segmentation and artery-vein classification, highlighting its robustness and versatility.

计算机断层扫描(CT)图像的肺血管分割在各种肺部疾病的诊断和治疗中起着至关重要的作用。尽管基于深度学习的方法近年来取得了显著的进展,但它们的性能往往受到缺乏高质量注释数据集的阻碍,其中肺血管的复杂解剖和形态使得手动注释具有挑战性,耗时且容易出错。为了解决这个问题,我们提出了PV25,这是第一个具有肺血管和气道精细配对注释的数据集。此外,我们提出了一种新的小管感知快速调整框架TPNet,用于在有限注释的情况下进行肺血管分割。具体而言,TPNet基于先进的冷冻分割主干,提出了可调的编码和解码网络,将学习管状结构作为迁移学习先验,弥合源和目标肺血管域之间的差距。具体来说,TPNet以编码器-解码器的方式构建,包括固定分割主干网、可调编码和解码网络。在编码阶段,我们开发了形态学驱动区域生长(MDRG)模块,利用血管的管状连通性来引导网络捕获肺血管的细粒度特征。在解码阶段,引入了相关导引(Cross-Correlation Guidance, CCG)模块,以粗到细的方式整合气道和血管结构之间的多尺度相关性。在多个数据集上进行的大量实验表明,TPNet在有限的训练数据下实现了最先进的肺血管分割性能。此外,TPNet在气道分割和动静脉分类等相关任务中表现出较强的性能,突出了其鲁棒性和通用性。
{"title":"Few-Shot Pulmonary Vessel Segmentation based on Tubular-Aware Prompt-Tuning.","authors":"Zijian Gao, Lai Jiang, Yichen Guo, Sukun Tian, Yuchun Sun, Mai Xu, Liyuan Tao","doi":"10.1109/TMI.2026.3662001","DOIUrl":"https://doi.org/10.1109/TMI.2026.3662001","url":null,"abstract":"<p><p>Segmentation of the pulmonary vessel from computed tomography (CT) images plays a crucial role in the diagnosis and treatment of various lung diseases. Although deep learning-based approaches have shown remarkable progress in recent years, their performance is often hindered by the lack of high-quality annotated datasets, in which the complex anatomy and morphology of pulmonary vessels make manual annotation challenging, time-consuming, and prone to errors. To address this, we propose PV25, the first dataset that features finely paired annotations of both pulmonary vessels and airways. Moreover, we propose TPNet, a novel tubular-aware prompt-tuning framework for pulmonary vessel segmentation under few-shot training with limited annotations. Specifically, based on an advanced and frozen segmentation backbone, TPNet proposes tunable encoding and decoding networks that learn tubular structures as transfer learning priors, bridging the gap between the source and target pulmonary vessel domains. Specifically, TPNet is built in an encoder-decoder manner, including the fixed segmentation backbone, tunable encoding and decoding networks. In encoding stage, the Morphology-Driven Region Growing (MDRG) module is developed to leverage the tubular connectivity of vessels to guide the network in capturing fine-grained features of pulmonary vessels. In decoding stage, the Cross-Correlation Guidance (CCG) module is introduced to integrate multi-scale correlations between airway and vessel structures in a coarse-to-fine manner. Extensive experiments conducted on multiple datasets demonstrate that TPNet achieves state-of-the-art performance in pulmonary vessel segmentation under limited training data. Besides, TPNet shows strong performance in related tasks such as airway segmentation and artery-vein classification, highlighting its robustness and versatility.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146133787","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Laboratory Test-Guided Medical Image Generation for Multi-Modal Disease Prediction. 实验室测试引导医学图像生成用于多模态疾病预测。
Pub Date : 2026-02-06 DOI: 10.1109/TMI.2026.3660978
Jingwen Xu, Fei Lyu, Ye Zhu, Pong C Yuen

The integration of laboratory tests and medical images is crucial in making accurate disease prediction. However, imaging data exhibits temporal sparsity, compared to frequently collected laboratory tests. This temporal sparsity limits effective multi-modal interaction, which in turn degrades the prediction accuracy.We address this issue by generating additional medical images at more time points, conditioned on the laboratory tests. Inspired by the pivotal role of organs in mediating laboratory tests and imaging abnormalities, we propose an Organ-Centric Modal-Shared Image Generator. It converts laboratory tests into imaging abnormalities through two key components: (1) Organ-Centric Graph: It positions organs as central nodes connecting laboratory tests and imaging abnormalities; and (2) Knowledge-Guided Modal-Shared Trajectory Module: It binds multi-modal features across time into a unified organ state trajectory. Experimental results demonstrate that our method improves multi-modal prediction performance across various diseases. Code is available at https://github.com/LyapunovStability/Lab_Guide_Med_Image_Gen.

将实验室检测与医学影像相结合是准确预测疾病的关键。然而,与经常收集的实验室测试相比,成像数据显示出时间稀疏性。这种时间稀疏性限制了有效的多模态相互作用,从而降低了预测精度。我们通过在更多时间点生成额外的医学图像来解决这个问题,这些图像以实验室测试为条件。受器官在介导实验室测试和成像异常中的关键作用的启发,我们提出了一个以器官为中心的模态共享图像生成器。它通过两个关键组件将实验室检查转化为成像异常:(1)器官中心图:它将器官定位为连接实验室检查和成像异常的中心节点;(2)知识引导模态-共享轨迹模块:将跨时间的多模态特征绑定为统一的器官状态轨迹。实验结果表明,该方法提高了多种疾病的多模态预测性能。代码可从https://github.com/LyapunovStability/Lab_Guide_Med_Image_Gen获得。
{"title":"Laboratory Test-Guided Medical Image Generation for Multi-Modal Disease Prediction.","authors":"Jingwen Xu, Fei Lyu, Ye Zhu, Pong C Yuen","doi":"10.1109/TMI.2026.3660978","DOIUrl":"https://doi.org/10.1109/TMI.2026.3660978","url":null,"abstract":"<p><p>The integration of laboratory tests and medical images is crucial in making accurate disease prediction. However, imaging data exhibits temporal sparsity, compared to frequently collected laboratory tests. This temporal sparsity limits effective multi-modal interaction, which in turn degrades the prediction accuracy.We address this issue by generating additional medical images at more time points, conditioned on the laboratory tests. Inspired by the pivotal role of organs in mediating laboratory tests and imaging abnormalities, we propose an Organ-Centric Modal-Shared Image Generator. It converts laboratory tests into imaging abnormalities through two key components: (1) Organ-Centric Graph: It positions organs as central nodes connecting laboratory tests and imaging abnormalities; and (2) Knowledge-Guided Modal-Shared Trajectory Module: It binds multi-modal features across time into a unified organ state trajectory. Experimental results demonstrate that our method improves multi-modal prediction performance across various diseases. Code is available at https://github.com/LyapunovStability/Lab_Guide_Med_Image_Gen.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146133847","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Moving Beyond Functional Connectivity: Time-Series Modeling for fMRI-Based Brain Disorder Classification. 超越功能连接:基于fmri的脑障碍分类的时间序列建模。
Pub Date : 2026-02-06 DOI: 10.1109/TMI.2026.3662157
Guoqi Yu, Xiaowei Hu, Angelica I Aviles-Rivero, Anqi Qiu, Shujun Wang

Functional magnetic resonance imaging (fMRI) enables non-invasive brain disorder classification by capturing blood-oxygen-level-dependent (BOLD) signals. However, most existing methods rely on functional connectivity (FC) via Pearson correlation, which reduces 4D BOLD signals to static 2D matrices-discarding temporal dynamics and capturing only linear inter-regional relationships. In this work, we benchmark state-of-the-art temporal models (e.g., time-series models: PatchTST, TimesNet, TimeMixer) on raw BOLD signals across five public datasets. Results show these models consistently outperform traditional FC-based approaches, highlighting the value of directly modeling temporal information such as cycle-like oscillatory fluctuations and drift-like slow baseline trends. Building on this insight, we propose DeCI, a simple yet effective framework that integrates two key principles: (i) Cycle and Drift Decomposition to disentangle cycle and drift within each ROI (Region of Interest); and (ii) Channel-Independence to model each ROI separately, improving robustness and reducing overfitting. Extensive experiments demonstrate that DeCI achieves superior classification accuracy and generalization compared to both FC-based and temporal baselines. Our findings advocate for a shift toward end-to-end temporal modeling in fMRI analysis to better capture complex brain dynamics. The code is available at https://github.com/Levi-Ackman/DeCI.

功能磁共振成像(fMRI)通过捕获血氧水平依赖(BOLD)信号实现非侵入性脑疾病分类。然而,大多数现有方法依赖于通过Pearson相关的功能连接(FC),这将4D BOLD信号减少到静态的2D矩阵,丢弃了时间动态,仅捕获线性区域间关系。在这项工作中,我们在五个公共数据集的原始BOLD信号上对最先进的时间模型(例如,时间序列模型:PatchTST, TimesNet, TimeMixer)进行基准测试。结果表明,这些模型始终优于传统的基于fc的方法,突出了直接建模时间信息(如周期振荡波动和漂移样慢基线趋势)的价值。基于这一见解,我们提出了DeCI,这是一个简单而有效的框架,集成了两个关键原则:(i)循环和漂移分解,以解开每个ROI(感兴趣区域)内的循环和漂移;(ii)渠道独立性,分别对每个ROI进行建模,提高鲁棒性并减少过拟合。大量的实验表明,与基于fc的基线和时间基线相比,DeCI具有更高的分类精度和泛化能力。我们的研究结果提倡在fMRI分析中转向端到端时间建模,以更好地捕捉复杂的大脑动态。代码可在https://github.com/Levi-Ackman/DeCI上获得。
{"title":"Moving Beyond Functional Connectivity: Time-Series Modeling for fMRI-Based Brain Disorder Classification.","authors":"Guoqi Yu, Xiaowei Hu, Angelica I Aviles-Rivero, Anqi Qiu, Shujun Wang","doi":"10.1109/TMI.2026.3662157","DOIUrl":"https://doi.org/10.1109/TMI.2026.3662157","url":null,"abstract":"<p><p>Functional magnetic resonance imaging (fMRI) enables non-invasive brain disorder classification by capturing blood-oxygen-level-dependent (BOLD) signals. However, most existing methods rely on functional connectivity (FC) via Pearson correlation, which reduces 4D BOLD signals to static 2D matrices-discarding temporal dynamics and capturing only linear inter-regional relationships. In this work, we benchmark state-of-the-art temporal models (e.g., time-series models: PatchTST, TimesNet, TimeMixer) on raw BOLD signals across five public datasets. Results show these models consistently outperform traditional FC-based approaches, highlighting the value of directly modeling temporal information such as cycle-like oscillatory fluctuations and drift-like slow baseline trends. Building on this insight, we propose DeCI, a simple yet effective framework that integrates two key principles: (i) Cycle and Drift Decomposition to disentangle cycle and drift within each ROI (Region of Interest); and (ii) Channel-Independence to model each ROI separately, improving robustness and reducing overfitting. Extensive experiments demonstrate that DeCI achieves superior classification accuracy and generalization compared to both FC-based and temporal baselines. Our findings advocate for a shift toward end-to-end temporal modeling in fMRI analysis to better capture complex brain dynamics. The code is available at https://github.com/Levi-Ackman/DeCI.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146133817","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LGFFM: A Localized and Globalized Frequency Fusion Model for Ultrasound Image Segmentation. LGFFM:一种局部和全球化的超声图像分割频率融合模型。
Pub Date : 2026-02-01 DOI: 10.1109/TMI.2025.3600327
Xiling Luo, Yi Wang, Le Ou-Yang

Accurate segmentation of ultrasound images plays a critical role in disease screening and diagnosis. Recently, neural network-based methods have garnered significant attention for their potential in improving ultrasound image segmentation. However, these methods still face significant challenges, primarily due to inherent issues in ultrasound images, such as low resolution, speckle noise, and artifacts. Additionally, ultrasound image segmentation encompasses a wide range of scenarios, including organ segmentation (e.g., cardiac and fetal head) and lesion segmentation (e.g., breast cancer and thyroid nodules), making the task highly diverse and complex. Existing methods are often designed for specific segmentation scenarios, which limits their flexibility and ability to meet the diverse needs across various scenarios. To address these challenges, we propose a novel Localized and Globalized Frequency Fusion Model (LGFFM) for ultrasound image segmentation. Specifically, we first design a Parallel Bi-Encoder (PBE) architecture that integrates Local Feature Blocks (LFB) and Global Feature Blocks (GLB) to enhance feature extraction. Additionally, we introduce a Frequency Domain Mapping Module (FDMM) to capture texture information, particularly high-frequency details such as edges. Finally, a Multi-Domain Fusion (MDF) method is developed to effectively integrate features across different domains. We conduct extensive experiments on eight representative public ultrasound datasets across four different types. The results demonstrate that LGFFM outperforms current state-of-the-art methods in both segmentation accuracy and generalization performance.

超声图像的准确分割对疾病的筛查和诊断起着至关重要的作用。近年来,基于神经网络的方法因其在改善超声图像分割方面的潜力而受到广泛关注。然而,这些方法仍然面临着巨大的挑战,主要是由于超声图像固有的问题,如低分辨率、斑点噪声和伪影。此外,超声图像分割包含广泛的场景,包括器官分割(如心脏和胎儿头)和病变分割(如乳腺癌和甲状腺结节),使任务高度多样化和复杂。现有的方法通常是为特定的分割场景设计的,这限制了它们的灵活性和满足不同场景不同需求的能力。为了解决这些挑战,我们提出了一种新的局部和全球化频率融合模型(LGFFM)用于超声图像分割。具体而言,我们首先设计了一个并行双编码器(PBE)架构,该架构集成了局部特征块(LFB)和全局特征块(GLB)来增强特征提取。此外,我们引入了频域映射模块(FDMM)来捕获纹理信息,特别是高频细节,如边缘。最后,提出了一种多域融合(Multi-Domain Fusion, MDF)方法来有效地整合不同域的特征。我们对四种不同类型的八个代表性公共超声数据集进行了广泛的实验。结果表明,LGFFM在分割精度和泛化性能方面都优于当前最先进的方法。
{"title":"LGFFM: A Localized and Globalized Frequency Fusion Model for Ultrasound Image Segmentation.","authors":"Xiling Luo, Yi Wang, Le Ou-Yang","doi":"10.1109/TMI.2025.3600327","DOIUrl":"10.1109/TMI.2025.3600327","url":null,"abstract":"<p><p>Accurate segmentation of ultrasound images plays a critical role in disease screening and diagnosis. Recently, neural network-based methods have garnered significant attention for their potential in improving ultrasound image segmentation. However, these methods still face significant challenges, primarily due to inherent issues in ultrasound images, such as low resolution, speckle noise, and artifacts. Additionally, ultrasound image segmentation encompasses a wide range of scenarios, including organ segmentation (e.g., cardiac and fetal head) and lesion segmentation (e.g., breast cancer and thyroid nodules), making the task highly diverse and complex. Existing methods are often designed for specific segmentation scenarios, which limits their flexibility and ability to meet the diverse needs across various scenarios. To address these challenges, we propose a novel Localized and Globalized Frequency Fusion Model (LGFFM) for ultrasound image segmentation. Specifically, we first design a Parallel Bi-Encoder (PBE) architecture that integrates Local Feature Blocks (LFB) and Global Feature Blocks (GLB) to enhance feature extraction. Additionally, we introduce a Frequency Domain Mapping Module (FDMM) to capture texture information, particularly high-frequency details such as edges. Finally, a Multi-Domain Fusion (MDF) method is developed to effectively integrate features across different domains. We conduct extensive experiments on eight representative public ultrasound datasets across four different types. The results demonstrate that LGFFM outperforms current state-of-the-art methods in both segmentation accuracy and generalization performance.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":"515-527"},"PeriodicalIF":0.0,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144884639","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
EndoRD-GS: Robust Deformable Endoscopic Scene Reconstruction via Gaussian Splatting. 基于高斯飞溅的鲁棒可变形内窥镜场景重建。
Pub Date : 2026-02-01 DOI: 10.1109/TMI.2025.3600253
Bingchen Gao, Jun Zhou, Jing Zou, Jing Qin

Real-time and realistic reconstruction of 3D dynamic surgical scenes from surgical videos is a novel and unique tool for surgical planning and intraoperative guidance. The 3D Gaussian splatting (GS), with its high rendering speed and reconstruction fidelity, has recently emerged as a promising technique for surgical scene reconstruction. However, existing GS-based methods still have two obvious shortcomings for realistic reconstruction. First, they largely struggle to capture localized yet intricate soft tissue deformations caused by complex instrument-tissue interactions. Second, they fail to model spatiotemporal coupling among Gaussian primitives for global adjustments during rapid perspective transformations, resulting in unstable reconstruction outputs. In this paper, we propose EndoRD-GS, an innovative approach that overcomes these two limitations through two core techniques: 1) periodic modulated Gaussian functions and 2) a new Biplane module. Specifically, our periodic modulated Gaussian functions incorporate meticulously designed modulations, significantly enhancing the representation of complex local tissue deformations. On the other hand, our Biplane module constructs spatiotemporal interactions among Gaussian primitives, enabling global adjustments and ensuring reliable scene reconstruction during rapid perspective transformations. Extensive experiments on three datasets demonstrate that our EndoRD-GS achieves superior performance in endoscopic scene reconstruction compared to state-of-the-art methods. The code is available at EndoRD-GS.

从手术视频中实时、逼真地重建三维动态手术场景是一种新颖而独特的手术计划和术中指导工具。三维高斯溅射(GS)以其较高的绘制速度和重建保真度,近年来成为一种很有前途的手术场景重建技术。然而,现有的基于gis的方法在真实重建方面仍然存在两个明显的不足。首先,他们在很大程度上难以捕捉由复杂的仪器与组织相互作用引起的局部但复杂的软组织变形。其次,在快速视角转换过程中,它们无法模拟高斯原语之间的时空耦合以进行全局调整,导致重建输出不稳定。在本文中,我们提出了一种创新的方法,通过两个核心技术克服了这两个限制:(1)周期调制高斯函数和(2)一个新的Biplane模块。具体来说,我们的周期调制高斯函数包含精心设计的调制,显着增强了复杂的局部组织变形的表示。另一方面,我们的Biplane模块构建高斯原语之间的时空交互,实现全局调整,并确保在快速视角转换期间可靠的场景重建。在三个数据集上进行的大量实验表明,与最先进的方法相比,我们的EndoRD-GS在内窥镜场景重建方面取得了卓越的性能。该代码可在endd - gs上获得。
{"title":"EndoRD-GS: Robust Deformable Endoscopic Scene Reconstruction via Gaussian Splatting.","authors":"Bingchen Gao, Jun Zhou, Jing Zou, Jing Qin","doi":"10.1109/TMI.2025.3600253","DOIUrl":"10.1109/TMI.2025.3600253","url":null,"abstract":"<p><p>Real-time and realistic reconstruction of 3D dynamic surgical scenes from surgical videos is a novel and unique tool for surgical planning and intraoperative guidance. The 3D Gaussian splatting (GS), with its high rendering speed and reconstruction fidelity, has recently emerged as a promising technique for surgical scene reconstruction. However, existing GS-based methods still have two obvious shortcomings for realistic reconstruction. First, they largely struggle to capture localized yet intricate soft tissue deformations caused by complex instrument-tissue interactions. Second, they fail to model spatiotemporal coupling among Gaussian primitives for global adjustments during rapid perspective transformations, resulting in unstable reconstruction outputs. In this paper, we propose EndoRD-GS, an innovative approach that overcomes these two limitations through two core techniques: 1) periodic modulated Gaussian functions and 2) a new Biplane module. Specifically, our periodic modulated Gaussian functions incorporate meticulously designed modulations, significantly enhancing the representation of complex local tissue deformations. On the other hand, our Biplane module constructs spatiotemporal interactions among Gaussian primitives, enabling global adjustments and ensuring reliable scene reconstruction during rapid perspective transformations. Extensive experiments on three datasets demonstrate that our EndoRD-GS achieves superior performance in endoscopic scene reconstruction compared to state-of-the-art methods. The code is available at EndoRD-GS.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":"528-541"},"PeriodicalIF":0.0,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144884638","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PAINT: Prior-Aided Alternate Iterative NeTwork for Ultra-Low-Dose CT Imaging Using Diffusion Model-Restored Sinogram. PAINT:基于扩散模型恢复Sinogram超低剂量CT成像的先验辅助交替迭代网络。
Pub Date : 2026-02-01 DOI: 10.1109/TMI.2025.3599508
Kaile Chen, Weikang Zhang, Ziheng Deng, Yufu Zhou, Jun Zhao

Obtaining multiple CT scans from the same patient is required in many clinical scenarios, such as lung nodule screening and image-guided radiation therapy. Repeated scans would expose patients to higher radiation dose and increase the risk of cancer. In this study, we aim to achieve ultra-low-dose imaging for subsequent scans by collecting extremely undersampled sinogram via regional few-view scanning, and preserve image quality utilizing the preceding fullsampled scan as prior. To fully exploit prior information, we propose a two-stage framework consisting of diffusion model-based sinogram restoration and deep learning-based unrolled iterative reconstruction. Specifically, the undersampled sinogram is first restored by a conditional diffusion model with sinogram-domain prior guidance. Then, we formulate the undersampled data reconstruction problem as an optimization problem combining fidelity terms for both undersampled and restored data, along with a regularization term based on image-domain prior. Next, we propose Prior-aided Alternate Iterative NeTwork (PAINT) to solve the optimization problem. PAINT alternately updates the undersampled or restored data fidelity term, and unrolls the iterations to integrate neural network-based prior regularization. In the case of 112 mm field of view in simulated data experiments, our proposed framework achieved superior performance in terms of CT value accuracy and image details preservation. Clinical data experiments also demonstrated that our proposed framework outperformed the comparison methods in artifact reduction and structure recovery.

在许多临床情况下,需要对同一患者进行多次CT扫描,例如肺结节筛查和图像引导放射治疗。重复扫描会使病人暴露在更高的辐射剂量下,增加患癌症的风险。在本研究中,我们的目标是通过区域少视点扫描收集极度欠采样的正弦图,为后续扫描实现超低剂量成像,并利用之前的全采样扫描保持图像质量。为了充分利用先验信息,我们提出了一个两阶段框架,包括基于扩散模型的正弦图恢复和基于深度学习的展开迭代重建。具体地说,欠采样的正弦图首先通过带有正弦图域先验引导的条件扩散模型恢复。然后,我们将欠采样数据重建问题表述为将欠采样和恢复数据的保真度项以及基于图像域先验的正则化项结合在一起的优化问题。接下来,我们提出了先验辅助交替迭代网络(PAINT)来解决优化问题。PAINT交替更新欠采样或恢复的数据保真度项,并展开迭代以集成基于神经网络的先验正则化。在模拟数据实验中,在112 mm视场的情况下,我们提出的框架在CT值精度和图像细节保存方面取得了优异的性能。临床数据实验也表明,我们提出的框架在伪影减少和结构恢复方面优于比较方法。
{"title":"PAINT: Prior-Aided Alternate Iterative NeTwork for Ultra-Low-Dose CT Imaging Using Diffusion Model-Restored Sinogram.","authors":"Kaile Chen, Weikang Zhang, Ziheng Deng, Yufu Zhou, Jun Zhao","doi":"10.1109/TMI.2025.3599508","DOIUrl":"10.1109/TMI.2025.3599508","url":null,"abstract":"<p><p>Obtaining multiple CT scans from the same patient is required in many clinical scenarios, such as lung nodule screening and image-guided radiation therapy. Repeated scans would expose patients to higher radiation dose and increase the risk of cancer. In this study, we aim to achieve ultra-low-dose imaging for subsequent scans by collecting extremely undersampled sinogram via regional few-view scanning, and preserve image quality utilizing the preceding fullsampled scan as prior. To fully exploit prior information, we propose a two-stage framework consisting of diffusion model-based sinogram restoration and deep learning-based unrolled iterative reconstruction. Specifically, the undersampled sinogram is first restored by a conditional diffusion model with sinogram-domain prior guidance. Then, we formulate the undersampled data reconstruction problem as an optimization problem combining fidelity terms for both undersampled and restored data, along with a regularization term based on image-domain prior. Next, we propose Prior-aided Alternate Iterative NeTwork (PAINT) to solve the optimization problem. PAINT alternately updates the undersampled or restored data fidelity term, and unrolls the iterations to integrate neural network-based prior regularization. In the case of 112 mm field of view in simulated data experiments, our proposed framework achieved superior performance in terms of CT value accuracy and image details preservation. Clinical data experiments also demonstrated that our proposed framework outperformed the comparison methods in artifact reduction and structure recovery.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":"434-447"},"PeriodicalIF":0.0,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144877631","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Ultrasound Autofocusing: Common Midpoint Phase Error Optimization via Differentiable Beamforming. 超声自动聚焦:基于可微波束形成的常见中点相位误差优化。
Pub Date : 2026-02-01 DOI: 10.1109/TMI.2025.3607875
Walter Simson, Louise Zhuang, Benjamin N Frey, Sergio J Sanabria, Jeremy J Dahl, Dongwoon Hyun

In ultrasound imaging, propagation of an acoustic wavefront through heterogeneous media causes phase aberrations that degrade the coherence of the reflected wavefront, leading to reduced image resolution and contrast. Adaptive imaging techniques attempt to correct this phase aberration and restore coherence, leading to improved focusing of the image. We propose an autofocusing paradigm for aberration correction in ultrasound imaging by fitting an acoustic velocity field to pressure measurements, via optimization of the common midpoint phase error (CMPE), using a straight-ray wave propagation model for beamforming in diffusely scattering media. We show that CMPE induced by heterogeneous acoustic velocity is a robust measure of phase aberration that can be used for acoustic autofocusing. CMPE is optimized iteratively using a differentiable beamforming approach to simultaneously improve the image focus while estimating the acoustic velocity field of the interrogated medium. The approach relies solely on wavefield measurements using a straight-ray integral solution of the two-way time-of-flight without explicit numerical time-stepping models of wave propagation. We demonstrate method performance through in silico simulations, in vitro phantom measurements, and in vivo mammalian models, showing practical applications in distributed aberration quantification, correction, and velocity estimation for medical ultrasound autofocusing.

在超声成像中,声波前通过异质介质的传播会引起相位像差,从而降低反射波前的相干性,导致图像分辨率和对比度降低。自适应成像技术试图纠正这种相位像差并恢复相干性,从而改善图像的聚焦。我们提出了一种自动聚焦模式,通过优化共同中点相位误差(CMPE),将声速场拟合到压力测量值中,从而校正超声成像中的像差,并使用漫射散射介质中波束形成的直线波传播模型。研究表明,由非均匀声速引起的CMPE是一种可靠的相位像差测量方法,可用于声学自动聚焦。利用可微波束形成方法对CMPE进行迭代优化,在估计被测介质声速场的同时提高了图像聚焦。该方法完全依赖于使用双向飞行时间的直线积分解的波场测量,而没有明确的波传播数值时间步进模型。我们通过硅模拟、体外幻影测量和体内哺乳动物模型展示了该方法的性能,展示了在医学超声自动聚焦的分布式像差量化、校正和速度估计方面的实际应用。
{"title":"Ultrasound Autofocusing: Common Midpoint Phase Error Optimization via Differentiable Beamforming.","authors":"Walter Simson, Louise Zhuang, Benjamin N Frey, Sergio J Sanabria, Jeremy J Dahl, Dongwoon Hyun","doi":"10.1109/TMI.2025.3607875","DOIUrl":"10.1109/TMI.2025.3607875","url":null,"abstract":"<p><p>In ultrasound imaging, propagation of an acoustic wavefront through heterogeneous media causes phase aberrations that degrade the coherence of the reflected wavefront, leading to reduced image resolution and contrast. Adaptive imaging techniques attempt to correct this phase aberration and restore coherence, leading to improved focusing of the image. We propose an autofocusing paradigm for aberration correction in ultrasound imaging by fitting an acoustic velocity field to pressure measurements, via optimization of the common midpoint phase error (CMPE), using a straight-ray wave propagation model for beamforming in diffusely scattering media. We show that CMPE induced by heterogeneous acoustic velocity is a robust measure of phase aberration that can be used for acoustic autofocusing. CMPE is optimized iteratively using a differentiable beamforming approach to simultaneously improve the image focus while estimating the acoustic velocity field of the interrogated medium. The approach relies solely on wavefield measurements using a straight-ray integral solution of the two-way time-of-flight without explicit numerical time-stepping models of wave propagation. We demonstrate method performance through in silico simulations, in vitro phantom measurements, and in vivo mammalian models, showing practical applications in distributed aberration quantification, correction, and velocity estimation for medical ultrasound autofocusing.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":"681-692"},"PeriodicalIF":0.0,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145031617","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Ring artifact reduction in photon counting CT using redundant sampling and autocalibration. 利用冗余采样和自动校准减少光子计数CT中的环形伪影。
Pub Date : 2026-01-26 DOI: 10.1109/TMI.2026.3658004
Scott S Hsieh, James Day, Xinchen Deng, Magdalena Bazalova-Carter

Ring artifacts in CT are caused by uncalibrated variations in detector pixels and are especially prevalent with emerging photon counting detectors (PCDs). Control of ring artifacts is conventionally accomplished by improving either hardware manufacturing or software correction algorithms. An alternative solution is detector autocalibration, in which two redundant samples of each line integral are acquired and used to dynamically calibrate the PCD. Autocalibration was first proposed by Hounsfield in 1977 and was demonstrated on the EMI Topaz prototype scanner in 1980, but details surrounding this implementation are sparse. We investigate a form of autocalibration that requires just two redundant acquisitions, which could be acquired using flying focal spot on a clinical scanner but is demonstrated here with a detector shift. We formulated autocalibration as an optimization problem to determine the relative gain factor of each pixel and tested it on scans of a chicken thigh specimen, resolution phantom, and a cylindrical phantom. Ring artifacts were significantly reduced. Some residual artifacts remained but could not be discriminated from the intrinsic temporal instability of our PCD modules. Autocalibration could facilitate the adoption of widespread photon counting CT by reducing ring artifacts, thermal management requirements, or stability requirements that are present today. Demonstration of autocalibration on a rotating gantry with flying focal spot remains future work.

CT中的环形伪影是由探测器像素的未校准变化引起的,在新兴的光子计数探测器(PCDs)中尤其普遍。环形工件的控制通常是通过改进硬件制造或软件校正算法来完成的。另一种解决方案是探测器自动校准,其中每个线积分获得两个冗余样本并用于动态校准PCD。自动校准最初是由Hounsfield于1977年提出的,并于1980年在EMI Topaz原型扫描仪上进行了演示,但围绕该实现的细节很少。我们研究了一种自动校准的形式,它只需要两个冗余的采集,可以使用临床扫描仪上的飞行焦点来获得,但这里用检测器移位来演示。我们将自动校准制定为一个优化问题,以确定每个像素的相对增益因子,并在鸡大腿标本、分辨率幻影和圆柱形幻影的扫描上进行了测试。环形伪影显著减少。一些残留的工件仍然存在,但不能从我们的PCD模块固有的时间不稳定性中区分出来。通过减少目前存在的环形伪影、热管理要求或稳定性要求,自动校准可以促进光子计数CT的广泛采用。在具有飞行焦斑的旋转龙门上进行自动校准的演示仍有待进一步的工作。
{"title":"Ring artifact reduction in photon counting CT using redundant sampling and autocalibration.","authors":"Scott S Hsieh, James Day, Xinchen Deng, Magdalena Bazalova-Carter","doi":"10.1109/TMI.2026.3658004","DOIUrl":"https://doi.org/10.1109/TMI.2026.3658004","url":null,"abstract":"<p><p>Ring artifacts in CT are caused by uncalibrated variations in detector pixels and are especially prevalent with emerging photon counting detectors (PCDs). Control of ring artifacts is conventionally accomplished by improving either hardware manufacturing or software correction algorithms. An alternative solution is detector autocalibration, in which two redundant samples of each line integral are acquired and used to dynamically calibrate the PCD. Autocalibration was first proposed by Hounsfield in 1977 and was demonstrated on the EMI Topaz prototype scanner in 1980, but details surrounding this implementation are sparse. We investigate a form of autocalibration that requires just two redundant acquisitions, which could be acquired using flying focal spot on a clinical scanner but is demonstrated here with a detector shift. We formulated autocalibration as an optimization problem to determine the relative gain factor of each pixel and tested it on scans of a chicken thigh specimen, resolution phantom, and a cylindrical phantom. Ring artifacts were significantly reduced. Some residual artifacts remained but could not be discriminated from the intrinsic temporal instability of our PCD modules. Autocalibration could facilitate the adoption of widespread photon counting CT by reducing ring artifacts, thermal management requirements, or stability requirements that are present today. Demonstration of autocalibration on a rotating gantry with flying focal spot remains future work.</p>","PeriodicalId":94033,"journal":{"name":"IEEE transactions on medical imaging","volume":"PP ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146055735","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE transactions on medical imaging
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1