首页 > 最新文献

IEEE Transactions on Medical Imaging最新文献

英文 中文
NGP-Net: a Lightweight Growth Prediction Network for Pulmonary Nodules. NGP-Net:一个轻量级肺结节生长预测网络。
IF 10.6 1区 医学 Q1 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2026-01-20 DOI: 10.1109/tmi.2026.3656184
Xinkai Tang,Zhiyao Luo,Feng Liu,Wencai Huang,Jiani Zou
Accurate monitoring of pulmonary nodules' growth is crucial for preventing lung cancer progression and improving patient outcomes. Yet, identifying high-risk nodules in computed tomography (CT) scans remains challenging due to subtle growth patterns, irregular follow-up intervals, and the limitations of current diagnostic tools. Existing methods often depend on single-timepoint analyses or assume fixed temporal intervals, constraining prediction to rigid scenarios. To address these limitations, we propose NGP-Net, a W-shaped architecture for dynamic nodule growth prediction from irregular longitudinal CT scans. NGP-Net introduces a Spatial-Temporal Encoding Module (STEM) that learns temporal dynamics directly from irregularly sampled data, and a dual-branch decoder that reconstructs high-fidelity nodule textures and shapes at arbitrary future timepoints. We further release a curated dataset of 378 chest CT scans from 103 patients with 226 pulmonary nodules, each followed across at least three timepoints spanning 2-64 months and annotated by seven radiologists. Extensive evaluation demonstrates that NGP-Net achieves state-of-the-art performance on this new dataset, obtaining the lowest mean square error of 6.13 × 10-3 (overall) and 1.28 × 10-4 (nodule-specific), with substantial improvements in Dice similarity coefficient (10.55%), peak signal-to-noise ratio (0.29 dB), and structural similarity index (5.94%). NGP-Net's robust and precise predictions across varied growth scenarios highlight its potential to support radiologists in clinical decision-making. The source code and dataset are publicly available at GitHub and Kaggle.
准确监测肺结节的生长对预防肺癌进展和改善患者预后至关重要。然而,由于细微的生长模式、不规则的随访间隔以及当前诊断工具的局限性,在计算机断层扫描(CT)中识别高风险结节仍然具有挑战性。现有的方法通常依赖于单时间点分析或假设固定的时间间隔,将预测限制在刚性情景中。为了解决这些限制,我们提出了NGP-Net,这是一种w形结构,用于从不规则纵向CT扫描中动态预测结节生长。NGP-Net引入了一个时空编码模块(STEM),可以直接从不规则采样数据中学习时间动态,以及一个双分支解码器,可以在任意未来时间点重建高保真的根茎纹理和形状。我们进一步发布了一个精心整理的数据集,其中包括103例226个肺结节患者的378次胸部CT扫描,每次扫描至少跨越2-64个月的三个时间点,并由7名放射科医生注释。广泛的评估表明,NGP-Net在这个新数据集上达到了最先进的性能,获得了最低的均方误差6.13 × 10-3(总体)和1.28 × 10-4(结节特异性),在Dice相似系数(10.55%),峰值信噪比(0.29 dB)和结构相似指数(5.94%)方面有了显着改善。NGP-Net对各种增长情景的稳健而精确的预测突出了它在支持放射科医生临床决策方面的潜力。源代码和数据集可以在GitHub和Kaggle上公开获得。
{"title":"NGP-Net: a Lightweight Growth Prediction Network for Pulmonary Nodules.","authors":"Xinkai Tang,Zhiyao Luo,Feng Liu,Wencai Huang,Jiani Zou","doi":"10.1109/tmi.2026.3656184","DOIUrl":"https://doi.org/10.1109/tmi.2026.3656184","url":null,"abstract":"Accurate monitoring of pulmonary nodules' growth is crucial for preventing lung cancer progression and improving patient outcomes. Yet, identifying high-risk nodules in computed tomography (CT) scans remains challenging due to subtle growth patterns, irregular follow-up intervals, and the limitations of current diagnostic tools. Existing methods often depend on single-timepoint analyses or assume fixed temporal intervals, constraining prediction to rigid scenarios. To address these limitations, we propose NGP-Net, a W-shaped architecture for dynamic nodule growth prediction from irregular longitudinal CT scans. NGP-Net introduces a Spatial-Temporal Encoding Module (STEM) that learns temporal dynamics directly from irregularly sampled data, and a dual-branch decoder that reconstructs high-fidelity nodule textures and shapes at arbitrary future timepoints. We further release a curated dataset of 378 chest CT scans from 103 patients with 226 pulmonary nodules, each followed across at least three timepoints spanning 2-64 months and annotated by seven radiologists. Extensive evaluation demonstrates that NGP-Net achieves state-of-the-art performance on this new dataset, obtaining the lowest mean square error of 6.13 × 10-3 (overall) and 1.28 × 10-4 (nodule-specific), with substantial improvements in Dice similarity coefficient (10.55%), peak signal-to-noise ratio (0.29 dB), and structural similarity index (5.94%). NGP-Net's robust and precise predictions across varied growth scenarios highlight its potential to support radiologists in clinical decision-making. The source code and dataset are publicly available at GitHub and Kaggle.","PeriodicalId":13418,"journal":{"name":"IEEE Transactions on Medical Imaging","volume":"6 1","pages":""},"PeriodicalIF":10.6,"publicationDate":"2026-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146005208","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ADHD Classification with GCN via Joint Feature Learning among Nodes and Edges. 基于节点和边间联合特征学习的GCN ADHD分类。
IF 10.6 1区 医学 Q1 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2026-01-20 DOI: 10.1109/tmi.2026.3656430
Xiaotong Wang,Yibin Tang,Yuan Gao,Xiaojing Meng,Ying Chen,Aimin Jiang
Brain functional connectivity networks (FCNs) derived from resting-state functional magnetic resonance imaging (rs-fMRI) data have been widely used to identify altered brain network patterns in attention-deficit/hyperactivity disorder (ADHD). Current graph neural network (GNN) approaches using FCNs predominantly emphasize node features while underutilizing edge information. Moreover, these GNN-based methods also inadequately represent dynamic interdependencies among evolving node features across network layers, limiting their diagnostic performance. We present a graph convolutional network via joint feature learning between nodes and edges (JNEL-GCN) that integrates neuroimaging features for ADHD classification and biomarker discovery. Our framework constructs dual graph representations: (1) a node graph using amplitude of low-frequency fluctuations (ALFF) measures across multiple frequency bands as nodal features, along with functional connectivity (FC) and node feature relationship matrices as edge attributes; (2) an edge graph derived through line graph theory, enabling the interchange of node and edge roles. By leveraging the dual-graph design, our model implements an alternating feature update mechanism with optimized graph convolution operations, facilitating feature hierarchical learning of node-edge relationships across network layers. Extensive experiments demonstrate remarkable performance, achieving 97.3% accuracy on ADHD200 and 97.1% on ABIDE-I datasets, significantly outperforming current benchmarks. Meanwhile, gradient-based biomarker analysis identifies significant regions in bilateral limbic and default mode networks associated with ADHD, aligning with the findings in existing literature. Therefore, this dual-graph approach advances neuroimaging-based diagnosis by comprehensively capturing dynamic network interactions, while providing interpretable biomarkers for clinical neuroscience applications.
来自静息状态功能磁共振成像(rs-fMRI)数据的脑功能连接网络(fns)已被广泛用于识别注意力缺陷/多动障碍(ADHD)患者改变的脑网络模式。目前使用fcn的图神经网络(GNN)方法主要强调节点特征,而未充分利用边缘信息。此外,这些基于gnn的方法也不能充分表征网络层中不断发展的节点特征之间的动态相互依赖性,从而限制了它们的诊断性能。我们提出了一个通过节点和边缘之间的联合特征学习(jnell - gcn)的图卷积网络,该网络集成了用于ADHD分类和生物标志物发现的神经影像学特征。我们的框架构建了对偶图表示:(1)使用跨多个频带的低频波动幅度(ALFF)度量作为节点特征,以及功能连通性(FC)和节点特征关系矩阵作为边缘属性的节点图;(2)通过线形图理论导出的边图,实现节点和边角色的互换。通过利用双图设计,我们的模型实现了一种具有优化图卷积操作的交替特征更新机制,促进了跨网络层节点-边缘关系的特征分层学习。大量的实验证明了卓越的性能,在ADHD200上达到97.3%的准确率,在ABIDE-I数据集上达到97.1%,显著优于当前的基准测试。同时,基于梯度的生物标志物分析确定了与ADHD相关的双侧边缘和默认模式网络的重要区域,与现有文献的发现一致。因此,这种双图方法通过全面捕获动态网络相互作用来推进基于神经影像学的诊断,同时为临床神经科学应用提供可解释的生物标志物。
{"title":"ADHD Classification with GCN via Joint Feature Learning among Nodes and Edges.","authors":"Xiaotong Wang,Yibin Tang,Yuan Gao,Xiaojing Meng,Ying Chen,Aimin Jiang","doi":"10.1109/tmi.2026.3656430","DOIUrl":"https://doi.org/10.1109/tmi.2026.3656430","url":null,"abstract":"Brain functional connectivity networks (FCNs) derived from resting-state functional magnetic resonance imaging (rs-fMRI) data have been widely used to identify altered brain network patterns in attention-deficit/hyperactivity disorder (ADHD). Current graph neural network (GNN) approaches using FCNs predominantly emphasize node features while underutilizing edge information. Moreover, these GNN-based methods also inadequately represent dynamic interdependencies among evolving node features across network layers, limiting their diagnostic performance. We present a graph convolutional network via joint feature learning between nodes and edges (JNEL-GCN) that integrates neuroimaging features for ADHD classification and biomarker discovery. Our framework constructs dual graph representations: (1) a node graph using amplitude of low-frequency fluctuations (ALFF) measures across multiple frequency bands as nodal features, along with functional connectivity (FC) and node feature relationship matrices as edge attributes; (2) an edge graph derived through line graph theory, enabling the interchange of node and edge roles. By leveraging the dual-graph design, our model implements an alternating feature update mechanism with optimized graph convolution operations, facilitating feature hierarchical learning of node-edge relationships across network layers. Extensive experiments demonstrate remarkable performance, achieving 97.3% accuracy on ADHD200 and 97.1% on ABIDE-I datasets, significantly outperforming current benchmarks. Meanwhile, gradient-based biomarker analysis identifies significant regions in bilateral limbic and default mode networks associated with ADHD, aligning with the findings in existing literature. Therefore, this dual-graph approach advances neuroimaging-based diagnosis by comprehensively capturing dynamic network interactions, while providing interpretable biomarkers for clinical neuroscience applications.","PeriodicalId":13418,"journal":{"name":"IEEE Transactions on Medical Imaging","volume":"275 1","pages":""},"PeriodicalIF":10.6,"publicationDate":"2026-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146005055","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Quantitative ultrasound imaging of bone: anatomical images, tissue structural quality, and pulsatile blood flow 骨的定量超声成像:解剖图像、组织结构质量和搏动血流
IF 10.6 1区 医学 Q1 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2026-01-19 DOI: 10.1109/tmi.2026.3655400
Gabrielle Laloy-Borgna, Nastassia Navasiolava, Pim Hutting, Andréa Bertona, Amadou S. Dia, Sébastien Salles, Anthony Augé, Alice Mazzolini, Quentin Grimal, Olivier Lucidarme, Hervé Locrelle, Jacques-Olivier Fortrat, Laurence Vico, Marc-Antoine Custaud, Guillaume Renaud
{"title":"Quantitative ultrasound imaging of bone: anatomical images, tissue structural quality, and pulsatile blood flow","authors":"Gabrielle Laloy-Borgna, Nastassia Navasiolava, Pim Hutting, Andréa Bertona, Amadou S. Dia, Sébastien Salles, Anthony Augé, Alice Mazzolini, Quentin Grimal, Olivier Lucidarme, Hervé Locrelle, Jacques-Olivier Fortrat, Laurence Vico, Marc-Antoine Custaud, Guillaume Renaud","doi":"10.1109/tmi.2026.3655400","DOIUrl":"https://doi.org/10.1109/tmi.2026.3655400","url":null,"abstract":"","PeriodicalId":13418,"journal":{"name":"IEEE Transactions on Medical Imaging","volume":"26 1","pages":""},"PeriodicalIF":10.6,"publicationDate":"2026-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146000520","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
B2Q-Net: Bidirectional Branch Query Network for Surgical Phase Recognition. B2Q-Net:面向手术相位识别的双向分支查询网络。
IF 10.6 1区 医学 Q1 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2026-01-16 DOI: 10.1109/tmi.2026.3654795
Wenjie Zhang,Zhiheng Li,Yue Bi,Xiao Jia,Ran Song,Yipeng Zhang,Wei Zhang
Surgical phase recognition (SPR) is essential for surgical workflow analysis and provides immediate guidance during procedures. Existing methods aggregate frame-level information into a global representation and treat the task as frame-wise classification. However, this pipeline lacks a feedback mechanism for integrating historical information into local temporal modeling. To address this limitation, we propose the Bidirectional Branch Query Network (B2Q-Net), which reformulates the SPR task as the bidirectional query between phase-level features and frame-level features. B2Q-Net incorporates historical information during the initialization of phase queries. This enables bidirectional information flow during iterative refinement of two-level feature maps between phases and frames. Furthermore, we introduce a dual-scale selector (DSS) to generate high-quality phase queries for the current video clip. These phase queries retrieve historical information from the proposed state space query (SSQ) module, which uses learnable tokens as the historical state space to preserve historical information. Extensive evaluations on three datasets demonstrate that B2Q-Net consistently outperforms state-of-the-art methods in recognition accuracy while achieving an inference speed of 106 fps. The B2Q-Net code is available at https://github.com/vsislab/B2Q-Net.
手术阶段识别(SPR)对于手术流程分析至关重要,并在手术过程中提供即时指导。现有方法将帧级信息聚合到全局表示中,并将任务视为逐帧分类。然而,该管道缺乏将历史信息集成到局部时间建模中的反馈机制。为了解决这一限制,我们提出了双向分支查询网络(B2Q-Net),它将SPR任务重新表述为相位级特征和帧级特征之间的双向查询。B2Q-Net在初始化阶段查询期间合并了历史信息。这使得在迭代细化阶段和框架之间的两级特征映射期间双向信息流成为可能。此外,我们引入了双尺度选择器(DSS)来为当前视频片段生成高质量的相位查询。这些阶段查询从建议的状态空间查询(SSQ)模块检索历史信息,该模块使用可学习的令牌作为历史状态空间来保存历史信息。对三个数据集的广泛评估表明,B2Q-Net在识别精度方面始终优于最先进的方法,同时实现106 fps的推理速度。B2Q-Net的代码可在https://github.com/vsislab/B2Q-Net上获得。
{"title":"B2Q-Net: Bidirectional Branch Query Network for Surgical Phase Recognition.","authors":"Wenjie Zhang,Zhiheng Li,Yue Bi,Xiao Jia,Ran Song,Yipeng Zhang,Wei Zhang","doi":"10.1109/tmi.2026.3654795","DOIUrl":"https://doi.org/10.1109/tmi.2026.3654795","url":null,"abstract":"Surgical phase recognition (SPR) is essential for surgical workflow analysis and provides immediate guidance during procedures. Existing methods aggregate frame-level information into a global representation and treat the task as frame-wise classification. However, this pipeline lacks a feedback mechanism for integrating historical information into local temporal modeling. To address this limitation, we propose the Bidirectional Branch Query Network (B2Q-Net), which reformulates the SPR task as the bidirectional query between phase-level features and frame-level features. B2Q-Net incorporates historical information during the initialization of phase queries. This enables bidirectional information flow during iterative refinement of two-level feature maps between phases and frames. Furthermore, we introduce a dual-scale selector (DSS) to generate high-quality phase queries for the current video clip. These phase queries retrieve historical information from the proposed state space query (SSQ) module, which uses learnable tokens as the historical state space to preserve historical information. Extensive evaluations on three datasets demonstrate that B2Q-Net consistently outperforms state-of-the-art methods in recognition accuracy while achieving an inference speed of 106 fps. The B2Q-Net code is available at https://github.com/vsislab/B2Q-Net.","PeriodicalId":13418,"journal":{"name":"IEEE Transactions on Medical Imaging","volume":"37 1","pages":""},"PeriodicalIF":10.6,"publicationDate":"2026-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145986341","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A High-Performance Self-Collimation SPECT for Small Animal Imaging. 用于小动物成像的高性能自准直SPECT。
IF 10.6 1区 医学 Q1 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2026-01-16 DOI: 10.1109/tmi.2026.3654599
Debin Zhang,Zhenlei Lyu,Tianpeng Xu,Peng Fan,Zerui Yu,Qiqi Ye,Yifan Hu,Jing Wu,Qingyang Wei,Xin Zhang,Qianqian Gan,Yang Xu,Li Wang,Rutao Yao,Min-Fu Yang,Zuo-Xiang He,Yaqiang Liu,Tianyu Ma
Stemmed from our novel single-photon imaging concept of detector self-collimation-which leverages detectors themselves as collimators to overcome the inherent resolution-sensitivity trade-off in conventional SPECT-this study presents the design and evaluation of the first full-ring self-collimation SPECT (SC-SPECT) scanner for small animal imaging. The system features four concentric detector rings and two interchangeable high-aperture-ratio tungsten collimator rings optimized for high-resolution (HR) and general-purpose (GP) imaging applications. Detector rings contain 480, 720, 960, and 1,200 evenly distributed GAGG(Ce) scintillators, each measuring 0.84 mm (tangential) × 6 mm (radial) × 20 mm (axial) and separated by 0.84-mm gaps to enable effective photon collimation. Inner detector rings and the collimator ring collectively provide collimation for photons reaching subsequent outer rings. Dual-end SiPM readouts facilitate axial depth-of-interaction measurements. Phantom and mouse studies are performed to assess the system's resolution, sensitivity, and field-of-view volume, and SC-SPECT demonstrates generally superior performance compared with state-of-the-art small-animal SPECT systems. Mouse bone images using 99mTc-MDP show CT-like resolution, clearly delineating detailed tracer uptake distributions within small structures such as mouse paws and skulls, indicating a significant technological advancement in small-animal SPECT imaging.
源于我们新颖的探测器自准直的单光子成像概念-利用探测器本身作为准直器来克服传统SPECT固有的分辨率和灵敏度权衡-本研究提出了用于小动物成像的第一台全环自准直SPECT (SC-SPECT)扫描仪的设计和评估。该系统具有四个同心圆探测器环和两个可互换的高孔径比钨准直器环,针对高分辨率(HR)和通用(GP)成像应用进行了优化。探测器环包含480、720、960和1200个均匀分布的GAGG(Ce)闪烁体,每个闪烁体的尺寸为0.84 mm(切向)× 6 mm(径向)× 20 mm(轴向),间隔0.84 mm,以实现有效的光子准直。内探测器环和准直器环共同为到达后续外环的光子提供准直。双端SiPM读数便于轴向相互作用深度测量。幻影和小鼠研究是为了评估系统的分辨率、灵敏度和视野体积,SC-SPECT与最先进的小动物SPECT系统相比,通常表现出优越的性能。使用99mTc-MDP的小鼠骨骼图像显示出与ct类似的分辨率,清晰地描绘了小鼠爪子和头骨等小结构内详细的示踪剂摄取分布,表明了小动物SPECT成像的重大技术进步。
{"title":"A High-Performance Self-Collimation SPECT for Small Animal Imaging.","authors":"Debin Zhang,Zhenlei Lyu,Tianpeng Xu,Peng Fan,Zerui Yu,Qiqi Ye,Yifan Hu,Jing Wu,Qingyang Wei,Xin Zhang,Qianqian Gan,Yang Xu,Li Wang,Rutao Yao,Min-Fu Yang,Zuo-Xiang He,Yaqiang Liu,Tianyu Ma","doi":"10.1109/tmi.2026.3654599","DOIUrl":"https://doi.org/10.1109/tmi.2026.3654599","url":null,"abstract":"Stemmed from our novel single-photon imaging concept of detector self-collimation-which leverages detectors themselves as collimators to overcome the inherent resolution-sensitivity trade-off in conventional SPECT-this study presents the design and evaluation of the first full-ring self-collimation SPECT (SC-SPECT) scanner for small animal imaging. The system features four concentric detector rings and two interchangeable high-aperture-ratio tungsten collimator rings optimized for high-resolution (HR) and general-purpose (GP) imaging applications. Detector rings contain 480, 720, 960, and 1,200 evenly distributed GAGG(Ce) scintillators, each measuring 0.84 mm (tangential) × 6 mm (radial) × 20 mm (axial) and separated by 0.84-mm gaps to enable effective photon collimation. Inner detector rings and the collimator ring collectively provide collimation for photons reaching subsequent outer rings. Dual-end SiPM readouts facilitate axial depth-of-interaction measurements. Phantom and mouse studies are performed to assess the system's resolution, sensitivity, and field-of-view volume, and SC-SPECT demonstrates generally superior performance compared with state-of-the-art small-animal SPECT systems. Mouse bone images using 99mTc-MDP show CT-like resolution, clearly delineating detailed tracer uptake distributions within small structures such as mouse paws and skulls, indicating a significant technological advancement in small-animal SPECT imaging.","PeriodicalId":13418,"journal":{"name":"IEEE Transactions on Medical Imaging","volume":"4 1","pages":""},"PeriodicalIF":10.6,"publicationDate":"2026-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145986343","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning Modality-Aware Representations: Adaptive Group-wise Interaction Network for Multimodal MRI Synthesis. 学习模态感知表征:多模态MRI合成的自适应群智能交互网络。
IF 10.6 1区 医学 Q1 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2026-01-15 DOI: 10.1109/tmi.2026.3654249
Tao Song,Yicheng Wu,Minhao Hu,Xiangde Luo,Linda Wei,Guotai Wang,Yi Guo,Feng Xu,Shaoting Zhang
Multimodal MR image synthesis aims to generate missing modality images by effectively fusing and mapping from a subset of available MRI modalities. Most existing methods adopt an image-to-image translation paradigm, treating multiple modalities as input channels. However, these approaches often yield sub-optimal results due to the inherent difficulty in achieving precise feature-or semantic-level alignment across modalities. To address these challenges, we propose an Adaptive Group-wise Interaction Network (AGI-Net) that explicitly models both inter-modality and intra-modality relationships for multimodal MR image synthesis. Specifically, feature channels are first partitioned into predefined groups, after which an adaptive rolling mechanism is applied to conventional convolutional kernels to better capture feature and semantic correspondences between different modalities. In parallel, a cross-group attention module is introduced to enable effective feature fusion across groups, thereby enhancing the network's representational capacity. We validate the proposed AGI-Net on the publicly available IXI and BraTS2023 datasets. Experimental results demonstrate that AGI-Net achieves state-of-the-art performance in multimodal MR image synthesis tasks, confirming the effectiveness of its modality-aware interaction design. We release the relevant code at: https://github.com/zunzhumu/Adaptive-Group-wise-Interaction-Network-for-Multimodal-MRI-Synthesis.git.
多模态磁共振图像合成旨在通过有效地融合和映射可用的MRI模态子集来生成缺失的模态图像。大多数现有方法采用图像到图像的翻译范式,将多种模态作为输入通道。然而,这些方法往往产生次优结果,因为在实现跨模式的精确特征或语义级对齐方面存在固有的困难。为了应对这些挑战,我们提出了一种自适应群体智能交互网络(AGI-Net),该网络明确地模拟了多模态MR图像合成的模态间和模态内关系。具体而言,该方法首先将特征通道划分为预定义的组,然后将自适应滚动机制应用于传统卷积核,以更好地捕获不同模态之间的特征和语义对应关系。同时,引入了跨群体关注模块,实现了有效的跨群体特征融合,从而增强了网络的表征能力。我们在公开可用的IXI和BraTS2023数据集上验证了提议的AGI-Net。实验结果表明,AGI-Net在多模态磁共振图像合成任务中达到了最先进的性能,证实了其模态感知交互设计的有效性。我们在https://github.com/zunzhumu/Adaptive-Group-wise-Interaction-Network-for-Multimodal-MRI-Synthesis.git上发布了相关代码。
{"title":"Learning Modality-Aware Representations: Adaptive Group-wise Interaction Network for Multimodal MRI Synthesis.","authors":"Tao Song,Yicheng Wu,Minhao Hu,Xiangde Luo,Linda Wei,Guotai Wang,Yi Guo,Feng Xu,Shaoting Zhang","doi":"10.1109/tmi.2026.3654249","DOIUrl":"https://doi.org/10.1109/tmi.2026.3654249","url":null,"abstract":"Multimodal MR image synthesis aims to generate missing modality images by effectively fusing and mapping from a subset of available MRI modalities. Most existing methods adopt an image-to-image translation paradigm, treating multiple modalities as input channels. However, these approaches often yield sub-optimal results due to the inherent difficulty in achieving precise feature-or semantic-level alignment across modalities. To address these challenges, we propose an Adaptive Group-wise Interaction Network (AGI-Net) that explicitly models both inter-modality and intra-modality relationships for multimodal MR image synthesis. Specifically, feature channels are first partitioned into predefined groups, after which an adaptive rolling mechanism is applied to conventional convolutional kernels to better capture feature and semantic correspondences between different modalities. In parallel, a cross-group attention module is introduced to enable effective feature fusion across groups, thereby enhancing the network's representational capacity. We validate the proposed AGI-Net on the publicly available IXI and BraTS2023 datasets. Experimental results demonstrate that AGI-Net achieves state-of-the-art performance in multimodal MR image synthesis tasks, confirming the effectiveness of its modality-aware interaction design. We release the relevant code at: https://github.com/zunzhumu/Adaptive-Group-wise-Interaction-Network-for-Multimodal-MRI-Synthesis.git.","PeriodicalId":13418,"journal":{"name":"IEEE Transactions on Medical Imaging","volume":"8 1","pages":""},"PeriodicalIF":10.6,"publicationDate":"2026-01-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145971767","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Domain Adaptive Multiple Instance Self-Training for Intraoperative Anomaly Detection. 术中异常检测的领域自适应多实例自训练。
IF 10.6 1区 医学 Q1 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2026-01-15 DOI: 10.1109/tmi.2026.3654087
Ziang Chen,Yiming Ding,Jianchang Zhao,Bo Yi,Jianguo Wei
Intraoperative anomalies cause deviations from the ideal surgical workflow, heightening the risk of consequential errors and complications. Their reliable recognition has traditionally relied on continuous surgeon monitoring, yet automated anomaly detection systems are now indispensable for the safe advancement of assistive and autonomous surgery. However, existing approaches struggle with domain shifts across surgical platforms and unpredictable scenarios in deformable surgical environments. To address this, we propose DA-MIST, a Domain Adaptive Multiple Instance Self-Training framework for weakly supervised anomaly detection. DA-MIST adopts a two-stage training strategy that combines multiple instance learning with self-training, enhanced by a scene-decoupled memory mechanism that disentangles state-irrelevant scene variations from memory banks, preserving only state-discriminative features for robust anomaly identification. Additionally, a state-aware dual-branch attention module integrates Gaussian dynamic and global self-attention for effective temporal reasoning. Evaluated on our newly compiled large-scale endoscopic video dataset encompassing seven representative anomalies, DA-MIST demonstrates strong adaptability across heterogeneous surgical domains, consistently reducing false alarms and enhancing anomaly localization accuracy. Our code and dataset will be available at: https://github.com/iamziang/DA-MIST.
术中异常导致偏离理想的手术流程,增加了相应错误和并发症的风险。传统上,他们的可靠识别依赖于连续的外科医生监测,但自动化异常检测系统现在对于辅助和自主手术的安全进步是不可或缺的。然而,现有的方法在手术平台的领域转移和不可预测的手术环境中挣扎。为了解决这个问题,我们提出了DA-MIST,一个用于弱监督异常检测的领域自适应多实例自训练框架。DA-MIST采用两阶段训练策略,将多实例学习与自训练相结合,并通过场景解耦记忆机制进行增强,该机制将与状态无关的场景变化从记忆库中分离出来,仅保留状态判别特征以进行鲁棒异常识别。此外,一个状态感知的双分支注意模块集成了高斯动态和全局自注意,用于有效的时间推理。在我们新编译的包含七个代表性异常的大规模内窥镜视频数据集上进行评估,DA-MIST在异质手术领域表现出强大的适应性,持续减少误报并提高异常定位的准确性。我们的代码和数据集可以在https://github.com/iamziang/DA-MIST上获得。
{"title":"Domain Adaptive Multiple Instance Self-Training for Intraoperative Anomaly Detection.","authors":"Ziang Chen,Yiming Ding,Jianchang Zhao,Bo Yi,Jianguo Wei","doi":"10.1109/tmi.2026.3654087","DOIUrl":"https://doi.org/10.1109/tmi.2026.3654087","url":null,"abstract":"Intraoperative anomalies cause deviations from the ideal surgical workflow, heightening the risk of consequential errors and complications. Their reliable recognition has traditionally relied on continuous surgeon monitoring, yet automated anomaly detection systems are now indispensable for the safe advancement of assistive and autonomous surgery. However, existing approaches struggle with domain shifts across surgical platforms and unpredictable scenarios in deformable surgical environments. To address this, we propose DA-MIST, a Domain Adaptive Multiple Instance Self-Training framework for weakly supervised anomaly detection. DA-MIST adopts a two-stage training strategy that combines multiple instance learning with self-training, enhanced by a scene-decoupled memory mechanism that disentangles state-irrelevant scene variations from memory banks, preserving only state-discriminative features for robust anomaly identification. Additionally, a state-aware dual-branch attention module integrates Gaussian dynamic and global self-attention for effective temporal reasoning. Evaluated on our newly compiled large-scale endoscopic video dataset encompassing seven representative anomalies, DA-MIST demonstrates strong adaptability across heterogeneous surgical domains, consistently reducing false alarms and enhancing anomaly localization accuracy. Our code and dataset will be available at: https://github.com/iamziang/DA-MIST.","PeriodicalId":13418,"journal":{"name":"IEEE Transactions on Medical Imaging","volume":"17 1","pages":""},"PeriodicalIF":10.6,"publicationDate":"2026-01-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145971769","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Energy-Threshold Bias Calculator: A Physics-Model Based Adaptive Correction Scheme for Photon-Counting CT 能量阈值偏置计算器:一种基于物理模型的光子计数CT自适应校正方案
IF 10.6 1区 医学 Q1 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2026-01-15 DOI: 10.1109/tmi.2026.3654612
Yuting Chen, Yuxiang Xing, Li Zhang, Zhi Deng, Hewei Gao
{"title":"Energy-Threshold Bias Calculator: A Physics-Model Based Adaptive Correction Scheme for Photon-Counting CT","authors":"Yuting Chen, Yuxiang Xing, Li Zhang, Zhi Deng, Hewei Gao","doi":"10.1109/tmi.2026.3654612","DOIUrl":"https://doi.org/10.1109/tmi.2026.3654612","url":null,"abstract":"","PeriodicalId":13418,"journal":{"name":"IEEE Transactions on Medical Imaging","volume":"84 1","pages":""},"PeriodicalIF":10.6,"publicationDate":"2026-01-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145972015","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Medical Microwave Imaging Using Physics-Guided Deep Learning Part 2: The Inverse Solver 医学微波成像使用物理引导的深度学习第2部分:逆求解器
IF 10.6 1区 医学 Q1 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2026-01-14 DOI: 10.1109/tmi.2026.3653974
L. Guo, A. Bialkowski, A. Abbosh
{"title":"Medical Microwave Imaging Using Physics-Guided Deep Learning Part 2: The Inverse Solver","authors":"L. Guo, A. Bialkowski, A. Abbosh","doi":"10.1109/tmi.2026.3653974","DOIUrl":"https://doi.org/10.1109/tmi.2026.3653974","url":null,"abstract":"","PeriodicalId":13418,"journal":{"name":"IEEE Transactions on Medical Imaging","volume":"141 1","pages":""},"PeriodicalIF":10.6,"publicationDate":"2026-01-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145972159","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
UltraMamba: Mamba-based Multimodal Ultrasound Image Adaptive Fusion for Breast Lesion Segmentation. 基于mamba的多模态超声图像自适应融合乳腺病变分割。
IF 10.6 1区 医学 Q1 COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS Pub Date : 2026-01-13 DOI: 10.1109/tmi.2026.3653779
Jiahui Huang,Jiaxin Huang,Mingdu Zhang,Qiong Wang,Xiao-Qing Pei,Ying Hu,Hao Chen,Yan Pang
Multimodal ultrasound imaging, combining B-mode ultrasound, shear wave velocity, and shear wave time, is crucial for diagnosing and treating breast lesions, providing insights into lesion characteristics and tissue properties. However, challenges arise from intermodal feature misalignment and attention shifts due to varied capture methods and an overemphasis on vibrant color data. To tackle these issues, we introduce two innovations: a novel segmentation framework and a comprehensive dataset. The UltraMamba framework utilizes bidirectional alignment between modalities and enhances region-specific information to improve breast lesion segmentation accuracy. Key components include the Cross-Modal Knowledge Interaction module for robust information exchange and the Region-Aware Feature Excitation module to focus on relevant features. We also present the BreLS dataset, the first two-dimensional multimodal ultrasound breast lesion dataset, with paired images from 506 cases, serving as a valuable resource for analysis. UltraMamba shows strong performance on the BreLS dataset, achieving a Dice Similarity Coefficient of 72.16% and an HD95 of 42.02 mm, reflecting improvements of 2.59% in DSC and a 6.78 mm reduction in HD95 compared to the second-best framework, MMCA-NET. These results highlight UltraMamba's potential to enhance segmentation accuracy in clinical settings, facilitating precise treatment planning and, ultimately, leading to improved outcomes. Code: https://github.com/deepang-ai/UltraMamba.
多模态超声成像结合b超、横波速度和横波时间,对乳腺病变的诊断和治疗至关重要,可以深入了解病变特征和组织性质。然而,由于不同的捕获方法和过分强调鲜艳的颜色数据,多式联运特征不对齐和注意力转移带来了挑战。为了解决这些问题,我们引入了两个创新:一个新的分割框架和一个全面的数据集。UltraMamba框架利用模式之间的双向对齐,增强区域特异性信息,以提高乳腺病变分割的准确性。关键组件包括跨模态知识交互模块,用于鲁棒信息交换;区域感知特征激励模块,用于关注相关特征。我们还提出了BreLS数据集,这是第一个二维多模态超声乳腺病变数据集,其中包含来自506例病例的成对图像,作为有价值的分析资源。UltraMamba在brres数据集上表现出色,实现了72.16%的Dice Similarity Coefficient和42.02 mm的HD95,与第二好的框架MMCA-NET相比,DSC提高了2.59%,HD95降低了6.78 mm。这些结果突出了ultramba在临床环境中提高分割准确性的潜力,促进了精确的治疗计划,并最终改善了结果。代码:https://github.com/deepang-ai/UltraMamba。
{"title":"UltraMamba: Mamba-based Multimodal Ultrasound Image Adaptive Fusion for Breast Lesion Segmentation.","authors":"Jiahui Huang,Jiaxin Huang,Mingdu Zhang,Qiong Wang,Xiao-Qing Pei,Ying Hu,Hao Chen,Yan Pang","doi":"10.1109/tmi.2026.3653779","DOIUrl":"https://doi.org/10.1109/tmi.2026.3653779","url":null,"abstract":"Multimodal ultrasound imaging, combining B-mode ultrasound, shear wave velocity, and shear wave time, is crucial for diagnosing and treating breast lesions, providing insights into lesion characteristics and tissue properties. However, challenges arise from intermodal feature misalignment and attention shifts due to varied capture methods and an overemphasis on vibrant color data. To tackle these issues, we introduce two innovations: a novel segmentation framework and a comprehensive dataset. The UltraMamba framework utilizes bidirectional alignment between modalities and enhances region-specific information to improve breast lesion segmentation accuracy. Key components include the Cross-Modal Knowledge Interaction module for robust information exchange and the Region-Aware Feature Excitation module to focus on relevant features. We also present the BreLS dataset, the first two-dimensional multimodal ultrasound breast lesion dataset, with paired images from 506 cases, serving as a valuable resource for analysis. UltraMamba shows strong performance on the BreLS dataset, achieving a Dice Similarity Coefficient of 72.16% and an HD95 of 42.02 mm, reflecting improvements of 2.59% in DSC and a 6.78 mm reduction in HD95 compared to the second-best framework, MMCA-NET. These results highlight UltraMamba's potential to enhance segmentation accuracy in clinical settings, facilitating precise treatment planning and, ultimately, leading to improved outcomes. Code: https://github.com/deepang-ai/UltraMamba.","PeriodicalId":13418,"journal":{"name":"IEEE Transactions on Medical Imaging","volume":"7 1","pages":""},"PeriodicalIF":10.6,"publicationDate":"2026-01-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145961412","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Transactions on Medical Imaging
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1