首页 > 最新文献

IEEE Transactions on Biomedical Engineering最新文献

英文 中文
Segment Anything Model 2: An Application to 2D and 3D Medical Images. 细分任何模型2:应用于2D和3D医学图像。
IF 4.5 2区 医学 Q2 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-12 DOI: 10.1109/TBME.2026.3653267
Haoyu Dong, Hanxue Gu, Yaqian Chen, Jichen Yang, Yuwen Chen, Maciej A Mazurowski

Segment Anything Model (SAM) has gained significant attention because of its ability to segment a variety of objects in images upon providing a prompt. Recently developed SAM 2 has extended this ability to video segmentation, and by substituting the third spatial dimension in 3D images for the time dimension in videos, it opens an opportunity to apply SAM 2 to 3D medical images. In this paper, we extensively evaluate SAM 2's ability to segment both 2D and 3D medical images using 80 prompt strategies across 21 medical imaging datasets, including 2D modalities (X-ray and ultrasound), 3D modalities (magnetic resonance imaging, computed tomography, and positron emission tomography), and surgical videos. We find that in the 2D setting, SAM 2 performs similarly to SAM, while in the 3D setting we observe that: (1) selecting the first mask is more effective than choosing the one with the highest confidence, (2) prompting the slice with the largest object appears is the most cost-effective strategy when only one slice is prompted, (3) box prompts result in higher performance than point prompts at a slightly higher annotation cost, (4) bidirectional propagation outperforms front-to-end propagation, (5) interactive annotation is rarely effective, (6) SAM 2, without fine-tuning, achieves 3D IoU from 0.32 with a single point prompt to 0.51 with a ground truth mask on one slice, and exceeds 0.8 on certain datasets when using box or ground-truth prompts, a level that begins to approach clinical usefulness. These findings demonstrate that SAM 2's ability to segment 3D medical images can be improved with our proposed strategies over the default ones, providing practical guidance for using SAM 2 for prompt-based 3D medical image segmentation.

任何物体分割模型(SAM)由于能够在提供提示的情况下分割图像中的各种物体而受到广泛关注。最近开发的SAM 2已经将这种能力扩展到视频分割,并且通过用3D图像中的第三个空间维度代替视频中的时间维度,它打开了将SAM 2应用于3D医学图像的机会。在本文中,我们广泛评估了SAM 2在21个医学成像数据集中使用80种提示策略分割2D和3D医学图像的能力,这些数据集包括2D模式(x射线和超声波)、3D模式(磁共振成像、计算机断层扫描和正电子发射断层扫描)和手术视频。我们发现,在2D环境中,SAM 2的表现与SAM相似,而在3D环境中,我们观察到:(1)选择第一个掩码比选择置信度最高的掩码更有效;(2)当只提示一个切片时,提示出现最大对象的切片是最经济有效的策略;(3)框提示比点提示性能更高,注释成本略高;(4)双向传播优于前端传播;(5)交互式注释很少有效;(6)没有微调的SAM 2。使用单点提示实现3D IoU从0.32到0.51,在一个切片上使用真实值掩膜,并且在使用盒或真实值提示时在某些数据集上超过0.8,这一水平开始接近临床有用性。这些结果表明,与默认策略相比,我们提出的策略可以提高SAM 2对3D医学图像的分割能力,为使用SAM 2进行基于提示的3D医学图像分割提供了实践指导。
{"title":"Segment Anything Model 2: An Application to 2D and 3D Medical Images.","authors":"Haoyu Dong, Hanxue Gu, Yaqian Chen, Jichen Yang, Yuwen Chen, Maciej A Mazurowski","doi":"10.1109/TBME.2026.3653267","DOIUrl":"https://doi.org/10.1109/TBME.2026.3653267","url":null,"abstract":"<p><p>Segment Anything Model (SAM) has gained significant attention because of its ability to segment a variety of objects in images upon providing a prompt. Recently developed SAM 2 has extended this ability to video segmentation, and by substituting the third spatial dimension in 3D images for the time dimension in videos, it opens an opportunity to apply SAM 2 to 3D medical images. In this paper, we extensively evaluate SAM 2's ability to segment both 2D and 3D medical images using 80 prompt strategies across 21 medical imaging datasets, including 2D modalities (X-ray and ultrasound), 3D modalities (magnetic resonance imaging, computed tomography, and positron emission tomography), and surgical videos. We find that in the 2D setting, SAM 2 performs similarly to SAM, while in the 3D setting we observe that: (1) selecting the first mask is more effective than choosing the one with the highest confidence, (2) prompting the slice with the largest object appears is the most cost-effective strategy when only one slice is prompted, (3) box prompts result in higher performance than point prompts at a slightly higher annotation cost, (4) bidirectional propagation outperforms front-to-end propagation, (5) interactive annotation is rarely effective, (6) SAM 2, without fine-tuning, achieves 3D IoU from 0.32 with a single point prompt to 0.51 with a ground truth mask on one slice, and exceeds 0.8 on certain datasets when using box or ground-truth prompts, a level that begins to approach clinical usefulness. These findings demonstrate that SAM 2's ability to segment 3D medical images can be improved with our proposed strategies over the default ones, providing practical guidance for using SAM 2 for prompt-based 3D medical image segmentation.</p>","PeriodicalId":13245,"journal":{"name":"IEEE Transactions on Biomedical Engineering","volume":"PP ","pages":""},"PeriodicalIF":4.5,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145959261","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SW-VEI-Net: A Physics-Informed Deep Neural Network for Shear Wave Viscoelasticity Imaging. SW-VEI-Net:用于横波粘弹性成像的物理信息深度神经网络。
IF 4.5 2区 医学 Q2 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-12 DOI: 10.1109/TBME.2026.3652121
Haoming Lin, Zhongjun Ma, Yunxiang Wang, Muqing Lin, Shuming Xu, Mian Chen, Minhua Lu, Siping Chen, Xin Chen

Quantitative viscoelasticity imaging via shear wave elastography (SWE) remains challenging due to complex wave physics and limitations of conventional reconstruction methods. To address this, we present SW-VEI-Net, a physics-informed neural network (PINN) that simultaneously reconstructs the shear elastic modulus and viscous modulus by integrating viscoelastic wave equations into a dual-network architecture. The framework employs a dual-loss function to balance data fidelity and physics-based regularization, significantly reducing reliance on empirical data while improving interpretability. Extensive validation on tissue-mimicking phantoms, rat liver fibrosis model, and clinical cases demonstrates that SW-VEI-Net outperforms state-of-the-art SWE methods. Compared to SWENet (a PINN-based method using a linear elastic model), SW-VEI-Net not only enables simultaneous assessment of shear elastic and viscous moduli, but also achieves higher accuracy in shear elastic modulus reconstruction. Furthermore, when benchmarked against the dispersion fitting (DF) method (based on a viscoelastic model), SW-VEI-Net produces comparable viscoelastic parameter maps while exhibiting enhanced robustness and consistency. For liver fibrosis staging, SW-VEI-Net achieves AUC values of 0.85 ($geq$F2) and 0.91 ($=$F4) based on elastic modulus classification, surpassing both SWENet (0.84, 0.85) and DF (0.78, 0.88). Additional validation in healthy volunteers shows strong agreement with a commercial ultrasound system. By synergizing deep learning with fundamental wave physics, this study represents a significant advancement in SWE, offering substantial clinical potential for early detection of hepatic fibrosis and malignant lesions through precise viscoelastic biomarker mapping.

由于复杂的波物理特性和传统重建方法的局限性,通过剪切波弹性成像(SWE)进行定量粘弹性成像仍然具有挑战性。为了解决这个问题,我们提出了SW-VEI-Net,这是一个物理信息神经网络(PINN),通过将粘弹性波动方程集成到双网络架构中,同时重建剪切弹性模量和粘性模量。该框架采用双损失函数来平衡数据保真度和基于物理的正则化,显著减少了对经验数据的依赖,同时提高了可解释性。对组织模拟模型、大鼠肝纤维化模型和临床病例的广泛验证表明,SW-VEI-Net优于最先进的SWE方法。与SWENet(基于pup的线性弹性模型方法)相比,SW-VEI-Net不仅可以同时评估剪切弹性模量和粘性模量,而且在剪切弹性模量重建方面具有更高的精度。此外,当与离散拟合(DF)方法(基于粘弹性模型)进行基准测试时,SW-VEI-Net可以生成可比的粘弹性参数图,同时表现出增强的鲁棒性和一致性。对于肝纤维化分期,基于弹性模量分类,SW-VEI-Net的AUC值分别为0.85 ($geq$ F2)和0.91 ($=$ F4),超过了SWENet(0.84, 0.85)和DF(0.78, 0.88)。在健康志愿者中进行的额外验证显示与商业超声系统有很强的一致性。通过将深度学习与基本波物理相结合,该研究代表了SWE的重大进步,通过精确的粘弹性生物标志物定位,为早期检测肝纤维化和恶性病变提供了巨大的临床潜力。
{"title":"SW-VEI-Net: A Physics-Informed Deep Neural Network for Shear Wave Viscoelasticity Imaging.","authors":"Haoming Lin, Zhongjun Ma, Yunxiang Wang, Muqing Lin, Shuming Xu, Mian Chen, Minhua Lu, Siping Chen, Xin Chen","doi":"10.1109/TBME.2026.3652121","DOIUrl":"https://doi.org/10.1109/TBME.2026.3652121","url":null,"abstract":"<p><p>Quantitative viscoelasticity imaging via shear wave elastography (SWE) remains challenging due to complex wave physics and limitations of conventional reconstruction methods. To address this, we present SW-VEI-Net, a physics-informed neural network (PINN) that simultaneously reconstructs the shear elastic modulus and viscous modulus by integrating viscoelastic wave equations into a dual-network architecture. The framework employs a dual-loss function to balance data fidelity and physics-based regularization, significantly reducing reliance on empirical data while improving interpretability. Extensive validation on tissue-mimicking phantoms, rat liver fibrosis model, and clinical cases demonstrates that SW-VEI-Net outperforms state-of-the-art SWE methods. Compared to SWENet (a PINN-based method using a linear elastic model), SW-VEI-Net not only enables simultaneous assessment of shear elastic and viscous moduli, but also achieves higher accuracy in shear elastic modulus reconstruction. Furthermore, when benchmarked against the dispersion fitting (DF) method (based on a viscoelastic model), SW-VEI-Net produces comparable viscoelastic parameter maps while exhibiting enhanced robustness and consistency. For liver fibrosis staging, SW-VEI-Net achieves AUC values of 0.85 ($geq$F2) and 0.91 ($=$F4) based on elastic modulus classification, surpassing both SWENet (0.84, 0.85) and DF (0.78, 0.88). Additional validation in healthy volunteers shows strong agreement with a commercial ultrasound system. By synergizing deep learning with fundamental wave physics, this study represents a significant advancement in SWE, offering substantial clinical potential for early detection of hepatic fibrosis and malignant lesions through precise viscoelastic biomarker mapping.</p>","PeriodicalId":13245,"journal":{"name":"IEEE Transactions on Biomedical Engineering","volume":"PP ","pages":""},"PeriodicalIF":4.5,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145959277","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Feasibility of dual probe pulse wave imaging of the abdominal aorta. 腹主动脉双探头脉冲波成像的可行性。
IF 4.5 2区 医学 Q2 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-12 DOI: 10.1109/TBME.2026.3651584
Larissa C Jansen, Richard G P Lopata, Hans-Martin Schwab

Pulse wave velocity (PWV) is an indirect measure of vessel stiffness, that has the potential to serve as a meaningful parameter for risk stratification of vascular diseases, such as abdominal aortic aneurysms (AAAs). However, assessing the PWV and pulse wave patterns in the complete abdominal aorta using ultrasound-based pulse wave imaging (PWI) is challenging due to the limited field of view (FOV) and contrast of a single ultrasound (US) probe. Hence, an approach is required that can capture distension of aortas with different levels of stiffness accurately in a large FOV. Therefore, we propose PWI based on dual probe, bistatic US. Single and dual probe ultrasound simulations were performed using finite element models of pressure waves propagating in aortas with different stiffness levels. Next, the approach was tested on an aorta and AAA mimicking phantom in a mock circulation setup. The simulation results show that the FOV, image quality, and PWV-estimation accuracy improve when using the dual probe approach (accuracy range: 94.9 - 99.8 $%$; R$^{2}$ range: 0.92 - 0.98) compared to conventional US (accuracy range: 12.6 - 93.9 $%$; R$^{2}$ range: 0.52 - 0.91). The approach was successfully expanded to the phantom study, which demonstrated expected wave patterns within a larger FOV. With dual probe PWI of the non-dilated phantom, the R$^{2}$-value improves (monostatic: 0.95; bistatic: 0.96) compared to use of single probe PWI (0.85). The proposed method shows to be promising for PWV-estimations in less compliant vessels with high wave speeds.

脉搏波速度(PWV)是血管刚度的间接测量,有可能作为血管疾病(如腹主动脉瘤(AAAs))风险分层的有意义参数。然而,由于单个超声(US)探头的视野(FOV)和对比度有限,使用基于超声的脉冲波成像(PWI)评估完整腹主动脉的PWV和脉冲波模式具有挑战性。因此,需要一种能够在大视场中准确捕捉不同硬度的主动脉扩张的方法。因此,我们提出了基于双探头、双稳态US的PWI。利用有限元模型对不同刚度水平主动脉内压力波的传播进行了单探头和双探头超声模拟。接下来,该方法在模拟循环装置中对主动脉和AAA模拟幻影进行了测试。仿真结果表明,与传统的US方法(精度范围:12.6 ~ 93.9 $%$;R$^{2}$范围:0.52 ~ 0.91)相比,采用双探头方法的FOV、图像质量和pwv估计精度(精度范围:94.9 ~ 99.8 $%$;R$^{2}$范围:0.92 ~ 0.98)得到了提高。该方法成功地扩展到幻影研究中,在更大的视场内展示了预期的波模式。与使用单探头PWI(0.85)相比,使用非扩张幻体的双探头PWI, R$^{2}$-值(单静:0.95;双静:0.96)得到改善。所提出的方法对于在不太适应的高波速船舶中进行pwv估计是有希望的。
{"title":"Feasibility of dual probe pulse wave imaging of the abdominal aorta.","authors":"Larissa C Jansen, Richard G P Lopata, Hans-Martin Schwab","doi":"10.1109/TBME.2026.3651584","DOIUrl":"https://doi.org/10.1109/TBME.2026.3651584","url":null,"abstract":"<p><p>Pulse wave velocity (PWV) is an indirect measure of vessel stiffness, that has the potential to serve as a meaningful parameter for risk stratification of vascular diseases, such as abdominal aortic aneurysms (AAAs). However, assessing the PWV and pulse wave patterns in the complete abdominal aorta using ultrasound-based pulse wave imaging (PWI) is challenging due to the limited field of view (FOV) and contrast of a single ultrasound (US) probe. Hence, an approach is required that can capture distension of aortas with different levels of stiffness accurately in a large FOV. Therefore, we propose PWI based on dual probe, bistatic US. Single and dual probe ultrasound simulations were performed using finite element models of pressure waves propagating in aortas with different stiffness levels. Next, the approach was tested on an aorta and AAA mimicking phantom in a mock circulation setup. The simulation results show that the FOV, image quality, and PWV-estimation accuracy improve when using the dual probe approach (accuracy range: 94.9 - 99.8 $%$; R$^{2}$ range: 0.92 - 0.98) compared to conventional US (accuracy range: 12.6 - 93.9 $%$; R$^{2}$ range: 0.52 - 0.91). The approach was successfully expanded to the phantom study, which demonstrated expected wave patterns within a larger FOV. With dual probe PWI of the non-dilated phantom, the R$^{2}$-value improves (monostatic: 0.95; bistatic: 0.96) compared to use of single probe PWI (0.85). The proposed method shows to be promising for PWV-estimations in less compliant vessels with high wave speeds.</p>","PeriodicalId":13245,"journal":{"name":"IEEE Transactions on Biomedical Engineering","volume":"PP ","pages":""},"PeriodicalIF":4.5,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145959297","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multimodal Spiking Neural Network With Generalized Distributive Law for Biosignal and Sensory Fusion. 具有广义分配律的多模态脉冲神经网络用于生物信号和感觉融合。
IF 4.5 2区 医学 Q2 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-12 DOI: 10.1109/TBME.2026.3653109
Zenan Huang, Bingrui Guo, Hailing Xu, Haojie Ruan, Donghui Guo

Multimodal signal fusion is a cornerstone of biomedical engineering and intelligent sensing, enabling holistic analysis of heterogeneous sources such as electroencephalography (EEG), peripheral signals, speech, and imaging data. However, integrating diverse modalities in a computationally efficient and biologically plausible manner remains a significant challenge. Transformer-based fusion architectures rely on global cross-attention to integrate multimodal information but incur high computational costs. In contrast, STDP-driven fully connected layers adopt local learning rules, which restrict their ability to autonomously form efficient sparse topologies for complex multimodal tasks. To address these issues, we propose a novel end-to-end framework-the Multimodal Spiking Neural Network (MSNN)-featuring a fusion module grounded in the Generalized Distributive Law (GDL). This principled mechanism provides an efficient and interpretable means of integrating heterogeneous biomedical and sensory signals. The MSNN further incorporates structure-adaptive leaky integrate-and-fire (SALIF) neurons, enabling dynamic optimization of sparse connectivity to enhance fusion efficiency. The proposed MSNN is validated on a range of datasets, demonstrating strong versatility: it achieves binary classification accuracies of 92.29% (valence) and 91.08% (arousal) on the DEAP dataset for affective state decoding and 99.77% on the WESAD dataset for stress detection, while delivering state-of-the-art performance on standard pattern recognition tasks (MNIST & TIDIGITS: 99.01%) and event-driven neuromorphic datasets (MNIST-DVS & N-TIDIGITS: 99.98%). These results demonstrate that MSNN offers an effective and energy-efficient solution for multimodal sensor fusion in biomedical and intelligent sensing applications.

多模态信号融合是生物医学工程和智能传感的基石,能够对脑电图(EEG)、外围信号、语音和成像数据等异构源进行整体分析。然而,以计算效率和生物学上合理的方式整合多种模式仍然是一个重大挑战。基于变压器的融合体系结构依赖全局交叉关注来集成多模态信息,但计算成本较高。相比之下,stdp驱动的全连接层采用局部学习规则,这限制了它们在复杂多模态任务中自主形成高效稀疏拓扑的能力。为了解决这些问题,我们提出了一种新的端到端框架-多模态峰值神经网络(MSNN)-其特征是基于广义分配律(GDL)的融合模块。这一原则机制提供了一种有效且可解释的整合异质生物医学和感官信号的方法。MSNN进一步引入了结构自适应的泄漏集成与火灾(SALIF)神经元,实现了稀疏连接的动态优化,以提高融合效率。所提出的MSNN在一系列数据集上进行了验证,显示出强大的通用性:它在情感状态解码的DEAP数据集上实现了92.29%(价态)和91.08%(唤醒)的二元分类准确率,在压力检测的WESAD数据集上实现了99.77%的二元分类准确率,同时在标准模式识别任务(MNIST和TIDIGITS: 99.01%)和事件驱动的神经形态数据集(MNIST- dvs和N-TIDIGITS: 99.98%)上提供了最先进的性能。这些结果表明,MSNN为生物医学和智能传感应用中的多模态传感器融合提供了一种有效且节能的解决方案。
{"title":"Multimodal Spiking Neural Network With Generalized Distributive Law for Biosignal and Sensory Fusion.","authors":"Zenan Huang, Bingrui Guo, Hailing Xu, Haojie Ruan, Donghui Guo","doi":"10.1109/TBME.2026.3653109","DOIUrl":"https://doi.org/10.1109/TBME.2026.3653109","url":null,"abstract":"<p><p>Multimodal signal fusion is a cornerstone of biomedical engineering and intelligent sensing, enabling holistic analysis of heterogeneous sources such as electroencephalography (EEG), peripheral signals, speech, and imaging data. However, integrating diverse modalities in a computationally efficient and biologically plausible manner remains a significant challenge. Transformer-based fusion architectures rely on global cross-attention to integrate multimodal information but incur high computational costs. In contrast, STDP-driven fully connected layers adopt local learning rules, which restrict their ability to autonomously form efficient sparse topologies for complex multimodal tasks. To address these issues, we propose a novel end-to-end framework-the Multimodal Spiking Neural Network (MSNN)-featuring a fusion module grounded in the Generalized Distributive Law (GDL). This principled mechanism provides an efficient and interpretable means of integrating heterogeneous biomedical and sensory signals. The MSNN further incorporates structure-adaptive leaky integrate-and-fire (SALIF) neurons, enabling dynamic optimization of sparse connectivity to enhance fusion efficiency. The proposed MSNN is validated on a range of datasets, demonstrating strong versatility: it achieves binary classification accuracies of 92.29% (valence) and 91.08% (arousal) on the DEAP dataset for affective state decoding and 99.77% on the WESAD dataset for stress detection, while delivering state-of-the-art performance on standard pattern recognition tasks (MNIST & TIDIGITS: 99.01%) and event-driven neuromorphic datasets (MNIST-DVS & N-TIDIGITS: 99.98%). These results demonstrate that MSNN offers an effective and energy-efficient solution for multimodal sensor fusion in biomedical and intelligent sensing applications.</p>","PeriodicalId":13245,"journal":{"name":"IEEE Transactions on Biomedical Engineering","volume":"PP ","pages":""},"PeriodicalIF":4.5,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145959214","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FocFormer-UNet: UNet With Focal Modulation and Transformers for Ultrasound Needle Tracking Using Photoacoustic Ground Truth. FocFormer-UNet: UNet与聚焦调制和变压器的超声针跟踪利用光声地面真相。
IF 4.5 2区 医学 Q2 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-12 DOI: 10.1109/TBME.2026.3652428
S M Mahim, Md Emamul Hossen, Manojit Pramanik

Ultrasound (US)-guided needle tracking is a critical procedure for various clinical diagnoses and treatment planning, highlighting the need for improved visualization methods to enhance accuracy. While deep learning (DL) techniques have been employed to boost needle visibility in US images, they often rely heavily on manual annotations or simulated datasets, which can introduce biases and limit real-world applicability. Photoacoustic (PA) imaging, known for its high contrast capabilities, offers a promising solution by providing superior needle visualization compared to conventional US images. In this work, we present FocFormer-UNet, a DL network that leverages PA images of the needle as ground truth for training, eliminating the need for manual annotations. This approach significantly improves needle localization accuracy in US images, reducing the reliance on time-consuming manual labeling. FocFormer-UNet achieves excellent needle localization accuracy, demonstrated by a modified Hausdorff distance of 1.43 1.23 and a targeting error of 1.22 1.14 on human clinical dataset, indicating minimal deviation from actual needle positions. Our method offers robust needle tracking across diverse US systems, improving the precision and reliability of US-guided needle insertion procedures. It holds great promise for advancing AI-driven clinical support tools in medical imaging. The following is the source code: https://github.com/DeeplearningBILAB/FocFormer-UNet. Open Science Framework (OSF) provides datasets and checkpoints at: https://osf.io/yxt9v/.

超声(US)引导的针头跟踪是各种临床诊断和治疗计划的关键程序,强调需要改进可视化方法以提高准确性。虽然深度学习(DL)技术已被用于提高美国图像的针尖可见性,但它们通常严重依赖于手动注释或模拟数据集,这可能会引入偏见并限制现实世界的适用性。光声成像(PA)以其高对比度能力而闻名,与传统的美国图像相比,它提供了一种很有前途的解决方案,提供了优越的针头可视化。在这项工作中,我们提出了FocFormer-UNet,这是一种深度学习网络,它利用针的PA图像作为训练的基础真理,消除了手动注释的需要。这种方法显著提高了美国图像中针头定位的准确性,减少了对耗时的人工标记的依赖。FocFormer-UNet实现了优异的针头定位精度,在人类临床数据集上,改进的Hausdorff距离为1.43 1.23,靶向误差为1.22 1.14,与实际针头位置的偏差最小。我们的方法在不同的美国系统中提供强大的针头跟踪,提高了美国引导的针头插入程序的精度和可靠性。它为推进医学成像中人工智能驱动的临床支持工具提供了巨大的希望。源代码如下:https://github.com/DeeplearningBILAB/FocFormer-UNet。开放科学框架(OSF)提供数据集和检查点:https://osf.io/yxt9v/。
{"title":"FocFormer-UNet: UNet With Focal Modulation and Transformers for Ultrasound Needle Tracking Using Photoacoustic Ground Truth.","authors":"S M Mahim, Md Emamul Hossen, Manojit Pramanik","doi":"10.1109/TBME.2026.3652428","DOIUrl":"https://doi.org/10.1109/TBME.2026.3652428","url":null,"abstract":"<p><p>Ultrasound (US)-guided needle tracking is a critical procedure for various clinical diagnoses and treatment planning, highlighting the need for improved visualization methods to enhance accuracy. While deep learning (DL) techniques have been employed to boost needle visibility in US images, they often rely heavily on manual annotations or simulated datasets, which can introduce biases and limit real-world applicability. Photoacoustic (PA) imaging, known for its high contrast capabilities, offers a promising solution by providing superior needle visualization compared to conventional US images. In this work, we present FocFormer-UNet, a DL network that leverages PA images of the needle as ground truth for training, eliminating the need for manual annotations. This approach significantly improves needle localization accuracy in US images, reducing the reliance on time-consuming manual labeling. FocFormer-UNet achieves excellent needle localization accuracy, demonstrated by a modified Hausdorff distance of 1.43 1.23 and a targeting error of 1.22 1.14 on human clinical dataset, indicating minimal deviation from actual needle positions. Our method offers robust needle tracking across diverse US systems, improving the precision and reliability of US-guided needle insertion procedures. It holds great promise for advancing AI-driven clinical support tools in medical imaging. The following is the source code: https://github.com/DeeplearningBILAB/FocFormer-UNet. Open Science Framework (OSF) provides datasets and checkpoints at: https://osf.io/yxt9v/.</p>","PeriodicalId":13245,"journal":{"name":"IEEE Transactions on Biomedical Engineering","volume":"PP ","pages":""},"PeriodicalIF":4.5,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145959234","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Dual-Energy CBCT With Reduced Scatter and Cone Beam Artifacts Using an X-Ray Source Array and Interlaced Spectral Filters. 利用x射线源阵列和隔行光谱滤波器减少散射和锥束伪影的双能CBCT。
IF 4.5 2区 医学 Q2 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-12 DOI: 10.1109/TBME.2026.3651411
Yuanming Hu, Boyuan Li, Shuang Xu, Christina R Inscoe, Donald A Tyndall, Yueh Z Lee, Jianping Lu, Otto Zhou

Objective: To design a dual-energy cone beam computed tomography (DE-CBCT) scanner with reduced scatter and cone beam artifacts.

Methods: The scanner designed for maxillofacial imaging comprises a carbon nanotube (CNT) X-ray source array with multiple focal spots ("sources") and an energy integrating flat panel detector (FPD). The X-ray photons from each focal spot were narrowly collimated in the axial direction and was filtered by interlaced low‑ and high‑energy spectral filters. Two sets of projection images were acquired by sequentially activating the X‑ray beams from each source in one gantry rotation. The projections were processed using a one-step inversion algorithm. An anthropomorphic head phantom, a Defrise phantom and a water-equivalent phantom containing calcium and iodine inserts were used to compare the performance of the new dual-energy multisource CBCT (DE-MS-CBCT) with a conventional DE-CBCT using the same air-kerma.

Results: The DE-MS-CBCT eliminated the cone beam artifacts, reduced the degree of cupping artifacts from 14.53% to 2.94%, and lowered the mean relative error of water density from 15.3% to 1.7%, while the accuracies for iodine and calcium densities were comparable. The contrast-noise-ratios (CNR) of the calcium and iodine inserts against the solid water increased by 4.8%-53.4%.

Conclusion: The DE‑MS‑CBCT reduces scatter and cone‑beam artifacts, increases the image CNR, and enhances accuracy of materials quantification without increasing X-ray exposure compared to the conventional DE-CBCT.

Significance: The results demonstrate a new DE-CBCT method with improved image quality and accuracy of materials quantification without the need for an energy sensitive detector or kV switching.

目的:设计一种减少散射和锥束伪影的双能锥束ct (DE-CBCT)扫描仪。方法:颌面部成像扫描仪由碳纳米管(CNT)多焦点x射线源阵列(“源”)和能量集成平板探测器(FPD)组成。来自每个焦点点的x射线光子在轴向上被窄准直,并由交错的低能和高能谱滤波器过滤。通过在一次龙门旋转中依次激活来自每个源的X射线束,获得两组投影图像。投影使用一步反演算法进行处理。使用人形头部幻像、Defrise幻像和含有钙和碘插入物的水等效幻像来比较新型双能多源CBCT (DE-MS-CBCT)与传统DE-CBCT的性能。结果:DE-MS-CBCT消除了锥束伪影,将拔罐伪影的程度从14.53%降低到2.94%,将水密度的平均相对误差从15.3%降低到1.7%,而碘密度和钙密度的精度相当。钙、碘填料对固体水的比噪比(CNR)提高了4.8% ~ 53.4%。结论:与传统DE-CBCT相比,DE- MS -CBCT在不增加x射线曝光的情况下减少了散射和锥束伪影,提高了图像的CNR,提高了材料定量的准确性。意义:结果证明了一种新的DE-CBCT方法,提高了图像质量和材料量化精度,而不需要能量敏感探测器或kV开关。
{"title":"A Dual-Energy CBCT With Reduced Scatter and Cone Beam Artifacts Using an X-Ray Source Array and Interlaced Spectral Filters.","authors":"Yuanming Hu, Boyuan Li, Shuang Xu, Christina R Inscoe, Donald A Tyndall, Yueh Z Lee, Jianping Lu, Otto Zhou","doi":"10.1109/TBME.2026.3651411","DOIUrl":"https://doi.org/10.1109/TBME.2026.3651411","url":null,"abstract":"<p><strong>Objective: </strong>To design a dual-energy cone beam computed tomography (DE-CBCT) scanner with reduced scatter and cone beam artifacts.</p><p><strong>Methods: </strong>The scanner designed for maxillofacial imaging comprises a carbon nanotube (CNT) X-ray source array with multiple focal spots (\"sources\") and an energy integrating flat panel detector (FPD). The X-ray photons from each focal spot were narrowly collimated in the axial direction and was filtered by interlaced low‑ and high‑energy spectral filters. Two sets of projection images were acquired by sequentially activating the X‑ray beams from each source in one gantry rotation. The projections were processed using a one-step inversion algorithm. An anthropomorphic head phantom, a Defrise phantom and a water-equivalent phantom containing calcium and iodine inserts were used to compare the performance of the new dual-energy multisource CBCT (DE-MS-CBCT) with a conventional DE-CBCT using the same air-kerma.</p><p><strong>Results: </strong>The DE-MS-CBCT eliminated the cone beam artifacts, reduced the degree of cupping artifacts from 14.53% to 2.94%, and lowered the mean relative error of water density from 15.3% to 1.7%, while the accuracies for iodine and calcium densities were comparable. The contrast-noise-ratios (CNR) of the calcium and iodine inserts against the solid water increased by 4.8%-53.4%.</p><p><strong>Conclusion: </strong>The DE‑MS‑CBCT reduces scatter and cone‑beam artifacts, increases the image CNR, and enhances accuracy of materials quantification without increasing X-ray exposure compared to the conventional DE-CBCT.</p><p><strong>Significance: </strong>The results demonstrate a new DE-CBCT method with improved image quality and accuracy of materials quantification without the need for an energy sensitive detector or kV switching.</p>","PeriodicalId":13245,"journal":{"name":"IEEE Transactions on Biomedical Engineering","volume":"PP ","pages":""},"PeriodicalIF":4.5,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145959189","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fusing Tabular Features and Deep Learning for Fetal Heart Rate Analysis: A Clinically Interpretable Model for Fetal Compromise Detection. 融合表格特征和深度学习胎儿心率分析:胎儿损伤检测的临床可解释模型。
IF 4.5 2区 医学 Q2 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-12 DOI: 10.1109/TBME.2026.3652309
Lochana Mendis, Debjyoti Karmakar, Marimuthu Palaniswami, Fiona Brownfoot, Emerson Keenan

Objective: Cardiotocography (CTG) is commonly used to monitor fetal heart rate (FHR) and assess fetal well-being during labor. However, its effectiveness in reducing adverse outcomes remains limited due to low sensitivity and high false-positive rates. This study aims to develop an interpretable deep learning model that fuses FHR time series with tabular clinical features to improve prediction of fetal compromise (umbilical artery pH $< $ 7.05).

Methods: We introduce Fusion ResNet, a novel architecture combining residual convolutional networks for FHR signal processing with a parallel neural network for tabular features. The model was trained and internally validated on a private dataset of 9,887 FHR recordings. External validation was performed on the open-access CTU-UHB dataset comprising 552 recordings. Model interpretability was evaluated using Shapley Additive Explanations (SHAP) and Gradient-Weighted Class Activation Mapping (Grad-CAM).

Results: Fusion ResNet achieved a mean area under the ROC curve (AUC) of 0.77 during internal cross-validation and a state-of-the-art AUC of 0.84 on the CTU-UHB dataset, outperforming existing deep learning approaches. SHAP analysis identified key clinical features contributing to predictions, while Grad-CAM highlighted salient FHR patterns linked to fetal compromise.

Conclusion: The proposed model enhances predictive accuracy while providing clinically meaningful explanations, enabling more transparent and reliable CTG interpretation.

Significance: This work demonstrates the potential of interpretable deep learning to improve fetal monitoring by integrating multimodal data, supporting timely and informed decision-making in obstetric care.

目的:心脏造影(CTG)是一种常用的监测胎儿心率(FHR)和评估胎儿健康的方法。然而,由于低敏感性和高假阳性率,其在减少不良后果方面的有效性仍然有限。本研究旨在开发一种可解释的深度学习模型,该模型融合了FHR时间序列和表格临床特征,以提高胎儿妥协(脐带动脉pH $< $ 7.05)的预测。方法:我们引入Fusion ResNet,这是一种结合残差卷积网络和并行神经网络处理表格特征的新架构。该模型在9,887个FHR记录的私人数据集上进行了训练和内部验证。对包括552条录音的开放获取的CTU-UHB数据集进行外部验证。采用Shapley加性解释(SHAP)和梯度加权类激活映射(Grad-CAM)评价模型的可解释性。结果:Fusion ResNet在内部交叉验证中获得了0.77的ROC曲线下平均面积(AUC),在CTU-UHB数据集上获得了0.84的最先进的AUC,优于现有的深度学习方法。SHAP分析确定了有助于预测的关键临床特征,而Grad-CAM强调了与胎儿损害相关的突出FHR模式。结论:提出的模型在提供有临床意义的解释的同时,提高了预测精度,使CTG解释更加透明和可靠。意义:这项工作证明了可解释深度学习的潜力,通过整合多模式数据来改善胎儿监测,支持及时和明智的产科护理决策。
{"title":"Fusing Tabular Features and Deep Learning for Fetal Heart Rate Analysis: A Clinically Interpretable Model for Fetal Compromise Detection.","authors":"Lochana Mendis, Debjyoti Karmakar, Marimuthu Palaniswami, Fiona Brownfoot, Emerson Keenan","doi":"10.1109/TBME.2026.3652309","DOIUrl":"https://doi.org/10.1109/TBME.2026.3652309","url":null,"abstract":"<p><strong>Objective: </strong>Cardiotocography (CTG) is commonly used to monitor fetal heart rate (FHR) and assess fetal well-being during labor. However, its effectiveness in reducing adverse outcomes remains limited due to low sensitivity and high false-positive rates. This study aims to develop an interpretable deep learning model that fuses FHR time series with tabular clinical features to improve prediction of fetal compromise (umbilical artery pH $< $ 7.05).</p><p><strong>Methods: </strong>We introduce Fusion ResNet, a novel architecture combining residual convolutional networks for FHR signal processing with a parallel neural network for tabular features. The model was trained and internally validated on a private dataset of 9,887 FHR recordings. External validation was performed on the open-access CTU-UHB dataset comprising 552 recordings. Model interpretability was evaluated using Shapley Additive Explanations (SHAP) and Gradient-Weighted Class Activation Mapping (Grad-CAM).</p><p><strong>Results: </strong>Fusion ResNet achieved a mean area under the ROC curve (AUC) of 0.77 during internal cross-validation and a state-of-the-art AUC of 0.84 on the CTU-UHB dataset, outperforming existing deep learning approaches. SHAP analysis identified key clinical features contributing to predictions, while Grad-CAM highlighted salient FHR patterns linked to fetal compromise.</p><p><strong>Conclusion: </strong>The proposed model enhances predictive accuracy while providing clinically meaningful explanations, enabling more transparent and reliable CTG interpretation.</p><p><strong>Significance: </strong>This work demonstrates the potential of interpretable deep learning to improve fetal monitoring by integrating multimodal data, supporting timely and informed decision-making in obstetric care.</p>","PeriodicalId":13245,"journal":{"name":"IEEE Transactions on Biomedical Engineering","volume":"PP ","pages":""},"PeriodicalIF":4.5,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145959295","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Shear Wave Anisotropic Imaging for Pennate Muscle Assessment Using a Tilted Supersonic Push with Elliptical Analytical Inversion. 横波各向异性成像在矢状肌评估中的应用倾斜超音速推力与椭圆解析反演。
IF 4.5 2区 医学 Q2 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-05 DOI: 10.1109/TBME.2026.3651219
Guo-Xuan Xu, Chien Chen, Chih-Chung Huang

Accurate muscle anisotropy assessment is crucial for understanding muscle mechanics and diagnosing pathologies. Shear wave (SW) elastography struggles with the varying fiber orientations in pennate muscles. Rotating the ultrasound probe offers a solution but is cumbersome in clinical practice. This study presents a tilted supersonic push (TSP) method with elliptical analytical inversion to overcome this limitation. TSP method can generate multi-angle (0°-15°) SWs within a single scan plane, creating an elliptical shear wave velocity (SWV) distribution that enables calculation of fiber-aligned and perpendicular SWVs without probe rotation. The TSP method's accuracy was validated through ex vivo experiments on porcine muscles, and in vivo studies on human gastrocnemius muscles. Results consistently demonstrated accurate SWV measurements, even in the presence of significant pennate angles. For instance, in ex vivo porcine muscles with a 25° pennate angle, TSP corrected longitudinal and transverse SWVs of 3.33 m/s and 2.15 m/s, respectively, consistent with reference values without pennate angle obtained via the traditional rotation method. Similarly, in vivo measurements on human gastrocnemius muscle showed longitudinal and transverse SWVs of 2.55 m/s and 1.21 m/s in a relaxed state, increasing to 4.07 m/s and 1.70 m/s during stretching. These findings highlight the method's ability to capture dynamic changes in muscle stiffness. The TSP method provides a clinically viable and robust approach for comprehensive muscle anisotropy assessment, especially in complex pennate muscles. This technique simplifies the measurement process and offers potential for improved diagnosis and management of musculoskeletal disorders.

准确的肌肉各向异性评估是理解肌肉力学和诊断病理的关键。横波(SW)弹性成像与羽状肌中不同的纤维方向作斗争。旋转超声探头提供了一种解决方案,但在临床实践中很麻烦。为了克服这一限制,本文提出了一种带有椭圆解析反演的倾斜超音速推力(TSP)方法。TSP方法可以在单个扫描平面内生成多角度(0°-15°)剪切波速(SWV),形成椭圆剪切波速(SWV)分布,可以在不旋转探针的情况下计算光纤对准和垂直的剪切波速。通过猪肌肉的离体实验和人腓肠肌的体内实验验证了TSP方法的准确性。结果一致证明了准确的SWV测量,即使存在显著的pennate角。例如,在猪离体肌肉中,当角为25°时,TSP校正的纵向swv和横向swv分别为3.33 m/s和2.15 m/s,与传统旋转方法在没有角的情况下得到的参考值一致。同样,人体腓肠肌的体内测量结果显示,在放松状态下,纵向和横向swv分别为2.55 m/s和1.21 m/s,在拉伸状态下分别增加到4.07 m/s和1.70 m/s。这些发现突出了该方法捕捉肌肉僵硬度动态变化的能力。TSP方法为综合肌肉各向异性评估提供了一种临床可行且稳健的方法,特别是在复杂的pennate肌中。这项技术简化了测量过程,并为改善肌肉骨骼疾病的诊断和管理提供了潜力。
{"title":"Shear Wave Anisotropic Imaging for Pennate Muscle Assessment Using a Tilted Supersonic Push with Elliptical Analytical Inversion.","authors":"Guo-Xuan Xu, Chien Chen, Chih-Chung Huang","doi":"10.1109/TBME.2026.3651219","DOIUrl":"https://doi.org/10.1109/TBME.2026.3651219","url":null,"abstract":"<p><p>Accurate muscle anisotropy assessment is crucial for understanding muscle mechanics and diagnosing pathologies. Shear wave (SW) elastography struggles with the varying fiber orientations in pennate muscles. Rotating the ultrasound probe offers a solution but is cumbersome in clinical practice. This study presents a tilted supersonic push (TSP) method with elliptical analytical inversion to overcome this limitation. TSP method can generate multi-angle (0°-15°) SWs within a single scan plane, creating an elliptical shear wave velocity (SWV) distribution that enables calculation of fiber-aligned and perpendicular SWVs without probe rotation. The TSP method's accuracy was validated through ex vivo experiments on porcine muscles, and in vivo studies on human gastrocnemius muscles. Results consistently demonstrated accurate SWV measurements, even in the presence of significant pennate angles. For instance, in ex vivo porcine muscles with a 25° pennate angle, TSP corrected longitudinal and transverse SWVs of 3.33 m/s and 2.15 m/s, respectively, consistent with reference values without pennate angle obtained via the traditional rotation method. Similarly, in vivo measurements on human gastrocnemius muscle showed longitudinal and transverse SWVs of 2.55 m/s and 1.21 m/s in a relaxed state, increasing to 4.07 m/s and 1.70 m/s during stretching. These findings highlight the method's ability to capture dynamic changes in muscle stiffness. The TSP method provides a clinically viable and robust approach for comprehensive muscle anisotropy assessment, especially in complex pennate muscles. This technique simplifies the measurement process and offers potential for improved diagnosis and management of musculoskeletal disorders.</p>","PeriodicalId":13245,"journal":{"name":"IEEE Transactions on Biomedical Engineering","volume":"PP ","pages":""},"PeriodicalIF":4.5,"publicationDate":"2026-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145905804","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Spatio-Temporal Progressive Attention Model for EEG Classification in Rapid Serial Visual Presentation Task. 快速序列视觉呈现任务脑电分类的时空递进注意模型。
IF 4.5 2区 医学 Q2 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-01 DOI: 10.1109/TBME.2025.3579491
Yang Li, Wei Liu, Tianzhi Feng, Fu Li, Chennan Wu, Boxun Fu, Zhifu Zhao, Xiaotian Wang, Guangming Shi

As a type of multi-dimensional sequential data, the spatial and temporal dependencies of electroencephalogram (EEG) signals should be further investigated. Thus, in this paper, we propose a novel spatial-temporalprogressive attention model (STPAM) to improve EEG classification in rapid serial visual presentation(RSVP) tasks. STPAM employs a progressive approach using three sequential spatial experts to learn brain region topology and mitigate interference from irrelevant areas. Each expert refines EEG electrode selection, guiding subsequent experts to focus on significant spatial information, thus enhancing signals from key regions. Subsequently, based on the above spatially-enhanced features, three temporal experts progressively capture temporal dependencies by focusing attention on crucial EEG time slices. Except for the above EEG classification method, in this paper, we build a novel Infrared RSVP Dataset (IRED) which is based on dim infrared images with small targets for the first time, and conduct extensive experiments on it. Experimental results demonstrate that STPAM outperforms all baselines, achieving 2.02% and 1.17% on the public dataset and IRED dataset, respectively.

脑电图信号作为一种多维序列数据,其时空依赖性有待进一步研究。为此,我们提出了一种新的时空递进注意模型(STPAM)来改进快速序列视觉呈现(RSVP)任务的脑电分类。STPAM采用循序渐进的方法,使用三个连续的空间专家来学习大脑区域拓扑并减轻不相关区域的干扰。每位专家对EEG电极的选择进行细化,引导后续专家关注有意义的空间信息,从而增强关键区域的信号。随后,三位时间专家基于上述空间增强特征,通过将注意力集中在关键的EEG时间片上,逐步捕获时间依赖性。除上述脑电分类方法外,本文首次构建了基于小目标弱红外图像的红外RSVP数据集(IRED),并对其进行了大量实验。实验结果表明,STPAM优于所有基线,在公共数据集和IRED数据集上分别达到2.02%和1.17%。
{"title":"Spatio-Temporal Progressive Attention Model for EEG Classification in Rapid Serial Visual Presentation Task.","authors":"Yang Li, Wei Liu, Tianzhi Feng, Fu Li, Chennan Wu, Boxun Fu, Zhifu Zhao, Xiaotian Wang, Guangming Shi","doi":"10.1109/TBME.2025.3579491","DOIUrl":"10.1109/TBME.2025.3579491","url":null,"abstract":"<p><p>As a type of multi-dimensional sequential data, the spatial and temporal dependencies of electroencephalogram (EEG) signals should be further investigated. Thus, in this paper, we propose a novel spatial-temporalprogressive attention model (STPAM) to improve EEG classification in rapid serial visual presentation(RSVP) tasks. STPAM employs a progressive approach using three sequential spatial experts to learn brain region topology and mitigate interference from irrelevant areas. Each expert refines EEG electrode selection, guiding subsequent experts to focus on significant spatial information, thus enhancing signals from key regions. Subsequently, based on the above spatially-enhanced features, three temporal experts progressively capture temporal dependencies by focusing attention on crucial EEG time slices. Except for the above EEG classification method, in this paper, we build a novel Infrared RSVP Dataset (IRED) which is based on dim infrared images with small targets for the first time, and conduct extensive experiments on it. Experimental results demonstrate that STPAM outperforms all baselines, achieving 2.02% and 1.17% on the public dataset and IRED dataset, respectively.</p>","PeriodicalId":13245,"journal":{"name":"IEEE Transactions on Biomedical Engineering","volume":"PP ","pages":"191-207"},"PeriodicalIF":4.5,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144283749","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Development and Kinematics Optimization of a Human-Compatible Rope-Driven Ankle Rehabilitation Robot Based on Foot-Ankle IFHA Identification. 基于足踝IFHA识别的人兼容绳驱动踝关节康复机器人研制及运动学优化。
IF 4.5 2区 医学 Q2 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-01 DOI: 10.1109/TBME.2025.3576841
Jingke Song, Jianjun Zhang, Jun Wei, Chenglei Liu, Xiankun Zhao, Cunjin Ai

To address the mismatch between current ankle rehabilitation robots and natural human motion, which affects rehabilitation efficacy, this paper uses screw theory and motion capture experiments to identify the instantaneous finite helical motion axis (IFHA) of the human ankle joint. It determines the distribution law of the IFHA and twist pitch (TP) of the ankle, and designs a human-machine motion compatible rope-driven ankle joint rehabilitation robot that meets the needs of human ankle joint rehabilitation. Firstly, human ankle motion trajectories are captured using the VICON system and IMU, and the experimental data are processed according to screw theory to obtain the distribution law of the IFHA and the range of TP. Secondly, the ankle joint's motion characteristics from the experiment inform the constraint characteristics of the rehabilitation mechanism, which are then mapped into a novel parallel rope-driven ankle rehabilitation robot to meet rehabilitation needs. Thirdly, the kinematic model of the novel mechanism is established, and its kinematic performance and singular configurations are analyzed based on the motion/force transmission index, guiding the optimization of the driving rope layout and mechanism scale parameters. Finally, an experimental platform is built to validate the human-machine motion compatibility, safety, comfort, and effectiveness of the rehabilitation robot.

针对目前踝关节康复机器人与人体自然运动不匹配而影响康复效果的问题,本文利用螺旋理论和运动捕捉实验对人体踝关节的瞬时有限螺旋运动轴(IFHA)进行了识别。确定了踝关节IFHA和扭距(TP)的分布规律,设计了一种满足人体踝关节康复需要的人机运动兼容的绳驱动踝关节康复机器人。首先,利用VICON系统和IMU捕获人体踝关节运动轨迹,并根据螺旋理论对实验数据进行处理,得到IFHA的分布规律和TP的范围;其次,根据实验得到的踝关节运动特征,给出康复机构的约束特征,并将约束特征映射到一种新型的并联绳驱动踝关节康复机器人中,以满足康复需求。第三,建立了新型机构的运动学模型,并基于运动/力传递指标分析了其运动性能和奇异构型,指导了驱动绳布置和机构尺度参数的优化。最后,搭建实验平台,验证康复机器人的人机运动兼容性、安全性、舒适性和有效性。
{"title":"Development and Kinematics Optimization of a Human-Compatible Rope-Driven Ankle Rehabilitation Robot Based on Foot-Ankle IFHA Identification.","authors":"Jingke Song, Jianjun Zhang, Jun Wei, Chenglei Liu, Xiankun Zhao, Cunjin Ai","doi":"10.1109/TBME.2025.3576841","DOIUrl":"https://doi.org/10.1109/TBME.2025.3576841","url":null,"abstract":"<p><p>To address the mismatch between current ankle rehabilitation robots and natural human motion, which affects rehabilitation efficacy, this paper uses screw theory and motion capture experiments to identify the instantaneous finite helical motion axis (IFHA) of the human ankle joint. It determines the distribution law of the IFHA and twist pitch (TP) of the ankle, and designs a human-machine motion compatible rope-driven ankle joint rehabilitation robot that meets the needs of human ankle joint rehabilitation. Firstly, human ankle motion trajectories are captured using the VICON system and IMU, and the experimental data are processed according to screw theory to obtain the distribution law of the IFHA and the range of TP. Secondly, the ankle joint's motion characteristics from the experiment inform the constraint characteristics of the rehabilitation mechanism, which are then mapped into a novel parallel rope-driven ankle rehabilitation robot to meet rehabilitation needs. Thirdly, the kinematic model of the novel mechanism is established, and its kinematic performance and singular configurations are analyzed based on the motion/force transmission index, guiding the optimization of the driving rope layout and mechanism scale parameters. Finally, an experimental platform is built to validate the human-machine motion compatibility, safety, comfort, and effectiveness of the rehabilitation robot.</p>","PeriodicalId":13245,"journal":{"name":"IEEE Transactions on Biomedical Engineering","volume":"73 1","pages":"40-53"},"PeriodicalIF":4.5,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145855793","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Transactions on Biomedical Engineering
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1