首页 > 最新文献

IEEE Transactions on Human-Machine Systems最新文献

英文 中文
Automotive Cockpit-Driving Integration for Human-Centric Autonomous Driving: A Survey 以人为中心的自动驾驶汽车座舱驾驶集成研究
IF 4.4 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-10-23 DOI: 10.1109/THMS.2025.3616947
Zhongpan Zhu;Shuaijie Zhao;Mobing Cai;Cheng Wang;Aimin Du
Intelligent driving aims to handle dynamic driving tasks in complex environments, while driver behavior onboard is less focused. In contrast, an intelligent cockpit mainly focuses on interacting with a driver, with limited connection to the driving scenarios. Since the driver onboard could affect the driving strategy significantly and thus have nonnegligible safety implications on an autonomous vehicle, a cockpit-driving integration (CDI) is generally essential to take the driver’s behavior and intention into account when shaping the driving strategy. However, no comprehensive review of current existing CDI technologies is conducted despite the significant role of CDI in safe driving. Therefore, we are motivated to summarize the state-of-the-art of CDI methods and investigate the development trends of CDI. To this end, we identify thoroughly current applications of CDI for the perception and decision-making of autonomous vehicles and highlight critical issues that urgently need to be addressed. Additionally, we propose a lifelong learning framework based on evolvable neural networks as solutions for future CDI. Finally, challenges and future work are discussed. The work provides useful insights for developers regarding designing safe and human-centric autonomous vehicles.
智能驾驶旨在处理复杂环境下的动态驾驶任务,而车载驾驶员的行为则不太关注。相比之下,智能驾驶舱主要侧重于与驾驶员的互动,与驾驶场景的联系有限。由于驾驶员对自动驾驶汽车的驾驶策略影响很大,因此对自动驾驶汽车的安全影响不容忽视,因此在制定驾驶策略时,将驾驶员的行为和意图考虑在内的驾驶舱驾驶集成(CDI)通常是必不可少的。然而,尽管CDI在安全驾驶中发挥着重要作用,但目前还没有对现有的CDI技术进行全面的审查。因此,我们有动力总结CDI方法的最新进展,并探讨CDI的发展趋势。为此,我们彻底确定了当前CDI在自动驾驶汽车感知和决策方面的应用,并强调了迫切需要解决的关键问题。此外,我们提出了一个基于可进化神经网络的终身学习框架作为未来CDI的解决方案。最后,对面临的挑战和今后的工作进行了讨论。这项工作为开发人员设计安全和以人为本的自动驾驶汽车提供了有用的见解。
{"title":"Automotive Cockpit-Driving Integration for Human-Centric Autonomous Driving: A Survey","authors":"Zhongpan Zhu;Shuaijie Zhao;Mobing Cai;Cheng Wang;Aimin Du","doi":"10.1109/THMS.2025.3616947","DOIUrl":"https://doi.org/10.1109/THMS.2025.3616947","url":null,"abstract":"Intelligent driving aims to handle dynamic driving tasks in complex environments, while driver behavior onboard is less focused. In contrast, an intelligent cockpit mainly focuses on interacting with a driver, with limited connection to the driving scenarios. Since the driver onboard could affect the driving strategy significantly and thus have nonnegligible safety implications on an autonomous vehicle, a cockpit-driving integration (CDI) is generally essential to take the driver’s behavior and intention into account when shaping the driving strategy. However, no comprehensive review of current existing CDI technologies is conducted despite the significant role of CDI in safe driving. Therefore, we are motivated to summarize the state-of-the-art of CDI methods and investigate the development trends of CDI. To this end, we identify thoroughly current applications of CDI for the perception and decision-making of autonomous vehicles and highlight critical issues that urgently need to be addressed. Additionally, we propose a lifelong learning framework based on evolvable neural networks as solutions for future CDI. Finally, challenges and future work are discussed. The work provides useful insights for developers regarding designing safe and human-centric autonomous vehicles.","PeriodicalId":48916,"journal":{"name":"IEEE Transactions on Human-Machine Systems","volume":"55 6","pages":"1016-1032"},"PeriodicalIF":4.4,"publicationDate":"2025-10-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145652146","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Neurosurgical Craniotomy Training System Based on Haptic Virtual Reality Simulation 基于触觉虚拟现实仿真的神经外科开颅训练系统
IF 4.4 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-10-17 DOI: 10.1109/THMS.2025.3616313
Guobin Zhang;Keliang Li;Qiyuan Sun;Wenqi Wu;Shuai Li;Zhenzhong Liu
Traditional neurosurgical training modes face challenges including high costs, limited resources, lengthy learning curves, and difficulties in personalized training. In this article, we developed an immersive neurosurgical craniotomy virtual training system (NeuroSimulator) that integrates haptic feedback, enabling comprehensive surgical skill learning through an operator-control interface. Specifically, we constructed the comprehensive neurosurgical craniotomy surgical procedural (CNCSP) dataset to guide operators in repetitive learning and personalized training of relevant surgical skills. To address surgical site model rendering complexity, we proposed an algorithm that integrates vertex curvature and edge-length cost calculation factors (VC&ECL-QEM), resolving the incompatibility between surgical area image rendering quality and efficiency. For intracranial soft tissue haptic deformation, we developed a hybrid soft tissue haptic deformation (HBD) model that combines mass-spring and volumetric elements, addressing the collapse and distortion issues of traditional models and achieving more realistic soft tissue haptic deformation. Experimental results demonstrate that VC&ECL-QEM can simplify nonsurgical area feature preservation while maintaining surgical site detail features, reflecting the effectiveness of model simplification. The HBD model focuses on improving soft tissue deformation realism and shows high consistency with finite element model deformation effects. A total of 83 participants highly recognized NeuroSimulator’s system performance in terms of operational compliance, rendering real-time performance, and deformation realism, achieving effective improvements in skill proficiency metrics including operation time, ineffective operations, guidance requests, and operation scores. NeuroSimulator provides an innovative, efficient, and practical solution for neurosurgical training and is expected to play an increasingly important role in medical education and clinical skill enhancement.
传统的神经外科培训模式面临着成本高、资源有限、学习曲线长、个性化培训困难等挑战。在本文中,我们开发了一个沉浸式神经外科开颅虚拟训练系统(NeuroSimulator),该系统集成了触觉反馈,通过操作员控制界面实现全面的手术技能学习。具体而言,我们构建了综合神经外科开颅手术程序(CNCSP)数据集,以指导操作员重复学习和个性化培训相关手术技能。针对手术部位模型绘制的复杂性,提出了一种融合顶点曲率和边长成本计算因素的算法(VC&ECL-QEM),解决了手术部位图像绘制质量与效率之间的不兼容问题。对于颅内软组织触觉变形,我们开发了一种结合质量-弹簧和体积元素的混合软组织触觉变形(HBD)模型,解决了传统模型的崩溃和变形问题,实现了更真实的软组织触觉变形。实验结果表明,VC&ECL-QEM可以在保留手术部位细节特征的同时简化非手术区域特征的保存,反映了模型简化的有效性。HBD模型注重提高软组织变形真实感,与有限元模型变形效果具有较高的一致性。共有83名参与者在操作遵从性、实时渲染性能和变形真实感方面高度认可NeuroSimulator的系统性能,并在操作时间、无效操作、指导请求和操作分数等技能熟练度指标上取得了有效改进。神经模拟器为神经外科培训提供了一种创新、高效、实用的解决方案,有望在医学教育和临床技能提高中发挥越来越重要的作用。
{"title":"A Neurosurgical Craniotomy Training System Based on Haptic Virtual Reality Simulation","authors":"Guobin Zhang;Keliang Li;Qiyuan Sun;Wenqi Wu;Shuai Li;Zhenzhong Liu","doi":"10.1109/THMS.2025.3616313","DOIUrl":"https://doi.org/10.1109/THMS.2025.3616313","url":null,"abstract":"Traditional neurosurgical training modes face challenges including high costs, limited resources, lengthy learning curves, and difficulties in personalized training. In this article, we developed an immersive neurosurgical craniotomy virtual training system (NeuroSimulator) that integrates haptic feedback, enabling comprehensive surgical skill learning through an operator-control interface. Specifically, we constructed the comprehensive neurosurgical craniotomy surgical procedural (CNCSP) dataset to guide operators in repetitive learning and personalized training of relevant surgical skills. To address surgical site model rendering complexity, we proposed an algorithm that integrates vertex curvature and edge-length cost calculation factors (VC&ECL-QEM), resolving the incompatibility between surgical area image rendering quality and efficiency. For intracranial soft tissue haptic deformation, we developed a hybrid soft tissue haptic deformation (HBD) model that combines mass-spring and volumetric elements, addressing the collapse and distortion issues of traditional models and achieving more realistic soft tissue haptic deformation. Experimental results demonstrate that VC&ECL-QEM can simplify nonsurgical area feature preservation while maintaining surgical site detail features, reflecting the effectiveness of model simplification. The HBD model focuses on improving soft tissue deformation realism and shows high consistency with finite element model deformation effects. A total of 83 participants highly recognized NeuroSimulator’s system performance in terms of operational compliance, rendering real-time performance, and deformation realism, achieving effective improvements in skill proficiency metrics including operation time, ineffective operations, guidance requests, and operation scores. NeuroSimulator provides an innovative, efficient, and practical solution for neurosurgical training and is expected to play an increasingly important role in medical education and clinical skill enhancement.","PeriodicalId":48916,"journal":{"name":"IEEE Transactions on Human-Machine Systems","volume":"55 6","pages":"953-962"},"PeriodicalIF":4.4,"publicationDate":"2025-10-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145652176","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IEEE Systems, Man, and Cybernetics Society Information IEEE系统、人与控制论学会信息
IF 4.4 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-10-10 DOI: 10.1109/THMS.2025.3614336
{"title":"IEEE Systems, Man, and Cybernetics Society Information","authors":"","doi":"10.1109/THMS.2025.3614336","DOIUrl":"https://doi.org/10.1109/THMS.2025.3614336","url":null,"abstract":"","PeriodicalId":48916,"journal":{"name":"IEEE Transactions on Human-Machine Systems","volume":"55 5","pages":"C3-C3"},"PeriodicalIF":4.4,"publicationDate":"2025-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11200021","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145255898","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IEEE Transactions on Human-Machine Systems Information for Authors IEEE人机系统信息汇刊
IF 4.4 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-10-10 DOI: 10.1109/THMS.2025.3614338
{"title":"IEEE Transactions on Human-Machine Systems Information for Authors","authors":"","doi":"10.1109/THMS.2025.3614338","DOIUrl":"https://doi.org/10.1109/THMS.2025.3614338","url":null,"abstract":"","PeriodicalId":48916,"journal":{"name":"IEEE Transactions on Human-Machine Systems","volume":"55 5","pages":"C4-C4"},"PeriodicalIF":4.4,"publicationDate":"2025-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11200023","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145255896","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Understanding and Predicting Temporal Visual Attention Influenced by Dynamic Highlights in Monitoring Task 动态高光对监测任务时间视觉注意影响的理解与预测
IF 4.4 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-10-08 DOI: 10.1109/THMS.2025.3614364
Zekun Wu;Anna Maria Feit
Monitoring interfaces are crucial for dynamic, high-stakes tasks where effective user attention is essential. Visual highlights can guide attention effectively, but may also introduce unintended disruptions. To investigate this, we examined how visual highlights affect users’ gaze behavior in a drone monitoring task, focusing on when, how long, and how much attention they draw. We found that highlighted areas exhibit distinct temporal characteristics compared to nonhighlighted ones, quantified using normalized saliency (NS) metrics. We found that highlights elicited immediate responses, with NS peaking quickly, but this shift came at the cost of reduced search efforts elsewhere, potentially impacting situational awareness. To predict these dynamic changes and support interface design, we developed the Highlight-Informed Saliency Model, which provides granular predictions of NS over time. These predictions enable evaluations of highlight effectiveness and inform the optimal timing and deployment of highlights in future monitoring interface designs, particularly for time-sensitive tasks.
监控界面对于动态的、高风险的任务至关重要,在这些任务中,有效的用户注意力是必不可少的。视觉亮点可以有效地引导注意力,但也可能带来意想不到的干扰。为了研究这一点,我们研究了在无人机监控任务中,视觉亮点如何影响用户的凝视行为,重点关注它们何时、多长时间和多少注意力。我们发现,与未突出显示的区域相比,突出显示的区域表现出明显的时间特征,并使用标准化显著性(NS)指标进行量化。我们发现,突出显示会立即引起响应,NS很快达到峰值,但这种转变是以减少其他地方的搜索努力为代价的,可能会影响态势感知。为了预测这些动态变化并支持界面设计,我们开发了Highlight-Informed Saliency Model,该模型提供了NS随时间变化的细粒度预测。这些预测能够评估突出显示的有效性,并在未来的监控界面设计中告知突出显示的最佳时机和部署,特别是对于时间敏感的任务。
{"title":"Understanding and Predicting Temporal Visual Attention Influenced by Dynamic Highlights in Monitoring Task","authors":"Zekun Wu;Anna Maria Feit","doi":"10.1109/THMS.2025.3614364","DOIUrl":"https://doi.org/10.1109/THMS.2025.3614364","url":null,"abstract":"Monitoring interfaces are crucial for dynamic, high-stakes tasks where effective user attention is essential. Visual highlights can guide attention effectively, but may also introduce unintended disruptions. To investigate this, we examined how visual highlights affect users’ gaze behavior in a drone monitoring task, focusing on when, how long, and how much attention they draw. We found that highlighted areas exhibit distinct temporal characteristics compared to nonhighlighted ones, quantified using normalized saliency (NS) metrics. We found that highlights elicited immediate responses, with NS peaking quickly, but this shift came at the cost of reduced search efforts elsewhere, potentially impacting situational awareness. To predict these dynamic changes and support interface design, we developed the Highlight-Informed Saliency Model, which provides granular predictions of NS over time. These predictions enable evaluations of highlight effectiveness and inform the optimal timing and deployment of highlights in future monitoring interface designs, particularly for time-sensitive tasks.","PeriodicalId":48916,"journal":{"name":"IEEE Transactions on Human-Machine Systems","volume":"55 6","pages":"1053-1063"},"PeriodicalIF":4.4,"publicationDate":"2025-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145652163","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Learning Model With Fine-Tuning for Generalized Few-Shot Activity Recognition 广义少镜头活动识别的深度学习微调模型
IF 4.4 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-10-08 DOI: 10.1109/THMS.2025.3613773
Wei Wang;Qingzhong Li
The problem we focused on in this article is sensor-based generalized few-shot activity recognition. In this problem, each of the predefined activity classes (i.e., base classes) has substantial training instances, while each of the new activity classes (i.e., novel classes) just has a few training instances. Both the base and the novel classes need to be recognized. Currently, just a few works focus on this problem, and no formal statement of the problem is provided. In this article, we provide a formal definition of the problem, and propose a method to address it. In the proposed method, adopting the strategy of fine-tuning deep learning models, a deep learning model is first learned with the base-class training instances, and then fine-tuned with resampled training instances from both the base and the novel classes. We evaluate our method with three publicly available datasets on 1-shot, 5-shot, and 10-shot learning tasks. The results on the evaluation metric of harmonic mean of the average per-class accuracy for the base classes and that for the novel classes show that, our method could outperform state-of-the-art methods. In addition, the time and resource cost of our method is moderate.
本文主要研究基于传感器的广义少射活动识别问题。在这个问题中,每个预定义的活动类(即基类)都有大量的训练实例,而每个新的活动类(即新类)只有几个训练实例。基础类和新类都需要被识别。目前,针对这一问题的研究文献不多,并没有对这一问题进行正式的阐述。在本文中,我们提供了问题的正式定义,并提出了解决它的方法。该方法采用深度学习模型微调策略,首先使用基本类训练实例学习深度学习模型,然后使用从基本类和新类中重新采样的训练实例对深度学习模型进行微调。我们用三个公开可用的数据集来评估我们的方法,分别是1次、5次和10次学习任务。对基类和新类的平均每类准确率的调和平均值评价指标的结果表明,我们的方法优于目前最先进的方法。此外,我们的方法的时间和资源成本是适度的。
{"title":"Deep Learning Model With Fine-Tuning for Generalized Few-Shot Activity Recognition","authors":"Wei Wang;Qingzhong Li","doi":"10.1109/THMS.2025.3613773","DOIUrl":"https://doi.org/10.1109/THMS.2025.3613773","url":null,"abstract":"The problem we focused on in this article is sensor-based generalized few-shot activity recognition. In this problem, each of the predefined activity classes (i.e., base classes) has substantial training instances, while each of the new activity classes (i.e., novel classes) just has a few training instances. Both the base and the novel classes need to be recognized. Currently, just a few works focus on this problem, and no formal statement of the problem is provided. In this article, we provide a formal definition of the problem, and propose a method to address it. In the proposed method, adopting the strategy of fine-tuning deep learning models, a deep learning model is first learned with the base-class training instances, and then fine-tuned with resampled training instances from both the base and the novel classes. We evaluate our method with three publicly available datasets on 1-shot, 5-shot, and 10-shot learning tasks. The results on the evaluation metric of harmonic mean of the average per-class accuracy for the base classes and that for the novel classes show that, our method could outperform state-of-the-art methods. In addition, the time and resource cost of our method is moderate.","PeriodicalId":48916,"journal":{"name":"IEEE Transactions on Human-Machine Systems","volume":"55 6","pages":"940-952"},"PeriodicalIF":4.4,"publicationDate":"2025-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145652155","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ST-GCN-AltFormer: Gesture Recognition With Spatial-Temporal Alternating Transformer ST-GCN-AltFormer:基于时空交流变压器的手势识别
IF 4.4 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-10-01 DOI: 10.1109/THMS.2025.3607961
Qing Pan;Jintao Zhu;Lingwei Zhang;Gangmin Ning;Luping Fang
In skeleton-based gesture recognition tasks, existing approaches based on graph convolutional networks (GCNs) struggle to capture the synergistic actions of nonadjacent graph nodes and the information conveyed by their long-range dependencies. Combining spatial and temporal transformers is a promising solution to address the limitation, inspired by the advantage of transformer in assessing nonadjacent long-range dependencies, but there lacks an effective strategy to integrate the spatial and temporal information extracted by these transformers. Therefore, this article proposes the spatial-temporal alternating graph convolution transformer (ST-GCN-AltFormer), which connects the spatial-temporal graph convolutional network (ST-GCN) with the spatial-temporal alternating transformer (AltFormer) architecture. In the AltFormer architecture, the spatial-temporal transformer branch employs a spatial transformer to capture information from specific frames, and uses a temporal transformer to analyze its evolution over the entire temporal range. Meanwhile, the temporal-spatial transformer branch extracts temporal information from specific nodes using a temporal transformer, and integrates it with a spatial transformer. The fusion enhances accurate spatial-temporal information extraction. Our method achieves superior performance compared to state-of-the-art methods, achieving accuracies of 97.5%, 95.8%, 94.3%, 92.8%, and 98.31% on the large-scale 3D hand gesture recognition (SHREC’17 Track), Dynamic Hand Gesture 14-28 (DHG-14/28), and leap motion dynamic hand gesture (LMDHG) dynamic gesture datasets, respectively.
在基于骨架的手势识别任务中,现有的基于图卷积网络(GCNs)的方法难以捕捉非相邻图节点的协同作用以及它们的远程依赖关系所传递的信息。由于变压器在评估非相邻远程依赖关系方面的优势,结合时空变压器是一种很有前途的解决方案,但缺乏有效的策略来整合这些变压器提取的时空信息。因此,本文提出了时空交替图卷积变压器(ST-GCN-AltFormer),将时空图卷积网络(ST-GCN)与时空交替变压器(AltFormer)架构连接起来。在AltFormer架构中,时空变压器分支使用空间变压器从特定帧捕获信息,并使用时间变压器分析其在整个时间范围内的演变。同时,时空变压器分支利用时间变压器提取特定节点的时间信息,并将其与空间变压器集成。这种融合增强了准确的时空信息提取。该方法在大规模3D手势识别(SHREC ' 17 Track)、动态手势14-28 (DHG-14/28)和跳跃动作动态手势(LMDHG)数据集上的准确率分别达到了97.5%、95.8%、94.3%、92.8%和98.31%。
{"title":"ST-GCN-AltFormer: Gesture Recognition With Spatial-Temporal Alternating Transformer","authors":"Qing Pan;Jintao Zhu;Lingwei Zhang;Gangmin Ning;Luping Fang","doi":"10.1109/THMS.2025.3607961","DOIUrl":"https://doi.org/10.1109/THMS.2025.3607961","url":null,"abstract":"In skeleton-based gesture recognition tasks, existing approaches based on graph convolutional networks (GCNs) struggle to capture the synergistic actions of nonadjacent graph nodes and the information conveyed by their long-range dependencies. Combining spatial and temporal transformers is a promising solution to address the limitation, inspired by the advantage of transformer in assessing nonadjacent long-range dependencies, but there lacks an effective strategy to integrate the spatial and temporal information extracted by these transformers. Therefore, this article proposes the spatial-temporal alternating graph convolution transformer (ST-GCN-AltFormer), which connects the spatial-temporal graph convolutional network (ST-GCN) with the spatial-temporal alternating transformer (AltFormer) architecture. In the AltFormer architecture, the spatial-temporal transformer branch employs a spatial transformer to capture information from specific frames, and uses a temporal transformer to analyze its evolution over the entire temporal range. Meanwhile, the temporal-spatial transformer branch extracts temporal information from specific nodes using a temporal transformer, and integrates it with a spatial transformer. The fusion enhances accurate spatial-temporal information extraction. Our method achieves superior performance compared to state-of-the-art methods, achieving accuracies of 97.5%, 95.8%, 94.3%, 92.8%, and 98.31% on the large-scale 3D hand gesture recognition (SHREC’17 Track), Dynamic Hand Gesture 14-28 (DHG-14/28), and leap motion dynamic hand gesture (LMDHG) dynamic gesture datasets, respectively.","PeriodicalId":48916,"journal":{"name":"IEEE Transactions on Human-Machine Systems","volume":"55 6","pages":"963-972"},"PeriodicalIF":4.4,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145652188","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Palpation Characteristics of an Instrumented Virtual Cricothyroidotomy Simulator. 器械虚拟环甲软骨切开术模拟器的触诊特点。
IF 4.4 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-10-01 Epub Date: 2025-08-18 DOI: 10.1109/thms.2025.3592791
Melih Turkseven, Trudi Di Qi, Ganesh Sankaranarayanan, Suvranu De

Cricothyroidotomy (CCT) is a critical, life-saving procedure requiring the identification of key neck landmarks through palpation. Interactive virtual simulation offers a promising, cost-effective approach to CCT training with high visual realism. However, developing the palpation skills necessary for CCT requires a haptic interface with tactile sensitivity comparable to human fingers. Such interfaces are often represented by plastic partial mannequins, which require further adaptation to integrate into virtual environments. This study introduces an instrumented physical palpation interface for CCT, integrated into a virtual surgical simulator, and tested on ten surgeons who practiced the procedure over a training period. Data on haptic interactions collected during the training was analyzed to evaluate participants' palpation skills and explore their force modulation strategies about landmark identification scores. Our findings suggest that trainees become more precise in their exploration over time, apply greater normal forces around target areas. Initial landmark identification performance influences adjustments in the overall applied pressure.

环甲状腺切开术(CCT)是一个关键的,挽救生命的程序,需要通过触诊识别关键的颈部标志。交互式虚拟仿真为CCT训练提供了一种具有较高视觉真实感的有前途的、经济有效的方法。然而,发展CCT所需的触诊技能需要具有与人类手指相当的触觉灵敏度的触觉界面。这种界面通常由塑料部分人体模型表示,需要进一步适应才能融入虚拟环境。本研究介绍了一种用于CCT的仪器物理触诊界面,集成到虚拟手术模拟器中,并对10名在培训期间实践该手术的外科医生进行了测试。对训练过程中收集的触觉交互数据进行分析,以评估参与者的触诊技能,并探讨他们对地标识别分数的力调制策略。我们的研究结果表明,随着时间的推移,受训者在探索时变得更加精确,在目标区域周围施加更大的法向力。最初的地标识别性能会影响总体施加压力的调整。
{"title":"Palpation Characteristics of an Instrumented Virtual Cricothyroidotomy Simulator.","authors":"Melih Turkseven, Trudi Di Qi, Ganesh Sankaranarayanan, Suvranu De","doi":"10.1109/thms.2025.3592791","DOIUrl":"10.1109/thms.2025.3592791","url":null,"abstract":"<p><p>Cricothyroidotomy (CCT) is a critical, life-saving procedure requiring the identification of key neck landmarks through palpation. Interactive virtual simulation offers a promising, cost-effective approach to CCT training with high visual realism. However, developing the palpation skills necessary for CCT requires a haptic interface with tactile sensitivity comparable to human fingers. Such interfaces are often represented by plastic partial mannequins, which require further adaptation to integrate into virtual environments. This study introduces an instrumented physical palpation interface for CCT, integrated into a virtual surgical simulator, and tested on ten surgeons who practiced the procedure over a training period. Data on haptic interactions collected during the training was analyzed to evaluate participants' palpation skills and explore their force modulation strategies about landmark identification scores. Our findings suggest that trainees become more precise in their exploration over time, apply greater normal forces around target areas. Initial landmark identification performance influences adjustments in the overall applied pressure.</p>","PeriodicalId":48916,"journal":{"name":"IEEE Transactions on Human-Machine Systems","volume":"55 5","pages":"876-885"},"PeriodicalIF":4.4,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12617396/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145543563","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Bimanual Manipulation of Steady-Hand Eye Robots With Adaptive Sclera Force Control: Cooperative Versus Teleoperation Strategies 具有自适应巩膜力控制的双手操作稳定手眼机器人:合作与远程操作策略
IF 4.4 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-09-18 DOI: 10.1109/THMS.2025.3605011
Mojtaba Esfandiari;Peter Gehlbach;Russell H. Taylor;Iulian I. Iordachita
Performing retinal vein cannulation (RVC) as a potential treatment for retinal vein occlusion without the assistance of a surgical robotic system is very challenging to do safely. The main limitation is the physiological hand tremor of surgeons. Robot-assisted eye surgery technology may resolve the problems of hand tremors and fatigue and improve the safety and precision of RVC. The steady-hand eye robot (SHER) is an admittance-based robotic system that can filter out hand tremors and enables ophthalmologists to manipulate a surgical instrument inside the eye cooperatively. However, the admittance-based cooperative control mode does not safely minimize the contact force between the surgical instrument and the sclera to prevent tissue damage. In addition, features such as haptic feedback or hand motion scaling, which can improve the safety and precision of surgery, require a teleoperation control framework. This work presents, for the first time in the field of robot-assisted retinal microsurgery research, a registration-free bimanual adaptive teleoperation (BMAT) control framework using SHER 2.0 and SHER 2.1 robotic systems. Both SHERs are integrated with an adaptive force control algorithm that dynamically and automatically minimizes the tool–sclera interaction forces, enforcing them within a safe limit. The scleral forces are measured using two fiber Bragg grating-based force-sensing tools. The performance of the proposed BMAT control framework is evaluated by comparison with a bimanual adaptive cooperative framework in a vessel-following experiment conducted under a surgical microscope. Experimental results demonstrate the effectiveness of the BMAT control framework in performing a safe bimanual telemanipulation of the eye without overstretching it, even in the absence of registration between the two robots.
在没有外科机器人系统的帮助下,将视网膜静脉插管作为视网膜静脉闭塞的潜在治疗方法是非常具有挑战性的。主要的限制是外科医生的生理性手颤。机器人辅助眼科手术技术可以解决手部震颤和疲劳问题,提高RVC的安全性和精度。稳定手眼机器人(SHER)是一种基于导纳的机器人系统,它可以过滤手部震颤,使眼科医生能够协同操作眼内手术器械。然而,基于导纳的协同控制模式并不能安全地将手术器械与巩膜之间的接触力最小化,从而防止组织损伤。此外,可以提高手术安全性和精度的触觉反馈或手部运动缩放等功能需要远程操作控制框架。本工作首次在机器人辅助视网膜显微外科研究领域提出了一个使用SHER 2.0和SHER 2.1机器人系统的无配准手动自适应遥操作(BMAT)控制框架。这两种SHERs都集成了一种自适应力控制算法,该算法可以动态、自动地最小化工具-巩膜相互作用力,将其强制控制在安全范围内。巩膜力测量使用两个光纤布拉格光栅为基础的力传感工具。在外科显微镜下进行的血管跟踪实验中,通过与双手自适应合作框架的比较,评估了所提出的BMAT控制框架的性能。实验结果证明了BMAT控制框架的有效性,即使在两个机器人之间没有配准的情况下,也可以在不过度拉伸的情况下进行安全的双手遥控操作。
{"title":"Bimanual Manipulation of Steady-Hand Eye Robots With Adaptive Sclera Force Control: Cooperative Versus Teleoperation Strategies","authors":"Mojtaba Esfandiari;Peter Gehlbach;Russell H. Taylor;Iulian I. Iordachita","doi":"10.1109/THMS.2025.3605011","DOIUrl":"https://doi.org/10.1109/THMS.2025.3605011","url":null,"abstract":"Performing retinal vein cannulation (RVC) as a potential treatment for retinal vein occlusion without the assistance of a surgical robotic system is very challenging to do safely. The main limitation is the physiological hand tremor of surgeons. Robot-assisted eye surgery technology may resolve the problems of hand tremors and fatigue and improve the safety and precision of RVC. The steady-hand eye robot (SHER) is an admittance-based robotic system that can filter out hand tremors and enables ophthalmologists to manipulate a surgical instrument inside the eye cooperatively. However, the admittance-based cooperative control mode does not safely minimize the contact force between the surgical instrument and the sclera to prevent tissue damage. In addition, features such as haptic feedback or hand motion scaling, which can improve the safety and precision of surgery, require a teleoperation control framework. This work presents, for the first time in the field of robot-assisted retinal microsurgery research, a registration-free bimanual adaptive teleoperation (BMAT) control framework using SHER 2.0 and SHER 2.1 robotic systems. Both SHERs are integrated with an adaptive force control algorithm that dynamically and automatically minimizes the tool–sclera interaction forces, enforcing them within a safe limit. The scleral forces are measured using two fiber Bragg grating-based force-sensing tools. The performance of the proposed BMAT control framework is evaluated by comparison with a bimanual adaptive cooperative framework in a vessel-following experiment conducted under a surgical microscope. Experimental results demonstrate the effectiveness of the BMAT control framework in performing a safe bimanual telemanipulation of the eye without overstretching it, even in the absence of registration between the two robots.","PeriodicalId":48916,"journal":{"name":"IEEE Transactions on Human-Machine Systems","volume":"55 6","pages":"909-919"},"PeriodicalIF":4.4,"publicationDate":"2025-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145652177","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
EEG Neurofeedback-Based Gait Motor Imagery Training in Lokomat Enhances Motor Rhythms in Complete Spinal Cord Injury 基于脑电神经反馈的步态运动想象训练增强完全性脊髓损伤的运动节律
IF 4.4 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-09-12 DOI: 10.1109/THMS.2025.3603548
Ericka R. da Silva Serafini;Cristian D. Guerrero-Mendez;Douglas M. Dunga;Teodiano F. Bastos-Filho;Anibal Cotrina Atencio;André F. O. de Azevedo Dantas;Caroline C. do Espírito Santo;Denis Delisle-Rodriguez
Robotic interventions combining neurofeedback (NFB) and motor imagery (MI) are emerging strategies to promote cortical reorganization and functional training in individuals with complete spinal cord injury (SCI). This study proposes an electroencephalogram-based NFB approach for MI training, designed to teach the MI-related brain rhythmics modulation in Lokomat. For the purposes of this study, NFB is defined as a visual feedback training scheme. The proposed system introduces a formulation to minimize the default cortical effects that Lokomat produces on the individual’s activity during passive walking. Two individuals with complete SCI tested the proposed NFB system, in order to relearn the modulation of Mu ($mu$ : 8–12 Hz) and Beta ($beta$ : 13–30 Hz) rhythms over Cz, while receiving gait training with full weight support across 12 sessions. Each session consisted of the following three stages: 1) 2 min walking without MI (baseline); 2) 5 min walking with MI and True NFB; and 3) 5 min walking with MI and Sham NFB. The latter two stages were randomized session-by-session. The findings suggest that the proposed NFB approach may promote cortical reorganization and support the restoration of sensorimotor functions. Significant differences were observed between cortical patterns during True NFB and Sham NFB, particularly in the last intervention sessions. These results confirm the positive impact of the NFB system on gait motor training by enabling individuals with complete SCI to learn how to modulate their motor rhythms in specific cortical areas.
结合神经反馈(NFB)和运动成像(MI)的机器人干预是促进完全性脊髓损伤(SCI)患者皮质重组和功能训练的新兴策略。本研究提出了一种基于脑电图的NFB方法用于心梗训练,旨在教授Lokomat与心梗相关的脑节律调节。在本研究中,NFB被定义为视觉反馈训练方案。所提出的系统引入了一种配方,以尽量减少Lokomat在被动行走期间对个人活动产生的默认皮质影响。两名完全性脊髓损伤患者测试了拟议的NFB系统,以重新学习Mu ($mu$: 8-12 Hz)和Beta ($beta$: 13-30 Hz)节奏在Cz上的调制,同时接受12次全重量支持的步态训练。每个疗程包括以下三个阶段:1)2分钟无心肌梗死步行(基线);2)步行5分钟,伴心肌梗死和True NFB;3) MI和Sham NFB组步行5分钟。后两个阶段是随机分组的。研究结果表明,NFB方法可能促进皮层重组并支持感觉运动功能的恢复。在真NFB和假NFB期间,特别是在最后的干预阶段,观察到皮层模式的显著差异。这些结果证实了NFB系统对步态运动训练的积极影响,使完全性脊髓损伤患者能够学习如何调节特定皮质区域的运动节律。
{"title":"EEG Neurofeedback-Based Gait Motor Imagery Training in Lokomat Enhances Motor Rhythms in Complete Spinal Cord Injury","authors":"Ericka R. da Silva Serafini;Cristian D. Guerrero-Mendez;Douglas M. Dunga;Teodiano F. Bastos-Filho;Anibal Cotrina Atencio;André F. O. de Azevedo Dantas;Caroline C. do Espírito Santo;Denis Delisle-Rodriguez","doi":"10.1109/THMS.2025.3603548","DOIUrl":"https://doi.org/10.1109/THMS.2025.3603548","url":null,"abstract":"Robotic interventions combining neurofeedback (NFB) and motor imagery (MI) are emerging strategies to promote cortical reorganization and functional training in individuals with complete spinal cord injury (SCI). This study proposes an electroencephalogram-based NFB approach for MI training, designed to teach the MI-related brain rhythmics modulation in Lokomat. For the purposes of this study, NFB is defined as a visual feedback training scheme. The proposed system introduces a formulation to minimize the default cortical effects that Lokomat produces on the individual’s activity during passive walking. Two individuals with complete SCI tested the proposed NFB system, in order to relearn the modulation of Mu (<inline-formula><tex-math>$mu$</tex-math></inline-formula> : 8–12 Hz) and Beta (<inline-formula><tex-math>$beta$</tex-math></inline-formula> : 13–30 Hz) rhythms over Cz, while receiving gait training with full weight support across 12 sessions. Each session consisted of the following three stages: 1) 2 min walking without MI (baseline); 2) 5 min walking with MI and True NFB; and 3) 5 min walking with MI and Sham NFB. The latter two stages were randomized session-by-session. The findings suggest that the proposed NFB approach may promote cortical reorganization and support the restoration of sensorimotor functions. Significant differences were observed between cortical patterns during True NFB and Sham NFB, particularly in the last intervention sessions. These results confirm the positive impact of the NFB system on gait motor training by enabling individuals with complete SCI to learn how to modulate their motor rhythms in specific cortical areas.","PeriodicalId":48916,"journal":{"name":"IEEE Transactions on Human-Machine Systems","volume":"55 6","pages":"983-992"},"PeriodicalIF":4.4,"publicationDate":"2025-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11163517","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145652152","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Transactions on Human-Machine Systems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1