首页 > 最新文献

IEEE Transactions on Human-Machine Systems最新文献

英文 中文
Call for Papers: IEEE Transactions on Human-Machine Systems 论文征集:IEEE人机系统汇刊
IF 4.4 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-26 DOI: 10.1109/THMS.2026.3656624
{"title":"Call for Papers: IEEE Transactions on Human-Machine Systems","authors":"","doi":"10.1109/THMS.2026.3656624","DOIUrl":"https://doi.org/10.1109/THMS.2026.3656624","url":null,"abstract":"","PeriodicalId":48916,"journal":{"name":"IEEE Transactions on Human-Machine Systems","volume":"56 1","pages":"192-192"},"PeriodicalIF":4.4,"publicationDate":"2026-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11364046","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146045349","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IEEE Systems, Man, and Cybernetics Society Information IEEE系统、人与控制论学会信息
IF 4.4 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-26 DOI: 10.1109/THMS.2026.3651175
{"title":"IEEE Systems, Man, and Cybernetics Society Information","authors":"","doi":"10.1109/THMS.2026.3651175","DOIUrl":"https://doi.org/10.1109/THMS.2026.3651175","url":null,"abstract":"","PeriodicalId":48916,"journal":{"name":"IEEE Transactions on Human-Machine Systems","volume":"56 1","pages":"C3-C3"},"PeriodicalIF":4.4,"publicationDate":"2026-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11364037","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146045308","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IEEE Transactions on Human-Machine Systems Information for Authors IEEE人机系统信息汇刊
IF 4.4 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-26 DOI: 10.1109/THMS.2026.3651173
{"title":"IEEE Transactions on Human-Machine Systems Information for Authors","authors":"","doi":"10.1109/THMS.2026.3651173","DOIUrl":"https://doi.org/10.1109/THMS.2026.3651173","url":null,"abstract":"","PeriodicalId":48916,"journal":{"name":"IEEE Transactions on Human-Machine Systems","volume":"56 1","pages":"C4-C4"},"PeriodicalIF":4.4,"publicationDate":"2026-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11364039","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146045352","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IEEE Systems, Man, and Cybernetics Society Information IEEE系统、人与控制论学会信息
IF 4.4 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-26 DOI: 10.1109/THMS.2026.3651171
{"title":"IEEE Systems, Man, and Cybernetics Society Information","authors":"","doi":"10.1109/THMS.2026.3651171","DOIUrl":"https://doi.org/10.1109/THMS.2026.3651171","url":null,"abstract":"","PeriodicalId":48916,"journal":{"name":"IEEE Transactions on Human-Machine Systems","volume":"56 1","pages":"C2-C2"},"PeriodicalIF":4.4,"publicationDate":"2026-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11364045","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146045332","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Effects of Readiness Deprivation on Takeover With Varying Time Budget in Conditional Automated Driving Scenarios 条件自动驾驶下准备剥夺对变时间预算接管的影响
IF 4.4 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-05 DOI: 10.1109/THMS.2025.3595269
Hsueh-Yi Lai;Ching-Hao Chou;Pen-Kuei Wang;Tse-Yi Kuo;Yong-Jhih Chen
As vehicle automation levels rise, future automated driving systems can enable drivers to engage in nondriving-related tasks (NDRTs). However, drivers remain responsible for driving safety. In the case of an automated vehicle failure, a driver must disengage from their NDRTs and take control of the vehicle. When fully engaged in NDRTs during Level 3 automation, NDRTs can compete for the resources for driving tasks, thereby depriving cognitive and physical readiness. This study investigated how compromised driver readiness affected takeover performance in situations with varying urgency levels. A simulated driving experiment was conducted with 32 participants in four states of readiness deprivation created through NDRT assignment, and two time budget levels were applied to represent multiple urgency scenarios. First, subjective ratings on readiness deprivation showed that depriving drivers of one form of readiness (i.e., cognitive or physical) adversely affected the other. Furthermore, retaining cognitive readiness may provide greater self-assessed utility. The NDRTs with similar interaction attributes generate comparable readiness deprivation ratings, offering a systematic way to evaluate their impact on takeover. Then, the impact of readiness deprivation on takeover performance varied significantly based on time budgets. With ample time, depriving drivers of physical or full readiness increased takeover time. However, these delayed actions, combined with stable lateral control, suggested a safe takeover strategy aimed at readiness recovery. Conversely, limited time hindered this recovery. Drivers performed takeovers despite impaired readiness, resulting in quicker but often abrupt post-takeover lateral movements. Notably, takeover actions were initiated once both cognitive and physical readiness were achieved, regardless of the time budget.
随着车辆自动化水平的提高,未来的自动驾驶系统可以让驾驶员从事与驾驶无关的任务(NDRTs)。然而,司机仍然对驾驶安全负责。在自动车辆发生故障的情况下,驾驶员必须脱离NDRTs并控制车辆。当在3级自动化过程中完全投入NDRTs时,NDRTs可能会争夺驾驶任务的资源,从而剥夺认知和身体准备。本研究调查了在不同紧急程度的情况下,受损的驾驶员准备如何影响接管绩效。在模拟驾驶实验中,32名被试通过NDRT分配产生4种准备剥夺状态,并采用2个时间预算水平来表示多种紧急情况。首先,对准备剥夺的主观评价表明,剥夺一种形式的准备(即认知或身体)会对另一种形式的准备产生不利影响。此外,保持认知准备可能提供更大的自我评估效用。具有相似交互属性的ndrt产生了类似的准备剥夺评级,提供了一种系统的方法来评估它们对接管的影响。在不同的时间预算下,准备剥夺对接管绩效的影响存在显著差异。由于时间充裕,剥夺司机的物理或完全准备增加接管时间。然而,这些延迟的行动,加上稳定的横向控制,表明了一种旨在恢复准备状态的安全收购策略。相反,有限的时间阻碍了这种恢复。尽管准备程度受损,但司机还是进行了收购,导致收购后的横向变动更快,但往往是突然的。值得注意的是,不管时间预算如何,一旦认知和身体准备就绪,接管行动就会开始。
{"title":"Effects of Readiness Deprivation on Takeover With Varying Time Budget in Conditional Automated Driving Scenarios","authors":"Hsueh-Yi Lai;Ching-Hao Chou;Pen-Kuei Wang;Tse-Yi Kuo;Yong-Jhih Chen","doi":"10.1109/THMS.2025.3595269","DOIUrl":"https://doi.org/10.1109/THMS.2025.3595269","url":null,"abstract":"As vehicle automation levels rise, future automated driving systems can enable drivers to engage in nondriving-related tasks (NDRTs). However, drivers remain responsible for driving safety. In the case of an automated vehicle failure, a driver must disengage from their NDRTs and take control of the vehicle. When fully engaged in NDRTs during Level 3 automation, NDRTs can compete for the resources for driving tasks, thereby depriving cognitive and physical readiness. This study investigated how compromised driver readiness affected takeover performance in situations with varying urgency levels. A simulated driving experiment was conducted with 32 participants in four states of readiness deprivation created through NDRT assignment, and two time budget levels were applied to represent multiple urgency scenarios. First, subjective ratings on readiness deprivation showed that depriving drivers of one form of readiness (i.e., cognitive or physical) adversely affected the other. Furthermore, retaining cognitive readiness may provide greater self-assessed utility. The NDRTs with similar interaction attributes generate comparable readiness deprivation ratings, offering a systematic way to evaluate their impact on takeover. Then, the impact of readiness deprivation on takeover performance varied significantly based on time budgets. With ample time, depriving drivers of physical or full readiness increased takeover time. However, these delayed actions, combined with stable lateral control, suggested a safe takeover strategy aimed at readiness recovery. Conversely, limited time hindered this recovery. Drivers performed takeovers despite impaired readiness, resulting in quicker but often abrupt post-takeover lateral movements. Notably, takeover actions were initiated once both cognitive and physical readiness were achieved, regardless of the time budget.","PeriodicalId":48916,"journal":{"name":"IEEE Transactions on Human-Machine Systems","volume":"56 1","pages":"171-181"},"PeriodicalIF":4.4,"publicationDate":"2026-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146045340","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
P3-FSLNet: A Compact Spatio-Temporal Model With Contrastive Few-Shot Learning for Subject-Independent P300 Detection in Devanagari Script-Based P300 Speller P3-FSLNet:一种基于对比少镜头学习的紧凑时空模型,用于基于Devanagari脚本的P300拼写检测
IF 4.4 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-01 DOI: 10.1109/THMS.2025.3639248
Vibha Bhandari;Narendra D. Londhe;Ghanahshyam B. Kshirsagar
Brain–computer interface (BCI) systems frequently necessitate time-intensive subject-specific calibration, thereby motivating the development of subject-independent P300 detection approaches. Existing methodologies that employ transfer learning and knowledge distillation encounter challenges with limited generalizability due to substantial intersubject variability and constrained data availability. Furthermore, their applicability is often questioned, as most lack comprehensive external validation and cross-script evaluation. To address these limitations, we introduce P3-FSLNet, a few-shot metalearning framework that integrates prototypical networks and contrastive learning within a spatial-temporal convolutional neural network augmented by dual-channel attention for the selection of relevant electroencephalogram channels. Episodic metatraining facilitates the transfer of robust knowledge across different subjects. Evaluated on a self-recorded Devanagari script dataset, P3-FSLNet attains a classification accuracy of 93.17%, surpassing state-of-the-art methods by 1%–14%, while simultaneously reducing trainable parameters by up to 400 times. External validation using English-script datasets from BCI Competition II and III confirms its robustness in cross-subject and cross-script generalization. These findings demonstrate the efficacy of P3-FSLNet and represent a substantial advancement toward the development of script-agnostic P300 spellers that are lightweight, scalable, and conducive to multilingual applications.
脑机接口(BCI)系统经常需要耗时的受试者特定校准,从而推动了与受试者无关的P300检测方法的发展。由于存在大量的学科间可变性和有限的数据可用性,采用迁移学习和知识蒸馏的现有方法面临着泛化能力有限的挑战。此外,它们的适用性经常受到质疑,因为大多数缺乏全面的外部验证和跨脚本评估。为了解决这些限制,我们引入了P3-FSLNet,这是一个少量元学习框架,它在时空卷积神经网络中集成了原型网络和对比学习,并通过双通道注意力来增强相关脑电图通道的选择。情景元训练有助于在不同学科之间转移稳健的知识。在一个自记录的Devanagari脚本数据集上进行评估,P3-FSLNet的分类准确率达到93.17%,比目前最先进的方法高出1%-14%,同时将可训练参数减少了400倍。使用来自BCI竞赛II和III的英语脚本数据集进行外部验证,证实了其在跨主题和跨脚本泛化方面的稳健性。这些发现证明了P3-FSLNet的有效性,并代表了在开发与脚本无关的P300拼写器方面取得的实质性进展,这些拼写器轻量级、可扩展且有利于多语言应用程序。
{"title":"P3-FSLNet: A Compact Spatio-Temporal Model With Contrastive Few-Shot Learning for Subject-Independent P300 Detection in Devanagari Script-Based P300 Speller","authors":"Vibha Bhandari;Narendra D. Londhe;Ghanahshyam B. Kshirsagar","doi":"10.1109/THMS.2025.3639248","DOIUrl":"https://doi.org/10.1109/THMS.2025.3639248","url":null,"abstract":"Brain–computer interface (BCI) systems frequently necessitate time-intensive subject-specific calibration, thereby motivating the development of subject-independent P300 detection approaches. Existing methodologies that employ transfer learning and knowledge distillation encounter challenges with limited generalizability due to substantial intersubject variability and constrained data availability. Furthermore, their applicability is often questioned, as most lack comprehensive external validation and cross-script evaluation. To address these limitations, we introduce P3-FSLNet, a few-shot metalearning framework that integrates prototypical networks and contrastive learning within a spatial-temporal convolutional neural network augmented by dual-channel attention for the selection of relevant electroencephalogram channels. Episodic metatraining facilitates the transfer of robust knowledge across different subjects. Evaluated on a self-recorded Devanagari script dataset, P3-FSLNet attains a classification accuracy of 93.17%, surpassing state-of-the-art methods by 1%–14%, while simultaneously reducing trainable parameters by up to 400 times. External validation using English-script datasets from BCI Competition II and III confirms its robustness in cross-subject and cross-script generalization. These findings demonstrate the efficacy of P3-FSLNet and represent a substantial advancement toward the development of script-agnostic P300 spellers that are lightweight, scalable, and conducive to multilingual applications.","PeriodicalId":48916,"journal":{"name":"IEEE Transactions on Human-Machine Systems","volume":"56 1","pages":"124-134"},"PeriodicalIF":4.4,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146045342","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An HCAI Methodological Framework: Putting it Into Action to Enable Human-Centered AI HCAI方法框架:将其付诸行动以实现以人为本的AI
IF 4.4 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-01 DOI: 10.1109/THMS.2025.3631590
Wei Xu;Zaifeng Gao;Marvin J. Dainoff
Human-centered artificial intelligence (HCAI) is a design philosophy that prioritizes humans in the design, development, deployment, and use of AI systems, aiming to maximize AI’s benefits while mitigating its negative impacts. Despite its growing prominence in literature, the lack of methodological guidance for its implementation poses challenges to HCAI practice. To address this gap, this article proposes a comprehensive HCAI methodological framework (HCAI-MF) comprising five key components: HCAI requirement hierarchy, approach and method taxonomy, process, interdisciplinary collaboration approach, and multilevel design paradigms. A case study demonstrates HCAI-MF’s practical implications, while the article also analyzes implementation challenges. Actionable recommendations and a “three-layer” HCAI implementation strategy are provided to address these challenges and guide future evolution of HCAI-MF. HCAI-MF is presented as a systematic and executable methodology capable of overcoming current gaps, enabling effective design, development, deployment, and use of AI systems, and advancing HCAI practice.
以人为本的人工智能(HCAI)是一种设计理念,它在设计、开发、部署和使用人工智能系统时优先考虑人类,旨在最大限度地提高人工智能的效益,同时减轻其负面影响。尽管它在文献中日益突出,但缺乏对其实施的方法指导,对HCAI实践提出了挑战。为了解决这一差距,本文提出了一个全面的HCAI方法框架(HCAI- mf),包括五个关键组成部分:HCAI需求层次、方法和方法分类法、过程、跨学科协作方法和多层次设计范式。一个案例研究展示了HCAI-MF的实际意义,同时本文还分析了实施挑战。提出了可操作的建议和“三层”HCAI实施战略,以应对这些挑战并指导HCAI- mf的未来发展。HCAI- mf是一种系统的、可执行的方法,能够克服当前的差距,实现人工智能系统的有效设计、开发、部署和使用,并推进HCAI实践。
{"title":"An HCAI Methodological Framework: Putting it Into Action to Enable Human-Centered AI","authors":"Wei Xu;Zaifeng Gao;Marvin J. Dainoff","doi":"10.1109/THMS.2025.3631590","DOIUrl":"https://doi.org/10.1109/THMS.2025.3631590","url":null,"abstract":"Human-centered artificial intelligence (HCAI) is a design philosophy that prioritizes humans in the design, development, deployment, and use of AI systems, aiming to maximize AI’s benefits while mitigating its negative impacts. Despite its growing prominence in literature, the lack of methodological guidance for its implementation poses challenges to HCAI practice. To address this gap, this article proposes a comprehensive HCAI methodological framework (HCAI-MF) comprising five key components: HCAI requirement hierarchy, approach and method taxonomy, process, interdisciplinary collaboration approach, and multilevel design paradigms. A case study demonstrates HCAI-MF’s practical implications, while the article also analyzes implementation challenges. Actionable recommendations and a “three-layer” HCAI implementation strategy are provided to address these challenges and guide future evolution of HCAI-MF. HCAI-MF is presented as a systematic and executable methodology capable of overcoming current gaps, enabling effective design, development, deployment, and use of AI systems, and advancing HCAI practice.","PeriodicalId":48916,"journal":{"name":"IEEE Transactions on Human-Machine Systems","volume":"56 1","pages":"78-94"},"PeriodicalIF":4.4,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146045348","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Estimating Workload for Supervisory Human–Robot Teams: An Initial Analysis of Meta-Learning 管理人机团队的工作量估算:元学习的初步分析
IF 4.4 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-31 DOI: 10.1109/THMS.2025.3640359
Joshua Bhagat Smith;Julie A. Adams
A robust understanding of a human's internal state can greatly improve human–robot teaming, as estimating the human teammates' workload can inform more dynamic robot adaptations. Existing workload estimation methods use standard machine learning techniques to model the relationships between physiological metrics and workload. However, such methods are not sufficient for adaptive systems, as standard machine learning techniques struggle to make accurate workload estimates when the human–robot team performs unknown tasks. A meta-learning-based workload estimation algorithm is introduced and an initial analysis is conducted to show how adapting a machine learning model's parameters using task-specific information can improve result in more accurate workload estimates for unknown tasks.
对人类内部状态的强大理解可以极大地改善人机合作,因为估计人类队友的工作量可以为更动态的机器人适应提供信息。现有的工作量估计方法使用标准的机器学习技术来模拟生理指标和工作量之间的关系。然而,这种方法对于自适应系统来说是不够的,因为当人机团队执行未知任务时,标准的机器学习技术很难做出准确的工作量估计。介绍了一种基于元学习的工作负载估计算法,并进行了初步分析,以展示如何使用特定于任务的信息调整机器学习模型的参数,从而提高对未知任务的更准确的工作负载估计。
{"title":"Estimating Workload for Supervisory Human–Robot Teams: An Initial Analysis of Meta-Learning","authors":"Joshua Bhagat Smith;Julie A. Adams","doi":"10.1109/THMS.2025.3640359","DOIUrl":"https://doi.org/10.1109/THMS.2025.3640359","url":null,"abstract":"A robust understanding of a human's internal state can greatly improve human–robot teaming, as estimating the human teammates' workload can inform more dynamic robot adaptations. Existing workload estimation methods use standard machine learning techniques to model the relationships between physiological metrics and workload. However, such methods are not sufficient for adaptive systems, as standard machine learning techniques struggle to make accurate workload estimates when the human–robot team performs unknown tasks. A meta-learning-based workload estimation algorithm is introduced and an initial analysis is conducted to show how adapting a machine learning model's parameters using task-specific information can improve result in more accurate workload estimates for unknown tasks.","PeriodicalId":48916,"journal":{"name":"IEEE Transactions on Human-Machine Systems","volume":"56 1","pages":"68-77"},"PeriodicalIF":4.4,"publicationDate":"2025-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146045350","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Nonauditory Schemes for Universal Information Access: Translating Braille into Vibrotactile Cues for the Blind 通用信息获取的非听觉方案:将盲文翻译成盲人的振动触觉提示
IF 4.4 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-22 DOI: 10.1109/THMS.2025.3631785
Zihan Tang;Aiguo Song
Despite the progress in human–computer interaction technology, the interaction methods that visually impaired individuals possess are still rudimentary. The widely used Text-to-Speech technology has issues such as privacy leaks in practical applications, and most of the new interaction designs proposed in recent years have limited application scenarios. Geared toward enriching interaction methods for users with visual impairments, this article explores the potential for translating Braille into vibrotactile cues as a way of conveying universal information. We designed a set of schemes and implemented them based on the vibration motor of a mobile phone. These schemes convert a single Braille character into several highly distinguishable vibration combinations, thereby conveying any information that Braille can express. Experiments were conducted on both sighted and visually impaired participants to evaluate the accuracy and efficiency. With a brief learning period of just 10 minutes, individuals can attain an accuracy rate greater than 95%, and the accuracy degradation remains minimal when playback speeds increase. By employing vibration motors to deliver comprehensive information, this framework shows promise for application in a wider range of technological devices.
尽管人机交互技术取得了进步,但视障人士的交互方式仍处于初级阶段。广泛使用的文本转语音技术在实际应用中存在隐私泄露等问题,近年来提出的大多数新型交互设计的应用场景有限。为了丰富视觉障碍用户的交互方法,本文探讨了将盲文翻译成振动触觉提示作为传递通用信息的一种方式的潜力。我们设计了一套基于手机振动电机的方案并进行了实现。这些方案将单个盲文字符转换为几个高度可区分的振动组合,从而传递盲文可以表达的任何信息。实验在视力正常和视力受损的参与者中进行,以评估准确性和效率。在短短10分钟的学习时间内,个体的准确率可以达到95%以上,而且当回放速度增加时,准确率的下降仍然最小。通过使用振动电机来传递全面的信息,该框架显示出在更广泛的技术设备中应用的前景。
{"title":"Nonauditory Schemes for Universal Information Access: Translating Braille into Vibrotactile Cues for the Blind","authors":"Zihan Tang;Aiguo Song","doi":"10.1109/THMS.2025.3631785","DOIUrl":"https://doi.org/10.1109/THMS.2025.3631785","url":null,"abstract":"Despite the progress in human–computer interaction technology, the interaction methods that visually impaired individuals possess are still rudimentary. The widely used Text-to-Speech technology has issues such as privacy leaks in practical applications, and most of the new interaction designs proposed in recent years have limited application scenarios. Geared toward enriching interaction methods for users with visual impairments, this article explores the potential for translating Braille into vibrotactile cues as a way of conveying universal information. We designed a set of schemes and implemented them based on the vibration motor of a mobile phone. These schemes convert a single Braille character into several highly distinguishable vibration combinations, thereby conveying any information that Braille can express. Experiments were conducted on both sighted and visually impaired participants to evaluate the accuracy and efficiency. With a brief learning period of just 10 minutes, individuals can attain an accuracy rate greater than 95%, and the accuracy degradation remains minimal when playback speeds increase. By employing vibration motors to deliver comprehensive information, this framework shows promise for application in a wider range of technological devices.","PeriodicalId":48916,"journal":{"name":"IEEE Transactions on Human-Machine Systems","volume":"56 1","pages":"12-21"},"PeriodicalIF":4.4,"publicationDate":"2025-12-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146045334","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Speech2Blend: A Hybrid Network for Speech-Driven 3-D Facial Animation by Learning Blendshape 通过学习Blendshape实现语音驱动的3d面部动画的混合网络
IF 4.4 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-22 DOI: 10.1109/THMS.2025.3640719
Lei Wang;Gongbin Chen;Feng Liu;Jiaji Wu;Jun Cheng
Recent advances in speech-driven facial animation have attracted significant interest across computer graphics, human–computer interaction systems, and immersive virtual reality applications. However, existing methods remain constrained by dependencies on specific reference videos or proprietary face mesh structures, limiting their applicability across diverse production pipelines and reducing compatibility with industry-standard animation workflows. To overcome these fundamental limitations in generalization and deployment flexibility, we propose Speech2Blend—an end-to-end hybrid convolutional-recurrent network that directly learns nonlinear speech-to-blendshape parameter mappings. This novel approach enables markerless speech-driven facial animation generation without restrictive inputs like video references or specialized facial rigs. Trained on the largest available digital human dataset (BEAT) and rigorously evaluated using three benchmark datasets with photorealistic visualization tools, Speech2Blend achieves state-of-the-art performance. It delivers superior audio-visual synchronization through learned temporal dynamics and reduces lip vertex error by 30% compared to existing baseline methods. These advances significantly lower production costs for virtual human speech animation while enabling cross-platform compatibility with common game engines and animation software.
语音驱动面部动画的最新进展引起了计算机图形学、人机交互系统和沉浸式虚拟现实应用的极大兴趣。然而,现有的方法仍然受到特定参考视频或专有面部网格结构的依赖的限制,限制了它们在不同生产管道中的适用性,并降低了与行业标准动画工作流程的兼容性。为了克服泛化和部署灵活性方面的这些基本限制,我们提出了一个端到端混合卷积循环网络,它直接学习非线性语音到混合形状参数映射。这种新颖的方法使无标记语音驱动的面部动画生成不需要限制性输入,如视频参考或专门的面部设备。在最大的可用数字人类数据集(BEAT)上进行训练,并使用三个基准数据集和逼真的可视化工具进行严格评估,Speech2Blend达到了最先进的性能。与现有的基线方法相比,它通过学习时间动态提供了卓越的视听同步,并减少了30%的唇顶点误差。这些进步大大降低了虚拟人类语音动画的制作成本,同时实现了与通用游戏引擎和动画软件的跨平台兼容性。
{"title":"Speech2Blend: A Hybrid Network for Speech-Driven 3-D Facial Animation by Learning Blendshape","authors":"Lei Wang;Gongbin Chen;Feng Liu;Jiaji Wu;Jun Cheng","doi":"10.1109/THMS.2025.3640719","DOIUrl":"https://doi.org/10.1109/THMS.2025.3640719","url":null,"abstract":"Recent advances in speech-driven facial animation have attracted significant interest across computer graphics, human–computer interaction systems, and immersive virtual reality applications. However, existing methods remain constrained by dependencies on specific reference videos or proprietary face mesh structures, limiting their applicability across diverse production pipelines and reducing compatibility with industry-standard animation workflows. To overcome these fundamental limitations in generalization and deployment flexibility, we propose Speech2Blend—an end-to-end hybrid convolutional-recurrent network that directly learns nonlinear speech-to-blendshape parameter mappings. This novel approach enables markerless speech-driven facial animation generation without restrictive inputs like video references or specialized facial rigs. Trained on the largest available digital human dataset (BEAT) and rigorously evaluated using three benchmark datasets with photorealistic visualization tools, Speech2Blend achieves state-of-the-art performance. It delivers superior audio-visual synchronization through learned temporal dynamics and reduces lip vertex error by 30% compared to existing baseline methods. These advances significantly lower production costs for virtual human speech animation while enabling cross-platform compatibility with common game engines and animation software.","PeriodicalId":48916,"journal":{"name":"IEEE Transactions on Human-Machine Systems","volume":"56 1","pages":"48-57"},"PeriodicalIF":4.4,"publicationDate":"2025-12-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146045351","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Transactions on Human-Machine Systems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1