首页 > 最新文献

IEEE Transactions on Cognitive and Developmental Systems最新文献

英文 中文
HDMTK: Full Integration of Hierarchical Decision-Making and Tactical Knowledge in Multiagent Adversarial Games HDMTK:多智能体对抗博弈中层次决策和战术知识的全面整合
IF 5 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-30 DOI: 10.1109/TCDS.2024.3470068
Wei Li;Boling Hu;Aiguo Song;Kaizhu Huang
In the field of adversarial games, existing decision-making algorithms primarily rely on reinforcement learning, which can theoretically adapt to diverse scenarios through trial and error. However, these algorithms often face the challenges of low effectiveness and slow convergence in complex wargame environments. Inspired by how human commanders make decisions, this article proposes a novel method named full integration of hierarchical decision-making and tactical knowledge (HDMTK). This method comprises an upper reinforcement learning module and a lower multiagent reinforcement learning (MARL) module. To enable agents to efficiently learn the cooperative strategy, in HDMTK, we separate the whole task into explainable subtasks and devise their corresponding subgoals for shaping the online rewards based on tactical knowledge. Experimental results on the wargame simulation platform “MiaoSuan” show that, compared to the advanced MARL methods, HDMTK exhibits superior performance and faster convergence in the complex scenarios.
在对抗博弈领域,现有的决策算法主要依赖于强化学习,理论上可以通过试错来适应不同的场景。然而,这些算法在复杂的兵棋环境中往往面临效率低、收敛速度慢的挑战。受人类指挥官决策方式的启发,本文提出了一种新的分层决策与战术知识完全集成方法(HDMTK)。该方法包括上部强化学习模块和下部多智能体强化学习模块。为了使智能体能够有效地学习合作策略,在HDMTK中,我们将整个任务分解为可解释的子任务,并根据战术知识设计相应的子目标来形成在线奖励。在“妙算”兵棋模拟平台上的实验结果表明,与先进的MARL方法相比,HDMTK在复杂场景下表现出更优越的性能和更快的收敛速度。
{"title":"HDMTK: Full Integration of Hierarchical Decision-Making and Tactical Knowledge in Multiagent Adversarial Games","authors":"Wei Li;Boling Hu;Aiguo Song;Kaizhu Huang","doi":"10.1109/TCDS.2024.3470068","DOIUrl":"https://doi.org/10.1109/TCDS.2024.3470068","url":null,"abstract":"In the field of adversarial games, existing decision-making algorithms primarily rely on reinforcement learning, which can theoretically adapt to diverse scenarios through trial and error. However, these algorithms often face the challenges of low effectiveness and slow convergence in complex wargame environments. Inspired by how human commanders make decisions, this article proposes a novel method named full integration of hierarchical decision-making and tactical knowledge (HDMTK). This method comprises an upper reinforcement learning module and a lower multiagent reinforcement learning (MARL) module. To enable agents to efficiently learn the cooperative strategy, in HDMTK, we separate the whole task into explainable subtasks and devise their corresponding subgoals for shaping the online rewards based on tactical knowledge. Experimental results on the wargame simulation platform “MiaoSuan” show that, compared to the advanced MARL methods, HDMTK exhibits superior performance and faster convergence in the complex scenarios.","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"17 3","pages":"465-479"},"PeriodicalIF":5.0,"publicationDate":"2024-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144213528","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Graph-Laplacian-Processing-Based Multimodal Localization Backend for Robots and Autonomous Systems 基于图拉普拉斯处理的机器人和自主系统多模态定位后端
IF 5 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-26 DOI: 10.1109/TCDS.2024.3468712
Nikos Piperigkos;Christos Anagnostopoulos;Aris S. Lalos;Petros Kapsalas;Duong Van Nguyen
Simultaneous localization and mapping (SLAM) for positioning of robots and autonomous systems (RASs) and mapping of their surrounding environments is a task of major significance in various applications. However, the main disadvantage of traditional SLAM is that the deployed backend modules suffer from accumulative error caused by sharp viewpoint changes, diverse weather conditions, etc. As such, to improve the localization accuracy of the moving agents, we propose a cost-effective and loosely coupled relocalization backend, deployed on top of original SLAM algorithms, which exploits the topologies of poses and landmarks generated either by camera, LiDAR, or mechanical sensors, to couple and fuse them. This novel fusion scheme enhances the decision-making ability and adaptability of autonomous systems, akin to human cognition, by elaborating graph Laplacian processing concept with Kalman filters. Initially designed for cooperative localization of active road users, this approach optimally combines multisensor information through graph signal processing and Bayesian estimation for self-positioning. Conducted experiments were focused on evaluating how our approach can improve the positioning of autonomous ground vehicles, as prominent examples of RASs equipped with sensing capabilities, in challenging outdoor environments. More specifically, experiments were carried out using the CARLA simulator to generate different types of driving trajectories and environmental conditions, as well as real automotive data captured by an operating vehicle in Langen, Germany. Evaluation study demonstrates that localization accuracy is greatly improved both in terms of overall trajectory error as well as loop closing accuracy for each sensor fusion configuration.
同时定位和测绘(SLAM)是机器人和自主系统(ras)定位及其周围环境测绘的一项重要任务,在各种应用中具有重要意义。然而,传统SLAM的主要缺点是所部署的后端模块存在视点急剧变化、多变天气条件等导致的累积误差。因此,为了提高移动代理的定位精度,我们提出了一种经济高效且松散耦合的重新定位后端,部署在原始SLAM算法之上,该算法利用相机、激光雷达或机械传感器生成的姿势和地标拓扑来耦合和融合它们。该融合方案通过利用卡尔曼滤波器阐述图拉普拉斯处理概念,增强了自治系统类似于人类认知的决策能力和自适应能力。该方法最初是为主动道路使用者的协同定位而设计的,通过图信号处理和贝叶斯估计将多传感器信息最佳地结合在一起进行自定位。所进行的实验重点是评估我们的方法如何改善自动地面车辆的定位,作为配备传感能力的自动驾驶车辆在具有挑战性的室外环境中的突出例子。更具体地说,使用CARLA模拟器进行实验,以生成不同类型的驾驶轨迹和环境条件,以及德国兰根一辆正在行驶的车辆捕获的真实汽车数据。评估研究表明,无论从总体轨迹误差还是闭环精度方面,每种传感器融合配置都大大提高了定位精度。
{"title":"Graph-Laplacian-Processing-Based Multimodal Localization Backend for Robots and Autonomous Systems","authors":"Nikos Piperigkos;Christos Anagnostopoulos;Aris S. Lalos;Petros Kapsalas;Duong Van Nguyen","doi":"10.1109/TCDS.2024.3468712","DOIUrl":"https://doi.org/10.1109/TCDS.2024.3468712","url":null,"abstract":"Simultaneous localization and mapping (SLAM) for positioning of robots and autonomous systems (RASs) and mapping of their surrounding environments is a task of major significance in various applications. However, the main disadvantage of traditional SLAM is that the deployed backend modules suffer from accumulative error caused by sharp viewpoint changes, diverse weather conditions, etc. As such, to improve the localization accuracy of the moving agents, we propose a cost-effective and loosely coupled relocalization backend, deployed on top of original SLAM algorithms, which exploits the topologies of poses and landmarks generated either by camera, LiDAR, or mechanical sensors, to couple and fuse them. This novel fusion scheme enhances the decision-making ability and adaptability of autonomous systems, akin to human cognition, by elaborating graph Laplacian processing concept with Kalman filters. Initially designed for cooperative localization of active road users, this approach optimally combines multisensor information through graph signal processing and Bayesian estimation for self-positioning. Conducted experiments were focused on evaluating how our approach can improve the positioning of autonomous ground vehicles, as prominent examples of RASs equipped with sensing capabilities, in challenging outdoor environments. More specifically, experiments were carried out using the CARLA simulator to generate different types of driving trajectories and environmental conditions, as well as real automotive data captured by an operating vehicle in Langen, Germany. Evaluation study demonstrates that localization accuracy is greatly improved both in terms of overall trajectory error as well as loop closing accuracy for each sensor fusion configuration.","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"17 2","pages":"436-453"},"PeriodicalIF":5.0,"publicationDate":"2024-09-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143761405","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Touch Gesture Recognition-Based Physical Human–Robot Interaction for Collaborative Tasks 基于触摸手势识别的人机协作任务物理交互
IF 5 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-24 DOI: 10.1109/TCDS.2024.3466553
Dawoon Jung;Chengyan Gu;Junmin Park;Joono Cheong
Human–robot collaboration (HRC) has recently attracted increasing attention as a vital component of next-generation automated manufacturing and assembly tasks, yet physical human–robot interaction (pHRI)—which is an inevitable component of collaboration—is often limited to rudimentary touches. This article therefore proposes a deep-learning-based pHRI method that utilizes predefined types of human touch gestures as intuitive communicative signs for collaborative tasks. To this end, a touch gesture network model is first designed upon the framework of the gated recurrent unit (GRU) network, which accepts a set of ground-truth dynamic responses (energy change, generalized momentum, and external joint torque) of robot manipulators under the action of known types of touch gestures and learns to predict the five representative touch gesture types and the corresponding link toward a random touch gesture input. After training the GRU-based touch gesture model using a collected dataset of dynamic responses of a robot manipulator, a total of 35 outputs (five gesture types with seven links each) is recognized with 96.94% accuracy. The experimental results of recognition accuracy correlated with the touch gesture types, and their strength results are shown to validate the performance and disclose the characteristics of the proposed touch gesture model. An example of an IKEA chair assembly task is also presented to demonstrate a collaborative task using the proposed touch gestures. By developing the proposed pHRI method and demonstrating its applicability, we expect that this method can help position physical interaction as one of the key modalities for communication in real-world HRC applications.
作为下一代自动化制造和装配任务的重要组成部分,人机协作(HRC)最近引起了越来越多的关注,然而物理人机交互(pHRI)——这是协作的一个不可避免的组成部分——往往仅限于基本的接触。因此,本文提出了一种基于深度学习的pHRI方法,该方法利用预定义类型的人类触摸手势作为协作任务的直观交流符号。为此,首先在门控递归单元(GRU)网络框架上设计了一个触摸手势网络模型,该模型接受机器人在已知类型的触摸手势作用下的一组真实动态响应(能量变化、广义动量和外部关节扭矩),并学习预测五种具有代表性的触摸手势类型及其与随机触摸手势输入的对应链接。利用收集到的机器人动态响应数据集对基于gru的触摸手势模型进行训练后,共识别出35个输出(5种手势类型,每种手势7个链接),准确率为96.94%。实验结果表明,识别精度与触摸手势类型及其强度相关,验证了所提触摸手势模型的性能,揭示了所提触摸手势模型的特点。还以宜家椅子组装任务为例,演示了使用提议的触摸手势的协作任务。通过开发所提出的pHRI方法并证明其适用性,我们期望该方法可以帮助将物理交互定位为现实世界HRC应用中通信的关键模式之一。
{"title":"Touch Gesture Recognition-Based Physical Human–Robot Interaction for Collaborative Tasks","authors":"Dawoon Jung;Chengyan Gu;Junmin Park;Joono Cheong","doi":"10.1109/TCDS.2024.3466553","DOIUrl":"https://doi.org/10.1109/TCDS.2024.3466553","url":null,"abstract":"Human–robot collaboration (HRC) has recently attracted increasing attention as a vital component of next-generation automated manufacturing and assembly tasks, yet physical human–robot interaction (pHRI)—which is an inevitable component of collaboration—is often limited to rudimentary touches. This article therefore proposes a deep-learning-based pHRI method that utilizes predefined types of human touch gestures as intuitive communicative signs for collaborative tasks. To this end, a touch gesture network model is first designed upon the framework of the gated recurrent unit (GRU) network, which accepts a set of ground-truth dynamic responses (energy change, generalized momentum, and external joint torque) of robot manipulators under the action of known types of touch gestures and learns to predict the five representative touch gesture types and the corresponding link toward a random touch gesture input. After training the GRU-based touch gesture model using a collected dataset of dynamic responses of a robot manipulator, a total of 35 outputs (five gesture types with seven links each) is recognized with 96.94% accuracy. The experimental results of recognition accuracy correlated with the touch gesture types, and their strength results are shown to validate the performance and disclose the characteristics of the proposed touch gesture model. An example of an IKEA chair assembly task is also presented to demonstrate a collaborative task using the proposed touch gestures. By developing the proposed pHRI method and demonstrating its applicability, we expect that this method can help position physical interaction as one of the key modalities for communication in real-world HRC applications.","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"17 2","pages":"421-435"},"PeriodicalIF":5.0,"publicationDate":"2024-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143761496","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing Dimensional Image Emotion Detection With a Low-Resource Dataset via Two-Stage Training 基于两阶段训练的低资源数据集增强多维图像情感检测
IF 5 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-23 DOI: 10.1109/TCDS.2024.3465602
SangEun Lee;Seoyun Kim;Yubeen Lee;Jufeng Yang;Eunil Park
Image emotion analysis has gained notable attention owing to the growing importance of computationally modeling human emotions. Most previous studies have focused on classifying the feelings evoked by an image into predefined emotion categories. Compared with these categorical approaches which cannot address the ambiguity and complexity of human emotions, recent studies have taken dimensional approaches to address these problems. However, there is still a limitation in that the number of dimensional datasets is significantly smaller for model training, compared with many available categorical datasets. We propose four types of frameworks that use categorical datasets to predict emotion values for a given image in the valence–arousal (VA) space. Specifically, our proposed framework is trained to predict continuous emotion values under the supervision of categorical labels. Extensive experiments demonstrate that our approach showed a positive correlation with the actual VA values of the dimensional dataset. In addition, our framework improves further when a small number of dimensional datasets are available for the fine-tuning process.
由于人类情感的计算建模越来越重要,图像情感分析得到了显著的关注。大多数先前的研究都集中在将图像所唤起的感觉分类为预定义的情感类别。与这些分类方法无法解决人类情感的模糊性和复杂性相比,最近的研究采用了维度方法来解决这些问题。然而,与许多可用的分类数据集相比,用于模型训练的维度数据集的数量明显更少,这仍然是一个限制。我们提出了四种类型的框架,它们使用分类数据集来预测给定图像在价-觉醒(VA)空间中的情感值。具体来说,我们提出的框架被训练成在分类标签的监督下预测连续的情感值。大量的实验表明,我们的方法与维度数据集的实际VA值呈正相关。此外,当少量维度数据集可用于微调过程时,我们的框架会进一步改进。
{"title":"Enhancing Dimensional Image Emotion Detection With a Low-Resource Dataset via Two-Stage Training","authors":"SangEun Lee;Seoyun Kim;Yubeen Lee;Jufeng Yang;Eunil Park","doi":"10.1109/TCDS.2024.3465602","DOIUrl":"https://doi.org/10.1109/TCDS.2024.3465602","url":null,"abstract":"Image emotion analysis has gained notable attention owing to the growing importance of computationally modeling human emotions. Most previous studies have focused on classifying the feelings evoked by an image into predefined emotion categories. Compared with these categorical approaches which cannot address the ambiguity and complexity of human emotions, recent studies have taken dimensional approaches to address these problems. However, there is still a limitation in that the number of dimensional datasets is significantly smaller for model training, compared with many available categorical datasets. We propose four types of frameworks that use categorical datasets to predict emotion values for a given image in the valence–arousal (VA) space. Specifically, our proposed framework is trained to predict continuous emotion values under the supervision of categorical labels. Extensive experiments demonstrate that our approach showed a positive correlation with the actual VA values of the dimensional dataset. In addition, our framework improves further when a small number of dimensional datasets are available for the fine-tuning process.","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"17 3","pages":"455-464"},"PeriodicalIF":5.0,"publicationDate":"2024-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144213666","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Embodied Perception Interaction, and Cognition for Wearable Robotics: A Survey 可穿戴机器人的嵌入式感知、交互和认知:调查
IF 5 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-18 DOI: 10.1109/tcds.2024.3463194
Xiaoyu Wu, Jiale Liang, Yiang Yu, Guoxin Li, Gary G. Yen, Haoyong Yu
{"title":"Embodied Perception Interaction, and Cognition for Wearable Robotics: A Survey","authors":"Xiaoyu Wu, Jiale Liang, Yiang Yu, Guoxin Li, Gary G. Yen, Haoyong Yu","doi":"10.1109/tcds.2024.3463194","DOIUrl":"https://doi.org/10.1109/tcds.2024.3463194","url":null,"abstract":"","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"48 1","pages":""},"PeriodicalIF":5.0,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142269922","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CS-SLAM: A lightweight semantic SLAM method for dynamic scenarios CS-SLAM:适用于动态场景的轻量级语义 SLAM 方法
IF 5 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-17 DOI: 10.1109/tcds.2024.3462651
Zhendong Guo, Na Dong, Zehui Zhang, Xiaoming Mai, Donghui Li
{"title":"CS-SLAM: A lightweight semantic SLAM method for dynamic scenarios","authors":"Zhendong Guo, Na Dong, Zehui Zhang, Xiaoming Mai, Donghui Li","doi":"10.1109/tcds.2024.3462651","DOIUrl":"https://doi.org/10.1109/tcds.2024.3462651","url":null,"abstract":"","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"49 1","pages":""},"PeriodicalIF":5.0,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142266493","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Edge-Centric Functional-Connectivity-Based Cofluctuation-Guided Subcortical Connectivity Network Construction 以边缘为中心、基于共波动引导的皮层下功能连接网络构建
IF 5 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-17 DOI: 10.1109/TCDS.2024.3462709
Qinrui Ling;Aiping Liu;Taomian Mi;Piu Chan;Xun Chen
Subcortical regions can be functionally organized into connectivity networks and are extensively communicated with the cortex via reciprocal connections. However, most current research on subcortical networks ignores these interconnections, and networks of the whole brain are of high dimensionality and computational complexity. In this article, we propose a novel cofluctuation-guided subcortical connectivity network construction model based on edge-centric functional connectivity (FC). It is capable of extracting the cofluctuations between the cortex and subcortex and constructing dynamic subcortical networks based on these interconnections. Blind source separation approaches with domain knowledge are designed for dimensionality reduction and feature extraction. Great reproducibility and reliability were achieved when applying our model to two sessions of functional magnetic resonance imaging (fMRI) data. Cortical areas having synchronous communications with the cortex were detected, which was unable to be revealed by traditional node-centric FC. Significant alterations in connectivity patterns were observed when dealing with fMRI of subjects with and without Parkinson's disease, which were further correlated to clinical scores. These validations demonstrated that our model provided a promising strategy for brain network construction, exhibiting great potential in clinical practice.
皮层下区域可以在功能上组织成连接网络,并通过相互连接与皮层广泛沟通。然而,目前大多数关于皮层下网络的研究忽略了这些相互联系,而且整个大脑的网络具有高维性和计算复杂性。在本文中,我们提出了一种基于边缘中心功能连接(FC)的共同波动引导下皮层下连接网络构建模型。它能够提取皮层和下皮层之间的共同波动,并基于这些相互联系构建动态的皮层下网络。设计了基于领域知识的盲源分离方法,用于降维和特征提取。将我们的模型应用于两次功能磁共振成像(fMRI)数据时,获得了很高的再现性和可靠性。检测到与皮层有同步通信的皮质区域,这是传统的以节点为中心的FC无法显示的。在处理帕金森病患者和非帕金森病患者的fMRI时,观察到连接模式的显著改变,这与临床评分进一步相关。这些验证表明,我们的模型为脑网络构建提供了一种很有前景的策略,在临床实践中具有很大的潜力。
{"title":"Edge-Centric Functional-Connectivity-Based Cofluctuation-Guided Subcortical Connectivity Network Construction","authors":"Qinrui Ling;Aiping Liu;Taomian Mi;Piu Chan;Xun Chen","doi":"10.1109/TCDS.2024.3462709","DOIUrl":"10.1109/TCDS.2024.3462709","url":null,"abstract":"Subcortical regions can be functionally organized into connectivity networks and are extensively communicated with the cortex via reciprocal connections. However, most current research on subcortical networks ignores these interconnections, and networks of the whole brain are of high dimensionality and computational complexity. In this article, we propose a novel cofluctuation-guided subcortical connectivity network construction model based on edge-centric functional connectivity (FC). It is capable of extracting the cofluctuations between the cortex and subcortex and constructing dynamic subcortical networks based on these interconnections. Blind source separation approaches with domain knowledge are designed for dimensionality reduction and feature extraction. Great reproducibility and reliability were achieved when applying our model to two sessions of functional magnetic resonance imaging (fMRI) data. Cortical areas having synchronous communications with the cortex were detected, which was unable to be revealed by traditional node-centric FC. Significant alterations in connectivity patterns were observed when dealing with fMRI of subjects with and without Parkinson's disease, which were further correlated to clinical scores. These validations demonstrated that our model provided a promising strategy for brain network construction, exhibiting great potential in clinical practice.","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"17 2","pages":"390-399"},"PeriodicalIF":5.0,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142269531","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Unveiling Thoughts: A Review of Advancements in EEG Brain Signal Decoding Into Text 揭开思想的面纱:脑电图脑信号解码为文本的进展回顾
IF 5 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-17 DOI: 10.1109/TCDS.2024.3462452
Saydul Akbar Murad;Nick Rahimi
The conversion of brain activity into text using electroencephalography (EEG) has gained significant traction in recent years. Many researchers are working to develop new models to decode EEG signals into text form. Although this area has shown promising developments, it still faces numerous challenges that necessitate further improvement. It is important to outline this area's recent developments and future research directions to provide a comprehensive understanding of the current state of technology, guide future research efforts, and enhance the effectiveness and accessibility of EEG-to-text systems. In this review article, we thoroughly summarize the progress in EEG-to-text conversion. First, we talk about how EEG-to-text technology has grown and what problems the field still faces. Second, we discuss existing techniques used in this field. This includes methods for collecting EEG data, the steps to process these signals, and the development of systems capable of translating these signals into coherent text. We conclude with potential future research directions, emphasizing the need for enhanced accuracy, reduced system constraints, and the exploration of novel applications across varied sectors. By addressing these aspects, this review aims to contribute to developing more accessible and effective brain–computer interface (BCI) technology for a broader user base.
近年来,利用脑电图(EEG)将大脑活动转化为文本的研究得到了极大的关注。许多研究人员正在努力开发新的模型,将脑电图信号解码成文本形式。虽然这一领域已显示出有希望的发展,但仍面临许多需要进一步改进的挑战。重要的是概述该领域的最新发展和未来的研究方向,以提供对当前技术状态的全面了解,指导未来的研究工作,并提高脑电图到文本系统的有效性和可访问性。在这篇综述文章中,我们全面总结了脑电图到文本转换的进展。首先,我们讨论脑电图文本转换技术是如何发展的,以及该领域仍然面临哪些问题。其次,我们讨论了该领域中使用的现有技术。这包括收集脑电图数据的方法,处理这些信号的步骤,以及能够将这些信号翻译成连贯文本的系统的开发。我们总结了潜在的未来研究方向,强调需要提高准确性,减少系统约束,以及探索不同领域的新应用。通过对这些方面的探讨,本文旨在为更广泛的用户基础开发更易于访问和有效的脑机接口(BCI)技术做出贡献。
{"title":"Unveiling Thoughts: A Review of Advancements in EEG Brain Signal Decoding Into Text","authors":"Saydul Akbar Murad;Nick Rahimi","doi":"10.1109/TCDS.2024.3462452","DOIUrl":"10.1109/TCDS.2024.3462452","url":null,"abstract":"The conversion of brain activity into text using electroencephalography (EEG) has gained significant traction in recent years. Many researchers are working to develop new models to decode EEG signals into text form. Although this area has shown promising developments, it still faces numerous challenges that necessitate further improvement. It is important to outline this area's recent developments and future research directions to provide a comprehensive understanding of the current state of technology, guide future research efforts, and enhance the effectiveness and accessibility of EEG-to-text systems. In this review article, we thoroughly summarize the progress in EEG-to-text conversion. First, we talk about how EEG-to-text technology has grown and what problems the field still faces. Second, we discuss existing techniques used in this field. This includes methods for collecting EEG data, the steps to process these signals, and the development of systems capable of translating these signals into coherent text. We conclude with potential future research directions, emphasizing the need for enhanced accuracy, reduced system constraints, and the exploration of novel applications across varied sectors. By addressing these aspects, this review aims to contribute to developing more accessible and effective brain–computer interface (BCI) technology for a broader user base.","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"17 1","pages":"61-76"},"PeriodicalIF":5.0,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142266494","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The Methodology of Quantitative Social Intention Evaluation and Robot Gaze Behavior Control in Multiobjects Scenario 多目标情景下定量社会意向评价与机器人凝视行为控制方法
IF 5 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-16 DOI: 10.1109/TCDS.2024.3461335
Haoyu Zhu;Xiaorui Liu;Hang Su;Wei Wang;Jinpeng Yu
This article focuses on the multiple objects selection problem for the robot in social scenarios, and proposes a novel methodology composed of quantitative social intention evaluation and gaze behavior control. For the social scenarios containing various persons and multimodal social cues, a combination of the entropy weight method (EWM) and gray correlation-order preference by similarity to the ideal solution (GC-TOPSIS) model is proposed to fuse the multimodal social cues, and evaluate the social intention of candidates. According to the quantitative evaluation of social intention, a robot can generate the interaction priority among multiple social candidates. To ensure this interaction selection mechanism in behavior level, an optimal control framework composed of model predictive controller (MPC) and online Gaussian process (GP) observer is employed to drive the eye-head coordinated gaze behavior of robot. Through the experiments conducted on the Xiaopang robot, the availability of the proposed methodology can be illustrated. This work enables robots to generate social behavior based on quantitative intention perception, which could bring the potential to explore the sensory principles and biomechanical mechanism underlying the human-robot interaction, and broaden the application of robot in the social scenario.
针对机器人在社交场景下的多目标选择问题,提出了一种由定量社交意图评估和凝视行为控制组成的新方法。针对包含多个人和多模态社会线索的社会场景,提出了一种结合熵权法(EWM)和灰色关联-排序理想解相似性偏好(GC-TOPSIS)模型的方法来融合多模态社会线索,并对候选人的社会意向进行评估。根据社会意向的定量评价,机器人可以生成多个社会候选者之间的交互优先级。为了保证这种行为层面的交互选择机制,采用由模型预测控制器(MPC)和在线高斯过程观测器(GP)组成的最优控制框架驱动机器人的眼-头协调注视行为。通过在小胖机器人上进行的实验,可以说明所提出方法的有效性。本研究使机器人能够基于定量意向感知产生社会行为,为探索人机交互的感觉原理和生物力学机制,拓宽机器人在社会场景中的应用提供了可能。
{"title":"The Methodology of Quantitative Social Intention Evaluation and Robot Gaze Behavior Control in Multiobjects Scenario","authors":"Haoyu Zhu;Xiaorui Liu;Hang Su;Wei Wang;Jinpeng Yu","doi":"10.1109/TCDS.2024.3461335","DOIUrl":"https://doi.org/10.1109/TCDS.2024.3461335","url":null,"abstract":"This article focuses on the multiple objects selection problem for the robot in social scenarios, and proposes a novel methodology composed of quantitative social intention evaluation and gaze behavior control. For the social scenarios containing various persons and multimodal social cues, a combination of the entropy weight method (EWM) and gray correlation-order preference by similarity to the ideal solution (GC-TOPSIS) model is proposed to fuse the multimodal social cues, and evaluate the social intention of candidates. According to the quantitative evaluation of social intention, a robot can generate the interaction priority among multiple social candidates. To ensure this interaction selection mechanism in behavior level, an optimal control framework composed of model predictive controller (MPC) and online Gaussian process (GP) observer is employed to drive the eye-head coordinated gaze behavior of robot. Through the experiments conducted on the Xiaopang robot, the availability of the proposed methodology can be illustrated. This work enables robots to generate social behavior based on quantitative intention perception, which could bring the potential to explore the sensory principles and biomechanical mechanism underlying the human-robot interaction, and broaden the application of robot in the social scenario.","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"17 2","pages":"400-409"},"PeriodicalIF":5.0,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143761402","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Mental Workload Assessment Using Deep Learning Models From EEG Signals: A Systematic Review 利用脑电图信号的深度学习模型评估心理负荷:系统综述
IF 5 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-16 DOI: 10.1109/TCDS.2024.3460750
Kunjira Kingphai;Yashar Moshfeghi
Mental workload (MWL) assessment is crucial in information systems (IS), impacting task performance, user experience, and system effectiveness. Deep learning offers promising techniques for MWL classification using electroencephalography (EEG), which monitors cognitive states dynamically and unobtrusively. Our research explores deep learning's potential and challenges in EEG-based MWL classification, focusing on training inputs, cross-validation methods, and classification problem types. We identify five types of EEG-based MWL classification: within-subject, cross subject, cross session, cross task, and combined cross task and cross subject. Success depends on managing dataset uniqueness, session and task variability, and artifact removal. Despite the potential, real-world applications are limited. Enhancements are necessary for self-reporting methods, universal preprocessing standards, and MWL assessment accuracy. Specifically, inaccuracies are inflated when data are shuffled before splitting to train and test sets, disrupting EEG signals’ temporal sequence. In contrast, methods such as the time-series cross validation and leave-session-out approach better preserve temporal integrity, offering more accurate model performance evaluations. Utilizing deep learning for EEG-based MWL assessment could significantly improve IS functionality and adaptability in real time based on user cognitive states.
心理工作量(MWL)评估在信息系统(is)中至关重要,影响任务性能、用户体验和系统有效性。深度学习为使用脑电图(EEG)进行MWL分类提供了有前途的技术,该技术动态且不引人注目地监测认知状态。我们的研究探讨了深度学习在基于脑电图的MWL分类中的潜力和挑战,重点关注训练输入、交叉验证方法和分类问题类型。我们确定了五种基于脑电图的MWL分类:主题内分类、跨主题分类、跨会话分类、跨任务分类以及跨任务和跨主题组合分类。成功取决于对数据集唯一性、会话和任务可变性以及工件移除的管理。尽管有潜力,但实际应用是有限的。自我报告方法、通用预处理标准和MWL评估准确性需要改进。具体来说,当数据在分割成训练集和测试集之前被洗牌时,不准确性会被夸大,从而破坏脑电图信号的时间序列。相比之下,时间序列交叉验证和离开会话方法等方法更好地保持了时间完整性,提供了更准确的模型性能评估。利用深度学习进行基于脑电图的MWL评估可以显著提高基于用户认知状态的IS功能和实时适应性。
{"title":"Mental Workload Assessment Using Deep Learning Models From EEG Signals: A Systematic Review","authors":"Kunjira Kingphai;Yashar Moshfeghi","doi":"10.1109/TCDS.2024.3460750","DOIUrl":"https://doi.org/10.1109/TCDS.2024.3460750","url":null,"abstract":"Mental workload (MWL) assessment is crucial in information systems (IS), impacting task performance, user experience, and system effectiveness. Deep learning offers promising techniques for MWL classification using electroencephalography (EEG), which monitors cognitive states dynamically and unobtrusively. Our research explores deep learning's potential and challenges in EEG-based MWL classification, focusing on training inputs, cross-validation methods, and classification problem types. We identify five types of EEG-based MWL classification: within-subject, cross subject, cross session, cross task, and combined cross task and cross subject. Success depends on managing dataset uniqueness, session and task variability, and artifact removal. Despite the potential, real-world applications are limited. Enhancements are necessary for self-reporting methods, universal preprocessing standards, and MWL assessment accuracy. Specifically, inaccuracies are inflated when data are shuffled before splitting to train and test sets, disrupting EEG signals’ temporal sequence. In contrast, methods such as the time-series cross validation and leave-session-out approach better preserve temporal integrity, offering more accurate model performance evaluations. Utilizing deep learning for EEG-based MWL assessment could significantly improve IS functionality and adaptability in real time based on user cognitive states.","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"17 1","pages":"40-60"},"PeriodicalIF":5.0,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143361081","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Transactions on Cognitive and Developmental Systems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1