首页 > 最新文献

IEEE Transactions on Cognitive and Developmental Systems最新文献

英文 中文
PDRL: Towards Deeper States and Further Behaviors in Unsupervised Skill Discovery by Progressive Diversity PDRL:渐进式多样性在无监督技能发现中的深层状态和进一步行为
IF 5 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-02 DOI: 10.1109/TCDS.2024.3471645
Ziming He;Chao Song;Jingchen Li;Haobin Shi
We present progressive diversity reinforcement learning (PDRL), an unsupervised reinforcement learning (URL) method for discovering diverse skills. PDRL encourages learning behaviors that span multiple steps, particularly by introducing “deeper states”—states that require a longer sequence of actions to reach without repetition. To address the challenges of weak skill diversity and weak exploration in partially observable environments, PDRL employs two indications for skill learning to foster exploration and skill diversity, emphasizing each observation and subtrajectory's accuracy compared to its predecessor. Skill latent variables are represented by mappings from states or trajectories, helping to distinguish and recover learned skills. This dual representation promotes exploration and skill diversity without additional modeling or prior knowledge. PDRL also integrates intrinsic rewards through a combination of observations and subtrajectories, effectively preventing skill duplication. Experiments across multiple benchmarks show that PDRL discovers a broader range of skills compared to existing methods. Additionally, pretraining with PDRL accelerates fine-tuning in goal-conditioned reinforcement learning (GCRL) tasks, as demonstrated in Fetch robotic manipulation tasks.
我们提出了渐进式多样性强化学习(PDRL),一种用于发现多样化技能的无监督强化学习(URL)方法。PDRL鼓励跨越多个步骤的学习行为,特别是通过引入“更深层次的状态”——需要更长的动作序列才能达到的状态,而不需要重复。为了解决在部分可观察环境中弱技能多样性和弱探索的挑战,PDRL采用了两种技能学习指示来促进探索和技能多样性,强调与前一个相比,每个观察和子轨迹的准确性。技能潜变量由状态或轨迹映射表示,有助于区分和恢复所学技能。这种双重表示促进了探索和技能多样性,而无需额外的建模或先验知识。PDRL还通过观察和子轨迹的结合整合了内在奖励,有效地防止了技能重复。跨多个基准测试的实验表明,与现有方法相比,PDRL发现了更广泛的技能范围。此外,PDRL预训练加速了目标条件强化学习(GCRL)任务的微调,如Fetch机器人操作任务所示。
{"title":"PDRL: Towards Deeper States and Further Behaviors in Unsupervised Skill Discovery by Progressive Diversity","authors":"Ziming He;Chao Song;Jingchen Li;Haobin Shi","doi":"10.1109/TCDS.2024.3471645","DOIUrl":"https://doi.org/10.1109/TCDS.2024.3471645","url":null,"abstract":"We present progressive diversity reinforcement learning (PDRL), an unsupervised reinforcement learning (URL) method for discovering diverse skills. PDRL encourages learning behaviors that span multiple steps, particularly by introducing “deeper states”—states that require a longer sequence of actions to reach without repetition. To address the challenges of weak skill diversity and weak exploration in partially observable environments, PDRL employs two indications for skill learning to foster exploration and skill diversity, emphasizing each observation and subtrajectory's accuracy compared to its predecessor. Skill latent variables are represented by mappings from states or trajectories, helping to distinguish and recover learned skills. This dual representation promotes exploration and skill diversity without additional modeling or prior knowledge. PDRL also integrates intrinsic rewards through a combination of observations and subtrajectories, effectively preventing skill duplication. Experiments across multiple benchmarks show that PDRL discovers a broader range of skills compared to existing methods. Additionally, pretraining with PDRL accelerates fine-tuning in goal-conditioned reinforcement learning (GCRL) tasks, as demonstrated in Fetch robotic manipulation tasks.","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"17 3","pages":"495-509"},"PeriodicalIF":5.0,"publicationDate":"2024-10-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144213661","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Functional Connectivity Patterns Learning for EEG-Based Emotion Recognition 基于脑电图的情感识别的功能连接模式学习
IF 5 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-30 DOI: 10.1109/TCDS.2024.3470248
Chongxing Shi;C. L. Philip Chen;Shuzhen Li;Tong Zhang
Neuroscience research reveals that different emotions are associated with different functional connectivity structures of brain regions. However, many existing electroencephalography (EEG)-based emotion recognition methods use these connectivity patterns broadly without distinguishing between specific emotions. Additionally, the nonstationarity of EEG signals often results in high variations across different periods, leading models to extract time-specific features instead of emotional features. This article proposes a functional connectivity patterns learning network (FCPL) for EEG-based emotion recognition to address these challenges. FCPL includes a coefficient branch, a graph construction module, and a period domain adversarial module. These components capture individual characteristics and specific emotional connectivity patterns and reduce period-related variations, respectively. FCPL achieves state-of-the-art results: 42.04%/28.81% for seven-class subject-dependent/independent experiments on the MPED dataset, 97.45%/89.88% for subject-dependent/independent experiments on the SEED dataset, and 95.98%/96.19% for valence/arousal subject-dependent experiments and 67.90%/65.60% for valence/arousal subject-independent experiments on the DREAMER dataset. This work advances the exploration of functional connectivity structures in EEG signals from coarse-grained emotion-related patterns to fine-grained emotional distinctions, promoting neuroscience, and EEG-based emotion recognition technologies.
神经科学研究表明,不同的情绪与大脑区域不同的功能连接结构有关。然而,许多现有的基于脑电图(EEG)的情绪识别方法广泛地使用这些连接模式,而没有区分特定的情绪。此外,脑电图信号的非平稳性往往导致不同时期的变化很大,导致模型提取的是特定时间的特征,而不是情感特征。本文提出了一种基于脑电图的情感识别功能连接模式学习网络(FCPL)来解决这些挑战。FCPL包括一个系数分支、一个图构造模块和一个周期域对抗模块。这些组成部分分别捕捉个人特征和特定的情感连接模式,并减少与时期相关的变化。FCPL在MPED数据集上的7类受试者依赖/独立实验结果为42.04%/28.81%,SEED数据集上的受试者依赖/独立实验结果为97.45%/89.88%,而在做梦者数据集上的效价/唤醒受试者依赖实验结果为95.98%/96.19%,效价/唤醒受试者独立实验结果为67.90%/65.60%。这项工作推进了从粗粒度情绪相关模式到细粒度情绪区分的脑电图信号功能连接结构的探索,促进了神经科学和基于脑电图的情绪识别技术的发展。
{"title":"Functional Connectivity Patterns Learning for EEG-Based Emotion Recognition","authors":"Chongxing Shi;C. L. Philip Chen;Shuzhen Li;Tong Zhang","doi":"10.1109/TCDS.2024.3470248","DOIUrl":"https://doi.org/10.1109/TCDS.2024.3470248","url":null,"abstract":"Neuroscience research reveals that different emotions are associated with different functional connectivity structures of brain regions. However, many existing electroencephalography (EEG)-based emotion recognition methods use these connectivity patterns broadly without distinguishing between specific emotions. Additionally, the nonstationarity of EEG signals often results in high variations across different periods, leading models to extract time-specific features instead of emotional features. This article proposes a functional connectivity patterns learning network (FCPL) for EEG-based emotion recognition to address these challenges. FCPL includes a coefficient branch, a graph construction module, and a period domain adversarial module. These components capture individual characteristics and specific emotional connectivity patterns and reduce period-related variations, respectively. FCPL achieves state-of-the-art results: 42.04%/28.81% for seven-class subject-dependent/independent experiments on the MPED dataset, 97.45%/89.88% for subject-dependent/independent experiments on the SEED dataset, and 95.98%/96.19% for valence/arousal subject-dependent experiments and 67.90%/65.60% for valence/arousal subject-independent experiments on the DREAMER dataset. This work advances the exploration of functional connectivity structures in EEG signals from coarse-grained emotion-related patterns to fine-grained emotional distinctions, promoting neuroscience, and EEG-based emotion recognition technologies.","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"17 3","pages":"480-494"},"PeriodicalIF":5.0,"publicationDate":"2024-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144213531","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HDMTK: Full Integration of Hierarchical Decision-Making and Tactical Knowledge in Multiagent Adversarial Games HDMTK:多智能体对抗博弈中层次决策和战术知识的全面整合
IF 5 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-30 DOI: 10.1109/TCDS.2024.3470068
Wei Li;Boling Hu;Aiguo Song;Kaizhu Huang
In the field of adversarial games, existing decision-making algorithms primarily rely on reinforcement learning, which can theoretically adapt to diverse scenarios through trial and error. However, these algorithms often face the challenges of low effectiveness and slow convergence in complex wargame environments. Inspired by how human commanders make decisions, this article proposes a novel method named full integration of hierarchical decision-making and tactical knowledge (HDMTK). This method comprises an upper reinforcement learning module and a lower multiagent reinforcement learning (MARL) module. To enable agents to efficiently learn the cooperative strategy, in HDMTK, we separate the whole task into explainable subtasks and devise their corresponding subgoals for shaping the online rewards based on tactical knowledge. Experimental results on the wargame simulation platform “MiaoSuan” show that, compared to the advanced MARL methods, HDMTK exhibits superior performance and faster convergence in the complex scenarios.
在对抗博弈领域,现有的决策算法主要依赖于强化学习,理论上可以通过试错来适应不同的场景。然而,这些算法在复杂的兵棋环境中往往面临效率低、收敛速度慢的挑战。受人类指挥官决策方式的启发,本文提出了一种新的分层决策与战术知识完全集成方法(HDMTK)。该方法包括上部强化学习模块和下部多智能体强化学习模块。为了使智能体能够有效地学习合作策略,在HDMTK中,我们将整个任务分解为可解释的子任务,并根据战术知识设计相应的子目标来形成在线奖励。在“妙算”兵棋模拟平台上的实验结果表明,与先进的MARL方法相比,HDMTK在复杂场景下表现出更优越的性能和更快的收敛速度。
{"title":"HDMTK: Full Integration of Hierarchical Decision-Making and Tactical Knowledge in Multiagent Adversarial Games","authors":"Wei Li;Boling Hu;Aiguo Song;Kaizhu Huang","doi":"10.1109/TCDS.2024.3470068","DOIUrl":"https://doi.org/10.1109/TCDS.2024.3470068","url":null,"abstract":"In the field of adversarial games, existing decision-making algorithms primarily rely on reinforcement learning, which can theoretically adapt to diverse scenarios through trial and error. However, these algorithms often face the challenges of low effectiveness and slow convergence in complex wargame environments. Inspired by how human commanders make decisions, this article proposes a novel method named full integration of hierarchical decision-making and tactical knowledge (HDMTK). This method comprises an upper reinforcement learning module and a lower multiagent reinforcement learning (MARL) module. To enable agents to efficiently learn the cooperative strategy, in HDMTK, we separate the whole task into explainable subtasks and devise their corresponding subgoals for shaping the online rewards based on tactical knowledge. Experimental results on the wargame simulation platform “MiaoSuan” show that, compared to the advanced MARL methods, HDMTK exhibits superior performance and faster convergence in the complex scenarios.","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"17 3","pages":"465-479"},"PeriodicalIF":5.0,"publicationDate":"2024-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144213528","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Graph-Laplacian-Processing-Based Multimodal Localization Backend for Robots and Autonomous Systems 基于图拉普拉斯处理的机器人和自主系统多模态定位后端
IF 5 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-26 DOI: 10.1109/TCDS.2024.3468712
Nikos Piperigkos;Christos Anagnostopoulos;Aris S. Lalos;Petros Kapsalas;Duong Van Nguyen
Simultaneous localization and mapping (SLAM) for positioning of robots and autonomous systems (RASs) and mapping of their surrounding environments is a task of major significance in various applications. However, the main disadvantage of traditional SLAM is that the deployed backend modules suffer from accumulative error caused by sharp viewpoint changes, diverse weather conditions, etc. As such, to improve the localization accuracy of the moving agents, we propose a cost-effective and loosely coupled relocalization backend, deployed on top of original SLAM algorithms, which exploits the topologies of poses and landmarks generated either by camera, LiDAR, or mechanical sensors, to couple and fuse them. This novel fusion scheme enhances the decision-making ability and adaptability of autonomous systems, akin to human cognition, by elaborating graph Laplacian processing concept with Kalman filters. Initially designed for cooperative localization of active road users, this approach optimally combines multisensor information through graph signal processing and Bayesian estimation for self-positioning. Conducted experiments were focused on evaluating how our approach can improve the positioning of autonomous ground vehicles, as prominent examples of RASs equipped with sensing capabilities, in challenging outdoor environments. More specifically, experiments were carried out using the CARLA simulator to generate different types of driving trajectories and environmental conditions, as well as real automotive data captured by an operating vehicle in Langen, Germany. Evaluation study demonstrates that localization accuracy is greatly improved both in terms of overall trajectory error as well as loop closing accuracy for each sensor fusion configuration.
同时定位和测绘(SLAM)是机器人和自主系统(ras)定位及其周围环境测绘的一项重要任务,在各种应用中具有重要意义。然而,传统SLAM的主要缺点是所部署的后端模块存在视点急剧变化、多变天气条件等导致的累积误差。因此,为了提高移动代理的定位精度,我们提出了一种经济高效且松散耦合的重新定位后端,部署在原始SLAM算法之上,该算法利用相机、激光雷达或机械传感器生成的姿势和地标拓扑来耦合和融合它们。该融合方案通过利用卡尔曼滤波器阐述图拉普拉斯处理概念,增强了自治系统类似于人类认知的决策能力和自适应能力。该方法最初是为主动道路使用者的协同定位而设计的,通过图信号处理和贝叶斯估计将多传感器信息最佳地结合在一起进行自定位。所进行的实验重点是评估我们的方法如何改善自动地面车辆的定位,作为配备传感能力的自动驾驶车辆在具有挑战性的室外环境中的突出例子。更具体地说,使用CARLA模拟器进行实验,以生成不同类型的驾驶轨迹和环境条件,以及德国兰根一辆正在行驶的车辆捕获的真实汽车数据。评估研究表明,无论从总体轨迹误差还是闭环精度方面,每种传感器融合配置都大大提高了定位精度。
{"title":"Graph-Laplacian-Processing-Based Multimodal Localization Backend for Robots and Autonomous Systems","authors":"Nikos Piperigkos;Christos Anagnostopoulos;Aris S. Lalos;Petros Kapsalas;Duong Van Nguyen","doi":"10.1109/TCDS.2024.3468712","DOIUrl":"https://doi.org/10.1109/TCDS.2024.3468712","url":null,"abstract":"Simultaneous localization and mapping (SLAM) for positioning of robots and autonomous systems (RASs) and mapping of their surrounding environments is a task of major significance in various applications. However, the main disadvantage of traditional SLAM is that the deployed backend modules suffer from accumulative error caused by sharp viewpoint changes, diverse weather conditions, etc. As such, to improve the localization accuracy of the moving agents, we propose a cost-effective and loosely coupled relocalization backend, deployed on top of original SLAM algorithms, which exploits the topologies of poses and landmarks generated either by camera, LiDAR, or mechanical sensors, to couple and fuse them. This novel fusion scheme enhances the decision-making ability and adaptability of autonomous systems, akin to human cognition, by elaborating graph Laplacian processing concept with Kalman filters. Initially designed for cooperative localization of active road users, this approach optimally combines multisensor information through graph signal processing and Bayesian estimation for self-positioning. Conducted experiments were focused on evaluating how our approach can improve the positioning of autonomous ground vehicles, as prominent examples of RASs equipped with sensing capabilities, in challenging outdoor environments. More specifically, experiments were carried out using the CARLA simulator to generate different types of driving trajectories and environmental conditions, as well as real automotive data captured by an operating vehicle in Langen, Germany. Evaluation study demonstrates that localization accuracy is greatly improved both in terms of overall trajectory error as well as loop closing accuracy for each sensor fusion configuration.","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"17 2","pages":"436-453"},"PeriodicalIF":5.0,"publicationDate":"2024-09-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143761405","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Touch Gesture Recognition-Based Physical Human–Robot Interaction for Collaborative Tasks 基于触摸手势识别的人机协作任务物理交互
IF 5 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-24 DOI: 10.1109/TCDS.2024.3466553
Dawoon Jung;Chengyan Gu;Junmin Park;Joono Cheong
Human–robot collaboration (HRC) has recently attracted increasing attention as a vital component of next-generation automated manufacturing and assembly tasks, yet physical human–robot interaction (pHRI)—which is an inevitable component of collaboration—is often limited to rudimentary touches. This article therefore proposes a deep-learning-based pHRI method that utilizes predefined types of human touch gestures as intuitive communicative signs for collaborative tasks. To this end, a touch gesture network model is first designed upon the framework of the gated recurrent unit (GRU) network, which accepts a set of ground-truth dynamic responses (energy change, generalized momentum, and external joint torque) of robot manipulators under the action of known types of touch gestures and learns to predict the five representative touch gesture types and the corresponding link toward a random touch gesture input. After training the GRU-based touch gesture model using a collected dataset of dynamic responses of a robot manipulator, a total of 35 outputs (five gesture types with seven links each) is recognized with 96.94% accuracy. The experimental results of recognition accuracy correlated with the touch gesture types, and their strength results are shown to validate the performance and disclose the characteristics of the proposed touch gesture model. An example of an IKEA chair assembly task is also presented to demonstrate a collaborative task using the proposed touch gestures. By developing the proposed pHRI method and demonstrating its applicability, we expect that this method can help position physical interaction as one of the key modalities for communication in real-world HRC applications.
作为下一代自动化制造和装配任务的重要组成部分,人机协作(HRC)最近引起了越来越多的关注,然而物理人机交互(pHRI)——这是协作的一个不可避免的组成部分——往往仅限于基本的接触。因此,本文提出了一种基于深度学习的pHRI方法,该方法利用预定义类型的人类触摸手势作为协作任务的直观交流符号。为此,首先在门控递归单元(GRU)网络框架上设计了一个触摸手势网络模型,该模型接受机器人在已知类型的触摸手势作用下的一组真实动态响应(能量变化、广义动量和外部关节扭矩),并学习预测五种具有代表性的触摸手势类型及其与随机触摸手势输入的对应链接。利用收集到的机器人动态响应数据集对基于gru的触摸手势模型进行训练后,共识别出35个输出(5种手势类型,每种手势7个链接),准确率为96.94%。实验结果表明,识别精度与触摸手势类型及其强度相关,验证了所提触摸手势模型的性能,揭示了所提触摸手势模型的特点。还以宜家椅子组装任务为例,演示了使用提议的触摸手势的协作任务。通过开发所提出的pHRI方法并证明其适用性,我们期望该方法可以帮助将物理交互定位为现实世界HRC应用中通信的关键模式之一。
{"title":"Touch Gesture Recognition-Based Physical Human–Robot Interaction for Collaborative Tasks","authors":"Dawoon Jung;Chengyan Gu;Junmin Park;Joono Cheong","doi":"10.1109/TCDS.2024.3466553","DOIUrl":"https://doi.org/10.1109/TCDS.2024.3466553","url":null,"abstract":"Human–robot collaboration (HRC) has recently attracted increasing attention as a vital component of next-generation automated manufacturing and assembly tasks, yet physical human–robot interaction (pHRI)—which is an inevitable component of collaboration—is often limited to rudimentary touches. This article therefore proposes a deep-learning-based pHRI method that utilizes predefined types of human touch gestures as intuitive communicative signs for collaborative tasks. To this end, a touch gesture network model is first designed upon the framework of the gated recurrent unit (GRU) network, which accepts a set of ground-truth dynamic responses (energy change, generalized momentum, and external joint torque) of robot manipulators under the action of known types of touch gestures and learns to predict the five representative touch gesture types and the corresponding link toward a random touch gesture input. After training the GRU-based touch gesture model using a collected dataset of dynamic responses of a robot manipulator, a total of 35 outputs (five gesture types with seven links each) is recognized with 96.94% accuracy. The experimental results of recognition accuracy correlated with the touch gesture types, and their strength results are shown to validate the performance and disclose the characteristics of the proposed touch gesture model. An example of an IKEA chair assembly task is also presented to demonstrate a collaborative task using the proposed touch gestures. By developing the proposed pHRI method and demonstrating its applicability, we expect that this method can help position physical interaction as one of the key modalities for communication in real-world HRC applications.","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"17 2","pages":"421-435"},"PeriodicalIF":5.0,"publicationDate":"2024-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143761496","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing Dimensional Image Emotion Detection With a Low-Resource Dataset via Two-Stage Training 基于两阶段训练的低资源数据集增强多维图像情感检测
IF 5 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-23 DOI: 10.1109/TCDS.2024.3465602
SangEun Lee;Seoyun Kim;Yubeen Lee;Jufeng Yang;Eunil Park
Image emotion analysis has gained notable attention owing to the growing importance of computationally modeling human emotions. Most previous studies have focused on classifying the feelings evoked by an image into predefined emotion categories. Compared with these categorical approaches which cannot address the ambiguity and complexity of human emotions, recent studies have taken dimensional approaches to address these problems. However, there is still a limitation in that the number of dimensional datasets is significantly smaller for model training, compared with many available categorical datasets. We propose four types of frameworks that use categorical datasets to predict emotion values for a given image in the valence–arousal (VA) space. Specifically, our proposed framework is trained to predict continuous emotion values under the supervision of categorical labels. Extensive experiments demonstrate that our approach showed a positive correlation with the actual VA values of the dimensional dataset. In addition, our framework improves further when a small number of dimensional datasets are available for the fine-tuning process.
由于人类情感的计算建模越来越重要,图像情感分析得到了显著的关注。大多数先前的研究都集中在将图像所唤起的感觉分类为预定义的情感类别。与这些分类方法无法解决人类情感的模糊性和复杂性相比,最近的研究采用了维度方法来解决这些问题。然而,与许多可用的分类数据集相比,用于模型训练的维度数据集的数量明显更少,这仍然是一个限制。我们提出了四种类型的框架,它们使用分类数据集来预测给定图像在价-觉醒(VA)空间中的情感值。具体来说,我们提出的框架被训练成在分类标签的监督下预测连续的情感值。大量的实验表明,我们的方法与维度数据集的实际VA值呈正相关。此外,当少量维度数据集可用于微调过程时,我们的框架会进一步改进。
{"title":"Enhancing Dimensional Image Emotion Detection With a Low-Resource Dataset via Two-Stage Training","authors":"SangEun Lee;Seoyun Kim;Yubeen Lee;Jufeng Yang;Eunil Park","doi":"10.1109/TCDS.2024.3465602","DOIUrl":"https://doi.org/10.1109/TCDS.2024.3465602","url":null,"abstract":"Image emotion analysis has gained notable attention owing to the growing importance of computationally modeling human emotions. Most previous studies have focused on classifying the feelings evoked by an image into predefined emotion categories. Compared with these categorical approaches which cannot address the ambiguity and complexity of human emotions, recent studies have taken dimensional approaches to address these problems. However, there is still a limitation in that the number of dimensional datasets is significantly smaller for model training, compared with many available categorical datasets. We propose four types of frameworks that use categorical datasets to predict emotion values for a given image in the valence–arousal (VA) space. Specifically, our proposed framework is trained to predict continuous emotion values under the supervision of categorical labels. Extensive experiments demonstrate that our approach showed a positive correlation with the actual VA values of the dimensional dataset. In addition, our framework improves further when a small number of dimensional datasets are available for the fine-tuning process.","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"17 3","pages":"455-464"},"PeriodicalIF":5.0,"publicationDate":"2024-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144213666","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Embodied Perception Interaction, and Cognition for Wearable Robotics: A Survey 可穿戴机器人的嵌入式感知、交互和认知:调查
IF 5 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-18 DOI: 10.1109/tcds.2024.3463194
Xiaoyu Wu, Jiale Liang, Yiang Yu, Guoxin Li, Gary G. Yen, Haoyong Yu
{"title":"Embodied Perception Interaction, and Cognition for Wearable Robotics: A Survey","authors":"Xiaoyu Wu, Jiale Liang, Yiang Yu, Guoxin Li, Gary G. Yen, Haoyong Yu","doi":"10.1109/tcds.2024.3463194","DOIUrl":"https://doi.org/10.1109/tcds.2024.3463194","url":null,"abstract":"","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"48 1","pages":""},"PeriodicalIF":5.0,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142269922","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CS-SLAM: A lightweight semantic SLAM method for dynamic scenarios CS-SLAM:适用于动态场景的轻量级语义 SLAM 方法
IF 5 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-17 DOI: 10.1109/tcds.2024.3462651
Zhendong Guo, Na Dong, Zehui Zhang, Xiaoming Mai, Donghui Li
{"title":"CS-SLAM: A lightweight semantic SLAM method for dynamic scenarios","authors":"Zhendong Guo, Na Dong, Zehui Zhang, Xiaoming Mai, Donghui Li","doi":"10.1109/tcds.2024.3462651","DOIUrl":"https://doi.org/10.1109/tcds.2024.3462651","url":null,"abstract":"","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"49 1","pages":""},"PeriodicalIF":5.0,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142266493","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Edge-Centric Functional-Connectivity-Based Cofluctuation-Guided Subcortical Connectivity Network Construction 以边缘为中心、基于共波动引导的皮层下功能连接网络构建
IF 5 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-17 DOI: 10.1109/TCDS.2024.3462709
Qinrui Ling;Aiping Liu;Taomian Mi;Piu Chan;Xun Chen
Subcortical regions can be functionally organized into connectivity networks and are extensively communicated with the cortex via reciprocal connections. However, most current research on subcortical networks ignores these interconnections, and networks of the whole brain are of high dimensionality and computational complexity. In this article, we propose a novel cofluctuation-guided subcortical connectivity network construction model based on edge-centric functional connectivity (FC). It is capable of extracting the cofluctuations between the cortex and subcortex and constructing dynamic subcortical networks based on these interconnections. Blind source separation approaches with domain knowledge are designed for dimensionality reduction and feature extraction. Great reproducibility and reliability were achieved when applying our model to two sessions of functional magnetic resonance imaging (fMRI) data. Cortical areas having synchronous communications with the cortex were detected, which was unable to be revealed by traditional node-centric FC. Significant alterations in connectivity patterns were observed when dealing with fMRI of subjects with and without Parkinson's disease, which were further correlated to clinical scores. These validations demonstrated that our model provided a promising strategy for brain network construction, exhibiting great potential in clinical practice.
皮层下区域可以在功能上组织成连接网络,并通过相互连接与皮层广泛沟通。然而,目前大多数关于皮层下网络的研究忽略了这些相互联系,而且整个大脑的网络具有高维性和计算复杂性。在本文中,我们提出了一种基于边缘中心功能连接(FC)的共同波动引导下皮层下连接网络构建模型。它能够提取皮层和下皮层之间的共同波动,并基于这些相互联系构建动态的皮层下网络。设计了基于领域知识的盲源分离方法,用于降维和特征提取。将我们的模型应用于两次功能磁共振成像(fMRI)数据时,获得了很高的再现性和可靠性。检测到与皮层有同步通信的皮质区域,这是传统的以节点为中心的FC无法显示的。在处理帕金森病患者和非帕金森病患者的fMRI时,观察到连接模式的显著改变,这与临床评分进一步相关。这些验证表明,我们的模型为脑网络构建提供了一种很有前景的策略,在临床实践中具有很大的潜力。
{"title":"Edge-Centric Functional-Connectivity-Based Cofluctuation-Guided Subcortical Connectivity Network Construction","authors":"Qinrui Ling;Aiping Liu;Taomian Mi;Piu Chan;Xun Chen","doi":"10.1109/TCDS.2024.3462709","DOIUrl":"10.1109/TCDS.2024.3462709","url":null,"abstract":"Subcortical regions can be functionally organized into connectivity networks and are extensively communicated with the cortex via reciprocal connections. However, most current research on subcortical networks ignores these interconnections, and networks of the whole brain are of high dimensionality and computational complexity. In this article, we propose a novel cofluctuation-guided subcortical connectivity network construction model based on edge-centric functional connectivity (FC). It is capable of extracting the cofluctuations between the cortex and subcortex and constructing dynamic subcortical networks based on these interconnections. Blind source separation approaches with domain knowledge are designed for dimensionality reduction and feature extraction. Great reproducibility and reliability were achieved when applying our model to two sessions of functional magnetic resonance imaging (fMRI) data. Cortical areas having synchronous communications with the cortex were detected, which was unable to be revealed by traditional node-centric FC. Significant alterations in connectivity patterns were observed when dealing with fMRI of subjects with and without Parkinson's disease, which were further correlated to clinical scores. These validations demonstrated that our model provided a promising strategy for brain network construction, exhibiting great potential in clinical practice.","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"17 2","pages":"390-399"},"PeriodicalIF":5.0,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142269531","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Unveiling Thoughts: A Review of Advancements in EEG Brain Signal Decoding Into Text 揭开思想的面纱:脑电图脑信号解码为文本的进展回顾
IF 5 3区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-09-17 DOI: 10.1109/TCDS.2024.3462452
Saydul Akbar Murad;Nick Rahimi
The conversion of brain activity into text using electroencephalography (EEG) has gained significant traction in recent years. Many researchers are working to develop new models to decode EEG signals into text form. Although this area has shown promising developments, it still faces numerous challenges that necessitate further improvement. It is important to outline this area's recent developments and future research directions to provide a comprehensive understanding of the current state of technology, guide future research efforts, and enhance the effectiveness and accessibility of EEG-to-text systems. In this review article, we thoroughly summarize the progress in EEG-to-text conversion. First, we talk about how EEG-to-text technology has grown and what problems the field still faces. Second, we discuss existing techniques used in this field. This includes methods for collecting EEG data, the steps to process these signals, and the development of systems capable of translating these signals into coherent text. We conclude with potential future research directions, emphasizing the need for enhanced accuracy, reduced system constraints, and the exploration of novel applications across varied sectors. By addressing these aspects, this review aims to contribute to developing more accessible and effective brain–computer interface (BCI) technology for a broader user base.
近年来,利用脑电图(EEG)将大脑活动转化为文本的研究得到了极大的关注。许多研究人员正在努力开发新的模型,将脑电图信号解码成文本形式。虽然这一领域已显示出有希望的发展,但仍面临许多需要进一步改进的挑战。重要的是概述该领域的最新发展和未来的研究方向,以提供对当前技术状态的全面了解,指导未来的研究工作,并提高脑电图到文本系统的有效性和可访问性。在这篇综述文章中,我们全面总结了脑电图到文本转换的进展。首先,我们讨论脑电图文本转换技术是如何发展的,以及该领域仍然面临哪些问题。其次,我们讨论了该领域中使用的现有技术。这包括收集脑电图数据的方法,处理这些信号的步骤,以及能够将这些信号翻译成连贯文本的系统的开发。我们总结了潜在的未来研究方向,强调需要提高准确性,减少系统约束,以及探索不同领域的新应用。通过对这些方面的探讨,本文旨在为更广泛的用户基础开发更易于访问和有效的脑机接口(BCI)技术做出贡献。
{"title":"Unveiling Thoughts: A Review of Advancements in EEG Brain Signal Decoding Into Text","authors":"Saydul Akbar Murad;Nick Rahimi","doi":"10.1109/TCDS.2024.3462452","DOIUrl":"10.1109/TCDS.2024.3462452","url":null,"abstract":"The conversion of brain activity into text using electroencephalography (EEG) has gained significant traction in recent years. Many researchers are working to develop new models to decode EEG signals into text form. Although this area has shown promising developments, it still faces numerous challenges that necessitate further improvement. It is important to outline this area's recent developments and future research directions to provide a comprehensive understanding of the current state of technology, guide future research efforts, and enhance the effectiveness and accessibility of EEG-to-text systems. In this review article, we thoroughly summarize the progress in EEG-to-text conversion. First, we talk about how EEG-to-text technology has grown and what problems the field still faces. Second, we discuss existing techniques used in this field. This includes methods for collecting EEG data, the steps to process these signals, and the development of systems capable of translating these signals into coherent text. We conclude with potential future research directions, emphasizing the need for enhanced accuracy, reduced system constraints, and the exploration of novel applications across varied sectors. By addressing these aspects, this review aims to contribute to developing more accessible and effective brain–computer interface (BCI) technology for a broader user base.","PeriodicalId":54300,"journal":{"name":"IEEE Transactions on Cognitive and Developmental Systems","volume":"17 1","pages":"61-76"},"PeriodicalIF":5.0,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142266494","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Transactions on Cognitive and Developmental Systems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1