首页 > 最新文献

Frontiers in Neurorobotics最新文献

英文 中文
Mining local and global spatiotemporal features for tactile object recognition 挖掘局部和全局时空特征,实现触觉物体识别
IF 3.1 4区 计算机科学 Q2 Computer Science Pub Date : 2024-05-03 DOI: 10.3389/fnbot.2024.1387428
Xiaoliang Qian, Wei Deng, Wei Wang, Yucui Liu, Liying Jiang
The tactile object recognition (TOR) is highly important for environmental perception of robots. The previous works usually utilize single scale convolution which cannot simultaneously extract local and global spatiotemporal features of tactile data, which leads to low accuracy in TOR task. To address above problem, this article proposes a local and global residual (LGR-18) network which is mainly consisted of multiple local and global convolution (LGC) blocks. An LGC block contains two pairs of local convolution (LC) and global convolution (GC) modules. The LC module mainly utilizes a temporal shift operation and a 2D convolution layer to extract local spatiotemporal features. The GC module extracts global spatiotemporal features by fusing multiple 1D and 2D convolutions which can expand the receptive field in temporal and spatial dimensions. Consequently, our LGR-18 network can extract local-global spatiotemporal features without using 3D convolutions which usually require a large number of parameters. The effectiveness of LC module, GC module and LGC block is verified by ablation studies. Quantitative comparisons with state-of-the-art methods reveal the excellent capability of our method.
触觉物体识别(TOR)对于机器人的环境感知非常重要。以往的研究通常采用单尺度卷积,无法同时提取触觉数据的局部和全局时空特征,导致 TOR 任务的准确率较低。针对上述问题,本文提出了一种局部和全局残差(LGR-18)网络,它主要由多个局部和全局卷积(LGC)块组成。一个 LGC 块包含两对局部卷积(LC)和全局卷积(GC)模块。LC 模块主要利用时移操作和二维卷积层来提取局部时空特征。全局卷积模块通过融合多个一维和二维卷积来提取全局时空特征,从而在时间和空间维度上扩展感受野。因此,我们的 LGR-18 网络可以提取局部-全局时空特征,而无需使用通常需要大量参数的三维卷积。消融研究验证了 LC 模块、GC 模块和 LGC 模块的有效性。与最先进方法的定量比较显示了我们方法的卓越能力。
{"title":"Mining local and global spatiotemporal features for tactile object recognition","authors":"Xiaoliang Qian, Wei Deng, Wei Wang, Yucui Liu, Liying Jiang","doi":"10.3389/fnbot.2024.1387428","DOIUrl":"https://doi.org/10.3389/fnbot.2024.1387428","url":null,"abstract":"The tactile object recognition (TOR) is highly important for environmental perception of robots. The previous works usually utilize single scale convolution which cannot simultaneously extract local and global spatiotemporal features of tactile data, which leads to low accuracy in TOR task. To address above problem, this article proposes a local and global residual (LGR-18) network which is mainly consisted of multiple local and global convolution (LGC) blocks. An LGC block contains two pairs of local convolution (LC) and global convolution (GC) modules. The LC module mainly utilizes a temporal shift operation and a 2D convolution layer to extract local spatiotemporal features. The GC module extracts global spatiotemporal features by fusing multiple 1D and 2D convolutions which can expand the receptive field in temporal and spatial dimensions. Consequently, our LGR-18 network can extract local-global spatiotemporal features without using 3D convolutions which usually require a large number of parameters. The effectiveness of LC module, GC module and LGC block is verified by ablation studies. Quantitative comparisons with state-of-the-art methods reveal the excellent capability of our method.","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-05-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140841107","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
3D hand pose and mesh estimation via a generic Topology-aware Transformer model 通过通用拓扑感知变换器模型进行三维手部姿态和网格估算
IF 3.1 4区 计算机科学 Q2 Computer Science Pub Date : 2024-05-03 DOI: 10.3389/fnbot.2024.1395652
Shaoqi Yu, Yintong Wang, Lili Chen, Xiaolin Zhang, Jiamao Li
In Human-Robot Interaction (HRI), accurate 3D hand pose and mesh estimation hold critical importance. However, inferring reasonable and accurate poses in severe self-occlusion and high self-similarity remains an inherent challenge. In order to alleviate the ambiguity caused by invisible and similar joints during HRI, we propose a new Topology-aware Transformer network named HandGCNFormer with depth image as input, incorporating prior knowledge of hand kinematic topology into the network while modeling long-range contextual information. Specifically, we propose a novel Graphformer decoder with an additional Node-offset Graph Convolutional layer (NoffGConv). The Graphformer decoder optimizes the synergy between the Transformer and GCN, capturing long-range dependencies and local topological connections between joints. On top of that, we replace the standard MLP prediction head with a novel Topology-aware head to better exploit local topological constraints for more reasonable and accurate poses. Our method achieves state-of-the-art 3D hand pose estimation performance on four challenging datasets, including Hands2017, NYU, ICVL, and MSRA. To further demonstrate the effectiveness and scalability of our proposed Graphformer Decoder and Topology aware head, we extend our framework to HandGCNFormer-Mesh for the 3D hand mesh estimation task. The extended framework efficiently integrates a shape regressor with the original Graphformer Decoder and Topology aware head, producing Mano parameters. The results on the HO-3D dataset, which contains various and challenging occlusions, show that our HandGCNFormer-Mesh achieves competitive results compared to previous state-of-the-art 3D hand mesh estimation methods.
在人机交互(HRI)中,精确的三维手部姿势和网格估计至关重要。然而,在严重自闭和高度自相似的情况下推断合理准确的姿势仍然是一个固有的挑战。为了减轻 HRI 过程中因关节不可见和相似而造成的模糊性,我们提出了一种名为 HandGCNFormer 的新拓扑感知变换器网络,以深度图像为输入,将手部运动拓扑的先验知识纳入网络,同时对远距离上下文信息进行建模。具体来说,我们提出了一种带有额外节点偏移图卷积层(NoffGConv)的新型 Graphformer 解码器。Graphformer 解码器优化了 Transformer 和 GCN 之间的协同作用,捕捉了关节之间的长距离依赖关系和局部拓扑连接。在此基础上,我们用新颖的拓扑感知头取代了标准的 MLP 预测头,从而更好地利用局部拓扑约束来获得更合理、更准确的姿势。我们的方法在四个具有挑战性的数据集(包括 Hands2017、NYU、ICVL 和 MSRA)上实现了最先进的 3D 手部姿态估计性能。为了进一步证明我们提出的 Graphformer 解码器和拓扑感知头的有效性和可扩展性,我们将框架扩展为 HandGCNFormer-Mesh,用于三维手部网格估计任务。扩展框架有效地将形状回归器与原始 Graphformer 解码器和拓扑感知头集成在一起,生成了马诺参数。HO-3D 数据集包含各种具有挑战性的遮挡物,在该数据集上的结果表明,与之前最先进的三维手部网格估计方法相比,我们的 HandGCNFormer-Mesh 取得了具有竞争力的结果。
{"title":"3D hand pose and mesh estimation via a generic Topology-aware Transformer model","authors":"Shaoqi Yu, Yintong Wang, Lili Chen, Xiaolin Zhang, Jiamao Li","doi":"10.3389/fnbot.2024.1395652","DOIUrl":"https://doi.org/10.3389/fnbot.2024.1395652","url":null,"abstract":"In Human-Robot Interaction (HRI), accurate 3D hand pose and mesh estimation hold critical importance. However, inferring reasonable and accurate poses in severe self-occlusion and high self-similarity remains an inherent challenge. In order to alleviate the ambiguity caused by invisible and similar joints during HRI, we propose a new Topology-aware Transformer network named HandGCNFormer with depth image as input, incorporating prior knowledge of hand kinematic topology into the network while modeling long-range contextual information. Specifically, we propose a novel Graphformer decoder with an additional Node-offset Graph Convolutional layer (NoffGConv). The Graphformer decoder optimizes the synergy between the Transformer and GCN, capturing long-range dependencies and local topological connections between joints. On top of that, we replace the standard MLP prediction head with a novel Topology-aware head to better exploit local topological constraints for more reasonable and accurate poses. Our method achieves state-of-the-art 3D hand pose estimation performance on four challenging datasets, including Hands2017, NYU, ICVL, and MSRA. To further demonstrate the effectiveness and scalability of our proposed Graphformer Decoder and Topology aware head, we extend our framework to HandGCNFormer-Mesh for the 3D hand mesh estimation task. The extended framework efficiently integrates a shape regressor with the original Graphformer Decoder and Topology aware head, producing Mano parameters. The results on the HO-3D dataset, which contains various and challenging occlusions, show that our HandGCNFormer-Mesh achieves competitive results compared to previous state-of-the-art 3D hand mesh estimation methods.","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-05-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140841369","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Curiosity model policy optimization for robotic manipulator tracking control with input saturation in uncertain environment 针对不确定环境下输入饱和的机器人机械手跟踪控制的好奇心模型策略优化
IF 3.1 4区 计算机科学 Q2 Computer Science Pub Date : 2024-05-01 DOI: 10.3389/fnbot.2024.1376215
Tu Wang, Fujie Wang, Zhongye Xie, Feiyan Qin
In uncertain environments with robot input saturation, both model-based reinforcement learning (MBRL) and traditional controllers struggle to perform control tasks optimally. In this study, an algorithmic framework of Curiosity Model Policy Optimization (CMPO) is proposed by combining curiosity and model-based approach, where tracking errors are reduced via training agents on control gains for traditional model-free controllers. To begin with, a metric for judging positive and negative curiosity is proposed. Constrained optimization is employed to update the curiosity ratio, which improves the efficiency of agent training. Next, the novelty distance buffer ratio is defined to reduce bias between the environment and the model. Finally, CMPO is simulated with traditional controllers and baseline MBRL algorithms in the robotic environment designed with non-linear rewards. The experimental results illustrate that the algorithm achieves superior tracking performance and generalization capabilities.
在机器人输入饱和的不确定环境中,基于模型的强化学习(MBRL)和传统控制器都难以以最佳方式执行控制任务。本研究结合好奇心和基于模型的方法,提出了好奇心模型策略优化(CMPO)的算法框架,通过训练传统无模型控制器控制增益上的代理来减少跟踪误差。首先,提出了判断正负好奇心的指标。利用约束优化来更新好奇心比率,从而提高了代理训练的效率。接着,定义了新奇距离缓冲比,以减少环境与模型之间的偏差。最后,在非线性奖励设计的机器人环境中,将 CMPO 与传统控制器和基准 MBRL 算法进行了仿真。实验结果表明,该算法实现了卓越的跟踪性能和泛化能力。
{"title":"Curiosity model policy optimization for robotic manipulator tracking control with input saturation in uncertain environment","authors":"Tu Wang, Fujie Wang, Zhongye Xie, Feiyan Qin","doi":"10.3389/fnbot.2024.1376215","DOIUrl":"https://doi.org/10.3389/fnbot.2024.1376215","url":null,"abstract":"In uncertain environments with robot input saturation, both model-based reinforcement learning (MBRL) and traditional controllers struggle to perform control tasks optimally. In this study, an algorithmic framework of Curiosity Model Policy Optimization (CMPO) is proposed by combining curiosity and model-based approach, where tracking errors are reduced via training agents on control gains for traditional model-free controllers. To begin with, a metric for judging positive and negative curiosity is proposed. Constrained optimization is employed to update the curiosity ratio, which improves the efficiency of agent training. Next, the novelty distance buffer ratio is defined to reduce bias between the environment and the model. Finally, CMPO is simulated with traditional controllers and baseline MBRL algorithms in the robotic environment designed with non-linear rewards. The experimental results illustrate that the algorithm achieves superior tracking performance and generalization capabilities.","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140841364","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improved object detection method for unmanned driving based on Transformers 基于变压器的改进型无人驾驶物体检测方法
IF 3.1 4区 计算机科学 Q2 Computer Science Pub Date : 2024-05-01 DOI: 10.3389/fnbot.2024.1342126
Huaqi Zhao, Xiang Peng, Su Wang, Jun-Bao Li, Jeng-Shyang Pan, Xiaoguang Su, Xiaomin Liu
The object detection method serves as the core technology within the unmanned driving perception module, extensively employed for detecting vehicles, pedestrians, traffic signs, and various objects. However, existing object detection methods still encounter three challenges in intricate unmanned driving scenarios: unsatisfactory performance in multi-scale object detection, inadequate accuracy in detecting small objects, and occurrences of false positives and missed detections in densely occluded environments. Therefore, this study proposes an improved object detection method for unmanned driving, leveraging Transformer architecture to address these challenges. First, a multi-scale Transformer feature extraction method integrated with channel attention is used to enhance the network's capability in extracting features across different scales. Second, a training method incorporating Query Denoising with Gaussian decay was employed to enhance the network's proficiency in learning representations of small objects. Third, a hybrid matching method combining Optimal Transport and Hungarian algorithms was used to facilitate the matching process between predicted and actual values, thereby enriching the network with more informative positive sample features. Experimental evaluations conducted on datasets including KITTI demonstrate that the proposed method achieves 3% higher mean Average Precision (mAP) than that of the existing methodologies.
物体检测方法是无人驾驶感知模块的核心技术,广泛用于检测车辆、行人、交通标志和各种物体。然而,在错综复杂的无人驾驶场景中,现有的物体检测方法仍然面临三个挑战:多尺度物体检测性能不理想、小物体检测精度不够、在密集遮挡环境中出现误报和漏报。因此,本研究提出了一种改进的无人驾驶物体检测方法,利用 Transformer 架构来应对这些挑战。首先,多尺度 Transformer 特征提取方法与通道注意相结合,增强了网络在不同尺度上提取特征的能力。其次,采用了一种结合查询去噪和高斯衰减的训练方法,以提高网络学习小物体表征的能力。第三,采用了一种结合了最优传输和匈牙利算法的混合匹配方法,以促进预测值和实际值之间的匹配过程,从而为网络提供更多信息丰富的正样本特征。在包括 KITTI 在内的数据集上进行的实验评估表明,所提出的方法比现有方法的平均精度(mAP)高出 3%。
{"title":"Improved object detection method for unmanned driving based on Transformers","authors":"Huaqi Zhao, Xiang Peng, Su Wang, Jun-Bao Li, Jeng-Shyang Pan, Xiaoguang Su, Xiaomin Liu","doi":"10.3389/fnbot.2024.1342126","DOIUrl":"https://doi.org/10.3389/fnbot.2024.1342126","url":null,"abstract":"The object detection method serves as the core technology within the unmanned driving perception module, extensively employed for detecting vehicles, pedestrians, traffic signs, and various objects. However, existing object detection methods still encounter three challenges in intricate unmanned driving scenarios: unsatisfactory performance in multi-scale object detection, inadequate accuracy in detecting small objects, and occurrences of false positives and missed detections in densely occluded environments. Therefore, this study proposes an improved object detection method for unmanned driving, leveraging Transformer architecture to address these challenges. First, a multi-scale Transformer feature extraction method integrated with channel attention is used to enhance the network's capability in extracting features across different scales. Second, a training method incorporating Query Denoising with Gaussian decay was employed to enhance the network's proficiency in learning representations of small objects. Third, a hybrid matching method combining Optimal Transport and Hungarian algorithms was used to facilitate the matching process between predicted and actual values, thereby enriching the network with more informative positive sample features. Experimental evaluations conducted on datasets including KITTI demonstrate that the proposed method achieves 3% higher mean Average Precision (mAP) than that of the existing methodologies.","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140841101","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The meta-learning method for the ensemble model based on situational meta-task 基于情景元任务的集合模型元学习方法
IF 3.1 4区 计算机科学 Q2 Computer Science Pub Date : 2024-04-26 DOI: 10.3389/fnbot.2024.1391247
Zhengchao Zhang, Lianke Zhou, Yuyang Wu, Nianbin Wang
IntroductionThe meta-learning methods have been widely used to solve the problem of few-shot learning. Generally, meta-learners are trained on a variety of tasks and then generalized to novel tasks.MethodsHowever, existing meta-learning methods do not consider the relationship between meta-tasks and novel tasks during the meta-training period, so that initial models of the meta-learner provide less useful meta-knowledge for the novel tasks. This leads to a weak generalization ability on novel tasks. Meanwhile, different initial models contain different meta-knowledge, which leads to certain differences in the learning effect of novel tasks during the meta-testing period. Therefore, this article puts forward a meta-optimization method based on situational meta-task construction and cooperation of multiple initial models. First, during the meta-training period, a method of constructing situational meta-task is proposed, and the selected candidate task sets provide more effective meta-knowledge for novel tasks. Then, during the meta-testing period, an ensemble model method based on meta-optimization is proposed to minimize the loss of inter-model cooperation in prediction, so that multiple models cooperation can realize the learning of novel tasks.ResultsThe above-mentioned methods are applied to popular few-shot character datasets and image recognition datasets. Furthermore, the experiment results indicate that the proposed method achieves good effects in few-shot classification tasks.DiscussionIn future work, we will extend our methods to provide more generalized and useful meta-knowledge to the model during the meta-training period when the novel few-shot tasks are completely invisible.
引言 元学习方法已被广泛用于解决少量学习问题。然而,现有的元学习方法在元训练期间没有考虑元任务与新任务之间的关系,因此元学习器的初始模型为新任务提供的元知识较少。这导致元学习器对新任务的泛化能力较弱。同时,不同的初始模型包含不同的元知识,这也导致元测试期对新任务的学习效果存在一定差异。因此,本文提出了一种基于情景元任务构建和多个初始模型合作的元优化方法。首先,在元训练期,提出一种构建情境元任务的方法,所选的候选任务集能为新任务提供更有效的元知识。然后,在元测试阶段,提出一种基于元优化的集合模型方法,最大限度地减少模型间合作预测的损失,从而实现多模型合作学习新任务。讨论在未来的工作中,我们将扩展我们的方法,以便在新颖的少量任务完全不可见的元训练期为模型提供更多通用的、有用的元知识。
{"title":"The meta-learning method for the ensemble model based on situational meta-task","authors":"Zhengchao Zhang, Lianke Zhou, Yuyang Wu, Nianbin Wang","doi":"10.3389/fnbot.2024.1391247","DOIUrl":"https://doi.org/10.3389/fnbot.2024.1391247","url":null,"abstract":"IntroductionThe meta-learning methods have been widely used to solve the problem of few-shot learning. Generally, meta-learners are trained on a variety of tasks and then generalized to novel tasks.MethodsHowever, existing meta-learning methods do not consider the relationship between meta-tasks and novel tasks during the meta-training period, so that initial models of the meta-learner provide less useful meta-knowledge for the novel tasks. This leads to a weak generalization ability on novel tasks. Meanwhile, different initial models contain different meta-knowledge, which leads to certain differences in the learning effect of novel tasks during the meta-testing period. Therefore, this article puts forward a meta-optimization method based on situational meta-task construction and cooperation of multiple initial models. First, during the meta-training period, a method of constructing situational meta-task is proposed, and the selected candidate task sets provide more effective meta-knowledge for novel tasks. Then, during the meta-testing period, an ensemble model method based on meta-optimization is proposed to minimize the loss of inter-model cooperation in prediction, so that multiple models cooperation can realize the learning of novel tasks.ResultsThe above-mentioned methods are applied to popular few-shot character datasets and image recognition datasets. Furthermore, the experiment results indicate that the proposed method achieves good effects in few-shot classification tasks.DiscussionIn future work, we will extend our methods to provide more generalized and useful meta-knowledge to the model during the meta-training period when the novel few-shot tasks are completely invisible.","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-04-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140806225","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Ontology based autonomous robot task processing framework 基于本体的自主机器人任务处理框架
IF 3.1 4区 计算机科学 Q2 Computer Science Pub Date : 2024-04-23 DOI: 10.3389/fnbot.2024.1401075
Yueguang Ge, Shaolin Zhang, Yinghao Cai, Tao Lu, Haitao Wang, Xiaolong Hui, Shuo Wang
IntroductionIn recent years, the perceptual capabilities of robots have been significantly enhanced. However, the task execution of the robots still lacks adaptive capabilities in unstructured and dynamic environments.MethodsIn this paper, we propose an ontology based autonomous robot task processing framework (ARTProF), to improve the robot's adaptability within unstructured and dynamic environments. ARTProF unifies ontological knowledge representation, reasoning, and autonomous task planning and execution into a single framework. The interface between the knowledge base and neural network-based object detection is first introduced in ARTProF to improve the robot's perception capabilities. A knowledge-driven manipulation operator based on Robot Operating System (ROS) is then designed to facilitate the interaction between the knowledge base and the robot's primitive actions. Additionally, an operation similarity model is proposed to endow the robot with the ability to generalize to novel objects. Finally, a dynamic task planning algorithm, leveraging ontological knowledge, equips the robot with adaptability to execute tasks in unstructured and dynamic environments.ResultsExperimental results on real-world scenarios and simulations demonstrate the effectiveness and efficiency of the proposed ARTProF framework.DiscussionIn future work, we will focus on refining the ARTProF framework by integrating neurosymbolic inference.
引言 近年来,机器人的感知能力得到了显著提升。本文提出了一种基于本体的机器人自主任务处理框架(ARTProF),以提高机器人在非结构化动态环境中的适应能力。ARTProF 将本体知识表示、推理、自主任务规划和执行统一到一个框架中。ARTProF 首次引入了知识库与基于神经网络的物体检测之间的接口,以提高机器人的感知能力。然后,设计了基于机器人操作系统(ROS)的知识驱动操纵操作器,以促进知识库与机器人基本操作之间的互动。此外,还提出了一种操作相似性模型,以赋予机器人对新物体进行泛化的能力。最后,一种利用本体知识的动态任务规划算法使机器人具备了在非结构化动态环境中执行任务的适应能力。讨论在未来的工作中,我们将重点通过整合神经符号推理来完善 ARTProF 框架。
{"title":"Ontology based autonomous robot task processing framework","authors":"Yueguang Ge, Shaolin Zhang, Yinghao Cai, Tao Lu, Haitao Wang, Xiaolong Hui, Shuo Wang","doi":"10.3389/fnbot.2024.1401075","DOIUrl":"https://doi.org/10.3389/fnbot.2024.1401075","url":null,"abstract":"IntroductionIn recent years, the perceptual capabilities of robots have been significantly enhanced. However, the task execution of the robots still lacks adaptive capabilities in unstructured and dynamic environments.MethodsIn this paper, we propose an ontology based autonomous robot task processing framework (ARTProF), to improve the robot's adaptability within unstructured and dynamic environments. ARTProF unifies ontological knowledge representation, reasoning, and autonomous task planning and execution into a single framework. The interface between the knowledge base and neural network-based object detection is first introduced in ARTProF to improve the robot's perception capabilities. A knowledge-driven manipulation operator based on Robot Operating System (ROS) is then designed to facilitate the interaction between the knowledge base and the robot's primitive actions. Additionally, an operation similarity model is proposed to endow the robot with the ability to generalize to novel objects. Finally, a dynamic task planning algorithm, leveraging ontological knowledge, equips the robot with adaptability to execute tasks in unstructured and dynamic environments.ResultsExperimental results on real-world scenarios and simulations demonstrate the effectiveness and efficiency of the proposed ARTProF framework.DiscussionIn future work, we will focus on refining the ARTProF framework by integrating neurosymbolic inference.","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-04-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140885414","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Residual learning-based robotic image analysis model for low-voltage distributed photovoltaic fault identification and positioning 基于残差学习的机器人图像分析模型,用于低压分布式光伏故障识别和定位
IF 3.1 4区 计算机科学 Q2 Computer Science Pub Date : 2024-04-22 DOI: 10.3389/fnbot.2024.1396979
Xudong Zhang, Yunlong Ge, Yifeng Wang, Jun Wang, Wenhao Wang, Lijun Lu
With the fast development of large-scale Photovoltaic (PV) plants, the automatic PV fault identification and positioning have become an important task for the PV intelligent systems, aiming to guarantee the safety, reliability, and productivity of large-scale PV plants. In this paper, we propose a residual learning-based robotic (UAV) image analysis model for low-voltage distributed PV fault identification and positioning. In our target scenario, the unmanned aerial vehicles (UAVs) are deployed to acquire moving images of low-voltage distributed PV power plants. To get desired robustness and accuracy of PV image detection, we integrate residual learning with attention mechanism into the UAV image analysis model based on you only look once v4 (YOLOv4) network. Then, we design the sophisticated multi-scale spatial pyramid fusion and use it to optimize the YOLOv4 network for the nuanced task of fault localization within PV arrays, where the Complete-IOU loss is incorporated in the predictive modeling phase, significantly enhancing the accuracy and efficiency of fault detection. A series of experimental comparisons in terms of the accuracy of fault positioning are conducted, and the experimental results verify the feasibility and effectiveness of the proposed model in dealing with the safety and reliability maintenance of low-voltage distributed PV systems.
随着大型光伏电站的快速发展,光伏故障自动识别与定位已成为光伏智能系统的一项重要任务,旨在保障大型光伏电站的安全性、可靠性和生产率。本文提出了一种基于残差学习的机器人(无人机)图像分析模型,用于低压分布式光伏故障识别和定位。在我们的目标场景中,无人飞行器(UAV)被部署来获取低压分布式光伏电站的移动图像。为了获得理想的光伏图像检测鲁棒性和准确性,我们在无人机图像分析模型中集成了基于 YOLOv4 网络的残差学习和注意力机制。然后,我们设计了复杂的多尺度空间金字塔融合技术,并将其用于优化 YOLOv4 网络,以完成光伏阵列内部故障定位的细微任务,其中 Complete-IOU 损失被纳入预测建模阶段,从而显著提高了故障检测的准确性和效率。在故障定位精度方面进行了一系列实验比较,实验结果验证了所提模型在处理低压分布式光伏系统安全性和可靠性维护方面的可行性和有效性。
{"title":"Residual learning-based robotic image analysis model for low-voltage distributed photovoltaic fault identification and positioning","authors":"Xudong Zhang, Yunlong Ge, Yifeng Wang, Jun Wang, Wenhao Wang, Lijun Lu","doi":"10.3389/fnbot.2024.1396979","DOIUrl":"https://doi.org/10.3389/fnbot.2024.1396979","url":null,"abstract":"With the fast development of large-scale Photovoltaic (PV) plants, the automatic PV fault identification and positioning have become an important task for the PV intelligent systems, aiming to guarantee the safety, reliability, and productivity of large-scale PV plants. In this paper, we propose a residual learning-based robotic (UAV) image analysis model for low-voltage distributed PV fault identification and positioning. In our target scenario, the unmanned aerial vehicles (UAVs) are deployed to acquire moving images of low-voltage distributed PV power plants. To get desired robustness and accuracy of PV image detection, we integrate residual learning with attention mechanism into the UAV image analysis model based on you only look once v4 (YOLOv4) network. Then, we design the sophisticated multi-scale spatial pyramid fusion and use it to optimize the YOLOv4 network for the nuanced task of fault localization within PV arrays, where the Complete-IOU loss is incorporated in the predictive modeling phase, significantly enhancing the accuracy and efficiency of fault detection. A series of experimental comparisons in terms of the accuracy of fault positioning are conducted, and the experimental results verify the feasibility and effectiveness of the proposed model in dealing with the safety and reliability maintenance of low-voltage distributed PV systems.","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-04-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140635325","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hybrid knowledge transfer for MARL based on action advising and experience sharing 基于行动建议和经验分享的混合型 MARL 知识转让
IF 3.1 4区 计算机科学 Q2 Computer Science Pub Date : 2024-04-10 DOI: 10.3389/fnbot.2024.1364587
Feng Liu, Dongqi Li, Jian Gao
Multiagent Reinforcement Learning (MARL) has been well adopted due to its exceptional ability to solve multiagent decision-making problems. To further enhance learning efficiency, knowledge transfer algorithms have been developed, among which experience-sharing-based and action-advising-based transfer strategies share the mainstream. However, it is notable that, although there exist many successful applications of both strategies, they are not flawless. For the long-developed action-advising-based methods (namely KT-AA, short for knowledge transfer based on action advising), their data efficiency and scalability are not satisfactory. As for the newly proposed experience-sharing-based knowledge transfer methods (KT-ES), although the shortcomings of KT-AA have been partially overcome, they are incompetent to correct specific bad decisions in the later learning stage. To leverage the superiority of both KT-AA and KT-ES, this study proposes KT-Hybrid, a hybrid knowledge transfer approach. In the early learning phase, KT-ES methods are employed, expecting better data efficiency from KT-ES to enhance the policy to a basic level as soon as possible. Later, we focus on correcting specific errors made by the basic policy, trying to use KT-AA methods to further improve the performance. Simulations demonstrate that the proposed KT-Hybrid outperforms well-received action-advising- and experience-sharing-based methods.
多代理强化学习(MARL)因其解决多代理决策问题的卓越能力而被广泛采用。为了进一步提高学习效率,人们开发了知识迁移算法,其中基于经验分享的迁移策略和基于行动建议的迁移策略成为主流。然而,值得注意的是,尽管这两种策略都有许多成功的应用,但它们并非完美无缺。就开发已久的基于行动建议的方法(即 KT-AA,是基于行动建议的知识转移的简称)而言,其数据效率和可扩展性并不令人满意。至于新近提出的基于经验分享的知识转移方法(KT-ES),虽然部分克服了 KT-AA 的缺点,但却无法纠正后期学习阶段的具体错误决策。为了充分利用 KT-AA 和 KT-ES 的优势,本研究提出了混合知识转移方法 KT-Hybrid。在早期学习阶段,我们采用 KT-ES 方法,期望 KT-ES 能提供更好的数据效率,从而尽快将政策提升到基本水平。之后,我们将重点放在纠正基本策略的特定错误上,尝试使用 KT-AA 方法进一步提高性能。模拟结果表明,所提出的 KT-Hybrid 方法优于广受好评的基于行动建议和经验分享的方法。
{"title":"Hybrid knowledge transfer for MARL based on action advising and experience sharing","authors":"Feng Liu, Dongqi Li, Jian Gao","doi":"10.3389/fnbot.2024.1364587","DOIUrl":"https://doi.org/10.3389/fnbot.2024.1364587","url":null,"abstract":"Multiagent Reinforcement Learning (MARL) has been well adopted due to its exceptional ability to solve multiagent decision-making problems. To further enhance learning efficiency, knowledge transfer algorithms have been developed, among which experience-sharing-based and action-advising-based transfer strategies share the mainstream. However, it is notable that, although there exist many successful applications of both strategies, they are not flawless. For the long-developed action-advising-based methods (namely KT-AA, short for knowledge transfer based on action advising), their data efficiency and scalability are not satisfactory. As for the newly proposed experience-sharing-based knowledge transfer methods (KT-ES), although the shortcomings of KT-AA have been partially overcome, they are incompetent to correct specific bad decisions in the later learning stage. To leverage the superiority of both KT-AA and KT-ES, this study proposes KT-Hybrid, a hybrid knowledge transfer approach. In the early learning phase, KT-ES methods are employed, expecting better data efficiency from KT-ES to enhance the policy to a basic level as soon as possible. Later, we focus on correcting specific errors made by the basic policy, trying to use KT-AA methods to further improve the performance. Simulations demonstrate that the proposed KT-Hybrid outperforms well-received action-advising- and experience-sharing-based methods.","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140885486","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Rail surface defect data enhancement method based on improved ACGAN 基于改进 ACGAN 的铁路表面缺陷数据增强方法
IF 3.1 4区 计算机科学 Q2 Computer Science Pub Date : 2024-04-09 DOI: 10.3389/fnbot.2024.1397369
He Zhendong, Gao Xiangyang, Liu Zhiyuan, An Xiaoyu, Zheng Anping
Rail surface defects present a significant safety concern in railway operations. However, the scarcity of data poses challenges for employing deep learning in defect detection. This study proposes an enhanced ACGAN augmentation method to address these issues. Residual blocks mitigate vanishing gradient problems, while a spectral norm regularization-constrained discriminator improves stability and image quality. Substituting the generator’s deconvolution layer with upsampling and convolution operations enhances computational efficiency. A gradient penalty mechanism based on regret values addresses gradient abnormality concerns. Experimental validation demonstrates superior image clarity and classification accuracy compared to ACGAN, with a 17.6% reduction in FID value. MNIST dataset experiments verify the model’s generalization ability. This approach offers practical value for real-world applications.
铁路表面缺陷是铁路运营中的一个重大安全问题。然而,数据的稀缺性给采用深度学习进行缺陷检测带来了挑战。本研究提出了一种增强型 ACGAN 增强方法来解决这些问题。残差块缓解了梯度消失问题,而频谱规范正则化约束判别器提高了稳定性和图像质量。用上采样和卷积操作取代生成器的解卷积层,提高了计算效率。基于遗憾值的梯度惩罚机制解决了梯度异常问题。实验验证表明,与 ACGAN 相比,该模型的图像清晰度和分类准确性更胜一筹,FID 值降低了 17.6%。MNIST 数据集实验验证了该模型的泛化能力。这种方法为现实世界的应用提供了实用价值。
{"title":"Rail surface defect data enhancement method based on improved ACGAN","authors":"He Zhendong, Gao Xiangyang, Liu Zhiyuan, An Xiaoyu, Zheng Anping","doi":"10.3389/fnbot.2024.1397369","DOIUrl":"https://doi.org/10.3389/fnbot.2024.1397369","url":null,"abstract":"Rail surface defects present a significant safety concern in railway operations. However, the scarcity of data poses challenges for employing deep learning in defect detection. This study proposes an enhanced ACGAN augmentation method to address these issues. Residual blocks mitigate vanishing gradient problems, while a spectral norm regularization-constrained discriminator improves stability and image quality. Substituting the generator’s deconvolution layer with upsampling and convolution operations enhances computational efficiency. A gradient penalty mechanism based on regret values addresses gradient abnormality concerns. Experimental validation demonstrates superior image clarity and classification accuracy compared to ACGAN, with a 17.6% reduction in FID value. MNIST dataset experiments verify the model’s generalization ability. This approach offers practical value for real-world applications.","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-04-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140584722","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Extended residual learning with one-shot imitation learning for robotic assembly in semi-structured environment 针对半结构化环境中机器人装配的一次模仿学习与扩展残差学习
IF 3.1 4区 计算机科学 Q2 Computer Science Pub Date : 2024-04-08 DOI: 10.3389/fnbot.2024.1355170
Chuang Wang, Chupeng Su, Baozheng Sun, Gang Chen, Longhan Xie
IntroductionRobotic assembly tasks require precise manipulation and coordination, often necessitating advanced learning techniques to achieve efficient and effective performance. While residual reinforcement learning with a base policy has shown promise in this domain, existing base policy approaches often rely on hand-designed full-state features and policies or extensive demonstrations, limiting their applicability in semi-structured environments.MethodsIn this study, we propose an innovative Object-Embodiment-Centric Imitation and Residual Reinforcement Learning (OEC-IRRL) approach that leverages an object-embodiment-centric (OEC) task representation to integrate vision models with imitation and residual learning. By utilizing a single demonstration and minimizing interactions with the environment, our method aims to enhance learning efficiency and effectiveness. The proposed method involves three key steps: creating an object-embodiment-centric task representation, employing imitation learning for a base policy using via-point movement primitives for generalization to different settings, and utilizing residual RL for uncertainty-aware policy refinement during the assembly phase.ResultsThrough a series of comprehensive experiments, we investigate the impact of the OEC task representation on base and residual policy learning and demonstrate the effectiveness of the method in semi-structured environments. Our results indicate that the approach, requiring only a single demonstration and less than 1.2 h of interaction, improves success rates by 46% and reduces assembly time by 25%.DiscussionThis research presents a promising avenue for robotic assembly tasks, providing a viable solution without the need for specialized expertise or custom fixtures.
引言机器人装配任务需要精确的操作和协调,通常需要先进的学习技术来实现高效和有效的性能。在本研究中,我们提出了一种创新的以对象-体现为中心的模仿和剩余强化学习(OEC-IRRL)方法,该方法利用以对象-体现为中心的任务表示法,将视觉模型与模仿和剩余学习整合在一起。通过利用单个演示和尽量减少与环境的交互,我们的方法旨在提高学习效率和效果。所提出的方法包括三个关键步骤:创建一个以物体为中心的任务表示法;利用模仿学习来制定基本策略,并通过点运动原语将其泛化到不同的设置中;以及在组装阶段利用残差 RL 来进行不确定性感知策略改进。结果通过一系列综合实验,我们研究了 OEC 任务表示法对基本策略和残差策略学习的影响,并在半结构化环境中展示了该方法的有效性。结果表明,这种方法只需要一次演示和不到 1.2 小时的交互,就能将成功率提高 46%,并将装配时间缩短 25%。讨论这项研究为机器人装配任务提供了一种前景广阔的途径,它提供了一种可行的解决方案,无需专业知识或定制夹具。
{"title":"Extended residual learning with one-shot imitation learning for robotic assembly in semi-structured environment","authors":"Chuang Wang, Chupeng Su, Baozheng Sun, Gang Chen, Longhan Xie","doi":"10.3389/fnbot.2024.1355170","DOIUrl":"https://doi.org/10.3389/fnbot.2024.1355170","url":null,"abstract":"IntroductionRobotic assembly tasks require precise manipulation and coordination, often necessitating advanced learning techniques to achieve efficient and effective performance. While residual reinforcement learning with a base policy has shown promise in this domain, existing base policy approaches often rely on hand-designed full-state features and policies or extensive demonstrations, limiting their applicability in semi-structured environments.MethodsIn this study, we propose an innovative Object-Embodiment-Centric Imitation and Residual Reinforcement Learning (OEC-IRRL) approach that leverages an object-embodiment-centric (OEC) task representation to integrate vision models with imitation and residual learning. By utilizing a single demonstration and minimizing interactions with the environment, our method aims to enhance learning efficiency and effectiveness. The proposed method involves three key steps: creating an object-embodiment-centric task representation, employing imitation learning for a base policy using via-point movement primitives for generalization to different settings, and utilizing residual RL for uncertainty-aware policy refinement during the assembly phase.ResultsThrough a series of comprehensive experiments, we investigate the impact of the OEC task representation on base and residual policy learning and demonstrate the effectiveness of the method in semi-structured environments. Our results indicate that the approach, requiring only a single demonstration and less than 1.2 h of interaction, improves success rates by 46% and reduces assembly time by 25%.DiscussionThis research presents a promising avenue for robotic assembly tasks, providing a viable solution without the need for specialized expertise or custom fixtures.","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":null,"pages":null},"PeriodicalIF":3.1,"publicationDate":"2024-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140810365","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Frontiers in Neurorobotics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1