首页 > 最新文献

Frontiers in Neurorobotics最新文献

英文 中文
Unmanned aerial vehicle based multi-person detection via deep neural network models. 基于深度神经网络模型的无人机多人检测。
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-17 eCollection Date: 2025-01-01 DOI: 10.3389/fnbot.2025.1582995
Mohammed Alshehri, Laiba Zahoor, Yahya AlQahtani, Abdulmonem Alshahrani, Dina Abdulaziz AlHammadi, Ahmad Jalal, Hui Liu

Introduction: Understanding human actions in complex environments is crucial for advancing applications in areas such as surveillance, robotics, and autonomous systems. Identifying actions from UAV-recorded videos becomes more challenging as the task presents unique challenges, including motion blur, dynamic background, lighting variations, and varying viewpoints. The presented work develops a deep learning system that recognizes multi-person behaviors from data gathered by UAVs. The proposed system provides higher recognition accuracy while maintaining robustness along with dynamic environmental adaptability through the integration of different features and neural network models. The study supports the wider development of neural network systems utilized in complicated contexts while creating intelligent UAV applications utilizing neural networks.

Method: The proposed study uses deep learning and feature extraction approaches to create a novel method to recognize various actions in UAV-recorded video. The proposed model improves identification capacities and system robustness by addressing motion dynamic problems and intricate environmental constraints, encouraging advancements in UAV-based neural network systems.

Results: We proposed a deep learning-based framework with feature extraction approaches that may effectively increase the accuracy and robustness of multi-person action recognition in the challenging scenarios. Compared to the existing approaches, our system achieved 91.50% on MOD20 dataset and 89.71% on Okutama-Action. These results do, in fact, show how useful neural network-based methods are for managing the limitations of UAV-based application.

Discussion: Results how that the proposed framework is indeed effective at multi-person action recognition under difficult UAV conditions.

简介:了解人类在复杂环境中的行为对于推进监控、机器人和自主系统等领域的应用至关重要。从无人机录制的视频中识别动作变得更具挑战性,因为任务提出了独特的挑战,包括运动模糊,动态背景,照明变化和不同的观点。本文开发了一种深度学习系统,该系统可以从无人机收集的数据中识别多人行为。该系统通过集成不同特征和神经网络模型,在保持鲁棒性和动态环境适应性的同时,提高了识别精度。该研究支持在复杂环境中使用神经网络系统的更广泛发展,同时利用神经网络创建智能无人机应用。方法:本研究采用深度学习和特征提取方法,创建了一种新的方法来识别无人机录制视频中的各种动作。该模型通过解决运动动态问题和复杂的环境约束,提高了识别能力和系统鲁棒性,促进了基于无人机的神经网络系统的发展。结果:我们提出了一个基于深度学习的框架和特征提取方法,可以有效地提高具有挑战性场景下多人动作识别的准确性和鲁棒性。与现有方法相比,我们的系统在MOD20数据集上的准确率为91.50%,在Okutama-Action数据集上的准确率为89.71%。事实上,这些结果确实表明,基于神经网络的方法对于管理基于无人机的应用的局限性是多么有用。讨论:结果表明所提出的框架在无人机困难条件下的多人动作识别是有效的。
{"title":"Unmanned aerial vehicle based multi-person detection via deep neural network models.","authors":"Mohammed Alshehri, Laiba Zahoor, Yahya AlQahtani, Abdulmonem Alshahrani, Dina Abdulaziz AlHammadi, Ahmad Jalal, Hui Liu","doi":"10.3389/fnbot.2025.1582995","DOIUrl":"https://doi.org/10.3389/fnbot.2025.1582995","url":null,"abstract":"<p><strong>Introduction: </strong>Understanding human actions in complex environments is crucial for advancing applications in areas such as surveillance, robotics, and autonomous systems. Identifying actions from UAV-recorded videos becomes more challenging as the task presents unique challenges, including motion blur, dynamic background, lighting variations, and varying viewpoints. The presented work develops a deep learning system that recognizes multi-person behaviors from data gathered by UAVs. The proposed system provides higher recognition accuracy while maintaining robustness along with dynamic environmental adaptability through the integration of different features and neural network models. The study supports the wider development of neural network systems utilized in complicated contexts while creating intelligent UAV applications utilizing neural networks.</p><p><strong>Method: </strong>The proposed study uses deep learning and feature extraction approaches to create a novel method to recognize various actions in UAV-recorded video. The proposed model improves identification capacities and system robustness by addressing motion dynamic problems and intricate environmental constraints, encouraging advancements in UAV-based neural network systems.</p><p><strong>Results: </strong>We proposed a deep learning-based framework with feature extraction approaches that may effectively increase the accuracy and robustness of multi-person action recognition in the challenging scenarios. Compared to the existing approaches, our system achieved 91.50% on MOD20 dataset and 89.71% on Okutama-Action. These results do, in fact, show how useful neural network-based methods are for managing the limitations of UAV-based application.</p><p><strong>Discussion: </strong>Results how that the proposed framework is indeed effective at multi-person action recognition under difficult UAV conditions.</p>","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"19 ","pages":"1582995"},"PeriodicalIF":2.6,"publicationDate":"2025-04-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12043872/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143997194","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HR-NeRF: advancing realism and accuracy in highlight scene representation. HR-NeRF:在高光场景表现中推进现实主义和准确性。
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-16 eCollection Date: 2025-01-01 DOI: 10.3389/fnbot.2025.1558948
Shufan Dai, Shanqin Wang

NeRF and its variants excel in novel view synthesis but struggle with scenes featuring specular highlights. To address this limitation, we introduce the Highlight Recovery Network (HRNet), a new architecture that enhances NeRF's ability to capture specular scenes. HRNet incorporates Swish activation functions, affine transformations, multilayer perceptrons (MLPs), and residual blocks, which collectively enable smooth non-linear transformations, adaptive feature scaling, and hierarchical feature extraction. The residual connections help mitigate the vanishing gradient problem, ensuring stable training. Despite the simplicity of HRNet's components, it achieves impressive results in recovering specular highlights. Additionally, a density voxel grid enhances model efficiency. Evaluations on four inward-facing benchmarks demonstrate that our approach outperforms NeRF and its variants, achieving a 3-5 dB PSNR improvement on each dataset while accurately capturing scene details. Furthermore, our method effectively preserves image details without requiring positional encoding, rendering a single scene in ~18 min on an NVIDIA RTX 3090 Ti GPU.

NeRF和它的变体擅长新颖的视图合成,但与具有镜面高光的场景斗争。为了解决这一限制,我们引入了高光恢复网络(HRNet),这是一种增强NeRF捕捉高光场景能力的新架构。HRNet结合了Swish激活函数、仿射变换、多层感知器(mlp)和残差块,它们共同实现了平滑的非线性变换、自适应特征缩放和分层特征提取。残差连接有助于缓解梯度消失问题,保证训练的稳定性。尽管HRNet的组件简单,但它在恢复镜面高光方面取得了令人印象深刻的结果。此外,密度体素网格提高了模型效率。对四个内向基准的评估表明,我们的方法优于NeRF及其变体,在准确捕获场景细节的同时,在每个数据集上实现了3-5 dB的PSNR改进。此外,我们的方法在不需要位置编码的情况下有效地保留了图像细节,在NVIDIA RTX 3090 Ti GPU上渲染单个场景约18分钟。
{"title":"HR-NeRF: advancing realism and accuracy in highlight scene representation.","authors":"Shufan Dai, Shanqin Wang","doi":"10.3389/fnbot.2025.1558948","DOIUrl":"https://doi.org/10.3389/fnbot.2025.1558948","url":null,"abstract":"<p><p>NeRF and its variants excel in novel view synthesis but struggle with scenes featuring specular highlights. To address this limitation, we introduce the Highlight Recovery Network (HRNet), a new architecture that enhances NeRF's ability to capture specular scenes. HRNet incorporates Swish activation functions, affine transformations, multilayer perceptrons (MLPs), and residual blocks, which collectively enable smooth non-linear transformations, adaptive feature scaling, and hierarchical feature extraction. The residual connections help mitigate the vanishing gradient problem, ensuring stable training. Despite the simplicity of HRNet's components, it achieves impressive results in recovering specular highlights. Additionally, a density voxel grid enhances model efficiency. Evaluations on four inward-facing benchmarks demonstrate that our approach outperforms NeRF and its variants, achieving a 3-5 dB PSNR improvement on each dataset while accurately capturing scene details. Furthermore, our method effectively preserves image details without requiring positional encoding, rendering a single scene in ~18 min on an NVIDIA RTX 3090 Ti GPU.</p>","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"19 ","pages":"1558948"},"PeriodicalIF":2.6,"publicationDate":"2025-04-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12041011/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143988032","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Editorial: Neural network models in autonomous robotics. 编辑:自主机器人中的神经网络模型。
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-08 eCollection Date: 2025-01-01 DOI: 10.3389/fnbot.2025.1587137
Long Cheng, Ying Mao, Tomas Ward
{"title":"Editorial: Neural network models in autonomous robotics.","authors":"Long Cheng, Ying Mao, Tomas Ward","doi":"10.3389/fnbot.2025.1587137","DOIUrl":"https://doi.org/10.3389/fnbot.2025.1587137","url":null,"abstract":"","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"19 ","pages":"1587137"},"PeriodicalIF":2.6,"publicationDate":"2025-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12011765/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143987861","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Editorial: Towards a novel paradigm in brain-inspired computer vision. 社论:迈向大脑启发计算机视觉的新范式。
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-07 eCollection Date: 2025-01-01 DOI: 10.3389/fnbot.2025.1592181
Xianmin Wang, Jing Li
{"title":"Editorial: Towards a novel paradigm in brain-inspired computer vision.","authors":"Xianmin Wang, Jing Li","doi":"10.3389/fnbot.2025.1592181","DOIUrl":"https://doi.org/10.3389/fnbot.2025.1592181","url":null,"abstract":"","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"19 ","pages":"1592181"},"PeriodicalIF":2.6,"publicationDate":"2025-04-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12009925/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143996715","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Study on the standardization method of radiotelephony communication in low-altitude airspace based on BART. 基于BART的低空空域无线电话通信标准化方法研究。
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-02 eCollection Date: 2025-01-01 DOI: 10.3389/fnbot.2025.1482327
Weijun Pan, Boyuan Han, Peiyuan Jiang

The development of air traffic control (ATC) automation has been constrained by the scarcity and low quality of communication data, particularly in low-altitude complex airspace, where non-standardized instructions frequently hinder training efficiency and operational safety. This paper proposes the BART-Reinforcement Learning (BRL) model, a deep reinforcement learning model based on the BART pre-trained language model, optimized through transfer learning and reinforcement learning techniques. The model was evaluated on multiple ATC datasets, including training flight data, civil aviation operational data, and standardized datasets generated from Radiotelephony Communications for Air Traffic Services. Evaluation metrics included ROUGE and semantic intent-based indicators, with comparative analysis against several baseline models. Experimental results demonstrate that BRL achieves a 10.5% improvement in overall accuracy on the training dataset with the highest degree of non-standardization, significantly outperforming the baseline models. Furthermore, comprehensive evaluations validate the model's effectiveness in standardizing various types of instructions. The findings suggest that reinforcement learning-based approaches have the potential to significantly enhance ATC automation, reducing communication inconsistencies, and improving training efficiency and operational safety. Future research may further optimize standardization by incorporating additional contextual factors into the model.

空中交通管制(ATC)自动化的发展一直受到通信数据稀缺和低质量的制约,特别是在低空复杂空域,非标准化指令经常影响训练效率和操作安全。本文提出了BART- reinforcement Learning (BRL)模型,这是一种基于BART预训练语言模型的深度强化学习模型,通过迁移学习和强化学习技术进行优化。该模型在多个ATC数据集上进行了评估,包括训练飞行数据、民航运营数据和由空中交通服务无线电话通信生成的标准化数据集。评估指标包括ROUGE和基于语义意图的指标,并对几个基线模型进行了比较分析。实验结果表明,在非标准化程度最高的训练数据集上,BRL的整体准确率提高了10.5%,显著优于基线模型。此外,综合评估验证了该模型在标准化各种类型指令方面的有效性。研究结果表明,基于强化学习的方法有可能显著提高ATC自动化,减少通信不一致,提高培训效率和操作安全性。未来的研究可能会通过将额外的上下文因素纳入模型来进一步优化标准化。
{"title":"Study on the standardization method of radiotelephony communication in low-altitude airspace based on BART.","authors":"Weijun Pan, Boyuan Han, Peiyuan Jiang","doi":"10.3389/fnbot.2025.1482327","DOIUrl":"https://doi.org/10.3389/fnbot.2025.1482327","url":null,"abstract":"<p><p>The development of air traffic control (ATC) automation has been constrained by the scarcity and low quality of communication data, particularly in low-altitude complex airspace, where non-standardized instructions frequently hinder training efficiency and operational safety. This paper proposes the BART-Reinforcement Learning (BRL) model, a deep reinforcement learning model based on the BART pre-trained language model, optimized through transfer learning and reinforcement learning techniques. The model was evaluated on multiple ATC datasets, including training flight data, civil aviation operational data, and standardized datasets generated from Radiotelephony Communications for Air Traffic Services. Evaluation metrics included ROUGE and semantic intent-based indicators, with comparative analysis against several baseline models. Experimental results demonstrate that BRL achieves a 10.5% improvement in overall accuracy on the training dataset with the highest degree of non-standardization, significantly outperforming the baseline models. Furthermore, comprehensive evaluations validate the model's effectiveness in standardizing various types of instructions. The findings suggest that reinforcement learning-based approaches have the potential to significantly enhance ATC automation, reducing communication inconsistencies, and improving training efficiency and operational safety. Future research may further optimize standardization by incorporating additional contextual factors into the model.</p>","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"19 ","pages":"1482327"},"PeriodicalIF":2.6,"publicationDate":"2025-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12000013/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144010427","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ModuCLIP: multi-scale CLIP framework for predicting foundation pit deformation in multi-modal robotic systems. ModuCLIP:多模态机器人系统中预测基坑变形的多尺度CLIP框架。
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-01 eCollection Date: 2025-01-01 DOI: 10.3389/fnbot.2025.1544694
Lin Wenbo, Li Tingting, Li Xiao

Introduction: Foundation pit deformation prediction is a critical aspect of underground engineering safety assessment, influencing construction quality and personnel safety. However, due to complex geological conditions and numerous environmental interference factors, traditional prediction methods struggle to achieve precise modeling. Conventional approaches, including numerical simulations, empirical formulas, and machine learning models, suffer from limitations such as high computational costs, poor generalization, or excessive dependence on specific data distributions. Recently, deep learning models, particularly cross-modal architectures, have demonstrated great potential in engineering applications. However, effectively integrating multi-modal data for improved prediction accuracy remains a significant challenge.

Methods: This study proposes a Multi-Scale Contrastive Language-Image Pretraining (CLP) framework, ModuCLIP, designed for foundation pit deformation prediction in multi-modal robotic systems. The framework leverages a self-supervised contrastive learning mechanism to integrate multi-source information, including images, textual descriptions, and sensor data, while employing a multi-scale feature learning approach to enhance adaptability to complex conditions. Experiments conducted on multiple foundation pit engineering datasets demonstrate that ModuCLIP outperforms existing methods in terms of prediction accuracy, generalization, and robustness.

Results and discussion: The findings suggest that this framework provides an efficient and precise solution for foundation pit deformation prediction while offering new insights into multi-modal robotic perception and engineering monitoring applications.

基坑变形预测是地下工程安全评价的一个重要方面,影响着施工质量和人员安全。然而,由于地质条件复杂,环境干扰因素众多,传统的预测方法难以实现精确的建模。传统的方法,包括数值模拟、经验公式和机器学习模型,都受到诸如计算成本高、泛化能力差或过度依赖特定数据分布等限制。最近,深度学习模型,特别是跨模态架构,在工程应用中显示出巨大的潜力。然而,如何有效地整合多模态数据以提高预测精度仍然是一个重大挑战。方法:本研究提出了一个多尺度对比语言-图像预训练(CLP)框架ModuCLIP,用于多模态机器人系统的基坑变形预测。该框架利用自监督对比学习机制来整合多源信息,包括图像、文本描述和传感器数据,同时采用多尺度特征学习方法来增强对复杂条件的适应性。在多个基坑工程数据集上进行的实验表明,ModuCLIP在预测精度、泛化和鲁棒性方面优于现有方法。结果与讨论:研究结果表明,该框架为基坑变形预测提供了高效、精确的解决方案,同时为多模态机器人感知和工程监测应用提供了新的见解。
{"title":"ModuCLIP: multi-scale CLIP framework for predicting foundation pit deformation in multi-modal robotic systems.","authors":"Lin Wenbo, Li Tingting, Li Xiao","doi":"10.3389/fnbot.2025.1544694","DOIUrl":"https://doi.org/10.3389/fnbot.2025.1544694","url":null,"abstract":"<p><strong>Introduction: </strong>Foundation pit deformation prediction is a critical aspect of underground engineering safety assessment, influencing construction quality and personnel safety. However, due to complex geological conditions and numerous environmental interference factors, traditional prediction methods struggle to achieve precise modeling. Conventional approaches, including numerical simulations, empirical formulas, and machine learning models, suffer from limitations such as high computational costs, poor generalization, or excessive dependence on specific data distributions. Recently, deep learning models, particularly cross-modal architectures, have demonstrated great potential in engineering applications. However, effectively integrating multi-modal data for improved prediction accuracy remains a significant challenge.</p><p><strong>Methods: </strong>This study proposes a Multi-Scale Contrastive Language-Image Pretraining (CLP) framework, ModuCLIP, designed for foundation pit deformation prediction in multi-modal robotic systems. The framework leverages a self-supervised contrastive learning mechanism to integrate multi-source information, including images, textual descriptions, and sensor data, while employing a multi-scale feature learning approach to enhance adaptability to complex conditions. Experiments conducted on multiple foundation pit engineering datasets demonstrate that ModuCLIP outperforms existing methods in terms of prediction accuracy, generalization, and robustness.</p><p><strong>Results and discussion: </strong>The findings suggest that this framework provides an efficient and precise solution for foundation pit deformation prediction while offering new insights into multi-modal robotic perception and engineering monitoring applications.</p>","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"19 ","pages":"1544694"},"PeriodicalIF":2.6,"publicationDate":"2025-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11996866/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143979300","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Erratum: Latent space improved masked reconstruction model for human skeleton-based action recognition. 基于人体骨骼动作识别的潜在空间改进掩码重建模型。
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-03-18 eCollection Date: 2025-01-01 DOI: 10.3389/fnbot.2025.1587250

[This corrects the article DOI: 10.3389/fnbot.2025.1482281.].

[这更正了文章DOI: 10.3389/fnbot.2025.1482281.]。
{"title":"Erratum: Latent space improved masked reconstruction model for human skeleton-based action recognition.","authors":"","doi":"10.3389/fnbot.2025.1587250","DOIUrl":"https://doi.org/10.3389/fnbot.2025.1587250","url":null,"abstract":"<p><p>[This corrects the article DOI: 10.3389/fnbot.2025.1482281.].</p>","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"19 ","pages":"1587250"},"PeriodicalIF":2.6,"publicationDate":"2025-03-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11958970/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143763816","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A distributed penalty-based zeroing neural network for time-varying optimization with both equality and inequality constraints and its application to cooperative control of redundant robot manipulators. 基于分布式惩罚的不等式约束时变优化归零神经网络及其在冗余机器人机械手协同控制中的应用。
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-03-17 eCollection Date: 2025-01-01 DOI: 10.3389/fnbot.2025.1553623
Liu He, Hui Cheng, Yunong Zhang

This study addresses the distributed optimization problem with time-varying objective functions and time-varying constraints in a multi-agent system (MAS). To tackle the distributed time-varying constrained optimization (DTVCO) problem, each agent in the MAS communicates with its neighbors while relying solely on local information, such as its own objective function and constraints, to compute the optimal solution. We propose a novel penalty-based zeroing neural network (PB-ZNN) to solve the continuous-time DTVCO (CTDTVCO) problem. The PB-ZNN model incorporates two penalty functions: The first penalizes agents for deviating from the states of their neighbors, driving all agents to reach a consensus, and the second penalizes agents for falling outside the feasible range, ensuring that the solutions of all agents remain within the constraints. The PB-ZNN model solves the CTDTVCO problem in a semi-centralized manner, where information exchange between agents is distributed, but computation is centralized. Building on the semi-centralized PB-ZNN model, we adopt the Euler formula to develop a distributed PB-ZNN (DPB-ZNN) algorithm for solving the discrete-time DTVCO (DTDTVCO) problem in a fully distributed manner. We present and prove the convergence theorems of the proposed PB-ZNN model and DPB-ZNN algorithm. The efficacy and accuracy of the DPB-ZNN algorithm are illustrated through numerical examples, including a simulation experiment applying the algorithm to the cooperative control of redundant manipulators.

研究了多智能体系统中具有时变目标函数和时变约束的分布式优化问题。为了解决分布式时变约束优化(DTVCO)问题,MAS中的每个智能体与相邻智能体通信,同时仅依靠自身的目标函数和约束等局部信息来计算最优解。针对连续时间DTVCO (CTDTVCO)问题,提出了一种新的基于惩罚的归零神经网络(PB-ZNN)。PB-ZNN模型包含两个惩罚函数:第一个惩罚agent偏离其邻居的状态,促使所有agent达成共识;第二个惩罚agent落在可行范围之外,确保所有agent的解保持在约束范围内。PB-ZNN模型以半集中式的方式解决了CTDTVCO问题,agent之间的信息交换是分布式的,计算是集中式的。在半集中式PB-ZNN模型的基础上,采用欧拉公式开发了一种分布式PB-ZNN (DPB-ZNN)算法,以完全分布式的方式解决离散时间DTDTVCO (DTDTVCO)问题。给出并证明了PB-ZNN模型和DPB-ZNN算法的收敛性定理。通过数值算例说明了DPB-ZNN算法的有效性和准确性,包括将该算法应用于冗余机械手协同控制的仿真实验。
{"title":"A distributed penalty-based zeroing neural network for time-varying optimization with both equality and inequality constraints and its application to cooperative control of redundant robot manipulators.","authors":"Liu He, Hui Cheng, Yunong Zhang","doi":"10.3389/fnbot.2025.1553623","DOIUrl":"10.3389/fnbot.2025.1553623","url":null,"abstract":"<p><p>This study addresses the distributed optimization problem with time-varying objective functions and time-varying constraints in a multi-agent system (MAS). To tackle the distributed time-varying constrained optimization (DTVCO) problem, each agent in the MAS communicates with its neighbors while relying solely on local information, such as its own objective function and constraints, to compute the optimal solution. We propose a novel penalty-based zeroing neural network (PB-ZNN) to solve the continuous-time DTVCO (CTDTVCO) problem. The PB-ZNN model incorporates two penalty functions: The first penalizes agents for deviating from the states of their neighbors, driving all agents to reach a consensus, and the second penalizes agents for falling outside the feasible range, ensuring that the solutions of all agents remain within the constraints. The PB-ZNN model solves the CTDTVCO problem in a semi-centralized manner, where information exchange between agents is distributed, but computation is centralized. Building on the semi-centralized PB-ZNN model, we adopt the Euler formula to develop a distributed PB-ZNN (DPB-ZNN) algorithm for solving the discrete-time DTVCO (DTDTVCO) problem in a fully distributed manner. We present and prove the convergence theorems of the proposed PB-ZNN model and DPB-ZNN algorithm. The efficacy and accuracy of the DPB-ZNN algorithm are illustrated through numerical examples, including a simulation experiment applying the algorithm to the cooperative control of redundant manipulators.</p>","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"19 ","pages":"1553623"},"PeriodicalIF":2.6,"publicationDate":"2025-03-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11955690/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143752251","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
High-efficiency sparse convolution operator for event-based cameras. 基于事件相机的高效稀疏卷积算子。
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-03-12 eCollection Date: 2025-01-01 DOI: 10.3389/fnbot.2025.1537673
Sen Zhang, Fusheng Zha, Xiangji Wang, Mantian Li, Wei Guo, Pengfei Wang, Xiaolin Li, Lining Sun

Event-based cameras are bio-inspired vision sensors that mimic the sparse and asynchronous activation of the animal retina, offering advantages such as low latency and low computational load in various robotic applications. However, despite their inherent sparsity, most existing visual processing algorithms are optimized for conventional standard cameras and dense images captured from them, resulting in computational redundancy and high latency when applied to event-based cameras. To address this gap, we propose a sparse convolution operator tailored for event-based cameras. By selectively skipping invalid sub-convolutions and efficiently reorganizing valid computations, our operator reduces computational workload by nearly 90% and achieves almost 2× acceleration in processing speed, while maintaining the same accuracy as dense convolution operators. This innovation unlocks the potential of event-based cameras in applications such as autonomous navigation, real-time object tracking, and industrial inspection, enabling low-latency and high-efficiency perception in resource-constrained robotic systems.

基于事件的相机是仿生视觉传感器,模仿动物视网膜的稀疏和异步激活,在各种机器人应用中提供低延迟和低计算负载等优势。然而,尽管它们具有固有的稀疏性,但大多数现有的视觉处理算法都针对传统的标准相机和从它们捕获的密集图像进行了优化,从而导致计算冗余和高延迟,当应用于基于事件的相机时。为了解决这一差距,我们提出了一种为基于事件的相机量身定制的稀疏卷积算子。通过选择性地跳过无效的子卷积并有效地重组有效的计算,我们的算子减少了近90%的计算工作量,实现了近2倍的处理速度加速,同时保持了与密集卷积算子相同的精度。这一创新释放了基于事件的相机在自主导航、实时目标跟踪和工业检测等应用中的潜力,在资源受限的机器人系统中实现了低延迟和高效率的感知。
{"title":"High-efficiency sparse convolution operator for event-based cameras.","authors":"Sen Zhang, Fusheng Zha, Xiangji Wang, Mantian Li, Wei Guo, Pengfei Wang, Xiaolin Li, Lining Sun","doi":"10.3389/fnbot.2025.1537673","DOIUrl":"10.3389/fnbot.2025.1537673","url":null,"abstract":"<p><p>Event-based cameras are bio-inspired vision sensors that mimic the sparse and asynchronous activation of the animal retina, offering advantages such as low latency and low computational load in various robotic applications. However, despite their inherent sparsity, most existing visual processing algorithms are optimized for conventional standard cameras and dense images captured from them, resulting in computational redundancy and high latency when applied to event-based cameras. To address this gap, we propose a sparse convolution operator tailored for event-based cameras. By selectively skipping invalid sub-convolutions and efficiently reorganizing valid computations, our operator reduces computational workload by nearly 90% and achieves almost 2× acceleration in processing speed, while maintaining the same accuracy as dense convolution operators. This innovation unlocks the potential of event-based cameras in applications such as autonomous navigation, real-time object tracking, and industrial inspection, enabling low-latency and high-efficiency perception in resource-constrained robotic systems.</p>","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"19 ","pages":"1537673"},"PeriodicalIF":2.6,"publicationDate":"2025-03-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11936924/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143718530","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PoseRL-Net: human pose analysis for motion training guided by robot vision. PoseRL-Net:机器人视觉指导下的运动训练的人体姿态分析。
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-03-05 eCollection Date: 2025-01-01 DOI: 10.3389/fnbot.2025.1531894
Bin Liu, Hui Wang

Objective: To address the limitations of traditional methods in human pose recognition, such as occlusions, lighting variations, and motion continuity, particularly in complex dynamic environments for seamless human-robot interaction.

Method: We propose PoseRL-Net, a deep learning-based pose recognition model that enhances accuracy and robustness in human pose estimation. PoseRL-Net integrates multiple components, including a Spatial-Temporal Graph Convolutional Network (STGCN), attention mechanism, Gated Recurrent Unit (GRU) module, pose refinement, and symmetry constraints. The STGCN extracts spatial and temporal features, the attention mechanism focuses on key pose features, the GRU ensures temporal consistency, and the refinement and symmetry constraints improve structural plausibility and stability.

Results: Extensive experiments conducted on the Human3.6M and MPI-INF-3DHP datasets demonstrate that PoseRL-Net outperforms existing state-of-the-art models on key metrics such as MPIPE and P-MPIPE, showcasing superior performance across various pose recognition tasks.

Conclusion: PoseRL-Net not only improves pose estimation accuracy but also provides crucial support for intelligent decision-making and motion planning in robots operating in dynamic and complex scenarios, offering significant practical value for collaborative robotics.

目的:解决传统的人体姿势识别方法的局限性,如遮挡、光照变化和运动连续性,特别是在复杂的动态环境中实现无缝人机交互。方法:我们提出了基于深度学习的姿态识别模型PoseRL-Net,该模型提高了人体姿态估计的准确性和鲁棒性。PoseRL-Net集成了多个组件,包括时空图卷积网络(STGCN)、注意机制、门控循环单元(GRU)模块、姿态优化和对称约束。STGCN提取空间和时间特征,关注机制关注关键姿态特征,GRU保证时间一致性,精细化和对称约束提高结构的合理性和稳定性。结果:在Human3.6M和MPI-INF-3DHP数据集上进行的大量实验表明,PoseRL-Net在MPIPE和P-MPIPE等关键指标上优于现有的最先进模型,在各种姿势识别任务中表现出卓越的性能。结论:PoseRL-Net不仅提高了姿态估计精度,而且为机器人在动态和复杂场景下的智能决策和运动规划提供了重要支持,对协作机器人具有重要的实用价值。
{"title":"PoseRL-Net: human pose analysis for motion training guided by robot vision.","authors":"Bin Liu, Hui Wang","doi":"10.3389/fnbot.2025.1531894","DOIUrl":"10.3389/fnbot.2025.1531894","url":null,"abstract":"<p><strong>Objective: </strong>To address the limitations of traditional methods in human pose recognition, such as occlusions, lighting variations, and motion continuity, particularly in complex dynamic environments for seamless human-robot interaction.</p><p><strong>Method: </strong>We propose PoseRL-Net, a deep learning-based pose recognition model that enhances accuracy and robustness in human pose estimation. PoseRL-Net integrates multiple components, including a Spatial-Temporal Graph Convolutional Network (STGCN), attention mechanism, Gated Recurrent Unit (GRU) module, pose refinement, and symmetry constraints. The STGCN extracts spatial and temporal features, the attention mechanism focuses on key pose features, the GRU ensures temporal consistency, and the refinement and symmetry constraints improve structural plausibility and stability.</p><p><strong>Results: </strong>Extensive experiments conducted on the Human3.6M and MPI-INF-3DHP datasets demonstrate that PoseRL-Net outperforms existing state-of-the-art models on key metrics such as MPIPE and P-MPIPE, showcasing superior performance across various pose recognition tasks.</p><p><strong>Conclusion: </strong>PoseRL-Net not only improves pose estimation accuracy but also provides crucial support for intelligent decision-making and motion planning in robots operating in dynamic and complex scenarios, offering significant practical value for collaborative robotics.</p>","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"19 ","pages":"1531894"},"PeriodicalIF":2.6,"publicationDate":"2025-03-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11920136/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143663204","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Frontiers in Neurorobotics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1