首页 > 最新文献

Frontiers in Neurorobotics最新文献

英文 中文
TS-Resformer: a model based on multimodal fusion for the classification of music signals. TS-Resformer:一种基于多模态融合的音乐信号分类模型。
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-13 eCollection Date: 2025-01-01 DOI: 10.3389/fnbot.2025.1568811
Yilin Zhang

The number of music of different genres is increasing year by year, and manual classification is costly and requires professionals in the field of music to manually design features, some of which lack the generality of music genre classification. Deep learning has had a large number of scientific research results in the field of music classification, but the existing deep learning methods still have the problems of insufficient extraction of music feature information, low accuracy rate of music genres, loss of time series information, and slow training. To address the problem that different music durations affect the accuracy of music genre classification, we form a Log Mel spectrum with music audio data of different cut durations. After discarding incomplete audio, we design data enhancement with different slicing durations and verify its effect on accuracy and training time through comparison experiments. Based on this, the audio signal is divided into frames, windowed and short-time Fourier transformed, and then the Log Mel spectrum is obtained by using the Mel filter and logarithmic compression. Aiming at the problems of loss of time information, insufficient feature extraction, and low classification accuracy in music genre classification, firstly, we propose a Res-Transformer model that fuses the residual network with the Transformer coding layer. The model consists of two branches, the left branch is an improved residual network, which enhances the spectral feature extraction ability and network expression ability and realizes the dimensionality reduction; the right branch uses four Transformer coding layers to extract the time-series information of the Log Mel spectrum. The output vectors of the two branches are spliced and input into the classifier to realize music genre classification. Then, to further improve the classification accuracy of the model, we propose the TS-Resformer model based on the Res-Transformer model, combined with different attention mechanisms, and design the time-frequency attention mechanism, which employs different scales of filters to fully extract the low-level music features from the two dimensions of time and frequency as the input to the time-frequency attention mechanism, respectively. Finally, experiments show that the accuracy of this method is 90.23% on the FMA-small dataset, which is an improvement in classification accuracy compared with the classical model.

不同体裁的音乐数量逐年增加,手工分类成本高,需要音乐领域的专业人员手工设计功能,其中一些缺乏音乐体裁分类的通用性。深度学习在音乐分类领域已经有了大量的科学研究成果,但现有的深度学习方法仍然存在音乐特征信息提取不足、音乐流派准确率低、时间序列信息丢失、训练缓慢等问题。为了解决不同音乐持续时间影响音乐类型分类准确性的问题,我们将不同剪切持续时间的音乐音频数据组成Log Mel谱。在丢弃不完整音频后,我们设计了不同切片持续时间的数据增强,并通过对比实验验证了其对准确率和训练时间的影响。在此基础上,对音频信号进行分帧、加窗和短时傅里叶变换,然后利用梅尔滤波和对数压缩得到对数梅尔谱。针对音乐类型分类中存在的时间信息丢失、特征提取不足、分类准确率低等问题,首先提出了残差网络与Transformer编码层融合的Res-Transformer模型;该模型由两个分支组成,左分支是一个改进的残差网络,增强了光谱特征提取能力和网络表达能力,实现了降维;右分支使用四个Transformer编码层提取Log Mel频谱的时间序列信息。将两个分支的输出向量拼接输入到分类器中,实现音乐类型的分类。然后,为了进一步提高模型的分类精度,我们在Res-Transformer模型的基础上提出了TS-Resformer模型,结合不同的注意机制,设计了时频注意机制,采用不同尺度的滤波器分别从时间和频率两个维度上充分提取低层次音乐特征作为时频注意机制的输入。最后,实验表明,该方法在fma小数据集上的分类准确率为90.23%,与经典模型相比,分类准确率有了很大提高。
{"title":"TS-Resformer: a model based on multimodal fusion for the classification of music signals.","authors":"Yilin Zhang","doi":"10.3389/fnbot.2025.1568811","DOIUrl":"10.3389/fnbot.2025.1568811","url":null,"abstract":"<p><p>The number of music of different genres is increasing year by year, and manual classification is costly and requires professionals in the field of music to manually design features, some of which lack the generality of music genre classification. Deep learning has had a large number of scientific research results in the field of music classification, but the existing deep learning methods still have the problems of insufficient extraction of music feature information, low accuracy rate of music genres, loss of time series information, and slow training. To address the problem that different music durations affect the accuracy of music genre classification, we form a Log Mel spectrum with music audio data of different cut durations. After discarding incomplete audio, we design data enhancement with different slicing durations and verify its effect on accuracy and training time through comparison experiments. Based on this, the audio signal is divided into frames, windowed and short-time Fourier transformed, and then the Log Mel spectrum is obtained by using the Mel filter and logarithmic compression. Aiming at the problems of loss of time information, insufficient feature extraction, and low classification accuracy in music genre classification, firstly, we propose a Res-Transformer model that fuses the residual network with the Transformer coding layer. The model consists of two branches, the left branch is an improved residual network, which enhances the spectral feature extraction ability and network expression ability and realizes the dimensionality reduction; the right branch uses four Transformer coding layers to extract the time-series information of the Log Mel spectrum. The output vectors of the two branches are spliced and input into the classifier to realize music genre classification. Then, to further improve the classification accuracy of the model, we propose the TS-Resformer model based on the Res-Transformer model, combined with different attention mechanisms, and design the time-frequency attention mechanism, which employs different scales of filters to fully extract the low-level music features from the two dimensions of time and frequency as the input to the time-frequency attention mechanism, respectively. Finally, experiments show that the accuracy of this method is 90.23% on the FMA-small dataset, which is an improvement in classification accuracy compared with the classical model.</p>","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"19 ","pages":"1568811"},"PeriodicalIF":2.6,"publicationDate":"2025-05-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12106318/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144157987","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Motion control and singular perturbation algorithms for lower limb rehabilitation robots. 下肢康复机器人运动控制与奇异摄动算法。
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-09 eCollection Date: 2025-01-01 DOI: 10.3389/fnbot.2025.1562519
Yanchun Xie, Anna Wang, Xue Zhao, Yang Jiang, Yao Wu, Hailong Yu

To better assist patients with lower limb injuries in their rehabilitation training, this paper focuses on motion control and singular perturbation algorithms and their practical applications. First, the paper conducts an in-depth analysis of the mechanical structure of such robots and establishes detailed kinematics and dynamics models. An optimal S-type planning algorithm is proposed, transforming the S-type planning into an iterative solution problem for efficient and accelerated trajectory planning using dynamic equations. This algorithm comprehensively considers joint range of motion, speed constraints, and dynamic conditions, ensuring the smoothness and continuity of motion trajectories. Second, a zero-force control method is introduced, incorporating friction terms into the traditional dynamic equations and utilizing the LuGre friction model for friction analysis to achieve zero-force control. Furthermore, to address the multi-scale dynamic system characteristics present in rehabilitation training, a control method based on singular perturbation theory is proposed. This method enhances the system's robustness and adaptability by simplifying the system model and optimizing controller design, enabling it to better accommodate complex motion requirements during rehabilitation. Finally, experiments verify the correctness of the kinematics and optimal S-type trajectory planning. In lower limb rehabilitation robots, zero-force control can better assist patients in rehabilitation training for lower limb injuries, while the singular perturbation method improves the accuracy, response speed, and robustness of the control system, allowing it to adapt to individual rehabilitation needs and complex motion patterns. The novelty of this paper lies in the integration of the singular perturbation method with the LuGre friction model, significantly enhancing the precision of joint dynamic control, and improving controller design through the introduction of a torque deviation feedback mechanism, thereby increasing system stability and response speed. Experimental results demonstrate significant improvements in tracking error and system response compared to traditional methods, providing patients with a more comfortable and safer rehabilitation experience.

为了更好地辅助下肢损伤患者进行康复训练,本文重点研究了运动控制和奇异摄动算法及其实际应用。首先,对该类机器人的机械结构进行了深入分析,建立了详细的运动学和动力学模型。提出了一种最优s型规划算法,将s型规划转化为基于动态方程的高效加速轨迹规划迭代求解问题。该算法综合考虑关节运动范围、速度约束和动态条件,保证了运动轨迹的平滑性和连续性。其次,引入零力控制方法,将摩擦项引入传统动力学方程,利用LuGre摩擦模型进行摩擦分析,实现零力控制。此外,针对康复训练中存在的多尺度动态系统特性,提出了一种基于奇异摄动理论的控制方法。该方法通过简化系统模型和优化控制器设计,增强了系统的鲁棒性和适应性,使其能够更好地适应康复过程中复杂的运动要求。最后,通过实验验证了运动学和最优s型轨迹规划的正确性。在下肢康复机器人中,零力控制可以更好地辅助患者进行下肢损伤的康复训练,而奇异摄动方法提高了控制系统的精度、响应速度和鲁棒性,使其能够适应个性化的康复需求和复杂的运动模式。本文的新颖之处在于将奇异摄动法与LuGre摩擦模型相结合,显著提高了关节动态控制的精度,并通过引入转矩偏差反馈机制改进了控制器设计,从而提高了系统的稳定性和响应速度。实验结果表明,与传统方法相比,该方法在跟踪误差和系统响应方面有显著改善,为患者提供了更舒适、更安全的康复体验。
{"title":"Motion control and singular perturbation algorithms for lower limb rehabilitation robots.","authors":"Yanchun Xie, Anna Wang, Xue Zhao, Yang Jiang, Yao Wu, Hailong Yu","doi":"10.3389/fnbot.2025.1562519","DOIUrl":"10.3389/fnbot.2025.1562519","url":null,"abstract":"<p><p>To better assist patients with lower limb injuries in their rehabilitation training, this paper focuses on motion control and singular perturbation algorithms and their practical applications. First, the paper conducts an in-depth analysis of the mechanical structure of such robots and establishes detailed kinematics and dynamics models. An optimal S-type planning algorithm is proposed, transforming the S-type planning into an iterative solution problem for efficient and accelerated trajectory planning using dynamic equations. This algorithm comprehensively considers joint range of motion, speed constraints, and dynamic conditions, ensuring the smoothness and continuity of motion trajectories. Second, a zero-force control method is introduced, incorporating friction terms into the traditional dynamic equations and utilizing the LuGre friction model for friction analysis to achieve zero-force control. Furthermore, to address the multi-scale dynamic system characteristics present in rehabilitation training, a control method based on singular perturbation theory is proposed. This method enhances the system's robustness and adaptability by simplifying the system model and optimizing controller design, enabling it to better accommodate complex motion requirements during rehabilitation. Finally, experiments verify the correctness of the kinematics and optimal S-type trajectory planning. In lower limb rehabilitation robots, zero-force control can better assist patients in rehabilitation training for lower limb injuries, while the singular perturbation method improves the accuracy, response speed, and robustness of the control system, allowing it to adapt to individual rehabilitation needs and complex motion patterns. The novelty of this paper lies in the integration of the singular perturbation method with the LuGre friction model, significantly enhancing the precision of joint dynamic control, and improving controller design through the introduction of a torque deviation feedback mechanism, thereby increasing system stability and response speed. Experimental results demonstrate significant improvements in tracking error and system response compared to traditional methods, providing patients with a more comfortable and safer rehabilitation experience.</p>","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"19 ","pages":"1562519"},"PeriodicalIF":2.6,"publicationDate":"2025-05-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12098330/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144142345","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FOCUS: object-centric world models for robotic manipulation. 焦点:机器人操作的以对象为中心的世界模型。
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-30 eCollection Date: 2025-01-01 DOI: 10.3389/fnbot.2025.1585386
Stefano Ferraro, Pietro Mazzaglia, Tim Verbelen, Bart Dhoedt

Understanding the world in terms of objects and the possible interactions with them is an important cognitive ability. However, current world models adopted in reinforcement learning typically lack this structure and represent the world state in a global latent vector. To address this, we propose FOCUS, a model-based agent that learns an object-centric world model. This novel representation also enables the design of an object-centric exploration mechanism, which encourages the agent to interact with objects and discover useful interactions. We benchmark FOCUS in several robotic manipulation settings, where we found that our method can be used to improve manipulation skills. The object-centric world model leads to more accurate predictions of the objects in the scene and it enables more efficient learning. The object-centric exploration strategy fosters interactions with the objects in the environment, such as reaching, moving, and rotating them, and it allows fast adaptation of the agent to sparse reward reinforcement learning tasks. Using a Franka Emika robot arm, we also showcase how FOCUS proves useful in real-world applications. Website: focus-manipulation.github.io.

从物体及其可能的相互作用的角度来理解世界是一种重要的认知能力。然而,目前强化学习中采用的世界模型通常缺乏这种结构,而是用全局潜在向量表示世界状态。为了解决这个问题,我们提出FOCUS,一个基于模型的智能体,它学习一个以对象为中心的世界模型。这种新颖的表示还支持以对象为中心的探索机制的设计,该机制鼓励代理与对象进行交互并发现有用的交互。我们在几个机器人操作设置中对FOCUS进行了基准测试,发现我们的方法可以用来提高操作技能。以对象为中心的世界模型可以更准确地预测场景中的对象,并实现更有效的学习。以对象为中心的探索策略促进了与环境中对象的交互,例如到达,移动和旋转它们,并且它允许代理快速适应稀疏奖励强化学习任务。通过使用Franka Emika机械臂,我们还展示了FOCUS在实际应用中的实用性。网站:focus-manipulation.github.io。
{"title":"FOCUS: object-centric world models for robotic manipulation.","authors":"Stefano Ferraro, Pietro Mazzaglia, Tim Verbelen, Bart Dhoedt","doi":"10.3389/fnbot.2025.1585386","DOIUrl":"https://doi.org/10.3389/fnbot.2025.1585386","url":null,"abstract":"<p><p>Understanding the world in terms of objects and the possible interactions with them is an important cognitive ability. However, current world models adopted in reinforcement learning typically lack this structure and represent the world state in a global latent vector. To address this, we propose FOCUS, a model-based agent that learns an object-centric world model. This novel representation also enables the design of an object-centric exploration mechanism, which encourages the agent to interact with objects and discover useful interactions. We benchmark FOCUS in several robotic manipulation settings, where we found that our method can be used to improve manipulation skills. The object-centric world model leads to more accurate predictions of the objects in the scene and it enables more efficient learning. The object-centric exploration strategy fosters interactions with the objects in the environment, such as reaching, moving, and rotating them, and it allows fast adaptation of the agent to sparse reward reinforcement learning tasks. Using a Franka Emika robot arm, we also showcase how FOCUS proves useful in real-world applications. Website: focus-manipulation.github.io.</p>","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"19 ","pages":"1585386"},"PeriodicalIF":2.6,"publicationDate":"2025-04-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12075287/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144077540","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Medium density EMG armband for gesture recognition. 用于手势识别的中密度肌电臂带。
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-30 eCollection Date: 2025-01-01 DOI: 10.3389/fnbot.2025.1531815
Eisa Aghchehli, Milad Jabbari, Chenfei Ma, Matthew Dyson, Kianoush Nazarpour

Electromyography (EMG) systems are essential for the advancement of neuroprosthetics and human-machine interfaces. However, the gap between low-density and high-density systems poses challenges to researchers in experiment design and knowledge transfer. Medium-density surface EMG systems offer a balanced alternative, providing greater spatial resolution than low-density systems while avoiding the complexity and cost of high-density arrays. In this study, we developed a research-friendly medium-density EMG system and evaluated its performance with eleven volunteers performing grasping tasks. To enhance decoding accuracy, we introduced a novel spatio-temporal convolutional neural network that integrates spatial information from additional EMG sensors with temporal dynamics. The results show that medium-density EMG sensors significantly improve classification accuracy compared to low-density systems while maintaining the same footprint. Furthermore, the proposed neural network outperforms traditional gesture decoding approaches. This work highlights the potential of medium-density EMG systems as a practical and effective solution, bridging the gap between low- and high-density systems. These findings pave the way for broader adoption in research and potential clinical applications.

肌电图(EMG)系统对于神经修复和人机界面的发展至关重要。然而,低密度和高密度系统之间的差距给研究人员在实验设计和知识转移方面带来了挑战。中密度表面肌电信号系统提供了一种平衡的选择,提供比低密度系统更高的空间分辨率,同时避免了高密度阵列的复杂性和成本。在这项研究中,我们开发了一个适合研究的中密度肌电图系统,并通过11名志愿者执行抓取任务来评估其性能。为了提高解码精度,我们引入了一种新的时空卷积神经网络,该网络将来自额外肌电传感器的空间信息与时间动态相结合。结果表明,与低密度系统相比,中等密度的肌电传感器在保持相同足迹的情况下显著提高了分类精度。此外,所提出的神经网络优于传统的手势解码方法。这项工作强调了中密度肌电图系统作为一种实用有效的解决方案的潜力,弥合了低密度和高密度系统之间的差距。这些发现为更广泛地应用于研究和潜在的临床应用铺平了道路。
{"title":"Medium density EMG armband for gesture recognition.","authors":"Eisa Aghchehli, Milad Jabbari, Chenfei Ma, Matthew Dyson, Kianoush Nazarpour","doi":"10.3389/fnbot.2025.1531815","DOIUrl":"https://doi.org/10.3389/fnbot.2025.1531815","url":null,"abstract":"<p><p>Electromyography (EMG) systems are essential for the advancement of neuroprosthetics and human-machine interfaces. However, the gap between low-density and high-density systems poses challenges to researchers in experiment design and knowledge transfer. Medium-density surface EMG systems offer a balanced alternative, providing greater spatial resolution than low-density systems while avoiding the complexity and cost of high-density arrays. In this study, we developed a research-friendly medium-density EMG system and evaluated its performance with eleven volunteers performing grasping tasks. To enhance decoding accuracy, we introduced a novel spatio-temporal convolutional neural network that integrates spatial information from additional EMG sensors with temporal dynamics. The results show that medium-density EMG sensors significantly improve classification accuracy compared to low-density systems while maintaining the same footprint. Furthermore, the proposed neural network outperforms traditional gesture decoding approaches. This work highlights the potential of medium-density EMG systems as a practical and effective solution, bridging the gap between low- and high-density systems. These findings pave the way for broader adoption in research and potential clinical applications.</p>","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"19 ","pages":"1531815"},"PeriodicalIF":2.6,"publicationDate":"2025-04-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12075175/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144077543","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Diagnosis of pneumonia from chest X-ray images using YOLO deep learning. 基于YOLO深度学习的胸片肺炎诊断。
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-28 eCollection Date: 2025-01-01 DOI: 10.3389/fnbot.2025.1576438
Yanchun Xie, Binbin Zhu, Yang Jiang, Bin Zhao, Hailong Yu

Early and accurate diagnosis of pneumonia is crucial to improve cure rates and reduce mortality. Traditional chest X-ray analysis relies on physician experience, which can lead to subjectivity and misdiagnosis. To address this, we propose a novel pneumonia diagnosis method using the Fast-YOLO deep learning network that we introduced. First, we constructed a pneumonia dataset containing five categories and applied image enhancement techniques to increase data diversity and improve the model's generalization ability. Next, the YOLOv11 network structure was redesigned to accommodate the complex features of pneumonia X-ray images. By integrating the C3k2 module, DCNv2, and DynamicConv, the Fast-YOLO network effectively enhanced feature representation and reduced computational complexity (FPS increased from 53 to 120). Experimental results subsequently show that our method outperforms other commonly used detection models in terms of accuracy, recall, and mAP, offering better real-time detection capability and clinical application potential.

肺炎的早期和准确诊断对于提高治愈率和降低死亡率至关重要。传统的胸部x线分析依赖于医生的经验,容易导致主观性和误诊。为了解决这个问题,我们提出了一种新的肺炎诊断方法,使用我们介绍的Fast-YOLO深度学习网络。首先,我们构建了包含五个类别的肺炎数据集,并应用图像增强技术来增加数据的多样性,提高模型的泛化能力。接下来,重新设计了YOLOv11网络结构,以适应肺炎x线图像的复杂特征。通过集成C3k2模块、DCNv2和DynamicConv, Fast-YOLO网络有效地增强了特征表示,降低了计算复杂度(FPS从53增加到120)。实验结果表明,该方法在准确率、召回率和mAP等方面均优于其他常用检测模型,具有更好的实时检测能力和临床应用潜力。
{"title":"Diagnosis of pneumonia from chest X-ray images using YOLO deep learning.","authors":"Yanchun Xie, Binbin Zhu, Yang Jiang, Bin Zhao, Hailong Yu","doi":"10.3389/fnbot.2025.1576438","DOIUrl":"https://doi.org/10.3389/fnbot.2025.1576438","url":null,"abstract":"<p><p>Early and accurate diagnosis of pneumonia is crucial to improve cure rates and reduce mortality. Traditional chest X-ray analysis relies on physician experience, which can lead to subjectivity and misdiagnosis. To address this, we propose a novel pneumonia diagnosis method using the Fast-YOLO deep learning network that we introduced. First, we constructed a pneumonia dataset containing five categories and applied image enhancement techniques to increase data diversity and improve the model's generalization ability. Next, the YOLOv11 network structure was redesigned to accommodate the complex features of pneumonia X-ray images. By integrating the C3k2 module, DCNv2, and DynamicConv, the Fast-YOLO network effectively enhanced feature representation and reduced computational complexity (FPS increased from 53 to 120). Experimental results subsequently show that our method outperforms other commonly used detection models in terms of accuracy, recall, and mAP, offering better real-time detection capability and clinical application potential.</p>","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"19 ","pages":"1576438"},"PeriodicalIF":2.6,"publicationDate":"2025-04-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12077197/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144077440","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-scale image edge detection based on spatial-frequency domain interactive attention. 基于空频域交互关注的多尺度图像边缘检测。
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-28 eCollection Date: 2025-01-01 DOI: 10.3389/fnbot.2025.1550939
Yongfei Guo, Bo Li, Wenyue Zhang, Weilong Dong

Due to the many difficulties in accurately locating edges or boundaries in images of animals, plants, buildings, and the like with complex backgrounds, edge detection has become one of the most challenging tasks in the field of computer vision and is also a key step in many computer vision applications. Although existing deep learning-based methods can detect the edges of images relatively well, when the image background is rather complex and the key target is small, accurately detecting the edge of the main body and removing background interference remains a daunting task. Therefore, this paper proposes a multi-scale edge detection network based on spatial-frequency domain interactive attention, aiming to achieve accurate detection of the edge of the main target on multiple scales. The use of the spatial-frequency domain interactive attention module can not only perform significant edge extraction by filtering out some interference in the frequency domain. Moreover, by utilizing the interaction between the frequency domain and the spatial domain, edge features at different scales can be extracted and analyzed more accurately. The obtained results are superior to the current edge detection networks in terms of performance indicators and output image quality.

由于在动物、植物、建筑物等具有复杂背景的图像中难以准确定位边缘或边界,因此边缘检测已成为计算机视觉领域最具挑战性的任务之一,也是许多计算机视觉应用的关键步骤。虽然现有的基于深度学习的方法可以较好地检测图像的边缘,但当图像背景较为复杂,关键目标较小时,准确检测主体边缘并去除背景干扰仍然是一项艰巨的任务。为此,本文提出了一种基于空频域交互关注的多尺度边缘检测网络,旨在实现多尺度上主目标边缘的精确检测。利用空频域交互注意模块不仅可以通过滤除频域的一些干扰进行显著的边缘提取。此外,利用频域和空域的相互作用,可以更准确地提取和分析不同尺度的边缘特征。所得结果在性能指标和输出图像质量方面都优于当前的边缘检测网络。
{"title":"Multi-scale image edge detection based on spatial-frequency domain interactive attention.","authors":"Yongfei Guo, Bo Li, Wenyue Zhang, Weilong Dong","doi":"10.3389/fnbot.2025.1550939","DOIUrl":"https://doi.org/10.3389/fnbot.2025.1550939","url":null,"abstract":"<p><p>Due to the many difficulties in accurately locating edges or boundaries in images of animals, plants, buildings, and the like with complex backgrounds, edge detection has become one of the most challenging tasks in the field of computer vision and is also a key step in many computer vision applications. Although existing deep learning-based methods can detect the edges of images relatively well, when the image background is rather complex and the key target is small, accurately detecting the edge of the main body and removing background interference remains a daunting task. Therefore, this paper proposes a multi-scale edge detection network based on spatial-frequency domain interactive attention, aiming to achieve accurate detection of the edge of the main target on multiple scales. The use of the spatial-frequency domain interactive attention module can not only perform significant edge extraction by filtering out some interference in the frequency domain. Moreover, by utilizing the interaction between the frequency domain and the spatial domain, edge features at different scales can be extracted and analyzed more accurately. The obtained results are superior to the current edge detection networks in terms of performance indicators and output image quality.</p>","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"19 ","pages":"1550939"},"PeriodicalIF":2.6,"publicationDate":"2025-04-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12066664/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144015855","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Editorial: Biomedical signals and artificial intelligence towards smart robots control strategies. 社论:生物医学信号和人工智能对智能机器人控制策略的影响。
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-25 eCollection Date: 2025-01-01 DOI: 10.3389/fnbot.2025.1570870
Hassene Seddik, Hassen Fourati, Chiraz Ben Jabeur, Nahla Khraief
{"title":"Editorial: Biomedical signals and artificial intelligence towards smart robots control strategies.","authors":"Hassene Seddik, Hassen Fourati, Chiraz Ben Jabeur, Nahla Khraief","doi":"10.3389/fnbot.2025.1570870","DOIUrl":"https://doi.org/10.3389/fnbot.2025.1570870","url":null,"abstract":"","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"19 ","pages":"1570870"},"PeriodicalIF":2.6,"publicationDate":"2025-04-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12063127/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143996823","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Unmanned aerial vehicle based multi-person detection via deep neural network models. 基于深度神经网络模型的无人机多人检测。
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-17 eCollection Date: 2025-01-01 DOI: 10.3389/fnbot.2025.1582995
Mohammed Alshehri, Laiba Zahoor, Yahya AlQahtani, Abdulmonem Alshahrani, Dina Abdulaziz AlHammadi, Ahmad Jalal, Hui Liu

Introduction: Understanding human actions in complex environments is crucial for advancing applications in areas such as surveillance, robotics, and autonomous systems. Identifying actions from UAV-recorded videos becomes more challenging as the task presents unique challenges, including motion blur, dynamic background, lighting variations, and varying viewpoints. The presented work develops a deep learning system that recognizes multi-person behaviors from data gathered by UAVs. The proposed system provides higher recognition accuracy while maintaining robustness along with dynamic environmental adaptability through the integration of different features and neural network models. The study supports the wider development of neural network systems utilized in complicated contexts while creating intelligent UAV applications utilizing neural networks.

Method: The proposed study uses deep learning and feature extraction approaches to create a novel method to recognize various actions in UAV-recorded video. The proposed model improves identification capacities and system robustness by addressing motion dynamic problems and intricate environmental constraints, encouraging advancements in UAV-based neural network systems.

Results: We proposed a deep learning-based framework with feature extraction approaches that may effectively increase the accuracy and robustness of multi-person action recognition in the challenging scenarios. Compared to the existing approaches, our system achieved 91.50% on MOD20 dataset and 89.71% on Okutama-Action. These results do, in fact, show how useful neural network-based methods are for managing the limitations of UAV-based application.

Discussion: Results how that the proposed framework is indeed effective at multi-person action recognition under difficult UAV conditions.

简介:了解人类在复杂环境中的行为对于推进监控、机器人和自主系统等领域的应用至关重要。从无人机录制的视频中识别动作变得更具挑战性,因为任务提出了独特的挑战,包括运动模糊,动态背景,照明变化和不同的观点。本文开发了一种深度学习系统,该系统可以从无人机收集的数据中识别多人行为。该系统通过集成不同特征和神经网络模型,在保持鲁棒性和动态环境适应性的同时,提高了识别精度。该研究支持在复杂环境中使用神经网络系统的更广泛发展,同时利用神经网络创建智能无人机应用。方法:本研究采用深度学习和特征提取方法,创建了一种新的方法来识别无人机录制视频中的各种动作。该模型通过解决运动动态问题和复杂的环境约束,提高了识别能力和系统鲁棒性,促进了基于无人机的神经网络系统的发展。结果:我们提出了一个基于深度学习的框架和特征提取方法,可以有效地提高具有挑战性场景下多人动作识别的准确性和鲁棒性。与现有方法相比,我们的系统在MOD20数据集上的准确率为91.50%,在Okutama-Action数据集上的准确率为89.71%。事实上,这些结果确实表明,基于神经网络的方法对于管理基于无人机的应用的局限性是多么有用。讨论:结果表明所提出的框架在无人机困难条件下的多人动作识别是有效的。
{"title":"Unmanned aerial vehicle based multi-person detection via deep neural network models.","authors":"Mohammed Alshehri, Laiba Zahoor, Yahya AlQahtani, Abdulmonem Alshahrani, Dina Abdulaziz AlHammadi, Ahmad Jalal, Hui Liu","doi":"10.3389/fnbot.2025.1582995","DOIUrl":"https://doi.org/10.3389/fnbot.2025.1582995","url":null,"abstract":"<p><strong>Introduction: </strong>Understanding human actions in complex environments is crucial for advancing applications in areas such as surveillance, robotics, and autonomous systems. Identifying actions from UAV-recorded videos becomes more challenging as the task presents unique challenges, including motion blur, dynamic background, lighting variations, and varying viewpoints. The presented work develops a deep learning system that recognizes multi-person behaviors from data gathered by UAVs. The proposed system provides higher recognition accuracy while maintaining robustness along with dynamic environmental adaptability through the integration of different features and neural network models. The study supports the wider development of neural network systems utilized in complicated contexts while creating intelligent UAV applications utilizing neural networks.</p><p><strong>Method: </strong>The proposed study uses deep learning and feature extraction approaches to create a novel method to recognize various actions in UAV-recorded video. The proposed model improves identification capacities and system robustness by addressing motion dynamic problems and intricate environmental constraints, encouraging advancements in UAV-based neural network systems.</p><p><strong>Results: </strong>We proposed a deep learning-based framework with feature extraction approaches that may effectively increase the accuracy and robustness of multi-person action recognition in the challenging scenarios. Compared to the existing approaches, our system achieved 91.50% on MOD20 dataset and 89.71% on Okutama-Action. These results do, in fact, show how useful neural network-based methods are for managing the limitations of UAV-based application.</p><p><strong>Discussion: </strong>Results how that the proposed framework is indeed effective at multi-person action recognition under difficult UAV conditions.</p>","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"19 ","pages":"1582995"},"PeriodicalIF":2.6,"publicationDate":"2025-04-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12043872/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143997194","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HR-NeRF: advancing realism and accuracy in highlight scene representation. HR-NeRF:在高光场景表现中推进现实主义和准确性。
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-16 eCollection Date: 2025-01-01 DOI: 10.3389/fnbot.2025.1558948
Shufan Dai, Shanqin Wang

NeRF and its variants excel in novel view synthesis but struggle with scenes featuring specular highlights. To address this limitation, we introduce the Highlight Recovery Network (HRNet), a new architecture that enhances NeRF's ability to capture specular scenes. HRNet incorporates Swish activation functions, affine transformations, multilayer perceptrons (MLPs), and residual blocks, which collectively enable smooth non-linear transformations, adaptive feature scaling, and hierarchical feature extraction. The residual connections help mitigate the vanishing gradient problem, ensuring stable training. Despite the simplicity of HRNet's components, it achieves impressive results in recovering specular highlights. Additionally, a density voxel grid enhances model efficiency. Evaluations on four inward-facing benchmarks demonstrate that our approach outperforms NeRF and its variants, achieving a 3-5 dB PSNR improvement on each dataset while accurately capturing scene details. Furthermore, our method effectively preserves image details without requiring positional encoding, rendering a single scene in ~18 min on an NVIDIA RTX 3090 Ti GPU.

NeRF和它的变体擅长新颖的视图合成,但与具有镜面高光的场景斗争。为了解决这一限制,我们引入了高光恢复网络(HRNet),这是一种增强NeRF捕捉高光场景能力的新架构。HRNet结合了Swish激活函数、仿射变换、多层感知器(mlp)和残差块,它们共同实现了平滑的非线性变换、自适应特征缩放和分层特征提取。残差连接有助于缓解梯度消失问题,保证训练的稳定性。尽管HRNet的组件简单,但它在恢复镜面高光方面取得了令人印象深刻的结果。此外,密度体素网格提高了模型效率。对四个内向基准的评估表明,我们的方法优于NeRF及其变体,在准确捕获场景细节的同时,在每个数据集上实现了3-5 dB的PSNR改进。此外,我们的方法在不需要位置编码的情况下有效地保留了图像细节,在NVIDIA RTX 3090 Ti GPU上渲染单个场景约18分钟。
{"title":"HR-NeRF: advancing realism and accuracy in highlight scene representation.","authors":"Shufan Dai, Shanqin Wang","doi":"10.3389/fnbot.2025.1558948","DOIUrl":"https://doi.org/10.3389/fnbot.2025.1558948","url":null,"abstract":"<p><p>NeRF and its variants excel in novel view synthesis but struggle with scenes featuring specular highlights. To address this limitation, we introduce the Highlight Recovery Network (HRNet), a new architecture that enhances NeRF's ability to capture specular scenes. HRNet incorporates Swish activation functions, affine transformations, multilayer perceptrons (MLPs), and residual blocks, which collectively enable smooth non-linear transformations, adaptive feature scaling, and hierarchical feature extraction. The residual connections help mitigate the vanishing gradient problem, ensuring stable training. Despite the simplicity of HRNet's components, it achieves impressive results in recovering specular highlights. Additionally, a density voxel grid enhances model efficiency. Evaluations on four inward-facing benchmarks demonstrate that our approach outperforms NeRF and its variants, achieving a 3-5 dB PSNR improvement on each dataset while accurately capturing scene details. Furthermore, our method effectively preserves image details without requiring positional encoding, rendering a single scene in ~18 min on an NVIDIA RTX 3090 Ti GPU.</p>","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"19 ","pages":"1558948"},"PeriodicalIF":2.6,"publicationDate":"2025-04-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12041011/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143988032","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Editorial: Neural network models in autonomous robotics. 编辑:自主机器人中的神经网络模型。
IF 2.6 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-08 eCollection Date: 2025-01-01 DOI: 10.3389/fnbot.2025.1587137
Long Cheng, Ying Mao, Tomas Ward
{"title":"Editorial: Neural network models in autonomous robotics.","authors":"Long Cheng, Ying Mao, Tomas Ward","doi":"10.3389/fnbot.2025.1587137","DOIUrl":"https://doi.org/10.3389/fnbot.2025.1587137","url":null,"abstract":"","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"19 ","pages":"1587137"},"PeriodicalIF":2.6,"publicationDate":"2025-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12011765/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143987861","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Frontiers in Neurorobotics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1