首页 > 最新文献

Frontiers in Neurorobotics最新文献

英文 中文
YOLOv8-ACU: improved YOLOv8-pose for facial acupoint detection YOLOv8-ACU:用于面部穴位检测的改进型 YOLOv8-pose
IF 3.1 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-01-17 DOI: 10.3389/fnbot.2024.1355857
Zijian Yuan, Pengwei Shao, Jinran Li, Yinuo Wang, Zixuan Zhu, Weijie Qiu, Buqun Chen, Yan Tang, Aiqing Han
Introduction

Acupoint localization is integral to Traditional Chinese Medicine (TCM) acupuncture diagnosis and treatment. Employing intelligent detection models for recognizing facial acupoints can substantially enhance localization accuracy.

Methods

This study introduces an advancement in the YOLOv8-pose keypoint detection algorithm, tailored for facial acupoints, and named YOLOv8-ACU. This model enhances acupoint feature extraction by integrating ECA attention, replaces the original neck module with a lighter Slim-neck module, and improves the loss function for GIoU.

Results

The YOLOv8-ACU model achieves impressive accuracy, with an mAP@0.5 of 97.5% and an mAP@0.5–0.95 of 76.9% on our self-constructed datasets. It also marks a reduction in model parameters by 0.44M, model size by 0.82 MB, and GFLOPs by 9.3%.

Discussion

With its enhanced recognition accuracy and efficiency, along with good generalization ability, YOLOv8-ACU provides significant reference value for facial acupoint localization and detection. This is particularly beneficial for Chinese medicine practitioners engaged in facial acupoint research and intelligent detection.

引言 穴位定位是中医针灸诊断和治疗不可或缺的一部分。本研究介绍了 YOLOv8 姿态关键点检测算法的改进版,该算法专为面部穴位量身定制,并命名为 YOLOv8-ACU。该模型通过整合 ECA 注意加强了穴位特征提取,用更轻的 Slim-neck 模块取代了原来的颈部模块,并改进了 GIoU 的损失函数。结果 YOLOv8-ACU 模型取得了令人印象深刻的准确性,在我们自建的数据集上,其 mAP@0.5 的准确率为 97.5%,mAP@0.5-0.95 的准确率为 76.9%。讨论YOLOv8-ACU提高了识别精度和效率,并具有良好的泛化能力,为面部穴位定位和检测提供了重要的参考价值。这对从事面部穴位研究和智能检测的中医工作者尤其有益。
{"title":"YOLOv8-ACU: improved YOLOv8-pose for facial acupoint detection","authors":"Zijian Yuan, Pengwei Shao, Jinran Li, Yinuo Wang, Zixuan Zhu, Weijie Qiu, Buqun Chen, Yan Tang, Aiqing Han","doi":"10.3389/fnbot.2024.1355857","DOIUrl":"https://doi.org/10.3389/fnbot.2024.1355857","url":null,"abstract":"<sec><title>Introduction</title><p>Acupoint localization is integral to Traditional Chinese Medicine (TCM) acupuncture diagnosis and treatment. Employing intelligent detection models for recognizing facial acupoints can substantially enhance localization accuracy.</p></sec><sec><title>Methods</title><p>This study introduces an advancement in the YOLOv8-pose keypoint detection algorithm, tailored for facial acupoints, and named YOLOv8-ACU. This model enhances acupoint feature extraction by integrating ECA attention, replaces the original neck module with a lighter Slim-neck module, and improves the loss function for GIoU.</p></sec><sec><title>Results</title><p>The YOLOv8-ACU model achieves impressive accuracy, with an mAP@0.5 of 97.5% and an mAP@0.5–0.95 of 76.9% on our self-constructed datasets. It also marks a reduction in model parameters by 0.44M, model size by 0.82 MB, and GFLOPs by 9.3%.</p></sec><sec><title>Discussion</title><p>With its enhanced recognition accuracy and efficiency, along with good generalization ability, YOLOv8-ACU provides significant reference value for facial acupoint localization and detection. This is particularly beneficial for Chinese medicine practitioners engaged in facial acupoint research and intelligent detection.</p></sec>","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"26 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2024-01-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139658451","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Bidirectional feature pyramid attention-based temporal convolutional network model for motor imagery electroencephalogram classification 用于运动图像脑电图分类的基于注意力的双向特征金字塔时空卷积网络模型
IF 3.1 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-01-15 DOI: 10.3389/fnbot.2024.1343249
Xinghe Xie, Liyan Chen, Shujia Qin, Fusheng Zha, Xinggang Fan
Introduction

As an interactive method gaining popularity, brain-computer interfaces (BCIs) aim to facilitate communication between the brain and external devices. Among the various research topics in BCIs, the classification of motor imagery using electroencephalography (EEG) signals has the potential to greatly improve the quality of life for people with disabilities.

Methods

This technology assists them in controlling computers or other devices like prosthetic limbs, wheelchairs, and drones. However, the current performance of EEG signal decoding is not sufficient for real-world applications based on Motor Imagery EEG (MI-EEG). To address this issue, this study proposes an attention-based bidirectional feature pyramid temporal convolutional network model for the classification task of MI-EEG. The model incorporates a multi-head self-attention mechanism to weigh significant features in the MI-EEG signals. It also utilizes a temporal convolution network (TCN) to separate high-level temporal features. The signals are enhanced using the sliding-window technique, and channel and time-domain information of the MI-EEG signals is extracted through convolution.

Results

Additionally, a bidirectional feature pyramid structure is employed to implement attention mechanisms across different scales and multiple frequency bands of the MI-EEG signals. The performance of our model is evaluated on the BCI Competition IV-2a dataset and the BCI Competition IV-2b dataset, and the results showed that our model outperformed the state-of-the-art baseline model, with an accuracy of 87.5 and 86.3% for the subject-dependent, respectively.

Discussion

In conclusion, the BFATCNet model offers a novel approach for EEG-based motor imagery classification in BCIs, effectively capturing relevant features through attention mechanisms and temporal convolutional networks. Its superior performance on the BCI Competition IV-2a and IV-2b datasets highlights its potential for real-world applications. However, its performance on other datasets may vary, necessitating further research on data augmentation techniques and integration with multiple modalities to enhance interpretability and generalization. Additionally, reducing computational complexity for real-time applications is an important area for future work.

引言 脑机接口(BCIs)作为一种日益流行的互动方法,旨在促进大脑与外部设备之间的交流。在 BCIs 的各种研究课题中,利用脑电图(EEG)信号对运动图像进行分类有望大大提高残疾人的生活质量。然而,目前的脑电信号解码性能不足以满足基于运动图像脑电图(MI-EEG)的实际应用。为解决这一问题,本研究针对 MI-EEG 的分类任务提出了一种基于注意力的双向特征金字塔时空卷积网络模型。该模型采用多头自我注意机制来权衡 MI-EEG 信号中的重要特征。它还利用时空卷积网络(TCN)来分离高级时空特征。此外,该模型还采用了双向特征金字塔结构,以在 MI-EEG 信号的不同尺度和多个频段上实施注意机制。我们的模型在 BCI Competition IV-2a 数据集和 BCI Competition IV-2b 数据集上进行了性能评估,结果表明我们的模型优于最先进的基线模型,依赖主体的准确率分别为 87.5% 和 86.3%。讨论总之,BFATCNet 模型为基于脑电图的 BCI 运动图像分类提供了一种新方法,它通过注意机制和时序卷积网络有效捕捉相关特征。它在 BCI Competition IV-2a 和 IV-2b 数据集上的优异表现凸显了其在现实世界中的应用潜力。然而,它在其他数据集上的表现可能会有所不同,这就需要进一步研究数据增强技术和与多种模式的整合,以提高可解释性和通用性。此外,降低实时应用的计算复杂性也是未来工作的一个重要领域。
{"title":"Bidirectional feature pyramid attention-based temporal convolutional network model for motor imagery electroencephalogram classification","authors":"Xinghe Xie, Liyan Chen, Shujia Qin, Fusheng Zha, Xinggang Fan","doi":"10.3389/fnbot.2024.1343249","DOIUrl":"https://doi.org/10.3389/fnbot.2024.1343249","url":null,"abstract":"<sec><title>Introduction</title><p>As an interactive method gaining popularity, brain-computer interfaces (BCIs) aim to facilitate communication between the brain and external devices. Among the various research topics in BCIs, the classification of motor imagery using electroencephalography (EEG) signals has the potential to greatly improve the quality of life for people with disabilities.</p></sec><sec><title>Methods</title><p>This technology assists them in controlling computers or other devices like prosthetic limbs, wheelchairs, and drones. However, the current performance of EEG signal decoding is not sufficient for real-world applications based on Motor Imagery EEG (MI-EEG). To address this issue, this study proposes an attention-based bidirectional feature pyramid temporal convolutional network model for the classification task of MI-EEG. The model incorporates a multi-head self-attention mechanism to weigh significant features in the MI-EEG signals. It also utilizes a temporal convolution network (TCN) to separate high-level temporal features. The signals are enhanced using the sliding-window technique, and channel and time-domain information of the MI-EEG signals is extracted through convolution.</p></sec><sec><title>Results</title><p>Additionally, a bidirectional feature pyramid structure is employed to implement attention mechanisms across different scales and multiple frequency bands of the MI-EEG signals. The performance of our model is evaluated on the BCI Competition IV-2a dataset and the BCI Competition IV-2b dataset, and the results showed that our model outperformed the state-of-the-art baseline model, with an accuracy of 87.5 and 86.3% for the subject-dependent, respectively.</p></sec><sec><title>Discussion</title><p>In conclusion, the BFATCNet model offers a novel approach for EEG-based motor imagery classification in BCIs, effectively capturing relevant features through attention mechanisms and temporal convolutional networks. Its superior performance on the BCI Competition IV-2a and IV-2b datasets highlights its potential for real-world applications. However, its performance on other datasets may vary, necessitating further research on data augmentation techniques and integration with multiple modalities to enhance interpretability and generalization. Additionally, reducing computational complexity for real-time applications is an important area for future work.</p></sec>","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"123 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2024-01-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139581115","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing hazardous material vehicle detection with advanced feature enhancement modules using HMV-YOLO 利用 HMV-YOLO 高级特征增强模块加强危险品车辆检测
IF 3.1 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-01-15 DOI: 10.3389/fnbot.2024.1351939
Ling Wang, Bushi Liu, Wei Shao, Zhe Li, Kailu Chang, Wenjie Zhu

The transportation of hazardous chemicals on roadways has raised significant safety concerns. Incidents involving these substances often lead to severe and devastating consequences. Consequently, there is a pressing need for real-time detection systems tailored for hazardous material vehicles. However, existing detection methods face challenges in accurately identifying smaller targets and achieving high precision. This paper introduces a novel solution, HMV-YOLO, an enhancement of the YOLOv7-tiny model designed to address these challenges. Within this model, two innovative modules, CBSG and G-ELAN, are introduced. The CBSG module's mathematical model incorporates components such as Convolution (Conv2d), Batch Normalization (BN), SiLU activation, and Global Response Normalization (GRN) to mitigate feature collapse issues and enhance neuron activity. The G-ELAN module, building upon CBSG, further advances feature fusion. Experimental results showcase the superior performance of the enhanced model compared to the original one across various evaluation metrics. This advancement shows great promise for practical applications, particularly in the context of real-time monitoring systems for hazardous material vehicles.

在公路上运输危险化学品引发了重大的安全问题。涉及这些物质的事故往往会导致严重的破坏性后果。因此,迫切需要为危险品车辆量身定制实时检测系统。然而,现有的检测方法在准确识别较小目标和实现高精度方面面临挑战。本文介绍了一种新型解决方案 HMV-YOLO,它是 YOLOv7-tiny 模型的增强版,旨在应对这些挑战。在该模型中,引入了两个创新模块:CBSG 和 G-ELAN。CBSG 模块的数学模型融合了卷积(Conv2d)、批量归一化(BN)、SiLU 激活和全局响应归一化(GRN)等组件,以缓解特征坍塌问题并增强神经元活动。G-ELAN 模块以 CBSG 为基础,进一步推进了特征融合。实验结果表明,与原始模型相比,增强型模型在各种评估指标上都表现出色。这一进步为实际应用,尤其是危险品车辆实时监控系统的应用带来了巨大希望。
{"title":"Enhancing hazardous material vehicle detection with advanced feature enhancement modules using HMV-YOLO","authors":"Ling Wang, Bushi Liu, Wei Shao, Zhe Li, Kailu Chang, Wenjie Zhu","doi":"10.3389/fnbot.2024.1351939","DOIUrl":"https://doi.org/10.3389/fnbot.2024.1351939","url":null,"abstract":"<p>The transportation of hazardous chemicals on roadways has raised significant safety concerns. Incidents involving these substances often lead to severe and devastating consequences. Consequently, there is a pressing need for real-time detection systems tailored for hazardous material vehicles. However, existing detection methods face challenges in accurately identifying smaller targets and achieving high precision. This paper introduces a novel solution, HMV-YOLO, an enhancement of the YOLOv7-tiny model designed to address these challenges. Within this model, two innovative modules, CBSG and G-ELAN, are introduced. The CBSG module's mathematical model incorporates components such as Convolution (Conv2d), Batch Normalization (BN), SiLU activation, and Global Response Normalization (GRN) to mitigate feature collapse issues and enhance neuron activity. The G-ELAN module, building upon CBSG, further advances feature fusion. Experimental results showcase the superior performance of the enhanced model compared to the original one across various evaluation metrics. This advancement shows great promise for practical applications, particularly in the context of real-time monitoring systems for hazardous material vehicles.</p>","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"14 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2024-01-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139581112","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Re-framing bio-plausible collision detection: identifying shared meta-properties through strategic prototyping 重新构建生物拟真碰撞检测:通过战略原型设计确定共享元特性
IF 3.1 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-01-12 DOI: 10.3389/fnbot.2024.1349498
Haotian Wu, Shigang Yue, Cheng Hu

Insects exhibit remarkable abilities in navigating complex natural environments, whether it be evading predators, capturing prey, or seeking out con-specifics, all of which rely on their compact yet reliable neural systems. We explore the field of bio-inspired robotic vision systems, focusing on the locust inspired Lobula Giant Movement Detector (LGMD) models. The existing LGMD models are thoroughly evaluated, identifying their common meta-properties that are essential for their functionality. This article reveals a common framework, characterized by layered structures and computational strategies, which is crucial for enhancing the capability of bio-inspired models for diverse applications. The result of this analysis is the Strategic Prototype, which embodies the identified meta-properties. It represents a modular and more flexible method for developing more responsive and adaptable robotic visual systems. The perspective highlights the potential of the Strategic Prototype: LGMD-Universally Prototype (LGMD-UP), the key to re-framing LGMD models and advancing our understanding and implementation of bio-inspired visual systems in robotics. It might open up more flexible and adaptable avenues for research and practical applications.

昆虫在复杂的自然环境中表现出非凡的导航能力,无论是躲避捕食者、捕捉猎物还是寻找同类,所有这些都依赖于它们紧凑而可靠的神经系统。我们探索了生物启发机器人视觉系统领域,重点是受蝗虫启发的小叶巨型运动探测器(LGMD)模型。我们对现有的 LGMD 模型进行了全面评估,找出了其功能所必需的共同元特性。这篇文章揭示了一个以分层结构和计算策略为特征的共同框架,这对于提高生物启发模型在各种应用中的能力至关重要。这一分析的结果就是战略原型,它体现了已确定的元特性。它代表了一种模块化、更灵活的方法,可用于开发反应更灵敏、适应性更强的机器人视觉系统。该视角强调了战略原型的潜力:LGMD-UP(LGMD-Universally Prototype)是重新构建 LGMD 模型的关键,也是推进我们对机器人生物启发视觉系统的理解和实施的关键。它可以为研究和实际应用开辟更加灵活、适应性更强的途径。
{"title":"Re-framing bio-plausible collision detection: identifying shared meta-properties through strategic prototyping","authors":"Haotian Wu, Shigang Yue, Cheng Hu","doi":"10.3389/fnbot.2024.1349498","DOIUrl":"https://doi.org/10.3389/fnbot.2024.1349498","url":null,"abstract":"<p>Insects exhibit remarkable abilities in navigating complex natural environments, whether it be evading predators, capturing prey, or seeking out con-specifics, all of which rely on their compact yet reliable neural systems. We explore the field of bio-inspired robotic vision systems, focusing on the locust inspired Lobula Giant Movement Detector (LGMD) models. The existing LGMD models are thoroughly evaluated, identifying their common meta-properties that are essential for their functionality. This article reveals a common framework, characterized by layered structures and computational strategies, which is crucial for enhancing the capability of bio-inspired models for diverse applications. The result of this analysis is the Strategic Prototype, which embodies the identified meta-properties. It represents a modular and more flexible method for developing more responsive and adaptable robotic visual systems. The perspective highlights the potential of the Strategic Prototype: LGMD-Universally Prototype (LGMD-UP), the key to re-framing LGMD models and advancing our understanding and implementation of bio-inspired visual systems in robotics. It might open up more flexible and adaptable avenues for research and practical applications.</p>","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"330 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2024-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139559140","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Editorial: Neurorobotics and strategies for adaptive human-machine interaction, volume II 社论:神经机器人与自适应人机交互战略》第二卷
IF 3.1 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-01-09 DOI: 10.3389/fnbot.2023.1354389
F. Cordella, S. Soekadar, L. Zollo
{"title":"Editorial: Neurorobotics and strategies for adaptive human-machine interaction, volume II","authors":"F. Cordella, S. Soekadar, L. Zollo","doi":"10.3389/fnbot.2023.1354389","DOIUrl":"https://doi.org/10.3389/fnbot.2023.1354389","url":null,"abstract":"","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"49 3","pages":""},"PeriodicalIF":3.1,"publicationDate":"2024-01-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139444898","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Editorial: Safety and security of robotic systems: intelligent algorithms 社论:机器人系统的安全保障:智能算法
IF 3.1 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-01-08 DOI: 10.3389/fnbot.2023.1342742
Chengwei Wu
{"title":"Editorial: Safety and security of robotic systems: intelligent algorithms","authors":"Chengwei Wu","doi":"10.3389/fnbot.2023.1342742","DOIUrl":"https://doi.org/10.3389/fnbot.2023.1342742","url":null,"abstract":"","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"26 13","pages":""},"PeriodicalIF":3.1,"publicationDate":"2024-01-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139445254","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Context-aware SAR image ship detection and recognition network 情境感知合成孔径雷达图像船舶检测和识别网络
IF 3.1 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-01-03 DOI: 10.3389/fnbot.2024.1293992
Chao Li, Chenke Yue, Hanfu Li, Zhile Wang

With the development of deep learning, synthetic aperture radar (SAR) ship detection and recognition based on deep learning have gained widespread application and advancement. However, there are still challenging issues, manifesting in two primary facets: firstly, the imaging mechanism of SAR results in significant noise interference, making it difficult to separate background noise from ship target features in complex backgrounds such as ports and urban areas; secondly, the heterogeneous scales of ship target features result in the susceptibility of smaller targets to information loss, rendering them elusive to detection. In this article, we propose a context-aware one-stage ship detection network that exhibits heightened sensitivity to scale variations and robust resistance to noise interference. Then we introduce a Local feature refinement module (LFRM), which utilizes multiple receptive fields of different sizes to extract local multi-scale information, followed by a two-branch channel-wise attention approach to obtain local cross-channel interactions. To minimize the effect of a complex background on the target, we design the global context aggregation module (GCAM) to enhance the feature representation of the target and suppress the interference of noise by acquiring long-range dependencies. Finally, we validate the effectiveness of our method on three publicly available SAR ship detection datasets, SAR-Ship-Dataset, high-resolution SAR images dataset (HRSID), and SAR ship detection dataset (SSDD). The experimental results show that our method is more competitive, with AP50s of 96.3, 93.3, and 96.2% on the three publicly available datasets, respectively.

随着深度学习的发展,基于深度学习的合成孔径雷达(SAR)船舶探测与识别技术获得了广泛的应用和进步。然而,目前仍存在一些具有挑战性的问题,主要表现在两个方面:一是合成孔径雷达的成像机制会产生明显的噪声干扰,使得在港口和城市等复杂背景下难以将背景噪声与船舶目标特征分离开来;二是船舶目标特征的异构尺度导致较小的目标容易受到信息丢失的影响,使其难以被检测到。在本文中,我们提出了一种上下文感知的单级船舶检测网络,该网络对尺度变化表现出更高的灵敏度,并具有强大的抗噪声干扰能力。然后,我们引入了局部特征细化模块(LFRM),该模块利用多个不同大小的感受野来提取局部多尺度信息,然后采用双分支通道关注方法来获取局部跨通道交互信息。为了尽量减少复杂背景对目标的影响,我们设计了全局上下文聚合模块(GCAM),通过获取长程依赖关系来增强目标的特征表示并抑制噪声干扰。最后,我们在三个公开的 SAR 船舶检测数据集(SAR-Ship-Dataset)、高分辨率 SAR 图像数据集(HRSID)和 SAR 船舶检测数据集(SSDD)上验证了我们的方法的有效性。实验结果表明,我们的方法更具竞争力,在三个公开数据集上的 AP50 分别为 96.3%、93.3% 和 96.2%。
{"title":"Context-aware SAR image ship detection and recognition network","authors":"Chao Li, Chenke Yue, Hanfu Li, Zhile Wang","doi":"10.3389/fnbot.2024.1293992","DOIUrl":"https://doi.org/10.3389/fnbot.2024.1293992","url":null,"abstract":"<p>With the development of deep learning, synthetic aperture radar (SAR) ship detection and recognition based on deep learning have gained widespread application and advancement. However, there are still challenging issues, manifesting in two primary facets: firstly, the imaging mechanism of SAR results in significant noise interference, making it difficult to separate background noise from ship target features in complex backgrounds such as ports and urban areas; secondly, the heterogeneous scales of ship target features result in the susceptibility of smaller targets to information loss, rendering them elusive to detection. In this article, we propose a context-aware one-stage ship detection network that exhibits heightened sensitivity to scale variations and robust resistance to noise interference. Then we introduce a Local feature refinement module (LFRM), which utilizes multiple receptive fields of different sizes to extract local multi-scale information, followed by a two-branch channel-wise attention approach to obtain local cross-channel interactions. To minimize the effect of a complex background on the target, we design the global context aggregation module (GCAM) to enhance the feature representation of the target and suppress the interference of noise by acquiring long-range dependencies. Finally, we validate the effectiveness of our method on three publicly available SAR ship detection datasets, SAR-Ship-Dataset, high-resolution SAR images dataset (HRSID), and SAR ship detection dataset (SSDD). The experimental results show that our method is more competitive, with AP50s of 96.3, 93.3, and 96.2% on the three publicly available datasets, respectively.</p>","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"83 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2024-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139476612","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ID-YOLOv7: an efficient method for insulator defect detection in power distribution network ID-YOLOv7:配电网绝缘子缺陷检测的高效方法
IF 3.1 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-12-31 DOI: 10.3389/fnbot.2023.1331427
Bojian Chen, Weihao Zhang, Wenbin Wu, Yiran Li, Zhuolei Chen, Chenglong Li

Insulators play a pivotal role in the reliability of power distribution networks, necessitating precise defect detection. However, compared with aerial insulator images of transmission network, insulator images of power distribution network contain more complex backgrounds and subtle insulator defects, it leads to high false detection rates and omission rates in current mainstream detection algorithms. In response, this study presents ID-YOLOv7, a tailored convolutional neural network. First, we design a novel Edge Detailed Shape Data Augmentation (EDSDA) method to enhance the model's sensitivity to insulator's edge shapes. Meanwhile, a Cross-Channel and Spatial Multi-Scale Attention (CCSMA) module is proposed, which can interactively model across different channels and spatial domains, to augment the network's attention to high-level insulator defect features. Second, we design a Re-BiC module to fuse multi-scale contextual features and reconstruct the Neck component, alleviating the issue of critical feature loss during inter-feature layer interaction in traditional FPN structures. Finally, we utilize the MPDIoU function to calculate the model's localization loss, effectively reducing redundant computational costs. We perform comprehensive experiments using the Su22kV_broken and PASCAL VOC 2007 datasets to validate our algorithm's effectiveness. On the Su22kV_broken dataset, our approach attains an 85.7% mAP on a single NVIDIA RTX 2080ti graphics card, marking a 7.2% increase over the original YOLOv7. On the PASCAL VOC 2007 dataset, we achieve an impressive 90.3% mAP at a processing speed of 53 FPS, showing a 2.9% improvement compared to the original YOLOv7.

绝缘子对配电网络的可靠性起着举足轻重的作用,因此需要精确的缺陷检测。然而,与输电网络的空中绝缘子图像相比,配电网络的绝缘子图像包含更复杂的背景和更细微的绝缘子缺陷,这导致当前主流检测算法的误检率和漏检率较高。为此,本研究提出了一种量身定制的卷积神经网络 ID-YOLOv7。首先,我们设计了一种新颖的边缘详细形状数据增强(EDSDA)方法,以提高模型对绝缘体边缘形状的灵敏度。同时,我们还提出了跨信道和空间多尺度关注(CCSMA)模块,该模块可以跨信道和空间域交互建模,以增强网络对高层次绝缘体缺陷特征的关注。其次,我们设计了一个 Re-BiC 模块,用于融合多尺度上下文特征并重建内克分量,从而缓解了传统 FPN 结构中特征层间交互过程中关键特征丢失的问题。最后,我们利用 MPDIoU 函数计算模型的定位损失,有效降低了冗余计算成本。我们使用 Su22kV_broken 和 PASCAL VOC 2007 数据集进行了综合实验,以验证我们算法的有效性。在 Su22kV_broken 数据集上,我们的方法在单 NVIDIA RTX 2080ti 显卡上实现了 85.7% 的 mAP,比原始 YOLOv7 提高了 7.2%。在 PASCAL VOC 2007 数据集上,我们以 53 FPS 的处理速度实现了令人印象深刻的 90.3% mAP,与原始 YOLOv7 相比提高了 2.9%。
{"title":"ID-YOLOv7: an efficient method for insulator defect detection in power distribution network","authors":"Bojian Chen, Weihao Zhang, Wenbin Wu, Yiran Li, Zhuolei Chen, Chenglong Li","doi":"10.3389/fnbot.2023.1331427","DOIUrl":"https://doi.org/10.3389/fnbot.2023.1331427","url":null,"abstract":"<p>Insulators play a pivotal role in the reliability of power distribution networks, necessitating precise defect detection. However, compared with aerial insulator images of transmission network, insulator images of power distribution network contain more complex backgrounds and subtle insulator defects, it leads to high false detection rates and omission rates in current mainstream detection algorithms. In response, this study presents ID-YOLOv7, a tailored convolutional neural network. First, we design a novel Edge Detailed Shape Data Augmentation (EDSDA) method to enhance the model's sensitivity to insulator's edge shapes. Meanwhile, a Cross-Channel and Spatial Multi-Scale Attention (CCSMA) module is proposed, which can interactively model across different channels and spatial domains, to augment the network's attention to high-level insulator defect features. Second, we design a Re-BiC module to fuse multi-scale contextual features and reconstruct the Neck component, alleviating the issue of critical feature loss during inter-feature layer interaction in traditional FPN structures. Finally, we utilize the MPDIoU function to calculate the model's localization loss, effectively reducing redundant computational costs. We perform comprehensive experiments using the Su22kV_broken and PASCAL VOC 2007 datasets to validate our algorithm's effectiveness. On the Su22kV_broken dataset, our approach attains an 85.7% mAP on a single NVIDIA RTX 2080ti graphics card, marking a 7.2% increase over the original YOLOv7. On the PASCAL VOC 2007 dataset, we achieve an impressive 90.3% mAP at a processing speed of 53 FPS, showing a 2.9% improvement compared to the original YOLOv7.</p>","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"8 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2023-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139469309","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Loop closure detection of visual SLAM based on variational autoencoder 基于变异自动编码器的视觉 SLAM 的闭环检测
IF 3.1 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-12-26 DOI: 10.3389/fnbot.2023.1301785
Shibin Song, Fengjie Yu, Xiaojie Jiang, Jie Zhu, Weihao Cheng, Xiao Fang

Loop closure detection is an important module for simultaneous localization and mapping (SLAM). Correct detection of loops can reduce the cumulative drift in positioning. Because traditional detection methods rely on handicraft features, false positive detections can occur when the environment changes, resulting in incorrect estimates and an inability to obtain accurate maps. In this research paper, a loop closure detection method based on a variational autoencoder (VAE) is proposed. It is intended to be used as a feature extractor to extract image features through neural networks to replace the handicraft features used in traditional methods. This method extracts a low-dimensional vector as the representation of the image. At the same time, the attention mechanism is added to the network and constraints are added to improve the loss function for better image representation. In the back-end feature matching process, geometric checking is used to filter out the wrong matching for the false positive problem. Finally, through numerical experiments, the proposed method is demonstrated to have a better precision-recall curve than the traditional method of the bag-of-words model and other deep learning methods and is highly robust to environmental changes. In addition, experiments on datasets from three different scenarios also demonstrate that the method can be applied in real-world scenarios and that it has a good performance.

环路闭合检测是同步定位和绘图(SLAM)的一个重要模块。正确的环路检测可以减少定位的累积漂移。由于传统的检测方法依赖于手工特征,当环境发生变化时,可能会出现假阳性检测,从而导致估计错误,无法获得准确的地图。本研究论文提出了一种基于变异自动编码器(VAE)的环路闭合检测方法。它旨在作为一种特征提取器,通过神经网络提取图像特征,以取代传统方法中使用的手工特征。该方法提取低维向量作为图像的表示。同时,在网络中加入注意力机制,并加入约束条件来改进损失函数,以获得更好的图像表示。在后端特征匹配过程中,利用几何校验过滤掉错误的匹配,以解决假阳性问题。最后,通过数值实验证明,与传统的字袋模型方法和其他深度学习方法相比,所提出的方法具有更好的精度-召回曲线,并且对环境变化具有很强的鲁棒性。此外,在三种不同场景的数据集上进行的实验也证明了该方法可应用于实际场景,并具有良好的性能。
{"title":"Loop closure detection of visual SLAM based on variational autoencoder","authors":"Shibin Song, Fengjie Yu, Xiaojie Jiang, Jie Zhu, Weihao Cheng, Xiao Fang","doi":"10.3389/fnbot.2023.1301785","DOIUrl":"https://doi.org/10.3389/fnbot.2023.1301785","url":null,"abstract":"<p>Loop closure detection is an important module for simultaneous localization and mapping (SLAM). Correct detection of loops can reduce the cumulative drift in positioning. Because traditional detection methods rely on handicraft features, false positive detections can occur when the environment changes, resulting in incorrect estimates and an inability to obtain accurate maps. In this research paper, a loop closure detection method based on a variational autoencoder (VAE) is proposed. It is intended to be used as a feature extractor to extract image features through neural networks to replace the handicraft features used in traditional methods. This method extracts a low-dimensional vector as the representation of the image. At the same time, the attention mechanism is added to the network and constraints are added to improve the loss function for better image representation. In the back-end feature matching process, geometric checking is used to filter out the wrong matching for the false positive problem. Finally, through numerical experiments, the proposed method is demonstrated to have a better precision-recall curve than the traditional method of the bag-of-words model and other deep learning methods and is highly robust to environmental changes. In addition, experiments on datasets from three different scenarios also demonstrate that the method can be applied in real-world scenarios and that it has a good performance.</p>","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"84 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2023-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139499694","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-UAV simultaneous target assignment and path planning based on deep reinforcement learning in dynamic multiple obstacles environments 动态多障碍环境中基于深度强化学习的多无人机同步目标分配和路径规划
IF 3.1 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2023-12-26 DOI: 10.3389/fnbot.2023.1302898
Xiaoran Kong, Yatong Zhou, Zhe Li, Shaohai Wang

Target assignment and path planning are crucial for the cooperativity of multiple unmanned aerial vehicles (UAV) systems. However, it is a challenge considering the dynamics of environments and the partial observability of UAVs. In this article, the problem of multi-UAV target assignment and path planning is formulated as a partially observable Markov decision process (POMDP), and a novel deep reinforcement learning (DRL)-based algorithm is proposed to address it. Specifically, a target assignment network is introduced into the twin-delayed deep deterministic policy gradient (TD3) algorithm to solve the target assignment problem and path planning problem simultaneously. The target assignment network executes target assignment for each step of UAVs, while the TD3 guides UAVs to plan paths for this step based on the assignment result and provides training labels for the optimization of the target assignment network. Experimental results demonstrate that the proposed approach can ensure an optimal complete target allocation and achieve a collision-free path for each UAV in three-dimensional (3D) dynamic multiple-obstacle environments, and present a superior performance in target completion and a better adaptability to complex environments compared with existing methods.

目标分配和路径规划对于多个无人飞行器(UAV)系统的协同工作至关重要。然而,考虑到环境的动态性和无人飞行器的部分可观测性,这是一个挑战。本文将多无人机目标分配和路径规划问题表述为部分可观测马尔可夫决策过程(POMDP),并提出了一种基于深度强化学习(DRL)的新型算法来解决该问题。具体来说,在双延迟深度确定性策略梯度(TD3)算法中引入了目标分配网络,以同时解决目标分配问题和路径规划问题。目标分配网络为无人机的每一步执行目标分配,而 TD3 则根据分配结果引导无人机规划这一步的路径,并为目标分配网络的优化提供训练标签。实验结果表明,在三维(3D)动态多障碍物环境中,所提出的方法能确保每个无人机获得最佳的完整目标分配并实现无碰撞路径,与现有方法相比,在完成目标方面性能更优,对复杂环境的适应性更好。
{"title":"Multi-UAV simultaneous target assignment and path planning based on deep reinforcement learning in dynamic multiple obstacles environments","authors":"Xiaoran Kong, Yatong Zhou, Zhe Li, Shaohai Wang","doi":"10.3389/fnbot.2023.1302898","DOIUrl":"https://doi.org/10.3389/fnbot.2023.1302898","url":null,"abstract":"<p>Target assignment and path planning are crucial for the cooperativity of multiple unmanned aerial vehicles (UAV) systems. However, it is a challenge considering the dynamics of environments and the partial observability of UAVs. In this article, the problem of multi-UAV target assignment and path planning is formulated as a partially observable Markov decision process (POMDP), and a novel deep reinforcement learning (DRL)-based algorithm is proposed to address it. Specifically, a target assignment network is introduced into the twin-delayed deep deterministic policy gradient (TD3) algorithm to solve the target assignment problem and path planning problem simultaneously. The target assignment network executes target assignment for each step of UAVs, while the TD3 guides UAVs to plan paths for this step based on the assignment result and provides training labels for the optimization of the target assignment network. Experimental results demonstrate that the proposed approach can ensure an optimal complete target allocation and achieve a collision-free path for each UAV in three-dimensional (3D) dynamic multiple-obstacle environments, and present a superior performance in target completion and a better adaptability to complex environments compared with existing methods.</p>","PeriodicalId":12628,"journal":{"name":"Frontiers in Neurorobotics","volume":"8 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2023-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139515443","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Frontiers in Neurorobotics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1