首页 > 最新文献

Journal of Field Robotics最新文献

英文 中文
Dynamic Modeling and Adaptive Fuzzy Control of Spatial Deformation Pneumatic Leg Actuators With Error Constraints 带误差约束的空间变形气动腿动器动力学建模与自适应模糊控制
IF 5.2 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-09-01 DOI: 10.1002/rob.70055
Zehao Qiu, Qingxiang Wu, Zhuoqing Liu, Yongchun Fang, Ning Sun

With the development of soft crawling robots, pneumatic soft actuators (PSAs) with complicated nonplanar structures are increasingly designed due to the capabilities of achieving compound robot movements. However, the current deformation accuracy of complex PSAs is insufficient to satisfy application requirements. Specifically, the dynamic properties of complex PSAs are still ambiguous and PSA deformation control methods remain lacking, which are caused by inherent characteristics, such as strong nonlinearity and hysteresis. To this end, based on self-fabricated spatial deformation pneumatic leg actuators (SDPLAs), an equivalent dynamic model derived from Euler–Lagrange equations and a modified hysteresis model are established to form the new complete model of SDPLA systems, which accurately describes the bending deformation and hysteresis of SDPLAs at the same time. Furthermore, a novel model-based adaptive fuzzy tracking controller is designed for SDPLAs, which addresses model uncertainties and unknown external torque, and achieves the accurate bending of each SDPLA part. Subsequently, the closed-loop stability is rigorously proven by the Lyapunov theory. Finally, a series of experiments validates the effectiveness of the established dynamic and hysteresis models, the tracking control performance on time-varying trajectories with different amplitudes, frequencies, and shapes, and the control robustness against disturbances.

随着软爬行机器人的发展,复杂非平面结构的气动软执行器(PSAs)被越来越多地设计出来,以实现机器人的复合运动。然而,目前复杂psa的变形精度还不能满足应用要求。具体而言,由于其固有的非线性和迟滞等特性,复杂PSA的动态特性仍然不明确,且缺乏变形控制方法。为此,基于自制空间变形气动腿致动器(SDPLA),建立了由欧拉-拉格朗日方程推导的等效动力学模型和修正的滞回模型,形成了新的完整的SDPLA系统模型,该模型能准确地同时描述SDPLA的弯曲变形和滞回。在此基础上,设计了一种基于模型的自适应模糊跟踪控制器,解决了模型不确定性和未知的外部转矩问题,实现了SDPLA各部件的精确弯曲。随后,用李亚普诺夫理论严格证明了闭环的稳定性。最后,通过一系列实验验证了所建立的动态和滞后模型的有效性,以及对不同幅度、频率和形状的时变轨迹的跟踪控制性能,以及对干扰的鲁棒性。
{"title":"Dynamic Modeling and Adaptive Fuzzy Control of Spatial Deformation Pneumatic Leg Actuators With Error Constraints","authors":"Zehao Qiu,&nbsp;Qingxiang Wu,&nbsp;Zhuoqing Liu,&nbsp;Yongchun Fang,&nbsp;Ning Sun","doi":"10.1002/rob.70055","DOIUrl":"https://doi.org/10.1002/rob.70055","url":null,"abstract":"<div>\u0000 \u0000 <p>With the development of soft crawling robots, pneumatic soft actuators (PSAs) with complicated nonplanar structures are increasingly designed due to the capabilities of achieving compound robot movements. However, the current deformation accuracy of complex PSAs is insufficient to satisfy application requirements. Specifically, the dynamic properties of complex PSAs are still ambiguous and PSA deformation control methods remain lacking, which are caused by inherent characteristics, such as strong nonlinearity and hysteresis. To this end, based on self-fabricated spatial deformation pneumatic leg actuators (SDPLAs), an equivalent dynamic model derived from Euler–Lagrange equations and a modified hysteresis model are established to form the <i>new</i> complete model of SDPLA systems, which accurately describes the bending deformation and hysteresis of SDPLAs at the same time. Furthermore, a <i>novel</i> model-based adaptive fuzzy tracking controller is designed for SDPLAs, which addresses model uncertainties and unknown external torque, and achieves the accurate bending of each SDPLA part. Subsequently, the closed-loop stability is rigorously proven by the Lyapunov theory. Finally, a series of experiments validates the effectiveness of the established dynamic and hysteresis models, the tracking control performance on time-varying trajectories with different amplitudes, frequencies, and shapes, and the control robustness against disturbances.</p>\u0000 </div>","PeriodicalId":192,"journal":{"name":"Journal of Field Robotics","volume":"43 2","pages":"622-635"},"PeriodicalIF":5.2,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146139126","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Part I—Modernizing Nemabot: AI-Supported Identification of Frankliniella occidentalis Damage for Enhanced Biological Control Efficiency 第一部分现代化Nemabot:人工智能支持的西富兰克林菌危害识别提高生物防治效率
IF 5.2 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-09-01 DOI: 10.1002/rob.70068
Hilal Erdoğan, Atilla Erdinç, Alperen Kaan Bütüner, Tufan Can Ulu, İ. Alper Susurluk, Edwin E. Lewis, Halil Ünal

Western Flower Thrips (Frankliniella occidentalis) is a significant agricultural pest causing substantial economic losses by damaging crops and acting as a vector for plant diseases. Traditional pest control methods relying on chemical pesticides pose environmental and health risks, necessitating alternative solutions. Entomopathogenic nematodes (EPNs) have emerged as a promising biological control agent. This study presents an AI-supported precision application system, Nemabot, designed to optimize EPN deployment based on thrips-induced bean leaf damage. In this study, agricultural disease detection was performed using the Multi-Otsu Thresholding method integrated into deep learning-based object detection and segmentation algorithms. The developed method enhances segmentation accuracy through image processing techniques, thereby increasing the precision in identifying infested regions. The model used in the study was optimized with a YOLO-based architecture during training and reinforced with various data augmentation techniques for segmenting bean leaves. The model's performance evaluation yielded mAP0.5 values of B: 0.9481 and M: 0.94981, while mAP0.5:0.95 values were B: 0.90887 and M: 0.90887. The precision and recall values were 1.0 and 0.99975, respectively, indicating the model's high sensitivity. Additionally, the low values of box_loss, segmentation_loss, and objectness_loss demonstrate that the model maintains a minimal error rate. The proposed approach offers higher accuracy and sensitivity than conventional segmentation methods, contributing significantly to agricultural disease detection applications.

西花蓟马(Frankliniella occidentalis)是一种重要的农业害虫,通过破坏作物和作为植物疾病的媒介,造成巨大的经济损失。依靠化学农药的传统虫害防治方法对环境和健康构成风险,需要替代解决方案。昆虫病原线虫(EPNs)已成为一种很有前途的生物防治剂。本研究提出了一个人工智能支持的精确应用系统Nemabot,旨在优化基于蓟马诱导的豆类叶片损伤的EPN部署。在本研究中,将Multi-Otsu阈值方法与基于深度学习的目标检测和分割算法相结合,进行农业病害检测。该方法通过图像处理技术提高了分割精度,从而提高了识别侵染区域的精度。研究中使用的模型在训练期间使用基于yolo的架构进行优化,并使用各种数据增强技术进行增强,以分割豆类叶片。模型性能评价的mAP0.5值为B: 0.9481和M: 0.94981, mAP0.5:0.95值为B: 0.90887和M: 0.90887。精密度和召回率分别为1.0和0.99975,表明该模型具有较高的灵敏度。此外,box_loss、segmentation_loss和objectness_loss的低值表明该模型保持了最小的错误率。该方法比传统的分割方法具有更高的准确性和灵敏度,对农业病害检测具有重要意义。
{"title":"Part I—Modernizing Nemabot: AI-Supported Identification of Frankliniella occidentalis Damage for Enhanced Biological Control Efficiency","authors":"Hilal Erdoğan,&nbsp;Atilla Erdinç,&nbsp;Alperen Kaan Bütüner,&nbsp;Tufan Can Ulu,&nbsp;İ. Alper Susurluk,&nbsp;Edwin E. Lewis,&nbsp;Halil Ünal","doi":"10.1002/rob.70068","DOIUrl":"https://doi.org/10.1002/rob.70068","url":null,"abstract":"<div>\u0000 \u0000 <p>Western Flower Thrips (<i>Frankliniella occidentalis</i>) is a significant agricultural pest causing substantial economic losses by damaging crops and acting as a vector for plant diseases. Traditional pest control methods relying on chemical pesticides pose environmental and health risks, necessitating alternative solutions. Entomopathogenic nematodes (EPNs) have emerged as a promising biological control agent. This study presents an AI-supported precision application system, Nemabot, designed to optimize EPN deployment based on thrips-induced bean leaf damage. In this study, agricultural disease detection was performed using the Multi-Otsu Thresholding method integrated into deep learning-based object detection and segmentation algorithms. The developed method enhances segmentation accuracy through image processing techniques, thereby increasing the precision in identifying infested regions. The model used in the study was optimized with a YOLO-based architecture during training and reinforced with various data augmentation techniques for segmenting bean leaves. The model's performance evaluation yielded mAP0.5 values of B: 0.9481 and M: 0.94981, while mAP0.5:0.95 values were B: 0.90887 and M: 0.90887. The precision and recall values were 1.0 and 0.99975, respectively, indicating the model's high sensitivity. Additionally, the low values of box_loss, segmentation_loss, and objectness_loss demonstrate that the model maintains a minimal error rate. The proposed approach offers higher accuracy and sensitivity than conventional segmentation methods, contributing significantly to agricultural disease detection applications.</p>\u0000 </div>","PeriodicalId":192,"journal":{"name":"Journal of Field Robotics","volume":"43 2","pages":"761-778"},"PeriodicalIF":5.2,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146139157","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Design and Implementation of an Underwater Structural Inspection Robot for High-Flow Environments 大流量水下结构检测机器人的设计与实现
IF 5.2 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-09-01 DOI: 10.1002/rob.70047
Qinyun Tang, Ying Du, Zhaojin Liu, Shuo Zhang, Qiang Zhao, Yingxuan Li, Liquan Wang, Tong Cui, Zibing Liu, Gang Wang

This paper introduces the design methodology and development of an inspection robot for submerged structures in high-flow aquatic environments. The robotic system employs dual counter-rotating vortex suction cups for surface adhesion and achieves omnidirectional mobility through four integrated steering-drive joints. Its reconfigurable articulated structure enables adaptive conformation to diverse structural geometries. Current regulatory frameworks explicitly mandate non-destructive testing (NDT) requirements for underwater infrastructure maintenance. Conventional suspended underwater robots face challenges in maintaining stable positional control and consistent surface contact, while existing adhesion-based systems demonstrate deficiencies in flow-field adaptability, maneuverability, and transitional capability within complex hydrodynamic conditions. Through systematic exploration of robotic design principles balancing adsorption forces and hydrodynamic resistance under intense flow disturbances, we prototyped and validated the system via experimental campaigns in both dam environments and controlled hydrodynamic test basins. Experimental results demonstrate the robot's capability to execute high-dexterity maneuvers including in-situ rotation, comb-pattern scanning, zigzag trajectories, and intersection curve paths. The system proves effective for inspecting both large-curvature-radius surfaces (e.g., dam facades) and small-diameter tubular structures (e.g., jacket node welds). Field trials confirm that even with simplified control architecture requiring minimal operator intervention, the robot successfully acquires high-resolution continuous imaging on dam surfaces and obtains valid phased array ultrasonic testing (PAUT) signals for weld inspection, demonstrating detectable sensitivity to artificially induced defects.

本文介绍了一种高流量水下结构物检测机器人的设计方法和研制过程。机器人系统采用双反向旋转涡流吸盘进行表面附着,并通过四个集成的转向驱动关节实现全方位移动。其可重构的铰接结构能够适应不同的结构几何形状。目前的监管框架明确规定了水下基础设施维护的无损检测(NDT)要求。传统的悬浮式水下机器人在保持稳定的位置控制和一致的表面接触方面面临挑战,而现有的基于黏附的系统在复杂水动力条件下的流场适应性、机动性和过渡能力方面存在不足。通过系统探索在强水流扰动下平衡吸附力和水动力阻力的机器人设计原理,我们在大坝环境和受控水动力试验盆地中进行了原型设计并验证了系统。实验结果表明,该机器人能够完成原位旋转、梳状扫描、之字形轨迹和交叉曲线路径等高灵巧动作。事实证明,该系统对大曲率半径表面(如大坝外立面)和小直径管状结构(如夹套节点焊缝)的检测都是有效的。现场试验证实,即使采用简化的控制架构,只需最少的操作员干预,机器人也能成功地获得大坝表面的高分辨率连续成像,并获得有效的相控阵超声检测(pat)信号,用于焊缝检测,显示出对人为缺陷的可检测灵敏度。
{"title":"Design and Implementation of an Underwater Structural Inspection Robot for High-Flow Environments","authors":"Qinyun Tang,&nbsp;Ying Du,&nbsp;Zhaojin Liu,&nbsp;Shuo Zhang,&nbsp;Qiang Zhao,&nbsp;Yingxuan Li,&nbsp;Liquan Wang,&nbsp;Tong Cui,&nbsp;Zibing Liu,&nbsp;Gang Wang","doi":"10.1002/rob.70047","DOIUrl":"https://doi.org/10.1002/rob.70047","url":null,"abstract":"<div>\u0000 \u0000 <p>This paper introduces the design methodology and development of an inspection robot for submerged structures in high-flow aquatic environments. The robotic system employs dual counter-rotating vortex suction cups for surface adhesion and achieves omnidirectional mobility through four integrated steering-drive joints. Its reconfigurable articulated structure enables adaptive conformation to diverse structural geometries. Current regulatory frameworks explicitly mandate non-destructive testing (NDT) requirements for underwater infrastructure maintenance. Conventional suspended underwater robots face challenges in maintaining stable positional control and consistent surface contact, while existing adhesion-based systems demonstrate deficiencies in flow-field adaptability, maneuverability, and transitional capability within complex hydrodynamic conditions. Through systematic exploration of robotic design principles balancing adsorption forces and hydrodynamic resistance under intense flow disturbances, we prototyped and validated the system via experimental campaigns in both dam environments and controlled hydrodynamic test basins. Experimental results demonstrate the robot's capability to execute high-dexterity maneuvers including in-situ rotation, comb-pattern scanning, zigzag trajectories, and intersection curve paths. The system proves effective for inspecting both large-curvature-radius surfaces (e.g., dam facades) and small-diameter tubular structures (e.g., jacket node welds). Field trials confirm that even with simplified control architecture requiring minimal operator intervention, the robot successfully acquires high-resolution continuous imaging on dam surfaces and obtains valid phased array ultrasonic testing (PAUT) signals for weld inspection, demonstrating detectable sensitivity to artificially induced defects.</p>\u0000 </div>","PeriodicalId":192,"journal":{"name":"Journal of Field Robotics","volume":"43 1","pages":"526-553"},"PeriodicalIF":5.2,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145792370","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Prioritized Real-Time UAV-Based Vessel Detection for Efficient Maritime Search 基于无人机的船舶实时检测优先级高效海上搜索
IF 5.2 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-09-01 DOI: 10.1002/rob.70048
Lyes Saad Saoud, Zikai Jia, Siyuan Yang, Muhayy Ud Din, Lakmal Seneviratne, Shaoming He, Irfan Hussain

Real-time vessel detection in maritime environments is crucial for diverse applications requiring speed and accuracy. Static camera views often introduce blind spots, compromising detection efficiency. This paper proposes a novel, real-time UAV-based system that uses a dynamic camera control strategy to address this limitation. This strategy leverages pre-defined search patterns, historical data (if available), and real-time sensor information (e.g., radar or LiDAR) to dynamically adjust the UAV's camera gimbal angles. This ensures comprehensive search area coverage while minimizing the risk of undetected vessels. Beyond dynamic camera control, our system incorporates a unique feature-based prioritization scheme for real-time target vessel identification. This scheme analyzes features extracted from captured images, including object size and shape. Additionally, movement analysis helps distinguish stationary objects from potential vessels. The combined approach of dynamic camera control and feature-based prioritization offers significant advantages. Firstly, it enhances search efficiency by systematically scanning the area and prioritizing promising candidates based on dynamic camera adjustments and feature analysis. Secondly, it improves detection accuracy by employing feature similarity (cosine similarity with a reference vessel stored in the system using a ResNet50 module) to reduce false positives and expedite target identification, especially in scenarios with multiple vessels. A comprehensive evaluation process has been conducted to validate the effectiveness of our proposed system in diverse simulated and real-world environments encompassing various conditions (weather, traffic density, background clutter). The results from this evaluation are highly promising and suggest the system's strong potential for real-time vessel detection in maritime environments.

海上环境中的实时船舶检测对于需要速度和准确性的各种应用至关重要。静态摄像机视图通常会引入盲点,影响检测效率。本文提出了一种新颖的、基于无人机的实时系统,该系统使用动态摄像机控制策略来解决这一限制。该策略利用预定义的搜索模式、历史数据(如果可用)和实时传感器信息(例如雷达或激光雷达)来动态调整无人机的相机万向架角度。这确保了全面的搜索区域覆盖,同时最大限度地降低了未被发现船只的风险。除了动态摄像机控制之外,我们的系统还采用了独特的基于特征的优先级方案,用于实时目标船舶识别。该方案分析从捕获图像中提取的特征,包括物体的大小和形状。此外,运动分析有助于区分静止物体和潜在血管。动态相机控制和基于特征的优先级结合的方法具有显著的优势。首先,通过系统地扫描区域,并根据动态相机调整和特征分析对候选区域进行优先排序,提高搜索效率;其次,它通过使用特征相似性(与使用ResNet50模块存储在系统中的参考血管的余弦相似性)来提高检测精度,以减少误报并加快目标识别,特别是在有多个血管的情况下。我们进行了一个全面的评估过程,以验证我们提出的系统在各种模拟和现实环境中的有效性,这些环境包括各种条件(天气、交通密度、背景杂波)。这次评估的结果非常有希望,并表明该系统在海上环境中实时船舶检测方面具有强大的潜力。
{"title":"Prioritized Real-Time UAV-Based Vessel Detection for Efficient Maritime Search","authors":"Lyes Saad Saoud,&nbsp;Zikai Jia,&nbsp;Siyuan Yang,&nbsp;Muhayy Ud Din,&nbsp;Lakmal Seneviratne,&nbsp;Shaoming He,&nbsp;Irfan Hussain","doi":"10.1002/rob.70048","DOIUrl":"https://doi.org/10.1002/rob.70048","url":null,"abstract":"<p>Real-time vessel detection in maritime environments is crucial for diverse applications requiring speed and accuracy. Static camera views often introduce blind spots, compromising detection efficiency. This paper proposes a novel, real-time UAV-based system that uses a dynamic camera control strategy to address this limitation. This strategy leverages pre-defined search patterns, historical data (if available), and real-time sensor information (e.g., radar or LiDAR) to dynamically adjust the UAV's camera gimbal angles. This ensures comprehensive search area coverage while minimizing the risk of undetected vessels. Beyond dynamic camera control, our system incorporates a unique feature-based prioritization scheme for real-time target vessel identification. This scheme analyzes features extracted from captured images, including object size and shape. Additionally, movement analysis helps distinguish stationary objects from potential vessels. The combined approach of dynamic camera control and feature-based prioritization offers significant advantages. Firstly, it enhances search efficiency by systematically scanning the area and prioritizing promising candidates based on dynamic camera adjustments and feature analysis. Secondly, it improves detection accuracy by employing feature similarity (cosine similarity with a reference vessel stored in the system using a ResNet50 module) to reduce false positives and expedite target identification, especially in scenarios with multiple vessels. A comprehensive evaluation process has been conducted to validate the effectiveness of our proposed system in diverse simulated and real-world environments encompassing various conditions (weather, traffic density, background clutter). The results from this evaluation are highly promising and suggest the system's strong potential for real-time vessel detection in maritime environments.</p>","PeriodicalId":192,"journal":{"name":"Journal of Field Robotics","volume":"43 2","pages":"561-577"},"PeriodicalIF":5.2,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/rob.70048","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146139159","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
From “Mirror Flower, Water Moon” to Multi-Task Visual Prospective Representation Learning for Unmanned Aerial Vehicles Indoor Mapless Navigation 从“镜花水月”到无人机室内无地图导航的多任务视觉前瞻表征学习
IF 5.2 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-09-01 DOI: 10.1002/rob.70057
Yingxiu Chang, Yongqiang Cheng, John Murray, Muhammad Khalid, Umar Manzoor

Vision-based deep learning models have been widely adopted in autonomous agents, such as unmanned aerial vehicles (UAVs), particularly in reactive control policies that serve as a key component of navigation systems. These policies enable agents to respond instantaneously to dynamic environments without relying on pre-existing maps. However, there remain open challenges to improve the agent's reactive control performance: (1) Is it possible and how to anticipate future states at the current moment to benefit control precision? (2) Is it possible and how can we anticipate future states for different sub-tasks when the agent's control consists of both discrete classification and continuous regression commands? Inspired by the Chinese idiom “Mirror Flower, Water Moon,” this paper hypothesizes that future states in the latent space can be learnt from sequential images using contrastive learning, and consequently proposes a light-weight Multi-task Visual Prospective Representation Learning (MulVPRL) framework for benefiting reactive control. Specifically, (1) This paper leverages the advantage of contrastive learning to correlate the representations obtained from the latest sequential images and one image in the future. (2) This paper constructs an integrated loss function of contrastive learning for classification and regression sub-tasks. The MulVPRL framework outperforms the benchmark models on the public HDIN and DroNet datasets, and obtained the best performance in real-world experiments (� � 46.9� � m� � ,� � 177� � svs� � . SOTA � � 27.3� � m� � ,� � 136� � s). Therefore, the multi-task contrastive learning of the light-weight MulVPRL framework enhances reactive control performance on a 2D plane, and demonstrates the potential to be integrated with various intelligent strategies, and implemented on ground vehicles.

基于视觉的深度学习模型已被广泛应用于自主代理,如无人驾驶飞行器(uav),特别是在作为导航系统关键组成部分的响应式控制策略中。这些策略使代理能够即时响应动态环境,而不依赖于预先存在的地图。然而,提高智能体的反应性控制性能仍然存在一些开放的挑战:(1)是否有可能以及如何预测当前时刻的未来状态以提高控制精度?(2)当智能体的控制同时由离散分类和连续回归命令组成时,我们是否可能以及如何预测不同子任务的未来状态?受中国成语“镜花水月”的启发,本文假设潜在空间的未来状态可以通过对比学习从序列图像中学习,并因此提出了一个轻量级的多任务视觉前瞻表征学习(MulVPRL)框架,有利于反应性控制。具体而言,(1)本文利用对比学习的优势,将最新的连续图像和未来的一个图像获得的表征关联起来。(2)构建了分类子任务和回归子任务的对比学习综合损失函数。MulVPRL框架在公共HDIN和DroNet数据集上优于基准模型,并在实际实验中获得了最佳性能(46.9 m)。177 SVS。SOTA 27.3米,136秒)。因此,轻型MulVPRL框架的多任务对比学习增强了二维平面上的无功控制性能,并展示了与各种智能策略集成并在地面车辆上实施的潜力。
{"title":"From “Mirror Flower, Water Moon” to Multi-Task Visual Prospective Representation Learning for Unmanned Aerial Vehicles Indoor Mapless Navigation","authors":"Yingxiu Chang,&nbsp;Yongqiang Cheng,&nbsp;John Murray,&nbsp;Muhammad Khalid,&nbsp;Umar Manzoor","doi":"10.1002/rob.70057","DOIUrl":"https://doi.org/10.1002/rob.70057","url":null,"abstract":"<p>Vision-based deep learning models have been widely adopted in autonomous agents, such as unmanned aerial vehicles (UAVs), particularly in reactive control policies that serve as a key component of navigation systems. These policies enable agents to respond instantaneously to dynamic environments without relying on pre-existing maps. However, there remain open challenges to improve the agent's reactive control performance: (1) Is it possible and how to anticipate future states at the current moment to benefit control precision? (2) Is it possible and how can we anticipate future states for different sub-tasks when the agent's control consists of both discrete classification and continuous regression commands? Inspired by the Chinese idiom “Mirror Flower, Water Moon,” this paper hypothesizes that future states in the latent space can be learnt from sequential images using contrastive learning, and consequently proposes a light-weight Multi-task Visual Prospective Representation Learning (MulVPRL) framework for benefiting reactive control. Specifically, (1) This paper leverages the advantage of contrastive learning to correlate the representations obtained from the latest sequential images and one image in the future. (2) This paper constructs an integrated loss function of contrastive learning for classification and regression sub-tasks. The MulVPRL framework outperforms the benchmark models on the public HDIN and DroNet datasets, and obtained the best performance in real-world experiments (<span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 \u0000 <mrow>\u0000 <mn>46.9</mn>\u0000 <mspace></mspace>\u0000 \u0000 <mi>m</mi>\u0000 \u0000 <mo>,</mo>\u0000 \u0000 <mn>177</mn>\u0000 \u0000 <mi>svs</mi>\u0000 \u0000 <mo>.</mo>\u0000 </mrow>\u0000 </mrow>\u0000 </semantics></math> SOTA <span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 \u0000 <mrow>\u0000 <mn>27.3</mn>\u0000 <mspace></mspace>\u0000 \u0000 <mi>m</mi>\u0000 \u0000 <mo>,</mo>\u0000 \u0000 <mn>136</mn>\u0000 <mspace></mspace>\u0000 \u0000 <mi>s</mi>\u0000 </mrow>\u0000 </mrow>\u0000 </semantics></math>). Therefore, the multi-task contrastive learning of the light-weight MulVPRL framework enhances reactive control performance on a 2D plane, and demonstrates the potential to be integrated with various intelligent strategies, and implemented on ground vehicles.</p>","PeriodicalId":192,"journal":{"name":"Journal of Field Robotics","volume":"43 2","pages":"655-680"},"PeriodicalIF":5.2,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/rob.70057","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146139161","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
INF-SLiM: Large-Scale Implicit Neural Fields for Semantic LiDAR Mapping of Embodied AI Agents INF-SLiM:大规模隐式神经场,用于嵌入人工智能代理的语义激光雷达映射
IF 5.2 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-09-01 DOI: 10.1002/rob.70058
Jianyuan Zhang, Zhiliu Yang, Meng Zhang, Hongyu Chen, Lianhui Zhao, Chen Liu

Large-scale semantic mapping is crucial for outdoor autonomous agents to perform high-level tasks such as planning and navigation. In this paper, we propose a novel method for large-scale 3D semantic reconstruction through implicit representations from posed LiDAR measurements alone. We first construct a neural fields via the octree and latent features, then decode implicit features into signed distance value and semantic information through shallow Multilayer Perceptrons (MLPs). We leverage radial window self-attention networks to predict the semantic labels of point clouds. We then jointly optimize the feature embedding and MLP parameters with a self-supervision paradigm for point-cloud geometry and a pseudo-supervision paradigm for semantic and panoptic labels. Subsequently, geometric structures and semantic categories for novel points in the unseen area are regressed, and the marching cubes method is exploited to subdivide and visualize scenes in the inferring stage, the labeled mesh is produced correspondingly. Experiments on two real-world datasets, SemanticKITTI and SemanticPOSS, demonstrate the superior segmentation efficiency and mapping effectiveness of our framework compared to existing 3D semantic LiDAR mapping methods.

大规模语义映射对于户外自主智能体执行规划和导航等高级任务至关重要。在本文中,我们提出了一种新的方法,用于大规模的三维语义重建,通过隐式表示从提出的激光雷达测量。首先通过八叉树和潜在特征构造神经场,然后通过浅层多层感知器(mlp)将隐式特征解码为带符号距离值和语义信息。我们利用径向窗自关注网络来预测点云的语义标签。然后,我们使用点云几何的自监督范式和语义和全景标签的伪监督范式共同优化特征嵌入和MLP参数。随后,对未见区域中新点的几何结构和语义类别进行回归,并在推理阶段利用行军立方体方法对场景进行细分和可视化,生成相应的标记网格。在两个真实数据集SemanticKITTI和SemanticPOSS上的实验表明,与现有的3D语义LiDAR映射方法相比,我们的框架具有更高的分割效率和映射有效性。
{"title":"INF-SLiM: Large-Scale Implicit Neural Fields for Semantic LiDAR Mapping of Embodied AI Agents","authors":"Jianyuan Zhang,&nbsp;Zhiliu Yang,&nbsp;Meng Zhang,&nbsp;Hongyu Chen,&nbsp;Lianhui Zhao,&nbsp;Chen Liu","doi":"10.1002/rob.70058","DOIUrl":"https://doi.org/10.1002/rob.70058","url":null,"abstract":"<div>\u0000 \u0000 <p>Large-scale semantic mapping is crucial for outdoor autonomous agents to perform high-level tasks such as planning and navigation. In this paper, we propose a novel method for large-scale 3D semantic reconstruction through implicit representations from posed LiDAR measurements alone. We first construct a neural fields via the octree and latent features, then decode implicit features into signed distance value and semantic information through shallow Multilayer Perceptrons (MLPs). We leverage radial window self-attention networks to predict the semantic labels of point clouds. We then jointly optimize the feature embedding and MLP parameters with a self-supervision paradigm for point-cloud geometry and a pseudo-supervision paradigm for semantic and panoptic labels. Subsequently, geometric structures and semantic categories for novel points in the unseen area are regressed, and the marching cubes method is exploited to subdivide and visualize scenes in the inferring stage, the labeled mesh is produced correspondingly. Experiments on two real-world datasets, SemanticKITTI and SemanticPOSS, demonstrate the superior segmentation efficiency and mapping effectiveness of our framework compared to existing 3D semantic LiDAR mapping methods.</p>\u0000 </div>","PeriodicalId":192,"journal":{"name":"Journal of Field Robotics","volume":"43 2","pages":"681-694"},"PeriodicalIF":5.2,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146139125","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Blind Walking Control of a Heavy-Duty Hexapod Robot Based on Optimal Force Allocation for Uneven and Sloped Terrains 基于最优力分配的重型六足机器人不均匀和倾斜地形盲行控制
IF 5.2 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-09-01 DOI: 10.1002/rob.70051
Yufei Liu, Yongyao Li, Lei Jiang, Bo Su, Boyang Xing, Peng Xu

Heavy-duty robots such as hexapod robots exhibit tremendous potential with their high payload ability and terrain adaptability. This paper introduces a control approach for heavy-duty hexapod robot that traverses uneven and transition stage from a flat surface to a sloped terrain without prior geometric knowledge of the environment. Using force sensors on the robot feet and position sensors at its joints, the transition from a flat surface to a sloped terrain is detected. Constraint models based on terrain geometric information and joint limit positions are presented. The models are derived to estimate the unknown slope gradient based on the landing feet positions. The constraints for different terrains have also been investigated. An optimal force allocation method under contact constraints is developed to minimize the linear and quadratic costs in the commands and constrained contact forces. Extensive experiments have been conducted, and the results have demonstrated the effectiveness of the proposed control approach.

六足机器人等重型机器人以其高载荷能力和地形适应性表现出巨大的潜力。本文介绍了一种重型六足机器人的控制方法,该方法可以在不事先了解环境几何知识的情况下,通过不平整和从平坦到倾斜的过渡阶段。利用机器人脚上的力传感器和关节上的位置传感器,检测从平坦表面到倾斜地形的过渡。提出了基于地形几何信息和关节极限位置的约束模型。推导了基于着陆脚位置的未知坡度估计模型。对不同地形的约束条件也进行了研究。提出了一种接触约束下的最优力分配方法,以最大限度地降低指令和约束接触力的线性和二次代价。大量的实验已经进行,结果证明了所提出的控制方法的有效性。
{"title":"Blind Walking Control of a Heavy-Duty Hexapod Robot Based on Optimal Force Allocation for Uneven and Sloped Terrains","authors":"Yufei Liu,&nbsp;Yongyao Li,&nbsp;Lei Jiang,&nbsp;Bo Su,&nbsp;Boyang Xing,&nbsp;Peng Xu","doi":"10.1002/rob.70051","DOIUrl":"https://doi.org/10.1002/rob.70051","url":null,"abstract":"<div>\u0000 \u0000 <p>Heavy-duty robots such as hexapod robots exhibit tremendous potential with their high payload ability and terrain adaptability. This paper introduces a control approach for heavy-duty hexapod robot that traverses uneven and transition stage from a flat surface to a sloped terrain without prior geometric knowledge of the environment. Using force sensors on the robot feet and position sensors at its joints, the transition from a flat surface to a sloped terrain is detected. Constraint models based on terrain geometric information and joint limit positions are presented. The models are derived to estimate the unknown slope gradient based on the landing feet positions. The constraints for different terrains have also been investigated. An optimal force allocation method under contact constraints is developed to minimize the linear and quadratic costs in the commands and constrained contact forces. Extensive experiments have been conducted, and the results have demonstrated the effectiveness of the proposed control approach.</p>\u0000 </div>","PeriodicalId":192,"journal":{"name":"Journal of Field Robotics","volume":"43 2","pages":"599-621"},"PeriodicalIF":5.2,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146139127","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Development of Large Wingspan Flapping-Wing Flying Robots With Improved Efficiency by Rigid-Flexible Coupling Flapping Mechanism 采用刚柔耦合扑翼机构提高效率的大翼展扑翼飞行机器人的研制
IF 5.2 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-08-28 DOI: 10.1002/rob.70054
Siping Zhong, Wenfu Xu, Jizhou Jiang, Zihao Wei, Erzhen Pan

The flight efficiency of bionic flapping-wing robots is largely influenced by the flapping mechanism. This paper presents the development of a large wingspan flapping-wing flying robot that utilizes a rigid-flexible coupling flapping mechanism to reduce energy consumption, improve flight efficiency, and enhance wind resistance. The rigid-flexible coupling flapping mechanism consists of a DC motor with a gearset and a flexible flapping mechanism based on torsion spring. This mechanism combines the high torque driving capability of a rigid mechanism with the energy storage capacity of a flexible mechanism. Synchronizing the passive deformation of the torsion spring with the periodic acceleration and deceleration of the wings during the flapping cycle enables the storage, transfer, and release of kinetic energy and torsional spring elastic strain energy. Simulation studies show that the proposed design effectively reduces the peak power required. A prototype with a wingspan of 1.8 meters and a mass of 1.0 kilograms was developed. Compared to a rigid mechanism, the proposed design significantly reduces the electrical power consumption, especially when achieving flapping frequencies exceeding 1.5 Hz, as validated through wind tunnel experiments. At a flapping frequency of 2.2 Hz, the maximum reduction in electrical power consumption reaches 14.8%. The flexible elements also increased the downstroke ratio, consequently enhancing thrust and propulsion efficiency. Outdoor flight experiments have demonstrated that the prototype propelled by the rigid-flexible coupling drive mechanism consumes only 81.56 W during cruising, which is 8.78 W (9.72%) less than the rigidly driven prototype.

扑翼仿生机器人的飞行效率在很大程度上受扑翼机构的影响。研制了一种大翼展扑翼飞行机器人,该机器人采用刚柔耦合扑翼机构,以降低能耗,提高飞行效率,增强抗风能力。刚柔耦合扑动机构由带齿轮组的直流电动机和基于扭簧的柔性扑动机构组成。该机构结合了刚性机构的高扭矩驱动能力和柔性机构的储能能力。将扭转弹簧的被动变形与机翼在扑动周期中的周期性加减速同步,实现了动能和扭转弹簧弹性应变能的储存、传递和释放。仿真研究表明,该设计有效地降低了所需的峰值功率。研制出翼展1.8米、质量1.0公斤的原型机。与刚性机构相比,该设计显著降低了电力消耗,特别是在实现超过1.5 Hz的扑动频率时,这一点已通过风洞实验得到验证。当扑动频率为2.2 Hz时,电能消耗的最大降幅可达14.8%。柔性元件还增加了下冲程比,从而提高了推力和推进效率。室外飞行实验表明,采用刚柔耦合驱动机构驱动的原型机巡航时能耗仅为81.56 W,比刚性驱动的原型机减少8.78 W(9.72%)。
{"title":"Development of Large Wingspan Flapping-Wing Flying Robots With Improved Efficiency by Rigid-Flexible Coupling Flapping Mechanism","authors":"Siping Zhong,&nbsp;Wenfu Xu,&nbsp;Jizhou Jiang,&nbsp;Zihao Wei,&nbsp;Erzhen Pan","doi":"10.1002/rob.70054","DOIUrl":"https://doi.org/10.1002/rob.70054","url":null,"abstract":"<div>\u0000 \u0000 <p>The flight efficiency of bionic flapping-wing robots is largely influenced by the flapping mechanism. This paper presents the development of a large wingspan flapping-wing flying robot that utilizes a rigid-flexible coupling flapping mechanism to reduce energy consumption, improve flight efficiency, and enhance wind resistance. The rigid-flexible coupling flapping mechanism consists of a DC motor with a gearset and a flexible flapping mechanism based on torsion spring. This mechanism combines the high torque driving capability of a rigid mechanism with the energy storage capacity of a flexible mechanism. Synchronizing the passive deformation of the torsion spring with the periodic acceleration and deceleration of the wings during the flapping cycle enables the storage, transfer, and release of kinetic energy and torsional spring elastic strain energy. Simulation studies show that the proposed design effectively reduces the peak power required. A prototype with a wingspan of 1.8 meters and a mass of 1.0 kilograms was developed. Compared to a rigid mechanism, the proposed design significantly reduces the electrical power consumption, especially when achieving flapping frequencies exceeding 1.5 Hz, as validated through wind tunnel experiments. At a flapping frequency of 2.2 Hz, the maximum reduction in electrical power consumption reaches 14.8%. The flexible elements also increased the downstroke ratio, consequently enhancing thrust and propulsion efficiency. Outdoor flight experiments have demonstrated that the prototype propelled by the rigid-flexible coupling drive mechanism consumes only 81.56 W during cruising, which is 8.78 W (9.72%) less than the rigidly driven prototype.</p>\u0000 </div>","PeriodicalId":192,"journal":{"name":"Journal of Field Robotics","volume":"43 1","pages":"513-525"},"PeriodicalIF":5.2,"publicationDate":"2025-08-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145792471","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Under-Ice Odometry Fusing Acoustic and Visual Information for Underwater Vehicles 水下航行器的冰下里程测量融合声学和视觉信息
IF 5.2 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-08-28 DOI: 10.1002/rob.70052
Du Haohan, Ma Teng, Li Ye, Li Ao, Xu Shuo, Wang Hao, Cao Jian, Liao Yulei

Ice cover hinders the deployment of acoustic arrays or satellites, making it challenging to locate underwater vehicles in the polar regions. Odometry based on acoustic or optical sensors has the potential to provide precise navigation results for vehicles. However, the robustness of single-sensor-based odometry is significantly compromised by the smooth and featureless nature of the ice-covered bottom, which lacks distinct points, lines, or surfaces. This paper presents a filter-based multi-sensor fusion underwater ice odometry system, integrating measurements from a stereo camera, a forward-looking sonar (FLS), a low-cost inertial measurement unit (IMU), and a depth gauge (DG). To balance real-time performance and accuracy, the Multi-State Constraint Kalman Filter (MSCKF) algorithm is employed to fuse inertial and stereo measurements for vehicle state estimation. Additionally, FLS images and relative depth measurements from the DG are utilized to enhance the accuracy and robustness of the state estimation. Specifically, a Fourier-based algorithm is proposed to estimate linear speed using global information from FLS images. A compensation factor is introduced to address the impact of FLS elevation angle on estimation accuracy. Furthermore, a residual-based adaptive method and forgetting factor are implemented to adjust the measurement noise covariance of linear speed and relative depth, improving dynamic state estimation precision. The proposed system was evaluated through tank and field experiments using a remotely operated vehicle (ROV). Results show it provides more accurate and robust under-ice navigation compared to state-of-the-art algorithms.

冰盖阻碍了声学阵列或卫星的部署,使得在极地地区定位水下航行器变得具有挑战性。基于声学或光学传感器的里程计有可能为车辆提供精确的导航结果。然而,基于单传感器的里程计的鲁棒性受到冰层覆盖的底部光滑和无特征的性质的严重损害,因为冰层覆盖的底部缺乏明显的点、线或表面。本文提出了一种基于滤波的多传感器融合水下冰层测程系统,该系统集成了立体摄像机、前视声纳(FLS)、低成本惯性测量单元(IMU)和深度计(DG)的测量数据。为了平衡实时性和准确性,采用多状态约束卡尔曼滤波(MSCKF)算法融合惯性和立体测量进行车辆状态估计。此外,利用FLS图像和DG的相对深度测量来提高状态估计的准确性和鲁棒性。具体而言,提出了一种基于傅里叶的算法,利用FLS图像的全局信息估计线性速度。引入补偿因子来解决FLS仰角对估计精度的影响。采用残差自适应方法和遗忘因子对测量噪声的线性速度和相对深度协方差进行调整,提高了动态估计精度。通过远程操作车辆(ROV)的油箱和现场实验对该系统进行了评估。结果表明,与最先进的算法相比,它提供了更准确、更稳健的冰下导航。
{"title":"Under-Ice Odometry Fusing Acoustic and Visual Information for Underwater Vehicles","authors":"Du Haohan,&nbsp;Ma Teng,&nbsp;Li Ye,&nbsp;Li Ao,&nbsp;Xu Shuo,&nbsp;Wang Hao,&nbsp;Cao Jian,&nbsp;Liao Yulei","doi":"10.1002/rob.70052","DOIUrl":"https://doi.org/10.1002/rob.70052","url":null,"abstract":"<div>\u0000 \u0000 <p>Ice cover hinders the deployment of acoustic arrays or satellites, making it challenging to locate underwater vehicles in the polar regions. Odometry based on acoustic or optical sensors has the potential to provide precise navigation results for vehicles. However, the robustness of single-sensor-based odometry is significantly compromised by the smooth and featureless nature of the ice-covered bottom, which lacks distinct points, lines, or surfaces. This paper presents a filter-based multi-sensor fusion underwater ice odometry system, integrating measurements from a stereo camera, a forward-looking sonar (FLS), a low-cost inertial measurement unit (IMU), and a depth gauge (DG). To balance real-time performance and accuracy, the Multi-State Constraint Kalman Filter (MSCKF) algorithm is employed to fuse inertial and stereo measurements for vehicle state estimation. Additionally, FLS images and relative depth measurements from the DG are utilized to enhance the accuracy and robustness of the state estimation. Specifically, a Fourier-based algorithm is proposed to estimate linear speed using global information from FLS images. A compensation factor is introduced to address the impact of FLS elevation angle on estimation accuracy. Furthermore, a residual-based adaptive method and forgetting factor are implemented to adjust the measurement noise covariance of linear speed and relative depth, improving dynamic state estimation precision. The proposed system was evaluated through tank and field experiments using a remotely operated vehicle (ROV). Results show it provides more accurate and robust under-ice navigation compared to state-of-the-art algorithms.</p>\u0000 </div>","PeriodicalId":192,"journal":{"name":"Journal of Field Robotics","volume":"43 1","pages":"486-512"},"PeriodicalIF":5.2,"publicationDate":"2025-08-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145792470","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
S 2 MAT : Simultaneous and Self-Reinforced Mapping and Tracking in Dynamic Urban Scenarios s2 MAT:动态城市场景中的同步和自增强映射与跟踪
IF 5.2 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-08-28 DOI: 10.1002/rob.70044
Tingxiang Fan, Bowen Shen, Yinqiang Zhang, Chuye Zhang, Lei Yang, Hua Chen, Wei Zhang, Jia Pan

Despite the increasing prevalence of robots in daily life, their navigation capabilities are still limited to environments with prior knowledge, such as a global map. To fully unlock the potential of robots, it is crucial to enable them to navigate in large-scale unknown and changing unstructured scenarios. This requires the robot to construct an accurate static map in real-time as it explores, while filtering out moving objects to ensure mapping accuracy and, if possible, achieving high-quality pedestrian tracking and collision avoidance. While existing methods can achieve individual goals of spatial mapping or dynamic object detection and tracking, there has been limited research on effectively integrating these two tasks, which are actually coupled and reciprocal. In this study, we propose a solution called S2MAT (simultaneous and self-reinforced mapping and tracking) that integrates a front-end dynamic object detection and tracking module with a back-end static mapping module. S2MAT leverages the close and reciprocal interplay between these two modules to efficiently and effectively solve the open problem of simultaneous tracking and mapping in highly dynamic scenarios. The proposed method is primarily designed for use with 3D LiDAR and offers a solution for real-time navigation in large-scale, unknown dynamic scenarios with a low computational cost, making it feasible for deployment on onboard computers equipped with only a single CPU. We conducted long-range experiments in real-world urban scenarios spanning over 7 km, which included challenging obstacles like pedestrians and other traffic agents. The successful navigation provides a comprehensive test of S2MAT's robustness, scalability, efficiency, quality, and its ability to benefit autonomous robots in wild scenarios without pre-built maps.

尽管机器人在日常生活中越来越普遍,但它们的导航能力仍然局限于具有先验知识的环境,例如全球地图。为了充分释放机器人的潜力,使它们能够在大规模未知和不断变化的非结构化场景中导航至关重要。这要求机器人在探索过程中实时构建精确的静态地图,同时过滤掉移动物体以确保地图精度,并在可能的情况下实现高质量的行人跟踪和避碰。虽然现有的方法可以实现空间映射或动态目标检测与跟踪的单独目标,但有效整合这两个任务的研究有限,这两个任务实际上是耦合和相互作用的。在本研究中,我们提出了一种称为S2MAT(同步和自增强映射和跟踪)的解决方案,该解决方案将前端动态目标检测和跟踪模块与后端静态映射模块集成在一起。S2MAT利用这两个模块之间的密切和相互作用,有效地解决了在高动态场景中同时跟踪和映射的开放问题。所提出的方法主要用于3D激光雷达,为大规模未知动态场景的实时导航提供了一种低计算成本的解决方案,使其能够部署在仅配备单个CPU的车载计算机上。我们在超过7公里的真实城市场景中进行了远程实验,其中包括行人和其他交通代理等具有挑战性的障碍物。成功的导航提供了对S2MAT的鲁棒性、可扩展性、效率、质量及其在没有预先构建地图的情况下使自主机器人受益的能力的全面测试。
{"title":"S\u0000 \u0000 2\u0000 \u0000 \u0000 MAT\u0000 \u0000 \u0000 \u0000 \u0000 : Simultaneous and Self-Reinforced Mapping and Tracking in Dynamic Urban Scenarios","authors":"Tingxiang Fan,&nbsp;Bowen Shen,&nbsp;Yinqiang Zhang,&nbsp;Chuye Zhang,&nbsp;Lei Yang,&nbsp;Hua Chen,&nbsp;Wei Zhang,&nbsp;Jia Pan","doi":"10.1002/rob.70044","DOIUrl":"https://doi.org/10.1002/rob.70044","url":null,"abstract":"<div>\u0000 \u0000 <p>Despite the increasing prevalence of robots in daily life, their navigation capabilities are still limited to environments with prior knowledge, such as a global map. To fully unlock the potential of robots, it is crucial to enable them to navigate in large-scale unknown and changing unstructured scenarios. This requires the robot to construct an accurate static map in real-time as it explores, while filtering out moving objects to ensure mapping accuracy and, if possible, achieving high-quality pedestrian tracking and collision avoidance. While existing methods can achieve individual goals of spatial mapping or dynamic object detection and tracking, there has been limited research on effectively integrating these two tasks, which are actually coupled and reciprocal. In this study, we propose a solution called S<sup>2</sup>MAT (simultaneous and self-reinforced mapping and tracking) that integrates a front-end dynamic object detection and tracking module with a back-end static mapping module. S<sup>2</sup>MAT leverages the close and reciprocal interplay between these two modules to efficiently and effectively solve the open problem of simultaneous tracking and mapping in highly dynamic scenarios. The proposed method is primarily designed for use with 3D LiDAR and offers a solution for real-time navigation in large-scale, unknown dynamic scenarios with a low computational cost, making it feasible for deployment on onboard computers equipped with only a single CPU. We conducted long-range experiments in real-world urban scenarios spanning over 7 km, which included challenging obstacles like pedestrians and other traffic agents. The successful navigation provides a comprehensive test of S<sup>2</sup>MAT's robustness, scalability, efficiency, quality, and its ability to benefit autonomous robots in wild scenarios without pre-built maps.</p></div>","PeriodicalId":192,"journal":{"name":"Journal of Field Robotics","volume":"43 1","pages":"454-485"},"PeriodicalIF":5.2,"publicationDate":"2025-08-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145792472","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Field Robotics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1