首页 > 最新文献

Journal of Field Robotics最新文献

英文 中文
Prioritized Real-Time UAV-Based Vessel Detection for Efficient Maritime Search 基于无人机的船舶实时检测优先级高效海上搜索
IF 5.2 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-09-01 DOI: 10.1002/rob.70048
Lyes Saad Saoud, Zikai Jia, Siyuan Yang, Muhayy Ud Din, Lakmal Seneviratne, Shaoming He, Irfan Hussain

Real-time vessel detection in maritime environments is crucial for diverse applications requiring speed and accuracy. Static camera views often introduce blind spots, compromising detection efficiency. This paper proposes a novel, real-time UAV-based system that uses a dynamic camera control strategy to address this limitation. This strategy leverages pre-defined search patterns, historical data (if available), and real-time sensor information (e.g., radar or LiDAR) to dynamically adjust the UAV's camera gimbal angles. This ensures comprehensive search area coverage while minimizing the risk of undetected vessels. Beyond dynamic camera control, our system incorporates a unique feature-based prioritization scheme for real-time target vessel identification. This scheme analyzes features extracted from captured images, including object size and shape. Additionally, movement analysis helps distinguish stationary objects from potential vessels. The combined approach of dynamic camera control and feature-based prioritization offers significant advantages. Firstly, it enhances search efficiency by systematically scanning the area and prioritizing promising candidates based on dynamic camera adjustments and feature analysis. Secondly, it improves detection accuracy by employing feature similarity (cosine similarity with a reference vessel stored in the system using a ResNet50 module) to reduce false positives and expedite target identification, especially in scenarios with multiple vessels. A comprehensive evaluation process has been conducted to validate the effectiveness of our proposed system in diverse simulated and real-world environments encompassing various conditions (weather, traffic density, background clutter). The results from this evaluation are highly promising and suggest the system's strong potential for real-time vessel detection in maritime environments.

海上环境中的实时船舶检测对于需要速度和准确性的各种应用至关重要。静态摄像机视图通常会引入盲点,影响检测效率。本文提出了一种新颖的、基于无人机的实时系统,该系统使用动态摄像机控制策略来解决这一限制。该策略利用预定义的搜索模式、历史数据(如果可用)和实时传感器信息(例如雷达或激光雷达)来动态调整无人机的相机万向架角度。这确保了全面的搜索区域覆盖,同时最大限度地降低了未被发现船只的风险。除了动态摄像机控制之外,我们的系统还采用了独特的基于特征的优先级方案,用于实时目标船舶识别。该方案分析从捕获图像中提取的特征,包括物体的大小和形状。此外,运动分析有助于区分静止物体和潜在血管。动态相机控制和基于特征的优先级结合的方法具有显著的优势。首先,通过系统地扫描区域,并根据动态相机调整和特征分析对候选区域进行优先排序,提高搜索效率;其次,它通过使用特征相似性(与使用ResNet50模块存储在系统中的参考血管的余弦相似性)来提高检测精度,以减少误报并加快目标识别,特别是在有多个血管的情况下。我们进行了一个全面的评估过程,以验证我们提出的系统在各种模拟和现实环境中的有效性,这些环境包括各种条件(天气、交通密度、背景杂波)。这次评估的结果非常有希望,并表明该系统在海上环境中实时船舶检测方面具有强大的潜力。
{"title":"Prioritized Real-Time UAV-Based Vessel Detection for Efficient Maritime Search","authors":"Lyes Saad Saoud,&nbsp;Zikai Jia,&nbsp;Siyuan Yang,&nbsp;Muhayy Ud Din,&nbsp;Lakmal Seneviratne,&nbsp;Shaoming He,&nbsp;Irfan Hussain","doi":"10.1002/rob.70048","DOIUrl":"https://doi.org/10.1002/rob.70048","url":null,"abstract":"<p>Real-time vessel detection in maritime environments is crucial for diverse applications requiring speed and accuracy. Static camera views often introduce blind spots, compromising detection efficiency. This paper proposes a novel, real-time UAV-based system that uses a dynamic camera control strategy to address this limitation. This strategy leverages pre-defined search patterns, historical data (if available), and real-time sensor information (e.g., radar or LiDAR) to dynamically adjust the UAV's camera gimbal angles. This ensures comprehensive search area coverage while minimizing the risk of undetected vessels. Beyond dynamic camera control, our system incorporates a unique feature-based prioritization scheme for real-time target vessel identification. This scheme analyzes features extracted from captured images, including object size and shape. Additionally, movement analysis helps distinguish stationary objects from potential vessels. The combined approach of dynamic camera control and feature-based prioritization offers significant advantages. Firstly, it enhances search efficiency by systematically scanning the area and prioritizing promising candidates based on dynamic camera adjustments and feature analysis. Secondly, it improves detection accuracy by employing feature similarity (cosine similarity with a reference vessel stored in the system using a ResNet50 module) to reduce false positives and expedite target identification, especially in scenarios with multiple vessels. A comprehensive evaluation process has been conducted to validate the effectiveness of our proposed system in diverse simulated and real-world environments encompassing various conditions (weather, traffic density, background clutter). The results from this evaluation are highly promising and suggest the system's strong potential for real-time vessel detection in maritime environments.</p>","PeriodicalId":192,"journal":{"name":"Journal of Field Robotics","volume":"43 2","pages":"561-577"},"PeriodicalIF":5.2,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/rob.70048","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146139159","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
From “Mirror Flower, Water Moon” to Multi-Task Visual Prospective Representation Learning for Unmanned Aerial Vehicles Indoor Mapless Navigation 从“镜花水月”到无人机室内无地图导航的多任务视觉前瞻表征学习
IF 5.2 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-09-01 DOI: 10.1002/rob.70057
Yingxiu Chang, Yongqiang Cheng, John Murray, Muhammad Khalid, Umar Manzoor

Vision-based deep learning models have been widely adopted in autonomous agents, such as unmanned aerial vehicles (UAVs), particularly in reactive control policies that serve as a key component of navigation systems. These policies enable agents to respond instantaneously to dynamic environments without relying on pre-existing maps. However, there remain open challenges to improve the agent's reactive control performance: (1) Is it possible and how to anticipate future states at the current moment to benefit control precision? (2) Is it possible and how can we anticipate future states for different sub-tasks when the agent's control consists of both discrete classification and continuous regression commands? Inspired by the Chinese idiom “Mirror Flower, Water Moon,” this paper hypothesizes that future states in the latent space can be learnt from sequential images using contrastive learning, and consequently proposes a light-weight Multi-task Visual Prospective Representation Learning (MulVPRL) framework for benefiting reactive control. Specifically, (1) This paper leverages the advantage of contrastive learning to correlate the representations obtained from the latest sequential images and one image in the future. (2) This paper constructs an integrated loss function of contrastive learning for classification and regression sub-tasks. The MulVPRL framework outperforms the benchmark models on the public HDIN and DroNet datasets, and obtained the best performance in real-world experiments (� � 46.9� � m� � ,� � 177� � svs� � . SOTA � � 27.3� � m� � ,� � 136� � s). Therefore, the multi-task contrastive learning of the light-weight MulVPRL framework enhances reactive control performance on a 2D plane, and demonstrates the potential to be integrated with various intelligent strategies, and implemented on ground vehicles.

基于视觉的深度学习模型已被广泛应用于自主代理,如无人驾驶飞行器(uav),特别是在作为导航系统关键组成部分的响应式控制策略中。这些策略使代理能够即时响应动态环境,而不依赖于预先存在的地图。然而,提高智能体的反应性控制性能仍然存在一些开放的挑战:(1)是否有可能以及如何预测当前时刻的未来状态以提高控制精度?(2)当智能体的控制同时由离散分类和连续回归命令组成时,我们是否可能以及如何预测不同子任务的未来状态?受中国成语“镜花水月”的启发,本文假设潜在空间的未来状态可以通过对比学习从序列图像中学习,并因此提出了一个轻量级的多任务视觉前瞻表征学习(MulVPRL)框架,有利于反应性控制。具体而言,(1)本文利用对比学习的优势,将最新的连续图像和未来的一个图像获得的表征关联起来。(2)构建了分类子任务和回归子任务的对比学习综合损失函数。MulVPRL框架在公共HDIN和DroNet数据集上优于基准模型,并在实际实验中获得了最佳性能(46.9 m)。177 SVS。SOTA 27.3米,136秒)。因此,轻型MulVPRL框架的多任务对比学习增强了二维平面上的无功控制性能,并展示了与各种智能策略集成并在地面车辆上实施的潜力。
{"title":"From “Mirror Flower, Water Moon” to Multi-Task Visual Prospective Representation Learning for Unmanned Aerial Vehicles Indoor Mapless Navigation","authors":"Yingxiu Chang,&nbsp;Yongqiang Cheng,&nbsp;John Murray,&nbsp;Muhammad Khalid,&nbsp;Umar Manzoor","doi":"10.1002/rob.70057","DOIUrl":"https://doi.org/10.1002/rob.70057","url":null,"abstract":"<p>Vision-based deep learning models have been widely adopted in autonomous agents, such as unmanned aerial vehicles (UAVs), particularly in reactive control policies that serve as a key component of navigation systems. These policies enable agents to respond instantaneously to dynamic environments without relying on pre-existing maps. However, there remain open challenges to improve the agent's reactive control performance: (1) Is it possible and how to anticipate future states at the current moment to benefit control precision? (2) Is it possible and how can we anticipate future states for different sub-tasks when the agent's control consists of both discrete classification and continuous regression commands? Inspired by the Chinese idiom “Mirror Flower, Water Moon,” this paper hypothesizes that future states in the latent space can be learnt from sequential images using contrastive learning, and consequently proposes a light-weight Multi-task Visual Prospective Representation Learning (MulVPRL) framework for benefiting reactive control. Specifically, (1) This paper leverages the advantage of contrastive learning to correlate the representations obtained from the latest sequential images and one image in the future. (2) This paper constructs an integrated loss function of contrastive learning for classification and regression sub-tasks. The MulVPRL framework outperforms the benchmark models on the public HDIN and DroNet datasets, and obtained the best performance in real-world experiments (<span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 \u0000 <mrow>\u0000 <mn>46.9</mn>\u0000 <mspace></mspace>\u0000 \u0000 <mi>m</mi>\u0000 \u0000 <mo>,</mo>\u0000 \u0000 <mn>177</mn>\u0000 \u0000 <mi>svs</mi>\u0000 \u0000 <mo>.</mo>\u0000 </mrow>\u0000 </mrow>\u0000 </semantics></math> SOTA <span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 \u0000 <mrow>\u0000 <mn>27.3</mn>\u0000 <mspace></mspace>\u0000 \u0000 <mi>m</mi>\u0000 \u0000 <mo>,</mo>\u0000 \u0000 <mn>136</mn>\u0000 <mspace></mspace>\u0000 \u0000 <mi>s</mi>\u0000 </mrow>\u0000 </mrow>\u0000 </semantics></math>). Therefore, the multi-task contrastive learning of the light-weight MulVPRL framework enhances reactive control performance on a 2D plane, and demonstrates the potential to be integrated with various intelligent strategies, and implemented on ground vehicles.</p>","PeriodicalId":192,"journal":{"name":"Journal of Field Robotics","volume":"43 2","pages":"655-680"},"PeriodicalIF":5.2,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/rob.70057","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146139161","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
INF-SLiM: Large-Scale Implicit Neural Fields for Semantic LiDAR Mapping of Embodied AI Agents INF-SLiM:大规模隐式神经场,用于嵌入人工智能代理的语义激光雷达映射
IF 5.2 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-09-01 DOI: 10.1002/rob.70058
Jianyuan Zhang, Zhiliu Yang, Meng Zhang, Hongyu Chen, Lianhui Zhao, Chen Liu

Large-scale semantic mapping is crucial for outdoor autonomous agents to perform high-level tasks such as planning and navigation. In this paper, we propose a novel method for large-scale 3D semantic reconstruction through implicit representations from posed LiDAR measurements alone. We first construct a neural fields via the octree and latent features, then decode implicit features into signed distance value and semantic information through shallow Multilayer Perceptrons (MLPs). We leverage radial window self-attention networks to predict the semantic labels of point clouds. We then jointly optimize the feature embedding and MLP parameters with a self-supervision paradigm for point-cloud geometry and a pseudo-supervision paradigm for semantic and panoptic labels. Subsequently, geometric structures and semantic categories for novel points in the unseen area are regressed, and the marching cubes method is exploited to subdivide and visualize scenes in the inferring stage, the labeled mesh is produced correspondingly. Experiments on two real-world datasets, SemanticKITTI and SemanticPOSS, demonstrate the superior segmentation efficiency and mapping effectiveness of our framework compared to existing 3D semantic LiDAR mapping methods.

大规模语义映射对于户外自主智能体执行规划和导航等高级任务至关重要。在本文中,我们提出了一种新的方法,用于大规模的三维语义重建,通过隐式表示从提出的激光雷达测量。首先通过八叉树和潜在特征构造神经场,然后通过浅层多层感知器(mlp)将隐式特征解码为带符号距离值和语义信息。我们利用径向窗自关注网络来预测点云的语义标签。然后,我们使用点云几何的自监督范式和语义和全景标签的伪监督范式共同优化特征嵌入和MLP参数。随后,对未见区域中新点的几何结构和语义类别进行回归,并在推理阶段利用行军立方体方法对场景进行细分和可视化,生成相应的标记网格。在两个真实数据集SemanticKITTI和SemanticPOSS上的实验表明,与现有的3D语义LiDAR映射方法相比,我们的框架具有更高的分割效率和映射有效性。
{"title":"INF-SLiM: Large-Scale Implicit Neural Fields for Semantic LiDAR Mapping of Embodied AI Agents","authors":"Jianyuan Zhang,&nbsp;Zhiliu Yang,&nbsp;Meng Zhang,&nbsp;Hongyu Chen,&nbsp;Lianhui Zhao,&nbsp;Chen Liu","doi":"10.1002/rob.70058","DOIUrl":"https://doi.org/10.1002/rob.70058","url":null,"abstract":"<div>\u0000 \u0000 <p>Large-scale semantic mapping is crucial for outdoor autonomous agents to perform high-level tasks such as planning and navigation. In this paper, we propose a novel method for large-scale 3D semantic reconstruction through implicit representations from posed LiDAR measurements alone. We first construct a neural fields via the octree and latent features, then decode implicit features into signed distance value and semantic information through shallow Multilayer Perceptrons (MLPs). We leverage radial window self-attention networks to predict the semantic labels of point clouds. We then jointly optimize the feature embedding and MLP parameters with a self-supervision paradigm for point-cloud geometry and a pseudo-supervision paradigm for semantic and panoptic labels. Subsequently, geometric structures and semantic categories for novel points in the unseen area are regressed, and the marching cubes method is exploited to subdivide and visualize scenes in the inferring stage, the labeled mesh is produced correspondingly. Experiments on two real-world datasets, SemanticKITTI and SemanticPOSS, demonstrate the superior segmentation efficiency and mapping effectiveness of our framework compared to existing 3D semantic LiDAR mapping methods.</p>\u0000 </div>","PeriodicalId":192,"journal":{"name":"Journal of Field Robotics","volume":"43 2","pages":"681-694"},"PeriodicalIF":5.2,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146139125","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Blind Walking Control of a Heavy-Duty Hexapod Robot Based on Optimal Force Allocation for Uneven and Sloped Terrains 基于最优力分配的重型六足机器人不均匀和倾斜地形盲行控制
IF 5.2 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-09-01 DOI: 10.1002/rob.70051
Yufei Liu, Yongyao Li, Lei Jiang, Bo Su, Boyang Xing, Peng Xu

Heavy-duty robots such as hexapod robots exhibit tremendous potential with their high payload ability and terrain adaptability. This paper introduces a control approach for heavy-duty hexapod robot that traverses uneven and transition stage from a flat surface to a sloped terrain without prior geometric knowledge of the environment. Using force sensors on the robot feet and position sensors at its joints, the transition from a flat surface to a sloped terrain is detected. Constraint models based on terrain geometric information and joint limit positions are presented. The models are derived to estimate the unknown slope gradient based on the landing feet positions. The constraints for different terrains have also been investigated. An optimal force allocation method under contact constraints is developed to minimize the linear and quadratic costs in the commands and constrained contact forces. Extensive experiments have been conducted, and the results have demonstrated the effectiveness of the proposed control approach.

六足机器人等重型机器人以其高载荷能力和地形适应性表现出巨大的潜力。本文介绍了一种重型六足机器人的控制方法,该方法可以在不事先了解环境几何知识的情况下,通过不平整和从平坦到倾斜的过渡阶段。利用机器人脚上的力传感器和关节上的位置传感器,检测从平坦表面到倾斜地形的过渡。提出了基于地形几何信息和关节极限位置的约束模型。推导了基于着陆脚位置的未知坡度估计模型。对不同地形的约束条件也进行了研究。提出了一种接触约束下的最优力分配方法,以最大限度地降低指令和约束接触力的线性和二次代价。大量的实验已经进行,结果证明了所提出的控制方法的有效性。
{"title":"Blind Walking Control of a Heavy-Duty Hexapod Robot Based on Optimal Force Allocation for Uneven and Sloped Terrains","authors":"Yufei Liu,&nbsp;Yongyao Li,&nbsp;Lei Jiang,&nbsp;Bo Su,&nbsp;Boyang Xing,&nbsp;Peng Xu","doi":"10.1002/rob.70051","DOIUrl":"https://doi.org/10.1002/rob.70051","url":null,"abstract":"<div>\u0000 \u0000 <p>Heavy-duty robots such as hexapod robots exhibit tremendous potential with their high payload ability and terrain adaptability. This paper introduces a control approach for heavy-duty hexapod robot that traverses uneven and transition stage from a flat surface to a sloped terrain without prior geometric knowledge of the environment. Using force sensors on the robot feet and position sensors at its joints, the transition from a flat surface to a sloped terrain is detected. Constraint models based on terrain geometric information and joint limit positions are presented. The models are derived to estimate the unknown slope gradient based on the landing feet positions. The constraints for different terrains have also been investigated. An optimal force allocation method under contact constraints is developed to minimize the linear and quadratic costs in the commands and constrained contact forces. Extensive experiments have been conducted, and the results have demonstrated the effectiveness of the proposed control approach.</p>\u0000 </div>","PeriodicalId":192,"journal":{"name":"Journal of Field Robotics","volume":"43 2","pages":"599-621"},"PeriodicalIF":5.2,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146139127","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Development of Large Wingspan Flapping-Wing Flying Robots With Improved Efficiency by Rigid-Flexible Coupling Flapping Mechanism 采用刚柔耦合扑翼机构提高效率的大翼展扑翼飞行机器人的研制
IF 5.2 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-08-28 DOI: 10.1002/rob.70054
Siping Zhong, Wenfu Xu, Jizhou Jiang, Zihao Wei, Erzhen Pan

The flight efficiency of bionic flapping-wing robots is largely influenced by the flapping mechanism. This paper presents the development of a large wingspan flapping-wing flying robot that utilizes a rigid-flexible coupling flapping mechanism to reduce energy consumption, improve flight efficiency, and enhance wind resistance. The rigid-flexible coupling flapping mechanism consists of a DC motor with a gearset and a flexible flapping mechanism based on torsion spring. This mechanism combines the high torque driving capability of a rigid mechanism with the energy storage capacity of a flexible mechanism. Synchronizing the passive deformation of the torsion spring with the periodic acceleration and deceleration of the wings during the flapping cycle enables the storage, transfer, and release of kinetic energy and torsional spring elastic strain energy. Simulation studies show that the proposed design effectively reduces the peak power required. A prototype with a wingspan of 1.8 meters and a mass of 1.0 kilograms was developed. Compared to a rigid mechanism, the proposed design significantly reduces the electrical power consumption, especially when achieving flapping frequencies exceeding 1.5 Hz, as validated through wind tunnel experiments. At a flapping frequency of 2.2 Hz, the maximum reduction in electrical power consumption reaches 14.8%. The flexible elements also increased the downstroke ratio, consequently enhancing thrust and propulsion efficiency. Outdoor flight experiments have demonstrated that the prototype propelled by the rigid-flexible coupling drive mechanism consumes only 81.56 W during cruising, which is 8.78 W (9.72%) less than the rigidly driven prototype.

扑翼仿生机器人的飞行效率在很大程度上受扑翼机构的影响。研制了一种大翼展扑翼飞行机器人,该机器人采用刚柔耦合扑翼机构,以降低能耗,提高飞行效率,增强抗风能力。刚柔耦合扑动机构由带齿轮组的直流电动机和基于扭簧的柔性扑动机构组成。该机构结合了刚性机构的高扭矩驱动能力和柔性机构的储能能力。将扭转弹簧的被动变形与机翼在扑动周期中的周期性加减速同步,实现了动能和扭转弹簧弹性应变能的储存、传递和释放。仿真研究表明,该设计有效地降低了所需的峰值功率。研制出翼展1.8米、质量1.0公斤的原型机。与刚性机构相比,该设计显著降低了电力消耗,特别是在实现超过1.5 Hz的扑动频率时,这一点已通过风洞实验得到验证。当扑动频率为2.2 Hz时,电能消耗的最大降幅可达14.8%。柔性元件还增加了下冲程比,从而提高了推力和推进效率。室外飞行实验表明,采用刚柔耦合驱动机构驱动的原型机巡航时能耗仅为81.56 W,比刚性驱动的原型机减少8.78 W(9.72%)。
{"title":"Development of Large Wingspan Flapping-Wing Flying Robots With Improved Efficiency by Rigid-Flexible Coupling Flapping Mechanism","authors":"Siping Zhong,&nbsp;Wenfu Xu,&nbsp;Jizhou Jiang,&nbsp;Zihao Wei,&nbsp;Erzhen Pan","doi":"10.1002/rob.70054","DOIUrl":"https://doi.org/10.1002/rob.70054","url":null,"abstract":"<div>\u0000 \u0000 <p>The flight efficiency of bionic flapping-wing robots is largely influenced by the flapping mechanism. This paper presents the development of a large wingspan flapping-wing flying robot that utilizes a rigid-flexible coupling flapping mechanism to reduce energy consumption, improve flight efficiency, and enhance wind resistance. The rigid-flexible coupling flapping mechanism consists of a DC motor with a gearset and a flexible flapping mechanism based on torsion spring. This mechanism combines the high torque driving capability of a rigid mechanism with the energy storage capacity of a flexible mechanism. Synchronizing the passive deformation of the torsion spring with the periodic acceleration and deceleration of the wings during the flapping cycle enables the storage, transfer, and release of kinetic energy and torsional spring elastic strain energy. Simulation studies show that the proposed design effectively reduces the peak power required. A prototype with a wingspan of 1.8 meters and a mass of 1.0 kilograms was developed. Compared to a rigid mechanism, the proposed design significantly reduces the electrical power consumption, especially when achieving flapping frequencies exceeding 1.5 Hz, as validated through wind tunnel experiments. At a flapping frequency of 2.2 Hz, the maximum reduction in electrical power consumption reaches 14.8%. The flexible elements also increased the downstroke ratio, consequently enhancing thrust and propulsion efficiency. Outdoor flight experiments have demonstrated that the prototype propelled by the rigid-flexible coupling drive mechanism consumes only 81.56 W during cruising, which is 8.78 W (9.72%) less than the rigidly driven prototype.</p>\u0000 </div>","PeriodicalId":192,"journal":{"name":"Journal of Field Robotics","volume":"43 1","pages":"513-525"},"PeriodicalIF":5.2,"publicationDate":"2025-08-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145792471","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Under-Ice Odometry Fusing Acoustic and Visual Information for Underwater Vehicles 水下航行器的冰下里程测量融合声学和视觉信息
IF 5.2 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-08-28 DOI: 10.1002/rob.70052
Du Haohan, Ma Teng, Li Ye, Li Ao, Xu Shuo, Wang Hao, Cao Jian, Liao Yulei

Ice cover hinders the deployment of acoustic arrays or satellites, making it challenging to locate underwater vehicles in the polar regions. Odometry based on acoustic or optical sensors has the potential to provide precise navigation results for vehicles. However, the robustness of single-sensor-based odometry is significantly compromised by the smooth and featureless nature of the ice-covered bottom, which lacks distinct points, lines, or surfaces. This paper presents a filter-based multi-sensor fusion underwater ice odometry system, integrating measurements from a stereo camera, a forward-looking sonar (FLS), a low-cost inertial measurement unit (IMU), and a depth gauge (DG). To balance real-time performance and accuracy, the Multi-State Constraint Kalman Filter (MSCKF) algorithm is employed to fuse inertial and stereo measurements for vehicle state estimation. Additionally, FLS images and relative depth measurements from the DG are utilized to enhance the accuracy and robustness of the state estimation. Specifically, a Fourier-based algorithm is proposed to estimate linear speed using global information from FLS images. A compensation factor is introduced to address the impact of FLS elevation angle on estimation accuracy. Furthermore, a residual-based adaptive method and forgetting factor are implemented to adjust the measurement noise covariance of linear speed and relative depth, improving dynamic state estimation precision. The proposed system was evaluated through tank and field experiments using a remotely operated vehicle (ROV). Results show it provides more accurate and robust under-ice navigation compared to state-of-the-art algorithms.

冰盖阻碍了声学阵列或卫星的部署,使得在极地地区定位水下航行器变得具有挑战性。基于声学或光学传感器的里程计有可能为车辆提供精确的导航结果。然而,基于单传感器的里程计的鲁棒性受到冰层覆盖的底部光滑和无特征的性质的严重损害,因为冰层覆盖的底部缺乏明显的点、线或表面。本文提出了一种基于滤波的多传感器融合水下冰层测程系统,该系统集成了立体摄像机、前视声纳(FLS)、低成本惯性测量单元(IMU)和深度计(DG)的测量数据。为了平衡实时性和准确性,采用多状态约束卡尔曼滤波(MSCKF)算法融合惯性和立体测量进行车辆状态估计。此外,利用FLS图像和DG的相对深度测量来提高状态估计的准确性和鲁棒性。具体而言,提出了一种基于傅里叶的算法,利用FLS图像的全局信息估计线性速度。引入补偿因子来解决FLS仰角对估计精度的影响。采用残差自适应方法和遗忘因子对测量噪声的线性速度和相对深度协方差进行调整,提高了动态估计精度。通过远程操作车辆(ROV)的油箱和现场实验对该系统进行了评估。结果表明,与最先进的算法相比,它提供了更准确、更稳健的冰下导航。
{"title":"Under-Ice Odometry Fusing Acoustic and Visual Information for Underwater Vehicles","authors":"Du Haohan,&nbsp;Ma Teng,&nbsp;Li Ye,&nbsp;Li Ao,&nbsp;Xu Shuo,&nbsp;Wang Hao,&nbsp;Cao Jian,&nbsp;Liao Yulei","doi":"10.1002/rob.70052","DOIUrl":"https://doi.org/10.1002/rob.70052","url":null,"abstract":"<div>\u0000 \u0000 <p>Ice cover hinders the deployment of acoustic arrays or satellites, making it challenging to locate underwater vehicles in the polar regions. Odometry based on acoustic or optical sensors has the potential to provide precise navigation results for vehicles. However, the robustness of single-sensor-based odometry is significantly compromised by the smooth and featureless nature of the ice-covered bottom, which lacks distinct points, lines, or surfaces. This paper presents a filter-based multi-sensor fusion underwater ice odometry system, integrating measurements from a stereo camera, a forward-looking sonar (FLS), a low-cost inertial measurement unit (IMU), and a depth gauge (DG). To balance real-time performance and accuracy, the Multi-State Constraint Kalman Filter (MSCKF) algorithm is employed to fuse inertial and stereo measurements for vehicle state estimation. Additionally, FLS images and relative depth measurements from the DG are utilized to enhance the accuracy and robustness of the state estimation. Specifically, a Fourier-based algorithm is proposed to estimate linear speed using global information from FLS images. A compensation factor is introduced to address the impact of FLS elevation angle on estimation accuracy. Furthermore, a residual-based adaptive method and forgetting factor are implemented to adjust the measurement noise covariance of linear speed and relative depth, improving dynamic state estimation precision. The proposed system was evaluated through tank and field experiments using a remotely operated vehicle (ROV). Results show it provides more accurate and robust under-ice navigation compared to state-of-the-art algorithms.</p>\u0000 </div>","PeriodicalId":192,"journal":{"name":"Journal of Field Robotics","volume":"43 1","pages":"486-512"},"PeriodicalIF":5.2,"publicationDate":"2025-08-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145792470","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
S 2 MAT : Simultaneous and Self-Reinforced Mapping and Tracking in Dynamic Urban Scenarios s2 MAT:动态城市场景中的同步和自增强映射与跟踪
IF 5.2 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-08-28 DOI: 10.1002/rob.70044
Tingxiang Fan, Bowen Shen, Yinqiang Zhang, Chuye Zhang, Lei Yang, Hua Chen, Wei Zhang, Jia Pan

Despite the increasing prevalence of robots in daily life, their navigation capabilities are still limited to environments with prior knowledge, such as a global map. To fully unlock the potential of robots, it is crucial to enable them to navigate in large-scale unknown and changing unstructured scenarios. This requires the robot to construct an accurate static map in real-time as it explores, while filtering out moving objects to ensure mapping accuracy and, if possible, achieving high-quality pedestrian tracking and collision avoidance. While existing methods can achieve individual goals of spatial mapping or dynamic object detection and tracking, there has been limited research on effectively integrating these two tasks, which are actually coupled and reciprocal. In this study, we propose a solution called S2MAT (simultaneous and self-reinforced mapping and tracking) that integrates a front-end dynamic object detection and tracking module with a back-end static mapping module. S2MAT leverages the close and reciprocal interplay between these two modules to efficiently and effectively solve the open problem of simultaneous tracking and mapping in highly dynamic scenarios. The proposed method is primarily designed for use with 3D LiDAR and offers a solution for real-time navigation in large-scale, unknown dynamic scenarios with a low computational cost, making it feasible for deployment on onboard computers equipped with only a single CPU. We conducted long-range experiments in real-world urban scenarios spanning over 7 km, which included challenging obstacles like pedestrians and other traffic agents. The successful navigation provides a comprehensive test of S2MAT's robustness, scalability, efficiency, quality, and its ability to benefit autonomous robots in wild scenarios without pre-built maps.

尽管机器人在日常生活中越来越普遍,但它们的导航能力仍然局限于具有先验知识的环境,例如全球地图。为了充分释放机器人的潜力,使它们能够在大规模未知和不断变化的非结构化场景中导航至关重要。这要求机器人在探索过程中实时构建精确的静态地图,同时过滤掉移动物体以确保地图精度,并在可能的情况下实现高质量的行人跟踪和避碰。虽然现有的方法可以实现空间映射或动态目标检测与跟踪的单独目标,但有效整合这两个任务的研究有限,这两个任务实际上是耦合和相互作用的。在本研究中,我们提出了一种称为S2MAT(同步和自增强映射和跟踪)的解决方案,该解决方案将前端动态目标检测和跟踪模块与后端静态映射模块集成在一起。S2MAT利用这两个模块之间的密切和相互作用,有效地解决了在高动态场景中同时跟踪和映射的开放问题。所提出的方法主要用于3D激光雷达,为大规模未知动态场景的实时导航提供了一种低计算成本的解决方案,使其能够部署在仅配备单个CPU的车载计算机上。我们在超过7公里的真实城市场景中进行了远程实验,其中包括行人和其他交通代理等具有挑战性的障碍物。成功的导航提供了对S2MAT的鲁棒性、可扩展性、效率、质量及其在没有预先构建地图的情况下使自主机器人受益的能力的全面测试。
{"title":"S\u0000 \u0000 2\u0000 \u0000 \u0000 MAT\u0000 \u0000 \u0000 \u0000 \u0000 : Simultaneous and Self-Reinforced Mapping and Tracking in Dynamic Urban Scenarios","authors":"Tingxiang Fan,&nbsp;Bowen Shen,&nbsp;Yinqiang Zhang,&nbsp;Chuye Zhang,&nbsp;Lei Yang,&nbsp;Hua Chen,&nbsp;Wei Zhang,&nbsp;Jia Pan","doi":"10.1002/rob.70044","DOIUrl":"https://doi.org/10.1002/rob.70044","url":null,"abstract":"<div>\u0000 \u0000 <p>Despite the increasing prevalence of robots in daily life, their navigation capabilities are still limited to environments with prior knowledge, such as a global map. To fully unlock the potential of robots, it is crucial to enable them to navigate in large-scale unknown and changing unstructured scenarios. This requires the robot to construct an accurate static map in real-time as it explores, while filtering out moving objects to ensure mapping accuracy and, if possible, achieving high-quality pedestrian tracking and collision avoidance. While existing methods can achieve individual goals of spatial mapping or dynamic object detection and tracking, there has been limited research on effectively integrating these two tasks, which are actually coupled and reciprocal. In this study, we propose a solution called S<sup>2</sup>MAT (simultaneous and self-reinforced mapping and tracking) that integrates a front-end dynamic object detection and tracking module with a back-end static mapping module. S<sup>2</sup>MAT leverages the close and reciprocal interplay between these two modules to efficiently and effectively solve the open problem of simultaneous tracking and mapping in highly dynamic scenarios. The proposed method is primarily designed for use with 3D LiDAR and offers a solution for real-time navigation in large-scale, unknown dynamic scenarios with a low computational cost, making it feasible for deployment on onboard computers equipped with only a single CPU. We conducted long-range experiments in real-world urban scenarios spanning over 7 km, which included challenging obstacles like pedestrians and other traffic agents. The successful navigation provides a comprehensive test of S<sup>2</sup>MAT's robustness, scalability, efficiency, quality, and its ability to benefit autonomous robots in wild scenarios without pre-built maps.</p></div>","PeriodicalId":192,"journal":{"name":"Journal of Field Robotics","volume":"43 1","pages":"454-485"},"PeriodicalIF":5.2,"publicationDate":"2025-08-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145792472","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Design and Experimental Evaluation of a Flexible Wheel-Foot Actuator in Beach Terrain for Amphibious Robots 水陆两栖机器人沙滩地形柔性轮足作动器设计与实验评价
IF 5.2 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-08-25 DOI: 10.1002/rob.70046
Mingxuan Ding, Dake Lu, Junlong Wang, Lingzhe Meng, Liquan Wang, Ling Li, Gang Wang

The efficacy of field exploration robots is contingent upon their capacity and proficiency in traversing unstructured and complex terrains. In the context of beach terrain, substantial variations in the mechanical properties of the substrate can be triggered by fluctuations in moisture levels within granular materials. This poses a substantial challenge to the effectiveness of actuators designed for amphibious crawling, which depend exclusively on plowing force after penetration or surface friction to adapt to such environments. In this study, we propose a cost-effective, adaptable wheel-foot actuator that is designed based on the mechanical properties of granular media. The configuration consists of a circular wheel as the primary structural element and a regular arrangement of flexible feet, which collectively provide a combined traction force of penetration and friction. A total of 3840 experiments were conducted under controlled conditions on both the flexible wheel-foot scheme and the grousered wheel scheme. The outcomes of these experiments enabled us to ascertain and validate the penetration difficulty tendency of wet granules with varying moisture contents. Secondly, a comparison of the transportation cost, normalized antislip coefficient, standard deviation of vertical displacement, energy consumption, and other parameters of 40 schemes under the same working conditions allowed us to determine the structural parameters, such as the number and stiffness of flexible feet. This comparison also demonstrated that the proposed scheme has significant advantages in suppressing slip and reducing energy consumption. In field tests conducted in an amphibious environment, the proposed scheme demonstrated a statistically significant reduction in slip rate (10.4%) and energy consumption (22.8%) compared to other tested schemes. Subsequent evaluations were conducted through tests of water-land transition capability, extrication ability, and the ability to navigate through harsh ground conditions. These evaluations confirmed the effectiveness of the proposed scheme.

野外勘探机器人的效能取决于它们穿越非结构化和复杂地形的能力和熟练程度。在海滩地形的情况下,颗粒材料内水分水平的波动可能引发基材机械性能的实质性变化。这对设计用于两栖爬行的致动器的有效性提出了实质性的挑战,这些致动器完全依赖于穿透后的犁力或表面摩擦来适应这种环境。在这项研究中,我们提出了一种基于颗粒介质力学特性设计的具有成本效益,适应性强的轮足驱动器。该结构由圆形车轮作为主要结构元件和规则排列的柔性脚组成,它们共同提供穿透和摩擦的联合牵引力。在控制条件下,对柔性轮足方案和槽轮方案进行了3840次试验。这些实验的结果使我们能够确定和验证不同含水率的湿颗粒的渗透困难趋势。其次,比较40种方案在相同工况下的运输成本、标准化防滑系数、垂直位移标准差、能耗等参数,确定柔性脚数量、刚度等结构参数。结果表明,该方案在抑制滑移和降低能耗方面具有显著的优势。在两栖环境中进行的现场测试中,与其他测试方案相比,该方案在统计上显着降低了滑移率(10.4%)和能耗(22.8%)。随后的评价是通过测试水陆过渡能力、排水能力和在恶劣地面条件下航行的能力进行的。这些评价证实了拟议方案的有效性。
{"title":"Design and Experimental Evaluation of a Flexible Wheel-Foot Actuator in Beach Terrain for Amphibious Robots","authors":"Mingxuan Ding,&nbsp;Dake Lu,&nbsp;Junlong Wang,&nbsp;Lingzhe Meng,&nbsp;Liquan Wang,&nbsp;Ling Li,&nbsp;Gang Wang","doi":"10.1002/rob.70046","DOIUrl":"https://doi.org/10.1002/rob.70046","url":null,"abstract":"<div>\u0000 \u0000 <p>The efficacy of field exploration robots is contingent upon their capacity and proficiency in traversing unstructured and complex terrains. In the context of beach terrain, substantial variations in the mechanical properties of the substrate can be triggered by fluctuations in moisture levels within granular materials. This poses a substantial challenge to the effectiveness of actuators designed for amphibious crawling, which depend exclusively on plowing force after penetration or surface friction to adapt to such environments. In this study, we propose a cost-effective, adaptable wheel-foot actuator that is designed based on the mechanical properties of granular media. The configuration consists of a circular wheel as the primary structural element and a regular arrangement of flexible feet, which collectively provide a combined traction force of penetration and friction. A total of 3840 experiments were conducted under controlled conditions on both the flexible wheel-foot scheme and the grousered wheel scheme. The outcomes of these experiments enabled us to ascertain and validate the penetration difficulty tendency of wet granules with varying moisture contents. Secondly, a comparison of the transportation cost, normalized antislip coefficient, standard deviation of vertical displacement, energy consumption, and other parameters of 40 schemes under the same working conditions allowed us to determine the structural parameters, such as the number and stiffness of flexible feet. This comparison also demonstrated that the proposed scheme has significant advantages in suppressing slip and reducing energy consumption. In field tests conducted in an amphibious environment, the proposed scheme demonstrated a statistically significant reduction in slip rate (10.4%) and energy consumption (22.8%) compared to other tested schemes. Subsequent evaluations were conducted through tests of water-land transition capability, extrication ability, and the ability to navigate through harsh ground conditions. These evaluations confirmed the effectiveness of the proposed scheme.</p>\u0000 </div>","PeriodicalId":192,"journal":{"name":"Journal of Field Robotics","volume":"43 1","pages":"421-441"},"PeriodicalIF":5.2,"publicationDate":"2025-08-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145792466","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Degeneracy-Resistant LiDAR-SLAM Algorithm Based on Geometric and Visual Features' Fusion 基于几何特征与视觉特征融合的抗退化LiDAR-SLAM算法
IF 5.2 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-08-25 DOI: 10.1002/rob.70026
Daping Yang, Wenzhong Shi, Shuyu Zhang, Mingyan Nie, Yitao Wei, Qiru Zhong, Shengyu Lu, Ameer Hamza Khan

Simultaneous localization and mapping (SLAM) is at the core of robotics automation, relying on sensors such as Laser Detection and Ranging (LiDAR) and cameras to digitally construct a robot's environment and determine its position within. LiDAR-based SLAM outperforms visual-SLAM, especially in low visibility and challenging lighting conditions. However, these systems still face challenges like scene degradation when dealing with feature-deficient degenerate environments such as long corridors or tunnels. Traditional LiDAR SLAM algorithms primarily focus on the extraction of geometric features from the scene, with less utilization of visual information, for example, LiDAR-generated reflectivity (also commonly referred to as intensity image) and depth imagery. In this study, we explore the potential of fusing both geometric and LiDAR-generated image features into the SLAM system in various forms, aiming to enhance the system's adaptability in diverse environments and its robustness against environment degeneracy. We propose a new multifeature-modality SLAM designed for robust real-time localization and mapping in challenging environments. Our method enhances and extracts visual features from LiDAR-generated images, which are then fused with geometric features through a holistic residual function for pose optimization. We also integrate a deep learning-based object removal algorithm to reduce sensitivity to moving objects and sensor noise. This article conducts an in-depth comparison of the proposed algorithm with several leading technologies in terms of scan matching accuracy, robustness, odometry, and mapping. The experimental results vividly showcase the superiority of our method in achieving high scan matching success rates and strong resilience against random outliers and Gaussian noise across various challenging scenarios, compared to the existing LiDAR SLAM methods that rely solely on geometric features. Extensive field experiments conducted on publicly available data sets, along with independently developed backpack-based and robotic platforms, validated the robustness and accuracy of the proposed approach in both indoor and outdoor environments. In 3D mapping, we quantified the precision of 3D points by comparing point clouds collected by high-precision Mobile Laser Scanning (MLS) and Terrestrial Laser Scanning (TLS). Our method outperforms in terms of absolute pose errors (APE) and point cloud matching quality. Based on the fitted Weibull distribution, the root mean square error (RMS) of point-to-plane distances improved by 20%. Additionally, ablation tests revealed the efficacy of different components within our system.

同时定位和绘图(SLAM)是机器人自动化的核心,依靠激光探测和测距(LiDAR)和摄像头等传感器以数字方式构建机器人的环境并确定其在其中的位置。基于激光雷达的SLAM优于视觉SLAM,特别是在低能见度和具有挑战性的照明条件下。然而,这些系统在处理特征缺乏的退化环境(如长走廊或隧道)时仍然面临着场景退化等挑战。传统的LiDAR SLAM算法主要侧重于从场景中提取几何特征,对视觉信息的利用较少,例如LiDAR生成的反射率(通常也称为强度图像)和深度图像。在本研究中,我们探索了将几何和激光雷达生成的图像特征以各种形式融合到SLAM系统中的潜力,旨在增强系统在不同环境中的适应性和对环境退化的鲁棒性。我们提出了一种新的多特征模态SLAM,用于在具有挑战性的环境中进行鲁棒的实时定位和映射。该方法从激光雷达生成的图像中增强和提取视觉特征,然后通过整体残差函数将视觉特征与几何特征融合,进行姿态优化。我们还集成了一种基于深度学习的物体去除算法,以降低对移动物体和传感器噪声的敏感性。本文在扫描匹配精度、鲁棒性、里程计和映射方面对所提出的算法与几种领先技术进行了深入的比较。实验结果生动地展示了我们的方法在各种具有挑战性的场景下实现高扫描匹配成功率和对随机异常值和高斯噪声的强弹性方面的优势,与现有的仅依赖几何特征的LiDAR SLAM方法相比。在公开可用的数据集上进行了大量的现场实验,以及独立开发的基于背包和机器人的平台,验证了该方法在室内和室外环境下的鲁棒性和准确性。在三维制图中,我们通过比较高精度移动激光扫描(MLS)和地面激光扫描(TLS)收集的点云,量化了三维点的精度。我们的方法在绝对位姿误差(APE)和点云匹配质量方面优于其他方法。基于拟合的威布尔分布,点平面距离的均方根误差(RMS)提高了20%。此外,消融测试揭示了我们系统中不同成分的功效。
{"title":"Degeneracy-Resistant LiDAR-SLAM Algorithm Based on Geometric and Visual Features' Fusion","authors":"Daping Yang,&nbsp;Wenzhong Shi,&nbsp;Shuyu Zhang,&nbsp;Mingyan Nie,&nbsp;Yitao Wei,&nbsp;Qiru Zhong,&nbsp;Shengyu Lu,&nbsp;Ameer Hamza Khan","doi":"10.1002/rob.70026","DOIUrl":"https://doi.org/10.1002/rob.70026","url":null,"abstract":"<div>\u0000 \u0000 <p>Simultaneous localization and mapping (SLAM) is at the core of robotics automation, relying on sensors such as Laser Detection and Ranging (LiDAR) and cameras to digitally construct a robot's environment and determine its position within. LiDAR-based SLAM outperforms visual-SLAM, especially in low visibility and challenging lighting conditions. However, these systems still face challenges like scene degradation when dealing with feature-deficient degenerate environments such as long corridors or tunnels. Traditional LiDAR SLAM algorithms primarily focus on the extraction of geometric features from the scene, with less utilization of visual information, for example, LiDAR-generated reflectivity (also commonly referred to as intensity image) and depth imagery. In this study, we explore the potential of fusing both geometric and LiDAR-generated image features into the SLAM system in various forms, aiming to enhance the system's adaptability in diverse environments and its robustness against environment degeneracy. We propose a new multifeature-modality SLAM designed for robust real-time localization and mapping in challenging environments. Our method enhances and extracts visual features from LiDAR-generated images, which are then fused with geometric features through a holistic residual function for pose optimization. We also integrate a deep learning-based object removal algorithm to reduce sensitivity to moving objects and sensor noise. This article conducts an in-depth comparison of the proposed algorithm with several leading technologies in terms of scan matching accuracy, robustness, odometry, and mapping. The experimental results vividly showcase the superiority of our method in achieving high scan matching success rates and strong resilience against random outliers and Gaussian noise across various challenging scenarios, compared to the existing LiDAR SLAM methods that rely solely on geometric features. Extensive field experiments conducted on publicly available data sets, along with independently developed backpack-based and robotic platforms, validated the robustness and accuracy of the proposed approach in both indoor and outdoor environments. In 3D mapping, we quantified the precision of 3D points by comparing point clouds collected by high-precision Mobile Laser Scanning (MLS) and Terrestrial Laser Scanning (TLS). Our method outperforms in terms of absolute pose errors (APE) and point cloud matching quality. Based on the fitted Weibull distribution, the root mean square error (RMS) of point-to-plane distances improved by 20%. Additionally, ablation tests revealed the efficacy of different components within our system.</p>\u0000 </div>","PeriodicalId":192,"journal":{"name":"Journal of Field Robotics","volume":"43 1","pages":"330-352"},"PeriodicalIF":5.2,"publicationDate":"2025-08-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145792462","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Development and Field Evaluation of a Hydraulic Arm Integrated With an E-Vehicle for Robotic Cotton Picker 摘棉机器人液压臂与电动车辆集成的研制与现场评价
IF 5.2 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-08-25 DOI: 10.1002/rob.70045
Rahul Yadav, Manjeet Singh, Aseem Verma, Naresh Kumar Chhuneja, Derminder Singh, Paramjit Singh

In cotton picking, there is a need for a robotic arm that can pick bolls efficiently with minimal loss and damage to immature bolls. This study addresses that gap by developing a hydraulic arm to work effectively in unstructured cotton fields. A three-degrees-of-freedom hydraulic arm was developed with vertical, radial, and rotational movements, equipped with a roller-type end-effector and vacuum suction system to pick cotton bolls from the plants and collect them into a storage tank. The performance of the hydraulic arm was evaluated in the field under varying conditions of plant-to-plant spacing (250, 350, and 450 mm) and at suction pressures (25, 50, and 75 mmHg). Key performance parameters such as picking efficiency, picking capacity, damage of green bolls, losses of cotton bolls, and trash content were evaluated. The results indicated that the maximum picking efficiency (92.79%) and capacity (567 bolls picked/h) were achieved at 450 mm plant-to-plant spacing and 75 mmHg suction pressure, with minimal damage of green bolls (126 bolls/ha) and losses of cotton bolls (3.79%). The lowest trash content (2.06%) was observed at 450 mm spacing and 25 mmHg suction pressure. Finally, the hydraulic arm achieved precise and accurate cotton picking in complex cotton fields, showing the best performance at 450 mm plant-to-plant spacing and 75 mmHg suction pressure.

在棉花采摘中,需要一种机械臂,能够有效地采摘棉铃,同时将未成熟棉铃的损失和损坏降到最低。这项研究通过开发一种液压臂来解决这一差距,使其在非结构化棉田中有效地工作。开发了一个具有垂直、径向和旋转运动的三自由度液压臂,配备了滚轮式末端执行器和真空吸吸系统,从植株上采摘棉铃并将其收集到储罐中。在不同的工厂间距(250、350和450 mm)和吸入压力(25、50和75 mmHg)条件下,对液压臂的性能进行了现场评估。对采收效率、采收能力、青铃损失率、棉铃损失率、垃圾含量等关键性能参数进行了评价。结果表明,在450mm株距和75mmhg吸压条件下,青铃损伤最小(126铃/ha),棉铃损失最小(3.79%),单株采摘效率最高(92.79%),产量最高(567铃/h)。在450 mm间距和25 mmHg吸压条件下,垃圾含量最低,为2.06%。最后,液压臂在复杂的棉田中实现了精确的棉花采摘,在450mm株间间距和75mmhg吸压下表现最佳。
{"title":"Development and Field Evaluation of a Hydraulic Arm Integrated With an E-Vehicle for Robotic Cotton Picker","authors":"Rahul Yadav,&nbsp;Manjeet Singh,&nbsp;Aseem Verma,&nbsp;Naresh Kumar Chhuneja,&nbsp;Derminder Singh,&nbsp;Paramjit Singh","doi":"10.1002/rob.70045","DOIUrl":"https://doi.org/10.1002/rob.70045","url":null,"abstract":"<div>\u0000 \u0000 <p>In cotton picking, there is a need for a robotic arm that can pick bolls efficiently with minimal loss and damage to immature bolls. This study addresses that gap by developing a hydraulic arm to work effectively in unstructured cotton fields. A three-degrees-of-freedom hydraulic arm was developed with vertical, radial, and rotational movements, equipped with a roller-type end-effector and vacuum suction system to pick cotton bolls from the plants and collect them into a storage tank. The performance of the hydraulic arm was evaluated in the field under varying conditions of plant-to-plant spacing (250, 350, and 450 mm) and at suction pressures (25, 50, and 75 mmHg). Key performance parameters such as picking efficiency, picking capacity, damage of green bolls, losses of cotton bolls, and trash content were evaluated. The results indicated that the maximum picking efficiency (92.79%) and capacity (567 bolls picked/h) were achieved at 450 mm plant-to-plant spacing and 75 mmHg suction pressure, with minimal damage of green bolls (126 bolls/ha) and losses of cotton bolls (3.79%). The lowest trash content (2.06%) was observed at 450 mm spacing and 25 mmHg suction pressure. Finally, the hydraulic arm achieved precise and accurate cotton picking in complex cotton fields, showing the best performance at 450 mm plant-to-plant spacing and 75 mmHg suction pressure.</p>\u0000 </div>","PeriodicalId":192,"journal":{"name":"Journal of Field Robotics","volume":"43 1","pages":"409-420"},"PeriodicalIF":5.2,"publicationDate":"2025-08-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145792464","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Field Robotics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1