首页 > 最新文献

IEEE Robotics and Automation Letters最新文献

英文 中文
RoadRunner M&M - Learning Multi-Range Multi-Resolution Traversability Maps for Autonomous Off-Road Navigation RoadRunner M&M - 学习多范围多分辨率可穿越性地图,用于自主越野导航
IF 4.6 2区 计算机科学 Q2 ROBOTICS Pub Date : 2024-11-01 DOI: 10.1109/LRA.2024.3490404
Manthan Patel;Jonas Frey;Deegan Atha;Patrick Spieler;Marco Hutter;Shehryar Khattak
Autonomous robot navigation in off–road environments requires a comprehensive understanding of the terrain geometry and traversability. The degraded perceptual conditions and sparse geometric information at longer ranges make the problem challenging especially when driving at high speeds. Furthermore, the sensing–to–mapping latency and the look–ahead map range can limit the maximum speed of the vehicle. Building on top of the recent work RoadRunner, in this work, we address the challenge of long-range ($pm 100 ,text{m}$) traversability estimation. Our RoadRunner (M&M) is an end-to-end learning-based framework that directly predicts the traversability and elevation maps at multiple ranges ($pm 50 ,text{m}$, $pm 100 ,text{m}$) and resolutions ($0.2 ,text{m}$, $0.8 ,text{m}$) taking as input multiple images and a LiDAR voxel map. Our method is trained in a self–supervised manner by leveraging the dense supervision signal generated by fusing predictions from an existing traversability estimation stack (X-Racer) in hindsight and satellite Digital Elevation Maps. RoadRunner M&M achieves a significant improvement of up to 50% for elevation mapping and 30% for traversability estimation over RoadRunner, and is able to predict in 30% more regions compared to X-Racer while achieving real–time performance. Experiments on various out–of–distribution datasets also demonstrate that our data-driven approach starts to generalize to novel unstructured environments. We integrate our proposed framework in closed–loop with the path planner to demonstrate autonomous high–speed off–road robotic navigation in challenging real–world environments.
机器人在越野环境中自主导航需要全面了解地形的几何形状和可穿越性。在较远距离上,感知条件的退化和几何信息的稀疏使问题变得具有挑战性,尤其是在高速行驶时。此外,从感知到绘图的延迟和前视地图范围也会限制车辆的最高速度。在最近的研究成果 RoadRunner 的基础上,我们在本研究中解决了远距离($pm 100 ,text{m}$)可穿越性估计的难题。我们的RoadRunner (M&M)是一个基于端到端学习的框架,可以直接预测多种范围($pm 50 text{m}$、$pm 100 text{m}$)和分辨率($0.2 text{m}$、$0.8 text{m}$)下的可穿越性和高程图,并将多幅图像和LiDAR体素图作为输入。我们的方法是以自我监督的方式进行训练的,即利用现有的可穿越性估计堆栈(X-Racer)在事后视角和卫星数字高程图中融合预测所产生的密集监督信号。与 RoadRunner 相比,RoadRunner M&M 在高程测绘和可穿越性估算方面分别实现了高达 50% 和 30% 的显著改进,与 X-Racer 相比,能够预测更多 30% 的区域,同时实现了实时性能。在各种非分布数据集上进行的实验还表明,我们的数据驱动方法开始适用于新型非结构化环境。我们将所提出的框架与路径规划器进行闭环整合,演示了在具有挑战性的真实环境中的自主高速越野机器人导航。
{"title":"RoadRunner M&M - Learning Multi-Range Multi-Resolution Traversability Maps for Autonomous Off-Road Navigation","authors":"Manthan Patel;Jonas Frey;Deegan Atha;Patrick Spieler;Marco Hutter;Shehryar Khattak","doi":"10.1109/LRA.2024.3490404","DOIUrl":"https://doi.org/10.1109/LRA.2024.3490404","url":null,"abstract":"Autonomous robot navigation in off–road environments requires a comprehensive understanding of the terrain geometry and traversability. The degraded perceptual conditions and sparse geometric information at longer ranges make the problem challenging especially when driving at high speeds. Furthermore, the sensing–to–mapping latency and the look–ahead map range can limit the maximum speed of the vehicle. Building on top of the recent work RoadRunner, in this work, we address the challenge of long-range (\u0000<inline-formula><tex-math>$pm 100 ,text{m}$</tex-math></inline-formula>\u0000) traversability estimation. Our RoadRunner (M&M) is an end-to-end learning-based framework that directly predicts the traversability and elevation maps at multiple ranges (\u0000<inline-formula><tex-math>$pm 50 ,text{m}$</tex-math></inline-formula>\u0000, \u0000<inline-formula><tex-math>$pm 100 ,text{m}$</tex-math></inline-formula>\u0000) and resolutions (\u0000<inline-formula><tex-math>$0.2 ,text{m}$</tex-math></inline-formula>\u0000, \u0000<inline-formula><tex-math>$0.8 ,text{m}$</tex-math></inline-formula>\u0000) taking as input multiple images and a LiDAR voxel map. Our method is trained in a self–supervised manner by leveraging the dense supervision signal generated by fusing predictions from an existing traversability estimation stack (X-Racer) in hindsight and satellite Digital Elevation Maps. RoadRunner M&M achieves a significant improvement of up to 50% for elevation mapping and 30% for traversability estimation over RoadRunner, and is able to predict in 30% more regions compared to X-Racer while achieving real–time performance. Experiments on various out–of–distribution datasets also demonstrate that our data-driven approach starts to generalize to novel unstructured environments. We integrate our proposed framework in closed–loop with the path planner to demonstrate autonomous high–speed off–road robotic navigation in challenging real–world environments.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"9 12","pages":"11425-11432"},"PeriodicalIF":4.6,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142636431","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Bimanual Grasp Synthesis for Dexterous Robot Hands 灵巧机器手的双臂抓握合成技术
IF 4.6 2区 计算机科学 Q2 ROBOTICS Pub Date : 2024-11-01 DOI: 10.1109/LRA.2024.3490393
Yanming Shao;Chenxi Xiao
Humans naturally perform bimanual skills to handle large and heavy objects. To enhance robots' object manipulation capabilities, generating effective bimanual grasp poses is essential. Nevertheless, bimanual grasp synthesis for dexterous hand manipulators remains underexplored. To bridge this gap, we propose the BimanGrasp algorithm for synthesizing bimanual grasps on 3D objects. The BimanGrasp algorithm generates grasp poses by optimizing an energy function that considers grasp stability and feasibility. Furthermore, the synthesized grasps are verified using the Isaac Gym physics simulation engine. These verified grasp poses form the BimanGrasp-Dataset, the first large-scale synthesized bimanual dexterous hand grasp pose dataset to our knowledge. The dataset comprises over 150k verified grasps on 900 objects, facilitating the synthesis of bimanual grasps through a data-driven approach. Last, we propose BimanGrasp-DDPM, a diffusion model trained on the BimanGrasp-Dataset. This model achieved a grasp synthesis success rate of 69.87% and significant acceleration in computational speed compared to BimanGrasp algorithm.
人类在处理大而重的物体时自然会使用双臂技能。为了提高机器人的物体操纵能力,生成有效的双臂抓握姿势至关重要。然而,针对灵巧手部机械手的双臂抓握合成仍未得到充分探索。为了弥补这一不足,我们提出了 BimanGrasp 算法,用于合成三维物体上的双臂抓握姿势。BimanGrasp 算法通过优化能量函数生成抓握姿势,该函数考虑了抓握的稳定性和可行性。此外,还使用 Isaac Gym 物理模拟引擎对合成的抓取姿势进行验证。这些经过验证的抓握姿势形成了 BimanGrasp 数据集,这是我们所知的首个大规模合成双灵巧手抓握姿势数据集。该数据集包含 900 个物体上超过 15 万个经过验证的抓取姿势,有助于通过数据驱动方法合成双手抓取姿势。最后,我们提出了在 BimanGrasp 数据集上训练的扩散模型 BimanGrasp-DDPM。与 BimanGrasp 算法相比,该模型的抓取合成成功率高达 69.87%,计算速度显著加快。
{"title":"Bimanual Grasp Synthesis for Dexterous Robot Hands","authors":"Yanming Shao;Chenxi Xiao","doi":"10.1109/LRA.2024.3490393","DOIUrl":"https://doi.org/10.1109/LRA.2024.3490393","url":null,"abstract":"Humans naturally perform bimanual skills to handle large and heavy objects. To enhance robots' object manipulation capabilities, generating effective bimanual grasp poses is essential. Nevertheless, bimanual grasp synthesis for dexterous hand manipulators remains underexplored. To bridge this gap, we propose the BimanGrasp algorithm for synthesizing bimanual grasps on 3D objects. The BimanGrasp algorithm generates grasp poses by optimizing an energy function that considers grasp stability and feasibility. Furthermore, the synthesized grasps are verified using the Isaac Gym physics simulation engine. These verified grasp poses form the BimanGrasp-Dataset, the first large-scale synthesized bimanual dexterous hand grasp pose dataset to our knowledge. The dataset comprises over 150k verified grasps on 900 objects, facilitating the synthesis of bimanual grasps through a data-driven approach. Last, we propose BimanGrasp-DDPM, a diffusion model trained on the BimanGrasp-Dataset. This model achieved a grasp synthesis success rate of 69.87% and significant acceleration in computational speed compared to BimanGrasp algorithm.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"9 12","pages":"11377-11384"},"PeriodicalIF":4.6,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142636537","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Flexible Affine Formation Control Based on Dynamic Hierarchical Reorganization 基于动态分层重组的灵活仿射编队控制
IF 4.6 2区 计算机科学 Q2 ROBOTICS Pub Date : 2024-11-01 DOI: 10.1109/LRA.2024.3490407
Yuzhu Li;Wei Dong
Current formations commonly rely on invariant hierarchical structures, such as predetermined leaders or enumerated formation shapes. These structures could be unidirectional and sluggish, constraining their flexibility and agility when encountering cluttered environments. To surmount these constraints, this work proposes a dynamic hierarchical reorganization approach with affine formation. Central to our approach is the fluid leadership and authority redistribution, for which we develop a minimum time-driven leadership evaluation algorithm and a power transition control algorithm. These algorithms facilitate autonomous leader selection and ensure smooth power transitions, enabling the swarm to adapt hierarchically in alignment with the external environment. Extensive simulations and real-world experiments validate the effectiveness of the proposed method. The formation of five aerial robots successfully performs dynamic hierarchical reorganizations, enabling the execution of complex tasks such as swerving maneuvers and navigating through hoops at velocities of up to 1.05m/s. Comparative experimental results further demonstrate the significant advantages of hierarchical reorganization in enhancing formation flexibility and agility, particularly during complex maneuvers such as U-turns. Notably, in the aforementioned real-world experiments, the proposed method reduces the flight path length by at least 33.8% compared to formations without hierarchical reorganization.
目前的编队通常依赖于不变的层次结构,如预先确定的领导或列举的编队形状。这些结构可能是单向和迟缓的,在遇到杂乱环境时限制了它们的灵活性和敏捷性。为了克服这些限制,本研究提出了一种仿射编队的动态分层重组方法。我们的方法的核心是流畅的领导力和权力再分配,为此我们开发了一种由最短时间驱动的领导力评估算法和一种权力转换控制算法。这些算法有利于自主选择领导者,确保权力平稳过渡,使蜂群能够根据外部环境进行分层调整。大量的模拟和实际实验验证了所提方法的有效性。由五个空中机器人组成的编队成功地进行了动态分层重组,从而能够执行复杂的任务,如以高达 1.05 米/秒的速度进行转弯机动和环形导航。对比实验结果进一步证明了分层重组在提高编队灵活性和敏捷性方面的显著优势,尤其是在 U 形转弯等复杂机动过程中。值得注意的是,在上述真实世界的实验中,与没有进行分层重组的编队相比,所提出的方法至少减少了 33.8% 的飞行路径长度。
{"title":"Flexible Affine Formation Control Based on Dynamic Hierarchical Reorganization","authors":"Yuzhu Li;Wei Dong","doi":"10.1109/LRA.2024.3490407","DOIUrl":"https://doi.org/10.1109/LRA.2024.3490407","url":null,"abstract":"Current formations commonly rely on invariant hierarchical structures, such as predetermined leaders or enumerated formation shapes. These structures could be unidirectional and sluggish, constraining their flexibility and agility when encountering cluttered environments. To surmount these constraints, this work proposes a dynamic hierarchical reorganization approach with affine formation. Central to our approach is the fluid leadership and authority redistribution, for which we develop a minimum time-driven leadership evaluation algorithm and a power transition control algorithm. These algorithms facilitate autonomous leader selection and ensure smooth power transitions, enabling the swarm to adapt hierarchically in alignment with the external environment. Extensive simulations and real-world experiments validate the effectiveness of the proposed method. The formation of five aerial robots successfully performs dynamic hierarchical reorganizations, enabling the execution of complex tasks such as swerving maneuvers and navigating through hoops at velocities of up to 1.05m/s. Comparative experimental results further demonstrate the significant advantages of hierarchical reorganization in enhancing formation flexibility and agility, particularly during complex maneuvers such as U-turns. Notably, in the aforementioned real-world experiments, the proposed method reduces the flight path length by at least 33.8% compared to formations without hierarchical reorganization.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"9 12","pages":"11290-11297"},"PeriodicalIF":4.6,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142598645","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PEP: Policy-Embedded Trajectory Planning for Autonomous Driving PEP:自主驾驶的政策嵌入式轨迹规划
IF 4.6 2区 计算机科学 Q2 ROBOTICS Pub Date : 2024-11-01 DOI: 10.1109/LRA.2024.3490377
Dongkun Zhang;Jiaming Liang;Sha Lu;Ke Guo;Qi Wang;Rong Xiong;Zhenwei Miao;Yue Wang
Autonomous driving demands proficient trajectory planning to ensure safety and comfort. This letter introduces Policy-Embedded Planner (PEP), a novel framework that enhances closed-loop performance of imitation learning (IL) based planners by embedding a neural policy for sequential ego pose generation, leveraging predicted trajectories of traffic agents. PEP addresses the challenges of distribution shift and causal confusion by decomposing multi-step planning into single-step policy rollouts, applying a coordinate transformation technique to simplify training. PEP allows for the parallel generation of multi-modal candidate trajectories and incorporates both neural and rule-based scoring functions for trajectory selection. To mitigate the negative effects of prediction error on closed-loop performance, we propose an information-mixing mechanism that alternates the utilization of traffic agents' predicted and ground-truth information during training. Experimental validations on nuPlan benchmark highlight PEP's superiority over IL- and rule-based state-of-the-art methods.
自动驾驶需要熟练的轨迹规划,以确保安全性和舒适性。这篇文章介绍了政策嵌入式规划器(PEP),这是一种新颖的框架,通过嵌入神经政策,利用交通参与者的预测轨迹,按顺序生成自我姿态,从而提高基于模仿学习(IL)的规划器的闭环性能。PEP 将多步规划分解为单步策略滚动,应用坐标变换技术简化训练,从而解决了分布偏移和因果混淆的难题。PEP 允许并行生成多模式候选轨迹,并结合神经和基于规则的评分函数进行轨迹选择。为了减轻预测误差对闭环性能的负面影响,我们提出了一种信息混合机制,在训练过程中交替使用交通代理的预测信息和地面实况信息。在 nuPlan 基准上的实验验证凸显了 PEP 优于基于 IL 和规则的最先进方法。
{"title":"PEP: Policy-Embedded Trajectory Planning for Autonomous Driving","authors":"Dongkun Zhang;Jiaming Liang;Sha Lu;Ke Guo;Qi Wang;Rong Xiong;Zhenwei Miao;Yue Wang","doi":"10.1109/LRA.2024.3490377","DOIUrl":"https://doi.org/10.1109/LRA.2024.3490377","url":null,"abstract":"Autonomous driving demands proficient trajectory planning to ensure safety and comfort. This letter introduces Policy-Embedded Planner (PEP), a novel framework that enhances closed-loop performance of imitation learning (IL) based planners by embedding a neural policy for sequential ego pose generation, leveraging predicted trajectories of traffic agents. PEP addresses the challenges of distribution shift and causal confusion by decomposing multi-step planning into single-step policy rollouts, applying a coordinate transformation technique to simplify training. PEP allows for the parallel generation of multi-modal candidate trajectories and incorporates both neural and rule-based scoring functions for trajectory selection. To mitigate the negative effects of prediction error on closed-loop performance, we propose an information-mixing mechanism that alternates the utilization of traffic agents' predicted and ground-truth information during training. Experimental validations on nuPlan benchmark highlight PEP's superiority over IL- and rule-based state-of-the-art methods.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"9 12","pages":"11361-11368"},"PeriodicalIF":4.6,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142636255","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TICMapNet: A Tightly Coupled Temporal Fusion Pipeline for Vectorized HD Map Learning TICMapNet:用于矢量化高清地图学习的紧密耦合时态融合管道
IF 4.6 2区 计算机科学 Q2 ROBOTICS Pub Date : 2024-11-01 DOI: 10.1109/LRA.2024.3490384
Wenzhao Qiu;Shanmin Pang;Hao Zhang;Jianwu Fang;Jianru Xue
High-Definition (HD) map construction is essential for autonomous driving to accurately understand the surrounding environment. Most existing methods rely on single-frame inputs to predict local map, which often fail to effectively capture the temporal correlations between frames. This limitation results in discontinuities and instability in the generated map.To tackle this limitation, we propose a Tightly Coupled temporal fusion Map Network (TICMapNet). TICMapNet breaks down the fusion process into three sub-problems: PV feature alignment, BEV feature adjustment, and Query feature fusion. By doing so, we effectively integrate temporal information at different stages through three plug-and-play modules, using the proposed tightly coupled strategy. Unlike traditional methods, our approach does not rely on camera extrinsic parameters, offering a new perspective for addressing the visual fusion challenge in the field of object detection. Experimental results show that TICMapNet significantly improves upon its single-frame baseline model, achieving at least a 7.0% increase in mAP using just two consecutive frames on the nuScenes dataset, while also showing generalizability across other tasks.
高清(HD)地图的构建对于自动驾驶准确了解周围环境至关重要。现有的大多数方法都依赖于单帧输入来预测局部地图,这往往无法有效捕捉帧与帧之间的时间相关性。为了解决这一问题,我们提出了紧密耦合时空融合地图网络(TICMapNet)。TICMapNet 将融合过程分解为三个子问题:PV 特征对齐、BEV 特征调整和查询特征融合。这样,我们通过三个即插即用的模块,利用提出的紧密耦合策略,有效地整合了不同阶段的时间信息。与传统方法不同,我们的方法不依赖相机外在参数,为解决物体检测领域的视觉融合难题提供了一个新的视角。实验结果表明,TICMapNet 显著改善了其单帧基线模型,在 nuScenes 数据集上仅使用两个连续帧就实现了至少 7.0% 的 mAP 提升,同时还显示了在其他任务中的通用性。
{"title":"TICMapNet: A Tightly Coupled Temporal Fusion Pipeline for Vectorized HD Map Learning","authors":"Wenzhao Qiu;Shanmin Pang;Hao Zhang;Jianwu Fang;Jianru Xue","doi":"10.1109/LRA.2024.3490384","DOIUrl":"https://doi.org/10.1109/LRA.2024.3490384","url":null,"abstract":"High-Definition (HD) map construction is essential for autonomous driving to accurately understand the surrounding environment. Most existing methods rely on single-frame inputs to predict local map, which often fail to effectively capture the temporal correlations between frames. This limitation results in discontinuities and instability in the generated map.To tackle this limitation, we propose a \u0000<italic>Ti</i>\u0000ghtly \u0000<italic>C</i>\u0000oupled temporal fusion \u0000<italic>Map</i>\u0000 \u0000<italic>Net</i>\u0000work (TICMapNet). TICMapNet breaks down the fusion process into three sub-problems: PV feature alignment, BEV feature adjustment, and Query feature fusion. By doing so, we effectively integrate temporal information at different stages through three plug-and-play modules, using the proposed tightly coupled strategy. Unlike traditional methods, our approach does not rely on camera extrinsic parameters, offering a new perspective for addressing the visual fusion challenge in the field of object detection. Experimental results show that TICMapNet significantly improves upon its single-frame baseline model, achieving at least a 7.0% increase in mAP using just two consecutive frames on the nuScenes dataset, while also showing generalizability across other tasks.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"9 12","pages":"11289-11296"},"PeriodicalIF":4.6,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142645394","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
S3E: A Multi-Robot Multimodal Dataset for Collaborative SLAM S3E:用于协作式 SLAM 的多机器人多模态数据集
IF 4.6 2区 计算机科学 Q2 ROBOTICS Pub Date : 2024-11-01 DOI: 10.1109/LRA.2024.3490402
Dapeng Feng;Yuhua Qi;Shipeng Zhong;Zhiqiang Chen;Qiming Chen;Hongbo Chen;Jin Wu;Jun Ma
The burgeoning demand for collaborative robotic systems to execute complex tasks collectively has intensified the research community's focus on advancing simultaneous localization and mapping (SLAM) in a cooperative context. Despite this interest, the scalability and diversity of existing datasets for collaborative trajectories remain limited, especially in scenarios with constrained perspectives where the generalization capabilities of Collaborative SLAM (C-SLAM) are critical for the feasibility of multi-agent missions. Addressing this gap, we introduce S3E, an expansive multimodal dataset. Captured by a fleet of unmanned ground vehicles traversing four distinct collaborative trajectory paradigms, S3E encompasses 13 outdoor and 5 indoor sequences. These sequences feature meticulously synchronized and spatially calibrated data streams, including 360-degree LiDAR point cloud, high-resolution stereo imagery, high-frequency inertial measurement units (IMU), and Ultra-wideband (UWB) relative observations. Our dataset not only surpasses previous efforts in scale, scene diversity, and data intricacy but also provides a thorough analysis and benchmarks for both collaborative and individual SLAM methodologies.
人们对协作机器人系统集体执行复杂任务的需求日益增长,因此研究界更加关注在协作环境中推进同步定位和绘图(SLAM)。尽管如此,现有协作轨迹数据集的可扩展性和多样性仍然有限,尤其是在视角受限的情况下,协作 SLAM(C-SLAM)的泛化能力对于多机器人任务的可行性至关重要。为了填补这一空白,我们引入了 S3E--一个广阔的多模态数据集。S3E 包含 13 个室外序列和 5 个室内序列,由穿越四种不同协作轨迹范例的无人地面飞行器编队拍摄。这些序列具有精心同步和空间校准的数据流,包括 360 度激光雷达点云、高分辨率立体图像、高频惯性测量单元 (IMU) 和超宽带 (UWB) 相对观测数据。我们的数据集不仅在规模、场景多样性和数据复杂性方面超越了以往的工作,而且还为协作式和单独的 SLAM 方法提供了全面的分析和基准。
{"title":"S3E: A Multi-Robot Multimodal Dataset for Collaborative SLAM","authors":"Dapeng Feng;Yuhua Qi;Shipeng Zhong;Zhiqiang Chen;Qiming Chen;Hongbo Chen;Jin Wu;Jun Ma","doi":"10.1109/LRA.2024.3490402","DOIUrl":"https://doi.org/10.1109/LRA.2024.3490402","url":null,"abstract":"The burgeoning demand for collaborative robotic systems to execute complex tasks collectively has intensified the research community's focus on advancing simultaneous localization and mapping (SLAM) in a cooperative context. Despite this interest, the scalability and diversity of existing datasets for collaborative trajectories remain limited, especially in scenarios with constrained perspectives where the generalization capabilities of Collaborative SLAM (C-SLAM) are critical for the feasibility of multi-agent missions. Addressing this gap, we introduce S3E, an expansive multimodal dataset. Captured by a fleet of unmanned ground vehicles traversing four distinct collaborative trajectory paradigms, S3E encompasses 13 outdoor and 5 indoor sequences. These sequences feature meticulously synchronized and spatially calibrated data streams, including 360-degree LiDAR point cloud, high-resolution stereo imagery, high-frequency inertial measurement units (IMU), and Ultra-wideband (UWB) relative observations. Our dataset not only surpasses previous efforts in scale, scene diversity, and data intricacy but also provides a thorough analysis and benchmarks for both collaborative and individual SLAM methodologies.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"9 12","pages":"11401-11408"},"PeriodicalIF":4.6,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142636540","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Constrained Dirichlet Distribution Policy: Guarantee Zero Constraint Violation Reinforcement Learning for Continuous Robotic Control 受约束的 Dirichlet 分布策略:保证零违反约束的连续机器人控制强化学习
IF 4.6 2区 计算机科学 Q2 ROBOTICS Pub Date : 2024-11-01 DOI: 10.1109/LRA.2024.3490392
Jianming Ma;Zhanxiang Cao;Yue Gao
Learning-based controllers show promising performances in robotic control tasks. However, they still present potential safety risks due to the difficulty in ensuring satisfaction of complex action constraints. We propose a novel action-constrained reinforcement learning method, which transforms the constrained action space into its dual space and uses Dirichlet distribution policy to guarantee strict constraint satisfaction as well as randomized exploration. We validate the proposed method in benchmark environments and in a real quadruped locomotion task. Our method outperforms other baselines with higher reward and faster inference speed. Results of the real robot experiments demonstrate the effectiveness and potential application of our method.
基于学习的控制器在机器人控制任务中表现出良好的性能。然而,由于难以确保满足复杂的动作约束,它们仍然存在潜在的安全风险。我们提出了一种新颖的行动约束强化学习方法,它将约束行动空间转化为其对偶空间,并使用 Dirichlet 分布策略来保证严格的约束满足以及随机探索。我们在基准环境和真实的四足运动任务中验证了所提出的方法。我们的方法以更高的回报和更快的推理速度超越了其他基线方法。真实机器人实验结果证明了我们方法的有效性和潜在应用价值。
{"title":"Constrained Dirichlet Distribution Policy: Guarantee Zero Constraint Violation Reinforcement Learning for Continuous Robotic Control","authors":"Jianming Ma;Zhanxiang Cao;Yue Gao","doi":"10.1109/LRA.2024.3490392","DOIUrl":"https://doi.org/10.1109/LRA.2024.3490392","url":null,"abstract":"Learning-based controllers show promising performances in robotic control tasks. However, they still present potential safety risks due to the difficulty in ensuring satisfaction of complex action constraints. We propose a novel action-constrained reinforcement learning method, which transforms the constrained action space into its dual space and uses Dirichlet distribution policy to guarantee strict constraint satisfaction as well as randomized exploration. We validate the proposed method in benchmark environments and in a real quadruped locomotion task. Our method outperforms other baselines with higher reward and faster inference speed. Results of the real robot experiments demonstrate the effectiveness and potential application of our method.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"9 12","pages":"11690-11697"},"PeriodicalIF":4.6,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142679287","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Physics-Guided Deep Learning Enabled Surrogate Modeling for Pneumatic Soft Robots 针对气动软机器人的物理引导式深度学习代理建模
IF 4.6 2区 计算机科学 Q2 ROBOTICS Pub Date : 2024-11-01 DOI: 10.1109/LRA.2024.3490258
Sameh I. Beaber;Zhen Liu;Ye Sun
Soft robots, formulated by soft and compliant materials, have grown significantly in recent years toward safe and adaptable operations and interactions with dynamic environments. Modeling the complex, nonlinear behaviors and controlling the deformable structures of soft robots present challenges. This study aims to establish a physics-guided deep learning (PGDL) computational framework that integrates physical models into deep learning framework as surrogate models for soft robots. Once trained, these models can replace computationally expensive numerical simulations to shorten the computation time and enable real-time control. This PGDL framework is among the first to integrate first principle physics of soft robots into deep learning toward highly accurate yet computationally affordable models for soft robot modeling and control. The proposed framework has been implemented and validated using three different pneumatic soft fingers with different behaviors and geometries, along with two training and testing approaches, to demonstrate its effectiveness and generalizability. The results showed that the mean square error (MSE) of predicted deformed curvature and the maximum and minimum deformation at various loading conditions were as low as $10^{-4}$ mm$^{2}$. The proposed PGDL framework is constructed from first principle physics and intrinsically can be applicable to various conditions by carefully considering the governing equations, auxiliary equations, and the corresponding boundary and initial conditions.
近年来,由柔软和顺应性材料构成的软体机器人在安全和适应性操作以及与动态环境的互动方面得到了长足发展。对软体机器人复杂的非线性行为建模和对其可变形结构的控制是一项挑战。本研究旨在建立一个物理引导的深度学习(PGDL)计算框架,将物理模型集成到深度学习框架中,作为软机器人的代理模型。一旦训练完成,这些模型就能取代计算成本高昂的数值模拟,从而缩短计算时间,实现实时控制。该 PGDL 框架是首个将软机器人的第一原理物理融入深度学习的框架之一,旨在为软机器人建模和控制建立高精度且计算成本低廉的模型。我们使用三种不同行为和几何形状的气动软手指以及两种训练和测试方法实施并验证了所提出的框架,以证明其有效性和可推广性。结果表明,在各种加载条件下,预测的变形曲率以及最大和最小变形的均方误差(MSE)低至 10^{-4}$ mm$^{2}$。所提出的 PGDL 框架是从第一原理物理学构建的,通过仔细考虑控制方程、辅助方程以及相应的边界和初始条件,本质上可适用于各种条件。
{"title":"Physics-Guided Deep Learning Enabled Surrogate Modeling for Pneumatic Soft Robots","authors":"Sameh I. Beaber;Zhen Liu;Ye Sun","doi":"10.1109/LRA.2024.3490258","DOIUrl":"https://doi.org/10.1109/LRA.2024.3490258","url":null,"abstract":"Soft robots, formulated by soft and compliant materials, have grown significantly in recent years toward safe and adaptable operations and interactions with dynamic environments. Modeling the complex, nonlinear behaviors and controlling the deformable structures of soft robots present challenges. This study aims to establish a physics-guided deep learning (PGDL) computational framework that integrates physical models into deep learning framework as surrogate models for soft robots. Once trained, these models can replace computationally expensive numerical simulations to shorten the computation time and enable real-time control. This PGDL framework is among the first to integrate first principle physics of soft robots into deep learning toward highly accurate yet computationally affordable models for soft robot modeling and control. The proposed framework has been implemented and validated using three different pneumatic soft fingers with different behaviors and geometries, along with two training and testing approaches, to demonstrate its effectiveness and generalizability. The results showed that the mean square error (MSE) of predicted deformed curvature and the maximum and minimum deformation at various loading conditions were as low as \u0000<inline-formula><tex-math>$10^{-4}$</tex-math></inline-formula>\u0000 mm\u0000<inline-formula><tex-math>$^{2}$</tex-math></inline-formula>\u0000. The proposed PGDL framework is constructed from first principle physics and intrinsically can be applicable to various conditions by carefully considering the governing equations, auxiliary equations, and the corresponding boundary and initial conditions.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"9 12","pages":"11441-11448"},"PeriodicalIF":4.6,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142636297","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Inferring Occluded Agent Behavior in Dynamic Games From Noise Corrupted Observations 从噪声破坏的观测结果推断动态游戏中的隐蔽代理行为
IF 4.6 2区 计算机科学 Q2 ROBOTICS Pub Date : 2024-11-01 DOI: 10.1109/LRA.2024.3490398
Tianyu Qiu;David Fridovich-Keil
In mobile robotics and autonomous driving, it is natural to model agent interactions as the Nash equilibrium of a noncooperative, dynamic game. These methods inherently rely on observations from sensors such as lidars and cameras to identify agents participating in the game and, therefore, have difficulty when some agents are occluded. To address this limitation, this paper presents an occlusion-aware game-theoretic inference method to estimate the locations of potentially occluded agents, and simultaneously infer the intentions of both visible and occluded agents, which best accounts for the observations of visible agents. Additionally, we propose a receding horizon planning strategy based on an occlusion-aware contingency game designed to navigate in scenarios with potentially occluded agents. Monte Carlo simulations validate our approach, demonstrating that it accurately estimates the game model and trajectories for both visible and occluded agents using noisy observations of visible agents. Our planning pipeline significantly enhances navigation safety when compared to occlusion-ignorant baseline as well.
在移动机器人和自动驾驶领域,将代理互动建模为非合作动态博弈的纳什均衡是很自然的。这些方法本质上依赖于激光雷达和摄像头等传感器的观测结果来识别参与博弈的代理,因此在某些代理被遮挡时会遇到困难。为了解决这一局限性,本文提出了一种具有遮挡感知能力的博弈论推理方法,用于估计可能被遮挡的代理的位置,并同时推断可见代理和被遮挡代理的意图,这种方法能最好地解释可见代理的观察结果。此外,我们还提出了一种基于隐蔽感知应急博弈的后退地平线规划策略,用于在可能存在隐蔽代理的场景中进行导航。蒙特卡罗模拟验证了我们的方法,证明它能利用对可见代理的噪声观测,准确估计可见代理和隐蔽代理的博弈模型和轨迹。与不考虑遮挡的基线相比,我们的规划管道也大大提高了导航安全性。
{"title":"Inferring Occluded Agent Behavior in Dynamic Games From Noise Corrupted Observations","authors":"Tianyu Qiu;David Fridovich-Keil","doi":"10.1109/LRA.2024.3490398","DOIUrl":"https://doi.org/10.1109/LRA.2024.3490398","url":null,"abstract":"In mobile robotics and autonomous driving, it is natural to model agent interactions as the Nash equilibrium of a noncooperative, dynamic game. These methods inherently rely on observations from sensors such as lidars and cameras to identify agents participating in the game and, therefore, have difficulty when some agents are occluded. To address this limitation, this paper presents an occlusion-aware game-theoretic inference method to estimate the locations of potentially occluded agents, and simultaneously infer the intentions of both visible and occluded agents, which best accounts for the observations of visible agents. Additionally, we propose a receding horizon planning strategy based on an occlusion-aware contingency game designed to navigate in scenarios with potentially occluded agents. Monte Carlo simulations validate our approach, demonstrating that it accurately estimates the game model and trajectories for both visible and occluded agents using noisy observations of visible agents. Our planning pipeline significantly enhances navigation safety when compared to occlusion-ignorant baseline as well.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"9 12","pages":"11489-11496"},"PeriodicalIF":4.6,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142636558","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Free-Init: Scan-Free, Motion-Free, and Correspondence-Free Initialization for Doppler LiDAR-Inertial Systems 自由初始化:多普勒激光雷达-惯性系统的无扫描、无运动和无对应初始化
IF 4.6 2区 计算机科学 Q2 ROBOTICS Pub Date : 2024-11-01 DOI: 10.1109/LRA.2024.3490395
Mingle Zhao;Jiahao Wang;Tianxiao Gao;Chengzhong Xu;Hui Kong
Robust initialization is crucial for online systems. In the letter, a high-frequency and resilient initialization framework is designed for LiDAR-inertial systems, leveraging both inertial sensors and Doppler LiDAR. The innovative FMCW Doppler LiDAR opens up a novel avenue for robotic sensing by capturing not only point range but also Doppler velocity via the intrinsic Doppler effect. By fusing point-wise Doppler velocity with inertial measurements under non-inertial kinematics, the proposed framework, Free-Init, eliminates reliance on motion undistortion of LiDAR scans, excitation motions, and map correspondences during the initialization phase. Free-Init is also plug-and-play compatible with typical LiDAR-inertial systems and is versatile to handle a wide range of initial motions when the system starts, including stationary, dynamic, and even violent motions. The embedded Doppler-inertial velocimeter ensures fast convergence and high-frequency performance, delivering outputs exceeding 10 kHz. Comprehensive experiments on diverse platforms and across myriad motion scenes validate the framework's effectiveness. The results demonstrate the superior performance of Free-Init, highlighting the necessity of fast, resilient, and dynamic initialization for online systems.
稳健的初始化对在线系统至关重要。信中利用惯性传感器和多普勒激光雷达,为激光雷达-惯性系统设计了一个高频和弹性初始化框架。创新的 FMCW 多普勒激光雷达不仅能捕捉点测距,还能通过固有的多普勒效应捕捉多普勒速度,为机器人传感开辟了一条新途径。通过在非惯性运动学条件下将点式多普勒速度与惯性测量相结合,所提出的 Free-Init 框架在初始化阶段消除了对激光雷达扫描运动失真、激励运动和地图对应关系的依赖。Free-Init 还可与典型的激光雷达-惯性系统即插即用,并具有多功能性,可在系统启动时处理各种初始运动,包括静态、动态甚至剧烈运动。嵌入式多普勒惯性测速仪可确保快速收敛和高频性能,输出频率超过 10 kHz。在不同平台和无数运动场景中进行的综合实验验证了该框架的有效性。实验结果证明了 Free-Init 的卓越性能,突出了在线系统进行快速、弹性和动态初始化的必要性。
{"title":"Free-Init: Scan-Free, Motion-Free, and Correspondence-Free Initialization for Doppler LiDAR-Inertial Systems","authors":"Mingle Zhao;Jiahao Wang;Tianxiao Gao;Chengzhong Xu;Hui Kong","doi":"10.1109/LRA.2024.3490395","DOIUrl":"https://doi.org/10.1109/LRA.2024.3490395","url":null,"abstract":"Robust initialization is crucial for online systems. In the letter, a high-frequency and resilient initialization framework is designed for LiDAR-inertial systems, leveraging both inertial sensors and Doppler LiDAR. The innovative FMCW Doppler LiDAR opens up a novel avenue for robotic sensing by capturing not only point range but also Doppler velocity via the intrinsic Doppler effect. By fusing point-wise Doppler velocity with inertial measurements under non-inertial kinematics, the proposed framework, Free-Init, eliminates reliance on motion undistortion of LiDAR scans, excitation motions, and map correspondences during the initialization phase. Free-Init is also plug-and-play compatible with typical LiDAR-inertial systems and is versatile to handle a wide range of initial motions when the system starts, including stationary, dynamic, and even violent motions. The embedded Doppler-inertial velocimeter ensures fast convergence and high-frequency performance, delivering outputs exceeding 10 kHz. Comprehensive experiments on diverse platforms and across myriad motion scenes validate the framework's effectiveness. The results demonstrate the superior performance of Free-Init, highlighting the necessity of fast, resilient, and dynamic initialization for online systems.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"9 12","pages":"11329-11336"},"PeriodicalIF":4.6,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142600423","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Robotics and Automation Letters
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1