首页 > 最新文献

Autonomous Robots最新文献

英文 中文
Wild visual navigation: fast traversability learning via pre-trained models and online self-supervision 野生视觉导航:通过预训练模型和在线自我监督快速遍历学习
IF 4.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-07-18 DOI: 10.1007/s10514-025-10202-x
Matias Mattamala, Jonas Frey, Piotr Libera, Nived Chebrolu, Georg Martius, Cesar Cadena, Marco Hutter, Maurice Fallon

Natural environments such as forests and grasslands are challenging for robotic navigation because of the false perception of rigid obstacles from high grass, twigs, or bushes. In this work, we present Wild Visual Navigation (WVN), an online self-supervised learning system for visual traversability estimation. The system is able to continuously adapt from a short human demonstration in the field, only using onboard sensing and computing. One of the key ideas to achieve this is the use of high-dimensional features from pre-trained self-supervised models, which implicitly encode semantic information that massively simplifies the learning task. Further, the development of an online scheme for supervision generator enables concurrent training and inference of the learned model in the wild. We demonstrate our approach through diverse real-world deployments in forests, parks, and grasslands. Our system is able to bootstrap the traversable terrain segmentation in less than 5 min of in-field training time, enabling the robot to navigate in complex, previously unseen outdoor terrains.

森林和草原等自然环境对机器人导航来说是一个挑战,因为它们会错误地感知来自高草、树枝或灌木丛的坚硬障碍物。在这项工作中,我们提出了野生视觉导航(WVN),一个用于视觉遍历估计的在线自监督学习系统。该系统仅使用机载传感和计算,就能从现场短暂的人类演示中持续适应。实现这一目标的关键思想之一是使用来自预训练的自监督模型的高维特征,它隐含地编码语义信息,从而大大简化了学习任务。此外,开发了一种在线监督生成器方案,使学习模型能够在野外进行并发训练和推理。我们通过在森林、公园和草原上的各种实际部署来展示我们的方法。我们的系统能够在不到5分钟的现场训练时间内引导可穿越的地形分割,使机器人能够在复杂的、以前看不见的室外地形中导航。
{"title":"Wild visual navigation: fast traversability learning via pre-trained models and online self-supervision","authors":"Matias Mattamala,&nbsp;Jonas Frey,&nbsp;Piotr Libera,&nbsp;Nived Chebrolu,&nbsp;Georg Martius,&nbsp;Cesar Cadena,&nbsp;Marco Hutter,&nbsp;Maurice Fallon","doi":"10.1007/s10514-025-10202-x","DOIUrl":"10.1007/s10514-025-10202-x","url":null,"abstract":"<div><p>Natural environments such as forests and grasslands are challenging for robotic navigation because of the false perception of rigid obstacles from high grass, twigs, or bushes. In this work, we present Wild Visual Navigation (WVN), an online self-supervised learning system for visual traversability estimation. The system is able to continuously adapt from a short human demonstration in the field, only using onboard sensing and computing. One of the key ideas to achieve this is the use of high-dimensional features from pre-trained self-supervised models, which implicitly encode semantic information that massively simplifies the learning task. Further, the development of an online scheme for supervision generator enables concurrent training and inference of the learned model in the wild. We demonstrate our approach through diverse real-world deployments in forests, parks, and grasslands. Our system is able to bootstrap the traversable terrain segmentation in less than 5 min of in-field training time, enabling the robot to navigate in complex, previously unseen outdoor terrains.</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"49 3","pages":""},"PeriodicalIF":4.3,"publicationDate":"2025-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10514-025-10202-x.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145166899","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Autonomous learning-free grasping and robot-to-robot handover of unknown objects 自主学习抓取和机器人对未知物体的切换
IF 4.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-06-28 DOI: 10.1007/s10514-025-10201-y
Yuwei Wu, Wanze Li, Zhiyang Liu, Weixiao Liu, Gregory S. Chirikjian

In this paper, we propose a learning-free approach for an autonomous robotic system to grasp, hand over, and regrasp previously unseen objects. The proposed framework includes two main components: a novel grasping detector to predict grasping poses directly from the point cloud and a reachability-aware handover planner to select the exchange pose and grasping poses for two robots. In the grasping detection stage, multiple superquadrics are first recovered at different positions within the object, representing the local geometric feature of the object. Our algorithm then exploits the tri-symmetry feature of superquadrics and synthesizes a list of antipodal grasps from each recovered superquadric. An evaluation model is designed to assess and quantify the quality of each grasp candidate. In the handover planning stage, the planner first selects grasping candidates that have high scores and a larger number of collision-free partners. Then the exchange location is computed by utilizing two signed distance fields (SDF) which model the reachability space for the pair of two robots. To evaluate the performance of the proposed method, we first run experiments on isolated and packed scenes to corroborate the effectiveness of our grasping detection method. Then the handover experiments are conducted on a dual-arm system with two 7 degrees of freedom (DoF) manipulators. The results indicate that our method shows better performance compared with the state-of-the-art, without the need for large amounts of training.

在本文中,我们提出了一种无需学习的方法,用于自主机器人系统抓取、移交和重新抓取以前看不见的物体。该框架包括两个主要组成部分:一种新型抓取检测器,用于直接从点云预测抓取姿态;另一种可达性感知切换规划器用于选择两个机器人的交换姿态和抓取姿态。在抓取检测阶段,首先在物体内部的不同位置恢复多个超二次曲面,代表物体的局部几何特征。然后,我们的算法利用超二次曲面的三对称特征,并从每个恢复的超二次曲面合成对映抓取列表。设计了一个评估模型来评估和量化每个硕士候选人的质量。在交接规划阶段,规划者首先选择得分高、无碰撞伙伴数量多的抓取对象。然后利用两个有符号距离域(SDF)计算交换位置,SDF对两个机器人的可达空间进行建模。为了评估所提出方法的性能,我们首先在孤立和拥挤的场景上进行实验,以证实我们的抓取检测方法的有效性。然后,在具有两个7自由度机械臂的双臂系统上进行了切换实验。结果表明,我们的方法在不需要大量训练的情况下,表现出比最先进的方法更好的性能。
{"title":"Autonomous learning-free grasping and robot-to-robot handover of unknown objects","authors":"Yuwei Wu,&nbsp;Wanze Li,&nbsp;Zhiyang Liu,&nbsp;Weixiao Liu,&nbsp;Gregory S. Chirikjian","doi":"10.1007/s10514-025-10201-y","DOIUrl":"10.1007/s10514-025-10201-y","url":null,"abstract":"<div><p>In this paper, we propose a learning-free approach for an autonomous robotic system to grasp, hand over, and regrasp previously unseen objects. The proposed framework includes two main components: a novel grasping detector to predict grasping poses directly from the point cloud and a reachability-aware handover planner to select the exchange pose and grasping poses for two robots. In the grasping detection stage, multiple superquadrics are first recovered at different positions within the object, representing the local geometric feature of the object. Our algorithm then exploits the tri-symmetry feature of superquadrics and synthesizes a list of antipodal grasps from each recovered superquadric. An evaluation model is designed to assess and quantify the quality of each grasp candidate. In the handover planning stage, the planner first selects grasping candidates that have high scores and a larger number of collision-free partners. Then the exchange location is computed by utilizing two signed distance fields (SDF) which model the reachability space for the pair of two robots. To evaluate the performance of the proposed method, we first run experiments on isolated and packed scenes to corroborate the effectiveness of our grasping detection method. Then the handover experiments are conducted on a dual-arm system with two 7 degrees of freedom (DoF) manipulators. The results indicate that our method shows better performance compared with the state-of-the-art, without the need for large amounts of training.</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"49 3","pages":""},"PeriodicalIF":4.3,"publicationDate":"2025-06-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10514-025-10201-y.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145170980","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-robot exploration for the CADRE mission CADRE任务的多机器人探索
IF 4.3 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-06-12 DOI: 10.1007/s10514-025-10199-3
Sharan Nayak, Grace Lim, Federico Rossi, Michael Otte, Jean-Pierre de la Croix

We present the design, implementation and testing of a multi-robot exploration algorithm for NASA’s upcoming Cooperative Autonomous Distributed Robotic Exploration (CADRE) lunar technology demonstration mission. The CADRE mission, among its various objectives, entails utilizing a trio of autonomous mobile robots to collaboratively explore and construct a map of a designated area of the lunar surface. Given the mission’s inherent constraints, including limited mission duration, constrained power resources, and restricted communication capabilities, we formulate an exploration algorithm to improve exploration efficiency, facilitate equitable workload distribution among individual agents, and minimize inter-robot communication. To achieve these requirements, we employ a semi-centralized exploration algorithm that partitions the unexplored area, regardless of its shape and size, into a series of non-overlapping partitions, assigning each partition to a specific robot for exploration. Each robot autonomously explores its designated region without intervention from other robots. We explore the design space of the proposed algorithm and evaluate its performance under diverse conditions in simulations. Finally, we validate the algorithm’s functionality through two sets of hardware experiments: the first utilizes prototype rovers using a ROS-based navigation software stack for feasibility testing, while the second employs high-fidelity development model rovers running CADRE’s custom flight-software stack for flight-like performance validation. Both sets of experiments are conducted in the Jet Propulsion Laboratory’s lunar-simulated rover testing facilities, demonstrating the algorithm’s robustness and readiness for lunar deployment.

我们提出了一种多机器人探索算法的设计、实现和测试,用于NASA即将进行的合作自主分布式机器人探索(CADRE)月球技术演示任务。在其众多目标中,CADRE任务需要利用三个自主移动机器人协同探索和构建月球表面指定区域的地图。考虑到任务持续时间有限、电力资源受限、通信能力受限等固有约束,本文提出了一种探索算法,以提高探索效率,促进个体智能体之间的公平工作量分配,并最大限度地减少机器人间的通信。为了实现这些要求,我们采用了一种半集中式的探索算法,将未探索的区域划分为一系列不重叠的分区,而不考虑其形状和大小,并将每个分区分配给特定的机器人进行探索。每个机器人在没有其他机器人干预的情况下自主探索其指定的区域。我们探索了该算法的设计空间,并在仿真中评估了其在不同条件下的性能。最后,我们通过两组硬件实验验证了算法的功能:第一组使用基于ros的导航软件堆栈利用原型漫游者进行可行性测试,而第二组使用运行CADRE自定义飞行软件堆栈的高保真开发模型漫游者进行飞行性能验证。两组实验都在喷气推进实验室的月球模拟月球车测试设施中进行,证明了该算法的鲁棒性和月球部署的就绪性。
{"title":"Multi-robot exploration for the CADRE mission","authors":"Sharan Nayak,&nbsp;Grace Lim,&nbsp;Federico Rossi,&nbsp;Michael Otte,&nbsp;Jean-Pierre de la Croix","doi":"10.1007/s10514-025-10199-3","DOIUrl":"10.1007/s10514-025-10199-3","url":null,"abstract":"<div><p>We present the design, implementation and testing of a multi-robot exploration algorithm for NASA’s upcoming Cooperative Autonomous Distributed Robotic Exploration (CADRE) lunar technology demonstration mission. The CADRE mission, among its various objectives, entails utilizing a trio of autonomous mobile robots to collaboratively explore and construct a map of a designated area of the lunar surface. Given the mission’s inherent constraints, including limited mission duration, constrained power resources, and restricted communication capabilities, we formulate an exploration algorithm to improve exploration efficiency, facilitate equitable workload distribution among individual agents, and minimize inter-robot communication. To achieve these requirements, we employ a semi-centralized exploration algorithm that partitions the unexplored area, regardless of its shape and size, into a series of non-overlapping partitions, assigning each partition to a specific robot for exploration. Each robot autonomously explores its designated region without intervention from other robots. We explore the design space of the proposed algorithm and evaluate its performance under diverse conditions in simulations. Finally, we validate the algorithm’s functionality through two sets of hardware experiments: the first utilizes prototype rovers using a ROS-based navigation software stack for feasibility testing, while the second employs high-fidelity development model rovers running CADRE’s custom flight-software stack for flight-like performance validation. Both sets of experiments are conducted in the Jet Propulsion Laboratory’s lunar-simulated rover testing facilities, demonstrating the algorithm’s robustness and readiness for lunar deployment.</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"49 2","pages":""},"PeriodicalIF":4.3,"publicationDate":"2025-06-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145164979","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Effective tracking of unknown clustered targets using a distributed team of mobile robots 使用分布式移动机器人团队有效跟踪未知集群目标
IF 3.7 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-24 DOI: 10.1007/s10514-025-10200-z
Jun Chen, Philip Dames, Shinkyu Park

Distributed multi-target tracking is a canonical task for multi-robot systems, encompassing applications from environmental monitoring to disaster response to surveillance. In many situations the unknown distribution of the targets in a search area is non-uniform, e.g., herds of animals moving together. This paper develops a novel distributed multi-robot multi-target tracking algorithm to effectively search for and track clustered targets. There are two key features. First, there are two parallel estimators, one to provide the best guess of the current states of targets and a second to provide a coarse, long-term distribution of clusters. Second, robots use the power diagram to divide the search space between agents in a way that effectively trades off between tracking detected targets within high density areas and searching for other potential targets. Extensive simulation experiments demonstrate the efficacy of the proposed method and show that it outperforms other approaches in tracking accuracy of clustered targets while maintain good performance for uniformly distributed targets.

分布式多目标跟踪是多机器人系统的典型任务,涵盖了从环境监测到灾难响应再到监视的应用。在许多情况下,未知目标在搜索区域的分布是不均匀的,例如,一群动物一起移动。为了有效地搜索和跟踪聚类目标,提出了一种新的分布式多机器人多目标跟踪算法。有两个关键特性。首先,有两个并行估计器,一个用于提供目标当前状态的最佳猜测,另一个用于提供聚类的粗略长期分布。其次,机器人使用功率图在代理之间划分搜索空间,从而有效地在高密度区域内跟踪检测到的目标和搜索其他潜在目标之间进行权衡。大量的仿真实验证明了该方法的有效性,并表明该方法在对均匀分布目标保持良好跟踪性能的同时,在对聚类目标的跟踪精度方面优于其他方法。
{"title":"Effective tracking of unknown clustered targets using a distributed team of mobile robots","authors":"Jun Chen,&nbsp;Philip Dames,&nbsp;Shinkyu Park","doi":"10.1007/s10514-025-10200-z","DOIUrl":"10.1007/s10514-025-10200-z","url":null,"abstract":"<div><p>Distributed multi-target tracking is a canonical task for multi-robot systems, encompassing applications from environmental monitoring to disaster response to surveillance. In many situations the unknown distribution of the targets in a search area is non-uniform, e.g., herds of animals moving together. This paper develops a novel distributed multi-robot multi-target tracking algorithm to effectively search for and track clustered targets. There are two key features. First, there are two parallel estimators, one to provide the best guess of the current states of targets and a second to provide a coarse, long-term distribution of clusters. Second, robots use the power diagram to divide the search space between agents in a way that effectively trades off between tracking detected targets within high density areas and searching for other potential targets. Extensive simulation experiments demonstrate the efficacy of the proposed method and show that it outperforms other approaches in tracking accuracy of clustered targets while maintain good performance for uniformly distributed targets.</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"49 2","pages":""},"PeriodicalIF":3.7,"publicationDate":"2025-05-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10514-025-10200-z.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144131723","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LSF-planner: a visual local planner for legged robots based on ground structure and feature information LSF-planner:基于地面结构和特征信息的有腿机器人视觉局部规划器
IF 3.7 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-12 DOI: 10.1007/s10514-025-10195-7
Teng Zhang, Xiangji Wang, Fusheng Zha, Fucheng Liu

Three-dimensional navigation of legged robots is crucial for field exploration and post-disaster rescue. Existing optimization-based local trajectory planners predominantly focus on obstacle avoidance, neglecting negative obstacles (e.g., pits) and varying ground features (e.g., different terrain types). Additionally, non-overlapping areas between the planned space in three-dimensional trajectory planning and the robot’s actual reachable space lead to decision-making issues between crossing and obstacle avoidance, making it challenging to differentiate between passable and hazardous areas, thus impacting navigation safety and stability. To address these limitations, we propose a novel visual local planner, LSF-Planner (Visual Local Planner for Legged Robots Based on Ground Structure and Feature Information). The LSF-Planner employs a multi-layer local perception map that integrates ground feature semantics, sensor range, and negative obstacles (e.g., voids detected by depth sensors) to construct a ground reliability representation. The Label2Grad method is introduced to convert this representation into gradient layers, incorporating a ground reliability penalty function into trajectory optimization. By incorporating constraints on the center of mass height and crossing angles, LSF-Planner effectively differentiates between traversable and hazardous areas. Experimental results show that LSF-Planner significantly outperforms existing methods in 3D trajectory planning, enhancing the navigation performance of legged robots in unstructured environments.

有腿机器人的三维导航对于野外勘探和灾后救援至关重要。现有的基于优化的局部轨迹规划主要关注避障,忽略了负面障碍物(如坑)和不同的地面特征(如不同的地形类型)。此外,三维轨迹规划中的规划空间与机器人实际可达空间之间的非重叠区域导致了穿越和避障的决策问题,使得区分可通过区域和危险区域变得困难,从而影响了导航的安全性和稳定性。为了解决这些限制,我们提出了一种新的视觉局部规划器,LSF-Planner(基于地面结构和特征信息的有腿机器人视觉局部规划器)。LSF-Planner采用多层局部感知图,该图集成了地面特征语义、传感器范围和负面障碍物(例如,深度传感器检测到的空洞),以构建地面可靠性表示。引入Label2Grad方法将此表示转换为梯度层,并将地面可靠性惩罚函数纳入轨迹优化。通过结合质心高度和交叉角度的约束,LSF-Planner可以有效区分可穿越区域和危险区域。实验结果表明,LSF-Planner在三维轨迹规划方面明显优于现有方法,提高了有腿机器人在非结构化环境中的导航性能。
{"title":"LSF-planner: a visual local planner for legged robots based on ground structure and feature information","authors":"Teng Zhang,&nbsp;Xiangji Wang,&nbsp;Fusheng Zha,&nbsp;Fucheng Liu","doi":"10.1007/s10514-025-10195-7","DOIUrl":"10.1007/s10514-025-10195-7","url":null,"abstract":"<div><p>Three-dimensional navigation of legged robots is crucial for field exploration and post-disaster rescue. Existing optimization-based local trajectory planners predominantly focus on obstacle avoidance, neglecting negative obstacles (e.g., pits) and varying ground features (e.g., different terrain types). Additionally, non-overlapping areas between the planned space in three-dimensional trajectory planning and the robot’s actual reachable space lead to decision-making issues between crossing and obstacle avoidance, making it challenging to differentiate between passable and hazardous areas, thus impacting navigation safety and stability. To address these limitations, we propose a novel visual local planner, LSF-Planner (Visual Local Planner for Legged Robots Based on Ground Structure and Feature Information). The LSF-Planner employs a multi-layer local perception map that integrates ground feature semantics, sensor range, and negative obstacles (e.g., voids detected by depth sensors) to construct a ground reliability representation. The Label2Grad method is introduced to convert this representation into gradient layers, incorporating a ground reliability penalty function into trajectory optimization. By incorporating constraints on the center of mass height and crossing angles, LSF-Planner effectively differentiates between traversable and hazardous areas. Experimental results show that LSF-Planner significantly outperforms existing methods in 3D trajectory planning, enhancing the navigation performance of legged robots in unstructured environments.</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"49 2","pages":""},"PeriodicalIF":3.7,"publicationDate":"2025-05-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143938522","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Shortest coordinated motions for square robots 方形机器人的最短协调运动
IF 3.7 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-08 DOI: 10.1007/s10514-025-10198-4
Guillermo Esteban, Dan Halperin, Rodrigo I. Silveira

We study the problem of determining minimum-length coordinated motions for two axis-aligned square robots translating in an obstacle-free plane: Given feasible start and goal configurations (feasible in the sense that the two squares are interior disjoint), find a continuous motion for the two squares from start to goal, comprising only robot-robot collision-free configurations, such that the total Euclidean distance traveled by the two squares is minimal among all possible such motions. In this paper we present an adaptation of the tools developed for the case of disks to the case of squares. We show that in certain aspects the case of squares is more complicated, requiring additional and more involved arguments over the case of disks.

我们研究了确定两个轴向正方形机器人在无障碍平面上平移的最小长度协调运动的问题:给定可行的起始和目标构型(在这两个正方形内部不相交的意义上是可行的),找到两个正方形从起始到目标的连续运动,只包含机器人-机器人无碰撞构型,使得两个正方形走过的总欧氏距离在所有可能的运动中最小。在本文中,我们提出了一个适应的工具开发的情况下,磁盘的情况下,以正方形的情况。我们表明,在某些方面,正方形的情况是更复杂的,需要额外的和更复杂的论证比磁盘的情况。
{"title":"Shortest coordinated motions for square robots","authors":"Guillermo Esteban,&nbsp;Dan Halperin,&nbsp;Rodrigo I. Silveira","doi":"10.1007/s10514-025-10198-4","DOIUrl":"10.1007/s10514-025-10198-4","url":null,"abstract":"<div><p>We study the problem of determining minimum-length coordinated motions for two axis-aligned square robots translating in an obstacle-free plane: Given feasible start and goal configurations (feasible in the sense that the two squares are interior disjoint), find a continuous motion for the two squares from start to goal, comprising only robot-robot collision-free configurations, such that the total Euclidean distance traveled by the two squares is minimal among all possible such motions. In this paper we present an adaptation of the tools developed for the case of disks to the case of squares. We show that in certain aspects the case of squares is more complicated, requiring additional and more involved arguments over the case of disks.\u0000</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"49 2","pages":""},"PeriodicalIF":3.7,"publicationDate":"2025-05-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10514-025-10198-4.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143925693","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FIMD: fast isolated marker detection for UV-based visual relative localisation in agile UAV swarms FIMD:敏捷无人机群中基于uv的视觉相对定位快速隔离标记检测
IF 3.7 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-06 DOI: 10.1007/s10514-025-10197-5
Vojtěch Vrba, Viktor Walter, Petr Štěpán, Martin Saska

A novel approach for the fast onboard detection of isolated markers for visual relative localisation of multiple teammates in agile UAV swarms is introduced in this paper. As the detection forms a key component of real-time localisation systems, a three-fold innovation is presented, consisting of an optimised procedure for CPUs, a GPU shader program, and a functionally equivalent FPGA streaming architecture. For the proposed CPU and GPU solutions, the mean processing time per pixel of input camera frames was accelerated by two to three orders of magnitude compared to the unoptimised state-of-the-art approach. For the localisation task, the proposed FPGA architecture offered the most significant overall acceleration by minimising the total delay from camera exposure to detection results. Additionally, the proposed solutions were evaluated on various 32-bit and 64-bit embedded platforms to demonstrate their efficiency, as well as their feasibility for applications using low-end UAVs and MAVs. Thus, it has become a crucial enabling technology for agile UAV swarming.

提出了一种针对敏捷无人机群中多成员视觉相对定位的孤立标记快速机载检测方法。由于检测是实时定位系统的关键组成部分,因此提出了三方面的创新,包括针对cpu的优化程序,GPU着色器程序和功能等效的FPGA流架构。对于提议的CPU和GPU解决方案,与未优化的最先进方法相比,输入相机帧的每像素平均处理时间加快了两到三个数量级。对于定位任务,提出的FPGA架构通过最大限度地减少从相机曝光到检测结果的总延迟,提供了最显著的整体加速。此外,所提出的解决方案在各种32位和64位嵌入式平台上进行了评估,以证明它们的效率,以及它们在低端无人机和MAVs应用中的可行性。因此,它已成为无人机敏捷蜂群的关键使能技术。
{"title":"FIMD: fast isolated marker detection for UV-based visual relative localisation in agile UAV swarms","authors":"Vojtěch Vrba,&nbsp;Viktor Walter,&nbsp;Petr Štěpán,&nbsp;Martin Saska","doi":"10.1007/s10514-025-10197-5","DOIUrl":"10.1007/s10514-025-10197-5","url":null,"abstract":"<div><p>A novel approach for the fast onboard detection of isolated markers for visual relative localisation of multiple teammates in agile UAV swarms is introduced in this paper. As the detection forms a key component of real-time localisation systems, a three-fold innovation is presented, consisting of an optimised procedure for CPUs, a GPU shader program, and a functionally equivalent FPGA streaming architecture. For the proposed CPU and GPU solutions, the mean processing time per pixel of input camera frames was accelerated by two to three orders of magnitude compared to the unoptimised state-of-the-art approach. For the localisation task, the proposed FPGA architecture offered the most significant overall acceleration by minimising the total delay from camera exposure to detection results. Additionally, the proposed solutions were evaluated on various 32-bit and 64-bit embedded platforms to demonstrate their efficiency, as well as their feasibility for applications using low-end UAVs and MAVs. Thus, it has become a crucial enabling technology for agile UAV swarming.\u0000</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"49 2","pages":""},"PeriodicalIF":3.7,"publicationDate":"2025-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10514-025-10197-5.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143913840","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deadlock-free, safe, and decentralized multi-robot navigation in social mini-games via discrete-time control barrier functions 无死锁,安全,分散的多机器人导航在社交小游戏通过离散时间控制障碍功能
IF 3.7 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-20 DOI: 10.1007/s10514-025-10194-8
Rohan Chandra, Vrushabh Zinage, Efstathios Bakolas, Peter Stone, Joydeep Biswas

We present an approach to ensure safe and deadlock-free navigation for decentralized multi-robot systems operating in constrained environments, including doorways and intersections. Although many solutions have been proposed that ensure safety and resolve deadlocks, optimally preventing deadlocks in a minimally invasive and decentralized fashion remains an open problem. We first formalize the objective as a non-cooperative, non-communicative, partially observable multi-robot navigation problem in constrained spaces with multiple conflicting agents, which we term as social mini-games. Formally, we solve a discrete-time optimal receding horizon control problem leveraging control barrier functions for safe long-horizon planning. Our approach to ensuring liveness rests on the insight that there exists barrier certificates that allow each robot to preemptively perturb their state in a minimally-invasive fashion onto liveness sets i.e. states where robots are deadlock-free. We evaluate our approach in simulation as well on physical robots using F1/10 robots, a Clearpath Jackal, as well as a Boston Dynamics Spot in a doorway, hallway, and corridor intersection scenario. Compared to both fully decentralized and centralized approaches with and without deadlock resolution capabilities, we demonstrate that our approach results in safer, more efficient, and smoother navigation, based on a comprehensive set of metrics including success rate, collision rate, stop time, change in velocity, path deviation, time-to-goal, and flow rate.

我们提出了一种方法来确保在受限环境下(包括门口和十字路口)运行的分散多机器人系统的安全和无死锁导航。尽管已经提出了许多确保安全和解决死锁的解决方案,但以微创和分散的方式最佳地防止死锁仍然是一个悬而未决的问题。我们首先将目标形式化为具有多个冲突代理的受限空间中的非合作、非交流、部分可观察的多机器人导航问题,我们将其称为社交迷你游戏。在形式上,我们利用控制障碍函数求解了一个安全的长期规划的离散时间最优后退水平控制问题。我们确保活动性的方法基于这样一种见解,即存在屏障证书,允许每个机器人以最小侵入的方式先发制人地干扰它们的状态到活动性集,即机器人无死锁的状态。我们在模拟中评估了我们的方法,以及使用F1/10机器人,Clearpath Jackal的物理机器人,以及门口,走廊和走廊交叉场景中的波士顿动力点。与具有或不具有死锁解决能力的完全分散式和集中式方法相比,我们证明,基于一系列综合指标,包括成功率、碰撞率、停止时间、速度变化、路径偏差、到达目标时间和流量,我们的方法可以实现更安全、更高效、更平稳的导航。
{"title":"Deadlock-free, safe, and decentralized multi-robot navigation in social mini-games via discrete-time control barrier functions","authors":"Rohan Chandra,&nbsp;Vrushabh Zinage,&nbsp;Efstathios Bakolas,&nbsp;Peter Stone,&nbsp;Joydeep Biswas","doi":"10.1007/s10514-025-10194-8","DOIUrl":"10.1007/s10514-025-10194-8","url":null,"abstract":"<div><p>We present an approach to ensure safe and deadlock-free navigation for decentralized multi-robot systems operating in constrained environments, including doorways and intersections. Although many solutions have been proposed that ensure safety and resolve deadlocks, optimally preventing deadlocks in a minimally invasive and decentralized fashion remains an open problem. We first formalize the objective as a non-cooperative, non-communicative, partially observable multi-robot navigation problem in constrained spaces with multiple conflicting agents, which we term as social mini-games. Formally, we solve a discrete-time optimal receding horizon control problem leveraging control barrier functions for safe long-horizon planning. Our approach to ensuring liveness rests on the insight that <i>there exists barrier certificates that allow each robot to preemptively perturb their state in a minimally-invasive fashion onto liveness sets i.e. states where robots are deadlock-free</i>. We evaluate our approach in simulation as well on physical robots using F1/10 robots, a Clearpath Jackal, as well as a Boston Dynamics Spot in a doorway, hallway, and corridor intersection scenario. Compared to both fully decentralized and centralized approaches with and without deadlock resolution capabilities, we demonstrate that our approach results in safer, more efficient, and smoother navigation, based on a comprehensive set of metrics including success rate, collision rate, stop time, change in velocity, path deviation, time-to-goal, and flow rate.</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"49 2","pages":""},"PeriodicalIF":3.7,"publicationDate":"2025-04-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10514-025-10194-8.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143850939","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fault-tolerant multi-robot localization: diagnostic decision-making with information theory and learning models 多机器人容错定位:利用信息论和学习模型进行诊断决策
IF 3.7 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-17 DOI: 10.1007/s10514-025-10196-6
Zaynab El Mawas, Cindy Cappelle, Maan El Badaoui El Najjar

In the domain of multi-robot systems, cooperative systems that are highly attuned and connected to their surroundings are becoming increasingly significant. This surge in interest highlights various challenges, especially regarding system integration and safety constraints. Our research contributes to the assurance of fault tolerance to avert abnormal behaviors and sustain reliable robot localization. In this paper, a mixed approach between data-driven and model-based for fault detection is introduced, within a decentralized architecture, thereby strengthening the system’s capacity to handle simultaneous sensor faults. Information theory-based fault indicators are developed by computing the Jensen-Shannon divergence ((D_{JS})) between state predictions and sensor-obtained corrections. This initiates a two-tiered data-driven mechanism: one layer employing Machine Learning for fault detection, and another distinct layer for fault isolation. The methodology’s efficacy is assessed using real data from the Turtlebot3 platform.

在多机器人系统领域,与周围环境高度协调和连接的协作系统变得越来越重要。这种兴趣的激增突出了各种挑战,特别是关于系统集成和安全约束。我们的研究有助于保证容错,避免异常行为和维持可靠的机器人定位。本文提出了一种基于数据驱动和基于模型的混合故障检测方法,该方法采用分散式结构,从而增强了系统处理传感器同步故障的能力。基于信息理论的故障指示器是通过计算状态预测和传感器获得的修正之间的Jensen-Shannon散度((D_{JS}))来开发的。这启动了一个两层数据驱动机制:一层使用机器学习进行故障检测,另一层用于故障隔离。使用来自Turtlebot3平台的真实数据评估该方法的有效性。
{"title":"Fault-tolerant multi-robot localization: diagnostic decision-making with information theory and learning models","authors":"Zaynab El Mawas,&nbsp;Cindy Cappelle,&nbsp;Maan El Badaoui El Najjar","doi":"10.1007/s10514-025-10196-6","DOIUrl":"10.1007/s10514-025-10196-6","url":null,"abstract":"<div><p>In the domain of multi-robot systems, cooperative systems that are highly attuned and connected to their surroundings are becoming increasingly significant. This surge in interest highlights various challenges, especially regarding system integration and safety constraints. Our research contributes to the assurance of fault tolerance to avert abnormal behaviors and sustain reliable robot localization. In this paper, a mixed approach between data-driven and model-based for fault detection is introduced, within a decentralized architecture, thereby strengthening the system’s capacity to handle simultaneous sensor faults. Information theory-based fault indicators are developed by computing the Jensen-Shannon divergence (<span>(D_{JS})</span>) between state predictions and sensor-obtained corrections. This initiates a two-tiered data-driven mechanism: one layer employing Machine Learning for fault detection, and another distinct layer for fault isolation. The methodology’s efficacy is assessed using real data from the Turtlebot3 platform.</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"49 2","pages":""},"PeriodicalIF":3.7,"publicationDate":"2025-04-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143840435","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Human2bot: learning zero-shot reward functions for robotic manipulation from human demonstrations Human2bot:从人类演示中学习机器人操作的零射击奖励函数
IF 3.7 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-04-15 DOI: 10.1007/s10514-025-10193-9
Yasir Salam, Yinbei Li, Jonas Herzog, Jiaqiang Yang

Developing effective reward functions is crucial for robot learning, as they guide behavior and facilitate adaptation to human-like tasks. We present Human2Bot (H2B), advancing the learning of such a generalized multi-task reward function that can be used zero-shot to execute unknown tasks in unseen environments. H2B is a newly designed task similarity estimation model that is trained on a large dataset of human videos. The model determines whether two videos from different environments represent the same task. At test time, the model serves as a reward function, evaluating how closely a robot’s execution matches the human demonstration. While previous approaches necessitate robot-specific data to learn reward functions or policies, our method can learn without any robot datasets. To achieve generalization in robotic environments, we incorporate a domain augmentation process that generates synthetic videos with varied visual appearances resembling simulation environments, alongside a multi-scale inter-frame attention mechanism that aligns human and robot task understanding. Finally, H2B is integrated with Visual Model Predictive Control (VMPC) to perform manipulation tasks in simulation and on the xARM6 robot in real-world settings. Our approach outperforms previous methods in simulated and real-world environments trained solely on human data, eliminating the need for privileged robot datasets.

开发有效的奖励功能对机器人学习至关重要,因为它们指导行为并促进适应类似人类的任务。我们提出了Human2Bot (H2B),推进了这种广义多任务奖励函数的学习,该函数可以在不可见的环境中使用零射击来执行未知任务。H2B是一种新设计的任务相似度估计模型,它是在一个大型人类视频数据集上训练的。该模型确定来自不同环境的两个视频是否代表相同的任务。在测试时,该模型作为奖励函数,评估机器人的执行与人类演示的接近程度。虽然以前的方法需要特定于机器人的数据来学习奖励函数或策略,但我们的方法可以在没有任何机器人数据集的情况下学习。为了在机器人环境中实现泛化,我们结合了一个域增强过程,该过程生成具有类似模拟环境的各种视觉外观的合成视频,以及一个多尺度帧间注意机制,使人类和机器人的任务理解保持一致。最后,H2B与视觉模型预测控制(VMPC)集成,在模拟和现实世界设置的xARM6机器人上执行操作任务。我们的方法在模拟和现实环境中优于以前的方法,这些方法只训练人类数据,消除了对特权机器人数据集的需求。
{"title":"Human2bot: learning zero-shot reward functions for robotic manipulation from human demonstrations","authors":"Yasir Salam,&nbsp;Yinbei Li,&nbsp;Jonas Herzog,&nbsp;Jiaqiang Yang","doi":"10.1007/s10514-025-10193-9","DOIUrl":"10.1007/s10514-025-10193-9","url":null,"abstract":"<div><p>Developing effective reward functions is crucial for robot learning, as they guide behavior and facilitate adaptation to human-like tasks. We present Human2Bot (H2B), advancing the learning of such a generalized multi-task reward function that can be used zero-shot to execute unknown tasks in unseen environments. H2B is a newly designed task similarity estimation model that is trained on a large dataset of human videos. The model determines whether two videos from different environments represent the same task. At test time, the model serves as a reward function, evaluating how closely a robot’s execution matches the human demonstration. While previous approaches necessitate robot-specific data to learn reward functions or policies, our method can learn without any robot datasets. To achieve generalization in robotic environments, we incorporate a domain augmentation process that generates synthetic videos with varied visual appearances resembling simulation environments, alongside a multi-scale inter-frame attention mechanism that aligns human and robot task understanding. Finally, H2B is integrated with Visual Model Predictive Control (VMPC) to perform manipulation tasks in simulation and on the xARM6 robot in real-world settings. Our approach outperforms previous methods in simulated and real-world environments trained solely on human data, eliminating the need for privileged robot datasets.</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"49 2","pages":""},"PeriodicalIF":3.7,"publicationDate":"2025-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143835536","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Autonomous Robots
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1