首页 > 最新文献

Autonomous Robots最新文献

英文 中文
View: visual imitation learning with waypoints 视图:带路径点的视觉模仿学习
IF 3.7 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-01-18 DOI: 10.1007/s10514-024-10188-y
Ananth Jonnavittula, Sagar Parekh, Dylan P. Losey

Robots can use visual imitation learning (VIL) to learn manipulation tasks from video demonstrations. However, translating visual observations into actionable robot policies is challenging due to the high-dimensional nature of video data. This challenge is further exacerbated by the morphological differences between humans and robots, especially when the video demonstrations feature humans performing tasks. To address these problems we introduce Visual Imitation lEarning with Waypoints (VIEW), an algorithm that significantly enhances the sample efficiency of human-to-robot VIL. VIEW achieves this efficiency using a multi-pronged approach: extracting a condensed prior trajectory that captures the demonstrator’s intent, employing an agent-agnostic reward function for feedback on the robot’s actions, and utilizing an exploration algorithm that efficiently samples around waypoints in the extracted trajectory. VIEW also segments the human trajectory into grasp and task phases to further accelerate learning efficiency. Through comprehensive simulations and real-world experiments, VIEW demonstrates improved performance compared to current state-of-the-art VIL methods. VIEW enables robots to learn manipulation tasks involving multiple objects from arbitrarily long video demonstrations. Additionally, it can learn standard manipulation tasks such as pushing or moving objects from a single video demonstration in under 30 min, with fewer than 20 real-world rollouts. Code and videos here: https://collab.me.vt.edu/view/

机器人可以使用视觉模仿学习(VIL)从视频演示中学习操作任务。然而,由于视频数据的高维性质,将视觉观察转化为可操作的机器人政策是具有挑战性的。人类和机器人之间的形态差异进一步加剧了这一挑战,特别是当视频演示以人类执行任务为特征时。为了解决这些问题,我们引入了带有路径点的视觉模仿学习(VIEW)算法,该算法显著提高了人对机器人VIL的采样效率。VIEW通过多管齐下的方法实现了这一效率:提取一个浓缩的先验轨迹,捕捉演示者的意图,采用一个代理不可知的奖励函数来反馈机器人的动作,并利用一个探索算法,在提取的轨迹中有效地对路点进行采样。VIEW还将人的轨迹划分为掌握和任务阶段,以进一步提高学习效率。通过全面的仿真和真实世界的实验,VIEW展示了与当前最先进的VIL方法相比,其性能有所提高。VIEW使机器人能够从任意长的视频演示中学习涉及多个对象的操作任务。此外,它可以在30分钟内从一个视频演示中学习标准的操作任务,例如推动或移动物体,而现实世界的演示次数不到20次。代码和视频在这里:https://collab.me.vt.edu/view/
{"title":"View: visual imitation learning with waypoints","authors":"Ananth Jonnavittula,&nbsp;Sagar Parekh,&nbsp;Dylan P. Losey","doi":"10.1007/s10514-024-10188-y","DOIUrl":"10.1007/s10514-024-10188-y","url":null,"abstract":"<div><p>Robots can use visual imitation learning (VIL) to learn manipulation tasks from video demonstrations. However, translating visual observations into actionable robot policies is challenging due to the high-dimensional nature of video data. This challenge is further exacerbated by the morphological differences between humans and robots, especially when the video demonstrations feature humans performing tasks. To address these problems we introduce <b>V</b>isual <b>I</b>mitation l<b>E</b>arning with <b>W</b>aypoints (VIEW), an algorithm that significantly enhances the sample efficiency of human-to-robot VIL. VIEW achieves this efficiency using a multi-pronged approach: extracting a condensed prior trajectory that captures the demonstrator’s intent, employing an agent-agnostic reward function for feedback on the robot’s actions, and utilizing an exploration algorithm that efficiently samples around waypoints in the extracted trajectory. VIEW also segments the human trajectory into grasp and task phases to further accelerate learning efficiency. Through comprehensive simulations and real-world experiments, VIEW demonstrates improved performance compared to current state-of-the-art VIL methods. VIEW enables robots to learn manipulation tasks involving multiple objects from arbitrarily long video demonstrations. Additionally, it can learn standard manipulation tasks such as pushing or moving objects from a single video demonstration in under 30 min, with fewer than 20 real-world rollouts. Code and videos here: https://collab.me.vt.edu/view/</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"49 1","pages":""},"PeriodicalIF":3.7,"publicationDate":"2025-01-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10514-024-10188-y.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142995247","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Safe and stable teleoperation of quadrotor UAVs under haptic shared autonomy 触觉共享自主下四旋翼无人机安全稳定的远程操作
IF 3.7 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-01-17 DOI: 10.1007/s10514-024-10186-0
Dawei Zhang, Roberto Tron

We present a novel approach that aims to address both safety and stability of a haptic teleoperation system within a framework of Haptic Shared Autonomy (HSA). We use Control Barrier Functions (CBFs) to generate the control input that follows the user’s input as closely as possible while guaranteeing safety. In the context of stability of the human-in-the-loop system, we limit the force feedback perceived by the user via a small (mathcal {L}_2)-gain, which is achieved by limiting the control and the force feedback via a differential constraint. Specifically, with the property of HSA, we propose two pathways to design the control and the force feedback: Sequential Control Force (SCF) and Joint Control Force (JCF). Both designs can achieve safety and stability but with different responses to the user’s commands. We conducted experimental simulations to evaluate and investigate the properties of the designed methods. We also tested the proposed method on a physical quadrotor UAV and a haptic interface.

我们提出了一种新的方法,旨在解决触觉共享自治(HSA)框架内触觉远程操作系统的安全性和稳定性。我们使用控制屏障函数(cbf)来生成尽可能紧跟用户输入的控制输入,同时保证安全性。在人在环系统稳定性的背景下,我们通过一个小的(mathcal {L}_2)增益来限制用户感知的力反馈,这是通过限制控制和力反馈来实现的微分约束。具体地说,根据HSA的特性,我们提出了两种设计控制和力反馈的途径:顺序控制力(SCF)和联合控制力(JCF)。两种设计都可以实现安全性和稳定性,但对用户命令的响应不同。我们进行了实验模拟来评估和研究所设计方法的性能。我们还在物理四旋翼无人机和触觉界面上测试了所提出的方法。
{"title":"Safe and stable teleoperation of quadrotor UAVs under haptic shared autonomy","authors":"Dawei Zhang,&nbsp;Roberto Tron","doi":"10.1007/s10514-024-10186-0","DOIUrl":"10.1007/s10514-024-10186-0","url":null,"abstract":"<div><p>We present a novel approach that aims to address both safety and stability of a haptic teleoperation system within a framework of Haptic Shared Autonomy (HSA). We use Control Barrier Functions (CBFs) to generate the control input that follows the user’s input as closely as possible while guaranteeing safety. In the context of stability of the human-in-the-loop system, we limit the force feedback perceived by the user via a small <span>(mathcal {L}_2)</span>-gain, which is achieved by limiting the control and the force feedback via a differential constraint. Specifically, with the property of HSA, we propose two pathways to design the control and the force feedback: Sequential Control Force (SCF) and Joint Control Force (JCF). Both designs can achieve safety and stability but with different responses to the user’s commands. We conducted experimental simulations to evaluate and investigate the properties of the designed methods. We also tested the proposed method on a physical quadrotor UAV and a haptic interface.</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"49 1","pages":""},"PeriodicalIF":3.7,"publicationDate":"2025-01-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142995138","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Synthesizing compact behavior trees for probabilistic robotics domains 概率机器人领域的紧凑行为树综合
IF 3.7 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-01-14 DOI: 10.1007/s10514-024-10187-z
Emily Scheide, Graeme Best, Geoffrey A. Hollinger

Complex robotics domains (e.g., remote exploration applications and scenarios involving interactions with humans) require encoding high-level mission specifications that consider uncertainty. Most current fielded systems in practice require humans to manually encode mission specifications in ways that require amounts of time and expertise that can become infeasible and limit mission scope. Therefore, we propose a method of automating the process of encoding mission specifications as behavior trees. In particular, we present an algorithm for synthesizing behavior trees that represent the optimal policy for a user-defined specification of a domain and problem in the Probabilistic Planning Domain Definition Language (PPDDL). Our algorithm provides access to behavior tree advantages including compactness and modularity, while alleviating the need for the time-intensive manual design of behavior trees, which requires substantial expert knowledge. Our method converts the PPDDL specification into solvable MDP matrices, simplifies the solution, i.e. policy, using Boolean algebra simplification, and converts this simplified policy to a compact behavior tree that can be executed by a robot. We present simulated experiments for a marine target search and response scenario and an infant-robot interaction for mobility domain. Our results demonstrate that the synthesized, simplified behavior trees have approximately between 15 x and 26 x fewer nodes and an average of between 8 x and 13 x fewer active conditions for selecting the active action than they would without simplification. These compactness and activity results suggest an increase in the interpretability and execution efficiency of the behavior trees synthesized by the proposed method. Additionally, our results demonstrate that this synthesis method is robust to a variety of user input mistakes, and we empirically confirm that the synthesized behavior trees perform equivalently to the optimal policy that they are constructed to logically represent.

复杂的机器人领域(例如,远程探索应用程序和涉及与人类交互的场景)需要编码考虑不确定性的高级任务规范。在实践中,大多数当前的现场系统需要人类手动编码任务规范,这种方式需要大量的时间和专业知识,这可能变得不可行并限制任务范围。因此,我们提出了一种将任务规范编码过程自动化为行为树的方法。特别是,我们提出了一种综合行为树的算法,这些行为树代表了概率规划领域定义语言(PPDDL)中用户定义的领域规范和问题的最佳策略。我们的算法提供了行为树的优点,包括紧凑性和模块化,同时减轻了需要大量专家知识的时间密集型人工设计行为树的需要。我们的方法将PPDDL规范转换为可解的MDP矩阵,使用布尔代数简化将解即策略简化,并将此简化策略转换为可由机器人执行的紧凑行为树。我们提出了一个海洋目标搜索和响应场景的模拟实验和一个婴儿-机器人在移动领域的交互。我们的结果表明,与没有简化的行为树相比,合成的、简化的行为树的节点大约减少了15到26倍,选择主动动作的活动条件平均减少了8到13倍。这些紧凑性和活动性的结果表明,该方法合成的行为树的可解释性和执行效率都有所提高。此外,我们的结果表明,这种合成方法对各种用户输入错误具有鲁棒性,并且我们经验地证实,合成行为树的性能等同于它们被构造为逻辑表示的最优策略。
{"title":"Synthesizing compact behavior trees for probabilistic robotics domains","authors":"Emily Scheide,&nbsp;Graeme Best,&nbsp;Geoffrey A. Hollinger","doi":"10.1007/s10514-024-10187-z","DOIUrl":"10.1007/s10514-024-10187-z","url":null,"abstract":"<div><p>Complex robotics domains (e.g., remote exploration applications and scenarios involving interactions with humans) require encoding high-level mission specifications that consider uncertainty. Most current fielded systems in practice require humans to manually encode mission specifications in ways that require amounts of time and expertise that can become infeasible and limit mission scope. Therefore, we propose a method of automating the process of encoding mission specifications as behavior trees. In particular, we present an algorithm for synthesizing behavior trees that represent the optimal policy for a user-defined specification of a domain and problem in the Probabilistic Planning Domain Definition Language (PPDDL). Our algorithm provides access to behavior tree advantages including compactness and modularity, while alleviating the need for the time-intensive manual design of behavior trees, which requires substantial expert knowledge. Our method converts the PPDDL specification into solvable MDP matrices, simplifies the solution, i.e. policy, using Boolean algebra simplification, and converts this simplified policy to a compact behavior tree that can be executed by a robot. We present simulated experiments for a marine target search and response scenario and an infant-robot interaction for mobility domain. Our results demonstrate that the synthesized, simplified behavior trees have approximately between 15 <span>x</span> and 26 <span>x</span> fewer nodes and an average of between 8 <span>x</span> and 13 <span>x</span> fewer active conditions for selecting the active action than they would without simplification. These compactness and activity results suggest an increase in the interpretability and execution efficiency of the behavior trees synthesized by the proposed method. Additionally, our results demonstrate that this synthesis method is robust to a variety of user input mistakes, and we empirically confirm that the synthesized behavior trees perform equivalently to the optimal policy that they are constructed to logically represent.\u0000</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"49 1","pages":""},"PeriodicalIF":3.7,"publicationDate":"2025-01-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142976351","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Integrative biomechanics of a human–robot carrying task: implications for future collaborative work 人-机器人承载任务的综合生物力学:对未来协同工作的启示
IF 3.7 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-01-09 DOI: 10.1007/s10514-024-10184-2
Verena Schuengel, Bjoern Braunstein, Fabian Goell, Daniel Braun, Nadine Reißner, Kirill Safronov, Christian Weiser, Jule Heieis, Kirsten Albracht

Patients with sarcopenia, who face difficulties in carrying heavy loads, may benefit from collaborative robotic assistance that is modeled after human–human interaction. The objective of this study is to describe the kinematics and spatio-temporal parameters during a collaborative carrying task involving both human and robotic partners. Fourteen subjects carried a table while moving forward with a human and a robotic partner. The movements were recorded using a three-dimensional motion capture system. The subjects successfully completed the task of carrying the table with the robot. No significant differences were found in the shoulder and elbow flexion/extension angles. In human–human dyads, the center of mass naturally oscillated vertically with an amplitude of approximately 2 cm. The here presented results of the human–human interaction serve as a model for the development of future robotic systems, designed for collaborative manipulation.

肌肉减少症患者在搬运重物时遇到困难,他们可能会受益于模仿人类互动的协作机器人协助。本研究的目的是描述在涉及人类和机器人伙伴的协作搬运任务中的运动学和时空参数。14名受试者与一个人和一个机器人搭档一起抬着桌子向前移动。这些动作是用三维动作捕捉系统记录下来的。受试者成功地完成了机器人搬运桌子的任务。肩关节和肘关节屈伸角度无显著差异。在人类二人组中,质心自然地垂直振荡,振幅约为2厘米。这里展示的人机交互结果可以作为未来机器人系统开发的模型,用于协作操作。
{"title":"Integrative biomechanics of a human–robot carrying task: implications for future collaborative work","authors":"Verena Schuengel,&nbsp;Bjoern Braunstein,&nbsp;Fabian Goell,&nbsp;Daniel Braun,&nbsp;Nadine Reißner,&nbsp;Kirill Safronov,&nbsp;Christian Weiser,&nbsp;Jule Heieis,&nbsp;Kirsten Albracht","doi":"10.1007/s10514-024-10184-2","DOIUrl":"10.1007/s10514-024-10184-2","url":null,"abstract":"<div><p>Patients with sarcopenia, who face difficulties in carrying heavy loads, may benefit from collaborative robotic assistance that is modeled after human–human interaction. The objective of this study is to describe the kinematics and spatio-temporal parameters during a collaborative carrying task involving both human and robotic partners. Fourteen subjects carried a table while moving forward with a human and a robotic partner. The movements were recorded using a three-dimensional motion capture system. The subjects successfully completed the task of carrying the table with the robot. No significant differences were found in the shoulder and elbow flexion/extension angles. In human–human dyads, the center of mass naturally oscillated vertically with an amplitude of approximately 2 cm. The here presented results of the human–human interaction serve as a model for the development of future robotic systems, designed for collaborative manipulation.\u0000</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"49 1","pages":""},"PeriodicalIF":3.7,"publicationDate":"2025-01-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10514-024-10184-2.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142938921","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Mori-zwanzig approach for belief abstraction with application to belief space planning Mori-zwanzig信念抽象方法及其在信念空间规划中的应用
IF 3.7 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-12-24 DOI: 10.1007/s10514-024-10185-1
Mengxue Hou, Tony X. Lin, Enlu Zhou, Fumin Zhang

We propose a learning-based method to extract symbolic representations of the belief state and its dynamics in order to solve planning problems in a continuous-state partially observable Markov decision processes (POMDP) problem. While existing approaches typically parameterize the continuous-state POMDP into a finite-dimensional Markovian model, they are unable to preserve fidelity of the abstracted model. To improve accuracy of the abstracted representation, we introduce a memory-dependent abstraction approach to mitigate the modeling error. The first major contribution of this paper is we propose a Neural Network based method to learn the non-Markovian transition model based on the Mori-Zwanzig (M-Z) formalism. Different from existing work in applying M-Z formalism to autonomous time-invariant systems, our approach is the first work generalizing the M-Z formalism to robotics, by addressing the non-Markovian modeling of the belief dynamics that is dependent on historical observations and actions. The second major contribution is we theoretically show that modeling the non-Markovian memory effect in the abstracted belief dynamics improves the modeling accuracy, which is the key benefit of the proposed algorithm. Simulation experiment of a belief space planning problem is provided to validate the performance of the proposed belief abstraction algorithms.

为了解决连续状态部分可观察马尔可夫决策过程(POMDP)中的规划问题,提出了一种基于学习的方法来提取信念状态及其动态的符号表示。虽然现有方法通常将连续状态POMDP参数化为有限维马尔可夫模型,但它们无法保持抽象模型的保真度。为了提高抽象表示的准确性,我们引入了一种依赖于内存的抽象方法来减少建模误差。本文的第一个主要贡献是我们提出了一种基于Mori-Zwanzig (M-Z)形式主义的基于神经网络的非马尔可夫转移模型学习方法。与将M-Z形式主义应用于自主时不变系统的现有工作不同,我们的方法是第一个将M-Z形式主义推广到机器人的工作,通过解决依赖于历史观察和行为的信念动力学的非马尔可夫建模。第二个主要贡献是我们从理论上证明了在抽象的信念动力学中建模非马尔可夫记忆效应提高了建模精度,这是该算法的主要优点。通过一个信念空间规划问题的仿真实验,验证了所提出的信念抽象算法的性能。
{"title":"Mori-zwanzig approach for belief abstraction with application to belief space planning","authors":"Mengxue Hou,&nbsp;Tony X. Lin,&nbsp;Enlu Zhou,&nbsp;Fumin Zhang","doi":"10.1007/s10514-024-10185-1","DOIUrl":"10.1007/s10514-024-10185-1","url":null,"abstract":"<div><p>We propose a learning-based method to extract symbolic representations of the belief state and its dynamics in order to solve planning problems in a continuous-state partially observable Markov decision processes (POMDP) problem. While existing approaches typically parameterize the continuous-state POMDP into a finite-dimensional Markovian model, they are unable to preserve fidelity of the abstracted model. To improve accuracy of the abstracted representation, we introduce a memory-dependent abstraction approach to mitigate the modeling error. The first major contribution of this paper is we propose a Neural Network based method to learn the non-Markovian transition model based on the Mori-Zwanzig (M-Z) formalism. Different from existing work in applying M-Z formalism to autonomous time-invariant systems, our approach is the first work generalizing the M-Z formalism to robotics, by addressing the non-Markovian modeling of the belief dynamics that is dependent on historical observations and actions. The second major contribution is we theoretically show that modeling the non-Markovian memory effect in the abstracted belief dynamics improves the modeling accuracy, which is the key benefit of the proposed algorithm. Simulation experiment of a belief space planning problem is provided to validate the performance of the proposed belief abstraction algorithms.\u0000</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"49 1","pages":""},"PeriodicalIF":3.7,"publicationDate":"2024-12-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10514-024-10185-1.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142880465","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multirotor nonlinear model predictive control based on visual servoing of evolving features 基于演化特征视觉伺服的多旋翼非线性模型预测控制
IF 3.7 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-11-28 DOI: 10.1007/s10514-024-10183-3
Sotirios N. Aspragkathos, Panagiotis Rousseas, George C. Karras, Kostas J. Kyriakopoulos

This article presents a Visual Servoing Nonlinear Model Predictive Control (NMPC) scheme for autonomously tracking a moving target using multirotor Unmanned Aerial Vehicles (UAVs). The scheme is developed for surveillance and tracking of contour-based areas with evolving features. NMPC is used to manage input and state constraints, while additional barrier functions are incorporated in order to ensure system safety and optimal performance. The proposed control scheme is designed based on the extraction and implementation of the full dynamic model of the features describing the target and the state variables. Real-time simulations and experiments using a quadrotor UAV equipped with a camera demonstrate the effectiveness of the proposed strategy.

本文介绍了一种视觉伺服非线性模型预测控制(NMPC)方案,用于使用多旋翼无人飞行器(UAV)自主跟踪移动目标。该方案是为监视和跟踪具有不断变化特征的等高线区域而开发的。NMPC 用于管理输入和状态约束,同时加入了额外的屏障功能,以确保系统安全和最佳性能。所提出的控制方案是在提取和实施描述目标和状态变量特征的全动态模型的基础上设计的。使用装有摄像头的四旋翼无人机进行的实时模拟和实验证明了所提策略的有效性。
{"title":"Multirotor nonlinear model predictive control based on visual servoing of evolving features","authors":"Sotirios N. Aspragkathos,&nbsp;Panagiotis Rousseas,&nbsp;George C. Karras,&nbsp;Kostas J. Kyriakopoulos","doi":"10.1007/s10514-024-10183-3","DOIUrl":"10.1007/s10514-024-10183-3","url":null,"abstract":"<div><p>This article presents a Visual Servoing Nonlinear Model Predictive Control (NMPC) scheme for autonomously tracking a moving target using multirotor Unmanned Aerial Vehicles (UAVs). The scheme is developed for surveillance and tracking of contour-based areas with evolving features. NMPC is used to manage input and state constraints, while additional barrier functions are incorporated in order to ensure system safety and optimal performance. The proposed control scheme is designed based on the extraction and implementation of the full dynamic model of the features describing the target and the state variables. Real-time simulations and experiments using a quadrotor UAV equipped with a camera demonstrate the effectiveness of the proposed strategy.</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"48 8","pages":""},"PeriodicalIF":3.7,"publicationDate":"2024-11-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142737228","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
BFAR: improving radar odometry estimation using a bounded false alarm rate detector BFAR:使用有界误报率检测器改进雷达测距估算
IF 3.7 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-11-19 DOI: 10.1007/s10514-024-10176-2
Anas Alhashimi, Daniel Adolfsson, Henrik Andreasson, Achim Lilienthal, Martin Magnusson

This work introduces a novel detector, bounded false-alarm rate (BFAR), for distinguishing true detections from noise in radar data, leading to improved accuracy in radar odometry estimation. Scanning frequency-modulated continuous wave (FMCW) radars can serve as valuable tools for localization and mapping under low visibility conditions. However, they tend to yield a higher level of noise in comparison to the more commonly employed lidars, thereby introducing additional challenges to the detection process. We propose a new radar target detector called BFAR which uses an affine transformation of the estimated noise level compared to the classical constant false-alarm rate (CFAR) detector. This transformation employs learned parameters that minimize the error in odometry estimation. Conceptually, BFAR can be viewed as an optimized blend of CFAR and fixed-level thresholding designed to minimize odometry estimation error. The strength of this approach lies in its simplicity. Only a single parameter needs to be learned from a training dataset when the affine transformation scale parameter is maintained. Compared to ad-hoc detectors, BFAR has the advantage of a specified upper-bound for the false-alarm probability, and better noise handling than CFAR. Repeatability tests show that BFAR yields highly repeatable detections with minimal redundancy. We have conducted simulations to compare the detection and false-alarm probabilities of BFAR with those of three baselines in non-homogeneous noise and varying target sizes. The results show that BFAR outperforms the other detectors. Moreover, We apply BFAR to the use case of radar odometry, and adapt a recent odometry pipeline, replacing its original conservative filtering with BFAR. In this way, we reduce the translation/rotation odometry errors/100 m from 1.3%/0.4(^circ ) to 1.12%/0.38(^circ ), and from 1.62%/0.57(^circ ) to 1.21%/0.32(^circ ), improving translation error by 14.2% and 25% on Oxford and Mulran public data sets, respectively.

这项工作介绍了一种新型检测器--有界误报率(BFAR),用于区分雷达数据中的真实检测和噪声,从而提高雷达测距估算的准确性。扫描频率调制连续波(FMCW)雷达是低能见度条件下进行定位和绘图的重要工具。然而,与更常用的激光雷达相比,扫描频率调制连续波(FMCW)雷达往往会产生更高水平的噪声,从而给探测过程带来额外的挑战。我们提出了一种名为 BFAR 的新型雷达目标检测器,与传统的恒定误报率(CFAR)检测器相比,该检测器对估计噪声水平进行了仿射变换。这种变换采用了学习到的参数,能最大限度地减少测距估计中的误差。从概念上讲,BFAR 可以看作是 CFAR 和固定阈值的优化组合,旨在最大限度地减少里程估算误差。这种方法的优势在于其简单性。在保持仿射变换比例参数的情况下,只需从训练数据集中学习一个参数。与临时检测器相比,BFAR 的优势在于为误报概率指定了上限,而且噪声处理能力比 CFAR 更强。可重复性测试表明,BFAR 能以最小的冗余产生高度可重复的检测结果。我们进行了仿真,比较了 BFAR 与三种基线在非均匀噪声和不同目标大小条件下的检测概率和误报概率。结果表明,BFAR 的性能优于其他探测器。此外,我们还将 BFAR 应用于雷达测距,并调整了最新的测距管道,用 BFAR 取代了原有的保守滤波。通过这种方法,我们将平移/旋转测度误差/100米从1.3%/0.4(^circ )降低到1.12%/0.38(^circ ),将平移误差/100米从1.62%/0.57(^circ )降低到1.21%/0.32(^circ ),在牛津和穆兰公共数据集上,平移误差分别改善了14.2%和25%。
{"title":"BFAR: improving radar odometry estimation using a bounded false alarm rate detector","authors":"Anas Alhashimi,&nbsp;Daniel Adolfsson,&nbsp;Henrik Andreasson,&nbsp;Achim Lilienthal,&nbsp;Martin Magnusson","doi":"10.1007/s10514-024-10176-2","DOIUrl":"10.1007/s10514-024-10176-2","url":null,"abstract":"<div><p>This work introduces a novel detector, bounded false-alarm rate (BFAR), for distinguishing true detections from noise in radar data, leading to improved accuracy in radar odometry estimation. Scanning frequency-modulated continuous wave (FMCW) radars can serve as valuable tools for localization and mapping under low visibility conditions. However, they tend to yield a higher level of noise in comparison to the more commonly employed lidars, thereby introducing additional challenges to the detection process. We propose a new radar target detector called BFAR which uses an affine transformation of the estimated noise level compared to the classical constant false-alarm rate (CFAR) detector. This transformation employs learned parameters that minimize the error in odometry estimation. Conceptually, BFAR can be viewed as an optimized blend of CFAR and fixed-level thresholding designed to minimize odometry estimation error. The strength of this approach lies in its simplicity. Only a single parameter needs to be learned from a training dataset when the affine transformation scale parameter is maintained. Compared to ad-hoc detectors, BFAR has the advantage of a specified upper-bound for the false-alarm probability, and better noise handling than CFAR. Repeatability tests show that BFAR yields highly repeatable detections with minimal redundancy. We have conducted simulations to compare the detection and false-alarm probabilities of BFAR with those of three baselines in non-homogeneous noise and varying target sizes. The results show that BFAR outperforms the other detectors. Moreover, We apply BFAR to the use case of radar odometry, and adapt a recent odometry pipeline, replacing its original conservative filtering with BFAR. In this way, we reduce the translation/rotation odometry errors/100 m from 1.3%/0.4<span>(^circ )</span> to 1.12%/0.38<span>(^circ )</span>, and from 1.62%/0.57<span>(^circ )</span> to 1.21%/0.32<span>(^circ )</span>, improving translation error by 14.2% and 25% on Oxford and Mulran public data sets, respectively.</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"48 8","pages":""},"PeriodicalIF":3.7,"publicationDate":"2024-11-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10514-024-10176-2.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142672386","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SAR: generalization of physiological agility and dexterity via synergistic action representation SAR:通过协同作用表示法概括生理敏捷性和灵巧性
IF 3.7 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-11-14 DOI: 10.1007/s10514-024-10182-4
Cameron Berg, Vittorio Caggiano, Vikash Kumar

Learning effective continuous control policies in high-dimensional systems, including musculoskeletal agents, remains a significant challenge. Over the course of biological evolution, organisms have developed robust mechanisms for overcoming this complexity to learn highly sophisticated strategies for motor control. What accounts for this robust behavioral flexibility? Modular control via muscle synergies, i.e. coordinated muscle co-contractions, is considered to be one putative mechanism that enables organisms to learn muscle control in a simplified and generalizable action space. Drawing inspiration from this evolved motor control strategy, we use physiologically accurate human hand and leg models as a testbed for determining the extent to which a Synergistic Action Representation (SAR) acquired from simpler tasks facilitates learning and generalization on more complex tasks. We find in both cases that SAR-exploiting policies significantly outperform end-to-end reinforcement learning. Policies trained with SAR were able to achieve robust locomotion on a diverse set of terrains (e.g., stairs, hills) with state-of-the-art sample efficiency (4 M total steps), while baseline approaches failed to learn any meaningful behaviors under the same training regime. Additionally, policies trained with SAR on in-hand 100-object manipulation task significantly outperformed (>70% success) baseline approaches (<20% success). Both SAR-exploiting policies were also found to generalize zero-shot to out-of-domain environmental conditions, while policies that did not adopt SAR failed to generalize. Finally, using a simulated robotic hand and humanoid agent, we establish the generality of SAR on broader high-dimensional control problems, solving tasks with greatly improved sample efficiency. To the best of our knowledge, this investigation is the first of its kind to present an end-to-end pipeline for discovering synergies and using this representation to learn high-dimensional continuous control across a wide diversity of tasks. Project website:https://sites.google.com/view/sar-rl

在高维系统(包括肌肉骨骼系统)中学习有效的连续控制策略仍然是一项重大挑战。在生物进化的过程中,生物已经发展出克服这种复杂性的强大机制,从而学会了高度复杂的运动控制策略。是什么造就了这种强大的行为灵活性?通过肌肉协同作用(即协调的肌肉共同收缩)进行的模块化控制被认为是一种推定机制,它使生物能够在简化和可泛化的动作空间中学习肌肉控制。从这种进化的运动控制策略中汲取灵感,我们使用生理上精确的人类手部和腿部模型作为试验平台,以确定从较简单任务中获得的协同动作表征(SAR)在多大程度上促进了对较复杂任务的学习和泛化。我们发现,在这两种情况下,利用 SAR 的策略都明显优于端到端强化学习。利用 SAR 训练的策略能够在各种地形(如楼梯、山丘)上实现稳健的运动,并具有最先进的采样效率(总步数为 400 万步),而基线方法在相同的训练机制下无法学习到任何有意义的行为。此外,在手持 100 个物体的操作任务中,使用 SAR 训练的策略明显优于基线方法(成功率为 70%)(成功率为 20%)。研究还发现,这两种利用合成孔径雷达的策略都能在域外环境条件下实现零误差泛化,而未采用合成孔径雷达的策略则无法实现泛化。最后,我们利用模拟机器人手和仿人代理,在更广泛的高维控制问题上确立了 SAR 的通用性,大大提高了解决任务的采样效率。据我们所知,这项研究首次提出了一个端到端的管道,用于发现协同效应,并利用这种表示学习各种任务的高维连续控制。项目网站:https://sites.google.com/view/sar-rl
{"title":"SAR: generalization of physiological agility and dexterity via synergistic action representation","authors":"Cameron Berg,&nbsp;Vittorio Caggiano,&nbsp;Vikash Kumar","doi":"10.1007/s10514-024-10182-4","DOIUrl":"10.1007/s10514-024-10182-4","url":null,"abstract":"<div><p>Learning effective continuous control policies in high-dimensional systems, including musculoskeletal agents, remains a significant challenge. Over the course of biological evolution, organisms have developed robust mechanisms for overcoming this complexity to learn highly sophisticated strategies for motor control. What accounts for this robust behavioral flexibility? Modular control via muscle synergies, i.e. coordinated muscle co-contractions, is considered to be one putative mechanism that enables organisms to learn muscle control in a simplified and generalizable action space. Drawing inspiration from this evolved motor control strategy, we use physiologically accurate human hand and leg models as a testbed for determining the extent to which a <i>Synergistic Action Representation</i> (<i>SAR</i>) acquired from simpler tasks facilitates learning and generalization on more complex tasks. We find in both cases that <i>SAR</i>-exploiting policies significantly outperform end-to-end reinforcement learning. Policies trained with <i>SAR</i> were able to achieve robust locomotion on a diverse set of terrains (e.g., stairs, hills) with state-of-the-art sample efficiency (4 M total steps), while baseline approaches failed to learn any meaningful behaviors under the same training regime. Additionally, policies trained with <i>SAR</i> on in-hand 100-object manipulation task significantly outperformed (&gt;70% success) baseline approaches (&lt;20% success). Both <i>SAR</i>-exploiting policies were also found to generalize zero-shot to out-of-domain environmental conditions, while policies that did not adopt <i>SAR</i> failed to generalize. Finally, using a simulated robotic hand and humanoid agent, we establish the generality of SAR on broader high-dimensional control problems, solving tasks with greatly improved sample efficiency. To the best of our knowledge, this investigation is the first of its kind to present an end-to-end pipeline for discovering synergies and using this representation to learn high-dimensional continuous control across a wide diversity of tasks. <b>Project website:</b>https://sites.google.com/view/sar-rl</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"48 8","pages":""},"PeriodicalIF":3.7,"publicationDate":"2024-11-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142636655","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimal policies for autonomous navigation in strong currents using fast marching trees 利用快速行进树在强水流中实现自主导航的最佳策略
IF 3.7 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-22 DOI: 10.1007/s10514-024-10179-z
Bernardo Martinez Rocamora Jr., Guilherme A. S. Pereira

Several applications require that unmanned vehicles, such as UAVs and AUVs, navigate environmental flows. While the flow can improve the vehicle’s efficiency when directed towards the goal, it may also cause feasibility problems when it is against the desired motion and is too strong to be counteracted by the vehicle. This paper proposes the flow-aware fast marching tree algorithm (FlowFMT*) to solve the optimal motion planning problem in generic three-dimensional flows. Our method creates either an optimal path from start to goal or, with a few modifications, a vector field-based policy that guides the vehicle from anywhere in its workspace to the goal. The basic idea of the proposed method is to replace the original neighborhood set used by FMT* with two sets that consider the reachability from/to each sampled position in the space. The new neighborhood sets are computed considering the flow and the maximum speed of the vehicle. Numerical results that compare our methods with the state-of-the-art optimal control solver illustrate the simplicity and correctness of the method.

在一些应用中,无人飞行器(如无人潜航器和自动潜航器)需要导航环境流。当环境流指向目标时,可以提高飞行器的效率,但当环境流与飞行器的运动目标相悖且强度过大时,也可能导致可行性问题。 本文提出了流量感知快速行进树算法(FlowFMT*)来解决一般三维流中的最优运动规划问题。我们的方法既可以创建一条从起点到目标的最优路径,也可以在稍作修改后创建一个基于矢量场的策略,引导车辆从其工作空间的任意位置到达目标。 所提方法的基本思想是用两个考虑空间中每个采样位置的可达性的邻域集取代 FMT* 使用的原始邻域集。新的邻域集在计算时考虑了流量和车辆的最大速度。将我们的方法与最先进的最优控制求解器进行比较的数值结果表明了该方法的简便性和正确性。
{"title":"Optimal policies for autonomous navigation in strong currents using fast marching trees","authors":"Bernardo Martinez Rocamora Jr.,&nbsp;Guilherme A. S. Pereira","doi":"10.1007/s10514-024-10179-z","DOIUrl":"10.1007/s10514-024-10179-z","url":null,"abstract":"<div><p>Several applications require that unmanned vehicles, such as UAVs and AUVs, navigate environmental flows. While the flow can improve the vehicle’s efficiency when directed towards the goal, it may also cause feasibility problems when it is against the desired motion and is too strong to be counteracted by the vehicle. This paper proposes the flow-aware fast marching tree algorithm (FlowFMT*) to solve the optimal motion planning problem in generic three-dimensional flows. Our method creates either an optimal path from start to goal or, with a few modifications, a vector field-based policy that guides the vehicle from anywhere in its workspace to the goal. The basic idea of the proposed method is to replace the original neighborhood set used by FMT* with two sets that consider the reachability from/to each sampled position in the space. The new neighborhood sets are computed considering the flow and the maximum speed of the vehicle. Numerical results that compare our methods with the state-of-the-art optimal control solver illustrate the simplicity and correctness of the method.</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"48 8","pages":""},"PeriodicalIF":3.7,"publicationDate":"2024-10-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142453025","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Correction: Planning under uncertainty for safe robot exploration using gaussian process prediction 更正:利用高斯过程预测在不确定条件下为机器人安全探索制定计划
IF 3.7 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-17 DOI: 10.1007/s10514-024-10181-5
Alex Stephens, Matthew Budd, Michal Staniaszek, Benoit Casseau, Paul Duckworth, Maurice Fallon, Nick Hawes, Bruno Lacerda
{"title":"Correction: Planning under uncertainty for safe robot exploration using gaussian process prediction","authors":"Alex Stephens,&nbsp;Matthew Budd,&nbsp;Michal Staniaszek,&nbsp;Benoit Casseau,&nbsp;Paul Duckworth,&nbsp;Maurice Fallon,&nbsp;Nick Hawes,&nbsp;Bruno Lacerda","doi":"10.1007/s10514-024-10181-5","DOIUrl":"10.1007/s10514-024-10181-5","url":null,"abstract":"","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"48 8","pages":""},"PeriodicalIF":3.7,"publicationDate":"2024-10-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10514-024-10181-5.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142447365","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Autonomous Robots
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1