首页 > 最新文献

IEEE Robotics and Automation Letters最新文献

英文 中文
Spider-Inspired Pneumatic Folding Membrane Soft Actuator 蜘蛛启发气动折叠膜软执行器
IF 4.6 2区 计算机科学 Q2 ROBOTICS Pub Date : 2024-12-23 DOI: 10.1109/LRA.2024.3521180
Shuo Liu;Shihao Shen;Yunshan Li;Ming Xu
Soft pneumatic joints are the main driving components of soft arthropod robots, but the problems of high start-up air pressure and low rotation angle limit the range of motion, control accuracy, and flexibility of the robots. Inspired by the hydraulic joints of spider leg, an innovative folding membrane soft joint actuator is designed to achieve an 80° rotation angle at 5.25 kPa air pressure, with an angle density of 15.23°/ kPa and a fatigue life of more than 10,000 times, significantly improving the output characteristics of pneumatic joints. This actuator's folding membrane is made of silicone-coated fiber fabric and is manufactured via integrated molding method. Furthermore, regarding the issue of ineffective expansion in actuators, the effects of various constraint layer materials on the actuator's output performance are investigated and discussed. It is found that the TPU-constrained actuator has a maximum output torque of 0.307 N·m, which is 7.7% and 13.7% greater than that of the fiber-constrained and unconstrained actuators, respectively. Meanwhile, the rebound time was 1.25 seconds, which was reduced by 0.23 seconds and 0.22 seconds respectively. Soft actuators can benefit from the design technique of the folding membrane soft joint actuator presented in the study, which can provide significant torque output at low air pressure and has superior fatigue characteristics.
软气动关节是软体节肢机器人的主要驱动部件,但其启动气压大、旋转角度小等问题限制了机器人的运动范围、控制精度和灵活性。受蜘蛛腿液压关节的启发,设计了一种创新的折叠膜软关节执行器,在5.25 kPa气压下实现80°旋转角度,角度密度为15.23°/ kPa,疲劳寿命超过1万次,显著提高了气动关节的输出特性。该驱动器的折叠膜由有机硅涂层纤维织物制成,并通过集成成型方法制造。此外,针对执行器中的无效展开问题,研究和讨论了不同约束层材料对执行器输出性能的影响。结果表明,tpu约束致动器的最大输出扭矩为0.307 N·m,分别比光纤约束和无光纤约束致动器大7.7%和13.7%。同时,反弹时间为1.25秒,分别缩短了0.23秒和0.22秒。软执行器可以受益于研究中提出的折叠膜软关节执行器的设计技术,该技术可以在低气压下提供显著的扭矩输出,并且具有优越的疲劳特性。
{"title":"Spider-Inspired Pneumatic Folding Membrane Soft Actuator","authors":"Shuo Liu;Shihao Shen;Yunshan Li;Ming Xu","doi":"10.1109/LRA.2024.3521180","DOIUrl":"https://doi.org/10.1109/LRA.2024.3521180","url":null,"abstract":"Soft pneumatic joints are the main driving components of soft arthropod robots, but the problems of high start-up air pressure and low rotation angle limit the range of motion, control accuracy, and flexibility of the robots. Inspired by the hydraulic joints of spider leg, an innovative folding membrane soft joint actuator is designed to achieve an 80° rotation angle at 5.25 kPa air pressure, with an angle density of 15.23°/ kPa and a fatigue life of more than 10,000 times, significantly improving the output characteristics of pneumatic joints. This actuator's folding membrane is made of silicone-coated fiber fabric and is manufactured via integrated molding method. Furthermore, regarding the issue of ineffective expansion in actuators, the effects of various constraint layer materials on the actuator's output performance are investigated and discussed. It is found that the TPU-constrained actuator has a maximum output torque of 0.307 N·m, which is 7.7% and 13.7% greater than that of the fiber-constrained and unconstrained actuators, respectively. Meanwhile, the rebound time was 1.25 seconds, which was reduced by 0.23 seconds and 0.22 seconds respectively. Soft actuators can benefit from the design technique of the folding membrane soft joint actuator presented in the study, which can provide significant torque output at low air pressure and has superior fatigue characteristics.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"10 2","pages":"1401-1408"},"PeriodicalIF":4.6,"publicationDate":"2024-12-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142912526","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning to Model and Plan for Wheeled Mobility on Vertically Challenging Terrain 学习模型和计划在垂直挑战地形轮式移动
IF 4.6 2区 计算机科学 Q2 ROBOTICS Pub Date : 2024-12-23 DOI: 10.1109/LRA.2024.3520919
Aniket Datar;Chenhui Pan;Xuesu Xiao
Most autonomous navigation systems assume wheeled robots are rigid bodies and their 2D planar workspaces can be divided into free spaces and obstacles. However, recent wheeled mobility research, showing that wheeled platforms have the potential of moving over vertically challenging terrain (e.g., rocky outcroppings, rugged boulders, and fallen tree trunks), invalidate both assumptions. Navigating off-road vehicle chassis with long suspension travel and low tire pressure in places where the boundary between obstacles and free spaces is blurry requires precise 3D modeling of the interaction between the chassis and the terrain, which is complicated by suspension and tire deformation, varying tire-terrain friction, vehicle weight distribution and momentum, etc. In this letter, we present a learning approach to model wheeled mobility, i.e., in terms of vehicle-terrain forward dynamics, and plan feasible, stable, and efficient motion to drive over vertically challenging terrain without rolling over or getting stuck. We present physical experiments on two wheeled robots and show that planning using our learned model can achieve up to 60% improvement in navigation success rate and 46% reduction in unstable chassis roll and pitch angles.
大多数自主导航系统假设轮式机器人为刚体,其二维平面工作空间可分为自由空间和障碍物。然而,最近的轮式移动研究表明,轮式平台具有在垂直挑战性地形上移动的潜力(例如,裸露的岩石、崎岖的巨石和倒下的树干),这两种假设都无效。在障碍物与自由空间边界模糊的地方,对长悬架行程、低胎压的越野车底盘进行导航,需要对底盘与地形的相互作用进行精确的三维建模,而悬架与轮胎的变形、胎地摩擦的变化、车辆的重量分布和动量等因素使底盘与地形的相互作用变得复杂。在这封信中,我们提出了一种学习方法来模拟轮式机动性,即从车辆-地形前向动力学的角度出发,并规划可行、稳定和有效的运动,以在垂直挑战性地形上行驶,而不会翻车或卡住。我们在两个轮式机器人上进行了物理实验,并表明使用我们学习的模型进行规划可以使导航成功率提高60%,减少46%的不稳定底盘横摇和俯仰角。
{"title":"Learning to Model and Plan for Wheeled Mobility on Vertically Challenging Terrain","authors":"Aniket Datar;Chenhui Pan;Xuesu Xiao","doi":"10.1109/LRA.2024.3520919","DOIUrl":"https://doi.org/10.1109/LRA.2024.3520919","url":null,"abstract":"Most autonomous navigation systems assume wheeled robots are rigid bodies and their 2D planar workspaces can be divided into free spaces and obstacles. However, recent wheeled mobility research, showing that wheeled platforms have the potential of moving over vertically challenging terrain (e.g., rocky outcroppings, rugged boulders, and fallen tree trunks), invalidate both assumptions. Navigating off-road vehicle chassis with long suspension travel and low tire pressure in places where the boundary between obstacles and free spaces is blurry requires precise 3D modeling of the interaction between the chassis and the terrain, which is complicated by suspension and tire deformation, varying tire-terrain friction, vehicle weight distribution and momentum, etc. In this letter, we present a learning approach to model wheeled mobility, i.e., in terms of vehicle-terrain forward dynamics, and plan feasible, stable, and efficient motion to drive over vertically challenging terrain without rolling over or getting stuck. We present physical experiments on two wheeled robots and show that planning using our learned model can achieve up to 60% improvement in navigation success rate and 46% reduction in unstable chassis roll and pitch angles.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"10 2","pages":"1505-1512"},"PeriodicalIF":4.6,"publicationDate":"2024-12-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142938312","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Trifocal Tensor and Relative Pose Estimation With Known Vertical Direction 已知垂直方向的三焦张量和相对姿态估计
IF 4.6 2区 计算机科学 Q2 ROBOTICS Pub Date : 2024-12-23 DOI: 10.1109/LRA.2024.3521181
Tao Li;Zhenbao Yu;Banglei Guan;Jianli Han;Weimin Lv;Friedrich Fraundorfer
This work presents two novel solvers for estimating the relative poses among views with known vertical directions. The vertical directions of camera views can be easily obtained using inertial measurement units (IMUs) which have been widely used in autonomous vehicles, mobile phones, and autonomous aerial vehicles (AAVs). Given the known vertical directions, our algorithms only need to solve for two rotation angles and two translation vectors. In this paper, a linear closed-form solution has been described, requiring only four point correspondences in three views. We also propose a minimal solution with three point correspondences using the latest Gröbner basis solver. Since the proposed methods require fewer point correspondences, they can be efficiently applied within the RANSAC framework for outliers removal and pose estimation in visual odometry. The proposed method has been tested on both synthetic data and real-world scenes from KITTI. The experimental results show that the accuracy of the estimated poses is superior to other alternative methods.
本文提出了两种新的求解方法来估计具有已知垂直方向的视图之间的相对姿态。使用惯性测量单元(imu)可以很容易地获得相机视图的垂直方向,惯性测量单元已广泛应用于自动驾驶汽车,手机和自主飞行器(aav)。给定已知的垂直方向,我们的算法只需要求解两个旋转角度和两个平移向量。本文描述了一个线性闭型解,它只需要三个视图中的四个点对应。我们还使用最新的Gröbner基求解器提出了具有三点对应的最小解。由于所提出的方法需要较少的点对应,因此可以在RANSAC框架内有效地应用于视觉里程测量中的异常值去除和姿态估计。该方法已经在KITTI的合成数据和真实场景上进行了测试。实验结果表明,该方法的姿态估计精度优于其他替代方法。
{"title":"Trifocal Tensor and Relative Pose Estimation With Known Vertical Direction","authors":"Tao Li;Zhenbao Yu;Banglei Guan;Jianli Han;Weimin Lv;Friedrich Fraundorfer","doi":"10.1109/LRA.2024.3521181","DOIUrl":"https://doi.org/10.1109/LRA.2024.3521181","url":null,"abstract":"This work presents two novel solvers for estimating the relative poses among views with known vertical directions. The vertical directions of camera views can be easily obtained using inertial measurement units (IMUs) which have been widely used in autonomous vehicles, mobile phones, and autonomous aerial vehicles (AAVs). Given the known vertical directions, our algorithms only need to solve for two rotation angles and two translation vectors. In this paper, a linear closed-form solution has been described, requiring only four point correspondences in three views. We also propose a minimal solution with three point correspondences using the latest Gröbner basis solver. Since the proposed methods require fewer point correspondences, they can be efficiently applied within the RANSAC framework for outliers removal and pose estimation in visual odometry. The proposed method has been tested on both synthetic data and real-world scenes from KITTI. The experimental results show that the accuracy of the estimated poses is superior to other alternative methods.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"10 2","pages":"1305-1312"},"PeriodicalIF":4.6,"publicationDate":"2024-12-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142912515","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On Robust Context-Aware Navigation for Autonomous Ground Vehicles 自主地面车辆鲁棒上下文感知导航研究
IF 4.6 2区 计算机科学 Q2 ROBOTICS Pub Date : 2024-12-23 DOI: 10.1109/LRA.2024.3520920
Paolo Forte;Himanshu Gupta;Henrik Andreasson;Uwe Köckemann;Achim J. Lilienthal
We propose a context-aware navigation framework designed to support the navigation of autonomous ground vehicles, including articulated ones. The proposed framework employs a behavior tree with novel nodes to manage the navigation tasks: planner and controller selections, path planning, path following, and recovery. It incorporates a weather detection system and configurable global path planning and controller strategy selectors implemented as behavior tree action nodes. These components are integrated into a sub-tree that supervises and manages available options and parameters for global planners and control strategies by evaluating map and real-time sensor data. The proposed approach offers three key benefits: overcoming the limitations of single planner strategies in challenging scenarios; ensuring efficient path planning by balancing between optimization and computational effort; and achieving smoother navigation by reducing path curvature and improving drivability. The performance of the proposed framework is analyzed empirically, and compared against state of the art navigation systems with single path planning strategies.
我们提出了一个上下文感知导航框架,旨在支持自主地面车辆的导航,包括铰接式车辆。该框架采用具有新颖节点的行为树来管理导航任务:规划器和控制器的选择、路径规划、路径跟踪和恢复。它包含一个天气检测系统和可配置的全局路径规划和控制器策略选择器,实现为行为树动作节点。这些组件集成到一个子树中,通过评估地图和实时传感器数据,监督和管理全球规划人员和控制策略的可用选项和参数。提出的方法有三个主要优点:克服了单一规划策略在挑战性场景中的局限性;通过平衡优化和计算量,确保有效的路径规划;并通过减少路径曲率和提高驾驶性能来实现更平稳的导航。对所提框架的性能进行了实证分析,并与当前采用单一路径规划策略的导航系统进行了比较。
{"title":"On Robust Context-Aware Navigation for Autonomous Ground Vehicles","authors":"Paolo Forte;Himanshu Gupta;Henrik Andreasson;Uwe Köckemann;Achim J. Lilienthal","doi":"10.1109/LRA.2024.3520920","DOIUrl":"https://doi.org/10.1109/LRA.2024.3520920","url":null,"abstract":"We propose a context-aware navigation framework designed to support the navigation of autonomous ground vehicles, including articulated ones. The proposed framework employs a behavior tree with novel nodes to manage the navigation tasks: planner and controller selections, path planning, path following, and recovery. It incorporates a weather detection system and configurable global path planning and controller strategy selectors implemented as behavior tree action nodes. These components are integrated into a sub-tree that supervises and manages available options and parameters for global planners and control strategies by evaluating map and real-time sensor data. The proposed approach offers three key benefits: overcoming the limitations of single planner strategies in challenging scenarios; ensuring efficient path planning by balancing between optimization and computational effort; and achieving smoother navigation by reducing path curvature and improving drivability. The performance of the proposed framework is analyzed empirically, and compared against state of the art navigation systems with single path planning strategies.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"10 2","pages":"1449-1456"},"PeriodicalIF":4.6,"publicationDate":"2024-12-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10812052","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142918211","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-Task Reinforcement Learning for Quadrotors 四旋翼飞行器的多任务强化学习
IF 4.6 2区 计算机科学 Q2 ROBOTICS Pub Date : 2024-12-23 DOI: 10.1109/LRA.2024.3520894
Jiaxu Xing;Ismail Geles;Yunlong Song;Elie Aljalbout;Davide Scaramuzza
Reinforcement learning (RL) has shown great effectiveness in quadrotor control, enabling specialized policies to develop even human-champion-level performance in single-task scenarios. However, these specialized policies often struggle with novel tasks, requiring a complete retraining of the policy from scratch. To address this limitation, this paper presents a novel multi-task reinforcement learning (MTRL) framework tailored for quadrotor control, leveraging the shared physical dynamics of the platform to enhance sample efficiency and task performance. By employing a multi-critic architecture and shared task encoders, our framework facilitates knowledge transfer across tasks, enabling a single policy to execute diverse maneuvers, including high-speed stabilization, velocity tracking, and autonomous racing. Our experimental results, validated both in simulation and real-world scenarios, demonstrate that our framework outperforms baseline approaches in terms of sample efficiency and overall task performance. Video is available at https://youtu.be/HfK9UT1OVnY.
强化学习(RL)在四旋翼飞行器控制中显示出极大的有效性,使专门的策略能够在单任务场景中开发出甚至是人类冠军级别的性能。然而,这些专门化的政策往往难以应对新的任务,需要从头开始对政策进行彻底的再培训。为了解决这一限制,本文提出了一种针对四旋翼控制量身定制的新型多任务强化学习(MTRL)框架,利用平台的共享物理动力学来提高采样效率和任务性能。通过采用多批评家架构和共享任务编码器,我们的框架促进了跨任务的知识转移,使单个策略能够执行各种机动,包括高速稳定、速度跟踪和自动赛车。我们的实验结果在模拟和现实场景中都得到了验证,表明我们的框架在样本效率和整体任务性能方面优于基线方法。视频可在https://youtu.be/HfK9UT1OVnY上获得。
{"title":"Multi-Task Reinforcement Learning for Quadrotors","authors":"Jiaxu Xing;Ismail Geles;Yunlong Song;Elie Aljalbout;Davide Scaramuzza","doi":"10.1109/LRA.2024.3520894","DOIUrl":"https://doi.org/10.1109/LRA.2024.3520894","url":null,"abstract":"Reinforcement learning (RL) has shown great effectiveness in quadrotor control, enabling specialized policies to develop even human-champion-level performance in single-task scenarios. However, these specialized policies often struggle with novel tasks, requiring a complete retraining of the policy from scratch. To address this limitation, this paper presents a novel multi-task reinforcement learning (MTRL) framework tailored for quadrotor control, leveraging the shared physical dynamics of the platform to enhance sample efficiency and task performance. By employing a multi-critic architecture and shared task encoders, our framework facilitates knowledge transfer across tasks, enabling a single policy to execute diverse maneuvers, including high-speed stabilization, velocity tracking, and autonomous racing. Our experimental results, validated both in simulation and real-world scenarios, demonstrate that our framework outperforms baseline approaches in terms of sample efficiency and overall task performance. Video is available at <uri>https://youtu.be/HfK9UT1OVnY</uri>.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"10 3","pages":"2112-2119"},"PeriodicalIF":4.6,"publicationDate":"2024-12-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142993213","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ConditionNET: Learning Preconditions and Effects for Execution Monitoring 条件网:学习的前提条件和执行监控的效果
IF 4.6 2区 计算机科学 Q2 ROBOTICS Pub Date : 2024-12-23 DOI: 10.1109/LRA.2024.3520916
Daniel Sliwowski;Dongheui Lee
The introduction of robots into everyday scenarios necessitates algorithms capable of monitoring the execution of tasks. In this letter, we propose ConditionNET, an approach for learning the preconditions and effects of actions in a fully data-driven manner. We develop an efficient vision-language model and introduce additional optimization objectives during training to optimize for consistent feature representations. ConditionNET explicitly models the dependencies between actions, preconditions, and effects, leading to improved performance. We evaluate our model on two robotic datasets, one of which we collected for this letter, containing 406 successful and 138 failed teleoperated demonstrations of a Franka Emika Panda robot performing tasks like pouring and cleaning the counter. We show in our experiments that ConditionNET outperforms all baselines on both anomaly detection and phase prediction tasks. Furthermore, we implement an action monitoring system on a real robot to demonstrate the practical applicability of the learned preconditions and effects. Our results highlight the potential of ConditionNET for enhancing the reliability and adaptability of robots in real-world environments.
将机器人引入日常场景需要能够监控任务执行的算法。在这封信中,我们提出了ConditionNET,这是一种以完全数据驱动的方式学习行动的前提条件和效果的方法。我们开发了一个高效的视觉语言模型,并在训练过程中引入了额外的优化目标,以优化一致的特征表示。ConditionNET显式地对操作、前提条件和效果之间的依赖关系进行建模,从而提高性能。我们在两个机器人数据集上评估了我们的模型,其中一个是我们为这封信收集的,包含406个成功和138个失败的远程操作演示,展示了弗兰卡·艾米卡熊猫机器人执行倒水和清洁柜台等任务。我们在实验中表明,在异常检测和相位预测任务上,ConditionNET优于所有基线。此外,我们在一个真实的机器人上实现了一个动作监控系统,以证明所学习的前提条件和效果的实际适用性。我们的研究结果强调了ConditionNET在提高机器人在现实环境中的可靠性和适应性方面的潜力。
{"title":"ConditionNET: Learning Preconditions and Effects for Execution Monitoring","authors":"Daniel Sliwowski;Dongheui Lee","doi":"10.1109/LRA.2024.3520916","DOIUrl":"https://doi.org/10.1109/LRA.2024.3520916","url":null,"abstract":"The introduction of robots into everyday scenarios necessitates algorithms capable of monitoring the execution of tasks. In this letter, we propose ConditionNET, an approach for learning the preconditions and effects of actions in a fully data-driven manner. We develop an efficient vision-language model and introduce additional optimization objectives during training to optimize for consistent feature representations. ConditionNET explicitly models the dependencies between actions, preconditions, and effects, leading to improved performance. We evaluate our model on two robotic datasets, one of which we collected for this letter, containing 406 successful and 138 failed teleoperated demonstrations of a Franka Emika Panda robot performing tasks like pouring and cleaning the counter. We show in our experiments that ConditionNET outperforms all baselines on both anomaly detection and phase prediction tasks. Furthermore, we implement an action monitoring system on a real robot to demonstrate the practical applicability of the learned preconditions and effects. Our results highlight the potential of ConditionNET for enhancing the reliability and adaptability of robots in real-world environments.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"10 2","pages":"1337-1344"},"PeriodicalIF":4.6,"publicationDate":"2024-12-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142912458","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
GARD: A Geometry-Informed and Uncertainty-Aware Baseline Method for Zero-Shot Roadside Monocular Object Detection GARD:一种几何信息和不确定性感知的零射击路边单眼目标检测基线方法
IF 4.6 2区 计算机科学 Q2 ROBOTICS Pub Date : 2024-12-23 DOI: 10.1109/LRA.2024.3520923
Yuru Peng;Beibei Wang;Zijian Yu;Lu Zhang;Jianmin Ji;Yu Zhang;Yanyong Zhang
Roadside camera-based perception methods are in high demand for developing efficient vehicle-infrastructure collaborative perception systems. By focusing on object-level depth prediction, we explore the potential benefits of integrating environmental priors into such systems and propose a geometry-based roadside per-object depth estimation algorithm dubbed GARD. The proposed method capitalizes on the inherent geometric properties of the pinhole camera model to derive depth as well as 3D positions for given 2D targets in roadside-view images, alleviating the need for computationally intensive end-to-end learning architectures for monocular 3D detection. Using only a pre-trained 2D detection model, our approach does not require vast amounts of scene-specific training data and shows superior generalization abilities across varying environments and camera setups, making it a practical and cost-effective solution for monocular 3D object detection.
基于路边摄像头的感知方法对于开发高效的车辆-基础设施协同感知系统具有很高的需求。通过关注目标级深度预测,我们探索了将环境先验整合到此类系统中的潜在好处,并提出了一种基于几何的路边每目标深度估计算法GARD。该方法利用针孔相机模型固有的几何特性,从路边视图图像中获得给定2D目标的深度和3D位置,从而减轻了单眼3D检测对计算密集型端到端学习架构的需求。仅使用预训练的2D检测模型,我们的方法不需要大量特定场景的训练数据,并且在不同的环境和相机设置中显示出卓越的泛化能力,使其成为单目3D物体检测的实用且经济高效的解决方案。
{"title":"GARD: A Geometry-Informed and Uncertainty-Aware Baseline Method for Zero-Shot Roadside Monocular Object Detection","authors":"Yuru Peng;Beibei Wang;Zijian Yu;Lu Zhang;Jianmin Ji;Yu Zhang;Yanyong Zhang","doi":"10.1109/LRA.2024.3520923","DOIUrl":"https://doi.org/10.1109/LRA.2024.3520923","url":null,"abstract":"Roadside camera-based perception methods are in high demand for developing efficient vehicle-infrastructure collaborative perception systems. By focusing on object-level depth prediction, we explore the potential benefits of integrating environmental priors into such systems and propose a geometry-based roadside per-object depth estimation algorithm dubbed GARD. The proposed method capitalizes on the inherent geometric properties of the pinhole camera model to derive depth as well as 3D positions for given 2D targets in roadside-view images, alleviating the need for computationally intensive end-to-end learning architectures for monocular 3D detection. Using only a pre-trained 2D detection model, our approach does not require vast amounts of scene-specific training data and shows superior generalization abilities across varying environments and camera setups, making it a practical and cost-effective solution for monocular 3D object detection.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"10 2","pages":"1297-1304"},"PeriodicalIF":4.6,"publicationDate":"2024-12-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142912470","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Targeted Hard Sample Synthesis Based on Estimated Pose and Occlusion Error for Improved Object Pose Estimation 基于估计姿态和遮挡误差的目标硬样本合成改进的目标姿态估计
IF 4.6 2区 计算机科学 Q2 ROBOTICS Pub Date : 2024-12-20 DOI: 10.1109/LRA.2024.3520436
Alan Li;Angela P. Schoellig
6D Object pose estimation is a fundamental component in robotics enabling efficient interaction with the environment. It is particularly challenging in bin-picking applications, where objects may be textureless and in difficult poses, and occlusion between objects of the same type may cause confusion even in well-trained models. We propose a novel method of hard example synthesis that is model-agnostic, using existing simulators and the modeling of pose error in both the camera-to-object viewsphere and occlusion space. Through evaluation of the model performance with respect to the distribution of object poses and occlusions, we discover regions of high error and generate realistic training samples to specifically target these regions. With our training approach, we demonstrate an improvement in correct detection rate of up to 20% across several ROBI-dataset objects using state-of-the-art pose estimation models.
6D物体姿态估计是机器人技术的一个基本组成部分,可以实现与环境的有效交互。这在垃圾箱拾取应用程序中尤其具有挑战性,其中对象可能是无纹理的,并且处于困难的姿势,并且即使在训练有素的模型中,同一类型对象之间的遮挡也可能导致混乱。我们提出了一种新的硬例综合方法,该方法是模型无关的,使用现有的模拟器和相机到目标视域和遮挡空间的姿态误差建模。通过评估模型在物体姿态和遮挡分布方面的性能,我们发现了高误差的区域,并生成了专门针对这些区域的真实训练样本。通过我们的训练方法,我们展示了使用最先进的姿态估计模型在多个robi数据集对象上的正确检测率提高了20%。
{"title":"Targeted Hard Sample Synthesis Based on Estimated Pose and Occlusion Error for Improved Object Pose Estimation","authors":"Alan Li;Angela P. Schoellig","doi":"10.1109/LRA.2024.3520436","DOIUrl":"https://doi.org/10.1109/LRA.2024.3520436","url":null,"abstract":"6D Object pose estimation is a fundamental component in robotics enabling efficient interaction with the environment. It is particularly challenging in bin-picking applications, where objects may be textureless and in difficult poses, and occlusion between objects of the same type may cause confusion even in well-trained models. We propose a novel method of hard example synthesis that is model-agnostic, using existing simulators and the modeling of pose error in both the camera-to-object viewsphere and occlusion space. Through evaluation of the model performance with respect to the distribution of object poses and occlusions, we discover regions of high error and generate realistic training samples to specifically target these regions. With our training approach, we demonstrate an improvement in correct detection rate of up to 20% across several ROBI-dataset objects using state-of-the-art pose estimation models.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"10 2","pages":"1281-1288"},"PeriodicalIF":4.6,"publicationDate":"2024-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142912366","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Domain Randomization for Learning to Navigate in Human Environments 在人类环境中学习导航的领域随机化
IF 4.6 2区 计算机科学 Q2 ROBOTICS Pub Date : 2024-12-20 DOI: 10.1109/LRA.2024.3521178
Nick Ah Sen;Dana Kulić;Pamela Carreno-Medrano
In shared human-robot environments, effective navigation requires robots to adapt to various pedestrian behaviors encountered in the real world. Most existing deep reinforcement learning algorithms for human-aware robot navigation typically assume that pedestrians adhere to a single walking behavior during training, limiting their practicality/performance in scenarios where pedestrians exhibit various types of behavior. In this work, we propose to enhance the generalization capabilities of human-aware robot navigation by employing Domain Randomization (DR) techniques to train navigation policies on a diverse range of simulated pedestrian behaviors with the hope of better generalization to the real world. We evaluate the effectiveness of our method by comparing the generalization capabilities of a robot navigation policy trained with and without DR, both in simulations and through a real-user study, focusing on adaptability to different pedestrian behaviors, performance in novel environments, and users' perceived comfort, sociability and naturalness. Our findings reveal that the use of DR significantly enhances the robot's social compliance in both simulated and real-life contexts.
在共享的人机环境中,有效的导航需要机器人适应现实世界中遇到的各种行人行为。大多数现有的用于人类感知机器人导航的深度强化学习算法通常假设行人在训练过程中坚持单一的行走行为,这限制了它们在行人表现出各种行为的场景中的实用性/性能。在这项工作中,我们提出通过使用领域随机化(DR)技术来训练基于各种模拟行人行为的导航策略,从而增强人类感知机器人导航的泛化能力,以期更好地泛化到现实世界。我们通过比较在模拟和真实用户研究中使用和不使用DR训练的机器人导航策略的泛化能力来评估我们方法的有效性,重点关注对不同行人行为的适应性,在新环境中的表现,以及用户感知的舒适度,社交性和自然性。我们的研究结果表明,DR的使用显著提高了机器人在模拟和现实环境中的社会依从性。
{"title":"Domain Randomization for Learning to Navigate in Human Environments","authors":"Nick Ah Sen;Dana Kulić;Pamela Carreno-Medrano","doi":"10.1109/LRA.2024.3521178","DOIUrl":"https://doi.org/10.1109/LRA.2024.3521178","url":null,"abstract":"In shared human-robot environments, effective navigation requires robots to adapt to various pedestrian behaviors encountered in the real world. Most existing deep reinforcement learning algorithms for human-aware robot navigation typically assume that pedestrians adhere to a single walking behavior during training, limiting their practicality/performance in scenarios where pedestrians exhibit various types of behavior. In this work, we propose to enhance the generalization capabilities of human-aware robot navigation by employing Domain Randomization (DR) techniques to train navigation policies on a diverse range of simulated pedestrian behaviors with the hope of better generalization to the real world. We evaluate the effectiveness of our method by comparing the generalization capabilities of a robot navigation policy trained with and without DR, both in simulations and through a real-user study, focusing on adaptability to different pedestrian behaviors, performance in novel environments, and users' perceived comfort, sociability and naturalness. Our findings reveal that the use of DR significantly enhances the robot's social compliance in both simulated and real-life contexts.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"10 2","pages":"1625-1632"},"PeriodicalIF":4.6,"publicationDate":"2024-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142937838","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CODE: Complete Coverage AAV Exploration Planner Using Dual-Type Viewpoints for Multi-Layer Complex Environments 代码:完全覆盖AAV勘探计划使用双重视点多层次复杂环境
IF 4.6 2区 计算机科学 Q2 ROBOTICS Pub Date : 2024-12-20 DOI: 10.1109/LRA.2024.3521179
Huazhang Zhu;Tian Lan;Shunzheng Ma;Xuan Zhao;Huiliang Shang;Ruijiao Li
We present an autonomous exploration method for autonomous aerial vehicles (AAVs) for three-dimensional (3D) exploration tasks. Our approach, utilizing a cooperation strategy between common viewpoints and frontier viewpoints, fully leverages the agility and flexibility of AAVs, demonstrating faster and more comprehensive exploration than the current state-of-the-art. Common viewpoints, specifically designed for AAVs exploration, are evenly distributed throughout the 3D space for 3D exploration tasks. Frontier viewpoints are positioned at the centroids of clusters of frontier points to help the AAV maintain motivation to explore unknown complex 3D environments and navigate through narrow corners and passages. This strategy allows the AAV to access every corner of the 3D environment. Additionally, our method includes a refined relocation mechanism for AAVs specifically. Experimental comparisons show that our method ensures complete exploration coverage in environments with complex terrain. Our method outperforms TARE DSVP, GBP and MBP by the coverage rate of 64%, 63%, 54% and 49% respectively in garage-D. In narrow tunnels, ours and DSVP are the only two evaluated methods that achieve complete coverage, with ours outperforming DSVP by 35% in exploration efficiency.
提出了一种用于自主飞行器(aav)三维(3D)探测任务的自主探测方法。我们的方法利用了共同视点和前沿视点之间的合作策略,充分利用了aav的敏捷性和灵活性,展示了比目前最先进的技术更快、更全面的探索。共同视点是专门为aav探索设计的,均匀分布在整个3D空间中,用于3D探索任务。前沿视点位于前沿点集群的质心,以帮助AAV保持探索未知复杂3D环境的动力,并在狭窄的角落和通道中导航。这种策略允许AAV进入3D环境的每个角落。此外,我们的方法还包括针对aav的精确定位机制。实验对比表明,该方法在复杂地形环境下能够保证完整的勘探覆盖。我们的方法在车库d的覆盖率分别为64%、63%、54%和49%,优于TARE DSVP、GBP和MBP。在狭窄的隧道中,我们的和DSVP是仅有的两种实现完全覆盖的评估方法,我们的勘探效率比DSVP高出35%。
{"title":"CODE: Complete Coverage AAV Exploration Planner Using Dual-Type Viewpoints for Multi-Layer Complex Environments","authors":"Huazhang Zhu;Tian Lan;Shunzheng Ma;Xuan Zhao;Huiliang Shang;Ruijiao Li","doi":"10.1109/LRA.2024.3521179","DOIUrl":"https://doi.org/10.1109/LRA.2024.3521179","url":null,"abstract":"We present an autonomous exploration method for autonomous aerial vehicles (AAVs) for three-dimensional (3D) exploration tasks. Our approach, utilizing a cooperation strategy between common viewpoints and frontier viewpoints, fully leverages the agility and flexibility of AAVs, demonstrating faster and more comprehensive exploration than the current state-of-the-art. Common viewpoints, specifically designed for AAVs exploration, are evenly distributed throughout the 3D space for 3D exploration tasks. Frontier viewpoints are positioned at the centroids of clusters of frontier points to help the AAV maintain motivation to explore unknown complex 3D environments and navigate through narrow corners and passages. This strategy allows the AAV to access every corner of the 3D environment. Additionally, our method includes a refined relocation mechanism for AAVs specifically. Experimental comparisons show that our method ensures complete exploration coverage in environments with complex terrain. Our method outperforms TARE DSVP, GBP and MBP by the coverage rate of 64%, 63%, 54% and 49% respectively in garage-D. In narrow tunnels, ours and DSVP are the only two evaluated methods that achieve complete coverage, with ours outperforming DSVP by 35% in exploration efficiency.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"10 2","pages":"1880-1887"},"PeriodicalIF":4.6,"publicationDate":"2024-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142975964","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Robotics and Automation Letters
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1