首页 > 最新文献

IEEE Robotics and Automation Letters最新文献

英文 中文
A Shared-Control Teleoperation System Based on Potential-Field-Constraint Prediction 基于势场约束预测的共享控制遥操作系统
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-12-11 DOI: 10.1109/LRA.2025.3643275
Pengpeng Li;Weihua Li;Zhenwei Lian;Hongjun Xing;Bindi You;Jianfeng Wang;Liang Ding
This letter presents a bilateral shared-control teleoperation system by establishing a human machine environment cooperative control framework to address performance degradation caused by communication delays and cluttered environments. Specifically, we develop a leader-side robot kinematic model with environment-induced artificial potential-field constraints to dynamically fuse leader intent and follower-side environmental information, enabling delay compensation of leader commands. Based on the compensated commands, a bilateral teleoperation controller with local prediction-error compensation term is designed, and an auxiliary potential-field-based shared-control term is incorporated via weighting coefficients to enhance the safety and stability of the wheeled mobile robot (WMR) in complex scenarios. Experimental validation in both simple and challenging environments demonstrates that the proposed method significantly reduces the task failure rate and collision risk, while improving control efficiency and user experience. And operators’ subjective assessments also show significant improvements in lower perceived workload and satisfaction.
本文通过建立人机环境协同控制框架,提出了一种双边共享控制远程操作系统,以解决通信延迟和混乱环境引起的性能下降问题。具体而言,我们建立了一个具有环境诱导的人工势场约束的领导者侧机器人运动学模型,以动态融合领导者意图和随从侧环境信息,实现领导者命令的延迟补偿。在补偿指令的基础上,设计了具有局部预测误差补偿项的双向遥操作控制器,并通过加权系数引入基于势场的辅助共享控制项,提高了轮式移动机器人在复杂场景下的安全性和稳定性。在简单和复杂环境下的实验验证表明,该方法显著降低了任务失败率和碰撞风险,同时提高了控制效率和用户体验。操作员的主观评估也显示出较低的感知工作量和满意度的显着改善。
{"title":"A Shared-Control Teleoperation System Based on Potential-Field-Constraint Prediction","authors":"Pengpeng Li;Weihua Li;Zhenwei Lian;Hongjun Xing;Bindi You;Jianfeng Wang;Liang Ding","doi":"10.1109/LRA.2025.3643275","DOIUrl":"https://doi.org/10.1109/LRA.2025.3643275","url":null,"abstract":"This letter presents a bilateral shared-control teleoperation system by establishing a human machine environment cooperative control framework to address performance degradation caused by communication delays and cluttered environments. Specifically, we develop a leader-side robot kinematic model with environment-induced artificial potential-field constraints to dynamically fuse leader intent and follower-side environmental information, enabling delay compensation of leader commands. Based on the compensated commands, a bilateral teleoperation controller with local prediction-error compensation term is designed, and an auxiliary potential-field-based shared-control term is incorporated via weighting coefficients to enhance the safety and stability of the wheeled mobile robot (WMR) in complex scenarios. Experimental validation in both simple and challenging environments demonstrates that the proposed method significantly reduces the task failure rate and collision risk, while improving control efficiency and user experience. And operators’ subjective assessments also show significant improvements in lower perceived workload and satisfaction.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 2","pages":"1634-1641"},"PeriodicalIF":5.3,"publicationDate":"2025-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145778434","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Intention-Guided Reinforcement Learning Approach With Dirichlet Energy Constraint for Heterogeneous Multi-Robot Cooperation 基于Dirichlet能量约束的意向引导强化学习方法研究异构多机器人合作
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-12-11 DOI: 10.1109/LRA.2025.3643304
Heteng Zhang;Yunjie Jia;Zihao Sun;Yong Song;Bao Pang;Xianfeng Yuan;Rui Song;Simon X. Yang
Multi-robot systems have demonstrated significant potential in accomplishing complex tasks, such as cooperative pursuit, search-and-rescue operations. The emergence of heterogeneous robots with diverse capabilities and characteristics shows superior adaptability compared with homogeneous teams. However, in practical applications, global information is typically inaccessible, and composite teams must contend with partial observability and coordination difficulties. To address the issue in heterogeneous multi-robot systems, we propose a novel Intention-Guided reinforcement learning approach with Dirichlet Energy constraint (IGDE). Specifically, an intention-guided module is designed to derive long-horizon strategies based solely on local observations, enabling foresighted decision-making. In addition, a Dirichlet energy constraint is incorporated into the communication process to enhance the diversity of environmental cognition among different classes of robots. Heterogeneous robots perform class-aware actions driven by distinct cognitive representations, thereby enhancing cooperative efficiency. Notably, our approach alleviates the need of prior knowledge and heterogeneity modeling. Extensive comparative experiments and ablation studies verify the effectiveness of the proposed framework. Additionally, real-world deployment is conducted to demonstrate the practicality.
多机器人系统在完成复杂任务方面已经显示出巨大的潜力,比如协同追击、搜索和救援行动。具有多种能力和特征的异质机器人的出现,比同质团队表现出更强的适应性。然而,在实际应用中,全局信息通常是不可访问的,并且组合团队必须应对部分可观察性和协调困难。为了解决异构多机器人系统中的问题,我们提出了一种新的基于Dirichlet能量约束(IGDE)的意图引导强化学习方法。具体而言,意图引导模块旨在仅根据当地观察得出长期战略,从而实现有远见的决策。此外,在通信过程中引入Dirichlet能量约束,增强了不同类别机器人之间环境认知的多样性。异构机器人在不同的认知表征驱动下进行类别感知行为,从而提高合作效率。值得注意的是,我们的方法减轻了对先验知识和异质性建模的需要。大量的对比实验和烧蚀研究验证了所提出框架的有效性。此外,还进行了实际部署以演示其实用性。
{"title":"An Intention-Guided Reinforcement Learning Approach With Dirichlet Energy Constraint for Heterogeneous Multi-Robot Cooperation","authors":"Heteng Zhang;Yunjie Jia;Zihao Sun;Yong Song;Bao Pang;Xianfeng Yuan;Rui Song;Simon X. Yang","doi":"10.1109/LRA.2025.3643304","DOIUrl":"https://doi.org/10.1109/LRA.2025.3643304","url":null,"abstract":"Multi-robot systems have demonstrated significant potential in accomplishing complex tasks, such as cooperative pursuit, search-and-rescue operations. The emergence of heterogeneous robots with diverse capabilities and characteristics shows superior adaptability compared with homogeneous teams. However, in practical applications, global information is typically inaccessible, and composite teams must contend with partial observability and coordination difficulties. To address the issue in heterogeneous multi-robot systems, we propose a novel <italic>I</i>ntention-<italic>G</i>uided reinforcement learning approach with <italic>D</i>irichlet <italic>E</i>nergy constraint (IGDE). Specifically, an intention-guided module is designed to derive long-horizon strategies based solely on local observations, enabling foresighted decision-making. In addition, a Dirichlet energy constraint is incorporated into the communication process to enhance the diversity of environmental cognition among different classes of robots. Heterogeneous robots perform class-aware actions driven by distinct cognitive representations, thereby enhancing cooperative efficiency. Notably, our approach alleviates the need of prior knowledge and heterogeneity modeling. Extensive comparative experiments and ablation studies verify the effectiveness of the proposed framework. Additionally, real-world deployment is conducted to demonstrate the practicality.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 2","pages":"1450-1457"},"PeriodicalIF":5.3,"publicationDate":"2025-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145778271","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Gentle Object Retraction in Dense Clutter Using Multimodal Force Sensing and Imitation Learning 基于多模态力传感和模仿学习的密集杂波中温和物体收缩
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-12-11 DOI: 10.1109/LRA.2025.3643332
Dane Brouwer;Joshua Citron;Heather Nolte;Jeannette Bohg;Mark Cutkosky
Dense collections of movable objects are common in everyday spaces—from cabinets in a home to shelves in a warehouse. Safely retracting objects from such collections is difficult for robots, yet people do it frequently, leveraging learned experience in tandem with vision and non-prehensile tactile sensing on the sides and backs of their hands and arms. We investigate the role of contact force sensing for training robots to gently reach into constrained clutter and extract objects. The available sensing modalities are 1) “eye-in-hand” vision, 2) proprioception, 3) non-prehensile triaxial tactile sensing, 4) contact wrenches estimated from joint torques, and 5) a measure of object acquisition obtained by monitoring the vacuum line of a suction cup. We use imitation learning to train policies from a set of demonstrations on randomly generated scenes, then conduct an ablation study of wrench and tactile information. We evaluate each policy’s performance across 40 unseen environment configurations. Policies employing any force sensing show fewer excessive force failures, an increased overall success rate, and faster completion times. The best performance is achieved using both tactile and wrench information, producing an 80% improvement above the baseline without force information.
从家里的橱柜到仓库的架子,密集的可移动物品在日常空间中很常见。对机器人来说,安全地从这样的集合中缩回物体是很困难的,但人们经常这样做,利用所学的经验,结合视觉和手和手臂两侧和背部的不可抓握的触觉。我们研究了接触力传感在训练机器人轻柔地进入受限杂波并提取物体方面的作用。可用的传感方式有:1)“手眼”视觉,2)本体感觉,3)不可卷曲的三轴触觉,4)根据关节扭矩估计的接触扳手,以及5)通过监测吸盘的真空线获得的物体获取测量。我们使用模仿学习从随机生成的场景的一组演示中训练策略,然后对扳手和触觉信息进行消融研究。我们在40个不可见的环境配置中评估每个策略的性能。采用任何力传感的策略可以减少过度的力故障,提高总体成功率,并缩短完井时间。使用触觉和扳手信息实现最佳性能,在没有力信息的情况下比基线提高80%。
{"title":"Gentle Object Retraction in Dense Clutter Using Multimodal Force Sensing and Imitation Learning","authors":"Dane Brouwer;Joshua Citron;Heather Nolte;Jeannette Bohg;Mark Cutkosky","doi":"10.1109/LRA.2025.3643332","DOIUrl":"https://doi.org/10.1109/LRA.2025.3643332","url":null,"abstract":"Dense collections of movable objects are common in everyday spaces—from cabinets in a home to shelves in a warehouse. Safely retracting objects from such collections is difficult for robots, yet people do it frequently, leveraging learned experience in tandem with vision and non-prehensile tactile sensing on the sides and backs of their hands and arms. We investigate the role of contact force sensing for training robots to gently reach into constrained clutter and extract objects. The available sensing modalities are 1) “eye-in-hand” vision, 2) proprioception, 3) non-prehensile triaxial tactile sensing, 4) contact wrenches estimated from joint torques, and 5) a measure of object acquisition obtained by monitoring the vacuum line of a suction cup. We use imitation learning to train policies from a set of demonstrations on randomly generated scenes, then conduct an ablation study of wrench and tactile information. We evaluate each policy’s performance across 40 unseen environment configurations. Policies employing any force sensing show fewer excessive force failures, an increased overall success rate, and faster completion times. The best performance is achieved using both tactile and wrench information, producing an 80% improvement above the baseline without force information.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 2","pages":"1578-1585"},"PeriodicalIF":5.3,"publicationDate":"2025-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145778325","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ColorMap-VIO: A Drift-Free Visual-Inertial Odometry in a Prior Colored Point Cloud Map ColorMap-VIO:一种基于先验彩色点云图的无漂移视觉惯性里程计
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-12-11 DOI: 10.1109/LRA.2025.3643289
Jie Xu;Xuanxuan Zhang;Yongxin Ma;Yixuan Li;Linji Wang;Xinhang Xu;Shenghai Yuan;Lihua Xie
Visual-inertial odometry (VIO) can estimate robot poses at high frequencies but suffers from accumulated drift over time. Incorporating point cloud maps offers a promising solution, yet existing registration methods between vision and point clouds are limited by heterogeneous feature alignment, leaving much information underutilized and resulting in reduced accuracy, poor robustness, and high computational cost. To address these challenges, this paper proposes a visual-inertial localization system based on color point cloud maps, consisting of two main components: color map construction and visual-inertial tracking. A gradient-based map sparsification strategy is employed during map construction to preserve salient features while reducing storage and computation. For localization, we propose an image pyramid-based visual photometric IESKF, which fuses IMU and photometric observations to estimate precise poses. Gradient-rich feature points are projected onto image pyramids across multiple resolutions to perform iterative updates, effectively avoiding local minima and improving accuracy. Experimental results show that our method achieves stable and accurate localization bounded by map precision, and demonstrates higher efficiency and robustness than existing map-based approaches.
视觉惯性里程计(VIO)可以在高频率下估计机器人的姿态,但随着时间的推移会受到累积漂移的影响。结合点云图提供了一个很好的解决方案,但现有的视觉和点云之间的配准方法受到异构特征对齐的限制,导致大量信息未得到充分利用,导致精度降低,鲁棒性差,计算成本高。为了解决这些问题,本文提出了一种基于彩色点云图的视觉惯性定位系统,该系统主要由彩色地图构建和视觉惯性跟踪两部分组成。在地图构建过程中采用基于梯度的地图稀疏化策略,在保留显著特征的同时减少存储和计算量。为了定位,我们提出了一种基于图像金字塔的视觉光度IESKF,它融合了IMU和光度观测来估计精确的姿态。将梯度丰富的特征点投影到多个分辨率的图像金字塔上进行迭代更新,有效地避免了局部极小值,提高了精度。实验结果表明,该方法在地图精度约束下实现了稳定、准确的定位,比现有的基于地图的定位方法具有更高的效率和鲁棒性。
{"title":"ColorMap-VIO: A Drift-Free Visual-Inertial Odometry in a Prior Colored Point Cloud Map","authors":"Jie Xu;Xuanxuan Zhang;Yongxin Ma;Yixuan Li;Linji Wang;Xinhang Xu;Shenghai Yuan;Lihua Xie","doi":"10.1109/LRA.2025.3643289","DOIUrl":"https://doi.org/10.1109/LRA.2025.3643289","url":null,"abstract":"Visual-inertial odometry (VIO) can estimate robot poses at high frequencies but suffers from accumulated drift over time. Incorporating point cloud maps offers a promising solution, yet existing registration methods between vision and point clouds are limited by heterogeneous feature alignment, leaving much information underutilized and resulting in reduced accuracy, poor robustness, and high computational cost. To address these challenges, this paper proposes a visual-inertial localization system based on color point cloud maps, consisting of two main components: color map construction and visual-inertial tracking. A gradient-based map sparsification strategy is employed during map construction to preserve salient features while reducing storage and computation. For localization, we propose an image pyramid-based visual photometric IESKF, which fuses IMU and photometric observations to estimate precise poses. Gradient-rich feature points are projected onto image pyramids across multiple resolutions to perform iterative updates, effectively avoiding local minima and improving accuracy. Experimental results show that our method achieves stable and accurate localization bounded by map precision, and demonstrates higher efficiency and robustness than existing map-based approaches.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 2","pages":"1570-1577"},"PeriodicalIF":5.3,"publicationDate":"2025-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145778326","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Development of a Novel 7-DOF Position-Orientation Decoupled Microsurgical Robot with Motorized Instruments for Microvascular Anastomosis 新型7-DOF位姿解耦微型血管吻合手术机器人的研制
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-12-11 DOI: 10.1109/LRA.2025.3643333
Dunfa Long;Shaoan Chen;Shuai Ao;Zhiqiang Zhang;Chengzhi Hu;Chaoyang Shi
This work introduces a novel compact 7-degree-of-freedom (7-DOF) microsurgical robot with position-orientation decoupling capacity for microvascular anastomosis. The proposed system employs a modular architecture combining a proximal displacement platform for 3D small-stroke translation and a distal compact remote center of motion (RCM) mechanism for wide-range orientation adjustment. This design meets the workspace requirements for microvascular anastomosis, requiring extensive orientation adjustments with minimal positional movement and reducing the system footprint. The parasitic motion reverse self-compensation method has been developed for motorized surgical instruments, effectively reducing operational resistance to improve precision. Theoretical analysis has been performed on both the RCM mechanism and motorized surgical instruments, and kinematics-based parameter optimization and data-driven calibration have been conducted to enhance superior performance. A prototype has been constructed, and its experimental validation demonstrated that the system achieved repeatability of 11.24 ± 2.31 μm (XY) and 12.46 ± 4.48 μm (YZ), and absolute positioning accuracy of 29.80 ± 12.27 μm (XY) and 37.02 ± 19.47 μm (YZ), meeting super-microsurgical requirements. Experiments that include needle-threading and stamen peeling tasks demonstrate the robot's superior dexterity and manipulation capabilities.
本文介绍了一种新型紧凑的7自由度显微手术机器人,具有微血管吻合的位置-方向解耦能力。该系统采用模块化结构,结合近端位移平台用于3D小行程平移和远端紧凑远程运动中心(RCM)机构用于大范围方向调整。该设计满足微血管吻合的工作空间要求,需要以最小的位置移动进行广泛的方向调整,并减少系统占地面积。针对电动手术器械,提出了一种寄生运动反向自补偿方法,有效地降低了手术阻力,提高了手术精度。对RCM机构和电动手术器械进行了理论分析,并进行了基于运动学的参数优化和数据驱动的校准,以提高其优越的性能。实验验证表明,该系统的重复性为11.24±2.31 μm (XY)和12.46±4.48 μm (YZ),绝对定位精度为29.80±12.27 μm (XY)和37.02±19.47 μm (YZ),满足超显微外科手术要求。包括穿针引线和剥花蕊任务在内的实验证明了机器人的卓越灵活性和操作能力。
{"title":"Development of a Novel 7-DOF Position-Orientation Decoupled Microsurgical Robot with Motorized Instruments for Microvascular Anastomosis","authors":"Dunfa Long;Shaoan Chen;Shuai Ao;Zhiqiang Zhang;Chengzhi Hu;Chaoyang Shi","doi":"10.1109/LRA.2025.3643333","DOIUrl":"https://doi.org/10.1109/LRA.2025.3643333","url":null,"abstract":"This work introduces a novel compact 7-degree-of-freedom (7-DOF) microsurgical robot with position-orientation decoupling capacity for microvascular anastomosis. The proposed system employs a modular architecture combining a proximal displacement platform for 3D small-stroke translation and a distal compact remote center of motion (RCM) mechanism for wide-range orientation adjustment. This design meets the workspace requirements for microvascular anastomosis, requiring extensive orientation adjustments with minimal positional movement and reducing the system footprint. The parasitic motion reverse self-compensation method has been developed for motorized surgical instruments, effectively reducing operational resistance to improve precision. Theoretical analysis has been performed on both the RCM mechanism and motorized surgical instruments, and kinematics-based parameter optimization and data-driven calibration have been conducted to enhance superior performance. A prototype has been constructed, and its experimental validation demonstrated that the system achieved repeatability of 11.24 ± 2.31 μm (XY) and 12.46 ± 4.48 μm (YZ), and absolute positioning accuracy of 29.80 ± 12.27 μm (XY) and 37.02 ± 19.47 μm (YZ), meeting super-microsurgical requirements. Experiments that include needle-threading and stamen peeling tasks demonstrate the robot's superior dexterity and manipulation capabilities.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 2","pages":"1866-1873"},"PeriodicalIF":5.3,"publicationDate":"2025-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145830774","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Self-Supervised Learning for Object Pose Estimation Through Active Real Sample Capture 基于主动真实样本捕获的目标姿态估计的自监督学习
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-12-11 DOI: 10.1109/LRA.2025.3643269
Alan Li;Angela P. Schoellig
6D Object pose estimation is a fundamental component in robotics enabling efficient interaction with the environment. In industrial bin-picking tasks, this problem becomes especially challenging due to difficult object poses, complex occlusions, and inter-object ambiguities. In this work, we propose a novel self-supervised method that automatically collects, labels, and fine-tunes on real images using an eye-in-hand camera setup. We leverage the mobile camera to first obtain reliable ground-truth estimates through multi-view pose estimation, allowing us to subsequently reposition the camera to capture and label real ‘hard case’ samples from the estimated scene. This process enables closure of the sim-to-real gap through large quantities of targeted real training data, generated by comparing differences in model performance between real and synthetically reconstructed scenes and informing the mobile camera on specific poses or areas for data capture. We surpass state-of-the-art performance on a challenging bin-picking benchmark: five out of seven objects surpass a 95% correct detection rate, compared to only one out of seven for previous methods.
6D物体姿态估计是机器人技术的一个基本组成部分,可以实现与环境的有效交互。在工业拾取垃圾箱的任务中,由于困难的物体姿势,复杂的遮挡和物体间的模糊性,这个问题变得特别具有挑战性。在这项工作中,我们提出了一种新颖的自监督方法,该方法使用眼手相机设置自动收集,标记和微调真实图像。我们利用移动相机首先通过多视图姿态估计获得可靠的地面真值估计,允许我们随后重新定位相机以捕获和标记来自估计场景的真实“硬案例”样本。该过程通过比较真实场景和合成重建场景之间模型性能的差异,并告知移动相机特定姿势或区域进行数据捕获,从而产生大量有针对性的真实训练数据,从而缩小模拟与真实之间的差距。在一项具有挑战性的捡垃圾桶基准测试中,我们的表现超过了最先进的水平:7个物体中有5个物体的正确率超过了95%,而之前的方法只有1 / 7。
{"title":"Self-Supervised Learning for Object Pose Estimation Through Active Real Sample Capture","authors":"Alan Li;Angela P. Schoellig","doi":"10.1109/LRA.2025.3643269","DOIUrl":"https://doi.org/10.1109/LRA.2025.3643269","url":null,"abstract":"6D Object pose estimation is a fundamental component in robotics enabling efficient interaction with the environment. In industrial bin-picking tasks, this problem becomes especially challenging due to difficult object poses, complex occlusions, and inter-object ambiguities. In this work, we propose a novel self-supervised method that automatically collects, labels, and fine-tunes on real images using an eye-in-hand camera setup. We leverage the mobile camera to first obtain reliable ground-truth estimates through multi-view pose estimation, allowing us to subsequently reposition the camera to capture and label real ‘hard case’ samples from the estimated scene. This process enables closure of the sim-to-real gap through large quantities of targeted real training data, generated by comparing differences in model performance between real and synthetically reconstructed scenes and informing the mobile camera on specific poses or areas for data capture. We surpass state-of-the-art performance on a challenging bin-picking benchmark: five out of seven objects surpass a 95% correct detection rate, compared to only one out of seven for previous methods.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 2","pages":"1954-1961"},"PeriodicalIF":5.3,"publicationDate":"2025-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145830867","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Cycle Clustering: An Algorithm for Multi-Depot Multi-Agent Collaborative Coverage in Structured Road Network 周期聚类:结构化路网中多站点多智能体协同覆盖算法
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-12-11 DOI: 10.1109/LRA.2025.3643277
Ruochen Li;Junkai Jiang;Jiongqi Wang;Shaobing Xu;Jianqiang Wang
Multi-depot multi-agent collaborative coverage is a representative problem in swarm intelligence, with broad applications in real-world scenarios. In this problem, multiple agents are initially located at different depots, which differs from the traditional problem setting, and are required to collaboratively cover a given region represented by a structured road network. The objective is to minimize the longest individual route among all agents. This problem is closely related to the $k$-Chinese Postman Problem ($k$-CPP), but is more complex due to the different depot constraint. This letter proposes a novel centralized algorithm, Cycle Clustering (CC), to solve this problem efficiently. The proposed method first transforms the original graph into an Eulerian graph, and then partitions the Eulerian graph into multiple small cycles, which are subsequently clustered and assigned to different agents. This design significantly reduces unnecessary route overhead. The algorithm’s time complexity and completeness are theoretically analyzed. Experimental results demonstrate that the proposed algorithm performs better than existing methods, reducing the average gap to the theoretical lower bound from 29.10% to 10.29%.
多仓库多智能体协同覆盖是群智能中的代表性问题,在现实场景中有着广泛的应用。在该问题中,不同于传统的问题设置,多个智能体最初位于不同的仓库,并且需要协作覆盖由结构化道路网络表示的给定区域。目标是最小化所有agent中最长的单个路径。这个问题与$k$-中国邮差问题($k$-CPP)密切相关,但由于不同的仓库约束而更加复杂。本文提出了一种新的集中式算法——循环聚类算法(CC)来有效地解决这一问题。该方法首先将原始图转换为欧拉图,然后将欧拉图划分为多个小循环,然后将这些小循环聚类并分配给不同的智能体。这种设计大大减少了不必要的路由开销。从理论上分析了该算法的时间复杂度和完备性。实验结果表明,该算法比现有方法性能更好,将理论下界的平均差距从29.10%减小到10.29%。
{"title":"Cycle Clustering: An Algorithm for Multi-Depot Multi-Agent Collaborative Coverage in Structured Road Network","authors":"Ruochen Li;Junkai Jiang;Jiongqi Wang;Shaobing Xu;Jianqiang Wang","doi":"10.1109/LRA.2025.3643277","DOIUrl":"https://doi.org/10.1109/LRA.2025.3643277","url":null,"abstract":"Multi-depot multi-agent collaborative coverage is a representative problem in swarm intelligence, with broad applications in real-world scenarios. In this problem, multiple agents are initially located at different depots, which differs from the traditional problem setting, and are required to collaboratively cover a given region represented by a structured road network. The objective is to minimize the longest individual route among all agents. This problem is closely related to the <inline-formula><tex-math>$k$</tex-math></inline-formula>-Chinese Postman Problem (<inline-formula><tex-math>$k$</tex-math></inline-formula>-CPP), but is more complex due to the different depot constraint. This letter proposes a novel centralized algorithm, Cycle Clustering (CC), to solve this problem efficiently. The proposed method first transforms the original graph into an Eulerian graph, and then partitions the Eulerian graph into multiple small cycles, which are subsequently clustered and assigned to different agents. This design significantly reduces unnecessary route overhead. The algorithm’s time complexity and completeness are theoretically analyzed. Experimental results demonstrate that the proposed algorithm performs better than existing methods, reducing the average gap to the theoretical lower bound from 29.10% to 10.29%.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 2","pages":"1554-1561"},"PeriodicalIF":5.3,"publicationDate":"2025-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145778424","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Incremental Generalized Hybrid A* 增量广义混合A*
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-12-11 DOI: 10.1109/LRA.2025.3643271
Sidharth Talia;Oren Salzman;Siddhartha Srinivasa
We address the problem of efficiently organizing search over very large trees, which arises in many applications ranging from autonomous driving to aerial vehicles. Here, we are motivated by off-road autonomy, where real-time planning is essential. Classical approaches use graphs of motion primitives and exploit dominance to mitigate the curse of dimensionality and prune expansions efficiently. However, for complex dynamics, repeatedly solving two-point boundary-value problems makes graph construction too slow for fast kinodynamic planning. Hybrid A* (HA*) addressed this challenge by searching over a tree of motion primitives and introducing approximate pruning using a grid-based dominance check. However, choosing the grid resolution is difficult: too coarse risks failure, while too fine leads to excessive expansions and slow planning. We propose Incremental Generalized Hybrid A* (IGHA*), an anytime tree-search framework that dynamically organizes vertex expansions without rigid pruning. IGHA* provably matches or outperforms HA*. For both on-road kinematic and off-road kinodynamic planning queries for a car-like robot, variants of IGHA* use $6times$ fewer expansions to the best solution compared to an optimized version of HA* (HA*M, an internal baseline). In simulated off-road experiments in a high-fidelity simulator, IGHA* outperforms HA*M when both are used in the loop with a model predictive controller. We demonstrate real-time performance both in simulation and on a small-scale off-road vehicle, enabling fast, robust planning under complex dynamics.
我们解决了在非常大的树上有效组织搜索的问题,这个问题出现在从自动驾驶到飞行器的许多应用中。在这里,我们的动力来自于越野自动驾驶,实时规划是必不可少的。经典方法使用运动原语图并利用优势来有效地减轻维数和剪枝展开的诅咒。然而,对于复杂的动力学,反复求解两点边值问题会使图的构建过于缓慢,不利于快速的动力学规划。Hybrid A* (HA*)通过搜索运动原语树并使用基于网格的优势检查引入近似修剪来解决这一挑战。然而,选择网格分辨率是困难的:太粗有失败的风险,而太细会导致过度扩张和缓慢的规划。我们提出了增量广义混合A* (IGHA*),这是一个随时树搜索框架,可以动态组织顶点扩展而不需要刚性修剪。IGHA*可以证明与HA*相匹配或优于HA*。对于类车机器人的道路运动学和越野运动学规划查询,IGHA*的变体与HA* (HA*M,内部基线)的优化版本相比,对最佳解决方案的扩展减少了6倍。在高保真模拟器的模拟越野实验中,当IGHA*与模型预测控制器一起用于环路时,IGHA*优于HA*M。我们在模拟和小型越野车辆上展示了实时性能,在复杂动态下实现快速,稳健的规划。
{"title":"Incremental Generalized Hybrid A*","authors":"Sidharth Talia;Oren Salzman;Siddhartha Srinivasa","doi":"10.1109/LRA.2025.3643271","DOIUrl":"https://doi.org/10.1109/LRA.2025.3643271","url":null,"abstract":"We address the problem of efficiently organizing search over very large trees, which arises in many applications ranging from autonomous driving to aerial vehicles. Here, we are motivated by off-road autonomy, where real-time planning is essential. Classical approaches use graphs of motion primitives and exploit dominance to mitigate the curse of dimensionality and prune expansions efficiently. However, for complex dynamics, repeatedly solving two-point boundary-value problems makes graph construction too slow for fast kinodynamic planning. Hybrid A* (<monospace>HA*</monospace>) addressed this challenge by searching over a tree of motion primitives and introducing approximate pruning using a grid-based dominance check. However, choosing the grid resolution is difficult: too coarse risks failure, while too fine leads to excessive expansions and slow planning. We propose Incremental Generalized Hybrid A* (<monospace>IGHA*</monospace>), an anytime tree-search framework that dynamically organizes vertex expansions without rigid pruning. <monospace>IGHA*</monospace> provably matches or outperforms <monospace>HA*</monospace>. For both on-road kinematic and off-road kinodynamic planning queries for a car-like robot, variants of <monospace>IGHA*</monospace> use <inline-formula><tex-math>$6times$</tex-math></inline-formula> fewer expansions to the best solution compared to an optimized version of <monospace>HA*</monospace> (<monospace>HA*M</monospace>, an internal baseline). In simulated off-road experiments in a high-fidelity simulator, <monospace>IGHA*</monospace> outperforms <monospace>HA*M</monospace> when both are used in the loop with a model predictive controller. We demonstrate real-time performance both in simulation and on a small-scale off-road vehicle, enabling fast, robust planning under complex dynamics.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 2","pages":"1586-1593"},"PeriodicalIF":5.3,"publicationDate":"2025-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145778441","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MSG-Loc: Multi-Label Likelihood-Based Semantic Graph Matching for Object-Level Global Localization 基于多标签似然的对象级全局定位语义图匹配
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-12-11 DOI: 10.1109/LRA.2025.3643293
Gihyeon Lee;Jungwoo Lee;Juwon Kim;Young-Sik Shin;Younggun Cho
Robots are often required to localize in environments with unknown object classes and semantic ambiguity. However, when performing global localization using semantic objects, high semantic ambiguity intensifies object misclassification and increases the likelihood of incorrect associations, which in turn can cause significant errors in the estimated pose. Thus, in this letter, we propose a multi-label likelihood-based semantic graph matching framework for object-level global localization. The key idea is to exploit multi-label graph representations, rather than single-label alternatives, to capture and leverage the inherent semantic context of object observations. Based on these representations, our approach enhances semantic correspondence across graphs by combining the likelihood of each node with the maximum likelihood of its neighbors via context-aware likelihood propagation. For rigorous validation, data association and pose estimation performance are evaluated under both closed-set and open-set detection configurations. In addition, we demonstrate the scalability of our approach to large-vocabulary object categories in both real-world indoor scenes and synthetic environments.
机器人经常需要在具有未知对象类和语义模糊的环境中进行定位。然而,当使用语义对象进行全局定位时,高度的语义模糊加剧了对象的错误分类,增加了错误关联的可能性,这反过来又会导致估计姿态的严重错误。因此,在这封信中,我们提出了一个用于对象级全局定位的基于多标签似然的语义图匹配框架。关键思想是利用多标签图表示,而不是单标签替代,来捕获和利用对象观察的固有语义上下文。基于这些表示,我们的方法通过上下文感知似然传播将每个节点的似然与其相邻节点的最大似然相结合,从而增强图之间的语义对应性。为了严格验证,在闭集和开集检测配置下对数据关联和姿态估计性能进行了评估。此外,我们还演示了我们的方法在现实世界室内场景和合成环境中对大词汇量对象类别的可扩展性。
{"title":"MSG-Loc: Multi-Label Likelihood-Based Semantic Graph Matching for Object-Level Global Localization","authors":"Gihyeon Lee;Jungwoo Lee;Juwon Kim;Young-Sik Shin;Younggun Cho","doi":"10.1109/LRA.2025.3643293","DOIUrl":"https://doi.org/10.1109/LRA.2025.3643293","url":null,"abstract":"Robots are often required to localize in environments with unknown object classes and semantic ambiguity. However, when performing global localization using semantic objects, high semantic ambiguity intensifies object misclassification and increases the likelihood of incorrect associations, which in turn can cause significant errors in the estimated pose. Thus, in this letter, we propose a multi-label likelihood-based semantic graph matching framework for object-level global localization. The key idea is to exploit multi-label graph representations, rather than single-label alternatives, to capture and leverage the inherent semantic context of object observations. Based on these representations, our approach enhances semantic correspondence across graphs by combining the likelihood of each node with the maximum likelihood of its neighbors via context-aware likelihood propagation. For rigorous validation, data association and pose estimation performance are evaluated under both closed-set and open-set detection configurations. In addition, we demonstrate the scalability of our approach to large-vocabulary object categories in both real-world indoor scenes and synthetic environments.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 2","pages":"2066-2073"},"PeriodicalIF":5.3,"publicationDate":"2025-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145886711","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Universal Framework for Extrinsic Calibration of Camera, Radar, and LiDAR 相机、雷达和激光雷达外部标定的通用框架
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2025-12-11 DOI: 10.1109/LRA.2025.3643292
Sijie Hu;Alessandro Goldwurm;Martin Mujica;Sylvain Cadou;Frédéric Lerasle
Accurate extrinsic calibration of camera, radar, and LiDAR is critical for multi-modal sensor fusion in autonomous vehicles and mobile robots. Existing methods typically perform pair-wise calibration and rely on specialized targets, limiting scalability and flexibility. We introduce a universal calibration framework based on an Iterative Best Match (IBM) algorithm that refines alignment by optimizing correspondences between sensors, eliminating traditional point-to-point matching. IBM naturally extends to simultaneous camera–LiDAR–radar calibration and leverages tracked natural targets (e.g., pedestrians) to establish cross-modal correspondences without predefined calibration markers. Experiments on a realistic multi-sensor platform (fisheye-camera, LiDAR, and radar) and the KITTI dataset validate the accuracy, robustness, and efficiency of our method.
相机、雷达和激光雷达的精确外部标定对于自动驾驶汽车和移动机器人中的多模态传感器融合至关重要。现有方法通常执行成对校准,依赖于专门的目标,限制了可扩展性和灵活性。我们介绍了一种基于迭代最佳匹配(IBM)算法的通用校准框架,该算法通过优化传感器之间的对应关系来改进校准,消除了传统的点对点匹配。IBM很自然地扩展到摄像头-激光雷达-雷达同步校准,并利用跟踪的自然目标(例如,行人)来建立跨模态对应,而无需预定义的校准标记。在真实的多传感器平台(鱼眼相机、激光雷达和雷达)和KITTI数据集上进行的实验验证了我们的方法的准确性、鲁棒性和效率。
{"title":"A Universal Framework for Extrinsic Calibration of Camera, Radar, and LiDAR","authors":"Sijie Hu;Alessandro Goldwurm;Martin Mujica;Sylvain Cadou;Frédéric Lerasle","doi":"10.1109/LRA.2025.3643292","DOIUrl":"https://doi.org/10.1109/LRA.2025.3643292","url":null,"abstract":"Accurate extrinsic calibration of camera, radar, and LiDAR is critical for multi-modal sensor fusion in autonomous vehicles and mobile robots. Existing methods typically perform pair-wise calibration and rely on specialized targets, limiting scalability and flexibility. We introduce a universal calibration framework based on an Iterative Best Match (IBM) algorithm that refines alignment by optimizing correspondences between sensors, eliminating traditional point-to-point matching. IBM naturally extends to simultaneous camera–LiDAR–radar calibration and leverages tracked natural targets (e.g., pedestrians) to establish cross-modal correspondences without predefined calibration markers. Experiments on a realistic multi-sensor platform (fisheye-camera, LiDAR, and radar) and the KITTI dataset validate the accuracy, robustness, and efficiency of our method.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 2","pages":"1842-1849"},"PeriodicalIF":5.3,"publicationDate":"2025-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145830919","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Robotics and Automation Letters
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1