首页 > 最新文献

IEEE Robotics and Automation Letters最新文献

英文 中文
A Kinesthetic Teaching Framework for Tasks With Contact Transitions and Time-Optimized Execution 具有接触过渡和时间优化执行任务的动觉教学框架
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-02-09 DOI: 10.1109/LRA.2026.3662532
Nikolas Thelenberg;Christian Ott
In kinesthetic teaching, a robot is manually guided by a human operator to demonstrate a task. Most methods focus on replaying the recorded motion, but are agnostic to contact transitions, which can be critical when interacting with rigid environments. To overcome this limitation, we propose a framework that allows to teach motions in free space as well as in contact while preventing fast unintended contact transitions. This is accomplished by exploiting a projection-based unilateral damping force that increases close to contact. We derive an explicit analytical expression for the damping characteristics to ensure a safe stop before the contact when no further forces act on the robot. Furthermore, after the teaching, the recorded motion data is utilized to generate a time-optimized trajectory based on convex optimization, in which the contact transitions are explicitly considered. We validated our framework in experiments with a torque-controlled manipulator.
在动觉教学中,机器人由人类操作员手动引导来演示任务。大多数方法侧重于重放记录的运动,但对接触过渡不可知,这在与刚性环境交互时是至关重要的。为了克服这一限制,我们提出了一个框架,允许在自由空间和接触中教授运动,同时防止快速的意外接触过渡。这是通过利用基于投影的单边阻尼力来实现的,该力在接近接触时增加。我们导出了一个明确的解析表达式,以确保在没有外力作用的情况下,机器人在接触前安全停止。此外,在教学结束后,利用记录的运动数据生成基于凸优化的时间优化轨迹,其中明确考虑了接触转移。我们在力矩控制机械手的实验中验证了我们的框架。
{"title":"A Kinesthetic Teaching Framework for Tasks With Contact Transitions and Time-Optimized Execution","authors":"Nikolas Thelenberg;Christian Ott","doi":"10.1109/LRA.2026.3662532","DOIUrl":"https://doi.org/10.1109/LRA.2026.3662532","url":null,"abstract":"In kinesthetic teaching, a robot is manually guided by a human operator to demonstrate a task. Most methods focus on replaying the recorded motion, but are agnostic to contact transitions, which can be critical when interacting with rigid environments. To overcome this limitation, we propose a framework that allows to teach motions in free space as well as in contact while preventing fast unintended contact transitions. This is accomplished by exploiting a projection-based unilateral damping force that increases close to contact. We derive an explicit analytical expression for the damping characteristics to ensure a safe stop before the contact when no further forces act on the robot. Furthermore, after the teaching, the recorded motion data is utilized to generate a time-optimized trajectory based on convex optimization, in which the contact transitions are explicitly considered. We validated our framework in experiments with a torque-controlled manipulator.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 4","pages":"3971-3978"},"PeriodicalIF":5.3,"publicationDate":"2026-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11373869","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146216643","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PA-MPPI: Perception-Aware Model Predictive Path Integral Control for Quadrotor Navigation in Unknown Environments PA-MPPI:未知环境下四旋翼飞行器导航的感知模型预测路径积分控制
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-02-09 DOI: 10.1109/LRA.2026.3662653
Yifan Zhai;Rudolf Reiter;Davide Scaramuzza
Quadrotor navigation in unknown environments is critical for practical missions such as search-and-rescue. Solving it requires addressing three key challenges: path-planning in non-convex free space due to obstacles, satisfying quadrotor-specific dynamics and objectives, and exploring unknown regions to expand the map. Recently, the Model Predictive Path Integral (MPPI) method has emerged as a promising solution that solves the first two challenges. By leveraging sampling-based optimization, it can effectively handle non-convex free space while directly optimizing over the full quadrotor dynamics, enabling the inclusion of quadrotor-specific costs such as energy consumption. However, MPPI has been limited to tracking control that only optimizes trajectories in a small neighbourhood around a reference trajectory, as it lacks the ability to explore unknown regions and plan alternative paths when blocked by large obstacles. To solve this issue, we introduce Perception-Aware MPPI (PA-MPPI). Here, perception-awareness is characterized by planning and adapting the trajectory online based on perception objectives. Specifically, when the goal is occluded, PA-MPPI's perception cost biases trajectories that can perceive unknown regions. This expands the mapped traversable space and increases the likelihood of finding alternative paths to the goal. Through hardware experiments, we demonstrate that PA-MPPI, running at 50 Hz, performs on par with the SOTA quadrotor navigation planner for unknown environments in our challenging test scenarios. In addition, we demonstrate that PA-MPPI can be used as a safe and robust action policy for navigation foundation models, which often provide goal poses that are not directly reachable.
在未知环境下的四旋翼导航对于诸如搜索和救援等实际任务至关重要。解决这一问题需要解决三个关键挑战:由于障碍导致的非凸自由空间中的路径规划,满足四旋翼飞行器特定的动力学和目标,以及探索未知区域以扩展地图。最近,模型预测路径积分(MPPI)方法作为解决前两个挑战的有希望的解决方案出现了。通过利用基于采样的优化,它可以有效地处理非凸自由空间,同时直接优化整个四旋翼动力学,从而包含四旋翼特定的成本,如能量消耗。然而,MPPI仅限于跟踪控制,仅在参考轨迹周围的小区域内优化轨迹,因为它缺乏探索未知区域和在被大型障碍物阻塞时规划替代路径的能力。为了解决这个问题,我们引入了感知感知MPPI (PA-MPPI)。在这里,感知-意识的特点是基于感知目标在线规划和适应轨迹。具体来说,当目标被遮挡时,PA-MPPI的感知成本会对能够感知未知区域的轨迹产生偏差。这扩展了映射的可穿越空间,并增加了找到通往目标的替代路径的可能性。通过硬件实验,我们证明了PA-MPPI在50 Hz下运行,在我们具有挑战性的测试场景中,与SOTA四旋翼导航规划器在未知环境中的表现相当。此外,我们证明了PA-MPPI可以作为导航基础模型的安全鲁棒动作策略,这些模型通常提供无法直接到达的目标姿态。
{"title":"PA-MPPI: Perception-Aware Model Predictive Path Integral Control for Quadrotor Navigation in Unknown Environments","authors":"Yifan Zhai;Rudolf Reiter;Davide Scaramuzza","doi":"10.1109/LRA.2026.3662653","DOIUrl":"https://doi.org/10.1109/LRA.2026.3662653","url":null,"abstract":"Quadrotor navigation in unknown environments is critical for practical missions such as search-and-rescue. Solving it requires addressing three key challenges: path-planning in non-convex free space due to obstacles, satisfying quadrotor-specific dynamics and objectives, and exploring unknown regions to expand the map. Recently, the Model Predictive Path Integral (MPPI) method has emerged as a promising solution that solves the first two challenges. By leveraging sampling-based optimization, it can effectively handle non-convex free space while directly optimizing over the full quadrotor dynamics, enabling the inclusion of quadrotor-specific costs such as energy consumption. However, MPPI has been limited to tracking control that only optimizes trajectories in a small neighbourhood around a reference trajectory, as it lacks the ability to explore unknown regions and plan alternative paths when blocked by large obstacles. To solve this issue, we introduce Perception-Aware MPPI (PA-MPPI). Here, perception-awareness is characterized by planning and adapting the trajectory online based on perception objectives. Specifically, when the goal is occluded, PA-MPPI's perception cost biases trajectories that can perceive unknown regions. This expands the mapped traversable space and increases the likelihood of finding alternative paths to the goal. Through hardware experiments, we demonstrate that PA-MPPI, running at 50 Hz, performs on par with the SOTA quadrotor navigation planner for unknown environments in our challenging test scenarios. In addition, we demonstrate that PA-MPPI can be used as a safe and robust action policy for navigation foundation models, which often provide goal poses that are not directly reachable.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 3","pages":"3804-3811"},"PeriodicalIF":5.3,"publicationDate":"2026-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146223611","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FlyAware: Inertia-Aware Aerial Manipulation via Vision-Based Estimation and Post-Grasp Adaptation 飞行感知:惯性感知空中操纵通过基于视觉的估计和后抓取适应
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-02-09 DOI: 10.1109/LRA.2026.3662562
Biyu Ye;Na Fan;Zhengping Fan;Weiliang Deng;Hongming Chen;Qifeng Chen;Ximin Lyu
Aerial manipulators (AMs) are gaining increasing attention in automated transportation and emergency services due to their superior dexterity compared to conventional multirotor drones. However, their practical deployment is challenged by the complexity of time-varying inertial parameters, which are highly sensitive to payload variations and manipulator configurations. Inspired by human strategies for interacting with unknown objects, this letter presents a novel onboard framework for robust aerial manipulation. The proposed system integrates a vision-based pre-grasp inertia estimation module with a post-grasp adaptation mechanism, enabling real-time estimation and adaptation of inertial dynamics. For control, we develop an inertia-aware adaptive control strategy based on gain scheduling, and assess its robustness via frequency-domain system identification. Our study provides new insights into post-grasp control for AMs, and real-world experiments validate the effectiveness and feasibility of the proposed framework.
与传统的多旋翼无人机相比,空中操纵器(AMs)由于其优越的灵活性,在自动化运输和应急服务中越来越受到关注。然而,它们的实际部署受到时变惯性参数复杂性的挑战,这些参数对有效载荷变化和机械臂配置高度敏感。受人类与未知物体互动策略的启发,这封信提出了一种新颖的机载框架,用于强大的空中操纵。该系统集成了基于视觉的抓前惯性估计模块和抓后自适应机制,实现了对惯性动力学的实时估计和自适应。在控制方面,我们开发了一种基于增益调度的惯性感知自适应控制策略,并通过频域系统辨识来评估其鲁棒性。我们的研究为AMs的抓后控制提供了新的见解,并且实际实验验证了所提出框架的有效性和可行性。
{"title":"FlyAware: Inertia-Aware Aerial Manipulation via Vision-Based Estimation and Post-Grasp Adaptation","authors":"Biyu Ye;Na Fan;Zhengping Fan;Weiliang Deng;Hongming Chen;Qifeng Chen;Ximin Lyu","doi":"10.1109/LRA.2026.3662562","DOIUrl":"https://doi.org/10.1109/LRA.2026.3662562","url":null,"abstract":"Aerial manipulators (AMs) are gaining increasing attention in automated transportation and emergency services due to their superior dexterity compared to conventional multirotor drones. However, their practical deployment is challenged by the complexity of time-varying inertial parameters, which are highly sensitive to payload variations and manipulator configurations. Inspired by human strategies for interacting with unknown objects, this letter presents a novel onboard framework for robust aerial manipulation. The proposed system integrates a vision-based pre-grasp inertia estimation module with a post-grasp adaptation mechanism, enabling real-time estimation and adaptation of inertial dynamics. For control, we develop an inertia-aware adaptive control strategy based on gain scheduling, and assess its robustness via frequency-domain system identification. Our study provides new insights into post-grasp control for AMs, and real-world experiments validate the effectiveness and feasibility of the proposed framework.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 3","pages":"3780-3787"},"PeriodicalIF":5.3,"publicationDate":"2026-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146223616","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
InteLiPlan: An Interactive Lightweight LLM-Based Planner for Domestic Robot Autonomy InteLiPlan:基于llm的交互式轻型家用机器人自主规划器
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-02-09 DOI: 10.1109/LRA.2026.3662577
Kim Tien Ly;Kai Lu;Ioannis Havoutis
We introduce an interactive LLM-based framework designed to enhance the autonomy and robustness of domestic robots, targeting embodied intelligence. Our approach reduces reliance on large-scale data and incorporates a robot-agnostic pipeline that embodies an LLM. Our framework, InteLiPlan, ensures that the LLM’s decision-making capabilities are effectively aligned with robotic functions, enhancing operational robustness and adaptability, while our human-in-the-loop mechanism allows for real-time human intervention when user instruction is required. We evaluate our method in both simulation and on the real robot platforms, including a Toyota Human Support Robot and an ANYmal D robot with a Unitree Z1 arm. Our method achieves a 95% success rate in the ‘fetch me’ task completion with failure recovery, highlighting its capability in both failure reasoning and task planning. InteLiPlan achieves comparable performance to state-of-the-art LLM-based robotics planners, while using only real-time onboard computing.
我们介绍了一个交互式的基于llm的框架,旨在增强家用机器人的自主性和鲁棒性,目标是具身智能。我们的方法减少了对大规模数据的依赖,并结合了一个包含LLM的机器人不可知管道。我们的InteLiPlan框架确保LLM的决策能力有效地与机器人功能保持一致,增强了操作的稳健性和适应性,而我们的人机环机制允许在需要用户指令时进行实时人工干预。我们在仿真和真实机器人平台上对我们的方法进行了评估,包括一个丰田人类支持机器人和一个带有Unitree Z1手臂的ANYmal D机器人。我们的方法在“取回我”任务的失败恢复中达到95%的成功率,突出了它在失败推理和任务规划方面的能力。InteLiPlan实现了与最先进的基于llm的机器人规划器相当的性能,同时仅使用实时板载计算。
{"title":"InteLiPlan: An Interactive Lightweight LLM-Based Planner for Domestic Robot Autonomy","authors":"Kim Tien Ly;Kai Lu;Ioannis Havoutis","doi":"10.1109/LRA.2026.3662577","DOIUrl":"https://doi.org/10.1109/LRA.2026.3662577","url":null,"abstract":"We introduce an interactive LLM-based framework designed to enhance the autonomy and robustness of domestic robots, targeting embodied intelligence. Our approach reduces reliance on large-scale data and incorporates a robot-agnostic pipeline that embodies an LLM. Our framework, <italic>InteLiPlan</i>, ensures that the LLM’s decision-making capabilities are effectively aligned with robotic functions, enhancing operational robustness and adaptability, while our human-in-the-loop mechanism allows for real-time human intervention when user instruction is required. We evaluate our method in both simulation and on the real robot platforms, including a Toyota Human Support Robot and an ANYmal D robot with a Unitree Z1 arm. Our method achieves a 95% success rate in the ‘fetch me’ task completion with failure recovery, highlighting its capability in both failure reasoning and task planning. <italic>InteLiPlan</i> achieves comparable performance to state-of-the-art LLM-based robotics planners, while using only real-time onboard computing.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 3","pages":"3875-3882"},"PeriodicalIF":5.3,"publicationDate":"2026-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146223663","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Labriform-Inspired Multi-Stable Soft Robotic Swimmer labriform启发的多稳定软机器人游泳者
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-02-09 DOI: 10.1109/LRA.2026.3662591
Jiaqiao Liang;Zefeng Xu;Peiyu Liu;Qiaosong Fan;Linjun Liu;Bin Xie;Ye Chen;Yitong Zhou
Pectoral-fin-based (labriform) swimming combines high-speed propulsion with agile maneuverability through rigid–flexible fin partitioning. Inspired by this principle, we present a multi-stable soft robotic swimmer composed of two wedge-shaped bistable actuators integrated with fin-like rigid–flexible morphologies. The bistable actuators generate large, rapid deformations for thrust production, while the compliant fin membranes enable drag-reducing feathering during recovery. An analytical model is developed to predict the shape of bistable actuators and is validated experimentally with a minimum prediction accuracy of 97.74%. Computational fluid dynamics (CFD) analysis reveals that bistable switching induces vortex dipole ejection, contributing to thrust generation. The proposed robot attains a maximum speed of 17.53 $text{cm}cdot text{s}^{-1}$ (1.10 $text{BL}cdot text{s}^{-1}$), a turning radius of 0.58 per body length, and a turning speed of 31.51$^circ$/s, highlighting our design in shaping both swimming speed and maneuverability. By integrating bistable actuation with bio-inspired fin morphologies, this work offers a principled design strategy for achieving fast and maneuverable swimming in robotic systems.
基于胸鳍的游泳通过刚柔分离,将高速推进与敏捷机动性相结合。受这一原理的启发,我们提出了一个多稳定的软体游泳机器人,由两个楔形双稳态驱动器组成,并结合了鳍状刚柔形态。双稳致动器产生大而快速的变形以产生推力,而柔顺的鳍膜可以在恢复过程中减少阻力。建立了预测双稳作动器形状的解析模型,并进行了实验验证,预测精度最低为97.74%。计算流体动力学(CFD)分析表明,双稳态开关诱导涡旋偶极子喷射,有助于推力的产生。该机器人的最大速度为17.53 $text{cm}cdot text{s}^{-1}$ (1.10 $text{BL}cdot text{s}^{-1}$),转弯半径为0.58 $/体长,转弯速度为31.51$^circ$/s,突出了我们在塑造游泳速度和机动性方面的设计。通过将双稳态驱动与仿生鳍形态相结合,这项工作为实现机器人系统中的快速机动游泳提供了原则性的设计策略。
{"title":"A Labriform-Inspired Multi-Stable Soft Robotic Swimmer","authors":"Jiaqiao Liang;Zefeng Xu;Peiyu Liu;Qiaosong Fan;Linjun Liu;Bin Xie;Ye Chen;Yitong Zhou","doi":"10.1109/LRA.2026.3662591","DOIUrl":"https://doi.org/10.1109/LRA.2026.3662591","url":null,"abstract":"Pectoral-fin-based (labriform) swimming combines high-speed propulsion with agile maneuverability through rigid–flexible fin partitioning. Inspired by this principle, we present a multi-stable soft robotic swimmer composed of two wedge-shaped bistable actuators integrated with fin-like rigid–flexible morphologies. The bistable actuators generate large, rapid deformations for thrust production, while the compliant fin membranes enable drag-reducing feathering during recovery. An analytical model is developed to predict the shape of bistable actuators and is validated experimentally with a minimum prediction accuracy of 97.74%. Computational fluid dynamics (CFD) analysis reveals that bistable switching induces vortex dipole ejection, contributing to thrust generation. The proposed robot attains a maximum speed of 17.53 <inline-formula><tex-math>$text{cm}cdot text{s}^{-1}$</tex-math></inline-formula> (1.10 <inline-formula><tex-math>$text{BL}cdot text{s}^{-1}$</tex-math></inline-formula>), a turning radius of 0.58 per body length, and a turning speed of 31.51<inline-formula><tex-math>$^circ$</tex-math></inline-formula>/s, highlighting our design in shaping both swimming speed and maneuverability. By integrating bistable actuation with bio-inspired fin morphologies, this work offers a principled design strategy for achieving fast and maneuverable swimming in robotic systems.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 4","pages":"4026-4033"},"PeriodicalIF":5.3,"publicationDate":"2026-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146216621","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HiMRAG: Hierarchical Multimodal Retrieval-Augmented Generation for Robot Task Planning 机器人任务规划的分层多模态检索增强生成
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-02-09 DOI: 10.1109/LRA.2026.3662582
Zhuoyi Zhang;Yixin Han;Renjun Li;Xiao Li
Data-driven methods offer promising solutions for robotic manipulation in human-centric environments, but enabling robots to operate complex appliances from natural language remains a significant challenge. The ambiguity of human instructions and the visual diversity of real-world objects make it difficult to generate precise and reliable action sequences. In this letter, we propose a hierarchical multimodal Retrieval-Augmented Generation (RAG) framework that fuses visual perception with language understanding. Our framework uses a vision-based module to identify an appliance and its documentation from a snapshot, then leverages a task-oriented RAG pipeline to process user instructions, retrieve relevant manual sections, and generate executable action sequences. We train and validate this framework on a custom dataset of microwave oven operation tasks and demonstrate its effectiveness, robustness, and practical viability through extensive virtual and physical experiments on a robotic platform.
数据驱动的方法为以人为中心的环境中的机器人操作提供了有前途的解决方案,但使机器人能够使用自然语言操作复杂的设备仍然是一个重大挑战。人类指令的模糊性和现实世界物体的视觉多样性使得生成精确可靠的动作序列变得困难。在这封信中,我们提出了一个分层的多模态检索-增强生成(RAG)框架,将视觉感知与语言理解融合在一起。我们的框架使用基于视觉的模块从快照中识别设备及其文档,然后利用面向任务的RAG管道处理用户指令,检索相关的手动部分,并生成可执行的操作序列。我们在微波炉操作任务的自定义数据集上训练和验证了该框架,并通过机器人平台上广泛的虚拟和物理实验证明了其有效性,鲁棒性和实际可行性。
{"title":"HiMRAG: Hierarchical Multimodal Retrieval-Augmented Generation for Robot Task Planning","authors":"Zhuoyi Zhang;Yixin Han;Renjun Li;Xiao Li","doi":"10.1109/LRA.2026.3662582","DOIUrl":"https://doi.org/10.1109/LRA.2026.3662582","url":null,"abstract":"Data-driven methods offer promising solutions for robotic manipulation in human-centric environments, but enabling robots to operate complex appliances from natural language remains a significant challenge. The ambiguity of human instructions and the visual diversity of real-world objects make it difficult to generate precise and reliable action sequences. In this letter, we propose a hierarchical multimodal Retrieval-Augmented Generation (RAG) framework that fuses visual perception with language understanding. Our framework uses a vision-based module to identify an appliance and its documentation from a snapshot, then leverages a task-oriented RAG pipeline to process user instructions, retrieve relevant manual sections, and generate executable action sequences. We train and validate this framework on a custom dataset of microwave oven operation tasks and demonstrate its effectiveness, robustness, and practical viability through extensive virtual and physical experiments on a robotic platform.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 3","pages":"3883-3890"},"PeriodicalIF":5.3,"publicationDate":"2026-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146223536","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Predicting Human Locomotion in Reduced Gravity via Deep Learning-Driven Musculoskeletal Simulation 通过深度学习驱动的肌肉骨骼模拟预测失重下的人体运动
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-02-09 DOI: 10.1109/LRA.2026.3662600
Mingyi Wang;Shuzhen Luo
Understanding human walking locomotion in reduced gravity could enhance astronaut mobility and improve the space exploration efficiency. However, existing studies often require high costs and significant time and resource commitments for natural locomotion studies. Here, we present a deep reinforcement learning (DRL)-based simulation framework that predicts locomotion patterns across reduced-gravity environments by learning control policies tailored to each gravity condition. This approach identifies optimal gait behaviors without extensive experimental data and can be extended to include assistive devices such as exoskeletons, enabling systematic studies of human–exoskeleton interaction and walking adaptation in reduced-gravity settings. To validate the simulation, we utilized a mechanical body-weight suspension system to replicate reduced gravity and conducted walking experiments under three reduced gravity levels. The stance phase (ST) decreased from 72.59% to 61.03% and the swing phase (SW) increased from 27.41% to 38.97%, with stride duration nearly constant. Under the exoskeleton assistance, ST decreased from 63.52% to 62.02%, and SW increased from 36.48% to 37.98%. Hip joint range of motion decreased consistently with gravity in both conditions. These trends closely matched experimental results, demonstrating the potential of DRL-based simulations for studying locomotion and assistive strategies in reduced gravity.
了解人类在失重条件下的行走运动,可以增强宇航员的机动性,提高太空探索效率。然而,现有的研究往往需要高成本和大量的时间和资源来研究自然运动。在这里,我们提出了一个基于深度强化学习(DRL)的仿真框架,该框架通过学习针对每种重力条件量身定制的控制策略来预测失重环境中的运动模式。该方法无需大量实验数据即可识别最佳步态行为,并且可以扩展到包括外骨骼等辅助设备,从而可以系统地研究人-外骨骼相互作用和减轻重力环境下的步行适应。为了验证仿真结果,我们利用机械体重悬挂系统来模拟失重,并在三种失重水平下进行了步行实验。在步幅基本不变的情况下,站姿阶段(ST)从72.59%下降到61.03%,摇摆阶段(SW)从27.41%上升到38.97%。在外骨骼辅助下,ST从63.52%下降到62.02%,SW从36.48%上升到37.98%。在这两种情况下,髋关节的活动范围都随着重力而下降。这些趋势与实验结果密切匹配,证明了基于drl的模拟在研究失重运动和辅助策略方面的潜力。
{"title":"Predicting Human Locomotion in Reduced Gravity via Deep Learning-Driven Musculoskeletal Simulation","authors":"Mingyi Wang;Shuzhen Luo","doi":"10.1109/LRA.2026.3662600","DOIUrl":"https://doi.org/10.1109/LRA.2026.3662600","url":null,"abstract":"Understanding human walking locomotion in reduced gravity could enhance astronaut mobility and improve the space exploration efficiency. However, existing studies often require high costs and significant time and resource commitments for natural locomotion studies. Here, we present a deep reinforcement learning (DRL)-based simulation framework that predicts locomotion patterns across reduced-gravity environments by learning control policies tailored to each gravity condition. This approach identifies optimal gait behaviors without extensive experimental data and can be extended to include assistive devices such as exoskeletons, enabling systematic studies of human–exoskeleton interaction and walking adaptation in reduced-gravity settings. To validate the simulation, we utilized a mechanical body-weight suspension system to replicate reduced gravity and conducted walking experiments under three reduced gravity levels. The stance phase (ST) decreased from 72.59% to 61.03% and the swing phase (SW) increased from 27.41% to 38.97%, with stride duration nearly constant. Under the exoskeleton assistance, ST decreased from 63.52% to 62.02%, and SW increased from 36.48% to 37.98%. Hip joint range of motion decreased consistently with gravity in both conditions. These trends closely matched experimental results, demonstrating the potential of DRL-based simulations for studying locomotion and assistive strategies in reduced gravity.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 3","pages":"3899-3906"},"PeriodicalIF":5.3,"publicationDate":"2026-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146223608","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Neuro-Symbolic Generation of Explanations for Robot Policies With Weighted Signal Temporal Logic 加权信号时间逻辑下机器人策略解释的神经符号生成
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-02-09 DOI: 10.1109/LRA.2026.3662977
Mikihisa Yuasa;Ramavarapu S. Sreenivas;Huy T. Tran
Learning-based policies have demonstrated success in many robotic applications, but often lack explainability. We propose a neuro-symbolic explanation framework that generates a weighted signal temporal logic (wSTL) specification which describes a robot policy in a human-interpretable form. Existing methods typically produce explanations that are verbose and inconsistent, which hinders explainability, and are loose, which limits meaningful insights. We address these issues by introducing a simplification process consisting of predicate filtering, regularization, and iterative pruning. We also introduce three explainability metrics—conciseness, consistency, and strictness—to assess explanation quality beyond conventional classification accuracy. Our method—TLNet—is validated in three simulated robotic environments, where it outperforms baselines in generating concise, consistent, and strict wSTL explanations without sacrificing accuracy. This work bridges policy learning and explainability through formal methods, contributing to more transparent decision-making in robotics.
基于学习的策略在许多机器人应用中取得了成功,但往往缺乏可解释性。我们提出了一个神经符号解释框架,该框架生成加权信号时间逻辑(wSTL)规范,该规范以人类可解释的形式描述机器人策略。现有的方法通常产生冗长和不一致的解释,这阻碍了可解释性,并且是松散的,这限制了有意义的见解。我们通过引入一个由谓词过滤、正则化和迭代修剪组成的简化过程来解决这些问题。我们还引入了三个可解释性指标——简洁性、一致性和严谨性——来评估超出常规分类准确性的解释质量。我们的方法tlnet在三个模拟机器人环境中进行了验证,在不牺牲准确性的情况下,它在生成简洁、一致和严格的wSTL解释方面优于基线。这项工作通过正式的方法将政策学习和可解释性联系起来,有助于机器人领域更透明的决策。
{"title":"Neuro-Symbolic Generation of Explanations for Robot Policies With Weighted Signal Temporal Logic","authors":"Mikihisa Yuasa;Ramavarapu S. Sreenivas;Huy T. Tran","doi":"10.1109/LRA.2026.3662977","DOIUrl":"https://doi.org/10.1109/LRA.2026.3662977","url":null,"abstract":"Learning-based policies have demonstrated success in many robotic applications, but often lack explainability. We propose a neuro-symbolic explanation framework that generates a weighted signal temporal logic (wSTL) specification which describes a robot policy in a human-interpretable form. Existing methods typically produce explanations that are verbose and inconsistent, which hinders explainability, and are loose, which limits meaningful insights. We address these issues by introducing a simplification process consisting of predicate filtering, regularization, and iterative pruning. We also introduce three explainability metrics—conciseness, consistency, and strictness—to assess explanation quality beyond conventional classification accuracy. Our method—<sc>TLNet</small>—is validated in three simulated robotic environments, where it outperforms baselines in generating concise, consistent, and strict wSTL explanations without sacrificing accuracy. This work bridges policy learning and explainability through formal methods, contributing to more transparent decision-making in robotics.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 4","pages":"3963-3970"},"PeriodicalIF":5.3,"publicationDate":"2026-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11386893","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146216611","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
From Pixels to Predicates: Learning Symbolic World Models via Pretrained VLMs 从像素到谓词:通过预训练vlm学习符号世界模型
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-02-09 DOI: 10.1109/LRA.2026.3662533
Ashay Athalye;Nishanth Kumar;Tom Silver;Yichao Liang;Jiuguang Wang;Tomás Lozano-Pérez;Leslie Pack Kaelbling
Our aim is to learn to solve long-horizon decision-making problems in complex robotics domains given low-level skills and a handful of demonstrations containing sequences of images. To this end, we focus on learning abstract symbolic world models that facilitate zero-shot generalization to novel goals via planning. A critical component of such models is the set of symbolic predicates that define properties of and relationships between objects. In this work, we leverage pretrained vision-language models (VLMs) to propose a large set of visual predicates potentially relevant for decision-making, and to evaluate those predicates directly from camera images. At training time, we pass the proposed predicates and demonstrations into an optimization-based model-learning algorithm to obtain an abstract symbolic world model that is defined in terms of a compact subset of the proposed predicates. At test time, given a novel goal in a novel setting, we use the VLM to construct a symbolic description of the current world state, and then use a search-based planning algorithm to find a sequence of low-level skills that achieves the goal. We demonstrate empirically across experiments in both simulation and the real world that our method can generalize aggressively, applying its learned world model to solve problems with varying visual backgrounds, types, numbers, and arrangements of objects, as well as novel goals and much longer horizons than those seen at training time.
我们的目标是在给定低水平技能和少量包含图像序列的演示的情况下,学习解决复杂机器人领域的长期决策问题。为此,我们专注于学习抽象的符号世界模型,这些模型可以通过计划促进对新目标的零概率泛化。这种模型的一个关键组件是一组符号谓词,用于定义对象的属性和对象之间的关系。在这项工作中,我们利用预训练的视觉语言模型(VLMs)来提出一组可能与决策相关的大量视觉谓词,并直接从相机图像中评估这些谓词。在训练时,我们将提出的谓词和演示传递到基于优化的模型学习算法中,以获得一个抽象的符号世界模型,该模型是根据提出的谓词的紧凑子集定义的。在测试时,给定一个新设置中的新目标,我们使用VLM构建当前世界状态的符号描述,然后使用基于搜索的规划算法找到实现目标的低级技能序列。我们通过模拟和现实世界的经验实验证明,我们的方法可以积极推广,应用其学习世界模型来解决具有不同视觉背景,类型,数量和物体排列的问题,以及新的目标和比训练时看到的更长的视野。
{"title":"From Pixels to Predicates: Learning Symbolic World Models via Pretrained VLMs","authors":"Ashay Athalye;Nishanth Kumar;Tom Silver;Yichao Liang;Jiuguang Wang;Tomás Lozano-Pérez;Leslie Pack Kaelbling","doi":"10.1109/LRA.2026.3662533","DOIUrl":"https://doi.org/10.1109/LRA.2026.3662533","url":null,"abstract":"Our aim is to learn to solve long-horizon decision-making problems in complex robotics domains given low-level skills and a handful of demonstrations containing sequences of images. To this end, we focus on learning abstract symbolic world models that facilitate zero-shot generalization to novel goals via planning. A critical component of such models is the set of symbolic <italic>predicates</i> that define properties of and relationships between objects. In this work, we leverage pretrained vision-language models (VLMs) to propose a large set of visual predicates potentially relevant for decision-making, and to evaluate those predicates directly from camera images. At training time, we pass the proposed predicates and demonstrations into an optimization-based model-learning algorithm to obtain an abstract symbolic world model that is defined in terms of a compact subset of the proposed predicates. At test time, given a novel goal in a novel setting, we use the VLM to construct a symbolic description of the current world state, and then use a search-based planning algorithm to find a sequence of low-level skills that achieves the goal. We demonstrate empirically across experiments in both simulation and the real world that our method can generalize aggressively, applying its learned world model to solve problems with varying visual backgrounds, types, numbers, and arrangements of objects, as well as novel goals and much longer horizons than those seen at training time.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 4","pages":"4002-4009"},"PeriodicalIF":5.3,"publicationDate":"2026-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146216528","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Design and Analysis of Hybrid Rigid-Soft Self-Aligning Index Finger Exoskeleton 刚软混合自对准食指外骨骼的设计与分析
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-02-09 DOI: 10.1109/LRA.2026.3662646
Yao Huang;Li Liu;Jian Sun;Bo Song
Disrupted hand motor functions may be restored through exoskeleton-assisted rehabilitation training. However, the variability of soft tissue in human joints or across individuals and development of an exoskeleton that combines human-machine motion compatibility and dynamic compliance pose persistent challenges. We introduce a hybrid single-motor-driven rigid–soft exoskeleton for the index finger to assist in rehabilitation training. A rigid parallel mechanism directly drives the soft component of the metacarpophalangeal (MCP) joint. In addition, we adopt an interlocking mechanism to induce deformation in leaf springs, enabling the coordinated flexion and extension of multiple joints. A motion analysis based on the modified Denavit–Hartenberg convention confirms that the proposed parallel mechanism can compensate for the misalignment displacement of the MCP joint. Based on the displacement and force applied to the soft component by the designed rigid parallel mechanism, kinematic and static analyses along with dimensional optimization are performed on a dual-segment parallel leaf spring. A prototype exoskeleton undergoing tests demonstrated Pearson correlation coefficients of 0.998, 0.991, 0.986, for the MCP, proximal and distal interphalangeal (PIP/DIP) joints, respectively. The corresponding joint flexion angles were 68.19°, 81.91°, and 41.64°. The exoskeleton self-aligns with the index finger joints, properly assisting the natural bending motion of the finger to meet rehabilitation training needs of patients. The proposed exoskeleton can assist with a fingertip force of 6.2 N, thereby satisfying grip requirements, while the reduced force on the dorsal surface of the index finger enhances comfort during use. The proposed solution is promising for developing hand exoskeletons.
中断的手部运动功能可以通过外骨骼辅助康复训练恢复。然而,人体关节或个体间软组织的可变性以及结合人机运动兼容性和动态顺应性的外骨骼的发展构成了持续的挑战。我们介绍了一个混合的单电机驱动的硬-软外骨骼的食指,以协助康复训练。刚性并联机构直接驱动掌指关节(MCP)的软组件。此外,我们采用联锁机制诱导钢板弹簧变形,实现多个关节的协调屈伸。基于改进的Denavit-Hartenberg惯例的运动分析证实了所提出的并联机构可以补偿MCP关节的错位位移。基于所设计的刚性并联机构对软构件施加的位移和力,对双节并联钢板弹簧进行了运动学和静力分析,并进行了尺寸优化。经测试的外骨骼原型显示,MCP、近端和远端指间关节(PIP/DIP)的Pearson相关系数分别为0.998、0.991和0.986。相应的关节屈曲角度分别为68.19°、81.91°和41.64°。外骨骼与食指关节自我对准,适当辅助手指自然弯曲运动,满足患者康复训练需求。所提出的外骨骼可以辅助6.2 N的指尖力,从而满足抓握要求,同时减少了食指背表面的力,提高了使用时的舒适性。提出的解决方案有望开发手外骨骼。
{"title":"Design and Analysis of Hybrid Rigid-Soft Self-Aligning Index Finger Exoskeleton","authors":"Yao Huang;Li Liu;Jian Sun;Bo Song","doi":"10.1109/LRA.2026.3662646","DOIUrl":"https://doi.org/10.1109/LRA.2026.3662646","url":null,"abstract":"Disrupted hand motor functions may be restored through exoskeleton-assisted rehabilitation training. However, the variability of soft tissue in human joints or across individuals and development of an exoskeleton that combines human-machine motion compatibility and dynamic compliance pose persistent challenges. We introduce a hybrid single-motor-driven rigid–soft exoskeleton for the index finger to assist in rehabilitation training. A rigid parallel mechanism directly drives the soft component of the metacarpophalangeal (MCP) joint. In addition, we adopt an interlocking mechanism to induce deformation in leaf springs, enabling the coordinated flexion and extension of multiple joints. A motion analysis based on the modified Denavit–Hartenberg convention confirms that the proposed parallel mechanism can compensate for the misalignment displacement of the MCP joint. Based on the displacement and force applied to the soft component by the designed rigid parallel mechanism, kinematic and static analyses along with dimensional optimization are performed on a dual-segment parallel leaf spring. A prototype exoskeleton undergoing tests demonstrated Pearson correlation coefficients of 0.998, 0.991, 0.986, for the MCP, proximal and distal interphalangeal (PIP/DIP) joints, respectively. The corresponding joint flexion angles were 68.19°, 81.91°, and 41.64°. The exoskeleton self-aligns with the index finger joints, properly assisting the natural bending motion of the finger to meet rehabilitation training needs of patients. The proposed exoskeleton can assist with a fingertip force of 6.2 N, thereby satisfying grip requirements, while the reduced force on the dorsal surface of the index finger enhances comfort during use. The proposed solution is promising for developing hand exoskeletons.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 4","pages":"4010-4017"},"PeriodicalIF":5.3,"publicationDate":"2026-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146216616","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Robotics and Automation Letters
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1