首页 > 最新文献

IEEE Robotics and Automation Letters最新文献

英文 中文
Neuro-Symbolic Generation of Explanations for Robot Policies With Weighted Signal Temporal Logic 加权信号时间逻辑下机器人策略解释的神经符号生成
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-02-09 DOI: 10.1109/LRA.2026.3662977
Mikihisa Yuasa;Ramavarapu S. Sreenivas;Huy T. Tran
Learning-based policies have demonstrated success in many robotic applications, but often lack explainability. We propose a neuro-symbolic explanation framework that generates a weighted signal temporal logic (wSTL) specification which describes a robot policy in a human-interpretable form. Existing methods typically produce explanations that are verbose and inconsistent, which hinders explainability, and are loose, which limits meaningful insights. We address these issues by introducing a simplification process consisting of predicate filtering, regularization, and iterative pruning. We also introduce three explainability metrics—conciseness, consistency, and strictness—to assess explanation quality beyond conventional classification accuracy. Our method—TLNet—is validated in three simulated robotic environments, where it outperforms baselines in generating concise, consistent, and strict wSTL explanations without sacrificing accuracy. This work bridges policy learning and explainability through formal methods, contributing to more transparent decision-making in robotics.
基于学习的策略在许多机器人应用中取得了成功,但往往缺乏可解释性。我们提出了一个神经符号解释框架,该框架生成加权信号时间逻辑(wSTL)规范,该规范以人类可解释的形式描述机器人策略。现有的方法通常产生冗长和不一致的解释,这阻碍了可解释性,并且是松散的,这限制了有意义的见解。我们通过引入一个由谓词过滤、正则化和迭代修剪组成的简化过程来解决这些问题。我们还引入了三个可解释性指标——简洁性、一致性和严谨性——来评估超出常规分类准确性的解释质量。我们的方法tlnet在三个模拟机器人环境中进行了验证,在不牺牲准确性的情况下,它在生成简洁、一致和严格的wSTL解释方面优于基线。这项工作通过正式的方法将政策学习和可解释性联系起来,有助于机器人领域更透明的决策。
{"title":"Neuro-Symbolic Generation of Explanations for Robot Policies With Weighted Signal Temporal Logic","authors":"Mikihisa Yuasa;Ramavarapu S. Sreenivas;Huy T. Tran","doi":"10.1109/LRA.2026.3662977","DOIUrl":"https://doi.org/10.1109/LRA.2026.3662977","url":null,"abstract":"Learning-based policies have demonstrated success in many robotic applications, but often lack explainability. We propose a neuro-symbolic explanation framework that generates a weighted signal temporal logic (wSTL) specification which describes a robot policy in a human-interpretable form. Existing methods typically produce explanations that are verbose and inconsistent, which hinders explainability, and are loose, which limits meaningful insights. We address these issues by introducing a simplification process consisting of predicate filtering, regularization, and iterative pruning. We also introduce three explainability metrics—conciseness, consistency, and strictness—to assess explanation quality beyond conventional classification accuracy. Our method—<sc>TLNet</small>—is validated in three simulated robotic environments, where it outperforms baselines in generating concise, consistent, and strict wSTL explanations without sacrificing accuracy. This work bridges policy learning and explainability through formal methods, contributing to more transparent decision-making in robotics.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 4","pages":"3963-3970"},"PeriodicalIF":5.3,"publicationDate":"2026-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11386893","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146216611","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
From Pixels to Predicates: Learning Symbolic World Models via Pretrained VLMs 从像素到谓词:通过预训练vlm学习符号世界模型
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-02-09 DOI: 10.1109/LRA.2026.3662533
Ashay Athalye;Nishanth Kumar;Tom Silver;Yichao Liang;Jiuguang Wang;Tomás Lozano-Pérez;Leslie Pack Kaelbling
Our aim is to learn to solve long-horizon decision-making problems in complex robotics domains given low-level skills and a handful of demonstrations containing sequences of images. To this end, we focus on learning abstract symbolic world models that facilitate zero-shot generalization to novel goals via planning. A critical component of such models is the set of symbolic predicates that define properties of and relationships between objects. In this work, we leverage pretrained vision-language models (VLMs) to propose a large set of visual predicates potentially relevant for decision-making, and to evaluate those predicates directly from camera images. At training time, we pass the proposed predicates and demonstrations into an optimization-based model-learning algorithm to obtain an abstract symbolic world model that is defined in terms of a compact subset of the proposed predicates. At test time, given a novel goal in a novel setting, we use the VLM to construct a symbolic description of the current world state, and then use a search-based planning algorithm to find a sequence of low-level skills that achieves the goal. We demonstrate empirically across experiments in both simulation and the real world that our method can generalize aggressively, applying its learned world model to solve problems with varying visual backgrounds, types, numbers, and arrangements of objects, as well as novel goals and much longer horizons than those seen at training time.
我们的目标是在给定低水平技能和少量包含图像序列的演示的情况下,学习解决复杂机器人领域的长期决策问题。为此,我们专注于学习抽象的符号世界模型,这些模型可以通过计划促进对新目标的零概率泛化。这种模型的一个关键组件是一组符号谓词,用于定义对象的属性和对象之间的关系。在这项工作中,我们利用预训练的视觉语言模型(VLMs)来提出一组可能与决策相关的大量视觉谓词,并直接从相机图像中评估这些谓词。在训练时,我们将提出的谓词和演示传递到基于优化的模型学习算法中,以获得一个抽象的符号世界模型,该模型是根据提出的谓词的紧凑子集定义的。在测试时,给定一个新设置中的新目标,我们使用VLM构建当前世界状态的符号描述,然后使用基于搜索的规划算法找到实现目标的低级技能序列。我们通过模拟和现实世界的经验实验证明,我们的方法可以积极推广,应用其学习世界模型来解决具有不同视觉背景,类型,数量和物体排列的问题,以及新的目标和比训练时看到的更长的视野。
{"title":"From Pixels to Predicates: Learning Symbolic World Models via Pretrained VLMs","authors":"Ashay Athalye;Nishanth Kumar;Tom Silver;Yichao Liang;Jiuguang Wang;Tomás Lozano-Pérez;Leslie Pack Kaelbling","doi":"10.1109/LRA.2026.3662533","DOIUrl":"https://doi.org/10.1109/LRA.2026.3662533","url":null,"abstract":"Our aim is to learn to solve long-horizon decision-making problems in complex robotics domains given low-level skills and a handful of demonstrations containing sequences of images. To this end, we focus on learning abstract symbolic world models that facilitate zero-shot generalization to novel goals via planning. A critical component of such models is the set of symbolic <italic>predicates</i> that define properties of and relationships between objects. In this work, we leverage pretrained vision-language models (VLMs) to propose a large set of visual predicates potentially relevant for decision-making, and to evaluate those predicates directly from camera images. At training time, we pass the proposed predicates and demonstrations into an optimization-based model-learning algorithm to obtain an abstract symbolic world model that is defined in terms of a compact subset of the proposed predicates. At test time, given a novel goal in a novel setting, we use the VLM to construct a symbolic description of the current world state, and then use a search-based planning algorithm to find a sequence of low-level skills that achieves the goal. We demonstrate empirically across experiments in both simulation and the real world that our method can generalize aggressively, applying its learned world model to solve problems with varying visual backgrounds, types, numbers, and arrangements of objects, as well as novel goals and much longer horizons than those seen at training time.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 4","pages":"4002-4009"},"PeriodicalIF":5.3,"publicationDate":"2026-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146216528","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Design and Analysis of Hybrid Rigid-Soft Self-Aligning Index Finger Exoskeleton 刚软混合自对准食指外骨骼的设计与分析
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-02-09 DOI: 10.1109/LRA.2026.3662646
Yao Huang;Li Liu;Jian Sun;Bo Song
Disrupted hand motor functions may be restored through exoskeleton-assisted rehabilitation training. However, the variability of soft tissue in human joints or across individuals and development of an exoskeleton that combines human-machine motion compatibility and dynamic compliance pose persistent challenges. We introduce a hybrid single-motor-driven rigid–soft exoskeleton for the index finger to assist in rehabilitation training. A rigid parallel mechanism directly drives the soft component of the metacarpophalangeal (MCP) joint. In addition, we adopt an interlocking mechanism to induce deformation in leaf springs, enabling the coordinated flexion and extension of multiple joints. A motion analysis based on the modified Denavit–Hartenberg convention confirms that the proposed parallel mechanism can compensate for the misalignment displacement of the MCP joint. Based on the displacement and force applied to the soft component by the designed rigid parallel mechanism, kinematic and static analyses along with dimensional optimization are performed on a dual-segment parallel leaf spring. A prototype exoskeleton undergoing tests demonstrated Pearson correlation coefficients of 0.998, 0.991, 0.986, for the MCP, proximal and distal interphalangeal (PIP/DIP) joints, respectively. The corresponding joint flexion angles were 68.19°, 81.91°, and 41.64°. The exoskeleton self-aligns with the index finger joints, properly assisting the natural bending motion of the finger to meet rehabilitation training needs of patients. The proposed exoskeleton can assist with a fingertip force of 6.2 N, thereby satisfying grip requirements, while the reduced force on the dorsal surface of the index finger enhances comfort during use. The proposed solution is promising for developing hand exoskeletons.
中断的手部运动功能可以通过外骨骼辅助康复训练恢复。然而,人体关节或个体间软组织的可变性以及结合人机运动兼容性和动态顺应性的外骨骼的发展构成了持续的挑战。我们介绍了一个混合的单电机驱动的硬-软外骨骼的食指,以协助康复训练。刚性并联机构直接驱动掌指关节(MCP)的软组件。此外,我们采用联锁机制诱导钢板弹簧变形,实现多个关节的协调屈伸。基于改进的Denavit-Hartenberg惯例的运动分析证实了所提出的并联机构可以补偿MCP关节的错位位移。基于所设计的刚性并联机构对软构件施加的位移和力,对双节并联钢板弹簧进行了运动学和静力分析,并进行了尺寸优化。经测试的外骨骼原型显示,MCP、近端和远端指间关节(PIP/DIP)的Pearson相关系数分别为0.998、0.991和0.986。相应的关节屈曲角度分别为68.19°、81.91°和41.64°。外骨骼与食指关节自我对准,适当辅助手指自然弯曲运动,满足患者康复训练需求。所提出的外骨骼可以辅助6.2 N的指尖力,从而满足抓握要求,同时减少了食指背表面的力,提高了使用时的舒适性。提出的解决方案有望开发手外骨骼。
{"title":"Design and Analysis of Hybrid Rigid-Soft Self-Aligning Index Finger Exoskeleton","authors":"Yao Huang;Li Liu;Jian Sun;Bo Song","doi":"10.1109/LRA.2026.3662646","DOIUrl":"https://doi.org/10.1109/LRA.2026.3662646","url":null,"abstract":"Disrupted hand motor functions may be restored through exoskeleton-assisted rehabilitation training. However, the variability of soft tissue in human joints or across individuals and development of an exoskeleton that combines human-machine motion compatibility and dynamic compliance pose persistent challenges. We introduce a hybrid single-motor-driven rigid–soft exoskeleton for the index finger to assist in rehabilitation training. A rigid parallel mechanism directly drives the soft component of the metacarpophalangeal (MCP) joint. In addition, we adopt an interlocking mechanism to induce deformation in leaf springs, enabling the coordinated flexion and extension of multiple joints. A motion analysis based on the modified Denavit–Hartenberg convention confirms that the proposed parallel mechanism can compensate for the misalignment displacement of the MCP joint. Based on the displacement and force applied to the soft component by the designed rigid parallel mechanism, kinematic and static analyses along with dimensional optimization are performed on a dual-segment parallel leaf spring. A prototype exoskeleton undergoing tests demonstrated Pearson correlation coefficients of 0.998, 0.991, 0.986, for the MCP, proximal and distal interphalangeal (PIP/DIP) joints, respectively. The corresponding joint flexion angles were 68.19°, 81.91°, and 41.64°. The exoskeleton self-aligns with the index finger joints, properly assisting the natural bending motion of the finger to meet rehabilitation training needs of patients. The proposed exoskeleton can assist with a fingertip force of 6.2 N, thereby satisfying grip requirements, while the reduced force on the dorsal surface of the index finger enhances comfort during use. The proposed solution is promising for developing hand exoskeletons.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 4","pages":"4010-4017"},"PeriodicalIF":5.3,"publicationDate":"2026-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146216616","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Stable Vision-Based Robot Kinematic Control With Deep Learning-Based Oriented Object Detector 基于深度学习的定向目标检测器的稳定视觉机器人运动控制
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-02-09 DOI: 10.1109/LRA.2026.3662576
Sitan Li;Chao Liu;Koji Matsuno;Chien Chern Cheah
Recent advances in machine learning and deep learning have significantly enhanced robot control by improving object detection and visual feature extraction. However, ensuring theoretical guarantees of stability and convergence in learning-enabled control systems remains a major challenge. In this paper, we propose a vision-based control framework that integrates a deep learning oriented-object detector with a Lyapunov-stable servo control law. The proposed method ensures provably stable convergence of the robot end-effector or its grasped object’s pose to a desired camera image region for both eye-in-hand and eye-to-hand configurations. Unlike existing deep learning based visual servoing methods, which either lack formal stability guarantees or ignore object orientation control, our approach incorporates object orientation into the control loop through a region-based method using quaternion representation and formally guarantees stability. We validated our framework on a 6-DoF UR5e manipulator performing cup insertion and centering tasks, demonstrating accurate and stable control in both camera setups.
机器学习和深度学习的最新进展通过改进物体检测和视觉特征提取显著增强了机器人的控制。然而,从理论上保证学习控制系统的稳定性和收敛性仍然是一个主要的挑战。在本文中,我们提出了一种基于视觉的控制框架,该框架集成了面向深度学习的对象检测器和李雅普诺夫稳定伺服控制律。该方法保证了机器人末端执行器或其被抓物体的姿态收敛到所需的相机图像区域的稳定性,无论是眼对手构型还是眼对手构型。与现有的基于深度学习的视觉伺服方法不同,这些方法要么缺乏正式的稳定性保证,要么忽略对象定向控制,我们的方法通过使用四元数表示的基于区域的方法将对象定向纳入控制回路,并正式保证稳定性。我们在6自由度UR5e机械手上验证了我们的框架,执行杯插入和定心任务,在两种相机设置中都展示了准确和稳定的控制。
{"title":"Stable Vision-Based Robot Kinematic Control With Deep Learning-Based Oriented Object Detector","authors":"Sitan Li;Chao Liu;Koji Matsuno;Chien Chern Cheah","doi":"10.1109/LRA.2026.3662576","DOIUrl":"https://doi.org/10.1109/LRA.2026.3662576","url":null,"abstract":"Recent advances in machine learning and deep learning have significantly enhanced robot control by improving object detection and visual feature extraction. However, ensuring theoretical guarantees of stability and convergence in learning-enabled control systems remains a major challenge. In this paper, we propose a vision-based control framework that integrates a deep learning oriented-object detector with a Lyapunov-stable servo control law. The proposed method ensures provably stable convergence of the robot end-effector or its grasped object’s pose to a desired camera image region for both eye-in-hand and eye-to-hand configurations. Unlike existing deep learning based visual servoing methods, which either lack formal stability guarantees or ignore object orientation control, our approach incorporates object orientation into the control loop through a region-based method using quaternion representation and formally guarantees stability. We validated our framework on a 6-DoF UR5e manipulator performing cup insertion and centering tasks, demonstrating accurate and stable control in both camera setups.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 3","pages":"3915-3922"},"PeriodicalIF":5.3,"publicationDate":"2026-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146223645","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Event-Triggered Indirect Herding Control of a Cooperative Agent 协同Agent的事件触发间接羊群控制
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-02-09 DOI: 10.1109/LRA.2026.3662559
Patrick M. Amy;Brandon C. Fallin;Jhyv N. Philor;Warren E. Dixon
This work explores the indirect herding control problem for a single pursuer agent regulating a single target agent to a goal location. To accommodate the constraints of sensing hardware, an event-triggered inter-agent influence model between the pursuer agent and target agent is considered. Motivated by fielded sensing systems, we present an event-triggered controller and trigger mechanism that satisfies a user-selected minimum inter-event time. The combined pursuer-target system is presented as a switched system that alternates between stable and unstable modes. A dwell-time analysis is completed to develop a closed-form solution for the maximum time the pursuer agent can allow the target agent to evolve in the unstable mode before requiring a control input update. The presented trigger function is designed to produce inter-event times that are upper-bounded by the maximum dwell time. The effectiveness of the proposed approach is demonstrated through both simulated and experimental studies, where a pursuer agent successfully regulates a target agent to a desired goal location.
本研究探讨了单个追踪体调节单个目标体到目标位置的间接羊群控制问题。为了适应传感硬件的约束,考虑了追踪体和目标体之间的事件触发agent间影响模型。在现场传感系统的激励下,我们提出了一个事件触发控制器和触发机制,满足用户选择的最小事件间时间。跟踪-目标组合系统是一个在稳定模式和不稳定模式之间交替的切换系统。在需要更新控制输入之前,跟踪代理可以允许目标代理在不稳定模式中进化的最大时间,完成了驻留时间分析,以开发封闭形式的解决方案。所提出的触发函数被设计为产生以最大停留时间为上限的事件间时间。通过模拟和实验研究证明了所提出方法的有效性,其中跟踪代理成功地将目标代理调节到期望的目标位置。
{"title":"Event-Triggered Indirect Herding Control of a Cooperative Agent","authors":"Patrick M. Amy;Brandon C. Fallin;Jhyv N. Philor;Warren E. Dixon","doi":"10.1109/LRA.2026.3662559","DOIUrl":"https://doi.org/10.1109/LRA.2026.3662559","url":null,"abstract":"This work explores the indirect herding control problem for a single pursuer agent regulating a single target agent to a goal location. To accommodate the constraints of sensing hardware, an event-triggered inter-agent influence model between the pursuer agent and target agent is considered. Motivated by fielded sensing systems, we present an event-triggered controller and trigger mechanism that satisfies a user-selected minimum inter-event time. The combined pursuer-target system is presented as a switched system that alternates between stable and unstable modes. A dwell-time analysis is completed to develop a closed-form solution for the maximum time the pursuer agent can allow the target agent to evolve in the unstable mode before requiring a control input update. The presented trigger function is designed to produce inter-event times that are upper-bounded by the maximum dwell time. The effectiveness of the proposed approach is demonstrated through both simulated and experimental studies, where a pursuer agent successfully regulates a target agent to a desired goal location.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 3","pages":"3828-3835"},"PeriodicalIF":5.3,"publicationDate":"2026-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146223659","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
From Obstacles to Etiquette: Robot Social Navigation With VLM-Informed Path Selection 从障碍到礼仪:基于vlm信息路径选择的机器人社交导航
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-02-09 DOI: 10.1109/LRA.2026.3662586
Zilin Fang;Anxing Xiao;David Hsu;Gim Hee Lee
Navigating socially in human environments requires more than satisfying geometric constraints, as collision-free paths may still interfere with ongoing activities or conflict with social norms. Addressing this challenge calls for analyzing interactions between agents and incorporating common-sense reasoning into planning. This paper presents a social robot navigation framework that integrates geometric planning with contextual social reasoning. The system first extracts obstacles and human dynamics to generate geometrically feasible candidate paths, then leverages a fine-tuned vision-language model (VLM) to evaluate these paths, informed by contextually grounded social expectations, selecting a socially optimized path for the controller. This task-specific VLM distills social reasoning from large foundation models into a smaller and efficient model, allowing the framework to perform real-time adaptation in diverse human–robot interaction contexts. Experiments in four social navigation contexts demonstrate that our method achieves the best overall performance with the lowest personal space violation duration, the minimal pedestrian-facing time, and no social zone intrusions.
在人类环境中进行社交导航需要的不仅仅是满足几何约束,因为无碰撞路径可能仍然会干扰正在进行的活动或与社会规范发生冲突。解决这一挑战需要分析代理之间的相互作用,并将常识性推理纳入规划。本文提出了一种将几何规划与情境社会推理相结合的社交机器人导航框架。该系统首先提取障碍物和人类动态,生成几何上可行的候选路径,然后利用微调的视觉语言模型(VLM)来评估这些路径,根据情境基础的社会期望,为控制器选择一条社会优化的路径。这个特定于任务的VLM将社会推理从大型基础模型提炼成一个更小、更高效的模型,允许框架在不同的人机交互环境中执行实时适应。在四种社交导航环境下的实验表明,我们的方法以最少的个人空间侵犯持续时间、最少的行人面对时间和无社交区域入侵达到了最佳的综合性能。
{"title":"From Obstacles to Etiquette: Robot Social Navigation With VLM-Informed Path Selection","authors":"Zilin Fang;Anxing Xiao;David Hsu;Gim Hee Lee","doi":"10.1109/LRA.2026.3662586","DOIUrl":"https://doi.org/10.1109/LRA.2026.3662586","url":null,"abstract":"Navigating socially in human environments requires more than satisfying geometric constraints, as collision-free paths may still interfere with ongoing activities or conflict with social norms. Addressing this challenge calls for analyzing interactions between agents and incorporating common-sense reasoning into planning. This paper presents a social robot navigation framework that integrates geometric planning with contextual social reasoning. The system first extracts obstacles and human dynamics to generate geometrically feasible candidate paths, then leverages a fine-tuned vision-language model (VLM) to evaluate these paths, informed by contextually grounded social expectations, selecting a socially optimized path for the controller. This task-specific VLM distills social reasoning from large foundation models into a smaller and efficient model, allowing the framework to perform real-time adaptation in diverse human–robot interaction contexts. Experiments in four social navigation contexts demonstrate that our method achieves the best overall performance with the lowest personal space violation duration, the minimal pedestrian-facing time, and no social zone intrusions.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 4","pages":"3947-3954"},"PeriodicalIF":5.3,"publicationDate":"2026-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146216636","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Gripper System With Soft Dielectric Elastomer Functional Units for Actuation, Sensing, and Signal Processing 具有软介电弹性体功能单元的抓手系统,用于驱动,传感和信号处理
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-02-09 DOI: 10.1109/LRA.2026.3662528
Junhao Ni;Moritz Scharff;Katherine Wilson;Hui Zhi Beh;Andreas Tairych;Andreas Richter;Iain Anderson;Gerald Gerlach;E.-F. Markus Vorrath
This paper presents a soft gripper system in which actuation, sensing, and signal processing are all realized using soft dielectric elastomer (DE) devices. The platform integrates three functional units: a multilayer dielectric elastomer actuator (DEA) module that drives a compliant two-finger gripper, two tactile sensors mounted on the fingertips for contact detection and coarse shape recognition, and a dielectric elastomer switch (DES) that enables soft electronics control. The DEA module consists of eight parallel multilayer actuators, each comprising four active layers and a nonlinear biasing spring to amplify stroke, and is powered by an embedded high-voltage supply operated at 3 kV. The DES operates by mechanically modulating the resistance of a stretchable piezoresistive electrode, providing reliable switching behavior under high voltage. Grasping tests demonstrate that the system achieves a maximum opening angle of 38$^{circ }$ and can safely manipulate delicate objects, including cherries and eggs, without damage. These results demonstrate the feasibility of advancing toward fully soft robotic systems by integrating DE-based components for actuation, sensing, and signal processing.
本文介绍了一种采用软介电弹性体(DE)器件实现驱动、传感和信号处理的软夹持器系统。该平台集成了三个功能单元:一个多层介电弹性体致动器(DEA)模块,驱动一个兼容的双指抓手,两个安装在指尖上的触觉传感器,用于接触检测和粗糙形状识别,以及一个介电弹性体开关(DES),实现软电子控制。DEA模块由8个并联多层执行器组成,每个执行器由4个有源层和一个非线性偏置弹簧组成,用于放大行程,并由一个3kv的嵌入式高压电源供电。DES通过机械调制可拉伸压阻电极的电阻来工作,在高压下提供可靠的开关行为。抓取测试表明,该系统的最大打开角度为38$^{circ}$,可以安全地操作包括樱桃和鸡蛋在内的精致物体而不会损坏。这些结果表明,通过集成基于de的驱动、传感和信号处理组件,向全软机器人系统迈进是可行的。
{"title":"A Gripper System With Soft Dielectric Elastomer Functional Units for Actuation, Sensing, and Signal Processing","authors":"Junhao Ni;Moritz Scharff;Katherine Wilson;Hui Zhi Beh;Andreas Tairych;Andreas Richter;Iain Anderson;Gerald Gerlach;E.-F. Markus Vorrath","doi":"10.1109/LRA.2026.3662528","DOIUrl":"https://doi.org/10.1109/LRA.2026.3662528","url":null,"abstract":"This paper presents a soft gripper system in which actuation, sensing, and signal processing are all realized using soft dielectric elastomer (DE) devices. The platform integrates three functional units: a multilayer dielectric elastomer actuator (DEA) module that drives a compliant two-finger gripper, two tactile sensors mounted on the fingertips for contact detection and coarse shape recognition, and a dielectric elastomer switch (DES) that enables soft electronics control. The DEA module consists of eight parallel multilayer actuators, each comprising four active layers and a nonlinear biasing spring to amplify stroke, and is powered by an embedded high-voltage supply operated at 3 kV. The DES operates by mechanically modulating the resistance of a stretchable piezoresistive electrode, providing reliable switching behavior under high voltage. Grasping tests demonstrate that the system achieves a maximum opening angle of 38<inline-formula><tex-math>$^{circ }$</tex-math></inline-formula> and can safely manipulate delicate objects, including cherries and eggs, without damage. These results demonstrate the feasibility of advancing toward fully soft robotic systems by integrating DE-based components for actuation, sensing, and signal processing.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 3","pages":"3598-3605"},"PeriodicalIF":5.3,"publicationDate":"2026-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11373904","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146175861","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DoGFlow: Self-Supervised LiDAR Scene Flow via Cross-Modal Doppler Guidance DoGFlow:通过交叉模态多普勒制导的自监督激光雷达场景流
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-02-09 DOI: 10.1109/LRA.2026.3662592
Ajinkya Khoche;Qingwen Zhang;Yixi Cai;Sina Sharif Mansouri;Patric Jensfelt
Accurate 3D scene flow estimation is critical for autonomous systems to navigate dynamic environments safely, but creating the necessary large-scale, manually annotated datasets remains a significant bottleneck for developing robust perception models. Current self-supervised methods struggle to match the performance of fully supervised approaches, especially in challenging long-range and adverse weather scenarios, while supervised methods are not scalable due to their reliance on expensive human labeling. We introduce DoGFlow, a novel self-supervised framework that recovers full 3D object motions for LiDAR scene flow estimation without requiring any manual ground truth annotations. This paper presents our cross-modal label transfer approach, where DoGFlow computes motion labels directly from 4D radar Doppler measurements and transfers them to the LiDAR domain using dynamic-aware association and ambiguity-resolved propagation. On the challenging MAN TruckScenes dataset, DoGFlow substantially outperforms existing self-supervised methods and improves label efficiency by enabling LiDAR backbones to achieve over 90% of fully supervised performance with only 10% of the ground truth data.
准确的3D场景流估计对于自主系统安全导航动态环境至关重要,但创建必要的大规模手动注释数据集仍然是开发鲁棒感知模型的重要瓶颈。目前的自我监督方法很难与完全监督方法的性能相匹配,特别是在具有挑战性的远程和恶劣天气场景中,而监督方法由于依赖昂贵的人工标记而无法扩展。我们介绍了DoGFlow,这是一个新颖的自监督框架,可以恢复LiDAR场景流估计的完整3D物体运动,而无需任何手动地面真值注释。本文介绍了我们的跨模态标签传输方法,其中DoGFlow直接从4D雷达多普勒测量中计算运动标签,并使用动态感知关联和模糊解决传播将其传输到LiDAR域。在具有挑战性的MAN TruckScenes数据集上,DoGFlow大大优于现有的自我监督方法,并通过使激光雷达骨干仅使用10%的地面真实数据就能实现90%以上的完全监督性能,从而提高了标签效率。
{"title":"DoGFlow: Self-Supervised LiDAR Scene Flow via Cross-Modal Doppler Guidance","authors":"Ajinkya Khoche;Qingwen Zhang;Yixi Cai;Sina Sharif Mansouri;Patric Jensfelt","doi":"10.1109/LRA.2026.3662592","DOIUrl":"https://doi.org/10.1109/LRA.2026.3662592","url":null,"abstract":"Accurate 3D scene flow estimation is critical for autonomous systems to navigate dynamic environments safely, but creating the necessary large-scale, manually annotated datasets remains a significant bottleneck for developing robust perception models. Current self-supervised methods struggle to match the performance of fully supervised approaches, especially in challenging long-range and adverse weather scenarios, while supervised methods are not scalable due to their reliance on expensive human labeling. We introduce DoGFlow, a novel self-supervised framework that recovers full 3D object motions for LiDAR scene flow estimation without requiring any manual ground truth annotations. This paper presents our cross-modal label transfer approach, where DoGFlow computes motion labels directly from 4D radar Doppler measurements and transfers them to the LiDAR domain using dynamic-aware association and ambiguity-resolved propagation. On the challenging MAN TruckScenes dataset, DoGFlow substantially outperforms existing self-supervised methods and improves label efficiency by enabling LiDAR backbones to achieve over 90% of fully supervised performance with only 10% of the ground truth data.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 3","pages":"3836-3843"},"PeriodicalIF":5.3,"publicationDate":"2026-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11373844","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146223613","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Robotic Piezo-Assisted Oocyte Penetration and Intracytoplasmic Sperm Injection 机器人压电辅助卵母细胞穿透和胞浆内单精子注射
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-02-09 DOI: 10.1109/LRA.2026.3662617
Yifan Huang;Haoyuan Gu;Ziteng Liu;Chunfeng Yue;Changsheng Dai
Cell penetration and intracellular injection are indispensable procedures in many cell surgery tasks. Because of the complex layered structure of oocytes, conventional piezo-assisted penetration often causesunavoidable cellular damage. Moreover, the small volume of single cells and the nonlinear dynamics involved in the injection process make it challenging to achieve precise and rapid delivery of the injected material within the cytoplasm. This letter presents an integrated robotic system for automated oocyte penetration and intracellular sperm injection that enhances precision and minimizes cellular damage. A deep neural network enables robust segmentation of oocyte structural layers under low-contrast and partially occluded conditions, providing reliable feedback for penetration optimization. A dynamic model was developed to describe the interaction between the pump motion and the fluidic response. A model-predictive controller is then designed to compensate for delays and pressure deviations, ensuring smooth and accurate sperm delivery. Experimental validation demonstrates that the proposed system achieves penetration with a deformation of 4.76 $pm$ 2.34 $mu$m and precise sperm positioning within $pm$ 5 pixels (overshoot below 8 pixels), while ensuring a post-injection survival rate of 82.8%.
细胞穿透和细胞内注射是许多细胞外科手术中不可缺少的步骤。由于卵母细胞复杂的层状结构,传统的压电辅助穿透通常会造成不可避免的细胞损伤。此外,单细胞的小体积和注射过程中涉及的非线性动力学使得在细胞质内实现注射物质的精确和快速递送具有挑战性。这封信介绍了一个集成的机器人系统,用于自动卵母细胞穿透和细胞内精子注射,提高了精度并最大限度地减少了细胞损伤。深度神经网络可以在低对比度和部分遮挡条件下对卵母细胞结构层进行鲁棒分割,为穿透优化提供可靠的反馈。建立了一个动力学模型来描述泵的运动和流体响应之间的相互作用。然后设计一个模型预测控制器来补偿延迟和压力偏差,确保精子顺利准确地输送。实验验证表明,该系统实现了渗透变形4.76下午 μm美元2.34美元内精子和精确定位下午5美元像素(过度低于8像素),同时确保post-injection存活率为82.8%。
{"title":"Robotic Piezo-Assisted Oocyte Penetration and Intracytoplasmic Sperm Injection","authors":"Yifan Huang;Haoyuan Gu;Ziteng Liu;Chunfeng Yue;Changsheng Dai","doi":"10.1109/LRA.2026.3662617","DOIUrl":"https://doi.org/10.1109/LRA.2026.3662617","url":null,"abstract":"Cell penetration and intracellular injection are indispensable procedures in many cell surgery tasks. Because of the complex layered structure of oocytes, conventional piezo-assisted penetration often causesunavoidable cellular damage. Moreover, the small volume of single cells and the nonlinear dynamics involved in the injection process make it challenging to achieve precise and rapid delivery of the injected material within the cytoplasm. This letter presents an integrated robotic system for automated oocyte penetration and intracellular sperm injection that enhances precision and minimizes cellular damage. A deep neural network enables robust segmentation of oocyte structural layers under low-contrast and partially occluded conditions, providing reliable feedback for penetration optimization. A dynamic model was developed to describe the interaction between the pump motion and the fluidic response. A model-predictive controller is then designed to compensate for delays and pressure deviations, ensuring smooth and accurate sperm delivery. Experimental validation demonstrates that the proposed system achieves penetration with a deformation of 4.76 <inline-formula><tex-math>$pm$</tex-math></inline-formula> 2.34 <inline-formula><tex-math>$mu$</tex-math></inline-formula>m and precise sperm positioning within <inline-formula><tex-math>$pm$</tex-math></inline-formula> 5 pixels (overshoot below 8 pixels), while ensuring a post-injection survival rate of 82.8%.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 3","pages":"3764-3771"},"PeriodicalIF":5.3,"publicationDate":"2026-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146223643","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MFE: A Multimodal Hand Exoskeleton With Interactive Force, Pressure and Thermo-Haptic Feedback MFE:具有交互力、压力和热触觉反馈的多模态手外骨骼
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-02-09 DOI: 10.1109/LRA.2026.3662616
Ziyuan Tang;Yitian Guo;Chenxi Xiao
Recent advancements in virtual reality and robotic teleoperation have greatly increased the variety of haptic information that must be conveyed to users. While existing haptic devices typically provide unimodal feedback to enhance situational awareness, a gap remains in their ability to deliver rich, multimodal sensory feedback encompassing force, pressure, and thermal sensations. To address this limitation, we present the Multimodal Feedback Exoskeleton (MFE), a hand exoskeleton designed to deliver hybrid haptic feedback. The MFE features 20 degrees of freedom for capturing hand pose. For force feedback, it employs an active mechanism capable of generating 3.5-8.1 N of pushing and pulling forces, enabling realistic interaction with deformable objects. The fingertips are equipped with flat actuators based on the electro-osmotic principle, providing pressure and vibration stimuli and achieving up to 2.47 kPa of contact pressure to render tactile sensations. For thermal feedback, the MFE integrates thermoelectric heat pumps capable of rendering temperatures from 10 $^circ$C to 55 $^circ$C. We validated the MFE by integrating it into a robotic teleoperation system using the X-Arm 6 and Inspire Hand manipulator. In user studies, participants successfully recognized and manipulated deformable objects and differentiated remote objects with varying temperatures. These results demonstrate that the MFE enhances situational awareness, as well as the usability and transparency of robotic teleoperation systems.
虚拟现实和机器人远程操作的最新进展大大增加了必须传达给用户的触觉信息的种类。虽然现有的触觉设备通常提供单模态反馈来增强态势感知,但它们提供丰富的多模态感官反馈的能力仍然存在差距,包括力、压力和热感觉。为了解决这一限制,我们提出了多模态反馈外骨骼(MFE),一种旨在提供混合触觉反馈的手外骨骼。MFE具有20个捕捉手部姿势的自由度。在力反馈方面,采用主动机构,能够产生3.5-8.1 N的推力和拉力,实现与可变形物体的真实交互。指尖装有基于电渗透原理的平面致动器,提供压力和振动刺激,达到2.47 kPa的接触压力,呈现触觉。对于热反馈,MFE集成了热电热泵,能够呈现温度从10 $^circ$C到55 $^circ$C。我们通过将其集成到使用X-Arm 6和Inspire Hand机械手的机器人远程操作系统中来验证MFE。在用户研究中,参与者成功地识别和操纵可变形的物体,并区分具有不同温度的远程物体。这些结果表明,MFE增强了态势感知能力,以及机器人遥操作系统的可用性和透明度。
{"title":"MFE: A Multimodal Hand Exoskeleton With Interactive Force, Pressure and Thermo-Haptic Feedback","authors":"Ziyuan Tang;Yitian Guo;Chenxi Xiao","doi":"10.1109/LRA.2026.3662616","DOIUrl":"https://doi.org/10.1109/LRA.2026.3662616","url":null,"abstract":"Recent advancements in virtual reality and robotic teleoperation have greatly increased the variety of haptic information that must be conveyed to users. While existing haptic devices typically provide unimodal feedback to enhance situational awareness, a gap remains in their ability to deliver rich, multimodal sensory feedback encompassing force, pressure, and thermal sensations. To address this limitation, we present the <bold>Multimodal Feedback Exoskeleton (MFE)</b>, a hand exoskeleton designed to deliver hybrid haptic feedback. The MFE features 20 degrees of freedom for capturing hand pose. For force feedback, it employs an active mechanism capable of generating 3.5-8.1 N of pushing and pulling forces, enabling realistic interaction with deformable objects. The fingertips are equipped with flat actuators based on the electro-osmotic principle, providing pressure and vibration stimuli and achieving up to 2.47 kPa of contact pressure to render tactile sensations. For thermal feedback, the MFE integrates thermoelectric heat pumps capable of rendering temperatures from 10 <inline-formula><tex-math>$^circ$</tex-math></inline-formula>C to 55 <inline-formula><tex-math>$^circ$</tex-math></inline-formula>C. We validated the MFE by integrating it into a robotic teleoperation system using the X-Arm 6 and Inspire Hand manipulator. In user studies, participants successfully recognized and manipulated deformable objects and differentiated remote objects with varying temperatures. These results demonstrate that the MFE enhances situational awareness, as well as the usability and transparency of robotic teleoperation systems.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 3","pages":"3756-3763"},"PeriodicalIF":5.3,"publicationDate":"2026-02-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146223672","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Robotics and Automation Letters
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1