首页 > 最新文献

IEEE Robotics and Automation Letters最新文献

英文 中文
Incremental 3D Crop Model Association for Real-Time Counting in Dense Orchards 用于密集果园实时计数的增量三维作物模型关联
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-02-06 DOI: 10.1109/LRA.2026.3662648
Daesung Park;KwangEun Ko;Dongbum Pyo;Jaehyeon Kang
Accurate real-time crop counting is essential for autonomous agricultural systems. However, existing methods often fail in dense plantings due to heavy foliage, irregular planting patterns, and frequent occlusions. While 2D tracking suffers from double-counting and 3D reconstruction requires offline processing, we propose a real-time crop counting framework that incrementally constructs global 3D crop instances during data collection. Each crop is modeled as a 3D oriented bounding box, initialized upon detection and updated with subsequent observations. To ensure robust association across frames, we employ 3D Generalized Intersection over Union (GIoU) for spatial matching and confidence-based filtering for validation, effectively reducing double-counting in dense orchards. Unlike prior methods, our approach supports on-the-fly counting without post-hoc reconstruction and performs reliably in unstructured field conditions. Experimental results demonstrate the accuracy and real-time capability of the proposed system in dense agricultural settings.
准确的实时作物计数对自主农业系统至关重要。然而,由于茂密的植被、不规则的种植模式和频繁的遮挡,现有的方法往往失败。虽然2D跟踪存在重复计数问题,3D重建需要离线处理,但我们提出了一种实时作物计数框架,该框架在数据收集过程中逐步构建全局3D作物实例。每个作物都被建模为一个三维定向的边界框,在检测时进行初始化,并根据后续观察进行更新。为了确保跨帧的鲁棒关联,我们采用3D广义交联(GIoU)进行空间匹配,并基于置信度的过滤进行验证,有效地减少了密集果园中的重复计数。与之前的方法不同,我们的方法支持即时计数,无需事后重建,并且在非结构化的现场条件下可靠地执行。实验结果证明了该系统在密集农业环境下的准确性和实时性。
{"title":"Incremental 3D Crop Model Association for Real-Time Counting in Dense Orchards","authors":"Daesung Park;KwangEun Ko;Dongbum Pyo;Jaehyeon Kang","doi":"10.1109/LRA.2026.3662648","DOIUrl":"https://doi.org/10.1109/LRA.2026.3662648","url":null,"abstract":"Accurate real-time crop counting is essential for autonomous agricultural systems. However, existing methods often fail in dense plantings due to heavy foliage, irregular planting patterns, and frequent occlusions. While 2D tracking suffers from double-counting and 3D reconstruction requires offline processing, we propose a real-time crop counting framework that incrementally constructs global 3D crop instances during data collection. Each crop is modeled as a 3D oriented bounding box, initialized upon detection and updated with subsequent observations. To ensure robust association across frames, we employ 3D Generalized Intersection over Union (GIoU) for spatial matching and confidence-based filtering for validation, effectively reducing double-counting in dense orchards. Unlike prior methods, our approach supports on-the-fly counting without post-hoc reconstruction and performs reliably in unstructured field conditions. Experimental results demonstrate the accuracy and real-time capability of the proposed system in dense agricultural settings.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 3","pages":"3860-3866"},"PeriodicalIF":5.3,"publicationDate":"2026-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146223602","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Contact-Aware Path Planning for Autonomous Neuroendovascular Navigation 自主神经血管内导航的接触感知路径规划
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-02-06 DOI: 10.1109/LRA.2026.3662633
Aabha Tamhankar;Ron Alterovitz;Ajit S. Puri;Giovanni Pittiglio
We propose a deterministic and time-efficient contact-aware path planner for neurovascular navigation. The algorithm leverages information from pre- and intra-operative images of the vessels to navigate pre-bent passive tools, by intelligently predicting and exploiting interactions with the anatomy. A kinematic model is derived and employed by the sampling-based planner for tree expansion that utilizes simplified motion primitives. This approach enables fast computation of the feasible path, with negligible loss in accuracy, as demonstrated in diverse and representative anatomies of the vessels. In these anatomical demonstrators, the algorithm shows a 100% convergence rate within 22.8 s in the worst case, with sub-millimeter tracking errors ($< {0.64},{mathrm{mm}}$), and is found effective on anatomical phantoms representative of $sim$94% of patients.
我们提出了一个确定性和时间高效的接触感知路径规划神经血管导航。该算法通过智能预测和利用与解剖结构的相互作用,利用术前和术中血管图像的信息来导航预弯曲的被动工具。基于采样的规划器利用简化的运动原语,导出了一种运动模型,并将其用于树的展开。这种方法能够快速计算可行路径,精度损失可以忽略不计,正如在不同的和有代表性的血管解剖中所证明的那样。在这些解剖演示中,该算法在最坏的情况下在22.8 s内显示出100%的收敛率,跟踪误差为亚毫米($< {0.64}},{ maththrm {mm}}$),并且被发现对代表$sim$94%的患者的解剖幻影有效。
{"title":"Contact-Aware Path Planning for Autonomous Neuroendovascular Navigation","authors":"Aabha Tamhankar;Ron Alterovitz;Ajit S. Puri;Giovanni Pittiglio","doi":"10.1109/LRA.2026.3662633","DOIUrl":"https://doi.org/10.1109/LRA.2026.3662633","url":null,"abstract":"We propose a deterministic and time-efficient contact-aware path planner for neurovascular navigation. The algorithm leverages information from pre- and intra-operative images of the vessels to navigate pre-bent passive tools, by intelligently predicting and exploiting interactions with the anatomy. A kinematic model is derived and employed by the sampling-based planner for tree expansion that utilizes simplified motion primitives. This approach enables fast computation of the feasible path, with negligible loss in accuracy, as demonstrated in diverse and representative anatomies of the vessels. In these anatomical demonstrators, the algorithm shows a 100% convergence rate within 22.8 s in the worst case, with sub-millimeter tracking errors (<inline-formula><tex-math>$&lt; {0.64},{mathrm{mm}}$</tex-math></inline-formula>), and is found effective on anatomical phantoms representative of <inline-formula><tex-math>$sim$</tex-math></inline-formula>94% of patients.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 4","pages":"4130-4137"},"PeriodicalIF":5.3,"publicationDate":"2026-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146223848","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Design and Analysis of a Robotic Dexterous Hand: Combining Linkage Driven and Pneumatic Actuation 机器人灵巧手的设计与分析:连杆驱动与气动驱动相结合
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-02-06 DOI: 10.1109/LRA.2026.3662585
Yunfan Zhang;Yi Gan
This letter introduces a 10-active degree-of-freedom (DoF) robotic dexterous hand incorporating 4 modular fingers and 1 opposable thumb. Each finger (except the middle one) has 2-active-DoF implemented through 1 active flexion-extension (F-E) Proximal Interphalangeal (PIP) joint, 1 passive F-E Distal Interphalangeal (DIP) joint, and 1 active abduction-adduction (A-A) Metacarpophalangeal (MCP) joint. Specifically, the F-E motion of the PIP joint is actuated by a linear motor, and that of the DIP joint is mechanically coupled through a link. In contrast, the MCP joint’s A-A motion is pneumatically actuated due to its lower force requirements. A mathematical model that captures the chamber wall interactions in three consecutive stages (i.e., before contact, contact initiation, and during contact) is established to relate the pneumatic actuation to the finger A-A motion. Additionally, the proposed thumb has 3-active-DoF driven by 3 separate motors, allowing it to perform opposition movements to the other fingers. In the grasp evaluation, our hand successfully reproduces 7 out of 10 Kapandji test and 25 out of 33 grasps defined by the GRASP taxonomy.
这封信介绍了一个10个主动自由度(DoF)的机器人灵巧手,包括4个模块化手指和1个对生拇指。每个手指(除中指外)通过1个主动屈伸(F-E)近端指间(PIP)关节、1个被动F-E远端指间(DIP)关节和1个主动外展-内收(A-A)掌指关节(MCP)关节实现2个主动自由度。具体来说,PIP关节的F-E运动由直线电机驱动,DIP关节的F-E运动通过一个连杆进行机械耦合。相比之下,MCP关节的A-A运动是气动驱动的,因为它对力的要求较低。建立了一个数学模型,捕捉了三个连续阶段(即接触前、接触开始和接触期间)的腔壁相互作用,将气动驱动与手指A-A运动联系起来。此外,拇指具有3个独立电机驱动的3主动自由度,允许它执行与其他手指相反的运动。在抓握评估中,我们的手成功地再现了10个Kapandji测试中的7个,以及由grasp分类法定义的33个抓握中的25个。
{"title":"Design and Analysis of a Robotic Dexterous Hand: Combining Linkage Driven and Pneumatic Actuation","authors":"Yunfan Zhang;Yi Gan","doi":"10.1109/LRA.2026.3662585","DOIUrl":"https://doi.org/10.1109/LRA.2026.3662585","url":null,"abstract":"This letter introduces a 10-active degree-of-freedom (DoF) robotic dexterous hand incorporating 4 modular fingers and 1 opposable thumb. Each finger (except the middle one) has 2-active-DoF implemented through 1 active flexion-extension (F-E) Proximal Interphalangeal (PIP) joint, 1 passive F-E Distal Interphalangeal (DIP) joint, and 1 active abduction-adduction (A-A) Metacarpophalangeal (MCP) joint. Specifically, the F-E motion of the PIP joint is actuated by a linear motor, and that of the DIP joint is mechanically coupled through a link. In contrast, the MCP joint’s A-A motion is pneumatically actuated due to its lower force requirements. A mathematical model that captures the chamber wall interactions in three consecutive stages (i.e., before contact, contact initiation, and during contact) is established to relate the pneumatic actuation to the finger A-A motion. Additionally, the proposed thumb has 3-active-DoF driven by 3 separate motors, allowing it to perform opposition movements to the other fingers. In the grasp evaluation, our hand successfully reproduces 7 out of 10 Kapandji test and 25 out of 33 grasps defined by the GRASP taxonomy.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 3","pages":"3844-3851"},"PeriodicalIF":5.3,"publicationDate":"2026-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146223746","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-modal sensing in colonoscopy: a data-driven approach. 结肠镜检查中的多模态传感:数据驱动的方法。
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-02-01 Epub Date: 2025-12-18 DOI: 10.1109/lra.2025.3645700
Viola Del Bono, Emma Capaldi, Anushka Kelshiker, Ayhan Aktas, Hiroyuki Aihara, Sheila Russo

Soft optical sensors hold potential for enhancing minimally invasive procedures like colonoscopy, yet their complex, multi-modal responses pose significant challenges. This work introduces a machine learning (ML) framework for real-time estimation of 3D shape and contact force in a soft robotic sleeve for colonoscopy. To overcome limitations of manual calibration and collect large datasets for ML, we developed an automated platform for collecting data across a range of orientations, curvatures, and contact forces. A cascaded ML architecture was implemented for sequential estimation of contact force and 3D shape, enabling an accuracy with errors of 4.7% for curvature, 2.37% for orientation, and 5.5% for force tracking. We also explored the potential of ML for contact localization by training a model to estimate contact intensity and location across 16 indenters distributed along the sleeve. The force intensity was estimated with an error ranging from 0.06 N to 0.31 N throughout the indenters. Despite the proximity of the contact points, the system achieved high localization performances, with 8 indenters reaching over 80% accuracy, demonstrating promising spatial resolution.

软光学传感器具有增强结肠镜检查等微创手术的潜力,但其复杂的多模态响应构成了重大挑战。这项工作介绍了一种机器学习(ML)框架,用于实时估计用于结肠镜检查的软机器人套筒中的3D形状和接触力。为了克服人工校准和收集大型ML数据集的局限性,我们开发了一个自动化平台,用于收集一系列方向,曲率和接触力的数据。采用级联ML架构对接触力和3D形状进行顺序估计,曲率误差为4.7%,方向误差为2.37%,力跟踪误差为5.5%。我们还通过训练一个模型来估计沿套筒分布的16个压头的接触强度和位置,探索了机器学习在接触定位方面的潜力。整个压头的力强度估计误差范围为0.06 N至0.31 N。尽管接触点很接近,但该系统实现了很高的定位性能,8个压痕的精度超过80%,展示了有希望的空间分辨率。
{"title":"Multi-modal sensing in colonoscopy: a data-driven approach.","authors":"Viola Del Bono, Emma Capaldi, Anushka Kelshiker, Ayhan Aktas, Hiroyuki Aihara, Sheila Russo","doi":"10.1109/lra.2025.3645700","DOIUrl":"10.1109/lra.2025.3645700","url":null,"abstract":"<p><p>Soft optical sensors hold potential for enhancing minimally invasive procedures like colonoscopy, yet their complex, multi-modal responses pose significant challenges. This work introduces a machine learning (ML) framework for real-time estimation of 3D shape and contact force in a soft robotic sleeve for colonoscopy. To overcome limitations of manual calibration and collect large datasets for ML, we developed an automated platform for collecting data across a range of orientations, curvatures, and contact forces. A cascaded ML architecture was implemented for sequential estimation of contact force and 3D shape, enabling an accuracy with errors of 4.7% for curvature, 2.37% for orientation, and 5.5% for force tracking. We also explored the potential of ML for contact localization by training a model to estimate contact intensity and location across 16 indenters distributed along the sleeve. The force intensity was estimated with an error ranging from 0.06 N to 0.31 N throughout the indenters. Despite the proximity of the contact points, the system achieved high localization performances, with 8 indenters reaching over 80% accuracy, demonstrating promising spatial resolution.</p>","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 2","pages":"2018-2025"},"PeriodicalIF":5.3,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12811025/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145998055","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IEEE Robotics and Automation Society Information IEEE机器人与自动化学会信息
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-01-28 DOI: 10.1109/LRA.2026.3656038
{"title":"IEEE Robotics and Automation Society Information","authors":"","doi":"10.1109/LRA.2026.3656038","DOIUrl":"https://doi.org/10.1109/LRA.2026.3656038","url":null,"abstract":"","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 2","pages":"C3-C3"},"PeriodicalIF":5.3,"publicationDate":"2026-01-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11367139","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146082026","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IEEE Robotics and Automation Letters Information for Authors IEEE机器人与自动化作者信函信息
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-01-28 DOI: 10.1109/LRA.2026.3656040
{"title":"IEEE Robotics and Automation Letters Information for Authors","authors":"","doi":"10.1109/LRA.2026.3656040","DOIUrl":"https://doi.org/10.1109/LRA.2026.3656040","url":null,"abstract":"","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 2","pages":"C4-C4"},"PeriodicalIF":5.3,"publicationDate":"2026-01-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11367250","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146082025","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HERE: Hierarchical Active Exploration of Radiance Field With Epistemic Uncertainty Minimization 这里:层次主动探索与认知不确定性最小化的辐射场
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-01-23 DOI: 10.1109/LRA.2026.3656771
Taekbeom Lee;Dabin Kim;Youngseok Jang;H. Jin Kim
We present HERE, an active 3D scene reconstruction framework based on neural radiance fields, enabling high-fidelity implicit mapping. Our approach centers around an active learning strategy for camera trajectory generation, driven by accurate identification of unseen regions, which supports efficient data acquisition and precise scene reconstruction. The key to our approach is epistemic uncertainty quantification based on evidential deep learning, which directly captures data insufficiency and exhibits a strong correlation with reconstruction errors. This allows our framework to more reliably identify unexplored or poorly reconstructed regions compared to existing methods, leading to more informed and targeted exploration. Additionally, we design a hierarchical exploration strategy that leverages learned epistemic uncertainty, where local planning extracts target viewpoints from high-uncertainty voxels based on visibility for trajectory generation, and global planning uses uncertainty to guide large-scale coverage for efficient and comprehensive reconstruction. The effectiveness of the proposed method in active 3D reconstruction is demonstrated by achieving higher reconstruction completeness compared to previous approaches on photorealistic simulated scenes across varying scales, while a hardware demonstration further validates its real-world applicability.
我们提出了一个基于神经辐射场的主动3D场景重建框架,实现了高保真隐式映射。我们的方法围绕相机轨迹生成的主动学习策略,通过对未见区域的准确识别来驱动,从而支持高效的数据采集和精确的场景重建。我们方法的关键是基于证据深度学习的认知不确定性量化,它直接捕获数据不足,并与重建误差表现出很强的相关性。与现有方法相比,这使得我们的框架能够更可靠地识别未勘探或重建不良的区域,从而实现更明智、更有针对性的勘探。此外,我们设计了一种利用学习认知不确定性的分层探索策略,其中局部规划基于可见性从高不确定性体素中提取目标视点进行轨迹生成,全局规划利用不确定性指导大规模覆盖以实现高效和全面的重建。通过在不同尺度的逼真模拟场景中实现更高的重建完整性,证明了该方法在主动三维重建中的有效性,而硬件演示进一步验证了其在现实世界中的适用性。
{"title":"HERE: Hierarchical Active Exploration of Radiance Field With Epistemic Uncertainty Minimization","authors":"Taekbeom Lee;Dabin Kim;Youngseok Jang;H. Jin Kim","doi":"10.1109/LRA.2026.3656771","DOIUrl":"https://doi.org/10.1109/LRA.2026.3656771","url":null,"abstract":"We present <italic>HERE</i>, an active 3D scene reconstruction framework based on neural radiance fields, enabling high-fidelity implicit mapping. Our approach centers around an active learning strategy for camera trajectory generation, driven by accurate identification of unseen regions, which supports efficient data acquisition and precise scene reconstruction. The key to our approach is epistemic uncertainty quantification based on evidential deep learning, which directly captures data insufficiency and exhibits a strong correlation with reconstruction errors. This allows our framework to more reliably identify unexplored or poorly reconstructed regions compared to existing methods, leading to more informed and targeted exploration. Additionally, we design a hierarchical exploration strategy that leverages learned epistemic uncertainty, where local planning extracts target viewpoints from high-uncertainty voxels based on visibility for trajectory generation, and global planning uses uncertainty to guide large-scale coverage for efficient and comprehensive reconstruction. The effectiveness of the proposed method in active 3D reconstruction is demonstrated by achieving higher reconstruction completeness compared to previous approaches on photorealistic simulated scenes across varying scales, while a hardware demonstration further validates its real-world applicability.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 3","pages":"3788-3795"},"PeriodicalIF":5.3,"publicationDate":"2026-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146223610","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Taxonomy for Evaluating Generalist Robot Manipulation Policies 评估通才机器人操作策略的分类学
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-01-22 DOI: 10.1109/LRA.2026.3656785
Jensen Gao;Suneel Belkhale;Sudeep Dasari;Ashwin Balakrishna;Dhruv Shah;Dorsa Sadigh
Machine learning for robot manipulation promises to unlock generalization to novel tasks and environments. But how should we measure the progress of these policies towards generalization? Evaluating and quantifying generalization is the Wild West of modern robotics, with each work proposing and measuring different types of generalization in their own, often difficult to reproduce settings. In this work, our goal is (1) to outline the forms of generalization we believe are important for robot manipulation in a comprehensive and fine-grained manner, and (2) to provide reproducible guidelines for measuring these notions of generalization. We first propose $bigstar$-Gen, a taxonomy of generalization for robot manipulation structured around visual, semantic, and behavioral generalization. Next, we instantiate $bigstar$-Gen with two case studies on real-world benchmarking: one based on open-source models and the Bridge V2 dataset, and another based on the bimanual ALOHA 2 platform that covers more dexterous and longer horizon tasks. Our case studies reveal many interesting insights: for example, we observe that open-source vision-language-action models often struggle with semantic generalization, despite pre-training on internet-scale language datasets.
用于机器人操作的机器学习有望解锁对新任务和环境的泛化。但是,我们应该如何衡量这些政策在泛化方面的进展呢?评估和量化泛化是现代机器人的狂野西部,每项工作都在自己的环境中提出和测量不同类型的泛化,通常很难重现。在这项工作中,我们的目标是(1)以全面和细粒度的方式概述我们认为对机器人操作很重要的泛化形式,以及(2)为测量这些泛化概念提供可重复的指导方针。我们首先提出$bigstar$-Gen,这是一个围绕视觉、语义和行为泛化构建的机器人操作泛化分类法。接下来,我们实例化$bigstar$-Gen与两个案例研究在现实世界的基准:一个基于开源模型和Bridge V2数据集,另一个基于双手动ALOHA 2平台,涵盖更灵活和更长远的任务。我们的案例研究揭示了许多有趣的见解:例如,我们观察到,尽管在互联网规模的语言数据集上进行了预训练,开源的视觉语言动作模型经常在语义泛化方面遇到困难。
{"title":"A Taxonomy for Evaluating Generalist Robot Manipulation Policies","authors":"Jensen Gao;Suneel Belkhale;Sudeep Dasari;Ashwin Balakrishna;Dhruv Shah;Dorsa Sadigh","doi":"10.1109/LRA.2026.3656785","DOIUrl":"https://doi.org/10.1109/LRA.2026.3656785","url":null,"abstract":"Machine learning for robot manipulation promises to unlock generalization to novel tasks and environments. But how should we measure the progress of these policies towards generalization? Evaluating and quantifying generalization is the Wild West of modern robotics, with each work proposing and measuring different types of generalization in their own, often difficult to reproduce settings. In this work, our goal is (1) to outline the forms of generalization we believe are important for robot manipulation in a comprehensive and fine-grained manner, and (2) to provide reproducible guidelines for measuring these notions of generalization. We first propose <inline-formula><tex-math>$bigstar$</tex-math></inline-formula>-Gen, a taxonomy of generalization for robot manipulation structured around visual, semantic, and behavioral generalization. Next, we instantiate <inline-formula><tex-math>$bigstar$</tex-math></inline-formula>-Gen with two case studies on real-world benchmarking: one based on open-source models and the Bridge V2 dataset, and another based on the bimanual ALOHA 2 platform that covers more dexterous and longer horizon tasks. Our case studies reveal many interesting insights: for example, we observe that open-source vision-language-action models often struggle with semantic generalization, despite pre-training on internet-scale language datasets.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 3","pages":"3182-3189"},"PeriodicalIF":5.3,"publicationDate":"2026-01-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146082219","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HoLoArm: Deformable Arms for Collision-Tolerant Quadrotor Flight HoLoArm:用于抗碰撞四旋翼飞行的可变形臂
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-01-22 DOI: 10.1109/LRA.2026.3656783
Quang Ngoc Pham;Jonas Eschmann;Yang Zhou;Alejandro Ojeda Olarte;Giuseppe Loianno;Van Anh Ho
The increasing use of drones in human-centric applications highlights the need for designs that can survive collisions and recover rapidly, minimizing risks to both humans and the environment. We present HoLoArm, a quadrotor with compliant arms inspired by the nodus structure of dragonfly wings. This design provides natural flexibility and resilience while preserving flight stability, which is further reinforced by the integration of a Reinforcement Learning (RL) control policy that enhances both recovery and hovering performance. Experimental results demonstrate that HoLoArm can passively deform in any direction, including axial one, and recover within 0.3–0.6 s depending on the direction and level of the impact. The drone can survive collisions at speeds up to 7.6 m/s and carry a 540 g payload while maintaining stable flight. This work contributes to the morphological design of soft aerial robots with high agility and reliable safety, enabling operation in cluttered and human shared environments, and lays the groundwork for future fully soft drones that integrate compliant structures with intelligent control.
无人机在以人为中心的应用中越来越多地使用,这凸显了对能够在碰撞中幸存并快速恢复的设计的需求,从而最大限度地降低了对人类和环境的风险。我们提出HoLoArm,一个四旋翼与柔性臂的灵感来自蜻蜓翅膀的结节结构。这种设计在保持飞行稳定性的同时提供了自然的灵活性和弹性,通过集成强化学习(RL)控制策略进一步加强了这一点,从而提高了恢复和悬停性能。实验结果表明,HoLoArm可以在包括轴向在内的任何方向被动变形,并根据冲击的方向和程度在0.3-0.6 s内恢复。这种无人机可以在速度高达7.6米/秒的碰撞中幸存下来,在保持稳定飞行的同时携带540克的有效载荷。这项工作有助于柔性空中机器人的形态设计,具有高敏捷性和可靠的安全性,使其能够在混乱和人类共享的环境中运行,并为未来将柔性结构与智能控制相结合的全软无人机奠定基础。
{"title":"HoLoArm: Deformable Arms for Collision-Tolerant Quadrotor Flight","authors":"Quang Ngoc Pham;Jonas Eschmann;Yang Zhou;Alejandro Ojeda Olarte;Giuseppe Loianno;Van Anh Ho","doi":"10.1109/LRA.2026.3656783","DOIUrl":"https://doi.org/10.1109/LRA.2026.3656783","url":null,"abstract":"The increasing use of drones in human-centric applications highlights the need for designs that can survive collisions and recover rapidly, minimizing risks to both humans and the environment. We present <italic>HoLoArm</i>, a quadrotor with compliant arms inspired by the <italic>nodus</i> structure of dragonfly wings. This design provides natural flexibility and resilience while preserving flight stability, which is further reinforced by the integration of a Reinforcement Learning (RL) control policy that enhances both recovery and hovering performance. Experimental results demonstrate that <italic>HoLoArm</i> can passively deform in any direction, including axial one, and recover within 0.3–0.6 s depending on the direction and level of the impact. The drone can survive collisions at speeds up to 7.6 m/s and carry a 540 g payload while maintaining stable flight. This work contributes to the morphological design of soft aerial robots with high agility and reliable safety, enabling operation in cluttered and human shared environments, and lays the groundwork for future fully soft drones that integrate compliant structures with intelligent control.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 3","pages":"3582-3589"},"PeriodicalIF":5.3,"publicationDate":"2026-01-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11361075","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146175805","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Energy-Efficient and Powerful Oscillator for Micro-Air Vehicles With Electronics-Free Flapping 无电子扑动微型飞行器的高能效强振子
IF 5.3 2区 计算机科学 Q2 ROBOTICS Pub Date : 2026-01-22 DOI: 10.1109/LRA.2026.3656791
Yongjian Zhao;Yuyan Qi;Jiaqi Shao;Bin Sun;Min Wang;Songyi Zhong;Yang Yang
High power density and energy efficiency are critical for achieving agile locomotion and sustained operation in miniature flapping-wing robots. Here, a pneumatic linear reciprocating oscillator is developed as an actuation solution. The oscillator leverages the Bernoulli principle to establish a positive feedback mechanism through coordinated interactions among a soft membrane, a piston, and the airflow. Experimental validation demonstrates that the oscillator-based flapping-wing robot can generate a lift of 0.43 N to enable take-off and sustained flight in unstructured environments. The minimal oscillation unit exhibits maximum input and output specific power of 710.5 W/kg and 220.7 W/kg, respectively, with peak energy conversion efficiency reaching 41.9% . This design represents a paradigm shift from conventional electromechanical systems, offering two fundamental advancements: (i) simplified robotic drive architectures through an oscillator-based mechanism, and (ii) a foundation for hybrid energy systems that reduce reliance on electricity.
高功率密度和高能量效率是微型扑翼机器人实现敏捷运动和持续运行的关键。本文提出了一种气动直线往复振荡器作为驱动方案。振荡器利用伯努利原理,通过软膜、活塞和气流之间的协调相互作用,建立正反馈机制。实验验证表明,基于振子的扑翼机器人可以产生0.43 N的升力,能够在非结构化环境中起飞和持续飞行。最小振荡单元的最大输入比功率为710.5 W/kg,最大输出比功率为220.7 W/kg,最大能量转换效率达到41.9%。这种设计代表了传统机电系统的范式转变,提供了两个基本的进步:(i)通过基于振荡器的机制简化了机器人驱动架构;(ii)为减少对电力依赖的混合能源系统奠定了基础。
{"title":"An Energy-Efficient and Powerful Oscillator for Micro-Air Vehicles With Electronics-Free Flapping","authors":"Yongjian Zhao;Yuyan Qi;Jiaqi Shao;Bin Sun;Min Wang;Songyi Zhong;Yang Yang","doi":"10.1109/LRA.2026.3656791","DOIUrl":"https://doi.org/10.1109/LRA.2026.3656791","url":null,"abstract":"High power density and energy efficiency are critical for achieving agile locomotion and sustained operation in miniature flapping-wing robots. Here, a pneumatic linear reciprocating oscillator is developed as an actuation solution. The oscillator leverages the Bernoulli principle to establish a positive feedback mechanism through coordinated interactions among a soft membrane, a piston, and the airflow. Experimental validation demonstrates that the oscillator-based flapping-wing robot can generate a lift of 0.43 N to enable take-off and sustained flight in unstructured environments. The minimal oscillation unit exhibits maximum input and output specific power of 710.5 W/kg and 220.7 W/kg, respectively, with peak energy conversion efficiency reaching 41.9% . This design represents a paradigm shift from conventional electromechanical systems, offering two fundamental advancements: (i) simplified robotic drive architectures through an oscillator-based mechanism, and (ii) a foundation for hybrid energy systems that reduce reliance on electricity.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"11 3","pages":"3174-3181"},"PeriodicalIF":5.3,"publicationDate":"2026-01-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146082069","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Robotics and Automation Letters
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1