Pub Date : 2025-11-20DOI: 10.1007/s10514-025-10230-7
Peleg Shefi, Amir Ayali, Gal A. Kaminka
In collective motion, perceptually-limited individuals move in an ordered manner, without centralized control. The perception of each individual is highly localized, as is its ability to interact with others. While natural collective motion is robust, most artificial swarms are brittle. This particularly occurs when vision is used as the sensing modality, due to ambiguities and information-loss inherent in visual perception. This paper presents mechanisms for robust collective motion inspired by studies of locusts. First, we develop a robust distance estimation method that combines visually perceived horizontal and vertical sizes of neighbors. Second, we introduce intermittent locomotion as a mechanism that allows robots to reliably detect peers that fail to keep up, and disrupt the motion of the swarm. We show how such faulty robots can be avoided in a manner that is robust to errors in classifying them as faulty. Through extensive physics-based simulation experiments, we show dramatic improvements to swarm resilience when using these techniques. We show these are relevant to both distance-based Avoid–Attract models, as well as to models relying on Alignment, in a wide range of experiment settings.
{"title":"Bugs with features: vision-based fault-tolerant collective motion inspired by nature","authors":"Peleg Shefi, Amir Ayali, Gal A. Kaminka","doi":"10.1007/s10514-025-10230-7","DOIUrl":"10.1007/s10514-025-10230-7","url":null,"abstract":"<div><p>In <i>collective motion</i>, perceptually-limited individuals move in an ordered manner, without centralized control. The perception of each individual is highly localized, as is its ability to interact with others. While natural collective motion is robust, most artificial swarms are <i>brittle</i>. This particularly occurs when vision is used as the sensing modality, due to ambiguities and information-loss inherent in visual perception. This paper presents mechanisms for robust collective motion inspired by studies of locusts. First, we develop a robust distance estimation method that combines visually perceived horizontal and vertical sizes of neighbors. Second, we introduce <i>intermittent locomotion</i> as a mechanism that allows robots to reliably detect peers that fail to keep up, and disrupt the motion of the swarm. We show how such faulty robots can be avoided in a manner that is robust to errors in classifying them as faulty. Through extensive physics-based simulation experiments, we show dramatic improvements to swarm resilience when using these techniques. We show these are relevant to both distance-based <i>Avoid–Attract</i> models, as well as to models relying on <i>Alignment</i>, in a wide range of experiment settings.</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"49 4","pages":""},"PeriodicalIF":4.3,"publicationDate":"2025-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10514-025-10230-7.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145561637","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-18DOI: 10.1007/s10514-025-10217-4
{"title":"Editorial note from the publisher","authors":"","doi":"10.1007/s10514-025-10217-4","DOIUrl":"10.1007/s10514-025-10217-4","url":null,"abstract":"","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"49 4","pages":""},"PeriodicalIF":4.3,"publicationDate":"2025-11-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145560840","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-15DOI: 10.1007/s10514-025-10226-3
Jiahe Chen, Kirstin Petersen
New settlements in remote environments require terrain modification, a task well suited for autonomous multi-robot systems. Simple, robust earthmover robots offer an inexpensive and scalable alternative to sophisticated construction robots. We present a mathematical model for such robots modifying continuous granular structures in 2D and develop both centralized and decentralized planning algorithms to achieve user-defined construction goals. These algorithms decompose long-horizon tasks into subtasks solvable using optimal transport theory and Wasserstein geodesics. Simulations across 100 randomly generated tasks show that a centralized controller with global information achieves on average 85% and 92% construction progress on untraversable and traversable terrains respectively, even with action noise. Multiple robots reduce overall travel distance by 70%, important because motion over the structure also disturbs it. The distributed algorithm—without global information—matches centralized performance on traversable terrain, reaching 93% progress. Increasing robot numbers accelerates convergence, lowers moved material, and raises convergence rates, though congestion can increase total travel distance. These results indicate that simple earthmover robots hold promise for construction tasks ranging from extraterrestrial habitat preparation to coastal protective berms.
{"title":"2D construction planning for swarms of simple earthmover robots","authors":"Jiahe Chen, Kirstin Petersen","doi":"10.1007/s10514-025-10226-3","DOIUrl":"10.1007/s10514-025-10226-3","url":null,"abstract":"<div><p>New settlements in remote environments require terrain modification, a task well suited for autonomous multi-robot systems. Simple, robust earthmover robots offer an inexpensive and scalable alternative to sophisticated construction robots. We present a mathematical model for such robots modifying continuous granular structures in 2D and develop both centralized and decentralized planning algorithms to achieve user-defined construction goals. These algorithms decompose long-horizon tasks into subtasks solvable using optimal transport theory and Wasserstein geodesics. Simulations across 100 randomly generated tasks show that a centralized controller with global information achieves on average 85% and 92% construction progress on untraversable and traversable terrains respectively, even with action noise. Multiple robots reduce overall travel distance by 70%, important because motion over the structure also disturbs it. The distributed algorithm—without global information—matches centralized performance on traversable terrain, reaching 93% progress. Increasing robot numbers accelerates convergence, lowers moved material, and raises convergence rates, though congestion can increase total travel distance. These results indicate that simple earthmover robots hold promise for construction tasks ranging from extraterrestrial habitat preparation to coastal protective berms.</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"49 4","pages":""},"PeriodicalIF":4.3,"publicationDate":"2025-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145560976","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-13DOI: 10.1007/s10514-025-10219-2
Zhichao Li, Yinzhuang Yi, Zhuolin Niu, Nikolay Atanasov
This paper considers the problem of autonomous mobile robot navigation in unknown environments with moving obstacles. We propose a new method to achieve environment-aware safe tracking (EAST) of robot motion plans that integrates an obstacle clearance cost for path planning, a convex reachable set for robot motion prediction, and safety constraints for dynamic obstacle avoidance. EAST adapts the motion of the robot according to the locally sensed environment geometry and dynamics, leading to fast motion in wide open areas and cautious behavior in narrow passages or near moving obstacles. Our control design uses a reference governor, a virtual dynamical system that guides the robot’s motion and decouples the path tracking and safety objectives. While reference governor methods have been used for safe tracking control in static environments, our key contribution is an extension to dynamic environments using convex optimization with control barrier function (CBF) constraints. Thus, our work establishes a connection between reference governor techniques and CBF techniques for safe control in dynamic environments. We validate our approach in simulated and real-world environments, featuring complex obstacle configurations and natural dynamic obstacle motion.
{"title":"EAST: environment-aware safe tracking for robot navigation in dynamic environments","authors":"Zhichao Li, Yinzhuang Yi, Zhuolin Niu, Nikolay Atanasov","doi":"10.1007/s10514-025-10219-2","DOIUrl":"10.1007/s10514-025-10219-2","url":null,"abstract":"<div><p>This paper considers the problem of autonomous mobile robot navigation in unknown environments with moving obstacles. We propose a new method to achieve environment-aware safe tracking (EAST) of robot motion plans that integrates an obstacle clearance cost for path planning, a convex reachable set for robot motion prediction, and safety constraints for dynamic obstacle avoidance. EAST adapts the motion of the robot according to the locally sensed environment geometry and dynamics, leading to fast motion in wide open areas and cautious behavior in narrow passages or near moving obstacles. Our control design uses a reference governor, a virtual dynamical system that guides the robot’s motion and decouples the path tracking and safety objectives. While reference governor methods have been used for safe tracking control in static environments, our key contribution is an extension to dynamic environments using convex optimization with control barrier function (CBF) constraints. Thus, our work establishes a connection between reference governor techniques and CBF techniques for safe control in dynamic environments. We validate our approach in simulated and real-world environments, featuring complex obstacle configurations and natural dynamic obstacle motion.</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"49 4","pages":""},"PeriodicalIF":4.3,"publicationDate":"2025-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145510358","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-12DOI: 10.1007/s10514-025-10206-7
William Yang, Michael Posa
When legged robots impact their environment executing dynamic motions, they undergo large changes in their velocities in a short amount of time. Measuring and applying feedback to these velocities is challenging, further complicated by uncertainty in the impact model and impact timing. This work proposes a general framework for adapting feedback control during impact by projecting the control objectives to a subspace that is invariant to the impact event. The resultant controller is robust to uncertainties in the impact event while maintaining maximum control authority over the impact-invariant subspace. We demonstrate the improved performance of the projection over other commonly used heuristics on a walking controller for a planar five-link-biped. The projection is also applied to jumping, box jumping, and running controllers for the compliant 3D bipedal robot, Cassie. The modification is easily applied to these various controllers and is a critical component to deploying on the physical robot. Code and video of the experiments are available at https://impact-invariant-control.github.io/.
{"title":"Impact-invariant control: maximizing control authority during impacts","authors":"William Yang, Michael Posa","doi":"10.1007/s10514-025-10206-7","DOIUrl":"10.1007/s10514-025-10206-7","url":null,"abstract":"<div><p>When legged robots impact their environment executing dynamic motions, they undergo large changes in their velocities in a short amount of time. Measuring and applying feedback to these velocities is challenging, further complicated by uncertainty in the impact model and impact timing. This work proposes a general framework for adapting feedback control during impact by projecting the control objectives to a subspace that is invariant to the impact event. The resultant controller is robust to uncertainties in the impact event while maintaining maximum control authority over the impact-invariant subspace. We demonstrate the improved performance of the projection over other commonly used heuristics on a walking controller for a planar five-link-biped. The projection is also applied to jumping, box jumping, and running controllers for the compliant 3D bipedal robot, Cassie. The modification is easily applied to these various controllers and is a critical component to deploying on the physical robot. Code and video of the experiments are available at https://impact-invariant-control.github.io/.</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"49 4","pages":""},"PeriodicalIF":4.3,"publicationDate":"2025-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10514-025-10206-7.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145510340","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Maintaining a robust communication network is crucial for the success of multi-robot online task planning. A key capability of such systems is the ability to repair the communication topology in the event of robot failures, thereby ensuring continued coordination. In this paper, we address the Fast k-Connectivity Restoration (FCR) problem, which seeks to restore a network’s k-connectivity with minimal robot movement. Here, a k-connected network refers to a topology that remains connected despite the removal of up to (k-1) nodes. We first formulate the FCR problem as a Quadratically Constrained Program (QCP), which yields optimal solutions but is computationally intractable for large-scale instances. To overcome this limitation, we propose EA-SCR, a scalable algorithm grounded in graph-theoretic principles, which leverages global network information to guide robot movements. Furthermore, we develop a learning-based approach, GNN-EA-SCR, which employs aggregation graph neural networks to learn a decentralized counterpart of EA-SCR, relying solely on local information exchanged among neighboring robots. Through empirical evaluation, we demonstrate that EA-SCR achieves solutions within 10% of the optimal while being orders of magnitude faster. Additionally, EA-SCR surpasses existing methods by 30% in terms of the FCR distance metric. For the learning-based solution, GNN-EA-SCR, we show it attains a success rate exceeding 90% and exhibits comparable maximum robot movement to EA-SCR.
保持一个健壮的通信网络是多机器人在线任务规划成功的关键。这种系统的一个关键能力是在机器人发生故障时修复通信拓扑的能力,从而确保持续的协调。在本文中,我们解决了快速k-连通性恢复(FCR)问题,该问题旨在以最小的机器人运动恢复网络的k-连通性。这里,k连接的网络指的是尽管移除了最多(k-1)节点,但仍保持连接的拓扑结构。我们首先将FCR问题表述为一个二次约束规划(QCP),它产生最优解,但对于大规模实例来说,计算上难以处理。为了克服这一限制,我们提出了EA-SCR,这是一种基于图论原理的可扩展算法,它利用全局网络信息来指导机器人运动。此外,我们开发了一种基于学习的方法,GNN-EA-SCR,它使用聚合图神经网络来学习EA-SCR的分散对应体,仅依赖于相邻机器人之间交换的局部信息。通过实证评估,我们证明EA-SCR在10以内实现了解决方案% of the optimal while being orders of magnitude faster. Additionally, EA-SCR surpasses existing methods by 30% in terms of the FCR distance metric. For the learning-based solution, GNN-EA-SCR, we show it attains a success rate exceeding 90% and exhibits comparable maximum robot movement to EA-SCR.
{"title":"Fast k-connectivity restoration in multi-robot systems for robust communication maintenance: algorithmic and learning-based solutions","authors":"Guangyao Shi, Md Ishat-E-Rabban, Griffin Bonner, Pratap Tokekar","doi":"10.1007/s10514-025-10224-5","DOIUrl":"10.1007/s10514-025-10224-5","url":null,"abstract":"<div><p>Maintaining a robust communication network is crucial for the success of multi-robot online task planning. A key capability of such systems is the ability to repair the communication topology in the event of robot failures, thereby ensuring continued coordination. In this paper, we address the Fast <i>k</i>-Connectivity Restoration (FCR) problem, which seeks to restore a network’s <i>k</i>-connectivity with minimal robot movement. Here, a <i>k</i>-connected network refers to a topology that remains connected despite the removal of up to <span>(k-1)</span> nodes. We first formulate the FCR problem as a Quadratically Constrained Program (QCP), which yields optimal solutions but is computationally intractable for large-scale instances. To overcome this limitation, we propose EA-SCR, a scalable algorithm grounded in graph-theoretic principles, which leverages global network information to guide robot movements. Furthermore, we develop a learning-based approach, GNN-EA-SCR, which employs aggregation graph neural networks to learn a decentralized counterpart of EA-SCR, relying solely on local information exchanged among neighboring robots. Through empirical evaluation, we demonstrate that EA-SCR achieves solutions within 10% of the optimal while being orders of magnitude faster. Additionally, EA-SCR surpasses existing methods by 30% in terms of the FCR distance metric. For the learning-based solution, GNN-EA-SCR, we show it attains a success rate exceeding 90% and exhibits comparable maximum robot movement to EA-SCR. </p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"49 4","pages":""},"PeriodicalIF":4.3,"publicationDate":"2025-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145456445","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-23DOI: 10.1007/s10514-025-10216-5
Erin Hedlund-Botti, Julianna Schalkwyk, Michael Johnson, Matthew Gombolay
Learning from Demonstration (LfD) algorithms seek to enable end-users to teach robots new skills through human demonstration of a task. Previous studies have analyzed how robot failure affects human trust, but not in the context of the human teaching the robot. In this paper, we investigate how human teachers react to robot failure in an LfD setting. We conduct a study in which participants teach a robot how to complete three tasks, using one of three instruction methods, while the robot is pre-programmed to either succeed or fail at the task. We find that when the robot fails, people trust the robot less ((p<.001)) and themselves less ((p=.003)) and they believe that others will trust them less ((p<.001)). Human teachers also have a lower impression of the robot and themselves ((p<.001)) and found the task more difficult when the robot fails ((p<.001)). Motion capture was found to be a less difficult instruction method than teleoperation ((p=.012)), while kinesthetic teaching gave the teachers the lowest impression of themselves compared to teleoperation ((p=.035)) and motion capture ((p=.001)). Importantly, a mediation analysis showed that people’s trust in themselves is heavily mediated by what they think that others—including the robot—think of them ((p<.001)). These results provide valuable insights to improving the human–robot relationship for LfD.
{"title":"The effects of robot learning on human teachers for learning from demonstration","authors":"Erin Hedlund-Botti, Julianna Schalkwyk, Michael Johnson, Matthew Gombolay","doi":"10.1007/s10514-025-10216-5","DOIUrl":"10.1007/s10514-025-10216-5","url":null,"abstract":"<div><p>Learning from Demonstration (LfD) algorithms seek to enable end-users to teach robots new skills through human demonstration of a task. Previous studies have analyzed how robot failure affects human trust, but not in the context of the human teaching the robot. In this paper, we investigate how human teachers react to robot failure in an LfD setting. We conduct a study in which participants teach a robot how to complete three tasks, using one of three instruction methods, while the robot is pre-programmed to either succeed or fail at the task. We find that when the robot fails, people trust the robot less (<span>(p<.001)</span>) and themselves less (<span>(p=.003)</span>) and they believe that others will trust them less (<span>(p<.001)</span>). Human teachers also have a lower impression of the robot and themselves (<span>(p<.001)</span>) and found the task more difficult when the robot fails (<span>(p<.001)</span>). Motion capture was found to be a less difficult instruction method than teleoperation (<span>(p=.012)</span>), while kinesthetic teaching gave the teachers the lowest impression of themselves compared to teleoperation (<span>(p=.035)</span>) and motion capture (<span>(p=.001)</span>). Importantly, a mediation analysis showed that people’s trust in themselves is heavily mediated by what they think that others—including the robot—think of them (<span>(p<.001)</span>). These results provide valuable insights to improving the human–robot relationship for LfD.</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"49 4","pages":""},"PeriodicalIF":4.3,"publicationDate":"2025-10-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10514-025-10216-5.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145352649","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
We introduce GEOTACT, the first robotic system capable of grasping and retrieving objects of potentially unknown shapes buried in a granular environment. While important in many applications, ranging from mining and exploration to search and rescue, this type of interaction with granular media is difficult due to the uncertainty stemming from visual occlusion and noisy contact signals. To address these challenges, we use a learning method relying exclusively on touch feedback, trained end-to-end with simulated sensor noise. We show that our problem formulation leads to the natural emergence of learned pushing behaviors that the manipulator uses to reduce uncertainty and funnel the object to a stable grasp despite spurious and noisy tactile readings. We introduce a training curriculum that bootstraps learning in simulated granular environments, enabling zero-shot transfer to real hardware. Despite being trained only on seven objects with primitive shapes, our method is shown to successfully retrieve 35 different objects, including rigid, deformable, and articulated objects with complex shapes.
{"title":"Tactile-based object retrieval from granular media","authors":"Jingxi Xu, Yinsen Jia, Dongxiao Yang, Patrick Meng, Xinyue Zhu, Zihan Guo, Shuran Song, Matei Ciocarlie","doi":"10.1007/s10514-025-10212-9","DOIUrl":"10.1007/s10514-025-10212-9","url":null,"abstract":"<div><p>We introduce GEOTACT, the first robotic system capable of grasping and retrieving objects of potentially unknown shapes buried in a granular environment. While important in many applications, ranging from mining and exploration to search and rescue, this type of interaction with granular media is difficult due to the uncertainty stemming from visual occlusion and noisy contact signals. To address these challenges, we use a learning method relying exclusively on touch feedback, trained end-to-end with simulated sensor noise. We show that our problem formulation leads to the natural emergence of learned pushing behaviors that the manipulator uses to reduce uncertainty and funnel the object to a stable grasp despite spurious and noisy tactile readings. We introduce a training curriculum that bootstraps learning in simulated granular environments, enabling zero-shot transfer to real hardware. Despite being trained only on seven objects with primitive shapes, our method is shown to successfully retrieve 35 different objects, including rigid, deformable, and articulated objects with complex shapes.</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"49 4","pages":""},"PeriodicalIF":4.3,"publicationDate":"2025-10-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145352648","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-11DOI: 10.1007/s10514-025-10220-9
Ioannis Faros, Herbert G. Tanner
This paper reports an attempt to model the system dynamics and estimate both the unknown internal control input and the state of a recently developed marine autonomous vehicle, the Jaiabot. Although the Jaiabot has shown promise in many applications, process and sensor noise necessitates state estimation and noise filtering. In this work, we present the first surge and heading linear dynamical model for Jaiabots derived from real data collected during field testing. An adaptive input estimation algorithm is implemented to accurately estimate the control input and hence the state. For validation, this approach is compared to the classical Kalman filter, highlighting its advantages in handling unknown control inputs.
{"title":"System identification and adaptive input estimation on the Jaiabot micro autonomous underwater vehicle","authors":"Ioannis Faros, Herbert G. Tanner","doi":"10.1007/s10514-025-10220-9","DOIUrl":"10.1007/s10514-025-10220-9","url":null,"abstract":"<div><p>This paper reports an attempt to model the system dynamics and estimate both the unknown internal control input and the state of a recently developed marine autonomous vehicle, the Jaiabot. Although the Jaiabot has shown promise in many applications, process and sensor noise necessitates state estimation and noise filtering. In this work, we present the first surge and heading linear dynamical model for Jaiabots derived from real data collected during field testing. An adaptive input estimation algorithm is implemented to accurately estimate the control input and hence the state. For validation, this approach is compared to the classical Kalman filter, highlighting its advantages in handling unknown control inputs.</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"49 4","pages":""},"PeriodicalIF":4.3,"publicationDate":"2025-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145256806","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This work addresses the challenge of grasping a target object in cluttered environments, even when it is partially visible or fully occluded. The proposed approach enables the manipulator to learn a sequence of strategic pushing actions that rearrange the scene to make the target object graspable. Our pipeline integrates image morphological processing with deep reinforcement learning (DRL), using GR-ConvNet to predict grasp points for the target. When the object is considered ungraspable, a soft actor-critic (SAC) model guides optimal pushing actions. A novel clutter map is introduced, encoding environmental clutter into a quantitative score that informs the decision-making process. The system shows improved performance with a discount factor ((gamma )) of 0.9, demonstrated through comparative analysis with and without the clutter map. We also compare models trained in discrete versus continuous action spaces to evaluate the impact of action space on DRL effectiveness. The pipeline generalizes well to diverse objects and integrates directly with hardware, requiring no additional training for real-world deployment.
{"title":"Autonomous robotic manipulation for grasping a target object in cluttered environments","authors":"Sanraj Lachhiramka, Pradeep J, Archanaa A. Chandaragi, Arjun Achar, Shikha Tripathi","doi":"10.1007/s10514-025-10214-7","DOIUrl":"10.1007/s10514-025-10214-7","url":null,"abstract":"<div><p>This work addresses the challenge of grasping a target object in cluttered environments, even when it is partially visible or fully occluded. The proposed approach enables the manipulator to learn a sequence of strategic pushing actions that rearrange the scene to make the target object graspable. Our pipeline integrates image morphological processing with deep reinforcement learning (DRL), using GR-ConvNet to predict grasp points for the target. When the object is considered ungraspable, a soft actor-critic (SAC) model guides optimal pushing actions. A novel clutter map is introduced, encoding environmental clutter into a quantitative score that informs the decision-making process. The system shows improved performance with a discount factor (<span>(gamma )</span>) of 0.9, demonstrated through comparative analysis with and without the clutter map. We also compare models trained in discrete versus continuous action spaces to evaluate the impact of action space on DRL effectiveness. The pipeline generalizes well to diverse objects and integrates directly with hardware, requiring no additional training for real-world deployment.</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"49 4","pages":""},"PeriodicalIF":4.3,"publicationDate":"2025-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145256165","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}