Pub Date : 2023-05-30DOI: 10.1007/s10514-023-10104-w
Baskın Şenbaşlar, Wolfgang Hönig, Nora Ayanian
Trajectory planning for multiple robots in shared environments is a challenging problem especially when there is limited communication available or no central entity. In this article, we present Real-time planning using Linear Spatial Separations, or RLSS: a real-time decentralized trajectory planning algorithm for cooperative multi-robot teams in static environments. The algorithm requires relatively few robot capabilities, namely sensing the positions of robots and obstacles without higher-order derivatives and the ability of distinguishing robots from obstacles. There is no communication requirement and the robots’ dynamic limits are taken into account. RLSS generates and solves convex quadratic optimization problems that are kinematically feasible and guarantees collision avoidance if the resulting problems are feasible. We demonstrate the algorithm’s performance in real-time in simulations and on physical robots. We compare RLSS to two state-of-the-art planners and show empirically that RLSS does avoid deadlocks and collisions in forest-like and maze-like environments, significantly improving prior work, which result in collisions and deadlocks in such environments.
{"title":"RLSS: real-time, decentralized, cooperative, networkless multi-robot trajectory planning using linear spatial separations","authors":"Baskın Şenbaşlar, Wolfgang Hönig, Nora Ayanian","doi":"10.1007/s10514-023-10104-w","DOIUrl":"10.1007/s10514-023-10104-w","url":null,"abstract":"<div><p>Trajectory planning for multiple robots in shared environments is a challenging problem especially when there is limited communication available or no central entity. In this article, we present Real-time planning using Linear Spatial Separations, or RLSS: a real-time decentralized trajectory planning algorithm for cooperative multi-robot teams in static environments. The algorithm requires relatively few robot capabilities, namely sensing the positions of robots and obstacles without higher-order derivatives and the ability of distinguishing robots from obstacles. There is no communication requirement and the robots’ dynamic limits are taken into account. RLSS generates and solves convex quadratic optimization problems that are kinematically feasible and guarantees collision avoidance if the resulting problems are feasible. We demonstrate the algorithm’s performance in real-time in simulations and on physical robots. We compare RLSS to two state-of-the-art planners and show empirically that RLSS does avoid deadlocks and collisions in forest-like and maze-like environments, significantly improving prior work, which result in collisions and deadlocks in such environments.</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"47 7","pages":"921 - 946"},"PeriodicalIF":3.5,"publicationDate":"2023-05-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10514-023-10104-w.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135478964","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Trajectory generation for biped robots is very complex due to the challenge posed by real-world uneven terrain. To address this complexity, this paper proposes a data-driven Gait model that can handle continuously changing conditions. Data-driven approaches are used to incorporate the joint relationships. Therefore, the deep learning methods are employed to develop seven different data-driven models, namely DNN, LSTM, GRU, BiLSTM, BiGRU, LSTM+GRU, and BiLSTM+BiGRU. The dataset used for training the Gait model consists of walking data from 10 able subjects on continuously changing inclines and speeds. The objective function incorporates the standard error from the inter-subject mean trajectory to guide the Gait model to not accurately follow the high variance points in the gait cycle, which helps in providing a smooth and continuous gait cycle. The results show that the proposed Gait models outperform the traditional finite state machine (FSM) and Basis models in terms of mean and maximum error summary statistics. In particular, the LSTM+GRU-based Gait model provides the best performance compared to other data-driven models.
{"title":"Data-driven gait model for bipedal locomotion over continuous changing speeds and inclines","authors":"Bharat Singh, Suchit Patel, Ankit Vijayvargiya, Rajesh Kumar","doi":"10.1007/s10514-023-10108-6","DOIUrl":"10.1007/s10514-023-10108-6","url":null,"abstract":"<div><p>Trajectory generation for biped robots is very complex due to the challenge posed by real-world uneven terrain. To address this complexity, this paper proposes a data-driven Gait model that can handle continuously changing conditions. Data-driven approaches are used to incorporate the joint relationships. Therefore, the deep learning methods are employed to develop seven different data-driven models, namely DNN, LSTM, GRU, BiLSTM, BiGRU, LSTM+GRU, and BiLSTM+BiGRU. The dataset used for training the Gait model consists of walking data from 10 able subjects on continuously changing inclines and speeds. The objective function incorporates the standard error from the inter-subject mean trajectory to guide the Gait model to not accurately follow the high variance points in the gait cycle, which helps in providing a smooth and continuous gait cycle. The results show that the proposed Gait models outperform the traditional finite state machine (FSM) and Basis models in terms of mean and maximum error summary statistics. In particular, the LSTM+GRU-based Gait model provides the best performance compared to other data-driven models.</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"47 6","pages":"753 - 769"},"PeriodicalIF":3.5,"publicationDate":"2023-05-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46959244","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-05-24DOI: 10.1007/s10514-023-10105-9
Sheng Cheng, Derek A. Paley
Monitoring and controlling a large-scale spatiotemporal process can be costly and dangerous for human operators, which can delegate the task to mobile robots for improved efficiency at a lower cost. The complex evolution of the spatiotemporal process and limited onboard resources of the robots motivate a holistic design of the robots’ actions to complete the tasks efficiently. This paper describes a cooperative framework for estimating and controlling a spatiotemporal process using a team of mobile robots that have limited onboard resources. We model the spatiotemporal process as a 2D diffusion equation that can characterize the intrinsic dynamics of the process with a partial differential equation (PDE). Measurement and actuation of the diffusion process are performed by mobile robots carrying sensors and actuators. The core of the framework is a nonlinear optimization problem, that simultaneously seeks the actuation and guidance of the robots to control the spatiotemporal process subject to the PDE dynamics. The limited onboard resources are formulated as inequality constraints on the actuation and speed of the robots. Extensive numerical studies analyze and evaluate the proposed framework using nondimensionalization and compare the optimal strategy to baseline strategies. The framework is demonstrated on an outdoor multi-quadrotor testbed using hardware-in-the-loop simulations.
{"title":"Cooperative estimation and control of a diffusion-based spatiotemporal process using mobile sensors and actuators","authors":"Sheng Cheng, Derek A. Paley","doi":"10.1007/s10514-023-10105-9","DOIUrl":"10.1007/s10514-023-10105-9","url":null,"abstract":"<div><p>Monitoring and controlling a large-scale spatiotemporal process can be costly and dangerous for human operators, which can delegate the task to mobile robots for improved efficiency at a lower cost. The complex evolution of the spatiotemporal process and limited onboard resources of the robots motivate a holistic design of the robots’ actions to complete the tasks efficiently. This paper describes a cooperative framework for estimating and controlling a spatiotemporal process using a team of mobile robots that have limited onboard resources. We model the spatiotemporal process as a 2D diffusion equation that can characterize the intrinsic dynamics of the process with a partial differential equation (PDE). Measurement and actuation of the diffusion process are performed by mobile robots carrying sensors and actuators. The core of the framework is a nonlinear optimization problem, that simultaneously seeks the actuation and guidance of the robots to control the spatiotemporal process subject to the PDE dynamics. The limited onboard resources are formulated as inequality constraints on the actuation and speed of the robots. Extensive numerical studies analyze and evaluate the proposed framework using nondimensionalization and compare the optimal strategy to baseline strategies. The framework is demonstrated on an outdoor multi-quadrotor testbed using hardware-in-the-loop simulations.</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"47 6","pages":"715 - 731"},"PeriodicalIF":3.5,"publicationDate":"2023-05-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44384325","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-04-27DOI: 10.1007/s10514-023-10095-8
Gunhild Elisabeth Berget, Jo Eidsvik, Morten Omholt Alver, Tor Arne Johansen
Discharge of mine tailings significantly impacts the ecological status of the sea. Methods to efficiently monitor the extent of dispersion is essential to protect sensitive areas. By combining underwater robotic sampling with ocean models, we can choose informative sampling sites and adaptively change the robot’s path based on in situ measurements to optimally map the tailings distribution near a seafill. This paper creates a stochastic spatio-temporal proxy model of dispersal dynamics using training data from complex numerical models. The proxy model consists of a spatio-temporal Gaussian process model based on an advection–diffusion stochastic partial differential equation. Informative sampling sites are chosen based on predictions from the proxy model using an objective function favoring areas with high uncertainty and high expected tailings concentrations. A simulation study and data from real-life experiments are presented.
{"title":"Dynamic stochastic modeling for adaptive sampling of environmental variables using an AUV","authors":"Gunhild Elisabeth Berget, Jo Eidsvik, Morten Omholt Alver, Tor Arne Johansen","doi":"10.1007/s10514-023-10095-8","DOIUrl":"10.1007/s10514-023-10095-8","url":null,"abstract":"<div><p>Discharge of mine tailings significantly impacts the ecological status of the sea. Methods to efficiently monitor the extent of dispersion is essential to protect sensitive areas. By combining underwater robotic sampling with ocean models, we can choose informative sampling sites and adaptively change the robot’s path based on in situ measurements to optimally map the tailings distribution near a seafill. This paper creates a stochastic spatio-temporal proxy model of dispersal dynamics using training data from complex numerical models. The proxy model consists of a spatio-temporal Gaussian process model based on an advection–diffusion stochastic partial differential equation. Informative sampling sites are chosen based on predictions from the proxy model using an objective function favoring areas with high uncertainty and high expected tailings concentrations. A simulation study and data from real-life experiments are presented.\u0000</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"47 4","pages":"483 - 502"},"PeriodicalIF":3.5,"publicationDate":"2023-04-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10514-023-10095-8.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45324722","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-04-27DOI: 10.1007/s10514-023-10102-y
Steven Adams, Daniel Jarne Ornia, Manuel Mazo Jr
We present a biologically inspired design for swarm foraging based on ant’s pheromone deployment, where the swarm is assumed to have very restricted capabilities. The robots do not require global or relative position measurements and the swarm is fully decentralized and needs no infrastructure in place. Additionally, the system only requires one-hop communication over the robot network, we do not make any assumptions about the connectivity of the communication graph and the transmission of information and computation is scalable versus the number of agents. This is done by letting the agents in the swarm act as foragers or as guiding agents (beacons). We present experimental results computed for a swarm of Elisa-3 robots on a simulator, and show how the swarm self-organizes to solve a foraging problem over an unknown environment, converging to trajectories around the shortest path, and test the approach on a real swarm of Elisa-3 robots. At last, we discuss the limitations of such a system and propose how the foraging efficiency can be increased.
{"title":"A self-guided approach for navigation in a minimalistic foraging robotic swarm","authors":"Steven Adams, Daniel Jarne Ornia, Manuel Mazo Jr","doi":"10.1007/s10514-023-10102-y","DOIUrl":"10.1007/s10514-023-10102-y","url":null,"abstract":"<div><p>We present a biologically inspired design for swarm foraging based on ant’s pheromone deployment, where the swarm is assumed to have very restricted capabilities. The robots do not require global or relative position measurements and the swarm is fully decentralized and needs no infrastructure in place. Additionally, the system only requires one-hop communication over the robot network, we do not make any assumptions about the connectivity of the communication graph and the transmission of information and computation is scalable versus the number of agents. This is done by letting the agents in the swarm act as foragers or as guiding agents (beacons). We present experimental results computed for a swarm of Elisa-3 robots on a simulator, and show how the swarm self-organizes to solve a foraging problem over an unknown environment, converging to trajectories around the shortest path, and test the approach on a real swarm of Elisa-3 robots. At last, we discuss the limitations of such a system and propose how the foraging efficiency can be increased.</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"47 7","pages":"905 - 920"},"PeriodicalIF":3.5,"publicationDate":"2023-04-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10514-023-10102-y.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43468383","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-04-25DOI: 10.1007/s10514-023-10103-x
Hao-Yun Chen, Pei-Han Huang, Li-Chen Fu
This paper propose a hierarchical path planning algorithm that first captures the local crowd movement around the robot using RGB camera combined with LiDAR and predicts the movement of people nearby the robot, and then generates appropriate global path for the robot using the global path planner with the crowd information. After deciding the global path, the low-level control system receives the prediction results of the crowd and high-level global path, and generates the actual speed control commands for the robot after considering the social norms. With the high accuracy of computer vision for human recognition and the high precision of LiDAR, the system is able to accurately track the surrounding human locations. Through high-level path planning, the robot can use different movement strategies in different scenarios, while the crowd prediction allows the robot to generate more efficient and socially acceptable paths. With this system, even in a highly dynamic environment caused by the crowd, the robot can still plan an appropriate path reach the destination without causing psychological discomfort to others successfully.
{"title":"Social crowd navigation of a mobile robot based on human trajectory prediction and hybrid sensing","authors":"Hao-Yun Chen, Pei-Han Huang, Li-Chen Fu","doi":"10.1007/s10514-023-10103-x","DOIUrl":"10.1007/s10514-023-10103-x","url":null,"abstract":"<div><p>This paper propose a hierarchical path planning algorithm that first captures the local crowd movement around the robot using RGB camera combined with LiDAR and predicts the movement of people nearby the robot, and then generates appropriate global path for the robot using the global path planner with the crowd information. After deciding the global path, the low-level control system receives the prediction results of the crowd and high-level global path, and generates the actual speed control commands for the robot after considering the social norms. With the high accuracy of computer vision for human recognition and the high precision of LiDAR, the system is able to accurately track the surrounding human locations. Through high-level path planning, the robot can use different movement strategies in different scenarios, while the crowd prediction allows the robot to generate more efficient and socially acceptable paths. With this system, even in a highly dynamic environment caused by the crowd, the robot can still plan an appropriate path reach the destination without causing psychological discomfort to others successfully.</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"47 4","pages":"339 - 351"},"PeriodicalIF":3.5,"publicationDate":"2023-04-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10514-023-10103-x.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"45514911","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-04-19DOI: 10.1007/s10514-023-10089-6
Tahiya Salam, M. Ani Hsieh
This paper presents a framework to enable a team of heterogeneous mobile robots to model and sense a multiscale system. We propose a coupled strategy, where robots of one type collect high-fidelity measurements at a slow time scale and robots of another type collect low-fidelity measurements at a fast time scale, for the purpose of fusing measurements together. The multiscale measurements are fused to create a model of a complex, nonlinear spatiotemporal process. The model helps determine optimal sensing locations and predict the evolution of the process. Key contributions are: (i) consolidation of multiple types of data into one cohesive model, (ii) fast determination of optimal sensing locations for mobile robots, and (iii) adaptation of models online for various monitoring scenarios. We illustrate the proposed framework by modeling and predicting the evolution of an artificial plasma cloud. We test our approach using physical marine robots adaptively sampling a process in a water tank.
{"title":"Heterogeneous robot teams for modeling and prediction of multiscale environmental processes","authors":"Tahiya Salam, M. Ani Hsieh","doi":"10.1007/s10514-023-10089-6","DOIUrl":"10.1007/s10514-023-10089-6","url":null,"abstract":"<div><p>This paper presents a framework to enable a team of heterogeneous mobile robots to model and sense a multiscale system. We propose a coupled strategy, where robots of one type collect high-fidelity measurements at a slow time scale and robots of another type collect low-fidelity measurements at a fast time scale, for the purpose of fusing measurements together. The multiscale measurements are fused to create a model of a complex, nonlinear spatiotemporal process. The model helps determine optimal sensing locations and predict the evolution of the process. Key contributions are: (i) consolidation of multiple types of data into one cohesive model, (ii) fast determination of optimal sensing locations for mobile robots, and (iii) adaptation of models online for various monitoring scenarios. We illustrate the proposed framework by modeling and predicting the evolution of an artificial plasma cloud. We test our approach using physical marine robots adaptively sampling a process in a water tank.</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"47 4","pages":"353 - 376"},"PeriodicalIF":3.5,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49418994","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-04-12DOI: 10.1007/s10514-023-10094-9
Eric Heiden, Miles Macklin, Yashraj Narang, Dieter Fox, Animesh Garg, Fabio Ramos
Robotic cutting of soft materials is critical for applications such as food processing, household automation, and surgical manipulation. As in other areas of robotics, simulators can facilitate controller verification, policy learning, and dataset generation. Moreover, differentiable simulators can enable gradient-based optimization, which is invaluable for calibrating simulation parameters and optimizing controllers. In this work, we present DiSECt: the first differentiable simulator for cutting soft materials. The simulator augments the finite element method with a continuous contact model based on signed distance fields, as well as a continuous damage model that inserts springs on opposite sides of the cutting plane and allows them to weaken until zero stiffness, enabling crack formation. Through various experiments, we evaluate the performance of the simulator. We first show that the simulator can be calibrated to match resultant forces and deformation fields from a state-of-the-art commercial solver and real-world cutting datasets, with generality across cutting velocities and object instances. We then show that Bayesian inference can be performed efficiently by leveraging the differentiability of the simulator, estimating posteriors over hundreds of parameters in a fraction of the time of derivative-free methods. Next, we illustrate that control parameters in the simulation can be optimized to minimize cutting forces via lateral slicing motions. Finally, we conduct experiments on a real robot arm equipped with a slicing knife to infer simulation parameters from force measurements. By optimizing the slicing motion of the knife, we show on fruit cutting scenarios that the average knife force can be reduced by more than (40%) compared to a vertical cutting motion. We publish code and additional materials on our project website at https://diff-cutting-sim.github.io.
{"title":"DiSECt: a differentiable simulator for parameter inference and control in robotic cutting","authors":"Eric Heiden, Miles Macklin, Yashraj Narang, Dieter Fox, Animesh Garg, Fabio Ramos","doi":"10.1007/s10514-023-10094-9","DOIUrl":"10.1007/s10514-023-10094-9","url":null,"abstract":"<div><p>Robotic cutting of soft materials is critical for applications such as food processing, household automation, and surgical manipulation. As in other areas of robotics, simulators can facilitate controller verification, policy learning, and dataset generation. Moreover, <i>differentiable</i> simulators can enable gradient-based optimization, which is invaluable for calibrating simulation parameters and optimizing controllers. In this work, we present DiSECt: the first differentiable simulator for cutting soft materials. The simulator augments the finite element method with a continuous contact model based on signed distance fields, as well as a continuous damage model that inserts springs on opposite sides of the cutting plane and allows them to weaken until zero stiffness, enabling crack formation. Through various experiments, we evaluate the performance of the simulator. We first show that the simulator can be calibrated to match resultant forces and deformation fields from a state-of-the-art commercial solver and real-world cutting datasets, with generality across cutting velocities and object instances. We then show that Bayesian inference can be performed efficiently by leveraging the differentiability of the simulator, estimating posteriors over hundreds of parameters in a fraction of the time of derivative-free methods. Next, we illustrate that control parameters in the simulation can be optimized to minimize cutting forces via lateral slicing motions. Finally, we conduct experiments on a real robot arm equipped with a slicing knife to infer simulation parameters from force measurements. By optimizing the slicing motion of the knife, we show on fruit cutting scenarios that the average knife force can be reduced by more than <span>(40%)</span> compared to a vertical cutting motion. We publish code and additional materials on our project website at https://diff-cutting-sim.github.io.\u0000</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"47 5","pages":"549 - 578"},"PeriodicalIF":3.5,"publicationDate":"2023-04-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10514-023-10094-9.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48242007","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-04-12DOI: 10.1007/s10514-023-10096-7
Rahaf Rahal, Amir M. Ghalamzan-E., Firas Abi-Farraj, Claudio Pacchierotti, Paolo Robuffo Giordano
Teleoperating robotic manipulators can be complicated and cognitively demanding for the human operator. Despite these difficulties, teleoperated robotic systems are still popular in several industrial applications, e.g., remote handling of hazardous material. In this context, we present a novel haptic shared control method for minimising the manipulator torque effort during remote manipulative actions in which an operator is assisted in selecting a suitable grasping pose for then displacing an object along a desired trajectory. Minimising torque is important because it reduces the system operating cost and extends the range of objects that can be manipulated. We demonstrate the effectiveness of the proposed approach in a series of representative real-world pick-and-place experiments as well as in a human subjects study. The reported results prove the effectiveness of our shared control vs. a standard teleoperation approach. We also find that haptic-only guidance performs better than visual-only guidance, although combining them together leads to the best overall results.
{"title":"Haptic-guided grasping to minimise torque effort during robotic telemanipulation","authors":"Rahaf Rahal, Amir M. Ghalamzan-E., Firas Abi-Farraj, Claudio Pacchierotti, Paolo Robuffo Giordano","doi":"10.1007/s10514-023-10096-7","DOIUrl":"10.1007/s10514-023-10096-7","url":null,"abstract":"<div><p>Teleoperating robotic manipulators can be complicated and cognitively demanding for the human operator. Despite these difficulties, teleoperated robotic systems are still popular in several industrial applications, e.g., remote handling of hazardous material. In this context, we present a novel haptic shared control method for minimising the manipulator torque effort during remote manipulative actions in which an operator is assisted in selecting a suitable grasping pose for then displacing an object along a desired trajectory. Minimising torque is important because it reduces the system operating cost and extends the range of objects that can be manipulated. We demonstrate the effectiveness of the proposed approach in a series of representative real-world pick-and-place experiments as well as in a human subjects study. The reported results prove the effectiveness of our shared control vs. a standard teleoperation approach. We also find that haptic-only guidance performs better than visual-only guidance, although combining them together leads to the best overall results.\u0000</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"47 4","pages":"405 - 423"},"PeriodicalIF":3.5,"publicationDate":"2023-04-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48097502","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-04-11DOI: 10.1007/s10514-023-10101-z
Dimitrios Dimou, José Santos-Victor, Plinio Moreno
We develop a conditional generative model to represent dexterous grasp postures of a robotic hand and use it to generate in-hand regrasp trajectories. Our model learns to encode the robotic grasp postures into a low-dimensional space, called Synergy Space, while taking into account additional information about the object such as its size and its shape category. We then generate regrasp trajectories through linear interpolation in this low-dimensional space. The result is that the hand configuration moves from one grasp type to another while keeping the object stable in the hand. We show that our model achieves higher success rate on in-hand regrasping compared to previous methods used for synergy extraction, by taking advantage of the grasp size conditional variable.
{"title":"Robotic hand synergies for in-hand regrasping driven by object information","authors":"Dimitrios Dimou, José Santos-Victor, Plinio Moreno","doi":"10.1007/s10514-023-10101-z","DOIUrl":"10.1007/s10514-023-10101-z","url":null,"abstract":"<div><p>We develop a conditional generative model to represent dexterous grasp postures of a robotic hand and use it to generate in-hand regrasp trajectories. Our model learns to encode the robotic grasp postures into a low-dimensional space, called Synergy Space, while taking into account additional information about the object such as its size and its shape category. We then generate regrasp trajectories through linear interpolation in this low-dimensional space. The result is that the hand configuration moves from one grasp type to another while keeping the object stable in the hand. We show that our model achieves higher success rate on in-hand regrasping compared to previous methods used for synergy extraction, by taking advantage of the grasp size conditional variable.</p></div>","PeriodicalId":55409,"journal":{"name":"Autonomous Robots","volume":"47 4","pages":"453 - 464"},"PeriodicalIF":3.5,"publicationDate":"2023-04-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10514-023-10101-z.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46694831","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}