{"title":"柔性四指抓手的抓取检测方法","authors":"Jianan Liang, Xingrui Bian, Lina Jia, Meiyan Liang, Ruiling Kong, Jinhua Zhang","doi":"10.1016/j.robot.2024.104721","DOIUrl":null,"url":null,"abstract":"<div><p>The flexible four-finger gripper, as a specialized robotic end-effector, is highly valued for its ability to passively adapt to the shape of objects and perform non-destructive grasping. However, the development of grasping detection algorithms for flexible four-finger grippers remains relatively unexplored. This paper addresses the unique characteristics of the flexible four-finger gripper by proposing a grasping detection method based on deep learning. Firstly, the Acute Angle Representation model (AAR-model), which is based on the structure of the flexible four-finger gripper and consists of grasp points and angles, is designed as the grasping representation model that reduces unnecessary rotations of the gripper and improves its versatility in grasping objects. Then, the Flexible Gripper Adaptive Attribute model (FGAA-model) is proposed to represent the grasping attributes of objects, calculate the grasp angles that meet the criteria of the AAR-model, and aggregate the AAR-models on the image data into a unified set, thereby circumventing the time-consuming process of pixel-level annotation. Finally, the Adaptive Grasping Neural Net (AGNN), which is based on Adaptive Feature Fusion and the Grasp Aware Network (AFFGA), is introduced by eliminating redundant network detection headers, fusing color and depth images as inputs, and incorporating a Series Atrous Spatial Pyramid (SASP) structure to produce more accurate grasp poses. Our method not only attains a remarkable accuracy of 97.62% on the Cornell dataset but also swiftly completes grasping detection within 25 ms. In practical robotic arm grasping tests, where a robot is outfitted with a flexible four-finger gripper, it successfully grasps unknown objects with a 96% success rate. These results underscore the reliability and real-time performance of our method, significantly enhancing the gripper's adaptability and precision when handling objects of varying sizes and shapes. This advancement provides a powerful technical solution for robots utilizing flexible four-finger grippers, enabling autonomous, real-time, and highly accurate grasping maneuvers. Moreover, it addresses the persistent challenge of the scarcity of efficient grasping detection techniques tailored for flexible four-finger grippers.</p></div>","PeriodicalId":49592,"journal":{"name":"Robotics and Autonomous Systems","volume":"178 ","pages":"Article 104721"},"PeriodicalIF":4.3000,"publicationDate":"2024-05-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A method for grasp detection of flexible four-finger gripper\",\"authors\":\"Jianan Liang, Xingrui Bian, Lina Jia, Meiyan Liang, Ruiling Kong, Jinhua Zhang\",\"doi\":\"10.1016/j.robot.2024.104721\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>The flexible four-finger gripper, as a specialized robotic end-effector, is highly valued for its ability to passively adapt to the shape of objects and perform non-destructive grasping. However, the development of grasping detection algorithms for flexible four-finger grippers remains relatively unexplored. This paper addresses the unique characteristics of the flexible four-finger gripper by proposing a grasping detection method based on deep learning. Firstly, the Acute Angle Representation model (AAR-model), which is based on the structure of the flexible four-finger gripper and consists of grasp points and angles, is designed as the grasping representation model that reduces unnecessary rotations of the gripper and improves its versatility in grasping objects. Then, the Flexible Gripper Adaptive Attribute model (FGAA-model) is proposed to represent the grasping attributes of objects, calculate the grasp angles that meet the criteria of the AAR-model, and aggregate the AAR-models on the image data into a unified set, thereby circumventing the time-consuming process of pixel-level annotation. Finally, the Adaptive Grasping Neural Net (AGNN), which is based on Adaptive Feature Fusion and the Grasp Aware Network (AFFGA), is introduced by eliminating redundant network detection headers, fusing color and depth images as inputs, and incorporating a Series Atrous Spatial Pyramid (SASP) structure to produce more accurate grasp poses. Our method not only attains a remarkable accuracy of 97.62% on the Cornell dataset but also swiftly completes grasping detection within 25 ms. In practical robotic arm grasping tests, where a robot is outfitted with a flexible four-finger gripper, it successfully grasps unknown objects with a 96% success rate. These results underscore the reliability and real-time performance of our method, significantly enhancing the gripper's adaptability and precision when handling objects of varying sizes and shapes. This advancement provides a powerful technical solution for robots utilizing flexible four-finger grippers, enabling autonomous, real-time, and highly accurate grasping maneuvers. Moreover, it addresses the persistent challenge of the scarcity of efficient grasping detection techniques tailored for flexible four-finger grippers.</p></div>\",\"PeriodicalId\":49592,\"journal\":{\"name\":\"Robotics and Autonomous Systems\",\"volume\":\"178 \",\"pages\":\"Article 104721\"},\"PeriodicalIF\":4.3000,\"publicationDate\":\"2024-05-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Robotics and Autonomous Systems\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0921889024001040\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"AUTOMATION & CONTROL SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Robotics and Autonomous Systems","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0921889024001040","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
A method for grasp detection of flexible four-finger gripper
The flexible four-finger gripper, as a specialized robotic end-effector, is highly valued for its ability to passively adapt to the shape of objects and perform non-destructive grasping. However, the development of grasping detection algorithms for flexible four-finger grippers remains relatively unexplored. This paper addresses the unique characteristics of the flexible four-finger gripper by proposing a grasping detection method based on deep learning. Firstly, the Acute Angle Representation model (AAR-model), which is based on the structure of the flexible four-finger gripper and consists of grasp points and angles, is designed as the grasping representation model that reduces unnecessary rotations of the gripper and improves its versatility in grasping objects. Then, the Flexible Gripper Adaptive Attribute model (FGAA-model) is proposed to represent the grasping attributes of objects, calculate the grasp angles that meet the criteria of the AAR-model, and aggregate the AAR-models on the image data into a unified set, thereby circumventing the time-consuming process of pixel-level annotation. Finally, the Adaptive Grasping Neural Net (AGNN), which is based on Adaptive Feature Fusion and the Grasp Aware Network (AFFGA), is introduced by eliminating redundant network detection headers, fusing color and depth images as inputs, and incorporating a Series Atrous Spatial Pyramid (SASP) structure to produce more accurate grasp poses. Our method not only attains a remarkable accuracy of 97.62% on the Cornell dataset but also swiftly completes grasping detection within 25 ms. In practical robotic arm grasping tests, where a robot is outfitted with a flexible four-finger gripper, it successfully grasps unknown objects with a 96% success rate. These results underscore the reliability and real-time performance of our method, significantly enhancing the gripper's adaptability and precision when handling objects of varying sizes and shapes. This advancement provides a powerful technical solution for robots utilizing flexible four-finger grippers, enabling autonomous, real-time, and highly accurate grasping maneuvers. Moreover, it addresses the persistent challenge of the scarcity of efficient grasping detection techniques tailored for flexible four-finger grippers.
期刊介绍:
Robotics and Autonomous Systems will carry articles describing fundamental developments in the field of robotics, with special emphasis on autonomous systems. An important goal of this journal is to extend the state of the art in both symbolic and sensory based robot control and learning in the context of autonomous systems.
Robotics and Autonomous Systems will carry articles on the theoretical, computational and experimental aspects of autonomous systems, or modules of such systems.