Shiyao Wang , Xiuping Liu , Charlie C.L. Wang , Jian Liu
{"title":"用于双臂抓握规划的物理感知迭代学习和突出图预测","authors":"Shiyao Wang , Xiuping Liu , Charlie C.L. Wang , Jian Liu","doi":"10.1016/j.cagd.2024.102298","DOIUrl":null,"url":null,"abstract":"<div><p>Learning the skill of human bimanual grasping can extend the capabilities of robotic systems when grasping large or heavy objects. However, it requires a much larger search space for grasp points than single-hand grasping and numerous bimanual grasping annotations for network learning, making both data-driven or analytical grasping methods inefficient and insufficient. We propose a framework for bimanual grasp saliency learning that aims to predict the contact points for bimanual grasping based on existing human single-handed grasping data. We learn saliency corresponding vectors through minimal bimanual contact annotations that establishes correspondences between grasp positions of both hands, capable of eliminating the need for training a large-scale bimanual grasp dataset. The existing single-handed grasp saliency value serves as the initial value for bimanual grasp saliency, and we learn a <em>saliency adjusted score</em> that adds the initial value to obtain the final bimanual grasp saliency value, capable of predicting preferred bimanual grasp positions from single-handed grasp saliency. We also introduce a physics-balance loss function and a physics-aware refinement module that enables physical grasp balance, capable of enhancing the generalization of unknown objects. Comprehensive experiments in simulation and comparisons on dexterous grippers have demonstrated that our method can achieve balanced bimanual grasping effectively.</p></div>","PeriodicalId":55226,"journal":{"name":"Computer Aided Geometric Design","volume":"111 ","pages":"Article 102298"},"PeriodicalIF":1.3000,"publicationDate":"2024-04-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Physics-aware iterative learning and prediction of saliency map for bimanual grasp planning\",\"authors\":\"Shiyao Wang , Xiuping Liu , Charlie C.L. Wang , Jian Liu\",\"doi\":\"10.1016/j.cagd.2024.102298\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>Learning the skill of human bimanual grasping can extend the capabilities of robotic systems when grasping large or heavy objects. However, it requires a much larger search space for grasp points than single-hand grasping and numerous bimanual grasping annotations for network learning, making both data-driven or analytical grasping methods inefficient and insufficient. We propose a framework for bimanual grasp saliency learning that aims to predict the contact points for bimanual grasping based on existing human single-handed grasping data. We learn saliency corresponding vectors through minimal bimanual contact annotations that establishes correspondences between grasp positions of both hands, capable of eliminating the need for training a large-scale bimanual grasp dataset. The existing single-handed grasp saliency value serves as the initial value for bimanual grasp saliency, and we learn a <em>saliency adjusted score</em> that adds the initial value to obtain the final bimanual grasp saliency value, capable of predicting preferred bimanual grasp positions from single-handed grasp saliency. We also introduce a physics-balance loss function and a physics-aware refinement module that enables physical grasp balance, capable of enhancing the generalization of unknown objects. Comprehensive experiments in simulation and comparisons on dexterous grippers have demonstrated that our method can achieve balanced bimanual grasping effectively.</p></div>\",\"PeriodicalId\":55226,\"journal\":{\"name\":\"Computer Aided Geometric Design\",\"volume\":\"111 \",\"pages\":\"Article 102298\"},\"PeriodicalIF\":1.3000,\"publicationDate\":\"2024-04-23\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Computer Aided Geometric Design\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0167839624000323\",\"RegionNum\":4,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"COMPUTER SCIENCE, SOFTWARE ENGINEERING\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computer Aided Geometric Design","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0167839624000323","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, SOFTWARE ENGINEERING","Score":null,"Total":0}
Physics-aware iterative learning and prediction of saliency map for bimanual grasp planning
Learning the skill of human bimanual grasping can extend the capabilities of robotic systems when grasping large or heavy objects. However, it requires a much larger search space for grasp points than single-hand grasping and numerous bimanual grasping annotations for network learning, making both data-driven or analytical grasping methods inefficient and insufficient. We propose a framework for bimanual grasp saliency learning that aims to predict the contact points for bimanual grasping based on existing human single-handed grasping data. We learn saliency corresponding vectors through minimal bimanual contact annotations that establishes correspondences between grasp positions of both hands, capable of eliminating the need for training a large-scale bimanual grasp dataset. The existing single-handed grasp saliency value serves as the initial value for bimanual grasp saliency, and we learn a saliency adjusted score that adds the initial value to obtain the final bimanual grasp saliency value, capable of predicting preferred bimanual grasp positions from single-handed grasp saliency. We also introduce a physics-balance loss function and a physics-aware refinement module that enables physical grasp balance, capable of enhancing the generalization of unknown objects. Comprehensive experiments in simulation and comparisons on dexterous grippers have demonstrated that our method can achieve balanced bimanual grasping effectively.
期刊介绍:
The journal Computer Aided Geometric Design is for researchers, scholars, and software developers dealing with mathematical and computational methods for the description of geometric objects as they arise in areas ranging from CAD/CAM to robotics and scientific visualization. The journal publishes original research papers, survey papers and with quick editorial decisions short communications of at most 3 pages. The primary objects of interest are curves, surfaces, and volumes such as splines (NURBS), meshes, subdivision surfaces as well as algorithms to generate, analyze, and manipulate them. This journal will report on new developments in CAGD and its applications, including but not restricted to the following:
-Mathematical and Geometric Foundations-
Curve, Surface, and Volume generation-
CAGD applications in Numerical Analysis, Computational Geometry, Computer Graphics, or Computer Vision-
Industrial, medical, and scientific applications.
The aim is to collect and disseminate information on computer aided design in one journal. To provide the user community with methods and algorithms for representing curves and surfaces. To illustrate computer aided geometric design by means of interesting applications. To combine curve and surface methods with computer graphics. To explain scientific phenomena by means of computer graphics. To concentrate on the interaction between theory and application. To expose unsolved problems of the practice. To develop new methods in computer aided geometry.