{"title":"一种有效的6D物体姿态估计颜色和几何特征融合模块","authors":"Jiangeng Li, Hong Liu, Gao Huang, Guoyu Zuo","doi":"10.1109/CYBER55403.2022.9907032","DOIUrl":null,"url":null,"abstract":"6D pose estimation is widely used in robot tasks such as sorting and grasping. RGB-D-based methods have recently attained brilliant success, but they are still susceptible to heavy occlusion. Our critical insight is that color and geometry information in RGBD images are two complementary data, and the crux of the pose estimation problem under occlusion is fully leveraging them. Towards this end, we propose a new color and geometry feature fusion module that can efficiently leverage two complementary data sources from RGB-D images. Unlike prior fusion methods, we conduct a two-stage fusion strategy to do color-depth fusion and local-global fusion successively. Specifically, we fuse the color features extracted from RGB images into the point cloud in the first stage. In the second stage, we extract local and global features from the fused point cloud using an ASSANet-like network and splice them together to obtain the final fusion features. We conducted experiments on the widely used LineMod and YCB-Video datasets, which shows that our method improves the prediction accuracy while reducing the training time.","PeriodicalId":34110,"journal":{"name":"IET Cybersystems and Robotics","volume":"23 1","pages":"83-88"},"PeriodicalIF":1.5000,"publicationDate":"2022-07-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"An Efficient Color and Geometric Feature Fusion Module for 6D Object Pose Estiamtion\",\"authors\":\"Jiangeng Li, Hong Liu, Gao Huang, Guoyu Zuo\",\"doi\":\"10.1109/CYBER55403.2022.9907032\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"6D pose estimation is widely used in robot tasks such as sorting and grasping. RGB-D-based methods have recently attained brilliant success, but they are still susceptible to heavy occlusion. Our critical insight is that color and geometry information in RGBD images are two complementary data, and the crux of the pose estimation problem under occlusion is fully leveraging them. Towards this end, we propose a new color and geometry feature fusion module that can efficiently leverage two complementary data sources from RGB-D images. Unlike prior fusion methods, we conduct a two-stage fusion strategy to do color-depth fusion and local-global fusion successively. Specifically, we fuse the color features extracted from RGB images into the point cloud in the first stage. In the second stage, we extract local and global features from the fused point cloud using an ASSANet-like network and splice them together to obtain the final fusion features. We conducted experiments on the widely used LineMod and YCB-Video datasets, which shows that our method improves the prediction accuracy while reducing the training time.\",\"PeriodicalId\":34110,\"journal\":{\"name\":\"IET Cybersystems and Robotics\",\"volume\":\"23 1\",\"pages\":\"83-88\"},\"PeriodicalIF\":1.5000,\"publicationDate\":\"2022-07-27\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IET Cybersystems and Robotics\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CYBER55403.2022.9907032\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"AUTOMATION & CONTROL SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IET Cybersystems and Robotics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CYBER55403.2022.9907032","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
An Efficient Color and Geometric Feature Fusion Module for 6D Object Pose Estiamtion
6D pose estimation is widely used in robot tasks such as sorting and grasping. RGB-D-based methods have recently attained brilliant success, but they are still susceptible to heavy occlusion. Our critical insight is that color and geometry information in RGBD images are two complementary data, and the crux of the pose estimation problem under occlusion is fully leveraging them. Towards this end, we propose a new color and geometry feature fusion module that can efficiently leverage two complementary data sources from RGB-D images. Unlike prior fusion methods, we conduct a two-stage fusion strategy to do color-depth fusion and local-global fusion successively. Specifically, we fuse the color features extracted from RGB images into the point cloud in the first stage. In the second stage, we extract local and global features from the fused point cloud using an ASSANet-like network and splice them together to obtain the final fusion features. We conducted experiments on the widely used LineMod and YCB-Video datasets, which shows that our method improves the prediction accuracy while reducing the training time.