Refining 6D Object Pose Predictions using Abstract Render-and-Compare

Arul Selvam Periyasamy, Max Schwarz, Sven Behnke
{"title":"Refining 6D Object Pose Predictions using Abstract Render-and-Compare","authors":"Arul Selvam Periyasamy, Max Schwarz, Sven Behnke","doi":"10.1109/Humanoids43949.2019.9035024","DOIUrl":null,"url":null,"abstract":"Robotic systems often require precise scene analysis capabilities, especially in unstructured, cluttered situations, as occurring in human-made environments. While current deep-learning based methods yield good estimates of object poses, they often struggle with large amounts of occlusion and do not take inter-object effects into account. Vision as inverse graphics is a promising concept for detailed scene analysis. A key element for this idea is a method for inferring scene parameter updates from the rasterized 2D scene. However, the rasterization process is notoriously difficult to invert, both due to the projection and occlusion process, but also due to secondary effects such as lighting or reflections. We propose to remove the latter from the process by mapping the rasterized image into an abstract feature space learned in a self-supervised way from pixel correspondences. Using only a light-weight inverse rendering module, this allows us to refine 6D object pose estimations in highly cluttered scenes by optimizing a simple pixel-wise difference in the abstract image representation. We evaluate our approach on the challenging YCB-Video dataset, where it yields large improvements and demonstrates a large basin of attraction towards the correct object poses.","PeriodicalId":404758,"journal":{"name":"2019 IEEE-RAS 19th International Conference on Humanoid Robots (Humanoids)","volume":"38 2","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"14","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 IEEE-RAS 19th International Conference on Humanoid Robots (Humanoids)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/Humanoids43949.2019.9035024","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 14

Abstract

Robotic systems often require precise scene analysis capabilities, especially in unstructured, cluttered situations, as occurring in human-made environments. While current deep-learning based methods yield good estimates of object poses, they often struggle with large amounts of occlusion and do not take inter-object effects into account. Vision as inverse graphics is a promising concept for detailed scene analysis. A key element for this idea is a method for inferring scene parameter updates from the rasterized 2D scene. However, the rasterization process is notoriously difficult to invert, both due to the projection and occlusion process, but also due to secondary effects such as lighting or reflections. We propose to remove the latter from the process by mapping the rasterized image into an abstract feature space learned in a self-supervised way from pixel correspondences. Using only a light-weight inverse rendering module, this allows us to refine 6D object pose estimations in highly cluttered scenes by optimizing a simple pixel-wise difference in the abstract image representation. We evaluate our approach on the challenging YCB-Video dataset, where it yields large improvements and demonstrates a large basin of attraction towards the correct object poses.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
精炼6D对象姿态预测使用抽象渲染和比较
机器人系统通常需要精确的场景分析能力,特别是在人工环境中发生的非结构化、混乱的情况下。虽然目前基于深度学习的方法可以很好地估计物体的姿势,但它们经常与大量遮挡作斗争,并且没有考虑到物体间的影响。视觉作为逆图形是一个很有前途的概念,用于详细的场景分析。这个想法的一个关键元素是从栅格化的2D场景中推断场景参数更新的方法。然而,栅格化过程是出了名的难以反转,既由于投影和遮挡过程,也由于次要影响,如照明或反射。我们建议通过将光栅化图像映射到以自监督方式从像素对应中学习的抽象特征空间来消除后者。仅使用轻量级的反向渲染模块,这允许我们通过优化抽象图像表示中的简单像素差异来优化高度混乱场景中的6D对象姿态估计。我们在具有挑战性的YCB-Video数据集上评估了我们的方法,在那里它产生了很大的改进,并展示了对正确物体姿势的巨大吸引力。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Position-Based Lateral Balance Control for Knee-Stretched Biped Robot Mechanistic Properties of Five-bar Parallel Mechanism for Leg Structure Based on Spring Loaded Inverted Pendulum A deep reinforcement learning based approach towards generating human walking behavior with a neuromuscular model Using Virtual Reality to Examine the Neural and Physiological Anxiety-Related Responses to Balance-Demanding Target-Reaching Leaning Tasks Motion Retargeting and Control for Teleoperated Physical Human-Robot Interaction
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1