aTGV-SF: Dense Variational Scene Flow through Projective Warping and Higher Order Regularization

David Ferstl, Christian Reinbacher, G. Riegler, M. Rüther, H. Bischof
{"title":"aTGV-SF: Dense Variational Scene Flow through Projective Warping and Higher Order Regularization","authors":"David Ferstl, Christian Reinbacher, G. Riegler, M. Rüther, H. Bischof","doi":"10.1109/3DV.2014.19","DOIUrl":null,"url":null,"abstract":"In this paper we present a novel method to accurately estimate the dense 3D motion field, known as scene flow, from depth and intensity acquisitions. The method is formulated as a convex energy optimization, where the motion warping of each scene point is estimated through a projection and back-projection directly in 3D space. We utilize higher order regularization which is weighted and directed according to the input data by an anisotropic diffusion tensor. Our formulation enables the calculation of a dense flow field which does not penalize smooth and non-rigid movements while aligning motion boundaries with strong depth boundaries. An efficient parallelization of the numerical algorithm leads to runtimes in the order of 1s and therefore enables the method to be used in a variety of applications. We show that this novel scene flow calculation outperforms existing approaches in terms of speed and accuracy. Furthermore, we demonstrate applications such as camera pose estimation and depth image super resolution, which are enabled by the high accuracy of the proposed method. We show these applications using modern depth sensors such as Microsoft Kinect or the PMD Nano Time-of-Flight sensor.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"120 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"10","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2014 2nd International Conference on 3D Vision","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/3DV.2014.19","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 10

Abstract

In this paper we present a novel method to accurately estimate the dense 3D motion field, known as scene flow, from depth and intensity acquisitions. The method is formulated as a convex energy optimization, where the motion warping of each scene point is estimated through a projection and back-projection directly in 3D space. We utilize higher order regularization which is weighted and directed according to the input data by an anisotropic diffusion tensor. Our formulation enables the calculation of a dense flow field which does not penalize smooth and non-rigid movements while aligning motion boundaries with strong depth boundaries. An efficient parallelization of the numerical algorithm leads to runtimes in the order of 1s and therefore enables the method to be used in a variety of applications. We show that this novel scene flow calculation outperforms existing approaches in terms of speed and accuracy. Furthermore, we demonstrate applications such as camera pose estimation and depth image super resolution, which are enabled by the high accuracy of the proposed method. We show these applications using modern depth sensors such as Microsoft Kinect or the PMD Nano Time-of-Flight sensor.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
aTGV-SF:基于投影扭曲和高阶正则化的密集变分场景流
在本文中,我们提出了一种新的方法来准确估计密集的三维运动场,即场景流,从深度和强度采集。该方法是一种凸能量优化方法,通过直接在三维空间中的投影和反投影来估计每个场景点的运动翘曲。我们利用高阶正则化,根据输入数据通过各向异性扩散张量进行加权和定向。我们的公式使密集流场的计算不惩罚平滑和非刚性运动,同时对准运动边界与强深度边界。数值算法的有效并行化导致运行时间为1的顺序,因此使该方法能够用于各种应用程序。我们表明,这种新的场景流计算在速度和准确性方面优于现有的方法。此外,我们还演示了相机姿态估计和深度图像超分辨率等应用,这些应用都是由该方法的高精度实现的。我们展示了这些应用使用现代深度传感器,如微软Kinect或PMD纳米飞行时间传感器。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Querying 3D Mesh Sequences for Human Action Retrieval Temporal Octrees for Compressing Dynamic Point Cloud Streams High-Quality Depth Recovery via Interactive Multi-view Stereo Iterative Closest Spectral Kernel Maps Close-Range Photometric Stereo with Point Light Sources
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1