首页 > 最新文献

2014 2nd International Conference on 3D Vision最新文献

英文 中文
aTGV-SF: Dense Variational Scene Flow through Projective Warping and Higher Order Regularization aTGV-SF:基于投影扭曲和高阶正则化的密集变分场景流
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.19
David Ferstl, Christian Reinbacher, G. Riegler, M. Rüther, H. Bischof
In this paper we present a novel method to accurately estimate the dense 3D motion field, known as scene flow, from depth and intensity acquisitions. The method is formulated as a convex energy optimization, where the motion warping of each scene point is estimated through a projection and back-projection directly in 3D space. We utilize higher order regularization which is weighted and directed according to the input data by an anisotropic diffusion tensor. Our formulation enables the calculation of a dense flow field which does not penalize smooth and non-rigid movements while aligning motion boundaries with strong depth boundaries. An efficient parallelization of the numerical algorithm leads to runtimes in the order of 1s and therefore enables the method to be used in a variety of applications. We show that this novel scene flow calculation outperforms existing approaches in terms of speed and accuracy. Furthermore, we demonstrate applications such as camera pose estimation and depth image super resolution, which are enabled by the high accuracy of the proposed method. We show these applications using modern depth sensors such as Microsoft Kinect or the PMD Nano Time-of-Flight sensor.
在本文中,我们提出了一种新的方法来准确估计密集的三维运动场,即场景流,从深度和强度采集。该方法是一种凸能量优化方法,通过直接在三维空间中的投影和反投影来估计每个场景点的运动翘曲。我们利用高阶正则化,根据输入数据通过各向异性扩散张量进行加权和定向。我们的公式使密集流场的计算不惩罚平滑和非刚性运动,同时对准运动边界与强深度边界。数值算法的有效并行化导致运行时间为1的顺序,因此使该方法能够用于各种应用程序。我们表明,这种新的场景流计算在速度和准确性方面优于现有的方法。此外,我们还演示了相机姿态估计和深度图像超分辨率等应用,这些应用都是由该方法的高精度实现的。我们展示了这些应用使用现代深度传感器,如微软Kinect或PMD纳米飞行时间传感器。
{"title":"aTGV-SF: Dense Variational Scene Flow through Projective Warping and Higher Order Regularization","authors":"David Ferstl, Christian Reinbacher, G. Riegler, M. Rüther, H. Bischof","doi":"10.1109/3DV.2014.19","DOIUrl":"https://doi.org/10.1109/3DV.2014.19","url":null,"abstract":"In this paper we present a novel method to accurately estimate the dense 3D motion field, known as scene flow, from depth and intensity acquisitions. The method is formulated as a convex energy optimization, where the motion warping of each scene point is estimated through a projection and back-projection directly in 3D space. We utilize higher order regularization which is weighted and directed according to the input data by an anisotropic diffusion tensor. Our formulation enables the calculation of a dense flow field which does not penalize smooth and non-rigid movements while aligning motion boundaries with strong depth boundaries. An efficient parallelization of the numerical algorithm leads to runtimes in the order of 1s and therefore enables the method to be used in a variety of applications. We show that this novel scene flow calculation outperforms existing approaches in terms of speed and accuracy. Furthermore, we demonstrate applications such as camera pose estimation and depth image super resolution, which are enabled by the high accuracy of the proposed method. We show these applications using modern depth sensors such as Microsoft Kinect or the PMD Nano Time-of-Flight sensor.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"120 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132407183","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Towards Illumination-Invariant 3D Reconstruction Using ToF RGB-D Cameras 利用ToF RGB-D相机实现光照不变三维重建
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.62
C. Kerl, Mohamed Souiai, Jürgen Sturm, D. Cremers
Creating textured 3D scans of indoor environments has experienced a large boost with the advent of cheap commodity depth sensors. However, the quality of the acquired 3D models is often impaired by color seams in the reconstruction due to varying illumination (e.g., Shadows or highlights) and object surfaces whose brightness and color vary with the viewpoint of the camera. In this paper, we propose a direct and simple method to estimate the pure albedo of the texture, which allows us to remove illumination effects from IR and color images. Our approach first computes the illumination-independent albedo in the IR domain, which we subsequently transfer to the color albedo. As shadows and highlights lead to over- and underexposed image regions with little or no color information, we apply an advanced optimization scheme to infer color information in the color albedo from neigh boring image regions. We demonstrate the applicability of our approach to various real-world scenes.
随着廉价商品深度传感器的出现,创建室内环境的纹理3D扫描经历了巨大的推动。然而,由于不同的照度(例如,阴影或高光)和物体表面的亮度和颜色随着相机的视点而变化,所获得的3D模型的质量通常会受到重建中的颜色接缝的影响。在本文中,我们提出了一种直接和简单的方法来估计纹理的纯反照率,使我们能够从红外和彩色图像中去除照明影响。我们的方法首先计算红外域中与光照无关的反照率,然后将其转换为彩色反照率。由于阴影和高光会导致曝光过度和曝光不足的图像区域很少或没有颜色信息,我们采用了一种先进的优化方案,从相邻的无聊图像区域推断颜色反照率中的颜色信息。我们展示了我们的方法在各种现实场景中的适用性。
{"title":"Towards Illumination-Invariant 3D Reconstruction Using ToF RGB-D Cameras","authors":"C. Kerl, Mohamed Souiai, Jürgen Sturm, D. Cremers","doi":"10.1109/3DV.2014.62","DOIUrl":"https://doi.org/10.1109/3DV.2014.62","url":null,"abstract":"Creating textured 3D scans of indoor environments has experienced a large boost with the advent of cheap commodity depth sensors. However, the quality of the acquired 3D models is often impaired by color seams in the reconstruction due to varying illumination (e.g., Shadows or highlights) and object surfaces whose brightness and color vary with the viewpoint of the camera. In this paper, we propose a direct and simple method to estimate the pure albedo of the texture, which allows us to remove illumination effects from IR and color images. Our approach first computes the illumination-independent albedo in the IR domain, which we subsequently transfer to the color albedo. As shadows and highlights lead to over- and underexposed image regions with little or no color information, we apply an advanced optimization scheme to infer color information in the color albedo from neigh boring image regions. We demonstrate the applicability of our approach to various real-world scenes.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127847850","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 31
Learning 3D Part Detection from Sparsely Labeled Data 从稀疏标记数据中学习3D零件检测
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.108
A. Makadia, M. E. Yümer
For large collections of 3D models, the ability to detect and localize parts of interest is necessary to provide search and visualization enhancements beyond simple high-level categorization. While current 3D labeling approaches rely on learning from fully labeled meshes, such training data is difficult to acquire at scale. In this work we explore learning to detect object parts from sparsely labeled data, i.e. we operate under the assumption that for any object part we have only one labeled vertex rather than a full region segmentation. Similarly, we also learn to output a single representative vertex for each detected part. Such localized predictions are useful for applications where visualization is important. Our approach relies heavily on exploiting the spatial configuration of parts on a model to drive the detection. Inspired by structured multi-class object detection models for images, we develop an algorithm that combines independently trained part classifiers with a structured SVM model, and show promising results on real-world textured 3D data.
对于大型3D模型集合,检测和定位感兴趣部分的能力是必要的,以提供简单的高级分类之外的搜索和可视化增强。虽然目前的3D标记方法依赖于从完全标记的网格中学习,但这种训练数据很难大规模获取。在这项工作中,我们探索学习从稀疏标记的数据中检测物体部分,即我们假设对于任何物体部分,我们只有一个标记的顶点,而不是一个完整的区域分割。类似地,我们还学习为每个检测到的部分输出单个代表性顶点。这种本地化预测对于可视化很重要的应用程序非常有用。我们的方法在很大程度上依赖于利用模型上零件的空间配置来驱动检测。受结构化多类图像目标检测模型的启发,我们开发了一种将独立训练的部分分类器与结构化支持向量机模型相结合的算法,并在现实世界的纹理3D数据上显示出令人满意的结果。
{"title":"Learning 3D Part Detection from Sparsely Labeled Data","authors":"A. Makadia, M. E. Yümer","doi":"10.1109/3DV.2014.108","DOIUrl":"https://doi.org/10.1109/3DV.2014.108","url":null,"abstract":"For large collections of 3D models, the ability to detect and localize parts of interest is necessary to provide search and visualization enhancements beyond simple high-level categorization. While current 3D labeling approaches rely on learning from fully labeled meshes, such training data is difficult to acquire at scale. In this work we explore learning to detect object parts from sparsely labeled data, i.e. we operate under the assumption that for any object part we have only one labeled vertex rather than a full region segmentation. Similarly, we also learn to output a single representative vertex for each detected part. Such localized predictions are useful for applications where visualization is important. Our approach relies heavily on exploiting the spatial configuration of parts on a model to drive the detection. Inspired by structured multi-class object detection models for images, we develop an algorithm that combines independently trained part classifiers with a structured SVM model, and show promising results on real-world textured 3D data.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130034767","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
A General Framework to Generate Sizing Systems from 3D Motion Data Applied to Face Mask Design 基于三维运动数据生成尺寸系统的通用框架在面罩设计中的应用
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.43
Timo Bolkart, P. Bose, Chang Shu, S. Wuhrer
For the design of mass-produced wearable objects for a population it is important to find a small number of sizes, called a sizing system, that will fit well on a wide range of individuals in the population. To obtain a sizing system that incorporates the shape of an identity along with its motion, we introduce a general framework to generate a sizing system for dynamic 3D motion data. Based on a registered 3D motion database a sizing system is computed for task-specific anthropometric measurements and tolerances, specified by designers. We generate the sizing system by transforming the problem into a box stabbing problem, which aims to find the lowest number of points stabbing a set of boxes. We use a standard computational geometry technique to solve this, it recursively computes the stabbing of lower-dimensional boxes. We apply our framework to a database of facial motion data for anthropometric measurements related to the design of face masks. We show the generalization capabilities of this sizing system on unseen data, and compute, for each size, a representative 3D shape that can be used by designers to produce a prototype model.
对于为人群设计大规模生产的可穿戴物品,重要的是要找到一个小数量的尺寸,称为尺寸系统,它将适合人群中广泛的个体。为了获得一个包含身份形状及其运动的尺寸系统,我们引入了一个通用框架来生成动态3D运动数据的尺寸系统。基于注册的3D运动数据库,计算出特定任务的人体测量和公差的尺寸系统,由设计师指定。我们通过将问题转化为一个刺穿盒子的问题来生成尺寸系统,该问题的目标是找到刺穿一组盒子的最小点数。我们使用标准的计算几何技术来解决这个问题,它递归地计算低维盒子的刺入。我们将我们的框架应用于面部运动数据数据库,用于与口罩设计相关的人体测量。我们展示了该尺寸系统在未见数据上的泛化能力,并为每种尺寸计算出具有代表性的3D形状,可供设计师用于制作原型模型。
{"title":"A General Framework to Generate Sizing Systems from 3D Motion Data Applied to Face Mask Design","authors":"Timo Bolkart, P. Bose, Chang Shu, S. Wuhrer","doi":"10.1109/3DV.2014.43","DOIUrl":"https://doi.org/10.1109/3DV.2014.43","url":null,"abstract":"For the design of mass-produced wearable objects for a population it is important to find a small number of sizes, called a sizing system, that will fit well on a wide range of individuals in the population. To obtain a sizing system that incorporates the shape of an identity along with its motion, we introduce a general framework to generate a sizing system for dynamic 3D motion data. Based on a registered 3D motion database a sizing system is computed for task-specific anthropometric measurements and tolerances, specified by designers. We generate the sizing system by transforming the problem into a box stabbing problem, which aims to find the lowest number of points stabbing a set of boxes. We use a standard computational geometry technique to solve this, it recursively computes the stabbing of lower-dimensional boxes. We apply our framework to a database of facial motion data for anthropometric measurements related to the design of face masks. We show the generalization capabilities of this sizing system on unseen data, and compute, for each size, a representative 3D shape that can be used by designers to produce a prototype model.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116204002","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Line Matching and Pose Estimation for Unconstrained Model-to-Image Alignment 无约束模型-图像对齐的直线匹配和姿态估计
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.27
K. Bhat, J. Heikkilä
This paper has two contributions in the context of line based camera pose estimation, 1) We propose a purely geometric approach to establish correspondence between 3D line segments in a given model and 2D line segments detected in an image, 2) We eliminate a degenerate case due to the type of rotation representation in arguably the best line based pose estimation method currently available. For establishing line correspondences we perform exhaustive search on the space of camera pose values till we obtain a pose (position and rotation) which is geometrically consistent with the given set of 2D, 3D lines. For this highly complex search we design a strategy which performs precomputations on the 3D model using separate set of constraints on position and rotation values. During runtime, the set of different rotation values are ranked independently and combined with each position values in the order of their ranking. Then successive geometric constraints which are much simpler when compared to computing reprojection error are used to eliminate incorrect pose values. We show that the ranking of rotation values reduces the number of trials needed by a huge factor and the simple geometric constraints avoid the need for computing the reprojection error in most cases. Though the execution time for the current MATLAB implementation is far from real time requirement, our method can be accelerated significantly by exploiting simplicity and parallelizability of the operations we employ. For eliminating the degenerate case in the state of art pose estimation method, we reformulate the rotation representation. We use unit quaternions instead of CGR parameters used by the method.
本文在基于线的相机姿态估计方面有两个贡献,1)我们提出了一种纯几何方法来建立给定模型中的3D线段与图像中检测到的2D线段之间的对应关系,2)我们消除了由于旋转表示类型而导致的退化情况,这可以说是目前可用的最佳基于线的姿态估计方法。为了建立线对应关系,我们对相机姿态值的空间进行穷举搜索,直到我们获得与给定的2D, 3D线条集几何上一致的姿态(位置和旋转)。对于这种高度复杂的搜索,我们设计了一种策略,该策略使用位置和旋转值的单独约束集对3D模型执行预计算。在运行时,不同旋转值的集合独立排序,并按照排序顺序与每个位置值组合。然后,使用比计算重投影误差简单得多的连续几何约束来消除不正确的位姿值。我们证明旋转值的排序大大减少了所需的试验次数,并且在大多数情况下,简单的几何约束避免了计算重投影误差的需要。虽然目前MATLAB实现的执行时间远未达到实时要求,但我们的方法可以通过利用我们使用的操作的简单性和并行性来显着加速。为了消除现有姿态估计方法中的退化情况,我们对旋转表示进行了重新表述。我们使用单位四元数代替方法使用的CGR参数。
{"title":"Line Matching and Pose Estimation for Unconstrained Model-to-Image Alignment","authors":"K. Bhat, J. Heikkilä","doi":"10.1109/3DV.2014.27","DOIUrl":"https://doi.org/10.1109/3DV.2014.27","url":null,"abstract":"This paper has two contributions in the context of line based camera pose estimation, 1) We propose a purely geometric approach to establish correspondence between 3D line segments in a given model and 2D line segments detected in an image, 2) We eliminate a degenerate case due to the type of rotation representation in arguably the best line based pose estimation method currently available. For establishing line correspondences we perform exhaustive search on the space of camera pose values till we obtain a pose (position and rotation) which is geometrically consistent with the given set of 2D, 3D lines. For this highly complex search we design a strategy which performs precomputations on the 3D model using separate set of constraints on position and rotation values. During runtime, the set of different rotation values are ranked independently and combined with each position values in the order of their ranking. Then successive geometric constraints which are much simpler when compared to computing reprojection error are used to eliminate incorrect pose values. We show that the ranking of rotation values reduces the number of trials needed by a huge factor and the simple geometric constraints avoid the need for computing the reprojection error in most cases. Though the execution time for the current MATLAB implementation is far from real time requirement, our method can be accelerated significantly by exploiting simplicity and parallelizability of the operations we employ. For eliminating the degenerate case in the state of art pose estimation method, we reformulate the rotation representation. We use unit quaternions instead of CGR parameters used by the method.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123913254","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
DT-SLAM: Deferred Triangulation for Robust SLAM DT-SLAM:鲁棒SLAM的延迟三角剖分
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.49
D. Herrera C., Kihwan Kim, Juho Kannala, K. Pulli, J. Heikkila
Obtaining a good baseline between different video frames is one of the key elements in vision-based monocular SLAM systems. However, if the video frames contain only a few 2D feature correspondences with a good baseline, or the camera only rotates without sufficient translation in the beginning, tracking and mapping becomes unstable. We introduce a real-time visual SLAM system that incrementally tracks individual 2D features, and estimates camera pose by using matched 2D features, regardless of the length of the baseline. Triangulating 2D features into 3D points is deferred until key frames with sufficient baseline for the features are available. Our method can also deal with pure rotational motions, and fuse the two types of measurements in a bundle adjustment step. Adaptive criteria for key frame selection are also introduced for efficient optimization and dealing with multiple maps. We demonstrate that our SLAM system improves camera pose estimates and robustness, even with purely rotational motions.
在不同视频帧之间获得良好的基线是基于视觉的单目SLAM系统的关键因素之一。但是,如果视频帧只包含少量具有良好基线的2D特征对应,或者摄像机一开始只是旋转而没有足够的平移,那么跟踪和映射就会变得不稳定。我们引入了一个实时视觉SLAM系统,该系统增量跟踪单个2D特征,并通过使用匹配的2D特征来估计相机姿势,而不考虑基线的长度。将2D特征三角化到3D点的过程会延迟到具有足够基线的关键帧可用时。该方法还可以处理纯旋转运动,并在一个束平差步骤中融合两种类型的测量。引入了关键帧选择的自适应准则,以实现高效优化和多地图处理。我们证明我们的SLAM系统提高了相机姿态估计和鲁棒性,即使是纯旋转运动。
{"title":"DT-SLAM: Deferred Triangulation for Robust SLAM","authors":"D. Herrera C., Kihwan Kim, Juho Kannala, K. Pulli, J. Heikkila","doi":"10.1109/3DV.2014.49","DOIUrl":"https://doi.org/10.1109/3DV.2014.49","url":null,"abstract":"Obtaining a good baseline between different video frames is one of the key elements in vision-based monocular SLAM systems. However, if the video frames contain only a few 2D feature correspondences with a good baseline, or the camera only rotates without sufficient translation in the beginning, tracking and mapping becomes unstable. We introduce a real-time visual SLAM system that incrementally tracks individual 2D features, and estimates camera pose by using matched 2D features, regardless of the length of the baseline. Triangulating 2D features into 3D points is deferred until key frames with sufficient baseline for the features are available. Our method can also deal with pure rotational motions, and fuse the two types of measurements in a bundle adjustment step. Adaptive criteria for key frame selection are also introduced for efficient optimization and dealing with multiple maps. We demonstrate that our SLAM system improves camera pose estimates and robustness, even with purely rotational motions.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"58 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132276905","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 49
Estimating Depth of Layered Structure Based on Multispectral Speckle Correlation 基于多光谱散斑相关的层状结构深度估计
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.90
T. Matsumura, Y. Mukaigawa, Y. Yagi
Some objects have layered structures in which a dynamic region is covered by a static layer. In this paper, we propose a new experiment for estimating the depth of the dynamic region using speckle analysis. The speckle is caused by the mutual interference of a coherence laser. We use two characteristics of the speckle. One is the temporal stability of the speckle pattern and the other is the wavelength dependency of the transmittance of the laser. We estimate the depth by computing correlations of speckle patterns using multispectral lasers. Experimental results using a simulated skin show that multispectral speckle correlation can be used for analyzing a layered structure.
一些对象具有分层结构,其中动态区域被静态层覆盖。本文提出了一种利用散斑分析估计动态区域深度的新方法。散斑是由相干激光的相互干扰引起的。我们利用了散斑的两个特性。一个是散斑图的时间稳定性,另一个是激光透过率的波长依赖性。我们利用多光谱激光计算散斑模式的相关性来估计深度。模拟皮肤的实验结果表明,多光谱散斑相关可以用于层状结构的分析。
{"title":"Estimating Depth of Layered Structure Based on Multispectral Speckle Correlation","authors":"T. Matsumura, Y. Mukaigawa, Y. Yagi","doi":"10.1109/3DV.2014.90","DOIUrl":"https://doi.org/10.1109/3DV.2014.90","url":null,"abstract":"Some objects have layered structures in which a dynamic region is covered by a static layer. In this paper, we propose a new experiment for estimating the depth of the dynamic region using speckle analysis. The speckle is caused by the mutual interference of a coherence laser. We use two characteristics of the speckle. One is the temporal stability of the speckle pattern and the other is the wavelength dependency of the transmittance of the laser. We estimate the depth by computing correlations of speckle patterns using multispectral lasers. Experimental results using a simulated skin show that multispectral speckle correlation can be used for analyzing a layered structure.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126534295","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Piecewise Planar Decomposition of 3D Point Clouds Obtained from Multiple Static RGB-D Cameras 多台静态RGB-D相机三维点云的分段平面分解
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.57
F. Barrera, N. Padoy
In this paper, we address the problem of segmenting a 3D point cloud obtained from several RGB-D cameras into a set of 3D piecewise planar regions. This is a fundamental problem in computer vision, whose solution is helpful for further scene analysis, such as support inference and object localisation. In existing planar segmentation approaches for point clouds, the point cloud originates from a single RGB-D view. There is however a growing interest to monitor environments with computer vision setups that contain a set of calibrated 3D cameras located around the scene. To fully exploit the multi-view aspect of such setups, we propose in this paper a novel approach to perform the planar piecewise segmentation directly in 3D. This approach, called Voxel-MRF (V-MRF), is based on discrete 3D Markov random fields, whose nodes correspond to scene voxels and whose labels represent 3D planes. The voxelization of the scene permits to cope with noisy depth measurements, while the MRF formulation provides a natural handling of the 3D spatial constraints during the optimisation. The approach results in a decomposition of the scene into a set of 3D planar patches. A by-product of the method is also a joint planar segmentation of the original images into planar regions with consistent labels across the views. We demonstrate the advantages of our approach using a benchmark dataset of objects with known geometry. We also present qualitative results on challenging data acquired by a multi-camera system installed in two operating rooms.
在本文中,我们解决了从几个RGB-D相机获得的三维点云分割成一组三维分段平面区域的问题。这是计算机视觉中的一个基本问题,其解决方案有助于进一步的场景分析,如支持推理和对象定位。在现有的点云平面分割方法中,点云来源于单个RGB-D视图。然而,越来越多的人对使用计算机视觉设置来监控环境感兴趣,这些设置包含一组位于场景周围的校准3D摄像机。为了充分利用这种设置的多视图方面,我们提出了一种在三维中直接执行平面分段分割的新方法。这种方法被称为体素- mrf (V-MRF),它基于离散的3D马尔可夫随机场,其节点对应于场景体素,其标签代表3D平面。场景体素化允许处理有噪声的深度测量,而MRF公式在优化过程中提供了对3D空间约束的自然处理。该方法将场景分解为一组3D平面补丁。该方法的一个副产品也是将原始图像联合平面分割成跨视图具有一致标签的平面区域。我们使用具有已知几何形状的对象的基准数据集来演示我们方法的优点。我们也提出了定性的结果,挑战性的数据,由安装在两个手术室的多摄像头系统获得。
{"title":"Piecewise Planar Decomposition of 3D Point Clouds Obtained from Multiple Static RGB-D Cameras","authors":"F. Barrera, N. Padoy","doi":"10.1109/3DV.2014.57","DOIUrl":"https://doi.org/10.1109/3DV.2014.57","url":null,"abstract":"In this paper, we address the problem of segmenting a 3D point cloud obtained from several RGB-D cameras into a set of 3D piecewise planar regions. This is a fundamental problem in computer vision, whose solution is helpful for further scene analysis, such as support inference and object localisation. In existing planar segmentation approaches for point clouds, the point cloud originates from a single RGB-D view. There is however a growing interest to monitor environments with computer vision setups that contain a set of calibrated 3D cameras located around the scene. To fully exploit the multi-view aspect of such setups, we propose in this paper a novel approach to perform the planar piecewise segmentation directly in 3D. This approach, called Voxel-MRF (V-MRF), is based on discrete 3D Markov random fields, whose nodes correspond to scene voxels and whose labels represent 3D planes. The voxelization of the scene permits to cope with noisy depth measurements, while the MRF formulation provides a natural handling of the 3D spatial constraints during the optimisation. The approach results in a decomposition of the scene into a set of 3D planar patches. A by-product of the method is also a joint planar segmentation of the original images into planar regions with consistent labels across the views. We demonstrate the advantages of our approach using a benchmark dataset of objects with known geometry. We also present qualitative results on challenging data acquired by a multi-camera system installed in two operating rooms.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115131764","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Semantic Parametric Reshaping of Human Body Models 人体模型的语义参数重构
Pub Date : 2014-12-01 DOI: 10.1109/3DV.2014.47
Yipin Yang, Yao Yu, Yu Zhou, S. Du, James Davis, Ruigang Yang
We develop a novel approach to generate human body models in a variety of shapes and poses via tuning semantic parameters. Our approach is investigated with datasets of up to 3000 scanned body models which have been placed in point to point correspondence. Correspondence is established by nonrigid deformation of a template mesh. The large dataset allows a local model to be learned robustly, in which individual parts of the human body can be accurately reshaped according to semantic parameters. We evaluate performance on two datasets and find that our model outperforms existing methods.
我们开发了一种通过调整语义参数来生成各种形状和姿势的人体模型的新方法。我们的方法研究了多达3000个扫描身体模型的数据集,这些模型已经被放置在点对点对应中。通过模板网格的非刚性变形建立对应关系。大数据集允许鲁棒学习局部模型,其中人体的各个部位可以根据语义参数精确地重塑。我们在两个数据集上评估了性能,发现我们的模型优于现有的方法。
{"title":"Semantic Parametric Reshaping of Human Body Models","authors":"Yipin Yang, Yao Yu, Yu Zhou, S. Du, James Davis, Ruigang Yang","doi":"10.1109/3DV.2014.47","DOIUrl":"https://doi.org/10.1109/3DV.2014.47","url":null,"abstract":"We develop a novel approach to generate human body models in a variety of shapes and poses via tuning semantic parameters. Our approach is investigated with datasets of up to 3000 scanned body models which have been placed in point to point correspondence. Correspondence is established by nonrigid deformation of a template mesh. The large dataset allows a local model to be learned robustly, in which individual parts of the human body can be accurately reshaped according to semantic parameters. We evaluate performance on two datasets and find that our model outperforms existing methods.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127827996","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 80
Accurate Geo-Registration by Ground-to-Aerial Image Matching 地空图像匹配的精确地理配准
Pub Date : 2014-12-01 DOI: 10.1109/3DV.2014.69
Qi Shan, Changchang Wu, B. Curless, Yasutaka Furukawa, Carlos Hernández, S. Seitz
We address the problem of geo-registering ground-based multi-view stereo models by ground-to-aerial image matching. The main contribution is a fully automated geo-registration pipeline with a novel viewpoint-dependent matching method that handles ground to aerial viewpoint variation. We conduct large-scale experiments which consist of many popular outdoor landmarks in Rome. The proposed approach demonstrates a high success rate for the task, and dramatically outperforms state-of-the-art techniques, yielding geo-registration at pixel-level accuracy.
我们通过地空图像匹配解决了地面多视角立体模型的地理配准问题。主要贡献是一个完全自动化的地理配准管道,具有一种新颖的视点相关匹配方法,可以处理地面到空中的视点变化。我们进行了大规模的实验,其中包括罗马许多受欢迎的户外地标。所提出的方法证明了任务的高成功率,并且大大优于最先进的技术,产生像素级精度的地理注册。
{"title":"Accurate Geo-Registration by Ground-to-Aerial Image Matching","authors":"Qi Shan, Changchang Wu, B. Curless, Yasutaka Furukawa, Carlos Hernández, S. Seitz","doi":"10.1109/3DV.2014.69","DOIUrl":"https://doi.org/10.1109/3DV.2014.69","url":null,"abstract":"We address the problem of geo-registering ground-based multi-view stereo models by ground-to-aerial image matching. The main contribution is a fully automated geo-registration pipeline with a novel viewpoint-dependent matching method that handles ground to aerial viewpoint variation. We conduct large-scale experiments which consist of many popular outdoor landmarks in Rome. The proposed approach demonstrates a high success rate for the task, and dramatically outperforms state-of-the-art techniques, yielding geo-registration at pixel-level accuracy.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127861025","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 99
期刊
2014 2nd International Conference on 3D Vision
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1