首页 > 最新文献

2016 Fourth International Conference on 3D Vision (3DV)最新文献

英文 中文
Rapid Hand Shape Reconstruction with Chebyshev Phase Shifting 切比雪夫相移快速手形重建
Pub Date : 2016-10-01 DOI: 10.1109/3DV.2016.24
Daniel Moreno, Wook-Yeon Hwang, G. Taubin
Human hand motion and shape sensing is an area of high interest in medical communities and for human interaction researchers. Measurement of small hand movements could help professionals to quantize the stage of conditions like Parkinson's Disease (PD) and Essential Tremor (ET). Similar data is also useful for designers of human interaction algorithms to infer information about hand pose and gesture recognition. In this paper we present a structured light sensor capable of measuring hand shape and color at 121 FPS. Our algorithm uses a novel structured light method developed by us, called Chebyshev Phase Shifting (CPS). This method uses a digital projector and a camera to create high-resolution color 3D models from sequences of color images. We show how to encode CPS patterns in three RGB images for a reduced acquisition time, enabling high speed capture. We have built a prototype to measure rapid trembling hands. Our results show our prototype accurately captures fast tremors similar to those of PD patients. Color 3D model sequences recorded at high speed with our sensor will be used to study hand kinematic properties in a future.
人体手部运动和形状感知是医学界和人类互动研究人员非常感兴趣的一个领域。对手部运动的测量可以帮助专业人员量化帕金森病(PD)和原发性震颤(ET)等疾病的阶段。类似的数据对于人类交互算法的设计者推断手部姿势和手势识别的信息也很有用。在本文中,我们提出了一种能够以121 FPS的速度测量手的形状和颜色的结构光传感器。我们的算法使用了一种由我们开发的新型结构光方法,称为切比雪夫相移(CPS)。这种方法使用数字投影仪和相机从彩色图像序列中创建高分辨率彩色3D模型。我们展示了如何在三个RGB图像中编码CPS模式,以减少采集时间,实现高速捕获。我们已经建立了一个原型来测量快速颤抖的手。我们的研究结果表明,我们的原型准确地捕捉到与PD患者相似的快速震颤。用我们的传感器高速记录的彩色3D模型序列将在未来用于研究手部运动特性。
{"title":"Rapid Hand Shape Reconstruction with Chebyshev Phase Shifting","authors":"Daniel Moreno, Wook-Yeon Hwang, G. Taubin","doi":"10.1109/3DV.2016.24","DOIUrl":"https://doi.org/10.1109/3DV.2016.24","url":null,"abstract":"Human hand motion and shape sensing is an area of high interest in medical communities and for human interaction researchers. Measurement of small hand movements could help professionals to quantize the stage of conditions like Parkinson's Disease (PD) and Essential Tremor (ET). Similar data is also useful for designers of human interaction algorithms to infer information about hand pose and gesture recognition. In this paper we present a structured light sensor capable of measuring hand shape and color at 121 FPS. Our algorithm uses a novel structured light method developed by us, called Chebyshev Phase Shifting (CPS). This method uses a digital projector and a camera to create high-resolution color 3D models from sequences of color images. We show how to encode CPS patterns in three RGB images for a reduced acquisition time, enabling high speed capture. We have built a prototype to measure rapid trembling hands. Our results show our prototype accurately captures fast tremors similar to those of PD patients. Color 3D model sequences recorded at high speed with our sensor will be used to study hand kinematic properties in a future.","PeriodicalId":425304,"journal":{"name":"2016 Fourth International Conference on 3D Vision (3DV)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127800115","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
A Large-Scale 3D Object Recognition Dataset 一个大规模的三维物体识别数据集
Pub Date : 2016-10-01 DOI: 10.1109/3DV.2016.16
Thomas Sølund, A. Buch, N. Krüger, H. Aanæs
This paper presents a new large scale dataset targeting evaluation of local shape descriptors and 3d object recognition algorithms. The dataset consists of point clouds and triangulated meshes from 292 physical scenes taken from 11 different views, a total of approximately 3204 views. Each of the physical scenes contain 10 occluded objects resulting in a dataset with 32040 unique object poses and 45 different object models. The 45 object models are full 360 degree models which are scanned with a high precision structured light scanner and a turntable. All the included objects belong to different geometric groups, concave, convex, cylindrical and flat 3D object models. The object models have varying amount of local geometric features to challenge existing local shape feature descriptors in terms of descriptiveness and robustness. The dataset is validated in a benchmark which evaluates the matching performance of 7 different state-of-the-art local shape descriptors. Further, we validate the dataset in a 3D object recognition pipeline. Our benchmark shows as expected that local shape feature descriptors without any global point relation across the surface have a poor matching performance with flat and cylindrical objects. It is our objective that this dataset contributes to the future development of next generation of 3D object recognition algorithms. The dataset is public available at http://roboimagedata.compute.dtu.dk/.
本文提出了一种新的大规模数据集,用于评估局部形状描述符和三维物体识别算法。该数据集由292个物理场景的点云和三角网格组成,这些场景取自11个不同的视图,总计约3204个视图。每个物理场景包含10个遮挡的对象,从而形成一个具有32040个独特对象姿势和45个不同对象模型的数据集。45个对象模型是完整的360度模型,用高精度结构光扫描仪和转盘扫描。所有包含的对象都属于不同的几何组,凹、凸、圆柱形和平面三维对象模型。对象模型具有不同数量的局部几何特征,在描述性和鲁棒性方面挑战了现有的局部形状特征描述符。该数据集在基准测试中得到验证,该基准测试评估了7种不同的最先进的局部形状描述符的匹配性能。此外,我们在3D对象识别管道中验证数据集。我们的基准测试结果表明,没有任何全局点关系的局部形状特征描述符与平面和圆柱形物体的匹配性能很差。我们的目标是这个数据集有助于下一代3D物体识别算法的未来发展。该数据集可在http://roboimagedata.compute.dtu.dk/上公开获取。
{"title":"A Large-Scale 3D Object Recognition Dataset","authors":"Thomas Sølund, A. Buch, N. Krüger, H. Aanæs","doi":"10.1109/3DV.2016.16","DOIUrl":"https://doi.org/10.1109/3DV.2016.16","url":null,"abstract":"This paper presents a new large scale dataset targeting evaluation of local shape descriptors and 3d object recognition algorithms. The dataset consists of point clouds and triangulated meshes from 292 physical scenes taken from 11 different views, a total of approximately 3204 views. Each of the physical scenes contain 10 occluded objects resulting in a dataset with 32040 unique object poses and 45 different object models. The 45 object models are full 360 degree models which are scanned with a high precision structured light scanner and a turntable. All the included objects belong to different geometric groups, concave, convex, cylindrical and flat 3D object models. The object models have varying amount of local geometric features to challenge existing local shape feature descriptors in terms of descriptiveness and robustness. The dataset is validated in a benchmark which evaluates the matching performance of 7 different state-of-the-art local shape descriptors. Further, we validate the dataset in a 3D object recognition pipeline. Our benchmark shows as expected that local shape feature descriptors without any global point relation across the surface have a poor matching performance with flat and cylindrical objects. It is our objective that this dataset contributes to the future development of next generation of 3D object recognition algorithms. The dataset is public available at http://roboimagedata.compute.dtu.dk/.","PeriodicalId":425304,"journal":{"name":"2016 Fourth International Conference on 3D Vision (3DV)","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121413401","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Exemplar-Based 3D Shape Segmentation in Point Clouds 基于范例的点云三维形状分割
Pub Date : 2016-10-01 DOI: 10.1109/3DV.2016.29
Rongqi Qiu, U. Neumann
This paper addresses the problem of automatic 3D shape segmentation in point cloud representation. Of particular interest are segmentations of noisy real scans, which is a difficult problem in previous works. To guide segmentation of target shape, a small set of pre-segmented exemplar shapes in the same category is adopted. The main idea is to register the target shape with exemplar shapes in a piece-wise rigid manner, so that pieces under the same rigid transformation are more likely to be in the same segment. To achieve this goal, an over-complete set of candidate transformations is generated in the first stage. Then, each transformation is treated as a label and an assignment is optimized over all points. The transformation labels, together with nearest-neighbor transferred segment labels, constitute final labels of target shapes. The method is not dependent on high-order features, and thus robust to noise as can be shown in the experiments on challenging datasets.
研究了点云表示中三维形状的自动分割问题。特别感兴趣的是真正分割吵闹的扫描,这是一个困难的问题在以前的作品。为了指导目标形状的分割,采用了一组相同类别的预分割样例形状。主要的思想是注册目标形状与范例形状分片刚性的方式,这部分在同样的刚性变换更有可能在同一段。为了实现这一目标,一组完整的候选人在第一阶段生成转换。然后,将每个转换视为一个标签,并在所有点上优化分配。转换标签和近邻转移部分标签,构成最终的目标形状的标签。该方法不依赖于高阶特征,因此对噪声具有鲁棒性,这可以在具有挑战性的数据集上的实验中得到证明。
{"title":"Exemplar-Based 3D Shape Segmentation in Point Clouds","authors":"Rongqi Qiu, U. Neumann","doi":"10.1109/3DV.2016.29","DOIUrl":"https://doi.org/10.1109/3DV.2016.29","url":null,"abstract":"This paper addresses the problem of automatic 3D shape segmentation in point cloud representation. Of particular interest are segmentations of noisy real scans, which is a difficult problem in previous works. To guide segmentation of target shape, a small set of pre-segmented exemplar shapes in the same category is adopted. The main idea is to register the target shape with exemplar shapes in a piece-wise rigid manner, so that pieces under the same rigid transformation are more likely to be in the same segment. To achieve this goal, an over-complete set of candidate transformations is generated in the first stage. Then, each transformation is treated as a label and an assignment is optimized over all points. The transformation labels, together with nearest-neighbor transferred segment labels, constitute final labels of target shapes. The method is not dependent on high-order features, and thus robust to noise as can be shown in the experiments on challenging datasets.","PeriodicalId":425304,"journal":{"name":"2016 Fourth International Conference on 3D Vision (3DV)","volume":"132 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122902606","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Single View 3D Reconstruction under an Uncalibrated Camera and an Unknown Mirror Sphere 未标定相机和未知镜球下的单视图三维重建
Pub Date : 2016-10-01 DOI: 10.1109/3DV.2016.50
K. Han, Kwan-Yee Kenneth Wong, Xiao Tan
In this paper, we develop a novel self-calibration method for single view 3D reconstruction using a mirror sphere. Unlike other mirror sphere based reconstruction methods, our method needs neither the intrinsic parameters of the camera, nor the position and radius of the sphere be known. Based on eigen decomposition of the matrix representing the conic image of the sphere and enforcing a repeated eignvalue constraint, we derive an analytical solution for recovering the focal length of the camera given its principal point. We then introduce a robust algorithm for estimating both the principal point and the focal length of the camera by minimizing the differences between focal lengths estimated from multiple images of the sphere. We also present a novel approach for estimating both the principal point and focal length of the camera in the case of just one single image of the sphere. With the estimated camera intrinsic parameters, the position(s) of the sphere can be readily retrieved from the eigen decomposition(s) and a scaled 3D reconstruction follows. Experimental results on both synthetic and real data are presented, which demonstrate the feasibility and accuracy of our approach.
在本文中,我们开发了一种新的自校准方法,用于使用镜球进行单视图三维重建。与其他基于镜球的重建方法不同,我们的方法既不需要相机的固有参数,也不需要知道球的位置和半径。基于表示球体圆锥像的矩阵的特征分解,并施加重复的特征值约束,我们导出了给定主点的相机焦距恢复的解析解。然后,我们引入了一种鲁棒算法,通过最小化从球体的多个图像估计的焦距之间的差异来估计相机的主点和焦距。我们还提出了一种新的方法来估计主点和焦距的相机的情况下,只有一个单一的图像的球体。利用估计的相机内部参数,可以很容易地从特征分解中提取球体的位置,然后进行缩放的三维重建。在合成数据和实际数据上的实验结果验证了该方法的可行性和准确性。
{"title":"Single View 3D Reconstruction under an Uncalibrated Camera and an Unknown Mirror Sphere","authors":"K. Han, Kwan-Yee Kenneth Wong, Xiao Tan","doi":"10.1109/3DV.2016.50","DOIUrl":"https://doi.org/10.1109/3DV.2016.50","url":null,"abstract":"In this paper, we develop a novel self-calibration method for single view 3D reconstruction using a mirror sphere. Unlike other mirror sphere based reconstruction methods, our method needs neither the intrinsic parameters of the camera, nor the position and radius of the sphere be known. Based on eigen decomposition of the matrix representing the conic image of the sphere and enforcing a repeated eignvalue constraint, we derive an analytical solution for recovering the focal length of the camera given its principal point. We then introduce a robust algorithm for estimating both the principal point and the focal length of the camera by minimizing the differences between focal lengths estimated from multiple images of the sphere. We also present a novel approach for estimating both the principal point and focal length of the camera in the case of just one single image of the sphere. With the estimated camera intrinsic parameters, the position(s) of the sphere can be readily retrieved from the eigen decomposition(s) and a scaled 3D reconstruction follows. Experimental results on both synthetic and real data are presented, which demonstrate the feasibility and accuracy of our approach.","PeriodicalId":425304,"journal":{"name":"2016 Fourth International Conference on 3D Vision (3DV)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127057287","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Discriminative Filters for Depth from Defocus 离焦深度的判别滤波器
Pub Date : 2016-10-01 DOI: 10.1109/3DV.2016.67
Fahim Mannan, M. Langer
Depth from defocus (DFD) requires estimating the depth dependent defocus blur at every pixel. Several approaches for accomplishing this have been proposed over the years. For a pair of images this is done by modeling the defocus relationship between the two differently defocused images and for single defocused images by relying on the the properties of the point spread function and the characteristics of the latent sharp image. We propose depth discriminative filters for DFD that can represent many of the widely used models such as the relative blur, Blur Equalization Technique, deconvolution based depth estimation, and subspace projection methods. We show that by optimizing the parameters of this general model we can obtain state-of-the-art result on synthetic and real defocused images with single or multiple defocused images with different apertures.
离焦深度(DFD)需要估计每个像素上与深度相关的离焦模糊。多年来,已经提出了实现这一目标的几种方法。对于一对图像,这是通过模拟两个不同散焦图像之间的散焦关系来完成的,对于单个散焦图像,则依赖于点扩散函数的性质和潜在锐图像的特征。我们提出了用于DFD的深度判别滤波器,可以代表许多广泛使用的模型,如相对模糊,模糊均衡技术,基于反卷积的深度估计和子空间投影方法。结果表明,通过对该模型的参数进行优化,可以对不同孔径的单幅或多幅离焦图像进行综合和真实离焦图像的处理。
{"title":"Discriminative Filters for Depth from Defocus","authors":"Fahim Mannan, M. Langer","doi":"10.1109/3DV.2016.67","DOIUrl":"https://doi.org/10.1109/3DV.2016.67","url":null,"abstract":"Depth from defocus (DFD) requires estimating the depth dependent defocus blur at every pixel. Several approaches for accomplishing this have been proposed over the years. For a pair of images this is done by modeling the defocus relationship between the two differently defocused images and for single defocused images by relying on the the properties of the point spread function and the characteristics of the latent sharp image. We propose depth discriminative filters for DFD that can represent many of the widely used models such as the relative blur, Blur Equalization Technique, deconvolution based depth estimation, and subspace projection methods. We show that by optimizing the parameters of this general model we can obtain state-of-the-art result on synthetic and real defocused images with single or multiple defocused images with different apertures.","PeriodicalId":425304,"journal":{"name":"2016 Fourth International Conference on 3D Vision (3DV)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127406703","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-Label Semantic 3D Reconstruction Using Voxel Blocks 使用体素块的多标签语义三维重建
Pub Date : 2016-10-01 DOI: 10.1109/3DV.2016.68
Ian Cherabier, Christian Häne, Martin R. Oswald, M. Pollefeys
Techniques that jointly perform dense 3D reconstruction and semantic segmentation have recently shown very promising results. One major restriction so far is that they can often only handle a very low number of semantic labels. This is mostly due to their high memory consumption caused by the necessity to store indicator variables for every label and transition. We propose a way to reduce the memory consumption of existing methods. Our approach is based on the observation that many semantic labels are only present at very localized positions in the scene, such as cars. Therefore this label does not need to be active at every location. We exploit this observation by dividing the scene into blocks in which generally only a subset of labels is active. By determining early on in the reconstruction process which labels need to be active in which block the memory consumption can be significantly reduced. In order to recover from mistakes we propose to update the set of active labels during the iterative optimization procedure based on the current solution. We also propose a way to initialize the set of active labels using a boosted classifier. In our experimental evaluation we show the reduction of memory usage quantitatively. Eventually, we show results of joint semantic 3D reconstruction and semantic segmentation with significantly more labels than previous approaches were able to handle.
联合执行密集三维重建和语义分割的技术最近显示出非常有希望的结果。到目前为止,一个主要的限制是它们通常只能处理非常少的语义标签。这主要是由于它们需要为每个标签和转换存储指示符变量而导致的高内存消耗。我们提出了一种减少现有方法的内存消耗的方法。我们的方法是基于这样的观察:许多语义标签只出现在场景中非常局部的位置,比如汽车。因此,这个标签不需要在每个位置都是激活的。我们通过将场景划分为块来利用这一观察结果,通常只有一部分标签是活跃的。通过在重建过程的早期确定哪些标签需要在哪个块中激活,可以显着减少内存消耗。为了从错误中恢复,我们提出在迭代优化过程中基于当前解更新活动标签集。我们还提出了一种使用增强分类器初始化活动标签集的方法。在我们的实验评估中,我们定量地展示了内存使用的减少。最后,我们展示了联合语义三维重建和语义分割的结果,与以前的方法相比,它们能够处理更多的标签。
{"title":"Multi-Label Semantic 3D Reconstruction Using Voxel Blocks","authors":"Ian Cherabier, Christian Häne, Martin R. Oswald, M. Pollefeys","doi":"10.1109/3DV.2016.68","DOIUrl":"https://doi.org/10.1109/3DV.2016.68","url":null,"abstract":"Techniques that jointly perform dense 3D reconstruction and semantic segmentation have recently shown very promising results. One major restriction so far is that they can often only handle a very low number of semantic labels. This is mostly due to their high memory consumption caused by the necessity to store indicator variables for every label and transition. We propose a way to reduce the memory consumption of existing methods. Our approach is based on the observation that many semantic labels are only present at very localized positions in the scene, such as cars. Therefore this label does not need to be active at every location. We exploit this observation by dividing the scene into blocks in which generally only a subset of labels is active. By determining early on in the reconstruction process which labels need to be active in which block the memory consumption can be significantly reduced. In order to recover from mistakes we propose to update the set of active labels during the iterative optimization procedure based on the current solution. We also propose a way to initialize the set of active labels using a boosted classifier. In our experimental evaluation we show the reduction of memory usage quantitatively. Eventually, we show results of joint semantic 3D reconstruction and semantic segmentation with significantly more labels than previous approaches were able to handle.","PeriodicalId":425304,"journal":{"name":"2016 Fourth International Conference on 3D Vision (3DV)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129324696","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 45
Camera Motion from Group Synchronization 相机运动从组同步
Pub Date : 2016-10-01 DOI: 10.1109/3DV.2016.64
F. Arrigoni, Andrea Fusiello, B. Rossi
This paper deals with the problem of estimating camera motion in the context of structure-from-motion. We describe a pipeline that consumes relative orientations and produces absolute orientations (i.e. camera position and attitude in an absolute reference frame). This pipeline exploits the concept of "group synchronization" in most of its stages, all of which entail direct solutions such as eigenvalue decompositions or linear least squares. A comprehensive introduction to the group synchronization problem is provided, and the proposed pipeline is evaluated on standard real datasets.
本文研究了基于运动构造的摄像机运动估计问题。我们描述了一个消耗相对方向并产生绝对方向的管道(即绝对参考系中的摄像机位置和姿态)。该管道在其大多数阶段利用了“组同步”的概念,所有这些都需要直接解决方案,如特征值分解或线性最小二乘。全面介绍了群同步问题,并在标准真实数据集上对所提出的管道进行了评估。
{"title":"Camera Motion from Group Synchronization","authors":"F. Arrigoni, Andrea Fusiello, B. Rossi","doi":"10.1109/3DV.2016.64","DOIUrl":"https://doi.org/10.1109/3DV.2016.64","url":null,"abstract":"This paper deals with the problem of estimating camera motion in the context of structure-from-motion. We describe a pipeline that consumes relative orientations and produces absolute orientations (i.e. camera position and attitude in an absolute reference frame). This pipeline exploits the concept of \"group synchronization\" in most of its stages, all of which entail direct solutions such as eigenvalue decompositions or linear least squares. A comprehensive introduction to the group synchronization problem is provided, and the proposed pipeline is evaluated on standard real datasets.","PeriodicalId":425304,"journal":{"name":"2016 Fourth International Conference on 3D Vision (3DV)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123713257","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Depth from Gradients in Dense Light Fields for Object Reconstruction 用于物体重建的密集光场梯度深度
Pub Date : 2016-10-01 DOI: 10.1109/3DV.2016.33
Kaan Yücer, Changil Kim, A. Sorkine-Hornung, O. Sorkine-Hornung
Objects with thin features and fine details are challenging for most multi-view stereo techniques, since such features occupy small volumes and are usually only visible in a small portion of the available views. In this paper, we present an efficient algorithm to reconstruct intricate objects using densely sampled light fields. At the heart of our technique lies a novel approach to compute per-pixel depth values by exploiting local gradient information in densely sampled light fields. This approach can generate accurate depth values for very thin features, and can be run for each pixel in parallel. We assess the reliability of our depth estimates using a novel two-sided photoconsistency measure, which can capture whether the pixel lies on a texture or a silhouette edge. This information is then used to propagate the depth estimates at high gradient regions to smooth parts of the views efficiently and reliably using edge-aware filtering. In the last step, the per-image depth values and color information are aggregated in 3D space using a voting scheme, allowing the reconstruction of a globally consistent mesh for the object. Our approach can process large video datasets very efficiently and at the same time generates high quality object reconstructions that compare favorably to the results of state-of-the-art multi-view stereo methods.
具有细特征和精细细节的对象对于大多数多视图立体技术来说是具有挑战性的,因为这些特征占用很小的体积,并且通常只在可用视图的一小部分可见。本文提出了一种利用密集采样光场重构复杂物体的有效算法。我们技术的核心在于一种新颖的方法,通过利用密集采样光场中的局部梯度信息来计算每像素深度值。这种方法可以为非常薄的特征生成精确的深度值,并且可以对每个像素并行运行。我们使用一种新的双面光一致性测量来评估深度估计的可靠性,该测量可以捕获像素是位于纹理还是轮廓边缘。然后使用这些信息在高梯度区域传播深度估计,使用边缘感知滤波高效可靠地平滑部分视图。在最后一步中,使用投票方案将每张图像的深度值和颜色信息聚合到3D空间中,从而允许为对象重建全局一致的网格。我们的方法可以非常有效地处理大型视频数据集,同时生成高质量的物体重建,与最先进的多视图立体方法的结果相比,效果更好。
{"title":"Depth from Gradients in Dense Light Fields for Object Reconstruction","authors":"Kaan Yücer, Changil Kim, A. Sorkine-Hornung, O. Sorkine-Hornung","doi":"10.1109/3DV.2016.33","DOIUrl":"https://doi.org/10.1109/3DV.2016.33","url":null,"abstract":"Objects with thin features and fine details are challenging for most multi-view stereo techniques, since such features occupy small volumes and are usually only visible in a small portion of the available views. In this paper, we present an efficient algorithm to reconstruct intricate objects using densely sampled light fields. At the heart of our technique lies a novel approach to compute per-pixel depth values by exploiting local gradient information in densely sampled light fields. This approach can generate accurate depth values for very thin features, and can be run for each pixel in parallel. We assess the reliability of our depth estimates using a novel two-sided photoconsistency measure, which can capture whether the pixel lies on a texture or a silhouette edge. This information is then used to propagate the depth estimates at high gradient regions to smooth parts of the views efficiently and reliably using edge-aware filtering. In the last step, the per-image depth values and color information are aggregated in 3D space using a voting scheme, allowing the reconstruction of a globally consistent mesh for the object. Our approach can process large video datasets very efficiently and at the same time generates high quality object reconstructions that compare favorably to the results of state-of-the-art multi-view stereo methods.","PeriodicalId":425304,"journal":{"name":"2016 Fourth International Conference on 3D Vision (3DV)","volume":"297 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125151147","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
Robust Recovery of Heavily Degraded Depth Measurements 严重退化深度测量的稳健恢复
Pub Date : 2016-10-01 DOI: 10.1109/3DV.2016.15
Gilad Drozdov, Yevgengy Shapiro, Guy Gilboa
The revolution of RGB-D sensors is advancing towards mobile platforms for robotics, autonomous vehicles and consumer hand-held devices. Strong pressures on power consumption and system price require new powerful algorithms that can robustly handle very low quality raw data. In this paper we demonstrate the ability to reliably recover depth measurements from a variety of highly degraded depth modalities, coupled with standard RGB imagery. The method is based on a regularizer which fuses super-pixel information with the total-generalized-variation (TGV) functional. We examine our algorithm on several different degradations, including new Intel's RealSense hand-held device, LiDAR-type data and ultra-sparse random sampling. In all modalities which are heavily degraded, our robust algorithm achieves superior performance over the state-ofthe-art. Additionally, a robust error measure based on Tukey's biweight metric is suggested, which is better at ranking algorithm performance since it does not reward blurry non-physical depth results.
RGB-D传感器的革命正在向机器人、自动驾驶汽车和消费者手持设备的移动平台推进。功耗和系统价格的巨大压力要求新的强大算法能够稳健地处理非常低质量的原始数据。在本文中,我们展示了从各种高度退化的深度模式中可靠地恢复深度测量的能力,再加上标准的RGB图像。该方法基于正则化器,将超像素信息与总广义变分(TGV)泛函融合。我们在几种不同的退化情况下检查了我们的算法,包括新的英特尔的RealSense手持设备,激光雷达类型的数据和超稀疏随机抽样。在所有严重退化的模式中,我们的鲁棒算法实现了优于最先进的性能。此外,提出了一种基于Tukey的双权重度量的鲁棒误差度量,由于它不奖励模糊的非物理深度结果,因此可以更好地对算法性能进行排名。
{"title":"Robust Recovery of Heavily Degraded Depth Measurements","authors":"Gilad Drozdov, Yevgengy Shapiro, Guy Gilboa","doi":"10.1109/3DV.2016.15","DOIUrl":"https://doi.org/10.1109/3DV.2016.15","url":null,"abstract":"The revolution of RGB-D sensors is advancing towards mobile platforms for robotics, autonomous vehicles and consumer hand-held devices. Strong pressures on power consumption and system price require new powerful algorithms that can robustly handle very low quality raw data. In this paper we demonstrate the ability to reliably recover depth measurements from a variety of highly degraded depth modalities, coupled with standard RGB imagery. The method is based on a regularizer which fuses super-pixel information with the total-generalized-variation (TGV) functional. We examine our algorithm on several different degradations, including new Intel's RealSense hand-held device, LiDAR-type data and ultra-sparse random sampling. In all modalities which are heavily degraded, our robust algorithm achieves superior performance over the state-ofthe-art. Additionally, a robust error measure based on Tukey's biweight metric is suggested, which is better at ranking algorithm performance since it does not reward blurry non-physical depth results.","PeriodicalId":425304,"journal":{"name":"2016 Fourth International Conference on 3D Vision (3DV)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128467286","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Detecting and Correcting Shadows in Urban Point Clouds and Image Collections 城市点云和图像集中的阴影检测与校正
Pub Date : 2016-10-01 DOI: 10.1109/3DV.2016.63
M. Guislain, Julie Digne, R. Chaine, Dimitri Kudelski, Pascal Lefebvre-Albaret, Lefebvre-Albaret. Detecting
LiDAR (Light Detection And Ranging) acquisition is a widespread method for measuring urban scenes, be it a small town neighborhood or an entire city. It is even more interesting when this acquisition is coupled with a collection of pictures registered with the data, permitting to recover the color information of the points. Yet, this added color can be perturbed by shadows that are very dependent on the sun direction and weather conditions during the acquisition. In this paper, we focus on the problem of automatically detecting and correcting the shadows from the LiDAR data by exploiting both the images and the point set laser reflectance. Building on the observation that shadow boundaries are characterized by both a significant color change and a stable laser reflectance, we propose to first detect shadow boundaries in the point set and then segment ground shadows using graph cuts in the image. Finally using a simplified illumination model we correct the shadows directly on the colored point sets. This joint exploitation of both the laser point set and the images renders our approach robust and efficient, avoiding user interaction.
激光雷达(光探测和测距)采集是一种广泛的测量城市场景的方法,无论是小城镇社区还是整个城市。更有趣的是,当这种获取与与数据注册的图片集合相结合时,允许恢复点的颜色信息。然而,在采集过程中,这种增加的颜色可能会受到阴影的干扰,阴影非常依赖于太阳的方向和天气条件。本文主要研究了利用图像和点集激光反射率对激光雷达数据中的阴影进行自动检测和校正的问题。在观察到阴影边界具有显著的颜色变化和稳定的激光反射率的基础上,我们建议首先在点集中检测阴影边界,然后使用图像中的图形切割来分割地面阴影。最后,使用简化的光照模型直接对彩色点集上的阴影进行校正。这种激光点集和图像的联合利用使我们的方法鲁棒和高效,避免了用户交互。
{"title":"Detecting and Correcting Shadows in Urban Point Clouds and Image Collections","authors":"M. Guislain, Julie Digne, R. Chaine, Dimitri Kudelski, Pascal Lefebvre-Albaret, Lefebvre-Albaret. Detecting","doi":"10.1109/3DV.2016.63","DOIUrl":"https://doi.org/10.1109/3DV.2016.63","url":null,"abstract":"LiDAR (Light Detection And Ranging) acquisition is a widespread method for measuring urban scenes, be it a small town neighborhood or an entire city. It is even more interesting when this acquisition is coupled with a collection of pictures registered with the data, permitting to recover the color information of the points. Yet, this added color can be perturbed by shadows that are very dependent on the sun direction and weather conditions during the acquisition. In this paper, we focus on the problem of automatically detecting and correcting the shadows from the LiDAR data by exploiting both the images and the point set laser reflectance. Building on the observation that shadow boundaries are characterized by both a significant color change and a stable laser reflectance, we propose to first detect shadow boundaries in the point set and then segment ground shadows using graph cuts in the image. Finally using a simplified illumination model we correct the shadows directly on the colored point sets. This joint exploitation of both the laser point set and the images renders our approach robust and efficient, avoiding user interaction.","PeriodicalId":425304,"journal":{"name":"2016 Fourth International Conference on 3D Vision (3DV)","volume":"26 10","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120922864","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
期刊
2016 Fourth International Conference on 3D Vision (3DV)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1