首页 > 最新文献

2016 Fourth International Conference on 3D Vision (3DV)最新文献

英文 中文
Optical Flow for Rigid Multi-Motion Scenes 刚性多运动场景的光流
Pub Date : 2016-10-01 DOI: 10.1109/3DV.2016.30
Tomas Gerlich, Jakob Eriksson
We observe that in many applications, the motion present in a scene is well characterized by a small number of (rigid) motion hypotheses. Based on this observation, we present rigid multi-motion optical flow (RMM). By restricting flow to one of several motion hypotheses, RMM produces more accurate optical flow than arbitrary motion models. We evaluate an algorithm based on RMM on a novel synthetic dataset, consisting of 12 photo-realistically rendered scenes containing rigid vehicular motion and a corresponding, exact, ground truth. On this dataset, we demonstrate a substantial advantage of RMM over general-purpose algorithms: going from 36% outliers with the DiscreteFlow algorithm, to 26% with ours, with a mean error reduction from 8.4px to 6.9px. We also perform qualitative evaluation on real-world imagery from traffic cameras.
我们观察到,在许多应用中,场景中存在的运动很好地表征为少量(刚性)运动假设。基于这一观察,我们提出了刚性多运动光流(RMM)。通过将流限制在几个运动假设中的一个,RMM产生比任意运动模型更精确的光流。我们在一个新的合成数据集上评估了一种基于RMM的算法,该数据集由12个逼真渲染的场景组成,其中包含刚性车辆运动和相应的,精确的地面真相。在这个数据集上,我们展示了RMM相对于通用算法的巨大优势:离散流算法的离群值从36%降至26%,平均误差从8.4px降至6.9px。我们还对来自交通摄像头的真实世界图像进行定性评估。
{"title":"Optical Flow for Rigid Multi-Motion Scenes","authors":"Tomas Gerlich, Jakob Eriksson","doi":"10.1109/3DV.2016.30","DOIUrl":"https://doi.org/10.1109/3DV.2016.30","url":null,"abstract":"We observe that in many applications, the motion present in a scene is well characterized by a small number of (rigid) motion hypotheses. Based on this observation, we present rigid multi-motion optical flow (RMM). By restricting flow to one of several motion hypotheses, RMM produces more accurate optical flow than arbitrary motion models. We evaluate an algorithm based on RMM on a novel synthetic dataset, consisting of 12 photo-realistically rendered scenes containing rigid vehicular motion and a corresponding, exact, ground truth. On this dataset, we demonstrate a substantial advantage of RMM over general-purpose algorithms: going from 36% outliers with the DiscreteFlow algorithm, to 26% with ours, with a mean error reduction from 8.4px to 6.9px. We also perform qualitative evaluation on real-world imagery from traffic cameras.","PeriodicalId":425304,"journal":{"name":"2016 Fourth International Conference on 3D Vision (3DV)","volume":"62 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132316697","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Learning a General-Purpose Confidence Measure Based on O(1) Features and a Smarter Aggregation Strategy for Semi Global Matching 基于O(1)特征的通用置信度学习和半全局匹配的智能聚合策略
Pub Date : 2016-10-01 DOI: 10.1109/3DV.2016.61
Matteo Poggi, S. Mattoccia
Inferring dense depth from stereo is crucial for several computer vision applications and Semi Global Matching (SGM) is often the preferred choice due to its good tradeoff between accuracy and computation requirements. Nevertheless, it suffers of two major issues: streaking artifacts caused by the Scanline Optimization (SO) approach, at the core of this algorithm, may lead to inaccurate results and the high memory footprint that may become prohibitive with high resolution images or devices with constrained resources. In this paper, we propose a smart scanline aggregation approach for SGM aimed at dealing with both issues. In particular, the contribution of this paper is threefold: i) leveraging on machine learning, proposes a novel generalpurpose confidence measure suited for any for stereo algorithm, based on O(1) features, that outperforms state of-the-art ii) taking advantage of this confidence measure proposes a smart aggregation strategy for SGM enabling significant improvements with a very small overhead iii) the overall strategy drastically reduces the memory footprint of SGM and, at the same time, improves its effectiveness and execution time. We provide extensive experimental results, including a cross-validation with multiple datasets (KITTI 2012, KITTI 2015 and Middlebury 2014).
从立体图像中推断密集深度对于许多计算机视觉应用至关重要,半全局匹配(SGM)通常是首选,因为它在精度和计算要求之间取得了良好的平衡。然而,它存在两个主要问题:该算法的核心是扫描线优化(SO)方法,这可能导致不准确的结果,以及高内存占用,这可能会使高分辨率图像或资源受限的设备变得令人难以接受。在本文中,我们提出了一种针对SGM的智能扫描线聚合方法,旨在解决这两个问题。本文的贡献主要体现在三个方面:i)利用机器学习,提出了一种新的通用置信度度量,适用于任何立体算法,基于O(1)特征,优于最先进的ii)利用这种置信度度量,为SGM提出了一种智能聚合策略,以非常小的开销实现显著改进iii)总体策略大大减少了SGM的内存占用,同时提高了其有效性和执行时间。我们提供了广泛的实验结果,包括多个数据集的交叉验证(KITTI 2012, KITTI 2015和Middlebury 2014)。
{"title":"Learning a General-Purpose Confidence Measure Based on O(1) Features and a Smarter Aggregation Strategy for Semi Global Matching","authors":"Matteo Poggi, S. Mattoccia","doi":"10.1109/3DV.2016.61","DOIUrl":"https://doi.org/10.1109/3DV.2016.61","url":null,"abstract":"Inferring dense depth from stereo is crucial for several computer vision applications and Semi Global Matching (SGM) is often the preferred choice due to its good tradeoff between accuracy and computation requirements. Nevertheless, it suffers of two major issues: streaking artifacts caused by the Scanline Optimization (SO) approach, at the core of this algorithm, may lead to inaccurate results and the high memory footprint that may become prohibitive with high resolution images or devices with constrained resources. In this paper, we propose a smart scanline aggregation approach for SGM aimed at dealing with both issues. In particular, the contribution of this paper is threefold: i) leveraging on machine learning, proposes a novel generalpurpose confidence measure suited for any for stereo algorithm, based on O(1) features, that outperforms state of-the-art ii) taking advantage of this confidence measure proposes a smart aggregation strategy for SGM enabling significant improvements with a very small overhead iii) the overall strategy drastically reduces the memory footprint of SGM and, at the same time, improves its effectiveness and execution time. We provide extensive experimental results, including a cross-validation with multiple datasets (KITTI 2012, KITTI 2015 and Middlebury 2014).","PeriodicalId":425304,"journal":{"name":"2016 Fourth International Conference on 3D Vision (3DV)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132419881","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 61
Structure from Category: A Generic and Prior-Less Approach 从类别结构:一个通用的和无先验的方法
Pub Date : 2016-10-01 DOI: 10.1109/3DV.2016.38
Chen Kong, Rui Zhu, Hamed Kiani Galoogahi, S. Lucey
Inferring the motion and shape of non-rigid objects from images has been widely explored by Non-Rigid Structure from Motion (NRSfM) algorithms. Despite their promising results, they often utilize additional constraints about the camera motion (e.g. temporal order) and the deformation of the object of interest, which are not always provided in real-world scenarios. This makes the application of NRSfM limited to very few deformable objects (e.g. human face and body). In this paper, we propose the concept of Structure from Category (SfC) to reconstruct 3D structure of generic objects solely from images with no shape and motion constraint (i.e. prior-less). Similar to the NRSfM approaches, SfC involves two steps: (i) correspondence, and (ii) inversion. Correspondence determines the location of key points across images of the same object category. Once established, the inverse problem of recovering the 3D structure from the 2D points is solved over an augmented sparse shape-space model. We validate our approach experimentally by reconstructing 3D structures of both synthetic and natural images, and demonstrate the superiority of our approach to the state-of-the-art low-rank NRSfM approaches.
从图像中推断非刚体物体的运动和形状已经被非刚体结构从运动(NRSfM)算法广泛探索。尽管他们的结果很有希望,但他们经常利用关于相机运动的额外约束(例如时间顺序)和感兴趣对象的变形,这些在现实世界中并不总是提供。这使得NRSfM的应用仅限于非常少的可变形对象(例如人脸和身体)。在本文中,我们提出了结构来自类别(SfC)的概念,用于仅从没有形状和运动约束(即无先验)的图像中重建一般物体的三维结构。与NRSfM方法类似,SfC涉及两个步骤:(i)对应,(ii)反转。对应关系决定了同一对象类别图像上关键点的位置。一旦建立,从二维点恢复三维结构的逆问题是在一个增广的稀疏形状空间模型上解决的。我们通过重建合成和自然图像的三维结构来验证我们的方法,并证明了我们的方法比最先进的低秩NRSfM方法的优越性。
{"title":"Structure from Category: A Generic and Prior-Less Approach","authors":"Chen Kong, Rui Zhu, Hamed Kiani Galoogahi, S. Lucey","doi":"10.1109/3DV.2016.38","DOIUrl":"https://doi.org/10.1109/3DV.2016.38","url":null,"abstract":"Inferring the motion and shape of non-rigid objects from images has been widely explored by Non-Rigid Structure from Motion (NRSfM) algorithms. Despite their promising results, they often utilize additional constraints about the camera motion (e.g. temporal order) and the deformation of the object of interest, which are not always provided in real-world scenarios. This makes the application of NRSfM limited to very few deformable objects (e.g. human face and body). In this paper, we propose the concept of Structure from Category (SfC) to reconstruct 3D structure of generic objects solely from images with no shape and motion constraint (i.e. prior-less). Similar to the NRSfM approaches, SfC involves two steps: (i) correspondence, and (ii) inversion. Correspondence determines the location of key points across images of the same object category. Once established, the inverse problem of recovering the 3D structure from the 2D points is solved over an augmented sparse shape-space model. We validate our approach experimentally by reconstructing 3D structures of both synthetic and natural images, and demonstrate the superiority of our approach to the state-of-the-art low-rank NRSfM approaches.","PeriodicalId":425304,"journal":{"name":"2016 Fourth International Conference on 3D Vision (3DV)","volume":"162 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134160425","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
Registration of Point Clouds Based on the Ratio of Bidirectional Distances 基于双向距离比的点云配准
Pub Date : 2016-10-01 DOI: 10.1109/3DV.2016.85
Jihua Zhu, Di Wang, Xiuxiu Bai, Huimin Lu, Congcong Jin, Zhongyu Li
Despite the fact that original Iterative Closest Point(ICP) algorithm has been widely used for registration, itcannot tackle the problem when two point clouds are par-tially overlapping. Accordingly, this paper proposes a ro-bust approach for the registration of partially overlappingpoint clouds. Given two initially posed clouds, it firstlybuilds up bilateral correspondence and computes bidirec-tional distances for each point in the data shape. Based onthe ratio of bidirectional distances, the exponential functionis selected and utilized to calculate the probability value,which can indicate whether the point pair belongs to theoverlapping part or not. Subsequently, the probability val-ue can be embedded into the least square function for reg-istration of partially overlapping point clouds and a novelvariant of ICP algorithm is presented to obtain the optimalrigid transformation. The proposed approach can achievegood registration of point clouds, even when their overlappercentage is low. Experimental results tested on public da-ta sets illustrate its superiority over previous approaches onrobustness.
尽管原始的迭代最近点(ICP)算法在配准中得到了广泛的应用,但它无法解决两个点云部分重叠的配准问题。据此,本文提出了一种局部重叠点云的准配方法。给定两个初始构成的云,它首先建立双边对应关系,并计算数据形状中每个点的双向距离。根据双向距离的比值,选择指数函数计算概率值,该概率值可以指示点对是否属于重叠部分。然后,将概率值嵌入到最小二乘函数中进行部分重叠点云的配准,并提出了一种新的ICP算法来获得最优刚体变换。该方法可以在点云重叠率较低的情况下实现较好的配准。在公共数据集上测试的实验结果表明,该方法在鲁棒性方面优于以往的方法。
{"title":"Registration of Point Clouds Based on the Ratio of Bidirectional Distances","authors":"Jihua Zhu, Di Wang, Xiuxiu Bai, Huimin Lu, Congcong Jin, Zhongyu Li","doi":"10.1109/3DV.2016.85","DOIUrl":"https://doi.org/10.1109/3DV.2016.85","url":null,"abstract":"Despite the fact that original Iterative Closest Point(ICP) algorithm has been widely used for registration, itcannot tackle the problem when two point clouds are par-tially overlapping. Accordingly, this paper proposes a ro-bust approach for the registration of partially overlappingpoint clouds. Given two initially posed clouds, it firstlybuilds up bilateral correspondence and computes bidirec-tional distances for each point in the data shape. Based onthe ratio of bidirectional distances, the exponential functionis selected and utilized to calculate the probability value,which can indicate whether the point pair belongs to theoverlapping part or not. Subsequently, the probability val-ue can be embedded into the least square function for reg-istration of partially overlapping point clouds and a novelvariant of ICP algorithm is presented to obtain the optimalrigid transformation. The proposed approach can achievegood registration of point clouds, even when their overlappercentage is low. Experimental results tested on public da-ta sets illustrate its superiority over previous approaches onrobustness.","PeriodicalId":425304,"journal":{"name":"2016 Fourth International Conference on 3D Vision (3DV)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115105106","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Automatic 3D Car Model Alignment for Mixed Image-Based Rendering 基于混合图像渲染的自动3D汽车模型对齐
Pub Date : 2016-10-01 DOI: 10.1109/3DV.2016.37
Rodrigo Ortiz Cayon, Abdelaziz Djelouah, Francisco Massa, Mathieu Aubry, G. Drettakis
Image-Based Rendering (IBR) allows good-quality free-viewpoint navigation in urban scenes, but suffers from artifacts on poorly reconstructed objects, e.g., reflective surfaces such as cars. To alleviate this problem, we propose a method that automatically identifies stock 3D models, aligns them in the 3D scene and performs morphing to better capture image contours. We do this by first adapting learning-based methods to detect and identify an object class/pose in images. We then propose a method which exploits all available information, namely partial and inaccurate 3D reconstruction, multi-view calibration, image contours and the 3D model to achieve accurate object alignment suitable for subsequent morphing. These steps provide models which are well-aligned in 3D and to contours in all the images of the multi-view dataset, allowing us to use the resulting model in our mixed IBR algorithm. Our results show significant improvement in image quality for free-viewpoint IBR, especially when moving far from the captured viewpoints.
基于图像的渲染(IBR)可以在城市场景中实现高质量的自由视点导航,但在重建效果较差的物体(如汽车等反射表面)上存在人工影响。为了缓解这一问题,我们提出了一种自动识别库存3D模型,在3D场景中对齐它们并执行变形以更好地捕获图像轮廓的方法。我们首先采用基于学习的方法来检测和识别图像中的对象类别/姿势。然后,我们提出了一种利用所有可用信息的方法,即部分和不准确的3D重建,多视图校准,图像轮廓和3D模型,以实现适合后续变形的精确目标对齐。这些步骤提供了在3D和多视图数据集中所有图像的轮廓中良好对齐的模型,允许我们在混合IBR算法中使用生成的模型。我们的研究结果表明,自由视点IBR的图像质量有了显著改善,尤其是在远离捕获视点的情况下。
{"title":"Automatic 3D Car Model Alignment for Mixed Image-Based Rendering","authors":"Rodrigo Ortiz Cayon, Abdelaziz Djelouah, Francisco Massa, Mathieu Aubry, G. Drettakis","doi":"10.1109/3DV.2016.37","DOIUrl":"https://doi.org/10.1109/3DV.2016.37","url":null,"abstract":"Image-Based Rendering (IBR) allows good-quality free-viewpoint navigation in urban scenes, but suffers from artifacts on poorly reconstructed objects, e.g., reflective surfaces such as cars. To alleviate this problem, we propose a method that automatically identifies stock 3D models, aligns them in the 3D scene and performs morphing to better capture image contours. We do this by first adapting learning-based methods to detect and identify an object class/pose in images. We then propose a method which exploits all available information, namely partial and inaccurate 3D reconstruction, multi-view calibration, image contours and the 3D model to achieve accurate object alignment suitable for subsequent morphing. These steps provide models which are well-aligned in 3D and to contours in all the images of the multi-view dataset, allowing us to use the resulting model in our mixed IBR algorithm. Our results show significant improvement in image quality for free-viewpoint IBR, especially when moving far from the captured viewpoints.","PeriodicalId":425304,"journal":{"name":"2016 Fourth International Conference on 3D Vision (3DV)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128613110","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Robust Feature-Preserving Denoising of 3D Point Clouds 三维点云的鲁棒特征保持去噪
Pub Date : 2016-10-01 DOI: 10.1109/3DV.2016.17
Sk. Mohammadul Haque, V. Govindu
The increased availability of point cloud data in recent years has lead to a concomitant requirement for high quality denoising methods. This is particularly the case with data obtained using depth cameras or from multi-view stereo reconstruction as both approaches result in noisy point clouds and include significant outliers. Most of the available denoising methods in the literature are not sufficiently robust to outliers and/or are unable to preserve fine-scale 3D features in the denoised representations. In this paper we propose an approach to point cloud denoising that is both robust to outliers and capable of preserving fine-scale 3D features. We identify and remove outliers by utilising a dissimilarity measure based on point positions and their corresponding normals. Subsequently, we use a robust approach to estimate surface point positions in a manner designed to preserve sharp and fine-scale 3D features. We demonstrate the efficacy of our approach and compare with similar methods in the literature by means of experiments on synthetic and real data including large-scale 3D reconstructions of heritage monuments.
近年来,随着点云数据可用性的增加,对高质量的去噪方法提出了更高的要求。使用深度相机或多视点立体重建获得的数据尤其如此,因为这两种方法都会产生嘈杂的点云,并包含显著的异常值。文献中大多数可用的去噪方法对异常值的鲁棒性不够,或者无法在去噪后的表示中保留精细尺度的3D特征。在本文中,我们提出了一种既对异常值鲁棒又能保留精细尺度三维特征的点云去噪方法。我们通过利用基于点位置及其相应法线的不相似性度量来识别和去除异常值。随后,我们使用一种鲁棒的方法来估计表面点的位置,以保持清晰和精细的3D特征。我们通过对文物古迹大规模三维重建的合成和真实数据进行实验,证明了我们方法的有效性,并与文献中的类似方法进行了比较。
{"title":"Robust Feature-Preserving Denoising of 3D Point Clouds","authors":"Sk. Mohammadul Haque, V. Govindu","doi":"10.1109/3DV.2016.17","DOIUrl":"https://doi.org/10.1109/3DV.2016.17","url":null,"abstract":"The increased availability of point cloud data in recent years has lead to a concomitant requirement for high quality denoising methods. This is particularly the case with data obtained using depth cameras or from multi-view stereo reconstruction as both approaches result in noisy point clouds and include significant outliers. Most of the available denoising methods in the literature are not sufficiently robust to outliers and/or are unable to preserve fine-scale 3D features in the denoised representations. In this paper we propose an approach to point cloud denoising that is both robust to outliers and capable of preserving fine-scale 3D features. We identify and remove outliers by utilising a dissimilarity measure based on point positions and their corresponding normals. Subsequently, we use a robust approach to estimate surface point positions in a manner designed to preserve sharp and fine-scale 3D features. We demonstrate the efficacy of our approach and compare with similar methods in the literature by means of experiments on synthetic and real data including large-scale 3D reconstructions of heritage monuments.","PeriodicalId":425304,"journal":{"name":"2016 Fourth International Conference on 3D Vision (3DV)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129082655","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Absolute Pose and Structure from Motion for Surfaces of Revolution: Minimal Problems Using Apparent Contours 绝对姿态和结构从运动的旋转曲面:最小的问题使用表观轮廓
Pub Date : 2016-10-01 DOI: 10.1109/3DV.2016.86
Cody J. Phillips, Kostas Daniilidis
The class of objects that can be represented by surfaces of revolution (SoRs) is highly prevalent in human work and living spaces. Due to their prevalence and convenient geometric properties, SoRs have been employed over the past thirty years for single-view camera calibration and pose estimation, and have been studied in terms of SoR object reconstruction and recognition. Such treatment has provided techniques for the automatic identification and classification of important SoR structures, such as apparent contours, cross sections, bitangent points, creases, and inflections. The presence of these structures are crucial to most SoR-based image metrology algorithms. This paper develops single-view and two-view pose recovery and reconstruction formulations that only require apparent contours, and no other SoR features.The primary objective of this paper is to present and experimentallyvalidate the minimal problems pertaining toSoR metrology from apparent contours. For a single view with a known reference model, this includes absolute pose recovery. For many views and no reference model this is extended to structure from motion (SfM). Assuming apparent contours as input that have been identified and segmented with reasonable accuracy, the minimal problems aredemonstrated to produce accurate SoR pose and shape results when used as part of a RANSAC-based hypothesis generation and evaluation pipeline.
在人类的工作和生活空间中,可以用旋转曲面(sor)来表示的物体类别非常普遍。由于其普遍存在和方便的几何特性,在过去的三十年中,SoR被用于单视图相机校准和姿态估计,并在SoR目标重建和识别方面进行了研究。这种处理为重要的SoR结构的自动识别和分类提供了技术,如表观轮廓、横截面、切点、折痕和弯曲。这些结构的存在对于大多数基于sor的图像测量算法至关重要。本文开发了单视图和双视图姿态恢复和重建公式,仅需要表观轮廓,而不需要其他SoR特征。本文的主要目的是提出和实验验证有关的最小问题,从表观轮廓测量。对于具有已知参考模型的单个视图,这包括绝对姿势恢复。对于许多视图和没有参考模型,这是扩展到结构从运动(SfM)。假设视轮廓作为输入,以合理的精度识别和分割,最小的问题被证明可以产生准确的SoR姿态和形状结果,当用作基于ransac的假设生成和评估管道的一部分时。
{"title":"Absolute Pose and Structure from Motion for Surfaces of Revolution: Minimal Problems Using Apparent Contours","authors":"Cody J. Phillips, Kostas Daniilidis","doi":"10.1109/3DV.2016.86","DOIUrl":"https://doi.org/10.1109/3DV.2016.86","url":null,"abstract":"The class of objects that can be represented by surfaces of revolution (SoRs) is highly prevalent in human work and living spaces. Due to their prevalence and convenient geometric properties, SoRs have been employed over the past thirty years for single-view camera calibration and pose estimation, and have been studied in terms of SoR object reconstruction and recognition. Such treatment has provided techniques for the automatic identification and classification of important SoR structures, such as apparent contours, cross sections, bitangent points, creases, and inflections. The presence of these structures are crucial to most SoR-based image metrology algorithms. This paper develops single-view and two-view pose recovery and reconstruction formulations that only require apparent contours, and no other SoR features.The primary objective of this paper is to present and experimentallyvalidate the minimal problems pertaining toSoR metrology from apparent contours. For a single view with a known reference model, this includes absolute pose recovery. For many views and no reference model this is extended to structure from motion (SfM). Assuming apparent contours as input that have been identified and segmented with reasonable accuracy, the minimal problems aredemonstrated to produce accurate SoR pose and shape results when used as part of a RANSAC-based hypothesis generation and evaluation pipeline.","PeriodicalId":425304,"journal":{"name":"2016 Fourth International Conference on 3D Vision (3DV)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130826788","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Multi-View Inpainting for Image-Based Scene Editing and Rendering 多视图绘画基于图像的场景编辑和渲染
Pub Date : 2016-10-01 DOI: 10.1109/3DV.2016.44
T. Thonat, Eli Shechtman, Sylvain Paris, G. Drettakis
We propose a method to remove objects such as people and cars from multi-view urban image datasets, enabling free-viewpoint IBR in the edited scenes. Our method combines information from multi-view 3D reconstruction with image inpainting techniques, by formulating the problem as an optimization of a global patch-based objective function. We use Image-Based Rendering (IBR) techniques to reproject information from neighboring views, and 3D multi-view stereo reconstruction to perform multiview coherent initialization for inpainting of pixels not filled by reprojection. Our algorithm performs multi-view consistent inpainting for color and 3D by blending reprojections with patch-based image inpainting. We run our algorithm on casually captured datasets, and Google StreetViewdata, removing objects cars, people and pillars, showing that our approach produces results of sufficient quality for free-viewpoint IBR on "cleaned up" scenes, as well as IBR scene editing, such as limited motion of real objects.
我们提出了一种从多视点城市图像数据集中去除人物和汽车等物体的方法,从而在编辑后的场景中实现自由视点IBR。我们的方法结合了多视图三维重建和图像绘制技术的信息,通过将问题表述为基于全局补丁的目标函数的优化。我们使用基于图像的渲染(IBR)技术来重新投影来自相邻视图的信息,并使用3D多视图立体重建来执行多视图相干初始化,以重新绘制未被重新投影填充的像素。我们的算法通过混合重投影和基于补丁的图像绘制来执行彩色和3D的多视图一致性绘制。我们在随机捕获的数据集和Google StreetViewdata上运行我们的算法,删除对象汽车,人和柱子,表明我们的方法产生了足够质量的结果,用于“清理”场景的自由视点IBR,以及IBR场景编辑,例如真实物体的有限运动。
{"title":"Multi-View Inpainting for Image-Based Scene Editing and Rendering","authors":"T. Thonat, Eli Shechtman, Sylvain Paris, G. Drettakis","doi":"10.1109/3DV.2016.44","DOIUrl":"https://doi.org/10.1109/3DV.2016.44","url":null,"abstract":"We propose a method to remove objects such as people and cars from multi-view urban image datasets, enabling free-viewpoint IBR in the edited scenes. Our method combines information from multi-view 3D reconstruction with image inpainting techniques, by formulating the problem as an optimization of a global patch-based objective function. We use Image-Based Rendering (IBR) techniques to reproject information from neighboring views, and 3D multi-view stereo reconstruction to perform multiview coherent initialization for inpainting of pixels not filled by reprojection. Our algorithm performs multi-view consistent inpainting for color and 3D by blending reprojections with patch-based image inpainting. We run our algorithm on casually captured datasets, and Google StreetViewdata, removing objects cars, people and pillars, showing that our approach produces results of sufficient quality for free-viewpoint IBR on \"cleaned up\" scenes, as well as IBR scene editing, such as limited motion of real objects.","PeriodicalId":425304,"journal":{"name":"2016 Fourth International Conference on 3D Vision (3DV)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115284977","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Face Reconstruction on Mobile Devices Using a Height Map Shape Model and Fast Regularization 基于高度映射形状模型和快速正则化的移动设备人脸重建
Pub Date : 2016-10-01 DOI: 10.1109/3DV.2016.59
Fabio Maninchedda, Christian Häne, Martin R. Oswald, M. Pollefeys
We present a system which is able to reconstruct human faces on mobile devices with only on-device processing using the sensors which are typically built into a current commodity smart phone. Such technology can for example be used for facial authentication purposes or as a fast preview for further post-processing. Our method uses recently proposed techniques which compute depth maps by passive multi-view stereo directly on the device. We propose an efficient method which recovers the geometry of the face from the typically noisy point cloud. First, we show that we can safely restrict the reconstruction to a 2.5D height map representation. Therefore we then propose a novel low dimensional height map shape model for faces which can be fitted to the input data efficiently even on a mobile phone. In order to be able to represent instance specific shape details, such as moles, we augment the reconstruction from the shape model with a distance map which can be regularized efficiently. We thoroughly evaluate our approach on synthetic and real data, thereby we use both high resolution depth data acquired using high quality multi-view stereo and depth data directly computed on mobile phones.
我们提出了一个系统,该系统能够在移动设备上重建人脸,仅使用当前商品智能手机中通常内置的传感器进行设备上处理。例如,这种技术可以用于面部认证目的或作为进一步后处理的快速预览。我们的方法使用了最近提出的技术,即直接在设备上通过被动多视立体来计算深度图。提出了一种从典型噪声点云中恢复人脸几何形状的有效方法。首先,我们证明了我们可以安全地将重建限制为2.5D高度地图表示。因此,我们提出了一种新颖的低维高度地图形状模型,该模型可以在手机上有效地拟合输入数据。为了能够表示实例特定的形状细节,例如痣,我们用一个可以有效正则化的距离图来增强形状模型的重建。我们在合成数据和真实数据上全面评估了我们的方法,因此我们既使用高质量多视角立体图像获得的高分辨率深度数据,也使用直接在手机上计算的深度数据。
{"title":"Face Reconstruction on Mobile Devices Using a Height Map Shape Model and Fast Regularization","authors":"Fabio Maninchedda, Christian Häne, Martin R. Oswald, M. Pollefeys","doi":"10.1109/3DV.2016.59","DOIUrl":"https://doi.org/10.1109/3DV.2016.59","url":null,"abstract":"We present a system which is able to reconstruct human faces on mobile devices with only on-device processing using the sensors which are typically built into a current commodity smart phone. Such technology can for example be used for facial authentication purposes or as a fast preview for further post-processing. Our method uses recently proposed techniques which compute depth maps by passive multi-view stereo directly on the device. We propose an efficient method which recovers the geometry of the face from the typically noisy point cloud. First, we show that we can safely restrict the reconstruction to a 2.5D height map representation. Therefore we then propose a novel low dimensional height map shape model for faces which can be fitted to the input data efficiently even on a mobile phone. In order to be able to represent instance specific shape details, such as moles, we augment the reconstruction from the shape model with a distance map which can be regularized efficiently. We thoroughly evaluate our approach on synthetic and real data, thereby we use both high resolution depth data acquired using high quality multi-view stereo and depth data directly computed on mobile phones.","PeriodicalId":425304,"journal":{"name":"2016 Fourth International Conference on 3D Vision (3DV)","volume":"99 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124931951","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Fast Obstacle Detection Using Sparse Edge-Based Disparity Maps 基于稀疏边缘的视差图的快速障碍物检测
Pub Date : 2016-10-01 DOI: 10.1109/3DV.2016.80
Dexmont Alejandro Peãa Carrillo, Alistair Sutherland
This paper presents a fast approach for computing image stixels from a sparse edge-based disparity map. The use of edge-based disparity maps speeds up the computation of the stixels as only a few pixels must be processed compared to approaches which use dense disparity maps. The proposed approach produces as output the stixels in one of the views of the stereo-pair and a segmentation of the edge-points into obstacle. Additionally the proposed approach allows the identification of partially occluded objects by allowing more than one stixel per image column. The proposed approach is fast to compute with no loss on accuracy.
本文提出了一种从稀疏边缘视差图中快速计算图像像素的方法。与使用密集视差图的方法相比,使用基于边缘的视差图加快了像素的计算速度,因为只需要处理几个像素。该方法在立体对的一个视图中产生像素作为输出,并将边缘点分割成障碍物。此外,所提出的方法允许通过每个图像列允许多个像素来识别部分遮挡的对象。该方法计算速度快,且精度无损失。
{"title":"Fast Obstacle Detection Using Sparse Edge-Based Disparity Maps","authors":"Dexmont Alejandro Peãa Carrillo, Alistair Sutherland","doi":"10.1109/3DV.2016.80","DOIUrl":"https://doi.org/10.1109/3DV.2016.80","url":null,"abstract":"This paper presents a fast approach for computing image stixels from a sparse edge-based disparity map. The use of edge-based disparity maps speeds up the computation of the stixels as only a few pixels must be processed compared to approaches which use dense disparity maps. The proposed approach produces as output the stixels in one of the views of the stereo-pair and a segmentation of the edge-points into obstacle. Additionally the proposed approach allows the identification of partially occluded objects by allowing more than one stixel per image column. The proposed approach is fast to compute with no loss on accuracy.","PeriodicalId":425304,"journal":{"name":"2016 Fourth International Conference on 3D Vision (3DV)","volume":"65 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121507380","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
期刊
2016 Fourth International Conference on 3D Vision (3DV)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1