首页 > 最新文献

2014 2nd International Conference on 3D Vision最新文献

英文 中文
A Data-Driven Regularization Model for Stereo and Flow 立体与流的数据驱动正则化模型
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.97
D. Wei, Ce Liu, W. Freeman
Data-driven techniques can reliably build semantic correspondence among images. In this paper, we present a new regularization model for stereo or flow through transferring the shape information of the disparity or flow from semantically matched patches in the training database. Compared to previous regularization models based on image appearance alone, we can better resolve local ambiguity of the disparity or flow by considering the semantic information without explicit object modeling. We incorporate this data-driven regularization model into a standard Markov Random Field (MRF) model, inferred with a gradient descent algorithm and learned with a discriminative learning approach. Compared to prior state-of-the-art methods, our full model achieves comparable or better results on the KITTI stereo and flow datasets, and improves results on the Sintel Flow dataset under an online estimation setting.
数据驱动技术可以可靠地建立图像之间的语义对应关系。本文提出了一种新的立体或流的正则化模型,该模型通过对训练数据库中语义匹配块的视差或流的形状信息进行转换。与以往单纯基于图像外观的正则化模型相比,我们在没有明确对象建模的情况下,通过考虑语义信息,可以更好地解决视差或流的局部模糊问题。我们将这个数据驱动的正则化模型合并到一个标准的马尔可夫随机场(MRF)模型中,使用梯度下降算法进行推断,并使用判别学习方法进行学习。与之前最先进的方法相比,我们的完整模型在KITTI立体和流动数据集上取得了相当或更好的结果,并在在线估计设置下改善了sinintel流动数据集的结果。
{"title":"A Data-Driven Regularization Model for Stereo and Flow","authors":"D. Wei, Ce Liu, W. Freeman","doi":"10.1109/3DV.2014.97","DOIUrl":"https://doi.org/10.1109/3DV.2014.97","url":null,"abstract":"Data-driven techniques can reliably build semantic correspondence among images. In this paper, we present a new regularization model for stereo or flow through transferring the shape information of the disparity or flow from semantically matched patches in the training database. Compared to previous regularization models based on image appearance alone, we can better resolve local ambiguity of the disparity or flow by considering the semantic information without explicit object modeling. We incorporate this data-driven regularization model into a standard Markov Random Field (MRF) model, inferred with a gradient descent algorithm and learned with a discriminative learning approach. Compared to prior state-of-the-art methods, our full model achieves comparable or better results on the KITTI stereo and flow datasets, and improves results on the Sintel Flow dataset under an online estimation setting.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115740362","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Classification of Vehicle Parts in Unstructured 3D Point Clouds 非结构化三维点云中车辆部件的分类
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.58
Allan Zelener, Philippos Mordohai, I. Stamos
Unprecedented amounts of 3D data can be acquired in urban environments, but their use for scene understanding is challenging due to varying data resolution and variability of objects in the same class. An additional challenge is due to the nature of the point clouds themselves, since they lack detailed geometric or semantic information that would aid scene understanding. In this paper we present a general algorithm for segmenting and jointly classifying object parts and the object itself. Our pipeline consists of local feature extraction, robust RANSAC part segmentation, part-level feature extraction, a structured model for parts in objects, and classification using state-of-the-art classifiers. We have tested this pipeline in a very challenging dataset that consists of real world scans of vehicles. Our contributions include the development of a segmentation and classification pipeline for objects and their parts, and a method for segmentation that is robust to the complexity of unstructured 3D points clouds, as well as a part ordering strategy for the sequential structured model and a joint feature representation between object parts.
在城市环境中可以获得前所未有的3D数据量,但由于同一类对象的不同数据分辨率和可变性,它们用于场景理解具有挑战性。另一个挑战是由于点云本身的性质,因为它们缺乏有助于场景理解的详细几何或语义信息。本文提出了一种用于分割和联合分类目标部分和目标本身的通用算法。我们的管道包括局部特征提取、鲁棒RANSAC部件分割、部件级特征提取、对象中部件的结构化模型,以及使用最先进的分类器进行分类。我们已经在一个非常具有挑战性的数据集中测试了这个管道,该数据集由真实世界的车辆扫描组成。我们的贡献包括开发对象及其部件的分割和分类管道,以及一种对非结构化3D点云的复杂性具有鲁棒性的分割方法,以及用于顺序结构化模型的部件排序策略和对象部件之间的联合特征表示。
{"title":"Classification of Vehicle Parts in Unstructured 3D Point Clouds","authors":"Allan Zelener, Philippos Mordohai, I. Stamos","doi":"10.1109/3DV.2014.58","DOIUrl":"https://doi.org/10.1109/3DV.2014.58","url":null,"abstract":"Unprecedented amounts of 3D data can be acquired in urban environments, but their use for scene understanding is challenging due to varying data resolution and variability of objects in the same class. An additional challenge is due to the nature of the point clouds themselves, since they lack detailed geometric or semantic information that would aid scene understanding. In this paper we present a general algorithm for segmenting and jointly classifying object parts and the object itself. Our pipeline consists of local feature extraction, robust RANSAC part segmentation, part-level feature extraction, a structured model for parts in objects, and classification using state-of-the-art classifiers. We have tested this pipeline in a very challenging dataset that consists of real world scans of vehicles. Our contributions include the development of a segmentation and classification pipeline for objects and their parts, and a method for segmentation that is robust to the complexity of unstructured 3D points clouds, as well as a part ordering strategy for the sequential structured model and a joint feature representation between object parts.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125423148","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
High-Quality Depth Recovery via Interactive Multi-view Stereo 通过交互式多视图立体高质量深度恢复
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.55
Weifeng Chen, Guofeng Zhang, Xiaojun Xiang, Jiaya Jia, H. Bao
Although multi-view stereo has been extensively studied during the past decades, automatically computing high-quality dense depth information from captured images/videos is still quite difficult. Many factors, such as serious occlusion, large texture less regions and strong reflection, easily cause erroneous depth recovery. In this paper, we present a novel semi-automatic multi-view stereo system, which can quickly create and repair depth from a monocular sequence taken by a freely moving camera. One of our main contributions is that we propose a novel multi-view stereo model incorporating prior constraints indicated by user interaction, which makes it possible to even handle Non-Lambertian surface that surely violates the photo-consistency constraint. Users only need to provide a coarse segmentation and a few user interactions, our system can automatically correct depth and refine boundary. With other priors and occlusion handling, the erroneous depth can be effectively corrected even for very challenging examples that are difficult for state-of-the-art methods.
尽管在过去的几十年里,人们对多视点立体图像进行了广泛的研究,但从捕获的图像/视频中自动计算高质量的密集深度信息仍然非常困难。严重的遮挡、大的纹理少的区域、强的反射等因素容易导致深度恢复错误。在本文中,我们提出了一种新的半自动多视点立体系统,该系统可以从自由移动的相机拍摄的单目序列中快速创建和修复深度。我们的主要贡献之一是我们提出了一种新的多视图立体模型,该模型结合了由用户交互指示的先验约束,这使得处理违反光一致性约束的非兰伯曲面成为可能。用户只需要提供一个粗略的分割和少量的用户交互,我们的系统就可以自动校正深度和细化边界。与其他先验和遮挡处理,错误的深度可以有效地纠正,即使是非常具有挑战性的例子,是最先进的方法是困难的。
{"title":"High-Quality Depth Recovery via Interactive Multi-view Stereo","authors":"Weifeng Chen, Guofeng Zhang, Xiaojun Xiang, Jiaya Jia, H. Bao","doi":"10.1109/3DV.2014.55","DOIUrl":"https://doi.org/10.1109/3DV.2014.55","url":null,"abstract":"Although multi-view stereo has been extensively studied during the past decades, automatically computing high-quality dense depth information from captured images/videos is still quite difficult. Many factors, such as serious occlusion, large texture less regions and strong reflection, easily cause erroneous depth recovery. In this paper, we present a novel semi-automatic multi-view stereo system, which can quickly create and repair depth from a monocular sequence taken by a freely moving camera. One of our main contributions is that we propose a novel multi-view stereo model incorporating prior constraints indicated by user interaction, which makes it possible to even handle Non-Lambertian surface that surely violates the photo-consistency constraint. Users only need to provide a coarse segmentation and a few user interactions, our system can automatically correct depth and refine boundary. With other priors and occlusion handling, the erroneous depth can be effectively corrected even for very challenging examples that are difficult for state-of-the-art methods.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"59 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114393387","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Building Modeling through Enclosure Reasoning 通过围合推理进行建筑建模
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.65
Adam Stambler, Daniel F. Huber
This paper introduces a method for automatically transforming a point cloud from a laser scanner into a volumetric 3D building model based on the new concept of enclosure reasoning. Rather than simply classifying and modeling building surfaces independently or with pair wise contextual relationships, this work introduces room, floor and building level reasoning. Enclosure reasoning premises that rooms are cycles of walls enclosing free interior space. These cycles should be of minimum description length (MDL) and obey the statistical priors expected for rooms. Floors and buildings then contain the best coverage of the mostly likely rooms. This allows the pipeline to generate higher fidelity models by performing modeling and recognition jointly over the entire building at once. The complete pipeline takes raw, registered laser scan surveys of a single building. It extracts the most likely smooth architectural surfaces, locates the building, and generates wall hypotheses. The algorithm then optimizes the model by growing, merging, and pruning these hypotheses to generate the most likely rooms, floors, and building in the presence of significant clutter.
本文介绍了一种基于封闭推理的新概念,将激光扫描仪上的点云自动转换为三维建筑体模型的方法。这项工作不是简单地对建筑表面进行独立分类和建模,也不是使用成对的上下文关系,而是引入了房间、楼层和建筑水平的推理。封闭推理的前提是房间是包围自由内部空间的墙壁循环。这些循环应该具有最小描述长度(MDL),并符合房间预期的统计先验。楼层和建筑物可以最好地覆盖最可能出现的房间。这允许管道通过在整个建筑上同时执行建模和识别来生成更高保真度的模型。完整的管道对一栋建筑进行原始的、注册的激光扫描调查。它提取最可能光滑的建筑表面,定位建筑,并产生墙壁假设。然后,算法通过增长、合并和修剪这些假设来优化模型,以在存在大量杂乱的情况下生成最可能的房间、楼层和建筑物。
{"title":"Building Modeling through Enclosure Reasoning","authors":"Adam Stambler, Daniel F. Huber","doi":"10.1109/3DV.2014.65","DOIUrl":"https://doi.org/10.1109/3DV.2014.65","url":null,"abstract":"This paper introduces a method for automatically transforming a point cloud from a laser scanner into a volumetric 3D building model based on the new concept of enclosure reasoning. Rather than simply classifying and modeling building surfaces independently or with pair wise contextual relationships, this work introduces room, floor and building level reasoning. Enclosure reasoning premises that rooms are cycles of walls enclosing free interior space. These cycles should be of minimum description length (MDL) and obey the statistical priors expected for rooms. Floors and buildings then contain the best coverage of the mostly likely rooms. This allows the pipeline to generate higher fidelity models by performing modeling and recognition jointly over the entire building at once. The complete pipeline takes raw, registered laser scan surveys of a single building. It extracts the most likely smooth architectural surfaces, locates the building, and generates wall hypotheses. The algorithm then optimizes the model by growing, merging, and pruning these hypotheses to generate the most likely rooms, floors, and building in the presence of significant clutter.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129413942","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
A Real-Time View-Dependent Shape Optimization for High Quality Free-Viewpoint Rendering of 3D Video 面向高质量自由视点3D视频渲染的实时视相关形状优化
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.28
S. Nobuhara, Wei Ning, T. Matsuyama
This paper is aimed at proposing a new high quality free-viewpoint rendering algorithm of 3D video. The main challenge on visualizing 3D video is how to utilize the original multi-view images used to estimate the 3D surface, and how to manage the mismatches between them due to calibration and reconstruction errors. The key idea to solve this problem is to optimize the 3D shape on a per-viewpoint basis on the fly. Given a virtual viewpoint for visualization, our algorithm optimizes the 3D shape so as to maximize the photo-consistency over the surface visible from the virtual viewpoint. An evaluation demonstrates that our method outperforms the state-of-the-art rendering qualitatively and quantitatively.
本文旨在提出一种新的高质量的三维视频自由视点绘制算法。如何利用原始的多视点图像来估计三维表面,以及如何处理由于校准和重建误差而导致的多视点图像之间的不匹配,是三维视频可视化的主要挑战。解决这个问题的关键思想是在每个视点的基础上实时优化3D形状。给定一个用于可视化的虚拟视点,我们的算法优化了三维形状,以最大限度地提高从虚拟视点可见的表面上的照片一致性。评估表明,我们的方法在质量和数量上都优于最先进的渲染。
{"title":"A Real-Time View-Dependent Shape Optimization for High Quality Free-Viewpoint Rendering of 3D Video","authors":"S. Nobuhara, Wei Ning, T. Matsuyama","doi":"10.1109/3DV.2014.28","DOIUrl":"https://doi.org/10.1109/3DV.2014.28","url":null,"abstract":"This paper is aimed at proposing a new high quality free-viewpoint rendering algorithm of 3D video. The main challenge on visualizing 3D video is how to utilize the original multi-view images used to estimate the 3D surface, and how to manage the mismatches between them due to calibration and reconstruction errors. The key idea to solve this problem is to optimize the 3D shape on a per-viewpoint basis on the fly. Given a virtual viewpoint for visualization, our algorithm optimizes the 3D shape so as to maximize the photo-consistency over the surface visible from the virtual viewpoint. An evaluation demonstrates that our method outperforms the state-of-the-art rendering qualitatively and quantitatively.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124577817","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Matching Many Identical Features of Planar Urban Facades Using Global Regularity 利用全局规则匹配平面城市立面许多相同特征
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.107
Eduardo B. Almeida, D. Cooper
Reasonable computation and accurate camera calibration require matching many interest points over long baselines. This is a difficult problem requiring better solutions than presently exist for urban scenes involving large buildings containing many windows since windows in a facade all have the same texture and, therefore, cannot be distinguished from one another based solely on appearance. Hence, the usual approach to feature detection and matching, such as use of SIFT, does not work in these scenes. A novel algorithm is introduced to provide correspondences for multiple repeating feature patterns seen under significant viewpoint changes. Most existing appearance-based algorithms cannot handle highly repetitive textures due to the match location ambiguity. However, the target structure provides a rich set of repeating features to be matched and tracked across multiple views, thus potentially improving camera estimation accuracy. The proposed method also exploits the geometric structure of regular grids of repeating features on planar surfaces.
合理的计算和精确的摄像机标定需要在长基线上匹配许多兴趣点。这是一个困难的问题,需要比目前存在的包含许多窗户的大型建筑的城市场景更好的解决方案,因为立面上的窗户都具有相同的纹理,因此不能仅仅基于外观来区分彼此。因此,通常的特征检测和匹配方法,如使用SIFT,在这些场景中不起作用。提出了一种新的算法,为视点显著变化下出现的多个重复特征模式提供对应关系。大多数现有的基于外观的算法由于匹配位置的模糊性而无法处理高度重复的纹理。然而,目标结构提供了一组丰富的重复特征,可以跨多个视图进行匹配和跟踪,从而潜在地提高相机估计精度。该方法还利用了平面上重复特征的规则网格的几何结构。
{"title":"Matching Many Identical Features of Planar Urban Facades Using Global Regularity","authors":"Eduardo B. Almeida, D. Cooper","doi":"10.1109/3DV.2014.107","DOIUrl":"https://doi.org/10.1109/3DV.2014.107","url":null,"abstract":"Reasonable computation and accurate camera calibration require matching many interest points over long baselines. This is a difficult problem requiring better solutions than presently exist for urban scenes involving large buildings containing many windows since windows in a facade all have the same texture and, therefore, cannot be distinguished from one another based solely on appearance. Hence, the usual approach to feature detection and matching, such as use of SIFT, does not work in these scenes. A novel algorithm is introduced to provide correspondences for multiple repeating feature patterns seen under significant viewpoint changes. Most existing appearance-based algorithms cannot handle highly repetitive textures due to the match location ambiguity. However, the target structure provides a rich set of repeating features to be matched and tracked across multiple views, thus potentially improving camera estimation accuracy. The proposed method also exploits the geometric structure of regular grids of repeating features on planar surfaces.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"112 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130377306","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Non-rigid registration with reliable distance field for dynamic shape completion 具有可靠距离场的非刚性配准,用于动态形状补全
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.111
Kent Fujiwara, Hiroshi Kawasaki, R. Sagawa, K. Ogawara, K. Ikeuchi
We propose a non-rigid registration method for completion of dynamic shapes with occlusion. Our method is based on the idea that an occluded region in a certain frame should be visible in another frame and that local regions should be moving rigidly when the motion is small. We achieve this with a novel reliable distance field (DF) for non-rigid registration with missing regions. We first fit a pseudo-surface onto the input shape using a surface reconstruction method. We then calculate the difference between the DF of the input shape and the pseudo-surface. We define the areas with large difference as unreliable, as these areas indicate that the original shape cannot be found nearby. We then conduct non-rigid registration using local rigid transformations to match the source and target data at visible regions and maintain the original shape as much as possible in occluded regions. The experimental results demonstrate that our method is capable of accurately filling in the missing regions using the shape information from prior or posterior frames. By sequentially processing the data, our method is also capable of completing an entire sequence with missing regions.
提出了一种非刚性配准方法,用于遮挡动态形状的补全。我们的方法是基于这样的思想,即某一帧中被遮挡的区域在另一帧中应该是可见的,并且当运动很小时,局部区域应该是刚性运动的。我们用一种新颖的可靠距离场(DF)来实现缺失区域的非刚性配准。我们首先使用曲面重建方法将伪曲面拟合到输入形状上。然后,我们计算输入形状的DF与伪曲面之间的差。我们将差异较大的区域定义为不可靠区域,因为这些区域表明附近找不到原始形状。然后,我们使用局部刚性变换进行非刚性配准,在可见区域匹配源数据和目标数据,在遮挡区域尽可能保持原始形状。实验结果表明,我们的方法能够利用先验或后验帧的形状信息准确地填充缺失区域。通过对数据进行顺序处理,我们的方法还能够完成缺失区域的整个序列。
{"title":"Non-rigid registration with reliable distance field for dynamic shape completion","authors":"Kent Fujiwara, Hiroshi Kawasaki, R. Sagawa, K. Ogawara, K. Ikeuchi","doi":"10.1109/3DV.2014.111","DOIUrl":"https://doi.org/10.1109/3DV.2014.111","url":null,"abstract":"We propose a non-rigid registration method for completion of dynamic shapes with occlusion. Our method is based on the idea that an occluded region in a certain frame should be visible in another frame and that local regions should be moving rigidly when the motion is small. We achieve this with a novel reliable distance field (DF) for non-rigid registration with missing regions. We first fit a pseudo-surface onto the input shape using a surface reconstruction method. We then calculate the difference between the DF of the input shape and the pseudo-surface. We define the areas with large difference as unreliable, as these areas indicate that the original shape cannot be found nearby. We then conduct non-rigid registration using local rigid transformations to match the source and target data at visible regions and maintain the original shape as much as possible in occluded regions. The experimental results demonstrate that our method is capable of accurately filling in the missing regions using the shape information from prior or posterior frames. By sequentially processing the data, our method is also capable of completing an entire sequence with missing regions.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"140 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121158327","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improving Sparse 3D Models for Man-Made Environments Using Line-Based 3D Reconstruction 基于线的三维重建改进人工环境稀疏三维模型
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.14
Manuel Hofer, Michael Maurer, H. Bischof
Traditional Structure-from-Motion (SfM) approaches work well for richly textured scenes with a high number of distinctive feature points. Since man-made environments often contain texture less objects, the resulting point cloud suffers from a low density in corresponding scene parts. The missing 3D information heavily affects all kinds of subsequent post-processing tasks (e.g. Meshing), and significantly decreases the visual appearance of the resulting 3D model. We propose a novel 3D reconstruction approach, which uses the output of conventional SfM pipelines to generate additional complementary 3D information, by exploiting line segments. We use appearance-less epipolar guided line matching to create a potentially large set of 3D line hypotheses, which are then verified using a global graph clustering procedure. We show that our proposed method outperforms the current state-of-the-art in terms of runtime and accuracy, as well as visual appearance of the resulting reconstructions.
传统的动态结构(SfM)方法对于具有大量特征点的丰富纹理场景效果良好。由于人造环境通常包含较少纹理的对象,因此生成的点云在相应的场景部分中密度较低。缺失的3D信息严重影响了后续的各种后处理任务(如网格划分),并显著降低了生成的3D模型的视觉外观。我们提出了一种新的三维重建方法,该方法通过利用线段,使用传统SfM管道的输出来生成额外的互补三维信息。我们使用较少外观的极极引导线匹配来创建一个潜在的大型3D线假设集,然后使用全局图聚类过程对其进行验证。我们表明,我们提出的方法在运行时间和准确性以及结果重建的视觉外观方面优于当前最先进的技术。
{"title":"Improving Sparse 3D Models for Man-Made Environments Using Line-Based 3D Reconstruction","authors":"Manuel Hofer, Michael Maurer, H. Bischof","doi":"10.1109/3DV.2014.14","DOIUrl":"https://doi.org/10.1109/3DV.2014.14","url":null,"abstract":"Traditional Structure-from-Motion (SfM) approaches work well for richly textured scenes with a high number of distinctive feature points. Since man-made environments often contain texture less objects, the resulting point cloud suffers from a low density in corresponding scene parts. The missing 3D information heavily affects all kinds of subsequent post-processing tasks (e.g. Meshing), and significantly decreases the visual appearance of the resulting 3D model. We propose a novel 3D reconstruction approach, which uses the output of conventional SfM pipelines to generate additional complementary 3D information, by exploiting line segments. We use appearance-less epipolar guided line matching to create a potentially large set of 3D line hypotheses, which are then verified using a global graph clustering procedure. We show that our proposed method outperforms the current state-of-the-art in terms of runtime and accuracy, as well as visual appearance of the resulting reconstructions.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"51 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114692449","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 41
LETHA: Learning from High Quality Inputs for 3D Pose Estimation in Low Quality Images LETHA:从低质量图像的3D姿态估计的高质量输入中学习
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.18
Adrián Peñate Sánchez, F. Moreno-Noguer, J. Andrade-Cetto, F. Fleuret
We introduce LETHA (Learning on Easy data, Test on Hard), a new learning paradigm consisting of building strong priors from high quality training data, and combining them with discriminative machine learning to deal with low-quality test data. Our main contribution is an implementation of that concept for pose estimation. We first automatically build a 3D model of the object of interest from high-definition images, and devise from it a pose-indexed feature extraction scheme. We then train a single classifier to process these feature vectors. Given a low quality test image, we visit many hypothetical poses, extract features consistently and evaluate the response of the classifier. Since this process uses locations recorded during learning, it does not require matching points anymore. We use a boosting procedure to train this classifier common to all poses, which is able to deal with missing features, due in this context to self-occlusion. Our results demonstrate that the method combines the strengths of global image representations, discriminative even for very tiny images, and the robustness to occlusions of approaches based on local feature point descriptors.
我们介绍了LETHA (Learning on Easy data, Test on Hard),这是一种新的学习范式,它包括从高质量的训练数据中构建强先验,并将它们与判别机器学习相结合,以处理低质量的测试数据。我们的主要贡献是实现了姿态估计的概念。我们首先从高清图像中自动建立感兴趣对象的三维模型,并设计了一个基于姿态索引的特征提取方案。然后我们训练一个分类器来处理这些特征向量。给定低质量的测试图像,我们访问许多假设的姿势,一致地提取特征并评估分类器的响应。由于这个过程使用的是在学习过程中记录的位置,因此不再需要匹配点。我们使用一个增强过程来训练这个对所有姿势都通用的分类器,它能够处理由于这种情况下的自遮挡而缺失的特征。我们的研究结果表明,该方法结合了全局图像表示的优势,即使对非常微小的图像也具有判别性,以及基于局部特征点描述符的方法对遮挡的鲁棒性。
{"title":"LETHA: Learning from High Quality Inputs for 3D Pose Estimation in Low Quality Images","authors":"Adrián Peñate Sánchez, F. Moreno-Noguer, J. Andrade-Cetto, F. Fleuret","doi":"10.1109/3DV.2014.18","DOIUrl":"https://doi.org/10.1109/3DV.2014.18","url":null,"abstract":"We introduce LETHA (Learning on Easy data, Test on Hard), a new learning paradigm consisting of building strong priors from high quality training data, and combining them with discriminative machine learning to deal with low-quality test data. Our main contribution is an implementation of that concept for pose estimation. We first automatically build a 3D model of the object of interest from high-definition images, and devise from it a pose-indexed feature extraction scheme. We then train a single classifier to process these feature vectors. Given a low quality test image, we visit many hypothetical poses, extract features consistently and evaluate the response of the classifier. Since this process uses locations recorded during learning, it does not require matching points anymore. We use a boosting procedure to train this classifier common to all poses, which is able to deal with missing features, due in this context to self-occlusion. Our results demonstrate that the method combines the strengths of global image representations, discriminative even for very tiny images, and the robustness to occlusions of approaches based on local feature point descriptors.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114833464","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Match Box: Indoor Image Matching via Box-Like Scene Estimation 火柴盒:室内图像匹配通过盒样场景估计
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.56
F. Srajer, A. Schwing, M. Pollefeys, T. Pajdla
Key point matching in images of indoor scenes traditionally employs features like SIFT, GIST and HOG. While those features work very well for two images related to each other by small camera transformations, we commonly observe a drop in performance for patches representing scene elements visualized from a very different perspective. Since increasing the space of considered local transformations for feature matching decreases their discriminative abilities, we propose a more global approach inspired by the recent success of monocular scene understanding. In particular we propose to reconstruct a box-like model of the scene from every single image and use it to rectify images before matching. We show that a monocular scene model reconstruction and rectification preceding standard feature matching significantly improves key point matching and dramatically improves reconstruction of difficult indoor scenes.
传统的室内场景图像关键点匹配采用SIFT、GIST、HOG等特征。虽然这些特征对于通过小相机变换相互关联的两幅图像非常有效,但我们通常会观察到从非常不同的角度呈现场景元素的补丁的性能下降。由于增加用于特征匹配的考虑局部变换的空间会降低它们的判别能力,我们提出了一种更全局的方法,灵感来自最近单目场景理解的成功。特别地,我们建议从每个单独的图像中重建一个场景的盒状模型,并用它在匹配之前对图像进行校正。我们发现,在标准特征匹配之前进行单目场景模型重建和校正,可以显著改善关键点匹配,并显著改善困难室内场景的重建。
{"title":"Match Box: Indoor Image Matching via Box-Like Scene Estimation","authors":"F. Srajer, A. Schwing, M. Pollefeys, T. Pajdla","doi":"10.1109/3DV.2014.56","DOIUrl":"https://doi.org/10.1109/3DV.2014.56","url":null,"abstract":"Key point matching in images of indoor scenes traditionally employs features like SIFT, GIST and HOG. While those features work very well for two images related to each other by small camera transformations, we commonly observe a drop in performance for patches representing scene elements visualized from a very different perspective. Since increasing the space of considered local transformations for feature matching decreases their discriminative abilities, we propose a more global approach inspired by the recent success of monocular scene understanding. In particular we propose to reconstruct a box-like model of the scene from every single image and use it to rectify images before matching. We show that a monocular scene model reconstruction and rectification preceding standard feature matching significantly improves key point matching and dramatically improves reconstruction of difficult indoor scenes.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115869832","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
期刊
2014 2nd International Conference on 3D Vision
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1