首页 > 最新文献

2014 2nd International Conference on 3D Vision最新文献

英文 中文
Multiple Hypothesis for Object Class Disambiguation from Multiple Observations 基于多重观测的对象类消歧的多重假设
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.101
Susana Brandão, M. Veloso, J. Costeira
The current paper addresses the problem of object identification from multiple3D partial views, collected from different view angles with the objective of disambiguating between similar objects. We assume a mobile robot equipped with a depth sensor that autonomously collects observations from an object from different positions, with no previous known pattern. The challenge is to efficiently combine the set of observations into a single classification. We approach the problem with a multiple hypothesis filter that allows to combine information from a sequence of observations given the robot movement. We further innovate by off-line learning neighborhoods between possible hypothesis based on the similarity of observations. Such neighborhoods translate directly the ambiguity between objects, and allow to transfer the knowledge of one object to the other. In this paper we introduce our algorithm, Multiple Hypothesis for Object Class Disambiguation from Multiple Observations, and evaluate its accuracy and efficiency.
本文解决了从不同视角收集的多三维局部视图中识别物体的问题,目的是消除相似物体之间的歧义。我们假设一个移动机器人配备了深度传感器,可以自主地从不同位置收集物体的观测数据,没有先前已知的模式。挑战在于如何有效地将一组观察结果合并到一个分类中。我们用一个多假设过滤器来解决这个问题,该过滤器允许从给定机器人运动的一系列观察中组合信息。我们通过离线学习基于观察相似性的可能假设之间的邻域来进一步创新。这样的邻域直接翻译了物体之间的模糊性,并允许将一个物体的知识转移到另一个物体。本文介绍了一种基于多重观测的多假设目标类消歧算法,并对其准确性和效率进行了评价。
{"title":"Multiple Hypothesis for Object Class Disambiguation from Multiple Observations","authors":"Susana Brandão, M. Veloso, J. Costeira","doi":"10.1109/3DV.2014.101","DOIUrl":"https://doi.org/10.1109/3DV.2014.101","url":null,"abstract":"The current paper addresses the problem of object identification from multiple3D partial views, collected from different view angles with the objective of disambiguating between similar objects. We assume a mobile robot equipped with a depth sensor that autonomously collects observations from an object from different positions, with no previous known pattern. The challenge is to efficiently combine the set of observations into a single classification. We approach the problem with a multiple hypothesis filter that allows to combine information from a sequence of observations given the robot movement. We further innovate by off-line learning neighborhoods between possible hypothesis based on the similarity of observations. Such neighborhoods translate directly the ambiguity between objects, and allow to transfer the knowledge of one object to the other. In this paper we introduce our algorithm, Multiple Hypothesis for Object Class Disambiguation from Multiple Observations, and evaluate its accuracy and efficiency.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126157719","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
A Structured Light Scanner for Hyper Flexible Industrial Automation 用于超柔性工业自动化的结构光扫描仪
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.53
K. Hansen, Jeppe Pedersen, Thomas Sølund, H. Aanæs, D. Kraft
A current trend in industrial automation implies a need for doing automatic scene understanding, from optical 3D sensors, which in turn imposes a need for a lightweight and reliable 3D optical sensor to be mounted on a collaborative robot e.g., Universal Robot UR5 or Kuka LWR. Here, we empirically evaluate the feasibility of structured light scanners for this purpose, by presenting a system optimized for this task. The system incorporates several recent advances in structured light scanning, such as Large-Gap Gray encoding for dealing with defocusing, automatic creation of illumination masks for noise removal, as well as employing a multi exposure approach dealing with different surface reflectance properties. In addition to this, we investigate expanding the traditional structured light setup to using three cameras, instead of one or two. Also, a novel method for fusing multiple exposures and camera pairs is given. We present an in-depth evaluation, that lead us to conclude, that this setup performs well on tasks relevant for an industrial environment, where many metallic and other surfaces with difficult reflectance properties are in abundance. We demonstrate, that the added components contribute to the robustness of the system. Hereby, we demonstrate that structured light scanning is a technology well suited for hyper flexible industrial automation, by proposing an appropriate system.
工业自动化的当前趋势意味着需要从光学3D传感器进行自动场景理解,这反过来又要求需要安装在协作机器人(例如Universal robot UR5或Kuka LWR)上的轻质可靠的3D光学传感器。在这里,我们通过提出一个针对该任务优化的系统,实证地评估了用于此目的的结构光扫描仪的可行性。该系统结合了结构光扫描的几个最新进展,例如用于处理散焦的大间隙灰度编码,用于去除噪声的照明掩模的自动创建,以及采用处理不同表面反射特性的多重曝光方法。除此之外,我们还研究了将传统的结构光设置扩展到使用三个相机,而不是一个或两个。同时,提出了一种多曝光和相机对融合的新方法。我们进行了深入的评估,得出结论,这种设置在工业环境中表现良好,工业环境中有许多具有困难反射特性的金属和其他表面。我们证明,添加的组件有助于系统的鲁棒性。因此,我们通过提出一个适当的系统来证明结构光扫描是一种非常适合超柔性工业自动化的技术。
{"title":"A Structured Light Scanner for Hyper Flexible Industrial Automation","authors":"K. Hansen, Jeppe Pedersen, Thomas Sølund, H. Aanæs, D. Kraft","doi":"10.1109/3DV.2014.53","DOIUrl":"https://doi.org/10.1109/3DV.2014.53","url":null,"abstract":"A current trend in industrial automation implies a need for doing automatic scene understanding, from optical 3D sensors, which in turn imposes a need for a lightweight and reliable 3D optical sensor to be mounted on a collaborative robot e.g., Universal Robot UR5 or Kuka LWR. Here, we empirically evaluate the feasibility of structured light scanners for this purpose, by presenting a system optimized for this task. The system incorporates several recent advances in structured light scanning, such as Large-Gap Gray encoding for dealing with defocusing, automatic creation of illumination masks for noise removal, as well as employing a multi exposure approach dealing with different surface reflectance properties. In addition to this, we investigate expanding the traditional structured light setup to using three cameras, instead of one or two. Also, a novel method for fusing multiple exposures and camera pairs is given. We present an in-depth evaluation, that lead us to conclude, that this setup performs well on tasks relevant for an industrial environment, where many metallic and other surfaces with difficult reflectance properties are in abundance. We demonstrate, that the added components contribute to the robustness of the system. Hereby, we demonstrate that structured light scanning is a technology well suited for hyper flexible industrial automation, by proposing an appropriate system.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128624983","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
High Resolution Projector for 3D Imaging 用于3D成像的高分辨率投影仪
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.86
M. Drouin, F. Blais, G. Godin
This paper proposes a hybrid software-hardware high-resolution projection system for 3D imaging based on fringe projection. The proposed solution combines the advantages of a digital projection with those of an analogue one. It is programmable and allows a high projection rate by avoiding mechanical displacements in the projection system. Moreover, it does not suffer from the limitations of digital systems such as the presence of inter-pixel gaps and limited resolution. The proposed projection system is relatively inexpensive to build since it is composed of a simple arrangement off-the-shelf components. The system is a combination of a low-resolution digital device such as a DMD, LCoS or LCD, some optical components and software to generate the fringe patterns. A prototype of a 3D scanner based on the proposed projection system is used to asses the fitness of the proposed technology.
提出了一种基于条纹投影的三维成像软硬件混合高分辨率投影系统。该方案结合了数字投影和模拟投影的优点。它是可编程的,通过避免投影系统中的机械位移,可以实现高投影率。此外,它不受数字系统的限制,如像素间间隙的存在和有限的分辨率。所提议的投影系统建造起来相对便宜,因为它是由一个简单的现成组件组成的。该系统是一个低分辨率的数字设备,如DMD, lco或LCD,一些光学元件和软件的组合,以产生条纹图案。基于所提出的投影系统的3D扫描仪原型被用来评估所提出的技术的适用性。
{"title":"High Resolution Projector for 3D Imaging","authors":"M. Drouin, F. Blais, G. Godin","doi":"10.1109/3DV.2014.86","DOIUrl":"https://doi.org/10.1109/3DV.2014.86","url":null,"abstract":"This paper proposes a hybrid software-hardware high-resolution projection system for 3D imaging based on fringe projection. The proposed solution combines the advantages of a digital projection with those of an analogue one. It is programmable and allows a high projection rate by avoiding mechanical displacements in the projection system. Moreover, it does not suffer from the limitations of digital systems such as the presence of inter-pixel gaps and limited resolution. The proposed projection system is relatively inexpensive to build since it is composed of a simple arrangement off-the-shelf components. The system is a combination of a low-resolution digital device such as a DMD, LCoS or LCD, some optical components and software to generate the fringe patterns. A prototype of a 3D scanner based on the proposed projection system is used to asses the fitness of the proposed technology.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125650812","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Gradient-Based Differential Approach for 3-D Motion Compensation in Interventional 2-D/3-D Image Fusion 介入二维/三维图像融合中基于梯度的三维运动补偿方法
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.45
Jian Wang, A. Borsdorf, B. Heigl, T. Köhler, J. Hornegger
In interventional radiology, preoperative 3-D volumes can be fused with intra-operative 2-D fluoroscopic images. Since the accuracy is crucial to the clinical usability of image fusion, patient motion resulting in misalignments has to be corrected during the procedure. In this paper, a novel gradient based differential approach is proposed to estimate the 3-D rigid motion from the 2-D tracking of contour points. The mathematical relationship between the 3-D differential motion and the 2-D motion is derived using the 3-D gradient, based on which a tracking-based motion compensation pipeline is introduced. Given the initial registration, the contour points are extracted and tracked along 2-D frames. The 3-D rigid motion is estimated using the iteratively re-weighted least square minimization to enhance the robustness. Our novel approach is evaluated on 10 datasets consisting of 1010 monoplane fluoroscopic images of a thorax phantom with 3-D rigid motion. Over all datasets, the maximum structure shift in the 2-D projection caused by the 3-D motion varies from 17.3 mm to 33.2 mm. Our approach reduces the 2-D structure shift to the range of 1.93 mm to 6.52 mm. For the most challenging longitudinal off-plane rotation, our approach achieves an average coverage of 79.9% regarding to the ground truth.
在介入放射学中,术前三维体积可以与术中二维透视图像融合。由于准确性对图像融合的临床可用性至关重要,因此必须在手术过程中纠正患者运动导致的不对准。本文提出了一种基于梯度的基于轮廓点二维跟踪的三维刚体运动估计方法。利用三维梯度推导了三维微分运动与二维运动之间的数学关系,并在此基础上引入了基于跟踪的运动补偿管道。给定初始配准,提取轮廓点并沿二维帧进行跟踪。采用迭代加权最小二乘最小化方法对三维刚体运动进行估计,增强了鲁棒性。我们的新方法在10个数据集上进行了评估,这些数据集包括1010个具有三维刚性运动的胸腔幻影的单面透视图像。在所有数据集中,三维运动引起的二维投影的最大结构位移从17.3 mm到33.2 mm不等。我们的方法将二维结构位移减小到1.93 mm到6.52 mm的范围内。对于最具挑战性的纵向离面旋转,我们的方法对地面真实度的平均覆盖率达到79.9%。
{"title":"Gradient-Based Differential Approach for 3-D Motion Compensation in Interventional 2-D/3-D Image Fusion","authors":"Jian Wang, A. Borsdorf, B. Heigl, T. Köhler, J. Hornegger","doi":"10.1109/3DV.2014.45","DOIUrl":"https://doi.org/10.1109/3DV.2014.45","url":null,"abstract":"In interventional radiology, preoperative 3-D volumes can be fused with intra-operative 2-D fluoroscopic images. Since the accuracy is crucial to the clinical usability of image fusion, patient motion resulting in misalignments has to be corrected during the procedure. In this paper, a novel gradient based differential approach is proposed to estimate the 3-D rigid motion from the 2-D tracking of contour points. The mathematical relationship between the 3-D differential motion and the 2-D motion is derived using the 3-D gradient, based on which a tracking-based motion compensation pipeline is introduced. Given the initial registration, the contour points are extracted and tracked along 2-D frames. The 3-D rigid motion is estimated using the iteratively re-weighted least square minimization to enhance the robustness. Our novel approach is evaluated on 10 datasets consisting of 1010 monoplane fluoroscopic images of a thorax phantom with 3-D rigid motion. Over all datasets, the maximum structure shift in the 2-D projection caused by the 3-D motion varies from 17.3 mm to 33.2 mm. Our approach reduces the 2-D structure shift to the range of 1.93 mm to 6.52 mm. For the most challenging longitudinal off-plane rotation, our approach achieves an average coverage of 79.9% regarding to the ground truth.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131425203","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Memory Efficient Stereoscopy from Light Fields 基于光场的高效立体记忆
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.12
Changil Kim, Ulrich Muller, H. Zimmer, Y. Pritch, A. Sorkine-Hornung, M. Gross
We address the problem of stereoscopic content generation from light fields using multi-perspective imaging. Our proposed method takes as input a light field and a target disparity map, and synthesizes a stereoscopic image pair by selecting light rays that fulfill the given target disparity constraints. We formulate this as a variational convex optimization problem. Compared to previous work, our method makes use of multi-view input to composite the new view with occlusions and disocclusions properly handled, does not require any correspondence information such as scene depth, is free from undesirable artifacts such as grid bias or image distortion, and is more efficiently solvable. In particular, our method is about ten times more memory efficient than the previous art, and is capable of processing higher resolution input. This is essential to make the proposed method practically applicable to realistic scenarios where HD content is standard. We demonstrate the effectiveness of our method experimentally.
我们解决了使用多视角成像从光场生成立体内容的问题。该方法以光场和目标视差图为输入,通过选择满足给定目标视差约束的光线合成立体像对。我们将其表述为一个变分凸优化问题。与以往的工作相比,我们的方法利用多视图输入来合成新视图,正确处理遮挡和解除遮挡,不需要任何对应信息,如场景深度,不需要网格偏差或图像失真等不良工件,并且更有效地求解。特别是,我们的方法比以前的技术内存效率高10倍,并且能够处理更高分辨率的输入。这对于使所提出的方法实际适用于高清内容标准的现实场景至关重要。我们用实验证明了方法的有效性。
{"title":"Memory Efficient Stereoscopy from Light Fields","authors":"Changil Kim, Ulrich Muller, H. Zimmer, Y. Pritch, A. Sorkine-Hornung, M. Gross","doi":"10.1109/3DV.2014.12","DOIUrl":"https://doi.org/10.1109/3DV.2014.12","url":null,"abstract":"We address the problem of stereoscopic content generation from light fields using multi-perspective imaging. Our proposed method takes as input a light field and a target disparity map, and synthesizes a stereoscopic image pair by selecting light rays that fulfill the given target disparity constraints. We formulate this as a variational convex optimization problem. Compared to previous work, our method makes use of multi-view input to composite the new view with occlusions and disocclusions properly handled, does not require any correspondence information such as scene depth, is free from undesirable artifacts such as grid bias or image distortion, and is more efficiently solvable. In particular, our method is about ten times more memory efficient than the previous art, and is capable of processing higher resolution input. This is essential to make the proposed method practically applicable to realistic scenarios where HD content is standard. We demonstrate the effectiveness of our method experimentally.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"92 4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123167209","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Temporal Octrees for Compressing Dynamic Point Cloud Streams 压缩动态点云流的时间八叉树
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.79
M. Slomp, Hiroshi Kawasaki, Furukawa Ryo, R. Sagawa
Range-based scanners built upon multiple cameras and projectors offer affordable, entire-shape and high-speed setups for 3D scanning. The point cloud streams produced by these devices require large amounts of storage space. Compressing these datasets is challenging since the capturing process may result in noise and surface irregularities, and consecutive frames can differ substantially in the overall point distribution. Exploiting spatial and temporal coherency is difficult on such conditions, but nonetheless crucial for achieving decent compression rates. This paper introduces a novel data structure, the temporal sparse voxel octree, capable of grouping spatio-temporal coherency of multiple point cloud streams into a single voxel hierarchy. In the data structure, a bit mask is attached to each node, existing nodes can then be reused at different frames by manipulating their bit masks, providing substantial memory savings. Although the technique yields some losses, the amount of loss can be controlled.
基于范围的扫描仪建立在多个摄像机和投影仪上,为3D扫描提供经济实惠,完整形状和高速设置。这些设备产生的点云流需要大量的存储空间。压缩这些数据集具有挑战性,因为捕获过程可能导致噪声和表面不规则,并且连续帧在整体点分布上可能存在很大差异。在这种情况下,利用空间和时间的相干性是很困难的,但对于实现体面的压缩率至关重要。本文介绍了一种新的数据结构——时间稀疏体素八叉树,它能够将多个点云流的时空相干性分组到单个体素层次结构中。在数据结构中,每个节点都附加一个位掩码,然后可以通过操作它们的位掩码在不同的帧中重用现有节点,从而节省大量内存。虽然该技术会产生一些损失,但损失的数量是可以控制的。
{"title":"Temporal Octrees for Compressing Dynamic Point Cloud Streams","authors":"M. Slomp, Hiroshi Kawasaki, Furukawa Ryo, R. Sagawa","doi":"10.1109/3DV.2014.79","DOIUrl":"https://doi.org/10.1109/3DV.2014.79","url":null,"abstract":"Range-based scanners built upon multiple cameras and projectors offer affordable, entire-shape and high-speed setups for 3D scanning. The point cloud streams produced by these devices require large amounts of storage space. Compressing these datasets is challenging since the capturing process may result in noise and surface irregularities, and consecutive frames can differ substantially in the overall point distribution. Exploiting spatial and temporal coherency is difficult on such conditions, but nonetheless crucial for achieving decent compression rates. This paper introduces a novel data structure, the temporal sparse voxel octree, capable of grouping spatio-temporal coherency of multiple point cloud streams into a single voxel hierarchy. In the data structure, a bit mask is attached to each node, existing nodes can then be reused at different frames by manipulating their bit masks, providing substantial memory savings. Although the technique yields some losses, the amount of loss can be controlled.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114078591","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Piecewise Planar and Non-Planar Segmentation of Large Complex 3D Urban Models 大型复杂三维城市模型的分段平面与非平面分割
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.88
A. Golbert, David Arnon, A. Sever
Advancements in computing power via Multi Core processors and GPUs have made large scale reconstruction modeling and real-time photorealistic rendering possible. However, in urban areas flat surfaces with little texture still challenge multiview algorithms. We present a method for planar area recognition and model correction while avoiding deformation of non-planar areas such as domes, pillars and plant matter. Our method works in object space, allows a global solution that is not affected by individual range map inaccuracies or poorly matched range maps. We describe a segmentation of the model into bounded planar and non-planar areas driven by a global error function incorporating model shape and original images texture. The error is minimized iteratively using locally restricted graph cuts and the model is corrected accordingly. The algorithm was run on various complex and challenging real-world urban scenes and synthetic photo-realistic images are created from novel viewpoints without noticeable deformities that are common to typical reconstructions.
通过多核处理器和gpu的计算能力的进步使得大规模重建建模和实时逼真渲染成为可能。然而,在城市地区,纹理较少的平坦表面仍然是多视图算法的挑战。提出了一种平面区域识别和模型校正的方法,同时避免了非平面区域(如圆顶、柱子和植物物质)的变形。我们的方法适用于对象空间,允许全局解决方案,不受个别范围图不准确或不匹配范围图的影响。我们描述了一个由模型形状和原始图像纹理结合的全局误差函数驱动的模型分割为有界平面和非平面区域。利用局部限制图切迭代最小化误差,并对模型进行相应的修正。该算法在各种复杂和具有挑战性的现实世界城市场景中运行,并从新颖的视点创建合成的逼真图像,而不会出现典型重建中常见的明显变形。
{"title":"Piecewise Planar and Non-Planar Segmentation of Large Complex 3D Urban Models","authors":"A. Golbert, David Arnon, A. Sever","doi":"10.1109/3DV.2014.88","DOIUrl":"https://doi.org/10.1109/3DV.2014.88","url":null,"abstract":"Advancements in computing power via Multi Core processors and GPUs have made large scale reconstruction modeling and real-time photorealistic rendering possible. However, in urban areas flat surfaces with little texture still challenge multiview algorithms. We present a method for planar area recognition and model correction while avoiding deformation of non-planar areas such as domes, pillars and plant matter. Our method works in object space, allows a global solution that is not affected by individual range map inaccuracies or poorly matched range maps. We describe a segmentation of the model into bounded planar and non-planar areas driven by a global error function incorporating model shape and original images texture. The error is minimized iteratively using locally restricted graph cuts and the model is corrected accordingly. The algorithm was run on various complex and challenging real-world urban scenes and synthetic photo-realistic images are created from novel viewpoints without noticeable deformities that are common to typical reconstructions.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116270360","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Merge2-3D: Combining Multiple Normal Maps with 3D Surfaces Merge2-3D:将多个法线贴图与3D表面结合
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.22
Sema Berkiten, Xinyi Fan, S. Rusinkiewicz
We propose an approach to enhance rough 3D geometry with fine details obtained from multiple normal maps. We begin with unaligned 2D normal maps and rough geometry, and automatically optimize the alignments through 2-step iterative registration algorithm. We then map the normals onto the surface, correcting and seamlessly blending them together. Finally, we optimize the geometry to produce high-quality 3D models that incorporate the high-frequency details from the normal maps. We demonstrate that our algorithm improves upon the results produced by some well-known algorithms: Poisson surface reconstruction [1] and the algorithm proposed by Nehab et al. [2].
我们提出了一种利用从多个法线贴图中获得的精细细节来增强粗糙的3D几何形状的方法。我们从未对齐的二维法线贴图和粗糙几何开始,通过两步迭代配准算法自动优化对齐。然后我们将法线映射到表面上,校正并无缝地将它们混合在一起。最后,我们优化几何结构,生成高质量的3D模型,其中包含来自法线贴图的高频细节。我们证明了我们的算法改进了一些著名算法产生的结果:泊松表面重建[1]和Nehab等人[2]提出的算法。
{"title":"Merge2-3D: Combining Multiple Normal Maps with 3D Surfaces","authors":"Sema Berkiten, Xinyi Fan, S. Rusinkiewicz","doi":"10.1109/3DV.2014.22","DOIUrl":"https://doi.org/10.1109/3DV.2014.22","url":null,"abstract":"We propose an approach to enhance rough 3D geometry with fine details obtained from multiple normal maps. We begin with unaligned 2D normal maps and rough geometry, and automatically optimize the alignments through 2-step iterative registration algorithm. We then map the normals onto the surface, correcting and seamlessly blending them together. Finally, we optimize the geometry to produce high-quality 3D models that incorporate the high-frequency details from the normal maps. We demonstrate that our algorithm improves upon the results produced by some well-known algorithms: Poisson surface reconstruction [1] and the algorithm proposed by Nehab et al. [2].","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121365537","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
An Integer Linear Programming Model for View Selection on Overlapping Camera Clusters 基于整数线性规划的重叠相机簇视图选择模型
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.25
Massimo Mauro, Hayko Riemenschneider, A. Signoroni, R. Leonardi, L. Gool
Multi-View Stereo (MVS) algorithms scale poorly on large image sets, and quickly become unfeasible to run on a single machine with limited memory. Typical solutions to lower the complexity include reducing the redundancy of the image set (view selection), and dividing the image set in groups to be processed independently (view clustering). A novel formulation for view selection is proposed here. We express the problem with an Integer Linear Programming (ILP) model, where cameras are modeled with binary variables, while the linear constraints enforce the completeness of the 3D reconstruction. The solution of the ILP leads to an optimal subset of selected cameras. As a second contribution, we integrate ILP camera selection with a view clustering approach which exploits Leveraged Affinity Propagation (LAP). LAP clustering can efficiently deal with large camera sets. We adapt the original algorithm so that it provides a set of overlapping clusters where the minimum and maximum sizes and the number of overlapping cameras can be specified. Evaluations on four different dataset show our solution provides significant complexity reductions and guarantees near-perfect coverage, making large reconstructions feasible even on a single machine.
多视图立体(MVS)算法在大型图像集上的可扩展性很差,并且很快就无法在内存有限的单个机器上运行。降低复杂性的典型解决方案包括减少图像集的冗余(视图选择)和将图像集分组独立处理(视图聚类)。本文提出了一种新的视图选择公式。我们用整数线性规划(ILP)模型来表达这个问题,其中相机用二进制变量建模,而线性约束强制三维重建的完整性。ILP的解可以得到所选相机的最优子集。作为第二个贡献,我们将ILP相机选择与利用杠杆亲和传播(LAP)的视图聚类方法集成在一起。LAP聚类可以有效地处理大型摄像机集。我们调整了原始算法,使其提供了一组重叠簇,其中可以指定最小和最大尺寸以及重叠相机的数量。对四个不同数据集的评估表明,我们的解决方案显著降低了复杂性,并保证了近乎完美的覆盖,即使在一台机器上也可以进行大型重建。
{"title":"An Integer Linear Programming Model for View Selection on Overlapping Camera Clusters","authors":"Massimo Mauro, Hayko Riemenschneider, A. Signoroni, R. Leonardi, L. Gool","doi":"10.1109/3DV.2014.25","DOIUrl":"https://doi.org/10.1109/3DV.2014.25","url":null,"abstract":"Multi-View Stereo (MVS) algorithms scale poorly on large image sets, and quickly become unfeasible to run on a single machine with limited memory. Typical solutions to lower the complexity include reducing the redundancy of the image set (view selection), and dividing the image set in groups to be processed independently (view clustering). A novel formulation for view selection is proposed here. We express the problem with an Integer Linear Programming (ILP) model, where cameras are modeled with binary variables, while the linear constraints enforce the completeness of the 3D reconstruction. The solution of the ILP leads to an optimal subset of selected cameras. As a second contribution, we integrate ILP camera selection with a view clustering approach which exploits Leveraged Affinity Propagation (LAP). LAP clustering can efficiently deal with large camera sets. We adapt the original algorithm so that it provides a set of overlapping clusters where the minimum and maximum sizes and the number of overlapping cameras can be specified. Evaluations on four different dataset show our solution provides significant complexity reductions and guarantees near-perfect coverage, making large reconstructions feasible even on a single machine.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"106 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122505491","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Close-Range Photometric Stereo with Point Light Sources 与点光源的近距离光度立体
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.68
Aaron Wetzler, R. Kimmel, A. Bruckstein, R. Mecca
Shape recovery based on shading variations of a lighted object was recently revisited with improvements that allow for the photometric stereo approach to serve as a competitive alternative for other shape reconstruction methods. However, most efforts of using photometric stereo tend to ignore some factors that are relevant in practical applications. The approach we consider tackles the photometric stereo reconstruction in the case of near-field imaging which means that both camera and light sources are close to the imaged object. The known challenges that characterize the problem involve perspective viewing geometry, attenuation of light and possibly missing regions. Here, we pay special attention to the question of how to faithfully model these aspects and by the same token design an efficient and robust numerical solver. We present a well-posed mathematical representation that integrates the above assumptions into a single coherent model. The surface reconstruction in our near-field scenario can then be executed efficiently in linear time. The merging strategy of the irradiance equations provided for each light source allows us to consider a characteristic expansion model which enables the direct computation of the surface. We evaluate several types of light attenuation models with nonuniform albedo and noise on synthetic data using four virtual sources. We also demonstrate the proposed method on surface reconstruction of real data using three images, each one taken with a different light source.
基于光照物体的阴影变化的形状恢复最近被重新审视,并进行了改进,使光度立体方法成为其他形状重建方法的竞争替代品。然而,大多数使用光度立体的努力往往忽略了一些在实际应用中相关的因素。我们考虑的方法解决了近场成像情况下的光度立体重建,这意味着相机和光源都接近被成像物体。已知的挑战包括透视几何、光衰减和可能缺失的区域。在这里,我们特别关注如何忠实地对这些方面进行建模,并通过同样的方式设计一个高效和鲁棒的数值求解器。我们提出了一个良好的数学表示,将上述假设集成到一个单一的连贯模型中。在我们的近场场景中,表面重建可以在线性时间内有效地执行。为每个光源提供的辐照度方程的合并策略使我们能够考虑能够直接计算表面的特征展开模型。利用四种虚拟光源,对几种具有非均匀反照率和噪声的光衰减模型进行了评价。我们还用三幅不同光源拍摄的图像对真实数据的表面重建进行了演示。
{"title":"Close-Range Photometric Stereo with Point Light Sources","authors":"Aaron Wetzler, R. Kimmel, A. Bruckstein, R. Mecca","doi":"10.1109/3DV.2014.68","DOIUrl":"https://doi.org/10.1109/3DV.2014.68","url":null,"abstract":"Shape recovery based on shading variations of a lighted object was recently revisited with improvements that allow for the photometric stereo approach to serve as a competitive alternative for other shape reconstruction methods. However, most efforts of using photometric stereo tend to ignore some factors that are relevant in practical applications. The approach we consider tackles the photometric stereo reconstruction in the case of near-field imaging which means that both camera and light sources are close to the imaged object. The known challenges that characterize the problem involve perspective viewing geometry, attenuation of light and possibly missing regions. Here, we pay special attention to the question of how to faithfully model these aspects and by the same token design an efficient and robust numerical solver. We present a well-posed mathematical representation that integrates the above assumptions into a single coherent model. The surface reconstruction in our near-field scenario can then be executed efficiently in linear time. The merging strategy of the irradiance equations provided for each light source allows us to consider a characteristic expansion model which enables the direct computation of the surface. We evaluate several types of light attenuation models with nonuniform albedo and noise on synthetic data using four virtual sources. We also demonstrate the proposed method on surface reconstruction of real data using three images, each one taken with a different light source.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114426859","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
期刊
2014 2nd International Conference on 3D Vision
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1