首页 > 最新文献

2014 2nd International Conference on 3D Vision最新文献

英文 中文
Hierarchical Co-Segmentation of Building Facades 建筑立面的分层共分割
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.26
Andelo Martinovic, L. Gool
We introduce a new system for automatic discovery of high-level structural representations of building facades. Under the assumption that each facade can be represented as a hierarchy of rectilinear subdivisions, our goal is to find the optimal direction of splitting, along with the number and positions of the split lines at each level of the tree. Unlike previous approaches, where each facade is analysed in isolation, we propose a joint analysis of a set of facade images. Initially, a co-segmentation approach is used to produce consistent decompositions across all facade images. Afterwards, a clustering step identifies semantically similar segments. Each cluster of similar segments is then used as the input for the joint segmentation in the next level of the hierarchy. We show that our approach produces consistent hierarchical segmentations on two different facade datasets. Furthermore, we argue that the discovered hierarchies capture essential structural information, which is demonstrated on the tasks of facade retrieval and virtual facade synthesis.
我们引入了一个新的系统,用于自动发现建筑立面的高层结构表示。假设每个立面都可以表示为直线细分的层次结构,我们的目标是找到最佳的分割方向,以及树的每一层的分割线的数量和位置。与以前的方法不同,每个立面都是单独分析的,我们提出了一组立面图像的联合分析。最初,使用一种共分割方法在所有立面图像中产生一致的分解。然后,聚类步骤识别语义上相似的片段。然后,每个相似片段的簇被用作层次结构下一层的联合分割的输入。我们展示了我们的方法在两个不同的外观数据集上产生一致的分层分割。此外,我们认为发现的层次结构捕获了基本的结构信息,这在立面检索和虚拟立面合成任务中得到了证明。
{"title":"Hierarchical Co-Segmentation of Building Facades","authors":"Andelo Martinovic, L. Gool","doi":"10.1109/3DV.2014.26","DOIUrl":"https://doi.org/10.1109/3DV.2014.26","url":null,"abstract":"We introduce a new system for automatic discovery of high-level structural representations of building facades. Under the assumption that each facade can be represented as a hierarchy of rectilinear subdivisions, our goal is to find the optimal direction of splitting, along with the number and positions of the split lines at each level of the tree. Unlike previous approaches, where each facade is analysed in isolation, we propose a joint analysis of a set of facade images. Initially, a co-segmentation approach is used to produce consistent decompositions across all facade images. Afterwards, a clustering step identifies semantically similar segments. Each cluster of similar segments is then used as the input for the joint segmentation in the next level of the hierarchy. We show that our approach produces consistent hierarchical segmentations on two different facade datasets. Furthermore, we argue that the discovered hierarchies capture essential structural information, which is demonstrated on the tasks of facade retrieval and virtual facade synthesis.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115498347","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Proceduralization of Buildings at City Scale 城市尺度建筑的程序化
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.31
Ilke Demir, Daniel G. Aliaga, Bedrich Benes
We present a framework for the conversion of existing 3D unstructured urban models into a compact procedural representation that enables model synthesis, querying, and simplification of large urban areas. During the de-instancing phase, a dissimilarity-based clustering is performed to obtain a set of building components and component types. During the proceduralization phase, the components are arranged into a context-free grammar, which can be directly edited or interactively manipulated. We applied our approach to convert several large city models, with up to 19,000 building components spanning over 180 km squares, into procedural models of a few thousand terminals, non-terminals, and 50-100 rules.
我们提出了一个框架,用于将现有的3D非结构化城市模型转换为紧凑的程序表示,从而实现大型城市地区的模型综合、查询和简化。在去实例化阶段,执行基于差异性的集群,以获得一组构建组件和组件类型。在程序化阶段,组件被安排成与上下文无关的语法,可以直接编辑或交互操作。我们运用我们的方法将几个大型城市模型转换为数千个终端,非终端和50-100个规则的程序模型,这些模型包含超过180平方公里的19,000个建筑组件。
{"title":"Proceduralization of Buildings at City Scale","authors":"Ilke Demir, Daniel G. Aliaga, Bedrich Benes","doi":"10.1109/3DV.2014.31","DOIUrl":"https://doi.org/10.1109/3DV.2014.31","url":null,"abstract":"We present a framework for the conversion of existing 3D unstructured urban models into a compact procedural representation that enables model synthesis, querying, and simplification of large urban areas. During the de-instancing phase, a dissimilarity-based clustering is performed to obtain a set of building components and component types. During the proceduralization phase, the components are arranged into a context-free grammar, which can be directly edited or interactively manipulated. We applied our approach to convert several large city models, with up to 19,000 building components spanning over 180 km squares, into procedural models of a few thousand terminals, non-terminals, and 50-100 rules.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114592889","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Distortion Driven Variational Multi-view Reconstruction 畸变驱动的变分多视图重构
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.99
Patricio A. Galindo, Rhaleb Zayer
This paper revisits variational multi-view stereo and identifies two issues pertaining to matching and view merging: i) regions with low visibility and relatively high depth variation are only resolved by the sole regularizer contribution. This often induces wrong matches which tend to bleed into neigh boring regions, and more importantly distort nearby features. ii) small matching errors can lead to overlapping surface layers which cannot be easily addressed by standard outlier removal techniques. In both scenarios, we rely on the analysis of the distortion of spatial and planar maps in order to improve the quality of the reconstruction. At the matching level, an anisotropic diffusion driven by spatial grid distortion is proposed to steer grid lines away from those problematic regions. At the merging level, advantage is taken of Lambert's cosine law to favor contributions from image areas where the cosine angle between the surface normal and the line of sight is maximal. Tests on standard benchmarks suggest a good blend between computational efficiency, ease of implementation, and reconstruction quality.
本文重新研究了变分多视图立体,并指出了与匹配和视图合并有关的两个问题:1)低能见度和相对高深度变化的区域只能由唯一的正则化器贡献来解决。这通常会导致错误的匹配,从而导致相邻的无聊区域,更重要的是会扭曲附近的特征。Ii)小的匹配误差会导致表面层重叠,这是标准的异常值去除技术无法轻易解决的。在这两种情况下,我们都依赖于空间和平面地图的变形分析来提高重建的质量。在匹配层面,提出了由空间网格畸变驱动的各向异性扩散,使网格线远离这些问题区域。在合并层面,利用朗伯特余弦定律的优势,有利于来自表面法线与视线之间的余弦角最大的图像区域的贡献。在标准基准测试上的测试表明,计算效率、易于实现和重构质量之间有很好的结合。
{"title":"Distortion Driven Variational Multi-view Reconstruction","authors":"Patricio A. Galindo, Rhaleb Zayer","doi":"10.1109/3DV.2014.99","DOIUrl":"https://doi.org/10.1109/3DV.2014.99","url":null,"abstract":"This paper revisits variational multi-view stereo and identifies two issues pertaining to matching and view merging: i) regions with low visibility and relatively high depth variation are only resolved by the sole regularizer contribution. This often induces wrong matches which tend to bleed into neigh boring regions, and more importantly distort nearby features. ii) small matching errors can lead to overlapping surface layers which cannot be easily addressed by standard outlier removal techniques. In both scenarios, we rely on the analysis of the distortion of spatial and planar maps in order to improve the quality of the reconstruction. At the matching level, an anisotropic diffusion driven by spatial grid distortion is proposed to steer grid lines away from those problematic regions. At the merging level, advantage is taken of Lambert's cosine law to favor contributions from image areas where the cosine angle between the surface normal and the line of sight is maximal. Tests on standard benchmarks suggest a good blend between computational efficiency, ease of implementation, and reconstruction quality.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114885141","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A 3D Shape Descriptor for Segmentation of Unstructured Meshes into Segment-Wise Coherent Mesh Series 一种非结构化网格分割成分段相干网格序列的三维形状描述符
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.20
T. Mukasa, S. Nobuhara, Tony Tung, T. Matsuyama
This paper presents a novel shape descriptor for topology-based segmentation of 3D video sequence. 3D video is a series of 3D meshes without temporal correspondences which benefit for applications including compression, motion analysis, and kinematic editing. In 3D video, both 3D mesh connectivities and the global surface topology can change frame by frame. This characteristic prevents from making accurate temporal correspondences through the entire 3D mesh series. To overcome this difficulty, we propose a two-step strategy which decomposes the entire sequence into a series of topologically coherent segments using our new shape descriptor, and then estimates temporal correspondences on a per-segment basis. We demonstrate the robustness and accuracy of the shape descriptor on real data which consist of large non-rigid motion and reconstruction errors.
提出了一种新的基于拓扑分割的三维视频序列形状描述符。3D视频是一系列没有时间对应的3D网格,有利于压缩、运动分析和运动编辑等应用。在3D视频中,3D网格连通性和全局表面拓扑结构都可以逐帧改变。这个特性阻止了通过整个3D网格系列进行精确的时间对应。为了克服这一困难,我们提出了一种两步策略,即使用我们的新形状描述符将整个序列分解为一系列拓扑一致的片段,然后在每个片段的基础上估计时间对应。在具有较大非刚体运动和重构误差的实际数据上证明了该形状描述符的鲁棒性和准确性。
{"title":"A 3D Shape Descriptor for Segmentation of Unstructured Meshes into Segment-Wise Coherent Mesh Series","authors":"T. Mukasa, S. Nobuhara, Tony Tung, T. Matsuyama","doi":"10.1109/3DV.2014.20","DOIUrl":"https://doi.org/10.1109/3DV.2014.20","url":null,"abstract":"This paper presents a novel shape descriptor for topology-based segmentation of 3D video sequence. 3D video is a series of 3D meshes without temporal correspondences which benefit for applications including compression, motion analysis, and kinematic editing. In 3D video, both 3D mesh connectivities and the global surface topology can change frame by frame. This characteristic prevents from making accurate temporal correspondences through the entire 3D mesh series. To overcome this difficulty, we propose a two-step strategy which decomposes the entire sequence into a series of topologically coherent segments using our new shape descriptor, and then estimates temporal correspondences on a per-segment basis. We demonstrate the robustness and accuracy of the shape descriptor on real data which consist of large non-rigid motion and reconstruction errors.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123665835","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Improved Techniques for Multi-view Registration with Motion Averaging 基于运动平均的多视图配准改进技术
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.23
Zhongyu Li, Jihua Zhu, Ke Lan, Chen Li, Chaowei Fang
Recently, motion averaging has been introduced as an effective means to solve multi-view registration problem. This approach utilizes the Lie-algebras to implement the averaging of many relative motions, each of which corresponds to the registration result of the scan pair involved in multi-view registration. Accordingly, a key question is how to obtain accurate registration between two partially overlapping scans. This paper presents a method to estimate the overlapping percentage between each scan pair involved in multi-view registration. What's more, it applies the trimmed iterative closest point (TrICP) algorithm to obtain accurate relative motions for the scan pairs including high overlapping percentage. Besides, it introduces the parallel computation to increase the efficiency of multi-view registration. Experimental results carried out with public data sets illustrate its superiority over previous approaches.
近年来,运动平均技术作为一种解决多视点配准问题的有效手段被引入。该方法利用lie代数实现对多个相对运动的平均,每个相对运动对应于多视图配准中涉及的扫描对的配准结果。因此,一个关键问题是如何在两个部分重叠的扫描之间获得准确的配准。提出了一种估计多视图配准中各扫描对重叠百分比的方法。针对重叠率较高的扫描对,采用裁剪迭代最近点(TrICP)算法获得精确的相对运动。此外,还引入了并行计算,提高了多视图配准的效率。在公共数据集上进行的实验结果表明,该方法优于以往的方法。
{"title":"Improved Techniques for Multi-view Registration with Motion Averaging","authors":"Zhongyu Li, Jihua Zhu, Ke Lan, Chen Li, Chaowei Fang","doi":"10.1109/3DV.2014.23","DOIUrl":"https://doi.org/10.1109/3DV.2014.23","url":null,"abstract":"Recently, motion averaging has been introduced as an effective means to solve multi-view registration problem. This approach utilizes the Lie-algebras to implement the averaging of many relative motions, each of which corresponds to the registration result of the scan pair involved in multi-view registration. Accordingly, a key question is how to obtain accurate registration between two partially overlapping scans. This paper presents a method to estimate the overlapping percentage between each scan pair involved in multi-view registration. What's more, it applies the trimmed iterative closest point (TrICP) algorithm to obtain accurate relative motions for the scan pairs including high overlapping percentage. Besides, it introduces the parallel computation to increase the efficiency of multi-view registration. Experimental results carried out with public data sets illustrate its superiority over previous approaches.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124618611","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
Iterative Closest Spectral Kernel Maps 迭代最接近谱核映射
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.24
A. Shtern, R. Kimmel
An important operation in geometry processing is finding the correspondences between pairs of shapes. Measures of dissimilarity between surfaces, has been found to be highly useful for nonrigid shape comparison. Here, we analyze the applicability of the spectral kernel distance, for solving the shape matching problem. To align the spectral kernels, we introduce the iterative closest spectral kernel maps (ICSKM) algorithm. The ICSKM algorithm farther extends the iterative closest point algorithm to the class of deformable shapes. The proposed method achieves state-of-the-art results on the Princeton isometric shape matching protocol applied, as usual, to the TOSCA and SCAPE benchmarks.
几何处理中的一项重要操作是找出图形对之间的对应关系。表面之间的不相似性的措施,已被发现是非刚性形状比较非常有用。在此,我们分析了谱核距离在解决形状匹配问题中的适用性。为了使谱核对齐,我们引入了迭代最接近谱核映射(ICSKM)算法。ICSKM算法将迭代最近点算法进一步扩展到可变形形状类。所提出的方法在应用于TOSCA和SCAPE基准的普林斯顿等距形状匹配协议上获得了最先进的结果。
{"title":"Iterative Closest Spectral Kernel Maps","authors":"A. Shtern, R. Kimmel","doi":"10.1109/3DV.2014.24","DOIUrl":"https://doi.org/10.1109/3DV.2014.24","url":null,"abstract":"An important operation in geometry processing is finding the correspondences between pairs of shapes. Measures of dissimilarity between surfaces, has been found to be highly useful for nonrigid shape comparison. Here, we analyze the applicability of the spectral kernel distance, for solving the shape matching problem. To align the spectral kernels, we introduce the iterative closest spectral kernel maps (ICSKM) algorithm. The ICSKM algorithm farther extends the iterative closest point algorithm to the class of deformable shapes. The proposed method achieves state-of-the-art results on the Princeton isometric shape matching protocol applied, as usual, to the TOSCA and SCAPE benchmarks.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114395015","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
A TV Prior for High-Quality Local Multi-view Stereo Reconstruction 高质量局部多视点立体重建的电视先验
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.76
Andreas Kuhn, H. Mayer, H. Hirschmüller, D. Scharstein
Local fusion of disparity maps allows fast parallel 3D modeling of large scenes that do not fit into main memory. While existing methods assume a constant disparity uncertainty, disparity errors typically vary spatially from tenths of pixels to several pixels. In this paper we propose a method that employs a set of Gaussians for different disparity classes, instead of a single error model with only one variance. The set of Gaussians is learned from the difference between generated disparity maps and ground-truth disparities. Pixels are assigned particular disparity classes based on a Total Variation (TV) feature measuring the local oscillation behavior of the 2D disparity map. This feature captures uncertainty caused for instance by lack of texture or fronto-parallel bias of the stereo method. Experimental results on several datasets in varying configurations demonstrate that our method yields improved performance both qualitatively and quantitatively.
视差图的局部融合允许对不适合主存储器的大型场景进行快速并行3D建模。虽然现有方法假设视差不确定性恒定,但视差误差通常在空间上从十分之一像素到几个像素不等。在本文中,我们提出了一种方法,对不同的视差类采用一组高斯分布,而不是只有一个方差的单一误差模型。高斯分布的集合是从生成的视差图和真差之间的差异中学习到的。基于测量二维视差图的局部振荡行为的总变化(TV)特征,像素被分配特定的视差类。这一特性捕捉了由于缺乏纹理或立体方法的正面平行偏差而引起的不确定性。在不同配置的数据集上的实验结果表明,我们的方法在定性和定量上都取得了改进的性能。
{"title":"A TV Prior for High-Quality Local Multi-view Stereo Reconstruction","authors":"Andreas Kuhn, H. Mayer, H. Hirschmüller, D. Scharstein","doi":"10.1109/3DV.2014.76","DOIUrl":"https://doi.org/10.1109/3DV.2014.76","url":null,"abstract":"Local fusion of disparity maps allows fast parallel 3D modeling of large scenes that do not fit into main memory. While existing methods assume a constant disparity uncertainty, disparity errors typically vary spatially from tenths of pixels to several pixels. In this paper we propose a method that employs a set of Gaussians for different disparity classes, instead of a single error model with only one variance. The set of Gaussians is learned from the difference between generated disparity maps and ground-truth disparities. Pixels are assigned particular disparity classes based on a Total Variation (TV) feature measuring the local oscillation behavior of the 2D disparity map. This feature captures uncertainty caused for instance by lack of texture or fronto-parallel bias of the stereo method. Experimental results on several datasets in varying configurations demonstrate that our method yields improved performance both qualitatively and quantitatively.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130055076","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 24
Least MSE Regression for View Synthesis 最小MSE回归的视图合成
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.29
Keita Takahashi, T. Fujii
View synthesis is the process of combining given multi-view images to generate an image from a new viewpoint. Assuming that each pixel of the new view is obtained as the weighted sum of the corresponding pixels from the input views, we focus on the problem of how to optimize the weight for each of the input views. Our weighting method is called least mean squared error (MSE) regression because it is formulated as a regression problem in which second order statistics among the viewpoints are exploited to minimize the MSE of the resulting image. More specifically, the affinity across the viewpoints is represented as a covariance and approximated using a linear model whose parameters are adapted for each dataset. By using the approximated covariance, the optimal weights can be successfully estimated. As a result, the weights derived using our method are data dependent and significantly differ from those obtained using current empirical methods such as distance penalty. Our method is still effective if the given correspondence is not completely accurate due to noise. We report on experimental results using several multi-view datasets to validate our theory and method.
视图合成是将给定的多视图图像组合在一起,从新的视点生成图像的过程。假设新视图的每个像素都是输入视图中相应像素的加权和,我们关注的问题是如何优化每个输入视图的权重。我们的加权方法被称为最小均方误差(MSE)回归,因为它被表述为一个回归问题,其中利用视点之间的二阶统计量来最小化结果图像的MSE。更具体地说,视点之间的亲和关系表示为协方差,并使用线性模型进行近似,该模型的参数适用于每个数据集。利用近似的协方差,可以成功地估计出最优权重。因此,使用我们的方法获得的权重依赖于数据,并且与使用当前经验方法(如距离惩罚)获得的权重有很大不同。如果给定的通信由于噪声而不完全准确,我们的方法仍然有效。我们报告了使用几个多视图数据集的实验结果来验证我们的理论和方法。
{"title":"Least MSE Regression for View Synthesis","authors":"Keita Takahashi, T. Fujii","doi":"10.1109/3DV.2014.29","DOIUrl":"https://doi.org/10.1109/3DV.2014.29","url":null,"abstract":"View synthesis is the process of combining given multi-view images to generate an image from a new viewpoint. Assuming that each pixel of the new view is obtained as the weighted sum of the corresponding pixels from the input views, we focus on the problem of how to optimize the weight for each of the input views. Our weighting method is called least mean squared error (MSE) regression because it is formulated as a regression problem in which second order statistics among the viewpoints are exploited to minimize the MSE of the resulting image. More specifically, the affinity across the viewpoints is represented as a covariance and approximated using a linear model whose parameters are adapted for each dataset. By using the approximated covariance, the optimal weights can be successfully estimated. As a result, the weights derived using our method are data dependent and significantly differ from those obtained using current empirical methods such as distance penalty. Our method is still effective if the given correspondence is not completely accurate due to noise. We report on experimental results using several multi-view datasets to validate our theory and method.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"62 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121113194","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Efficient Multiview Stereo by Random-Search and Propagation 基于随机搜索和传播的高效多视点立体视觉
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.35
Youngjung Uh, Y. Matsushita, H. Byun
We present an efficient multi-view 3D reconstruction method based on randomization and propagation scheme. Our method progressively refines 3D point estimates by randomly perturbing the initial guess of 3D points and propagates photo-consistent ones to their neighbors. In contrast to previous refinement methods that perform local optimization for a better photo-consistency, our randomization approach takes lucky matchings for reducing the computational complexity. Experiments show favorable efficiency of the proposed method with the accuracy that is close to the state-of-the-art methods.
提出了一种基于随机化和传播的高效多视图三维重建方法。我们的方法通过随机干扰3D点的初始猜测来逐步改进3D点估计,并将照片一致的点传播给它们的邻居。与之前的优化方法进行局部优化以获得更好的照片一致性相比,我们的随机化方法采用幸运匹配来降低计算复杂度。实验表明,该方法具有良好的效率,精度接近现有方法。
{"title":"Efficient Multiview Stereo by Random-Search and Propagation","authors":"Youngjung Uh, Y. Matsushita, H. Byun","doi":"10.1109/3DV.2014.35","DOIUrl":"https://doi.org/10.1109/3DV.2014.35","url":null,"abstract":"We present an efficient multi-view 3D reconstruction method based on randomization and propagation scheme. Our method progressively refines 3D point estimates by randomly perturbing the initial guess of 3D points and propagates photo-consistent ones to their neighbors. In contrast to previous refinement methods that perform local optimization for a better photo-consistency, our randomization approach takes lucky matchings for reducing the computational complexity. Experiments show favorable efficiency of the proposed method with the accuracy that is close to the state-of-the-art methods.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"39 2","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120883729","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Recovering Correct Reconstructions from Indistinguishable Geometry 从难以区分的几何中恢复正确的重建
Pub Date : 2014-12-08 DOI: 10.1109/3DV.2014.84
Jared Heinly, Enrique Dunn, Jan-Michael Frahm
Structure-from-motion (SFM) is widely utilized to generate 3D reconstructions from unordered photo-collections. However, in the presence of non unique, symmetric, or otherwise indistinguishable structure, SFM techniques often incorrectly reconstruct the final model. We propose a method that not only determines if an error is present, but automatically corrects the error in order to produce a correct representation of the scene. We find that by exploiting the co-occurrence information present in the scene's geometry, we can successfully isolate the 3D points causing the incorrect result. This allows us to split an incorrect reconstruction into error-free sub-models that we then correctly merge back together. Our experimental results show that our technique is efficient, robust to a variety of scenes, and outperforms existing methods.
运动结构(SFM)被广泛用于从无序的照片集合中生成三维重建。然而,在存在非唯一的、对称的或其他不可区分的结构时,SFM技术通常会错误地重建最终模型。我们提出了一种方法,不仅可以确定是否存在错误,还可以自动纠正错误,以便生成正确的场景表示。我们发现,通过利用场景几何中存在的共现信息,我们可以成功地隔离导致错误结果的3D点。这允许我们将不正确的重建拆分为没有错误的子模型,然后我们将它们正确地合并回一起。实验结果表明,该方法对各种场景具有较强的鲁棒性,并且优于现有方法。
{"title":"Recovering Correct Reconstructions from Indistinguishable Geometry","authors":"Jared Heinly, Enrique Dunn, Jan-Michael Frahm","doi":"10.1109/3DV.2014.84","DOIUrl":"https://doi.org/10.1109/3DV.2014.84","url":null,"abstract":"Structure-from-motion (SFM) is widely utilized to generate 3D reconstructions from unordered photo-collections. However, in the presence of non unique, symmetric, or otherwise indistinguishable structure, SFM techniques often incorrectly reconstruct the final model. We propose a method that not only determines if an error is present, but automatically corrects the error in order to produce a correct representation of the scene. We find that by exploiting the co-occurrence information present in the scene's geometry, we can successfully isolate the 3D points causing the incorrect result. This allows us to split an incorrect reconstruction into error-free sub-models that we then correctly merge back together. Our experimental results show that our technique is efficient, robust to a variety of scenes, and outperforms existing methods.","PeriodicalId":275516,"journal":{"name":"2014 2nd International Conference on 3D Vision","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115279514","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
期刊
2014 2nd International Conference on 3D Vision
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1