首页 > 最新文献

2013 IEEE Conference on Computer Vision and Pattern Recognition最新文献

英文 中文
Segment-Tree Based Cost Aggregation for Stereo Matching 基于分割树的成本聚合立体匹配
Pub Date : 2017-02-27 DOI: 10.1109/ICASSP.2017.7952512
Peng Yao, Hua Zhang, Yanbing Xue, Mian Zhou, Guangping Xu, Zan Gao, Shengyong Chen
This paper presents a novel tree-based cost aggregation method for dense stereo matching. Instead of employing the minimum spanning tree (MST) and its variants, a new tree structure, "Segment-Tree", is proposed for non-local matching cost aggregation. Conceptually, the segment-tree is constructed in a three-step process: first, the pixels are grouped into a set of segments with the reference color or intensity image, second, a tree graph is created for each segment, and in the final step, these independent segment graphs are linked to form the segment-tree structure. In practice, this tree can be efficiently built in time nearly linear to the number of the image pixels. Compared to MST where the graph connectivity is determined with local edge weights, our method introduces some 'non-local' decision rules: the pixels in one perceptually consistent segment are more likely to share similar disparities, and therefore their connectivity within the segment should be first enforced in the tree construction process. The matching costs are then aggregated over the tree within two passes. Performance evaluation on 19 Middlebury data sets shows that the proposed method is comparable to previous state-of-the-art aggregation methods in disparity accuracy and processing speed. Furthermore, the tree structure can be refined with the estimated disparities, which leads to consistent scene segmentation and significantly better aggregation results.
提出了一种基于树的密集立体匹配成本聚合方法。针对非局部匹配成本聚合问题,提出了一种新的树结构“分段树”,取代了最小生成树及其变体。从概念上讲,段树的构建分为三个步骤:首先,像素被分成一组具有参考颜色或强度图像的段,其次,为每个段创建一个树状图,在最后一步,这些独立的段图被连接起来形成段树结构。在实践中,该树可以在与图像像素数近似线性的时间内有效地构建。与MST相比,图的连通性是由局部边缘权重决定的,我们的方法引入了一些“非局部”决策规则:在一个感知一致的段中的像素更有可能共享相似的差异,因此它们在段内的连通性应该首先在树构建过程中强制执行。然后,匹配成本在两个通道内对树进行汇总。对19个Middlebury数据集的性能评估表明,所提出的方法在视差精度和处理速度方面与以前最先进的聚合方法相当。此外,可以利用估计的差异对树结构进行细化,从而实现一致的场景分割和明显更好的聚合效果。
{"title":"Segment-Tree Based Cost Aggregation for Stereo Matching","authors":"Peng Yao, Hua Zhang, Yanbing Xue, Mian Zhou, Guangping Xu, Zan Gao, Shengyong Chen","doi":"10.1109/ICASSP.2017.7952512","DOIUrl":"https://doi.org/10.1109/ICASSP.2017.7952512","url":null,"abstract":"This paper presents a novel tree-based cost aggregation method for dense stereo matching. Instead of employing the minimum spanning tree (MST) and its variants, a new tree structure, \"Segment-Tree\", is proposed for non-local matching cost aggregation. Conceptually, the segment-tree is constructed in a three-step process: first, the pixels are grouped into a set of segments with the reference color or intensity image, second, a tree graph is created for each segment, and in the final step, these independent segment graphs are linked to form the segment-tree structure. In practice, this tree can be efficiently built in time nearly linear to the number of the image pixels. Compared to MST where the graph connectivity is determined with local edge weights, our method introduces some 'non-local' decision rules: the pixels in one perceptually consistent segment are more likely to share similar disparities, and therefore their connectivity within the segment should be first enforced in the tree construction process. The matching costs are then aggregated over the tree within two passes. Performance evaluation on 19 Middlebury data sets shows that the proposed method is comparable to previous state-of-the-art aggregation methods in disparity accuracy and processing speed. Furthermore, the tree structure can be refined with the estimated disparities, which leads to consistent scene segmentation and significantly better aggregation results.","PeriodicalId":6343,"journal":{"name":"2013 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-02-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87048196","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 137
Tag Taxonomy Aware Dictionary Learning for Region Tagging 基于标签分类的区域标注字典学习
Pub Date : 2013-06-23 DOI: 10.1109/CVPR.2013.54
Jingjing Zheng, Zhuolin Jiang
Tags of image regions are often arranged in a hierarchical taxonomy based on their semantic meanings. In this paper, using the given tag taxonomy, we propose to jointly learn multi-layer hierarchical dictionaries and corresponding linear classifiers for region tagging. Specifically, we generate a node-specific dictionary for each tag node in the taxonomy, and then concatenate the node-specific dictionaries from each level to construct a level-specific dictionary. The hierarchical semantic structure among tags is preserved in the relationship among node-dictionaries. Simultaneously, the sparse codes obtained using the level-specific dictionaries are summed up as the final feature representation to design a linear classifier. Our approach not only makes use of sparse codes obtained from higher levels to help learn the classifiers for lower levels, but also encourages the tag nodes from lower levels that have the same parent tag node to implicitly share sparse codes obtained from higher levels. Experimental results using three benchmark datasets show that the proposed approach yields the best performance over recently proposed methods.
图像区域的标签通常根据其语义进行分层分类。本文在给定标签分类的基础上,提出联合学习多层层次字典和相应的线性分类器进行区域标注。具体来说,我们为分类法中的每个标记节点生成一个特定于节点的字典,然后将每个级别的特定于节点的字典连接起来,以构造一个特定于级别的字典。在节点字典之间的关系中保留了标签之间的层次语义结构。同时,将使用特定级别字典获得的稀疏码汇总为最终的特征表示,以设计线性分类器。我们的方法不仅利用从更高层次获得的稀疏代码来帮助学习更低层次的分类器,而且还鼓励具有相同父标记节点的较低层次的标记节点隐式共享从更高层次获得的稀疏代码。使用三个基准数据集的实验结果表明,该方法比最近提出的方法具有最好的性能。
{"title":"Tag Taxonomy Aware Dictionary Learning for Region Tagging","authors":"Jingjing Zheng, Zhuolin Jiang","doi":"10.1109/CVPR.2013.54","DOIUrl":"https://doi.org/10.1109/CVPR.2013.54","url":null,"abstract":"Tags of image regions are often arranged in a hierarchical taxonomy based on their semantic meanings. In this paper, using the given tag taxonomy, we propose to jointly learn multi-layer hierarchical dictionaries and corresponding linear classifiers for region tagging. Specifically, we generate a node-specific dictionary for each tag node in the taxonomy, and then concatenate the node-specific dictionaries from each level to construct a level-specific dictionary. The hierarchical semantic structure among tags is preserved in the relationship among node-dictionaries. Simultaneously, the sparse codes obtained using the level-specific dictionaries are summed up as the final feature representation to design a linear classifier. Our approach not only makes use of sparse codes obtained from higher levels to help learn the classifiers for lower levels, but also encourages the tag nodes from lower levels that have the same parent tag node to implicitly share sparse codes obtained from higher levels. Experimental results using three benchmark datasets show that the proposed approach yields the best performance over recently proposed methods.","PeriodicalId":6343,"journal":{"name":"2013 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2013-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73491711","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Simultaneous Super-Resolution of Depth and Images Using a Single Camera 同时超分辨率的深度和图像使用一个单一的相机
Pub Date : 2013-06-23 DOI: 10.1109/CVPR.2013.43
Hee Seok Lee, Kyoung Mu Lee
In this paper, we propose a convex optimization framework for simultaneous estimation of super-resolved depth map and images from a single moving camera. The pixel measurement error in 3D reconstruction is directly related to the resolution of the images at hand. In turn, even a small measurement error can cause significant errors in reconstructing 3D scene structure or camera pose. Therefore, enhancing image resolution can be an effective solution for securing the accuracy as well as the resolution of 3D reconstruction. In the proposed method, depth map estimation and image super-resolution are formulated in a single energy minimization framework with a convex function and solved efficiently by a first-order primal-dual algorithm. Explicit inter-frame pixel correspondences are not required for our super-resolution procedure, thus we can avoid a huge computation time and obtain improved depth map in the accuracy and resolution as well as high-resolution images with reasonable time. The superiority of our algorithm is demonstrated by presenting the improved depth map accuracy, image super-resolution results, and camera pose estimation.
在本文中,我们提出了一个凸优化框架,用于同时估计来自单个移动摄像机的超分辨深度图和图像。三维重建中的像素测量误差直接关系到手头图像的分辨率。反过来,即使是很小的测量误差也会导致重建3D场景结构或相机姿态的重大误差。因此,提高图像分辨率是保证三维重建精度和分辨率的有效解决方案。在该方法中,深度图估计和图像超分辨率在一个带有凸函数的单一能量最小化框架中表述,并通过一阶原始对偶算法高效地求解。我们的超分辨率过程不需要明确的帧间像素对应关系,从而避免了大量的计算时间,在精度和分辨率上得到了提高的深度图,在合理的时间内得到了高分辨率的图像。通过提高深度图精度、图像超分辨率结果和相机姿态估计,证明了该算法的优越性。
{"title":"Simultaneous Super-Resolution of Depth and Images Using a Single Camera","authors":"Hee Seok Lee, Kyoung Mu Lee","doi":"10.1109/CVPR.2013.43","DOIUrl":"https://doi.org/10.1109/CVPR.2013.43","url":null,"abstract":"In this paper, we propose a convex optimization framework for simultaneous estimation of super-resolved depth map and images from a single moving camera. The pixel measurement error in 3D reconstruction is directly related to the resolution of the images at hand. In turn, even a small measurement error can cause significant errors in reconstructing 3D scene structure or camera pose. Therefore, enhancing image resolution can be an effective solution for securing the accuracy as well as the resolution of 3D reconstruction. In the proposed method, depth map estimation and image super-resolution are formulated in a single energy minimization framework with a convex function and solved efficiently by a first-order primal-dual algorithm. Explicit inter-frame pixel correspondences are not required for our super-resolution procedure, thus we can avoid a huge computation time and obtain improved depth map in the accuracy and resolution as well as high-resolution images with reasonable time. The superiority of our algorithm is demonstrated by presenting the improved depth map accuracy, image super-resolution results, and camera pose estimation.","PeriodicalId":6343,"journal":{"name":"2013 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2013-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75669127","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
Probabilistic Label Trees for Efficient Large Scale Image Classification 基于概率标记树的高效大规模图像分类
Pub Date : 2013-06-23 DOI: 10.1109/CVPR.2013.114
Baoyuan Liu, Fereshteh Sadeghi, M. Tappen, O. Shamir, Ce Liu
Large-scale recognition problems with thousands of classes pose a particular challenge because applying the classifier requires more computation as the number of classes grows. The label tree model integrates classification with the traversal of the tree so that complexity grows logarithmically. In this paper, we show how the parameters of the label tree can be found using maximum likelihood estimation. This new probabilistic learning technique produces a label tree with significantly improved recognition accuracy.
具有数千个类的大规模识别问题提出了一个特别的挑战,因为随着类数量的增加,应用分类器需要更多的计算。标签树模型将分类与树的遍历集成在一起,从而使复杂性呈对数增长。在本文中,我们展示了如何使用最大似然估计来找到标签树的参数。这种新的概率学习技术产生了一个显著提高识别精度的标签树。
{"title":"Probabilistic Label Trees for Efficient Large Scale Image Classification","authors":"Baoyuan Liu, Fereshteh Sadeghi, M. Tappen, O. Shamir, Ce Liu","doi":"10.1109/CVPR.2013.114","DOIUrl":"https://doi.org/10.1109/CVPR.2013.114","url":null,"abstract":"Large-scale recognition problems with thousands of classes pose a particular challenge because applying the classifier requires more computation as the number of classes grows. The label tree model integrates classification with the traversal of the tree so that complexity grows logarithmically. In this paper, we show how the parameters of the label tree can be found using maximum likelihood estimation. This new probabilistic learning technique produces a label tree with significantly improved recognition accuracy.","PeriodicalId":6343,"journal":{"name":"2013 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2013-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75696811","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 82
Subspace Interpolation via Dictionary Learning for Unsupervised Domain Adaptation 基于字典学习的无监督域自适应子空间插值
Pub Date : 2013-06-23 DOI: 10.1109/CVPR.2013.95
Jie Ni, Qiang Qiu, R. Chellappa
Domain adaptation addresses the problem where data instances of a source domain have different distributions from that of a target domain, which occurs frequently in many real life scenarios. This work focuses on unsupervised domain adaptation, where labeled data are only available in the source domain. We propose to interpolate subspaces through dictionary learning to link the source and target domains. These subspaces are able to capture the intrinsic domain shift and form a shared feature representation for cross domain recognition. Further, we introduce a quantitative measure to characterize the shift between two domains, which enables us to select the optimal domain to adapt to the given multiple source domains. We present experiments on face recognition across pose, illumination and blur variations, cross dataset object recognition, and report improved performance over the state of the art.
域适应解决了源域的数据实例与目标域的数据实例具有不同分布的问题,这种情况在许多实际场景中经常发生。这项工作的重点是无监督域自适应,其中标记数据仅在源域中可用。我们提出通过字典学习插值子空间来连接源域和目标域。这些子空间能够捕获固有的域位移,并形成跨域识别的共享特征表示。此外,我们引入了一种定量度量来表征两个域之间的转移,使我们能够选择最优域来适应给定的多个源域。我们提出了跨姿态、光照和模糊变化、跨数据集对象识别的人脸识别实验,并报告了在当前状态下改进的性能。
{"title":"Subspace Interpolation via Dictionary Learning for Unsupervised Domain Adaptation","authors":"Jie Ni, Qiang Qiu, R. Chellappa","doi":"10.1109/CVPR.2013.95","DOIUrl":"https://doi.org/10.1109/CVPR.2013.95","url":null,"abstract":"Domain adaptation addresses the problem where data instances of a source domain have different distributions from that of a target domain, which occurs frequently in many real life scenarios. This work focuses on unsupervised domain adaptation, where labeled data are only available in the source domain. We propose to interpolate subspaces through dictionary learning to link the source and target domains. These subspaces are able to capture the intrinsic domain shift and form a shared feature representation for cross domain recognition. Further, we introduce a quantitative measure to characterize the shift between two domains, which enables us to select the optimal domain to adapt to the given multiple source domains. We present experiments on face recognition across pose, illumination and blur variations, cross dataset object recognition, and report improved performance over the state of the art.","PeriodicalId":6343,"journal":{"name":"2013 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2013-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"72608948","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 188
Manhattan Junction Catalogue for Spatial Reasoning of Indoor Scenes 室内场景空间推理曼哈顿枢纽目录
Pub Date : 2013-06-23 DOI: 10.1109/CVPR.2013.394
S. Ramalingam, Jaishanker K. Pillai, Arpit Jain, Yuichi Taguchi
Junctions are strong cues for understanding the geometry of a scene. In this paper, we consider the problem of detecting junctions and using them for recovering the spatial layout of an indoor scene. Junction detection has always been challenging due to missing and spurious lines. We work in a constrained Manhattan world setting where the junctions are formed by only line segments along the three principal orthogonal directions. Junctions can be classified into several categories based on the number and orientations of the incident line segments. We provide a simple and efficient voting scheme to detect and classify these junctions in real images. Indoor scenes are typically modeled as cuboids and we formulate the problem of the cuboid layout estimation as an inference problem in a conditional random field. Our formulation allows the incorporation of junction features and the training is done using structured prediction techniques. We outperform other single view geometry estimation methods on standard datasets.
连接点是理解场景几何的有力线索。在本文中,我们考虑的问题是检测节点,并利用它们来恢复室内场景的空间布局。由于缺失线和杂散线的存在,连接点检测一直是一个挑战。我们在一个受约束的曼哈顿世界环境中工作,在那里,连接处仅由沿着三个主要正交方向的线段组成。根据入射线段的数量和方向,结点可以分为几类。我们提供了一种简单有效的投票方案来检测和分类真实图像中的这些连接。室内场景通常被建模为长方体,我们将长方体布局估计问题表述为条件随机场中的推理问题。我们的公式允许结合连接特征,并且使用结构化预测技术完成训练。我们在标准数据集上优于其他单视图几何估计方法。
{"title":"Manhattan Junction Catalogue for Spatial Reasoning of Indoor Scenes","authors":"S. Ramalingam, Jaishanker K. Pillai, Arpit Jain, Yuichi Taguchi","doi":"10.1109/CVPR.2013.394","DOIUrl":"https://doi.org/10.1109/CVPR.2013.394","url":null,"abstract":"Junctions are strong cues for understanding the geometry of a scene. In this paper, we consider the problem of detecting junctions and using them for recovering the spatial layout of an indoor scene. Junction detection has always been challenging due to missing and spurious lines. We work in a constrained Manhattan world setting where the junctions are formed by only line segments along the three principal orthogonal directions. Junctions can be classified into several categories based on the number and orientations of the incident line segments. We provide a simple and efficient voting scheme to detect and classify these junctions in real images. Indoor scenes are typically modeled as cuboids and we formulate the problem of the cuboid layout estimation as an inference problem in a conditional random field. Our formulation allows the incorporation of junction features and the training is done using structured prediction techniques. We outperform other single view geometry estimation methods on standard datasets.","PeriodicalId":6343,"journal":{"name":"2013 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2013-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74802054","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 59
Robust Multi-resolution Pedestrian Detection in Traffic Scenes 交通场景中鲁棒多分辨率行人检测
Pub Date : 2013-06-23 DOI: 10.1109/CVPR.2013.390
Junjie Yan, Xucong Zhang, Zhen Lei, Shengcai Liao, S. Li
The serious performance decline with decreasing resolution is the major bottleneck for current pedestrian detection techniques. In this paper, we take pedestrian detection in different resolutions as different but related problems, and propose a Multi-Task model to jointly consider their commonness and differences. The model contains resolution aware transformations to map pedestrians in different resolutions to a common space, where a shared detector is constructed to distinguish pedestrians from background. For model learning, we present a coordinate descent procedure to learn the resolution aware transformations and deformable part model (DPM) based detector iteratively. In traffic scenes, there are many false positives located around vehicles, therefore, we further build a context model to suppress them according to the pedestrian-vehicle relationship. The context model can be learned automatically even when the vehicle annotations are not available. Our method reduces the mean miss rate to 60% for pedestrians taller than 30 pixels on the Caltech Pedestrian Benchmark, which noticeably outperforms previous state-of-the-art (71%).
随着分辨率的降低,性能严重下降是当前行人检测技术的主要瓶颈。本文将不同分辨率下的行人检测视为不同但相关的问题,并提出了一个多任务模型来综合考虑它们的共性和差异性。该模型包含分辨率感知转换,将不同分辨率的行人映射到公共空间,在公共空间中构建共享检测器来区分行人和背景。在模型学习方面,我们提出了一种坐标下降方法来迭代学习分辨率感知变换和基于检测器的可变形部分模型。在交通场景中,车辆周围存在许多误报,因此,我们进一步根据行人-车辆关系建立上下文模型来抑制误报。即使在没有车辆注释的情况下,上下文模型也可以自动学习。在加州理工学院行人基准测试中,我们的方法将身高超过30像素的行人的平均失分率降低到60%,明显优于之前的先进技术(71%)。
{"title":"Robust Multi-resolution Pedestrian Detection in Traffic Scenes","authors":"Junjie Yan, Xucong Zhang, Zhen Lei, Shengcai Liao, S. Li","doi":"10.1109/CVPR.2013.390","DOIUrl":"https://doi.org/10.1109/CVPR.2013.390","url":null,"abstract":"The serious performance decline with decreasing resolution is the major bottleneck for current pedestrian detection techniques. In this paper, we take pedestrian detection in different resolutions as different but related problems, and propose a Multi-Task model to jointly consider their commonness and differences. The model contains resolution aware transformations to map pedestrians in different resolutions to a common space, where a shared detector is constructed to distinguish pedestrians from background. For model learning, we present a coordinate descent procedure to learn the resolution aware transformations and deformable part model (DPM) based detector iteratively. In traffic scenes, there are many false positives located around vehicles, therefore, we further build a context model to suppress them according to the pedestrian-vehicle relationship. The context model can be learned automatically even when the vehicle annotations are not available. Our method reduces the mean miss rate to 60% for pedestrians taller than 30 pixels on the Caltech Pedestrian Benchmark, which noticeably outperforms previous state-of-the-art (71%).","PeriodicalId":6343,"journal":{"name":"2013 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2013-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"72675360","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 187
Blur Processing Using Double Discrete Wavelet Transform 基于双离散小波变换的模糊处理
Pub Date : 2013-06-23 DOI: 10.1109/CVPR.2013.145
Yi Zhang, Keigo Hirakawa
We propose a notion of double discrete wavelet transform (DDWT) that is designed to sparsify the blurred image and the blur kernel simultaneously. DDWT greatly enhances our ability to analyze, detect, and process blur kernels and blurry images-the proposed framework handles both global and spatially varying blur kernels seamlessly, and unifies the treatment of blur caused by object motion, optical defocus, and camera shake. To illustrate the potential of DDWT in computer vision and image processing, we develop example applications in blur kernel estimation, deblurring, and near-blur-invariant image feature extraction.
我们提出了一种双离散小波变换(DDWT)的概念,旨在同时对模糊图像和模糊核进行稀疏化。DDWT极大地提高了我们分析、检测和处理模糊核和模糊图像的能力——所提出的框架无缝地处理全局和空间变化的模糊核,并统一处理由物体运动、光学离焦和相机抖动引起的模糊。为了说明DDWT在计算机视觉和图像处理中的潜力,我们开发了模糊核估计,去模糊和近模糊不变图像特征提取的示例应用程序。
{"title":"Blur Processing Using Double Discrete Wavelet Transform","authors":"Yi Zhang, Keigo Hirakawa","doi":"10.1109/CVPR.2013.145","DOIUrl":"https://doi.org/10.1109/CVPR.2013.145","url":null,"abstract":"We propose a notion of double discrete wavelet transform (DDWT) that is designed to sparsify the blurred image and the blur kernel simultaneously. DDWT greatly enhances our ability to analyze, detect, and process blur kernels and blurry images-the proposed framework handles both global and spatially varying blur kernels seamlessly, and unifies the treatment of blur caused by object motion, optical defocus, and camera shake. To illustrate the potential of DDWT in computer vision and image processing, we develop example applications in blur kernel estimation, deblurring, and near-blur-invariant image feature extraction.","PeriodicalId":6343,"journal":{"name":"2013 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2013-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"72683702","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 51
Dense Non-rigid Point-Matching Using Random Projections 使用随机投影的密集非刚性点匹配
Pub Date : 2013-06-23 DOI: 10.1109/CVPR.2013.375
Raffay Hamid, D. DeCoste, Chih-Jen Lin
We present a robust and efficient technique for matching dense sets of points undergoing non-rigid spatial transformations. Our main intuition is that the subset of points that can be matched with high confidence should be used to guide the matching procedure for the rest. We propose a novel algorithm that incorporates these high-confidence matches as a spatial prior to learn a discriminative subspace that simultaneously encodes both the feature similarity as well as their spatial arrangement. Conventional subspace learning usually requires spectral decomposition of the pair-wise distance matrix across the point-sets, which can become inefficient even for moderately sized problems. To this end, we propose the use of random projections for approximate subspace learning, which can provide significant time improvements at the cost of minimal precision loss. This efficiency gain allows us to iteratively find and remove high-confidence matches from the point sets, resulting in high recall. To show the effectiveness of our approach, we present a systematic set of experiments and results for the problem of dense non-rigid image-feature matching.
我们提出了一种鲁棒和高效的方法来匹配经过非刚性空间变换的密集点集。我们的主要直觉是,可以高置信度匹配的点子集应该用来指导其余点的匹配过程。我们提出了一种新的算法,该算法将这些高置信度匹配作为空间先验来学习一个判别子空间,该子空间同时编码特征相似性及其空间排列。传统的子空间学习通常需要对点集上的成对距离矩阵进行谱分解,即使对于中等规模的问题,这种方法也会变得效率低下。为此,我们建议使用随机投影进行近似子空间学习,它可以以最小的精度损失为代价提供显著的时间改进。这种效率的提高使我们能够迭代地从点集中找到并删除高置信度的匹配,从而获得高召回率。为了证明我们方法的有效性,我们提供了一组系统的实验和结果,用于密集非刚性图像特征匹配问题。
{"title":"Dense Non-rigid Point-Matching Using Random Projections","authors":"Raffay Hamid, D. DeCoste, Chih-Jen Lin","doi":"10.1109/CVPR.2013.375","DOIUrl":"https://doi.org/10.1109/CVPR.2013.375","url":null,"abstract":"We present a robust and efficient technique for matching dense sets of points undergoing non-rigid spatial transformations. Our main intuition is that the subset of points that can be matched with high confidence should be used to guide the matching procedure for the rest. We propose a novel algorithm that incorporates these high-confidence matches as a spatial prior to learn a discriminative subspace that simultaneously encodes both the feature similarity as well as their spatial arrangement. Conventional subspace learning usually requires spectral decomposition of the pair-wise distance matrix across the point-sets, which can become inefficient even for moderately sized problems. To this end, we propose the use of random projections for approximate subspace learning, which can provide significant time improvements at the cost of minimal precision loss. This efficiency gain allows us to iteratively find and remove high-confidence matches from the point sets, resulting in high recall. To show the effectiveness of our approach, we present a systematic set of experiments and results for the problem of dense non-rigid image-feature matching.","PeriodicalId":6343,"journal":{"name":"2013 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2013-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79342251","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Axially Symmetric 3D Pots Configuration System Using Axis of Symmetry and Break Curve 利用对称轴和断裂曲线的轴对称三维罐形系统
Pub Date : 2013-06-23 DOI: 10.1109/CVPR.2013.40
Kilho Son, Eduardo B. Almeida, D. Cooper
This paper introduces a novel approach for reassembling pot sherds found at archaeological excavation sites, for the purpose of reconstructing clay pots that had been made on a wheel. These pots and the sherds into which they have broken are axially symmetric. The reassembly process can be viewed as 3D puzzle solving or generalized cylinder learning from broken fragments. The estimation exploits both local and semi-global geometric structure, thus making it a fundamental problem of geometry estimation from noisy fragments in computer vision and pattern recognition. The data used are densely digitized 3D laser scans of each fragment's outer surface. The proposed reassembly system is automatic and functions when the pile of available fragments is from one or multiple pots, and even when pieces are missing from any pot. The geometric structure used are curves on the pot along which the surface had broken and the silhouette of a pot with respect to an axis, called axis-profile curve (APC). For reassembling multiple pots with or without missing pieces, our algorithm estimates the APC from each fragment, then reassembles into configurations the ones having distinctive APC. Further growth of configurations is based on adding remaining fragments such that their APC and break curves are consistent with those of a configuration. The method is novel, more robust and handles the largest numbers of fragments to date.
本文介绍了一种重组考古发掘现场发现的陶罐碎片的新方法,目的是重建在车轮上制作的陶罐。这些罐子和它们破碎成的碎片是轴对称的。重组过程可以看作是3D解谜或从破碎的碎片中学习广义的圆柱体。该算法同时利用了局部和半全局的几何结构,是计算机视觉和模式识别中基于噪声碎片的几何估计的一个基本问题。使用的数据是每个碎片外表面的密集数字化3D激光扫描。所提出的重组系统是自动的,当可用的碎片堆来自一个或多个罐时,甚至当任何罐中都有碎片丢失时,它都能起作用。所使用的几何结构是罐表面沿破裂的曲线和罐相对于轴的轮廓,称为轴廓曲线(APC)。对于有或没有缺失碎片的多个罐,我们的算法从每个碎片中估计APC,然后重新组装成具有不同APC的配置。构型的进一步增长是基于添加剩余的片段,使它们的APC和断裂曲线与构型的一致。该方法新颖,鲁棒性更强,处理的片段数量最多。
{"title":"Axially Symmetric 3D Pots Configuration System Using Axis of Symmetry and Break Curve","authors":"Kilho Son, Eduardo B. Almeida, D. Cooper","doi":"10.1109/CVPR.2013.40","DOIUrl":"https://doi.org/10.1109/CVPR.2013.40","url":null,"abstract":"This paper introduces a novel approach for reassembling pot sherds found at archaeological excavation sites, for the purpose of reconstructing clay pots that had been made on a wheel. These pots and the sherds into which they have broken are axially symmetric. The reassembly process can be viewed as 3D puzzle solving or generalized cylinder learning from broken fragments. The estimation exploits both local and semi-global geometric structure, thus making it a fundamental problem of geometry estimation from noisy fragments in computer vision and pattern recognition. The data used are densely digitized 3D laser scans of each fragment's outer surface. The proposed reassembly system is automatic and functions when the pile of available fragments is from one or multiple pots, and even when pieces are missing from any pot. The geometric structure used are curves on the pot along which the surface had broken and the silhouette of a pot with respect to an axis, called axis-profile curve (APC). For reassembling multiple pots with or without missing pieces, our algorithm estimates the APC from each fragment, then reassembles into configurations the ones having distinctive APC. Further growth of configurations is based on adding remaining fragments such that their APC and break curves are consistent with those of a configuration. The method is novel, more robust and handles the largest numbers of fragments to date.","PeriodicalId":6343,"journal":{"name":"2013 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2013-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85487967","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 43
期刊
2013 IEEE Conference on Computer Vision and Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1