首页 > 最新文献

2009 IEEE Conference on Computer Vision and Pattern Recognition最新文献

英文 中文
Cancelable iris biometrics and using Error Correcting Codes to reduce variability in biometric data 可取消虹膜生物特征和使用纠错码来减少生物特征数据的可变性
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206646
S. Kanade, D. Petrovska-Delacrétaz, B. Dorizzi
With the increasing use of biometrics, more and more concerns are being raised about the privacy of the personal biometric data. Conventional biometric systems store biometric templates in a database. This may lead to the possibility of tracking personal information stored in one database by getting access to another database through cross-database matching. Moreover, biometric data are permanently associated with the user. Hence if stolen, they are lost permanently and become unusable in that system and possibly in all other systems based on that biometrics. In order to overcome this non-revocability of biometrics, we propose a two factor scheme to generate cancelable iris templates using iris-biometric and password. We employ a user specific shuffling key to shuffle the iris codes. Additionally, we introduce a novel way to use error correcting codes (ECC) to reduce the variabilities in biometric data. The shuffling scheme increases the impostor Hamming distance leaving genuine Hamming distance intact while the ECC reduce the Hamming distance for genuine comparisons by a larger amount than for the impostor comparisons. This results in better separation between genuine and impostor users which improves the verification performance. The shuffling key is protected by a password which makes the system truly revocable. The biometric data is stored in a protected form which protects the privacy. The proposed scheme reduces the equal error rate (EER) of the system by more than 90% (e.g., from 1.70% to 0.057% on the NIST-ICE database).
随着生物识别技术应用的日益广泛,个人生物识别数据的隐私问题越来越受到人们的关注。传统的生物识别系统将生物识别模板存储在数据库中。这可能导致通过跨数据库匹配访问另一个数据库来跟踪存储在一个数据库中的个人信息的可能性。此外,生物特征数据与用户永久关联。因此,如果被盗,它们将永久丢失,并且在该系统中无法使用,并且可能在基于该生物识别的所有其他系统中也无法使用。为了克服生物特征的不可撤销性,我们提出了一种利用虹膜生物特征和密码生成可取消虹膜模板的双因素方案。我们使用用户特定的洗牌密钥来洗牌虹膜代码。此外,我们还介绍了一种使用纠错码(ECC)的新方法来减少生物特征数据的可变性。洗牌方案增加了冒名顶替者的汉明距离,使真正的汉明距离保持不变,而ECC减少了真正比较的汉明距离,其幅度大于冒名顶替者比较。这样可以更好地区分真实用户和冒名顶替用户,从而提高验证性能。变换密钥由密码保护,使系统真正可撤销。生物特征数据以保护隐私的保护形式存储。该方案将系统的等错误率(EER)降低了90%以上(例如,在NIST-ICE数据库上从1.70%降低到0.057%)。
{"title":"Cancelable iris biometrics and using Error Correcting Codes to reduce variability in biometric data","authors":"S. Kanade, D. Petrovska-Delacrétaz, B. Dorizzi","doi":"10.1109/CVPR.2009.5206646","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206646","url":null,"abstract":"With the increasing use of biometrics, more and more concerns are being raised about the privacy of the personal biometric data. Conventional biometric systems store biometric templates in a database. This may lead to the possibility of tracking personal information stored in one database by getting access to another database through cross-database matching. Moreover, biometric data are permanently associated with the user. Hence if stolen, they are lost permanently and become unusable in that system and possibly in all other systems based on that biometrics. In order to overcome this non-revocability of biometrics, we propose a two factor scheme to generate cancelable iris templates using iris-biometric and password. We employ a user specific shuffling key to shuffle the iris codes. Additionally, we introduce a novel way to use error correcting codes (ECC) to reduce the variabilities in biometric data. The shuffling scheme increases the impostor Hamming distance leaving genuine Hamming distance intact while the ECC reduce the Hamming distance for genuine comparisons by a larger amount than for the impostor comparisons. This results in better separation between genuine and impostor users which improves the verification performance. The shuffling key is protected by a password which makes the system truly revocable. The biometric data is stored in a protected form which protects the privacy. The proposed scheme reduces the equal error rate (EER) of the system by more than 90% (e.g., from 1.70% to 0.057% on the NIST-ICE database).","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126695730","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 86
On edge detection on surfaces 关于曲面的边缘检测
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206517
Michael Kolomenkin, I. Shimshoni, A. Tal
Edge detection in images has been a fundamental problem in computer vision from its early days. Edge detection on surfaces, on the other hand, has received much less attention. The most common edges on surfaces are ridges and valleys, used for processing range images in computer vision, as well as for non-photo realistic rendering in computer graphics. We propose a new type of edges on surfaces, termed relief edges. Intuitively, the surface can be considered as an unknown smooth manifold, on top of which a local height image is placed. Relief edges are the edges of this local image. We show how to compute these edges from the local differential geometric surface properties, by fitting a local edge model to the surface. We also show how the underlying manifold and the local images can be roughly approximated and exploited in the edge detection process. Last but not least, we demonstrate the application of relief edges to artifact illustration in archaeology.
图像的边缘检测从早期开始就是计算机视觉的一个基本问题。另一方面,表面上的边缘检测受到的关注较少。表面上最常见的边缘是脊和谷,用于处理计算机视觉中的范围图像,以及计算机图形学中的非照片逼真渲染。我们提出了一种新的曲面边,称为浮雕边。直观地,表面可以看作是一个未知的光滑流形,在其上放置一个局部高度图像。浮雕边缘是这个局部图像的边缘。我们展示了如何通过拟合表面的局部边缘模型,从局部微分几何表面属性计算这些边缘。我们还展示了如何在边缘检测过程中大致近似和利用底层流形和局部图像。最后但并非最不重要的是,我们展示浮雕边缘在考古文物插图中的应用。
{"title":"On edge detection on surfaces","authors":"Michael Kolomenkin, I. Shimshoni, A. Tal","doi":"10.1109/CVPR.2009.5206517","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206517","url":null,"abstract":"Edge detection in images has been a fundamental problem in computer vision from its early days. Edge detection on surfaces, on the other hand, has received much less attention. The most common edges on surfaces are ridges and valleys, used for processing range images in computer vision, as well as for non-photo realistic rendering in computer graphics. We propose a new type of edges on surfaces, termed relief edges. Intuitively, the surface can be considered as an unknown smooth manifold, on top of which a local height image is placed. Relief edges are the edges of this local image. We show how to compute these edges from the local differential geometric surface properties, by fitting a local edge model to the surface. We also show how the underlying manifold and the local images can be roughly approximated and exploited in the edge detection process. Last but not least, we demonstrate the application of relief edges to artifact illustration in archaeology.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130577249","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 72
View-invariant dynamic texture recognition using a bag of dynamical systems 使用一组动态系统的视不变动态纹理识别
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206847
Avinash Ravichandran, Rizwan Ahmed Chaudhry, R. Vidal
In this paper, we consider the problem of categorizing videos of dynamic textures under varying view-point. We propose to model each video with a collection of linear dynamics systems (LDSs) describing the dynamics of spatiotemporal video patches. This bag of systems (BoS) representation is analogous to the bag of features (BoF) representation, except that we use LDSs as feature descriptors. This poses several technical challenges to the BoF framework. Most notably, LDSs do not live in a Euclidean space, hence novel methods for clustering LDSs and computing codewords of LDSs need to be developed. Our framework makes use of nonlinear dimensionality reduction and clustering techniques combined with the Martin distance for LDSs for tackling these issues. Our experiments show that our BoS approach can be used for recognizing dynamic textures in challenging scenarios, which could not be handled by existing dynamic texture recognition methods.
本文研究了动态纹理视频在不同视点下的分类问题。我们建议用一组描述时空视频补丁动态的线性动态系统(lds)对每个视频进行建模。这种系统包(BoS)表示类似于特征包(BoF)表示,只不过我们使用lds作为特征描述符。这对BoF框架提出了几个技术挑战。最值得注意的是,lds不存在于欧几里得空间中,因此需要开发新的lds聚类方法和lds码字计算方法。我们的框架利用非线性降维和聚类技术结合lds的马丁距离来解决这些问题。实验表明,该方法可用于识别现有动态纹理识别方法无法处理的复杂场景下的动态纹理。
{"title":"View-invariant dynamic texture recognition using a bag of dynamical systems","authors":"Avinash Ravichandran, Rizwan Ahmed Chaudhry, R. Vidal","doi":"10.1109/CVPR.2009.5206847","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206847","url":null,"abstract":"In this paper, we consider the problem of categorizing videos of dynamic textures under varying view-point. We propose to model each video with a collection of linear dynamics systems (LDSs) describing the dynamics of spatiotemporal video patches. This bag of systems (BoS) representation is analogous to the bag of features (BoF) representation, except that we use LDSs as feature descriptors. This poses several technical challenges to the BoF framework. Most notably, LDSs do not live in a Euclidean space, hence novel methods for clustering LDSs and computing codewords of LDSs need to be developed. Our framework makes use of nonlinear dimensionality reduction and clustering techniques combined with the Martin distance for LDSs for tackling these issues. Our experiments show that our BoS approach can be used for recognizing dynamic textures in challenging scenarios, which could not be handled by existing dynamic texture recognition methods.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124238226","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 132
Real-time learning of accurate patch rectification 实时学习准确的补丁整改
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206794
Stefan Hinterstoißer, Oliver Kutter, Nassir Navab, P. Fua, V. Lepetit
Recent work showed that learning-based patch rectification methods are both faster and more reliable than affine region methods. Unfortunately, their performance improvements are founded in a computationally expensive offline learning stage, which is not possible for applications such as SLAM. In this paper we propose an approach whose training stage is fast enough to be performed at run-time without the loss of accuracy or robustness. To this end, we developed a very fast method to compute the mean appearances of the feature points over sets of small variations that span the range of possible camera viewpoints. Then, by simply matching incoming feature points against these mean appearances, we get a coarse estimate of the viewpoint that is refined afterwards. Because there is no need to compute descriptors for the input image, the method is very fast at run-time. We demonstrate our approach on tracking-by-detection for SLAM, real-time object detection and pose estimation applications.
最近的研究表明,基于学习的补丁校正方法比仿射区域方法更快、更可靠。不幸的是,它们的性能改进是建立在计算昂贵的离线学习阶段,这对于SLAM等应用程序是不可能的。在本文中,我们提出了一种训练阶段足够快的方法,可以在不损失准确性和鲁棒性的情况下在运行时执行。为此,我们开发了一种非常快速的方法来计算跨越可能的相机视点范围的小变化集的特征点的平均外观。然后,通过简单地将输入的特征点与这些平均外观进行匹配,我们得到一个粗略的视点估计,然后进行细化。由于不需要计算输入图像的描述符,因此该方法在运行时非常快。我们展示了SLAM、实时目标检测和姿态估计应用的检测跟踪方法。
{"title":"Real-time learning of accurate patch rectification","authors":"Stefan Hinterstoißer, Oliver Kutter, Nassir Navab, P. Fua, V. Lepetit","doi":"10.1109/CVPR.2009.5206794","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206794","url":null,"abstract":"Recent work showed that learning-based patch rectification methods are both faster and more reliable than affine region methods. Unfortunately, their performance improvements are founded in a computationally expensive offline learning stage, which is not possible for applications such as SLAM. In this paper we propose an approach whose training stage is fast enough to be performed at run-time without the loss of accuracy or robustness. To this end, we developed a very fast method to compute the mean appearances of the feature points over sets of small variations that span the range of possible camera viewpoints. Then, by simply matching incoming feature points against these mean appearances, we get a coarse estimate of the viewpoint that is refined afterwards. Because there is no need to compute descriptors for the input image, the method is very fast at run-time. We demonstrate our approach on tracking-by-detection for SLAM, real-time object detection and pose estimation applications.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116369779","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 34
Removing partial blur in a single image 去除单个图像中的部分模糊
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206625
Shengyang Dai, Ying Wu
Removing image partial blur is of great practical importance. However, as existing recovery techniques usually assume a one-layer clear image model, they can not characterize the actual generation process of partial blurs. In this paper, a two-layer image model is investigated. Based on the study of partial blur generation process, a novel recovery technique is proposed for a single input image. Both foreground and background layers are recovered simultaneously with the help of the matting technique, powerful image prior models, and user assistance. The effectiveness of the proposed approach is demonstrated by extensive experiments on image recovery and synthesis on real data.
消除图像局部模糊具有重要的实际意义。然而,由于现有的恢复技术通常假设一个单层的清晰图像模型,它们不能表征部分模糊的实际生成过程。本文研究了一种双层图像模型。在研究局部模糊生成过程的基础上,提出了一种新的单输入图像恢复技术。在消光技术、强大的图像先验模型和用户辅助的帮助下,前景层和背景层同时恢复。在实际数据上进行了大量的图像恢复和合成实验,证明了该方法的有效性。
{"title":"Removing partial blur in a single image","authors":"Shengyang Dai, Ying Wu","doi":"10.1109/CVPR.2009.5206625","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206625","url":null,"abstract":"Removing image partial blur is of great practical importance. However, as existing recovery techniques usually assume a one-layer clear image model, they can not characterize the actual generation process of partial blurs. In this paper, a two-layer image model is investigated. Based on the study of partial blur generation process, a novel recovery technique is proposed for a single input image. Both foreground and background layers are recovered simultaneously with the help of the matting technique, powerful image prior models, and user assistance. The effectiveness of the proposed approach is demonstrated by extensive experiments on image recovery and synthesis on real data.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114776078","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 59
A unified model of specular and diffuse reflectance for rough, glossy surfaces 粗糙光滑表面的镜面反射和漫反射的统一模型
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206498
W. Smith, E. Hancock
In this paper we consider diffuse and specular reflectance from surfaces modeled as distributions of glossy microfacets. In contrast to previous work, we describe the relative contribution of both of these components in the same terms, namely with resource to Fresnel theory. This results in a more highly constrained model with a reduced number of parameters. Also, the need for ad hoc and physically meaningless specular and diffuse reflectance coefficients is removed. This ensures that the conservation of energy is obeyed and only physically plausible mixtures of the two components are allowed. In our model, both specular and diffuse reflectance are related to the roughness and refractive index of the surface. We show how physically meaningful parameters of a surface can be measured from uncalibrated imagery and that our model fits observed BRDF data more accurately than comparable existing models.
在本文中,我们考虑漫反射和镜面反射从表面建模为光滑的微面分布。与以前的工作相反,我们以相同的术语描述了这两个组件的相对贡献,即与菲涅耳理论的资源。这将产生一个具有更少参数数量的高度约束的模型。此外,需要特别的和物理上无意义的镜面和漫反射系数被删除。这确保了能量守恒定律得到遵守,并且只允许这两种成分在物理上合理的混合。在我们的模型中,镜面和漫反射都与表面的粗糙度和折射率有关。我们展示了如何从未校准的图像中测量表面的物理意义参数,并且我们的模型比可比的现有模型更准确地拟合观测到的BRDF数据。
{"title":"A unified model of specular and diffuse reflectance for rough, glossy surfaces","authors":"W. Smith, E. Hancock","doi":"10.1109/CVPR.2009.5206498","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206498","url":null,"abstract":"In this paper we consider diffuse and specular reflectance from surfaces modeled as distributions of glossy microfacets. In contrast to previous work, we describe the relative contribution of both of these components in the same terms, namely with resource to Fresnel theory. This results in a more highly constrained model with a reduced number of parameters. Also, the need for ad hoc and physically meaningless specular and diffuse reflectance coefficients is removed. This ensures that the conservation of energy is obeyed and only physically plausible mixtures of the two components are allowed. In our model, both specular and diffuse reflectance are related to the roughness and refractive index of the surface. We show how physically meaningful parameters of a surface can be measured from uncalibrated imagery and that our model fits observed BRDF data more accurately than comparable existing models.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114862349","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Shape priors and discrete MRFs for knowledge-based segmentation 基于知识分割的形状先验和离散mrf
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206649
A. Besbes, N. Komodakis, G. Langs, N. Paragios
In this paper we introduce a new approach to knowledge-based segmentation. Our method consists of a novel representation to model shape variations as well as an efficient inference procedure to fit the model to new data. The considered shape model is similarity-invariant and refers to an incomplete graph that consists of intra and intercluster connections representing the inter-dependencies of control points. The clusters are determined according to the co-dependencies of the deformations of the control points within the training set. The connections between the components of a cluster represent the local structure while the connections between the clusters account for the global structure. The distributions of the normalized distances between the connected control points encode the prior model. During search, this model is used together with a discrete Markov random field (MRF) based segmentation, where the unknown variables are the positions of the control points in the image domain. To encode the image support, a Voronoi decomposition of the domain is considered and regional based statistics are used. The resulting model is computationally efficient, can encode complex statistical models of shape variations and benefits from the image support of the entire spatial domain.
本文提出了一种新的基于知识的分割方法。我们的方法包括一种新的形状变化模型表示和一种有效的推理程序来拟合新的数据模型。所考虑的形状模型是相似不变的,它指的是由表示控制点相互依赖关系的簇内和簇间连接组成的不完全图。聚类是根据训练集中控制点变形的相互依赖性来确定的。集群组件之间的连接代表了局部结构,而集群之间的连接代表了全局结构。连接控制点之间归一化距离的分布对先验模型进行编码。在搜索过程中,该模型与基于离散马尔可夫随机场(MRF)的分割一起使用,其中未知变量是控制点在图像域中的位置。为了编码图像支持,考虑了域的Voronoi分解和基于区域的统计。该模型计算效率高,可以对形状变化的复杂统计模型进行编码,并受益于整个空间域的图像支持。
{"title":"Shape priors and discrete MRFs for knowledge-based segmentation","authors":"A. Besbes, N. Komodakis, G. Langs, N. Paragios","doi":"10.1109/CVPR.2009.5206649","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206649","url":null,"abstract":"In this paper we introduce a new approach to knowledge-based segmentation. Our method consists of a novel representation to model shape variations as well as an efficient inference procedure to fit the model to new data. The considered shape model is similarity-invariant and refers to an incomplete graph that consists of intra and intercluster connections representing the inter-dependencies of control points. The clusters are determined according to the co-dependencies of the deformations of the control points within the training set. The connections between the components of a cluster represent the local structure while the connections between the clusters account for the global structure. The distributions of the normalized distances between the connected control points encode the prior model. During search, this model is used together with a discrete Markov random field (MRF) based segmentation, where the unknown variables are the positions of the control points in the image domain. To encode the image support, a Voronoi decomposition of the domain is considered and regional based statistics are used. The resulting model is computationally efficient, can encode complex statistical models of shape variations and benefits from the image support of the entire spatial domain.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128068743","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 56
Spatiotemporal stereo via spatiotemporal quadric element (stequel) matching 通过时空二次元(stequel)匹配实现时空立体
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206728
Mikhail Sizintsev, Richard P. Wildes
Spatiotemporal stereo is concerned with the recovery of the 3D structure of a dynamic scene from a temporal sequence of multiview images. This paper presents a novel method for computing temporally coherent disparity maps from a sequence of binocular images through an integrated consideration of image spacetime structure and without explicit recovery of motion. The approach is based on matching spatiotemporal quadric elements (stequels) between views, as it is shown that this matching primitive provides a natural way to encapsulate both local spatial and temporal structure for disparity estimation. Empirical evaluation with laboratory based imagery with ground truth and more typical natural imagery shows that the approach provides considerable benefit in comparison to alternative methods for enforcing temporal coherence in disparity estimation.
时空立体关注的是从多视图图像的时间序列中恢复动态场景的3D结构。本文提出了一种综合考虑图像时空结构而不需要明确运动恢复的双目图像序列计算时间相干视差图的新方法。该方法基于视图之间的时空二次元匹配,因为这种匹配原语提供了一种自然的方式来封装局部时空结构以进行视差估计。对具有地面真实图像的实验室图像和更典型的自然图像进行的经验评估表明,与其他方法相比,该方法在视差估计中提供了相当大的优势。
{"title":"Spatiotemporal stereo via spatiotemporal quadric element (stequel) matching","authors":"Mikhail Sizintsev, Richard P. Wildes","doi":"10.1109/CVPR.2009.5206728","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206728","url":null,"abstract":"Spatiotemporal stereo is concerned with the recovery of the 3D structure of a dynamic scene from a temporal sequence of multiview images. This paper presents a novel method for computing temporally coherent disparity maps from a sequence of binocular images through an integrated consideration of image spacetime structure and without explicit recovery of motion. The approach is based on matching spatiotemporal quadric elements (stequels) between views, as it is shown that this matching primitive provides a natural way to encapsulate both local spatial and temporal structure for disparity estimation. Empirical evaluation with laboratory based imagery with ground truth and more typical natural imagery shows that the approach provides considerable benefit in comparison to alternative methods for enforcing temporal coherence in disparity estimation.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132320146","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 30
Image registration by minimization of residual complexity 残差复杂度最小化的图像配准方法
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206571
A. Myronenko, Xubo B. Song
Accurate definition of similarity measure is a key component in image registration. Most commonly used intensity-based similarity measures rely on the assumptions of independence and stationarity of the intensities from pixel to pixel. Such measures cannot capture the complex interactions among the pixel intensities, and often result in less satisfactory registration performances, especially in the presence of nonstationary intensity distortions. We propose a novel similarity measure that accounts for intensity non-stationarities and complex spatially-varying intensity distortions. We derive the similarity measure by analytically solving for the intensity correction field and its adaptive regularization. The final measure can be interpreted as one that favors a registration with minimum compression complexity of the residual image between the two registered images. This measure produces accurate registration results on both artificial and real-world problems that we have tested, whereas many other state-of-the-art similarity measures have failed to do so.
准确定义相似度是图像配准的关键。最常用的基于强度的相似性度量依赖于像素之间强度的独立性和平稳性的假设。这种方法不能捕获像素强度之间复杂的相互作用,并且通常导致不太令人满意的配准性能,特别是在存在非平稳强度扭曲的情况下。我们提出了一种新的相似性度量,该度量考虑了强度非平稳性和复杂的空间变化强度扭曲。通过解析求解强度校正场及其自适应正则化,推导出相似性测度。最后的度量可以被解释为一个倾向于在两个配准图像之间的残差图像具有最小压缩复杂度的配准。这种方法在我们测试过的人工问题和现实问题上都产生了准确的注册结果,而许多其他最先进的相似度方法都没有做到这一点。
{"title":"Image registration by minimization of residual complexity","authors":"A. Myronenko, Xubo B. Song","doi":"10.1109/CVPR.2009.5206571","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206571","url":null,"abstract":"Accurate definition of similarity measure is a key component in image registration. Most commonly used intensity-based similarity measures rely on the assumptions of independence and stationarity of the intensities from pixel to pixel. Such measures cannot capture the complex interactions among the pixel intensities, and often result in less satisfactory registration performances, especially in the presence of nonstationary intensity distortions. We propose a novel similarity measure that accounts for intensity non-stationarities and complex spatially-varying intensity distortions. We derive the similarity measure by analytically solving for the intensity correction field and its adaptive regularization. The final measure can be interpreted as one that favors a registration with minimum compression complexity of the residual image between the two registered images. This measure produces accurate registration results on both artificial and real-world problems that we have tested, whereas many other state-of-the-art similarity measures have failed to do so.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131592526","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 61
Fast human detection in crowded scenes by contour integration and local shape estimation 基于轮廓积分和局部形状估计的拥挤场景快速人体检测
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206564
Csaba Beleznai, H. Bischof
The complexity of human detection increases significantly with a growing density of humans populating a scene. This paper presents a Bayesian detection framework using shape and motion cues to obtain a maximum a posteriori (MAP) solution for human configurations consisting of many, possibly occluded pedestrians viewed by a stationary camera. The paper contains two novel contributions for the human detection task: 1. computationally efficient detection based on shape templates using contour integration by means of integral images which are built by oriented string scans; (2) a non-parametric approach using an approximated version of the shape context descriptor which generates informative object parts and infers the presence of humans despite occlusions. The outputs of the two detectors are used to generate a spatial configuration of hypothesized human body locations. The configuration is iteratively optimized while taking into account the depth ordering and occlusion status of the hypotheses. The method achieves fast computation times even in complex scenarios with a high density of people. Its validity is demonstrated on a substantial amount of image data using the CAVIAR and our own datasets. Evaluation results and comparison with state of the art are presented.
随着场景中人口密度的增加,人类检测的复杂性也会显著增加。本文提出了一个贝叶斯检测框架,使用形状和运动线索来获得一个最大后验(MAP)解决方案,用于由固定摄像机观察到的由许多可能被遮挡的行人组成的人类配置。本文对人类检测任务有两个新的贡献:1。基于轮廓积分的基于形状模板的高效检测方法,采用定向串扫描生成的积分图像(2)使用形状上下文描述符的近似版本的非参数方法,该方法生成信息丰富的物体部分,并推断出尽管有遮挡,但人类的存在。两个探测器的输出用于生成假设人体位置的空间配置。在考虑假设的深度排序和遮挡状态的情况下,对配置进行迭代优化。即使在人口密集的复杂场景下,该方法也能实现快速的计算速度。它的有效性证明了大量的图像数据使用CAVIAR和我们自己的数据集。给出了评价结果,并与现有技术进行了比较。
{"title":"Fast human detection in crowded scenes by contour integration and local shape estimation","authors":"Csaba Beleznai, H. Bischof","doi":"10.1109/CVPR.2009.5206564","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206564","url":null,"abstract":"The complexity of human detection increases significantly with a growing density of humans populating a scene. This paper presents a Bayesian detection framework using shape and motion cues to obtain a maximum a posteriori (MAP) solution for human configurations consisting of many, possibly occluded pedestrians viewed by a stationary camera. The paper contains two novel contributions for the human detection task: 1. computationally efficient detection based on shape templates using contour integration by means of integral images which are built by oriented string scans; (2) a non-parametric approach using an approximated version of the shape context descriptor which generates informative object parts and infers the presence of humans despite occlusions. The outputs of the two detectors are used to generate a spatial configuration of hypothesized human body locations. The configuration is iteratively optimized while taking into account the depth ordering and occlusion status of the hypotheses. The method achieves fast computation times even in complex scenarios with a high density of people. Its validity is demonstrated on a substantial amount of image data using the CAVIAR and our own datasets. Evaluation results and comparison with state of the art are presented.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129394622","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 61
期刊
2009 IEEE Conference on Computer Vision and Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1