首页 > 最新文献

2007 IEEE Conference on Computer Vision and Pattern Recognition最新文献

英文 中文
Secure Biometric Templates from Fingerprint-Face Features 从指纹-面部特征安全的生物识别模板
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383385
Y. Sutcu, Qiming Li, N. Memon
Since biometric data cannot be easily replaced or revoked, it is important that biometric templates used in biometric applications should be constructed and stored in a secure way, such that attackers would not be able to forge biometric data easily even when the templates are compromised. This is a challenging goal since biometric data are "noisy" by nature, and the matching algorithms are often complex, which make it difficult to apply traditional cryptographic techniques, especially when multiple modalities are considered. In this paper, we consider a "fusion " of a minutiae-based fingerprint authentication scheme and an SVD-based face authentication scheme, and show that by employing a recently proposed cryptographic primitive called "secure sketch ", and a known geometric transformation on minutiae, we can make it easier to combine different modalities, and at the same time make it computationally infeasible to forge an "original" combination of fingerprint and face image that passes the authentication. We evaluate the effectiveness of our scheme using real fingerprints and face images from publicly available sources.
由于生物特征数据不容易被替换或撤销,因此在生物特征应用中使用的生物特征模板应该以安全的方式构建和存储,这样即使模板被破坏,攻击者也无法轻易伪造生物特征数据。这是一个具有挑战性的目标,因为生物特征数据本质上是“嘈杂的”,匹配算法通常是复杂的,这使得难以应用传统的加密技术,特别是当考虑多种模式时。在本文中,我们考虑一个“融合”minutiae-based指纹身份验证方案和一个SVD-based面临身份验证方案,并表明,采用最近提议加密原语称为“安全示意图”,和一个已知的几何变换在细节上,我们可以让它更容易结合不同的方法,同时使它来打造一个“原始”却是不可行的组合指纹和面部的形象通过身份验证。我们使用公开来源的真实指纹和人脸图像来评估我们方案的有效性。
{"title":"Secure Biometric Templates from Fingerprint-Face Features","authors":"Y. Sutcu, Qiming Li, N. Memon","doi":"10.1109/CVPR.2007.383385","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383385","url":null,"abstract":"Since biometric data cannot be easily replaced or revoked, it is important that biometric templates used in biometric applications should be constructed and stored in a secure way, such that attackers would not be able to forge biometric data easily even when the templates are compromised. This is a challenging goal since biometric data are \"noisy\" by nature, and the matching algorithms are often complex, which make it difficult to apply traditional cryptographic techniques, especially when multiple modalities are considered. In this paper, we consider a \"fusion \" of a minutiae-based fingerprint authentication scheme and an SVD-based face authentication scheme, and show that by employing a recently proposed cryptographic primitive called \"secure sketch \", and a known geometric transformation on minutiae, we can make it easier to combine different modalities, and at the same time make it computationally infeasible to forge an \"original\" combination of fingerprint and face image that passes the authentication. We evaluate the effectiveness of our scheme using real fingerprints and face images from publicly available sources.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126163241","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 129
Learning Color Names from Real-World Images 从真实世界的图像中学习颜色名称
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383218
Joost van de Weijer, C. Schmid, J. Verbeek
Within a computer vision context color naming is the action of assigning linguistic color labels to image pixels. In general, research on color naming applies the following paradigm: a collection of color chips is labelled with color names within a well-defined experimental setup by multiple test subjects. The collected data set is subsequently used to label RGB values in real-world images with a color name. Apart from the fact that this collection process is time consuming, it is unclear to what extent color naming within a controlled setup is representative for color naming in real-world images. Therefore we propose to learn color names from real-world images. Furthermore, we avoid test subjects by using Google Image to collect a data set. Due to limitations of Google Image this data set contains a substantial quantity of wrongly labelled data. The color names are learned using a PLSA model adapted to this task. Experimental results show that color names learned from real-world images significantly outperform color names learned from labelled color chips on retrieval and classification.
在计算机视觉上下文中,颜色命名是为图像像素分配语言颜色标签的动作。一般来说,对颜色命名的研究采用以下范式:在一个定义良好的实验设置中,由多个测试对象标记一组颜色芯片的颜色名称。收集的数据集随后用于在真实图像中标记具有颜色名称的RGB值。除了这个收集过程很耗时这一事实外,还不清楚受控设置中的颜色命名在多大程度上代表了真实图像中的颜色命名。因此,我们建议从现实世界的图像中学习颜色名称。此外,我们通过使用Google Image收集数据集来避免测试对象。由于谷歌图像的限制,该数据集包含大量错误标记的数据。颜色名称是使用适用于此任务的PLSA模型学习的。实验结果表明,从真实图像中学习到的颜色名称在检索和分类上明显优于从标记的颜色芯片中学习到的颜色名称。
{"title":"Learning Color Names from Real-World Images","authors":"Joost van de Weijer, C. Schmid, J. Verbeek","doi":"10.1109/CVPR.2007.383218","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383218","url":null,"abstract":"Within a computer vision context color naming is the action of assigning linguistic color labels to image pixels. In general, research on color naming applies the following paradigm: a collection of color chips is labelled with color names within a well-defined experimental setup by multiple test subjects. The collected data set is subsequently used to label RGB values in real-world images with a color name. Apart from the fact that this collection process is time consuming, it is unclear to what extent color naming within a controlled setup is representative for color naming in real-world images. Therefore we propose to learn color names from real-world images. Furthermore, we avoid test subjects by using Google Image to collect a data set. Due to limitations of Google Image this data set contains a substantial quantity of wrongly labelled data. The color names are learned using a PLSA model adapted to this task. Experimental results show that color names learned from real-world images significantly outperform color names learned from labelled color chips on retrieval and classification.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125295554","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 199
High-Speed Measurement of BRDF using an Ellipsoidal Mirror and a Projector 利用椭球镜和投影仪高速测量BRDF
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383467
Y. Mukaigawa, K. Sumino, Y. Yagi
Measuring BRDF (bi-directional reflectance distribution function) requires huge amounts of time because a target object must be illuminated from all incident angles and the reflected lights must be measured from all reflected angles. In this paper, we present a high-speed method to measure BRDFs using an ellipsoidal mirror and a projector. Our method makes it possible to change incident angles without a mechanical drive. Moreover, the omni-directional reflected lights from the object can be measured by one static camera at once. Our prototype requires only fifty minutes to measure anisotropic BRDFs, even if the lighting interval is one degree.
测量BRDF(双向反射分布函数)需要大量的时间,因为目标物体必须从所有入射角照射,并且必须从所有反射角测量反射光。在本文中,我们提出了一种使用椭球镜和投影仪来高速测量brdf的方法。我们的方法可以在没有机械驱动的情况下改变入射角。此外,物体的全向反射光可以用一台静态相机同时测量。我们的原型只需要50分钟来测量各向异性brdf,即使照明间隔是一度。
{"title":"High-Speed Measurement of BRDF using an Ellipsoidal Mirror and a Projector","authors":"Y. Mukaigawa, K. Sumino, Y. Yagi","doi":"10.1109/CVPR.2007.383467","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383467","url":null,"abstract":"Measuring BRDF (bi-directional reflectance distribution function) requires huge amounts of time because a target object must be illuminated from all incident angles and the reflected lights must be measured from all reflected angles. In this paper, we present a high-speed method to measure BRDFs using an ellipsoidal mirror and a projector. Our method makes it possible to change incident angles without a mechanical drive. Moreover, the omni-directional reflected lights from the object can be measured by one static camera at once. Our prototype requires only fifty minutes to measure anisotropic BRDFs, even if the lighting interval is one degree.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"137 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126621132","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 38
A Nine-point Algorithm for Estimating Para-Catadioptric Fundamental Matrices 拟反射性基本矩阵的九点估计算法
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383065
Christopher Geyer, Henrik Stewénius
We present a minimal-point algorithm for finding fundamental matrices for catadioptric cameras of the parabolic type. Central catadioptric cameras-an optical combination of a mirror and a lens that yields an imaging device equivalent within hemispheres to perspective cameras-have found wide application in robotics, tele-immersion and providing enhanced situational awareness for remote operation. We use an uncalibrated structure-from-motion framework developed for these cameras to consider the problem of estimating the fundamental matrix for such cameras. We present a solution that can compute the para-catadioptirc fundamental matrix with nine point correspondences, the smallest number possible. We compare this algorithm to alternatives and show some results of using the algorithm in conjunction with random sample consensus (RANSAC).
提出了一种求抛物型反射相机基本矩阵的最小点算法。中央反射式相机——一种由镜子和镜头组成的光学组合,在半球内产生相当于透视相机的成像设备——在机器人、远程沉浸和为远程操作提供增强的态势感知方面得到了广泛应用。我们使用为这些相机开发的非校准运动结构框架来考虑估计这些相机的基本矩阵的问题。我们给出了一种解,它可以计算具有最小可能的9点对应的准抛射基本矩阵。我们将该算法与替代算法进行比较,并展示了将该算法与随机样本共识(RANSAC)结合使用的一些结果。
{"title":"A Nine-point Algorithm for Estimating Para-Catadioptric Fundamental Matrices","authors":"Christopher Geyer, Henrik Stewénius","doi":"10.1109/CVPR.2007.383065","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383065","url":null,"abstract":"We present a minimal-point algorithm for finding fundamental matrices for catadioptric cameras of the parabolic type. Central catadioptric cameras-an optical combination of a mirror and a lens that yields an imaging device equivalent within hemispheres to perspective cameras-have found wide application in robotics, tele-immersion and providing enhanced situational awareness for remote operation. We use an uncalibrated structure-from-motion framework developed for these cameras to consider the problem of estimating the fundamental matrix for such cameras. We present a solution that can compute the para-catadioptirc fundamental matrix with nine point correspondences, the smallest number possible. We compare this algorithm to alternatives and show some results of using the algorithm in conjunction with random sample consensus (RANSAC).","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"55 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126627898","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 28
Partially Occluded Object-Specific Segmentation in View-Based Recognition 基于视图的识别中部分遮挡的特定对象分割
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383268
Minsu Cho, Kyoung Mu Lee
We present a novel object-specific segmentation method which can be used in view-based object recognition systems. Previous object segmentation approaches generate inexact results especially in partially occluded and cluttered environment because their top-down strategies fail to explain the details of various specific objects. On the contrary, our segmentation method efficiently exploits the information of the matched model views in view-based recognition because the aligned model view to the input image can serve as the best top-down cue for object segmentation. In this paper, we cast the problem of partially occluded object segmentation as that of labelling displacement and foreground status simultaneously for each pixel between the aligned model view and an input image. The problem is formulated by a maximum a posteriori Markov random field (MAP-MRF) model which minimizes a particular energy function. Our method overcomes complex occlusion and clutter and provides accurate segmentation boundaries by combining a bottom-up segmentation cue together. We demonstrate the efficiency and robustness of it by experimental results on various objects under occluded and cluttered environments.
提出了一种新的目标分割方法,可用于基于视图的目标识别系统。以往的目标分割方法由于其自上而下的策略无法解释各种特定对象的细节,导致结果不精确,特别是在部分遮挡和混乱的环境中。相反,我们的分割方法有效地利用了基于视图的识别中匹配模型视图的信息,因为与输入图像对齐的模型视图可以作为最佳的自上而下的对象分割线索。在本文中,我们将部分遮挡的目标分割问题视为对齐模型视图和输入图像之间的每个像素同时标记位移和前景状态的问题。该问题是由一个极大后验马尔可夫随机场(MAP-MRF)模型来表达的,该模型最小化了一个特定的能量函数。该方法克服了复杂的遮挡和杂波,结合自下而上的分割线索提供了准确的分割边界。通过对遮挡和杂乱环境下各种目标的实验结果,验证了该方法的有效性和鲁棒性。
{"title":"Partially Occluded Object-Specific Segmentation in View-Based Recognition","authors":"Minsu Cho, Kyoung Mu Lee","doi":"10.1109/CVPR.2007.383268","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383268","url":null,"abstract":"We present a novel object-specific segmentation method which can be used in view-based object recognition systems. Previous object segmentation approaches generate inexact results especially in partially occluded and cluttered environment because their top-down strategies fail to explain the details of various specific objects. On the contrary, our segmentation method efficiently exploits the information of the matched model views in view-based recognition because the aligned model view to the input image can serve as the best top-down cue for object segmentation. In this paper, we cast the problem of partially occluded object segmentation as that of labelling displacement and foreground status simultaneously for each pixel between the aligned model view and an input image. The problem is formulated by a maximum a posteriori Markov random field (MAP-MRF) model which minimizes a particular energy function. Our method overcomes complex occlusion and clutter and provides accurate segmentation boundaries by combining a bottom-up segmentation cue together. We demonstrate the efficiency and robustness of it by experimental results on various objects under occluded and cluttered environments.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126759863","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
A Robust Warping Method for Fingerprint Matching 一种用于指纹匹配的鲁棒翘曲方法
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383391
Dongjin Kwon, I. Yun, Sang Uk Lee
This paper presents a robust warping method for minutiae based fingerprint matching approaches. In this method, a deformable fingerprint surface is described using a triangular mesh model. For given two extracted minutiae sets and their correspondences, the proposed method constructs an energy function using a robust correspondence energy estimator and smoothness measuring of the mesh model. We obtain a convergent deformation pattern using an efficient gradient based energy optimization method. This energy optimization approach deals successfully with deformation errors caused by outliers, which are more difficult problems for the thin-plate spline (TPS) model. The proposed method is fast and the run-time performance is comparable with the method based on the TPS model. In the experiments, we provide a visual inspection of warping results on given correspondences and quantitative results using database.
提出了一种基于细节特征的指纹匹配鲁棒翘曲方法。该方法采用三角形网格模型来描述可变形的指纹表面。对于给定的两个提取的细节集及其对应关系,该方法利用鲁棒对应能量估计和网格模型的平滑性度量来构造能量函数。我们利用一种有效的基于梯度的能量优化方法得到了一个收敛的变形模式。该能量优化方法成功地解决了薄板样条(TPS)模型中较难解决的异常值引起的变形误差问题。该方法速度快,运行时性能与基于TPS模型的方法相当。在实验中,我们提供了对给定对应的翘曲结果的视觉检查和使用数据库的定量结果。
{"title":"A Robust Warping Method for Fingerprint Matching","authors":"Dongjin Kwon, I. Yun, Sang Uk Lee","doi":"10.1109/CVPR.2007.383391","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383391","url":null,"abstract":"This paper presents a robust warping method for minutiae based fingerprint matching approaches. In this method, a deformable fingerprint surface is described using a triangular mesh model. For given two extracted minutiae sets and their correspondences, the proposed method constructs an energy function using a robust correspondence energy estimator and smoothness measuring of the mesh model. We obtain a convergent deformation pattern using an efficient gradient based energy optimization method. This energy optimization approach deals successfully with deformation errors caused by outliers, which are more difficult problems for the thin-plate spline (TPS) model. The proposed method is fast and the run-time performance is comparable with the method based on the TPS model. In the experiments, we provide a visual inspection of warping results on given correspondences and quantitative results using database.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115031432","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Scaled Motion Dynamics for Markerless Motion Capture 缩放运动动力学无标记运动捕捉
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383128
B. Rosenhahn, T. Brox, H. Seidel
This work proposes a way to use a-priori knowledge on motion dynamics for markerless human motion capture (MoCap). Specifically, we match tracked motion patterns to training patterns in order to predict states in successive frames. Thereby, modeling the motion by means of twists allows for a proper scaling of the prior. Consequently, there is no need for training data of different frame rates or velocities. Moreover, the method allows to combine very different motion patterns. Experiments in indoor and outdoor scenarios demonstrate the continuous tracking of familiar motion patterns in case of artificial frame drops or in situations insufficiently constrained by the image data.
这项工作提出了一种使用运动动力学先验知识的方法,用于无标记的人类运动捕捉(MoCap)。具体来说,我们将跟踪的运动模式与训练模式相匹配,以预测连续帧中的状态。因此,通过扭曲对运动进行建模,可以适当地缩放先验。因此,不需要训练不同帧率或速度的数据。此外,该方法允许组合非常不同的运动模式。室内和室外场景的实验表明,在人工帧下降或图像数据约束不足的情况下,可以持续跟踪熟悉的运动模式。
{"title":"Scaled Motion Dynamics for Markerless Motion Capture","authors":"B. Rosenhahn, T. Brox, H. Seidel","doi":"10.1109/CVPR.2007.383128","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383128","url":null,"abstract":"This work proposes a way to use a-priori knowledge on motion dynamics for markerless human motion capture (MoCap). Specifically, we match tracked motion patterns to training patterns in order to predict states in successive frames. Thereby, modeling the motion by means of twists allows for a proper scaling of the prior. Consequently, there is no need for training data of different frame rates or velocities. Moreover, the method allows to combine very different motion patterns. Experiments in indoor and outdoor scenarios demonstrate the continuous tracking of familiar motion patterns in case of artificial frame drops or in situations insufficiently constrained by the image data.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115053142","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 63
Bilattice-based Logical Reasoning for Human Detection 基于双栅格的人类检测逻辑推理
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383133
V. Shet, J. Neumann, Visvanathan Ramesh, L. Davis
The capacity to robustly detect humans in video is a critical component of automated visual surveillance systems. This paper describes a bilattice based logical reasoning approach that exploits contextual information and knowledge about interactions between humans, and augments it with the output of different low level detectors for human detection. Detections from low level parts-based detectors are treated as logical facts and used to reason explicitly about the presence or absence of humans in the scene. Positive and negative information from different sources, as well as uncertainties from detections and logical rules, are integrated within the bilattice framework. This approach also generates proofs or justifications for each hypothesis it proposes. These justifications (or lack thereof) are further employed by the system to explain and validate, or reject potential hypotheses. This allows the system to explicitly reason about complex interactions between humans and handle occlusions. These proofs are also available to the end user as an explanation of why the system thinks a particular hypothesis is actually a human. We employ a boosted cascade of gradient histograms based detector to detect individual body parts. We have applied this framework to analyze the presence of humans in static images from different datasets.
鲁棒检测视频中的人的能力是自动视觉监控系统的关键组成部分。本文描述了一种基于双边格的逻辑推理方法,该方法利用人类之间相互作用的上下文信息和知识,并通过不同低水平检测器的输出来增强它,用于人类检测。来自低级部件检测器的检测被视为逻辑事实,并用于明确地推断场景中是否存在人类。来自不同来源的正面和负面信息,以及来自检测和逻辑规则的不确定性,被整合在双边框架内。这种方法也为它提出的每个假设产生证明或证明。这些证明(或缺乏证明)被系统进一步用来解释和验证,或拒绝潜在的假设。这使得系统可以明确地推理人类之间复杂的相互作用并处理闭塞。最终用户也可以使用这些证明来解释为什么系统认为特定的假设实际上是人类。我们采用增强级联梯度直方图为基础的检测器来检测单个身体部位。我们已经应用这个框架来分析来自不同数据集的静态图像中人类的存在。
{"title":"Bilattice-based Logical Reasoning for Human Detection","authors":"V. Shet, J. Neumann, Visvanathan Ramesh, L. Davis","doi":"10.1109/CVPR.2007.383133","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383133","url":null,"abstract":"The capacity to robustly detect humans in video is a critical component of automated visual surveillance systems. This paper describes a bilattice based logical reasoning approach that exploits contextual information and knowledge about interactions between humans, and augments it with the output of different low level detectors for human detection. Detections from low level parts-based detectors are treated as logical facts and used to reason explicitly about the presence or absence of humans in the scene. Positive and negative information from different sources, as well as uncertainties from detections and logical rules, are integrated within the bilattice framework. This approach also generates proofs or justifications for each hypothesis it proposes. These justifications (or lack thereof) are further employed by the system to explain and validate, or reject potential hypotheses. This allows the system to explicitly reason about complex interactions between humans and handle occlusions. These proofs are also available to the end user as an explanation of why the system thinks a particular hypothesis is actually a human. We employ a boosted cascade of gradient histograms based detector to detect individual body parts. We have applied this framework to analyze the presence of humans in static images from different datasets.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"155 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116393667","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 124
Learning the Compositional Nature of Visual Objects 学习视觉对象的组成性质
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383154
B. Ommer, J. Buhmann
The compositional nature of visual objects significantly limits their representation complexity and renders learning of structured object models tractable. Adopting this modeling strategy we both (i) automatically decompose objects into a hierarchy of relevant compositions and we (ii) learn such a compositional representation for each category without supervision. The compositional structure supports feature sharing already on the lowest level of small image patches. Compositions are represented as probability distributions over their constituent parts and the relations between them. The global shape of objects is captured by a graphical model which combines all compositions. Inference based on the underlying statistical model is then employed to obtain a category level object recognition system. Experiments on large standard benchmark datasets underline the competitive recognition performance of this approach and they provide insights into the learned compositional structure of objects.
视觉对象的组成特性极大地限制了其表示的复杂性,并使结构化对象模型的学习变得易于处理。采用这种建模策略,我们(i)自动将对象分解为相关组合的层次结构,(ii)在没有监督的情况下为每个类别学习这样的组合表示。该组合结构支持已经在最低级别的小图像补丁上的特征共享。组合用其组成部分及其之间关系的概率分布来表示。物体的整体形状由图形模型捕获,图形模型结合了所有组成部分。然后利用基于底层统计模型的推理得到类别级目标识别系统。在大型标准基准数据集上的实验强调了该方法的竞争性识别性能,并提供了对学习到的对象组成结构的见解。
{"title":"Learning the Compositional Nature of Visual Objects","authors":"B. Ommer, J. Buhmann","doi":"10.1109/CVPR.2007.383154","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383154","url":null,"abstract":"The compositional nature of visual objects significantly limits their representation complexity and renders learning of structured object models tractable. Adopting this modeling strategy we both (i) automatically decompose objects into a hierarchy of relevant compositions and we (ii) learn such a compositional representation for each category without supervision. The compositional structure supports feature sharing already on the lowest level of small image patches. Compositions are represented as probability distributions over their constituent parts and the relations between them. The global shape of objects is captured by a graphical model which combines all compositions. Inference based on the underlying statistical model is then employed to obtain a category level object recognition system. Experiments on large standard benchmark datasets underline the competitive recognition performance of this approach and they provide insights into the learned compositional structure of objects.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"27 7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116408361","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 60
The Hierarchical Isometric Self-Organizing Map for Manifold Representation 流形表示的层次等距自组织映射
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383402
Haiying Guan, M. Turk
We present an algorithm, Hierarchical ISOmetric Self-Organizing Map (H-ISOSOM), for a concise, organized manifold representation of complex, non-linear, large scale, high-dimensional input data in a low dimensional space. The main contribution of our algorithm is threefold. First, we modify the previous ISOSOM algorithm by a local linear interpolation (LLl) technique, which maps the data samples from low dimensional space back to high dimensional space and makes the complete mapping pseudo-invertible. The modified-ISOSOM (M-ISOSOM) follows the global geometric structure of the data, and also preserves local geometric relations to reduce the nonlinear mapping distortion and make the learning more accurate. Second, we propose the H-ISOSOM algorithm for the computational complexity problem of Isomap, SOM and LLI and the nonlinear complexity problem of the highly twisted manifold. H-ISOSOM learns an organized structure of a non-convex, large scale manifold and represents it by a set of hierarchical organized maps. The hierarchical structure follows a coarse-to-fine strategy. According to the coarse global structure, it "unfolds " the manifold at the coarse level and decomposes the sample data into small patches, then iteratively learns the nonlinearity of each patch in finer levels. The algorithm simultaneously reorganizes and clusters the data samples in a low dimensional space to obtain the concise representation. Third, we give quantitative comparisons of the proposed method with similar methods on standard data sets. Finally, we apply H-ISOSOM to the problem of appearance-based hand pose estimation. Encouraging experimental results validate the effectiveness and efficiency of H-ISOSOM.
我们提出了一种算法,层次等长自组织映射(H-ISOSOM),用于在低维空间中对复杂、非线性、大规模、高维输入数据进行简洁、有组织的流形表示。我们的算法的主要贡献有三个方面。首先,采用局部线性插值(LLl)技术对ISOSOM算法进行改进,将数据样本从低维空间映射回高维空间,使完全映射伪可逆。改进的isosom (M-ISOSOM)在遵循数据全局几何结构的同时,也保留了局部几何关系,减少了非线性映射失真,使学习更加准确。其次,针对Isomap、SOM和LLI的计算复杂性问题以及高扭曲流形的非线性复杂性问题,提出了H-ISOSOM算法。H-ISOSOM学习非凸、大规模流形的组织结构,并通过一组分层组织映射来表示它。层次结构遵循从粗到精的策略。它根据粗糙的全局结构,在粗糙的层次上“展开”流形,将样本数据分解成小块,然后在更精细的层次上迭代学习每个小块的非线性。该算法同时在低维空间对数据样本进行重组和聚类,以获得简洁的表示。第三,我们将所提出的方法与标准数据集上的类似方法进行了定量比较。最后,我们将H-ISOSOM应用于基于外观的手部姿态估计问题。令人鼓舞的实验结果验证了H-ISOSOM的有效性和效率。
{"title":"The Hierarchical Isometric Self-Organizing Map for Manifold Representation","authors":"Haiying Guan, M. Turk","doi":"10.1109/CVPR.2007.383402","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383402","url":null,"abstract":"We present an algorithm, Hierarchical ISOmetric Self-Organizing Map (H-ISOSOM), for a concise, organized manifold representation of complex, non-linear, large scale, high-dimensional input data in a low dimensional space. The main contribution of our algorithm is threefold. First, we modify the previous ISOSOM algorithm by a local linear interpolation (LLl) technique, which maps the data samples from low dimensional space back to high dimensional space and makes the complete mapping pseudo-invertible. The modified-ISOSOM (M-ISOSOM) follows the global geometric structure of the data, and also preserves local geometric relations to reduce the nonlinear mapping distortion and make the learning more accurate. Second, we propose the H-ISOSOM algorithm for the computational complexity problem of Isomap, SOM and LLI and the nonlinear complexity problem of the highly twisted manifold. H-ISOSOM learns an organized structure of a non-convex, large scale manifold and represents it by a set of hierarchical organized maps. The hierarchical structure follows a coarse-to-fine strategy. According to the coarse global structure, it \"unfolds \" the manifold at the coarse level and decomposes the sample data into small patches, then iteratively learns the nonlinearity of each patch in finer levels. The algorithm simultaneously reorganizes and clusters the data samples in a low dimensional space to obtain the concise representation. Third, we give quantitative comparisons of the proposed method with similar methods on standard data sets. Finally, we apply H-ISOSOM to the problem of appearance-based hand pose estimation. Encouraging experimental results validate the effectiveness and efficiency of H-ISOSOM.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121881890","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
期刊
2007 IEEE Conference on Computer Vision and Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1