首页 > 最新文献

2007 IEEE Conference on Computer Vision and Pattern Recognition最新文献

英文 中文
3D Probabilistic Feature Point Model for Object Detection and Recognition 用于目标检测与识别的三维概率特征点模型
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383284
S. Romdhani, T. Vetter
This paper presents a novel statistical shape model that can be used to detect and localise feature points of a class of objects in images. The shape model is inspired from the 3D morphable model (3DMM) and has the property to be viewpoint invariant. This shape model is used to estimate the probability of the position of a feature point given the position of reference feature points, accounting for the uncertainty of the position of the reference points and of the intrinsic variability of the class of objects. The viewpoint invariant detection algorithm maximises a foreground/background likelihood ratio of the relative position of the feature points, their appearance, scale, orientation and occlusion state. Computational efficiency is obtained by using the Bellman principle and an early rejection rule based on 3D to 2D projection constraints. Evaluations of the detection algorithm on the CMU-P1E face images and on a large set of non-face images show high levels of accuracy (zero false alarms for more than 90% detection rate). As well as locating feature points, the detection algorithm also estimates the pose of the object and a few shape parameters. It is shown that it can be used to initialise a 3DMM fitting algorithm and thus enables a fully automatic viewpoint and lighting invariant image analysis solution.
本文提出了一种新的统计形状模型,用于检测和定位图像中一类物体的特征点。形状模型受三维变形模型(3DMM)的启发,具有视点不变的特性。该形状模型用于在给定参考特征点位置的情况下估计特征点位置的概率,考虑了参考点位置的不确定性和物体类别的内在可变性。视点不变检测算法最大限度地提高了特征点的相对位置、外观、规模、方向和遮挡状态的前景/背景似然比。利用Bellman原理和基于三维到二维投影约束的早期拒绝规则来提高计算效率。对CMU-P1E人脸图像和大量非人脸图像的检测算法的评估显示出高水平的准确性(超过90%的检测率为零误报)。在定位特征点的同时,该检测算法还对目标的姿态和一些形状参数进行估计。结果表明,该算法可用于初始化3DMM拟合算法,从而实现全自动视点和光照不变图像分析解决方案。
{"title":"3D Probabilistic Feature Point Model for Object Detection and Recognition","authors":"S. Romdhani, T. Vetter","doi":"10.1109/CVPR.2007.383284","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383284","url":null,"abstract":"This paper presents a novel statistical shape model that can be used to detect and localise feature points of a class of objects in images. The shape model is inspired from the 3D morphable model (3DMM) and has the property to be viewpoint invariant. This shape model is used to estimate the probability of the position of a feature point given the position of reference feature points, accounting for the uncertainty of the position of the reference points and of the intrinsic variability of the class of objects. The viewpoint invariant detection algorithm maximises a foreground/background likelihood ratio of the relative position of the feature points, their appearance, scale, orientation and occlusion state. Computational efficiency is obtained by using the Bellman principle and an early rejection rule based on 3D to 2D projection constraints. Evaluations of the detection algorithm on the CMU-P1E face images and on a large set of non-face images show high levels of accuracy (zero false alarms for more than 90% detection rate). As well as locating feature points, the detection algorithm also estimates the pose of the object and a few shape parameters. It is shown that it can be used to initialise a 3DMM fitting algorithm and thus enables a fully automatic viewpoint and lighting invariant image analysis solution.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130841268","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 24
Iterative MAP and ML Estimations for Image Segmentation 图像分割的迭代MAP和ML估计
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383007
Shifeng Chen, Liangliang Cao, Jianzhuang Liu, Xiaoou Tang
Image segmentation plays an important role in computer vision and image analysis. In this paper, the segmentation problem is formulated as a labeling problem under a probability maximization framework. To estimate the label configuration, an iterative optimization scheme is proposed to alternately carry out the maximum a posteriori (MAP) estimation and the maximum-likelihood (ML) estimation. The MAP estimation problem is modeled with Markov random fields (MRFs). A graph-cut algorithm is used to find the solution to the MAP-MRF estimation. The ML estimation is achieved by finding the means of region features. Our algorithm can automatically segment an image into regions with relevant textures or colors without the need to know the number of regions in advance. In addition, under the same framework, it can be extended to another algorithm that extracts objects of a particular class from a group of images. Extensive experiments have shown the effectiveness of our approach.
图像分割在计算机视觉和图像分析中占有重要的地位。本文将分割问题表述为概率最大化框架下的标注问题。为了估计标签配置,提出了一种迭代优化方案,交替进行最大后验估计(MAP)和最大似然估计(ML)。利用马尔可夫随机场(mrf)对MAP估计问题进行建模。采用图切算法求解MAP-MRF估计问题。机器学习估计是通过寻找区域特征的均值来实现的。我们的算法可以自动将图像分割成具有相关纹理或颜色的区域,而无需事先知道区域的数量。此外,在相同的框架下,它还可以扩展为从一组图像中提取特定类对象的另一种算法。大量的实验证明了我们方法的有效性。
{"title":"Iterative MAP and ML Estimations for Image Segmentation","authors":"Shifeng Chen, Liangliang Cao, Jianzhuang Liu, Xiaoou Tang","doi":"10.1109/CVPR.2007.383007","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383007","url":null,"abstract":"Image segmentation plays an important role in computer vision and image analysis. In this paper, the segmentation problem is formulated as a labeling problem under a probability maximization framework. To estimate the label configuration, an iterative optimization scheme is proposed to alternately carry out the maximum a posteriori (MAP) estimation and the maximum-likelihood (ML) estimation. The MAP estimation problem is modeled with Markov random fields (MRFs). A graph-cut algorithm is used to find the solution to the MAP-MRF estimation. The ML estimation is achieved by finding the means of region features. Our algorithm can automatically segment an image into regions with relevant textures or colors without the need to know the number of regions in advance. In addition, under the same framework, it can be extended to another algorithm that extracts objects of a particular class from a group of images. Extensive experiments have shown the effectiveness of our approach.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"199 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130418175","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Thermal-Visible Video Fusion for Moving Target Tracking and Pedestrian Classification 热视视频融合运动目标跟踪与行人分类
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383444
A. Leykin, Yang Ran, R. Hammoud
The paper presents a fusion-tracker and pedestrian classifier for color and thermal cameras. The tracker builds a background model as a multi-modal distribution of colors and temperatures. It is constructed as a particle filter that makes a number of informed reversible transformations to sample the model probability space in order to maximize posterior probability of the scene model. Observation likelihoods of moving objects account their 3D locations with respect to the camera and occlusions by other tracked objects as well as static obstacles. After capturing the coordinates and dimensions of moving objects we apply a pedestrian classifier based on periodic gait analysis. To separate humans from other moving objects, such as cars, we detect, in human gait, a symmetrical double helical pattern, that can then be analyzed using the Frieze Group theory. The results of tracking on color and thermal sequences demonstrate that our algorithm is robust to illumination noise and performs well in the outdoor environments.
提出了一种适用于彩色相机和热像仪的融合跟踪器和行人分类器。跟踪器将背景模型构建为颜色和温度的多模态分布。它被构造为一个粒子滤波器,通过多次知情的可逆变换对模型概率空间进行采样,以最大化场景模型的后验概率。移动物体的观察可能性考虑了它们相对于相机和其他跟踪物体以及静态障碍物的遮挡的3D位置。在捕获运动物体的坐标和尺寸后,应用基于周期步态分析的行人分类器。为了将人类与其他运动物体(如汽车)区分开来,我们在人类的步态中检测到一种对称的双螺旋模式,然后可以使用Frieze群论对其进行分析。对颜色序列和热序列的跟踪结果表明,该算法对光照噪声具有较强的鲁棒性,在室外环境下具有良好的性能。
{"title":"Thermal-Visible Video Fusion for Moving Target Tracking and Pedestrian Classification","authors":"A. Leykin, Yang Ran, R. Hammoud","doi":"10.1109/CVPR.2007.383444","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383444","url":null,"abstract":"The paper presents a fusion-tracker and pedestrian classifier for color and thermal cameras. The tracker builds a background model as a multi-modal distribution of colors and temperatures. It is constructed as a particle filter that makes a number of informed reversible transformations to sample the model probability space in order to maximize posterior probability of the scene model. Observation likelihoods of moving objects account their 3D locations with respect to the camera and occlusions by other tracked objects as well as static obstacles. After capturing the coordinates and dimensions of moving objects we apply a pedestrian classifier based on periodic gait analysis. To separate humans from other moving objects, such as cars, we detect, in human gait, a symmetrical double helical pattern, that can then be analyzed using the Frieze Group theory. The results of tracking on color and thermal sequences demonstrate that our algorithm is robust to illumination noise and performs well in the outdoor environments.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130452970","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 69
Probabilistic visibility for multi-view stereo 多视点立体的概率可见性
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383193
Carlos Hernández, George Vogiatzis, R. Cipolla
We present a new formulation to multi-view stereo that treats the problem as probabilistic 3D segmentation. Previous work has used the stereo photo-consistency criterion as a detector of the boundary between the 3D scene and the surrounding empty space. Here we show how the same criterion can also provide a foreground/background model that can predict if a 3D location is inside or outside the scene. This model replaces the commonly used naive foreground model based on ballooning which is known to perform poorly in concavities. We demonstrate how the probabilistic visibility is linked to previous work on depth-map fusion and we present a multi-resolution graph-cut implementation using the new ballooning term that is very efficient both in terms of computation time and memory requirements.
提出了一种新的多视点立体图像分割方法,将多视点立体图像分割问题视为概率三维分割问题。以前的工作使用立体照片一致性准则作为3D场景和周围空白空间之间边界的检测器。在这里,我们展示了相同的标准如何提供前景/背景模型,可以预测3D位置是在场景内还是在场景外。该模型取代了常用的基于气球的朴素前景模型,该模型在凹陷中表现不佳。我们演示了概率可见性如何与深度图融合的先前工作相关联,并使用新的气球术语提出了一个多分辨率图切割实现,该实现在计算时间和内存需求方面都非常高效。
{"title":"Probabilistic visibility for multi-view stereo","authors":"Carlos Hernández, George Vogiatzis, R. Cipolla","doi":"10.1109/CVPR.2007.383193","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383193","url":null,"abstract":"We present a new formulation to multi-view stereo that treats the problem as probabilistic 3D segmentation. Previous work has used the stereo photo-consistency criterion as a detector of the boundary between the 3D scene and the surrounding empty space. Here we show how the same criterion can also provide a foreground/background model that can predict if a 3D location is inside or outside the scene. This model replaces the commonly used naive foreground model based on ballooning which is known to perform poorly in concavities. We demonstrate how the probabilistic visibility is linked to previous work on depth-map fusion and we present a multi-resolution graph-cut implementation using the new ballooning term that is very efficient both in terms of computation time and memory requirements.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"141 4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129175311","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 117
Algorithms for Batch Matrix Factorization with Application to Structure-from-Motion 批矩阵分解算法及其在运动构造中的应用
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383062
J. Tardif, A. Bartoli, Martin Trudeau, Nicolas Guilbert, S. Roy
Matrix factorization is a key component for solving several computer vision problems. It is particularly challenging in the presence of missing or erroneous data, which often arise in structure-from-motion. We propose batch algorithms for matrix factorization. They are based on closure and basis constraints, that are used either on the cameras or the structure, leading to four possible algorithms. The constraints are robustly computed from complete measurement sub-matrices with e.g. random data sampling. The cameras and 3D structure are then recovered through linear least squares. Prior information about the scene such as identical camera positions or orientations, smooth camera trajectory, known 3D points and coplanarity of some 3D points can be directly incorporated. We demonstrate our algorithms on challenging image sequences with tracking error and more than 95% missing data.
矩阵分解是解决许多计算机视觉问题的关键组成部分。在运动结构中经常出现数据缺失或错误的情况下,这尤其具有挑战性。我们提出了矩阵分解的批处理算法。它们基于闭包和基约束,用于摄像机或结构,导致四种可能的算法。约束由完整测量子矩阵鲁棒计算,例如随机数据采样。然后通过线性最小二乘法恢复相机和三维结构。可以直接结合场景的先验信息,如相同的相机位置或方向,平滑的相机轨迹,已知的3D点和一些3D点的共平面性。我们在具有跟踪误差和超过95%缺失数据的具有挑战性的图像序列上演示了我们的算法。
{"title":"Algorithms for Batch Matrix Factorization with Application to Structure-from-Motion","authors":"J. Tardif, A. Bartoli, Martin Trudeau, Nicolas Guilbert, S. Roy","doi":"10.1109/CVPR.2007.383062","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383062","url":null,"abstract":"Matrix factorization is a key component for solving several computer vision problems. It is particularly challenging in the presence of missing or erroneous data, which often arise in structure-from-motion. We propose batch algorithms for matrix factorization. They are based on closure and basis constraints, that are used either on the cameras or the structure, leading to four possible algorithms. The constraints are robustly computed from complete measurement sub-matrices with e.g. random data sampling. The cameras and 3D structure are then recovered through linear least squares. Prior information about the scene such as identical camera positions or orientations, smooth camera trajectory, known 3D points and coplanarity of some 3D points can be directly incorporated. We demonstrate our algorithms on challenging image sequences with tracking error and more than 95% missing data.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"30 3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123206874","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 44
Recognizing Human Activities from Silhouettes: Motion Subspace and Factorial Discriminative Graphical Model 从轮廓识别人类活动:运动子空间和析因判别图形模型
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383298
Liang Wang, D. Suter
We describe a probabilistic framework for recognizing human activities in monocular video based on simple silhouette observations in this paper. The methodology combines kernel principal component analysis (KPCA) based feature extraction and factorial conditional random field (FCRF) based motion modeling. Silhouette data is represented more compactly by nonlinear dimensionality reduction that explores the underlying structure of the articulated action space and preserves explicit temporal orders in projection trajectories of motions. FCRF models temporal sequences in multiple interacting ways, thus increasing joint accuracy by information sharing, with the ideal advantages of discriminative models over generative ones (e.g., relaxing independence assumption between observations and the ability to effectively incorporate both overlapping features and long-range dependencies). The experimental results on two recent datasets have shown that the proposed framework can not only accurately recognize human activities with temporal, intra-and inter-person variations, but also is considerably robust to noise and other factors such as partial occlusion and irregularities in motion styles.
本文描述了一种基于简单轮廓观察的单目视频中人类活动识别的概率框架。该方法结合了基于核主成分分析(KPCA)的特征提取和基于析因条件随机场(FCRF)的运动建模。轮廓数据通过非线性降维更紧凑地表示,该降维探索了关节动作空间的潜在结构,并保留了运动投影轨迹中的明确时间顺序。FCRF以多种交互方式对时间序列进行建模,从而通过信息共享提高联合精度,具有判别模型相对于生成模型的理想优势(例如,放宽观测值之间的独立性假设,能够有效地结合重叠特征和长期依赖关系)。在最近的两个数据集上的实验结果表明,所提出的框架不仅可以准确识别具有时间、内部和人与人之间变化的人类活动,而且对噪声和其他因素(如部分遮挡和运动风格的不规则性)具有相当的鲁棒性。
{"title":"Recognizing Human Activities from Silhouettes: Motion Subspace and Factorial Discriminative Graphical Model","authors":"Liang Wang, D. Suter","doi":"10.1109/CVPR.2007.383298","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383298","url":null,"abstract":"We describe a probabilistic framework for recognizing human activities in monocular video based on simple silhouette observations in this paper. The methodology combines kernel principal component analysis (KPCA) based feature extraction and factorial conditional random field (FCRF) based motion modeling. Silhouette data is represented more compactly by nonlinear dimensionality reduction that explores the underlying structure of the articulated action space and preserves explicit temporal orders in projection trajectories of motions. FCRF models temporal sequences in multiple interacting ways, thus increasing joint accuracy by information sharing, with the ideal advantages of discriminative models over generative ones (e.g., relaxing independence assumption between observations and the ability to effectively incorporate both overlapping features and long-range dependencies). The experimental results on two recent datasets have shown that the proposed framework can not only accurately recognize human activities with temporal, intra-and inter-person variations, but also is considerably robust to noise and other factors such as partial occlusion and irregularities in motion styles.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123322736","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 230
On the Direct Estimation of the Fundamental Matrix 关于基本矩阵的直接估计
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383064
Yaser Sheikh, Asaad Hakeem, M. Shah
The fundamental matrix is a central construct in the analysis of images captured from a pair of cameras and many feature-based methods have been proposed for its computation. In this paper, we propose a direct method for estimating the fundamental matrix where the motion between the frames is small (e.g. between successive frames of a video). To achieve this, a warping function is presented for the fundamental matrix by using the brightness constancy constraint in conjunction with geometric constraints. Using this warping function, an iterative hierarchical algorithm is described to recover accurate estimates of the fundamental matrix. We present results of experimentation to evaluate the performance of the proposed approach and demonstrate improved accuracy in the computation of the fundamental matrix.
基本矩阵是对从一对相机捕获的图像进行分析的核心结构,许多基于特征的计算方法已经被提出。在本文中,我们提出了一种估计帧之间运动较小(例如视频的连续帧之间)的基本矩阵的直接方法。为了实现这一目标,通过将亮度常数约束与几何约束相结合,提出了基本矩阵的翘曲函数。利用这种扭曲函数,描述了一种迭代分层算法来恢复基本矩阵的准确估计。我们提出了实验结果,以评估所提出的方法的性能,并证明提高了基本矩阵计算的准确性。
{"title":"On the Direct Estimation of the Fundamental Matrix","authors":"Yaser Sheikh, Asaad Hakeem, M. Shah","doi":"10.1109/CVPR.2007.383064","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383064","url":null,"abstract":"The fundamental matrix is a central construct in the analysis of images captured from a pair of cameras and many feature-based methods have been proposed for its computation. In this paper, we propose a direct method for estimating the fundamental matrix where the motion between the frames is small (e.g. between successive frames of a video). To achieve this, a warping function is presented for the fundamental matrix by using the brightness constancy constraint in conjunction with geometric constraints. Using this warping function, an iterative hierarchical algorithm is described to recover accurate estimates of the fundamental matrix. We present results of experimentation to evaluate the performance of the proposed approach and demonstrate improved accuracy in the computation of the fundamental matrix.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123362971","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Using Group Prior to Identify People in Consumer Images 使用群体优先识别消费者形象中的人物
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383492
Andrew C. Gallagher, Tsuhan Chen
While face recognition techniques have rapidly advanced in the last few years, most of the work is in the domain of security applications. For consumer imaging applications, person recognition is an important tool that is useful for searching and retrieving images from a personal image collection. It has been shown that when recognizing a single person in an image, a maximum likelihood classifier requires the prior probability for each candidate individual. In this paper, we extend this idea and describe the benefits of using a group prior for identifying people in consumer images with multiple people. The group prior describes the probability of a group of individuals appearing together in an image. In our application, we have a subset of ambiguously labeled images for a consumer image collection, where we seek to identify all of the people in the collection. We describe a simple algorithm for resolving the ambiguous labels. We show that despite errors in resolving ambiguous labels, useful classifiers can be trained with the resolved labels. Recognition performance is further improved with a group prior learned from the ambiguous labels. In summary, by modeling the relationships between the people with the group prior, we improve classification performance.
虽然人脸识别技术在过去几年中得到了迅速发展,但大部分工作都是在安全领域的应用。对于消费者成像应用来说,人物识别是一个重要的工具,用于从个人图像集合中搜索和检索图像。研究表明,当识别图像中的单个人时,最大似然分类器需要每个候选个体的先验概率。在本文中,我们扩展了这一想法,并描述了使用群体先验来识别具有多个人的消费者图像中的人的好处。群体先验描述了一组个体在图像中一起出现的概率。在我们的应用程序中,我们有一个消费者图像集合的模糊标记图像子集,我们试图识别集合中的所有人。我们描述了一种解决歧义标签的简单算法。我们表明,尽管在解决歧义标签时存在错误,但可以使用解决的标签训练有用的分类器。利用从歧义标签中学习到的组先验进一步提高了识别性能。总之,通过对具有组先验的人之间的关系进行建模,我们提高了分类性能。
{"title":"Using Group Prior to Identify People in Consumer Images","authors":"Andrew C. Gallagher, Tsuhan Chen","doi":"10.1109/CVPR.2007.383492","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383492","url":null,"abstract":"While face recognition techniques have rapidly advanced in the last few years, most of the work is in the domain of security applications. For consumer imaging applications, person recognition is an important tool that is useful for searching and retrieving images from a personal image collection. It has been shown that when recognizing a single person in an image, a maximum likelihood classifier requires the prior probability for each candidate individual. In this paper, we extend this idea and describe the benefits of using a group prior for identifying people in consumer images with multiple people. The group prior describes the probability of a group of individuals appearing together in an image. In our application, we have a subset of ambiguously labeled images for a consumer image collection, where we seek to identify all of the people in the collection. We describe a simple algorithm for resolving the ambiguous labels. We show that despite errors in resolving ambiguous labels, useful classifiers can be trained with the resolved labels. Recognition performance is further improved with a group prior learned from the ambiguous labels. In summary, by modeling the relationships between the people with the group prior, we improve classification performance.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"76 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126212676","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 58
Sensor and Data Systems, Audio-Assisted Cameras and Acoustic Doppler Sensors 传感器和数据系统,声控相机和声学多普勒传感器
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383533
K. Kalgaonkar, P. Smaragdis, B. Raj
In this chapter we present two technologies for sensing and surveillance -audio-assisted cameras and acoustic Doppler sensors for gait recognition.
在本章中,我们介绍了两种用于传感和监视的技术——音频辅助摄像机和用于步态识别的声学多普勒传感器。
{"title":"Sensor and Data Systems, Audio-Assisted Cameras and Acoustic Doppler Sensors","authors":"K. Kalgaonkar, P. Smaragdis, B. Raj","doi":"10.1109/CVPR.2007.383533","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383533","url":null,"abstract":"In this chapter we present two technologies for sensing and surveillance -audio-assisted cameras and acoustic Doppler sensors for gait recognition.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126376650","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Real-time Automatic Deceit Detection from Involuntary Facial Expressions 基于非自愿面部表情的实时自动欺骗检测
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383383
Zhi Zhang, Vartika Singh, T. E. Slowe, S. Tulyakov, V. Govindaraju
Being the most broadly used tool for deceit measurement, the polygraph is a limited method as it suffers from human operator subjectivity and the fact that target subjects are aware of the measurement, which invites the opportunity to alter their behavior or plan counter-measures in advance. The approach presented in this paper attempts to circumvent these problems by unobtrusively and automatically measuring several prior identified deceit indicators (DIs) based upon involuntary, so-called reliable facial expressions through computer vision analysis of image sequences in real time. Reliable expressions are expressions said by the psychology community to be impossible for a significant percentage of the population to convincingly simulate, without feeling a true inner felt emotion. The strategy is to detect the difference between those expressions which arise from internal emotion, implying verity, and those expressions which are simulated, implying deceit. First, a group of facial action units (AUs) related to the reliable expressions are detected based on distance and texture based features. The DIs then can be measured and finally a decision of deceit or verity will be made accordingly. The performance of this proposed approach is evaluated by its real time implementation for deceit detection.
作为最广泛使用的欺骗测量工具,测谎仪是一种有限的方法,因为它受到操作员的主观性和目标对象知道测量的事实,这就有机会改变他们的行为或提前计划应对措施。本文提出的方法试图通过计算机视觉实时分析图像序列,通过非自愿的所谓可靠的面部表情,不显眼地自动测量几个预先识别的欺骗指标(DIs),从而规避这些问题。可靠的表情是指心理学界认为,如果没有真实的内心感受,很大一部分人不可能令人信服地模仿出的表情。这个策略就是要分辨出那些发自内心的、暗示真实的表情和那些模拟的、暗示欺骗的表情之间的区别。首先,基于距离和纹理特征检测与可靠表情相关的一组面部动作单元。然后可以测量DIs,并最终据此做出欺骗或真实的决定。通过对欺骗检测的实时实现来评估该方法的性能。
{"title":"Real-time Automatic Deceit Detection from Involuntary Facial Expressions","authors":"Zhi Zhang, Vartika Singh, T. E. Slowe, S. Tulyakov, V. Govindaraju","doi":"10.1109/CVPR.2007.383383","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383383","url":null,"abstract":"Being the most broadly used tool for deceit measurement, the polygraph is a limited method as it suffers from human operator subjectivity and the fact that target subjects are aware of the measurement, which invites the opportunity to alter their behavior or plan counter-measures in advance. The approach presented in this paper attempts to circumvent these problems by unobtrusively and automatically measuring several prior identified deceit indicators (DIs) based upon involuntary, so-called reliable facial expressions through computer vision analysis of image sequences in real time. Reliable expressions are expressions said by the psychology community to be impossible for a significant percentage of the population to convincingly simulate, without feeling a true inner felt emotion. The strategy is to detect the difference between those expressions which arise from internal emotion, implying verity, and those expressions which are simulated, implying deceit. First, a group of facial action units (AUs) related to the reliable expressions are detected based on distance and texture based features. The DIs then can be measured and finally a decision of deceit or verity will be made accordingly. The performance of this proposed approach is evaluated by its real time implementation for deceit detection.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126090416","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 38
期刊
2007 IEEE Conference on Computer Vision and Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1