首页 > 最新文献

2007 IEEE Conference on Advanced Video and Signal Based Surveillance最新文献

英文 中文
Model-based human posture estimation for gesture analysis in an opportunistic fusion smart camera network 机会融合智能摄像机网络中基于模型的人体姿态估计
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425353
Chen Wu, H. Aghajan
In multi-camera networks rich visual data is provided both spatially and temporally. In this paper a method of human posture estimation is described incorporating the concept of an opportunistic fusion framework aiming to employ manifold sources of visual information across space, time, and feature levels. One motivation for the proposed method is to reduce raw visual data in a single camera to elliptical parameterized segments for efficient communication between cameras. A 3D human body model is employed as the convergence point of spatiotemporal and feature fusion. It maintains both geometric parameters of the human posture and the adoptively learned appearance attributes, all of which are updated from the three dimensions of space, time and features of the opportunistic fusion. In sufficient confidence levels parameters of the 3D human body model are again used as feedback to aid subsequent in-node vision analysis. Color distribution registered in the model is used to initialize segmentation. Perceptually Organized Expectation Maximization (POEM) is then applied to refine color segments with observations from a single camera. Geometric configuration of the 3D skeleton is estimated by Particle Swarm Optimization (PSO).
在多摄像机网络中,提供了丰富的空间和时间视觉数据。本文描述了一种人体姿态估计方法,该方法结合了机会融合框架的概念,旨在利用跨越空间、时间和特征级别的多种视觉信息来源。该方法的一个动机是将单个摄像机中的原始视觉数据简化为椭圆参数化段,以便于摄像机之间的有效通信。采用三维人体模型作为时空和特征融合的收敛点。它既保留了人体姿势的几何参数,又保留了自适应学习的外观属性,这些属性都是从空间、时间和机会融合的特征三个维度更新的。在足够的置信水平下,三维人体模型的参数再次用作反馈,以帮助后续的节点内视觉分析。使用模型中注册的颜色分布初始化分割。然后应用感知组织期望最大化(POEM)来细化从单个相机观察到的颜色段。采用粒子群算法对三维骨架的几何构型进行估计。
{"title":"Model-based human posture estimation for gesture analysis in an opportunistic fusion smart camera network","authors":"Chen Wu, H. Aghajan","doi":"10.1109/AVSS.2007.4425353","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425353","url":null,"abstract":"In multi-camera networks rich visual data is provided both spatially and temporally. In this paper a method of human posture estimation is described incorporating the concept of an opportunistic fusion framework aiming to employ manifold sources of visual information across space, time, and feature levels. One motivation for the proposed method is to reduce raw visual data in a single camera to elliptical parameterized segments for efficient communication between cameras. A 3D human body model is employed as the convergence point of spatiotemporal and feature fusion. It maintains both geometric parameters of the human posture and the adoptively learned appearance attributes, all of which are updated from the three dimensions of space, time and features of the opportunistic fusion. In sufficient confidence levels parameters of the 3D human body model are again used as feedback to aid subsequent in-node vision analysis. Color distribution registered in the model is used to initialize segmentation. Perceptually Organized Expectation Maximization (POEM) is then applied to refine color segments with observations from a single camera. Geometric configuration of the 3D skeleton is estimated by Particle Swarm Optimization (PSO).","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"75 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121755431","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 38
Multitarget association and tracking in 3-D space based on particle filter with joint multitarget probability density 基于联合多目标概率密度粒子滤波的三维空间多目标关联与跟踪
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425374
Jinseok Lee, Byung Guk Kim, S. Cho, Sangjin Hong, W. Cho
This paper addresses the problem of 3-dimensional (3D) multitarget tracking using particle filter with the joint multitarget probability density (JMPD) technique. The estimation allows the nonlinear target motion with unlabeled measurement association as well as non-Gaussian target state densities. In addition, we decompose the 3D formulation into multiple 2D particle filters that operate on the 2D planes. Both selection and combining of the 2D particle filters for 3D tracking are presented and discussed. Finally, we analyze the tracking and association performance of the proposed approach especially in the cases of multitarget crossing and overlapping.
本文研究了结合联合多目标概率密度(JMPD)技术的粒子滤波三维多目标跟踪问题。该估计允许具有非标记测量关联的非线性目标运动和非高斯目标状态密度。此外,我们将3D公式分解为多个在2D平面上运行的2D粒子过滤器。提出并讨论了用于三维跟踪的二维粒子滤波器的选择和组合。最后,分析了该方法在多目标交叉和重叠情况下的跟踪和关联性能。
{"title":"Multitarget association and tracking in 3-D space based on particle filter with joint multitarget probability density","authors":"Jinseok Lee, Byung Guk Kim, S. Cho, Sangjin Hong, W. Cho","doi":"10.1109/AVSS.2007.4425374","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425374","url":null,"abstract":"This paper addresses the problem of 3-dimensional (3D) multitarget tracking using particle filter with the joint multitarget probability density (JMPD) technique. The estimation allows the nonlinear target motion with unlabeled measurement association as well as non-Gaussian target state densities. In addition, we decompose the 3D formulation into multiple 2D particle filters that operate on the 2D planes. Both selection and combining of the 2D particle filters for 3D tracking are presented and discussed. Finally, we analyze the tracking and association performance of the proposed approach especially in the cases of multitarget crossing and overlapping.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124291888","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Compact representation and probabilistic classification of human actions in videos 视频中人类行为的紧凑表示和概率分类
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425334
C. Colombo, Dario Comanducci, A. Bimbo
This paper addresses the problem of classifying human actions in a video sequence. A representation eigenspace approach based on the PCA algorithm is used to train the classifier according to an incremental learning scheme based on a "one action, one eigenspace" approach. Before dimensionality reduction, a high dimensional description of each frame of the video sequence is constructed, based on foreground blob analysis. Classification is performed by matching incrementally the reduced representation of the test image sequence against each of the learned ones, and accumulating matching scores according to a probabilistic framework, until a decision is obtained. Experimental results with real video sequences are presented and discussed.
本文研究了视频序列中人类行为的分类问题。采用基于PCA算法的表征特征空间方法,根据基于“一个动作,一个特征空间”方法的增量学习方案训练分类器。在降维之前,基于前景斑点分析,构建视频序列每帧的高维描述。分类是通过将测试图像序列的简化表示与每个学习到的图像序列进行增量匹配,并根据概率框架累积匹配分数,直到获得决策。给出了真实视频序列的实验结果并进行了讨论。
{"title":"Compact representation and probabilistic classification of human actions in videos","authors":"C. Colombo, Dario Comanducci, A. Bimbo","doi":"10.1109/AVSS.2007.4425334","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425334","url":null,"abstract":"This paper addresses the problem of classifying human actions in a video sequence. A representation eigenspace approach based on the PCA algorithm is used to train the classifier according to an incremental learning scheme based on a \"one action, one eigenspace\" approach. Before dimensionality reduction, a high dimensional description of each frame of the video sequence is constructed, based on foreground blob analysis. Classification is performed by matching incrementally the reduced representation of the test image sequence against each of the learned ones, and accumulating matching scores according to a probabilistic framework, until a decision is obtained. Experimental results with real video sequences are presented and discussed.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123215248","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
3-D model-based people detection & tracking 基于三维模型的人员检测和跟踪
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425375
G. Garibotto
The paper describes a method for people detection and tracking from multi-camera views. The proposed approach is based on 3D models of the person shape, where motion tracking is carried out in 3D space with re-projection onto calibrated images to perform target validation according to a prediction-verification paradigm. Multiple cameras with partial overlap can be used to cover a much wider area. The referred examples are based on the data base from PETS 2006 video sequences and a data base from EU-ISCAPS demonstration environment.
本文介绍了一种多摄像机视角下的人物检测与跟踪方法。所提出的方法基于人体形状的3D模型,其中运动跟踪在3D空间中进行,并根据预测-验证范式重新投影到校准图像上以执行目标验证。部分重叠的多个摄像机可以用来覆盖更广泛的区域。所引用的实例是基于pet 2006视频序列数据库和EU-ISCAPS演示环境的数据库。
{"title":"3-D model-based people detection & tracking","authors":"G. Garibotto","doi":"10.1109/AVSS.2007.4425375","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425375","url":null,"abstract":"The paper describes a method for people detection and tracking from multi-camera views. The proposed approach is based on 3D models of the person shape, where motion tracking is carried out in 3D space with re-projection onto calibrated images to perform target validation according to a prediction-verification paradigm. Multiple cameras with partial overlap can be used to cover a much wider area. The referred examples are based on the data base from PETS 2006 video sequences and a data base from EU-ISCAPS demonstration environment.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121214153","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Sphere detection and tracking for a space capturing operation 用于空间捕获操作的球体检测和跟踪
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425307
M. Kharbat, N. Aouf, A. Tsourdos, B. White
Capture mechanisms are used to transfer objects between two vehicles in the space with no physical contact. A sphere (canister) detection and tracking method using an enhanced Hough transform technique and Hinfin filter is proposed. The presented system aims to assist in the capture operation, currently investigated the European Space Agency and other partners, and to be used in space missions as an alternative to docking or berthing operations. Test results show the robustness and reliability of the proposed method. They also demonstrate the low computational and memory complexities needed.
捕捉机制用于在没有物理接触的情况下在空间中的两辆车之间转移物体。提出了一种基于增强霍夫变换技术和Hinfin滤波的球(罐)检测与跟踪方法。提出的系统旨在协助捕获操作,目前正在研究欧洲空间局和其他合作伙伴,并将用于空间任务,作为对接或停泊操作的替代方案。实验结果表明了该方法的鲁棒性和可靠性。它们还展示了所需的低计算和内存复杂性。
{"title":"Sphere detection and tracking for a space capturing operation","authors":"M. Kharbat, N. Aouf, A. Tsourdos, B. White","doi":"10.1109/AVSS.2007.4425307","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425307","url":null,"abstract":"Capture mechanisms are used to transfer objects between two vehicles in the space with no physical contact. A sphere (canister) detection and tracking method using an enhanced Hough transform technique and Hinfin filter is proposed. The presented system aims to assist in the capture operation, currently investigated the European Space Agency and other partners, and to be used in space missions as an alternative to docking or berthing operations. Test results show the robustness and reliability of the proposed method. They also demonstrate the low computational and memory complexities needed.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121437844","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Sign language detection using 3D visual cues 使用3D视觉线索的手语检测
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425350
J. Lichtenauer, G. T. Holt, E. Hendriks, M. Reinders
A 3D visual hand gesture recognition method is proposed that detects correctly performed signs from stereo camera input. Hand tracking is based on skin detection with an adaptive chrominance model to get high accuracy. Informative high level motion properties are extracted to simplify the classification task. Each example is mapped onto a fixed reference sign by Dynamic Time Warping, to get precise time correspondences. The classification is done by combining weak classifiers based on robust statistics. Each base classifier assumes a uniform distribution of a single feature, determined by winsorization on the noisy training set. The operating point of the classifier is determined by stretching the uniform distributions of the base classifiers instead of changing the threshold on the total posterior likelihood. In a cross validation with 120 signs performed by 70 different persons, 95% of the test signs were correctly detected at a false positive rate of 5%.
提出了一种三维视觉手势识别方法,该方法可以检测立体摄像机输入的正确手势。手部跟踪是基于皮肤检测的自适应色度模型,以获得较高的精度。提取信息丰富的高级运动属性以简化分类任务。通过动态时间翘曲将每个示例映射到固定的参考符号上,以获得精确的时间对应。分类是通过结合基于鲁棒统计的弱分类器来完成的。每个基分类器假设单个特征的均匀分布,通过对有噪声的训练集进行加权化来确定。分类器的工作点是通过拉伸基本分类器的均匀分布来确定的,而不是改变总后验似然的阈值。在由70个不同的人执行的120个标志的交叉验证中,95%的测试标志被正确检测,假阳性率为5%。
{"title":"Sign language detection using 3D visual cues","authors":"J. Lichtenauer, G. T. Holt, E. Hendriks, M. Reinders","doi":"10.1109/AVSS.2007.4425350","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425350","url":null,"abstract":"A 3D visual hand gesture recognition method is proposed that detects correctly performed signs from stereo camera input. Hand tracking is based on skin detection with an adaptive chrominance model to get high accuracy. Informative high level motion properties are extracted to simplify the classification task. Each example is mapped onto a fixed reference sign by Dynamic Time Warping, to get precise time correspondences. The classification is done by combining weak classifiers based on robust statistics. Each base classifier assumes a uniform distribution of a single feature, determined by winsorization on the noisy training set. The operating point of the classifier is determined by stretching the uniform distributions of the base classifiers instead of changing the threshold on the total posterior likelihood. In a cross validation with 120 signs performed by 70 different persons, 95% of the test signs were correctly detected at a false positive rate of 5%.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"54 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116998094","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
2D face pose normalisation using a 3D morphable model 使用3D变形模型的2D面部姿势归一化
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425285
J. Tena, Raymond S. Smith, M. Hamouz, J. Kittler, A. Hilton, J. Illingworth
The ever growing need for improved security, surveillance and identity protection, calls for the creation of evermore reliable and robust face recognition technology that is scalable and can be deployed in all kinds of environments without compromising its effectiveness. In this paper we study the impact that pose correction has on the performance of 2D face recognition. To measure the effect, we use a state of the art 2D recognition algorithm. The pose correction is performed by means of 3D morphable model. Our results on the non frontal XM2VTS database showed that pose correction can improve recognition rates up to 30%.
对安全、监控和身份保护的需求不断增长,要求创造更可靠、更强大的面部识别技术,这种技术具有可扩展性,可以部署在各种环境中,而不会影响其有效性。本文研究了姿态校正对二维人脸识别性能的影响。为了测量效果,我们使用了最先进的二维识别算法。采用三维变形模型进行姿态校正。我们在非正面XM2VTS数据库上的结果表明,姿态校正可以将识别率提高30%。
{"title":"2D face pose normalisation using a 3D morphable model","authors":"J. Tena, Raymond S. Smith, M. Hamouz, J. Kittler, A. Hilton, J. Illingworth","doi":"10.1109/AVSS.2007.4425285","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425285","url":null,"abstract":"The ever growing need for improved security, surveillance and identity protection, calls for the creation of evermore reliable and robust face recognition technology that is scalable and can be deployed in all kinds of environments without compromising its effectiveness. In this paper we study the impact that pose correction has on the performance of 2D face recognition. To measure the effect, we use a state of the art 2D recognition algorithm. The pose correction is performed by means of 3D morphable model. Our results on the non frontal XM2VTS database showed that pose correction can improve recognition rates up to 30%.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131251730","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Experiments with patch-based object classification 基于patch的目标分类实验
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425294
R. Wijnhoven, P. D. With
We present and experiment with a patch-based algorithm for the purpose of object classification in video surveillance. A feature vector is calculated based on template matching of a large set of image patches, within detected regions-of-interest (ROIs, also called blobs), of moving objects. Instead of matching direct image pixels, we use Gabor-filtered versions of the input image at several scales. We present results for a new typical video surveillance dataset containing over 9,000 object images. Additionally, we show results for the PETS 2001 dataset and another dataset from literature. Because our algorithm is not invariant to the object orientation, the set was split into four subsets with different orientation. We show the improvements, resulting from taking the object orientation into account. Using 50 training samples or higher, our resulting detection rate is on the average above 95%, which improves with the orientation consideration to 98%. Because of the inherent scalability of the algorithm, an embedded system implementation is well within reach.
本文提出并实验了一种基于补丁的视频监控目标分类算法。在检测到的运动物体的感兴趣区域(roi,也称为blobs)内,基于大量图像补丁的模板匹配计算特征向量。我们不是直接匹配图像像素,而是在几个尺度上使用gabor滤波版本的输入图像。我们给出了一个新的典型视频监控数据集的结果,该数据集包含超过9000个目标图像。此外,我们展示了PETS 2001数据集和另一个文献数据集的结果。由于算法对物体的方向不是不变的,因此将集合分成四个方向不同的子集。我们展示了由于考虑了面向对象而产生的改进。使用50个或更多的训练样本,我们得到的检测率平均在95%以上,考虑方向后提高到98%。由于该算法具有固有的可扩展性,因此可以在嵌入式系统中实现。
{"title":"Experiments with patch-based object classification","authors":"R. Wijnhoven, P. D. With","doi":"10.1109/AVSS.2007.4425294","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425294","url":null,"abstract":"We present and experiment with a patch-based algorithm for the purpose of object classification in video surveillance. A feature vector is calculated based on template matching of a large set of image patches, within detected regions-of-interest (ROIs, also called blobs), of moving objects. Instead of matching direct image pixels, we use Gabor-filtered versions of the input image at several scales. We present results for a new typical video surveillance dataset containing over 9,000 object images. Additionally, we show results for the PETS 2001 dataset and another dataset from literature. Because our algorithm is not invariant to the object orientation, the set was split into four subsets with different orientation. We show the improvements, resulting from taking the object orientation into account. Using 50 training samples or higher, our resulting detection rate is on the average above 95%, which improves with the orientation consideration to 98%. Because of the inherent scalability of the algorithm, an embedded system implementation is well within reach.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134516182","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Face recognition using non-linear image reconstruction 基于非线性图像重建的人脸识别
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425354
S. Duffner, Christophe Garcia
We present a face recognition technique based on a special type of convolutional neural network that is trained to extract characteristic features from face images and reconstruct the corresponding reference face images which are chosen beforehand for each individual to recognize. The reconstruction is realized by a so-called "bottle-neck" neural network that learns to project face images into a low-dimensional vector space and to reconstruct the respective reference images from the projected vectors. In contrast to methods based on the Principal Component Analysis (PCA), the Linear Discriminant Analysis (LDA) etc., the projection is non-linear and depends on the choice of the reference images. Moreover, local and global processing are closely interconnected and the respective parameters are conjointly learnt. Having trained the neural network, new face images can then be classified by comparing the respective projected vectors. We experimentally show that the choice of the reference images influences the final recognition performance and that this method outperforms linear projection methods in terms of precision and robustness.
本文提出了一种基于卷积神经网络的人脸识别技术,该技术通过训练从人脸图像中提取特征特征,并重建相应的参考人脸图像,这些图像是预先选择的,供每个人识别。重建是通过所谓的“瓶颈”神经网络实现的,该网络学习将人脸图像投影到低维向量空间中,并从投影向量中重建相应的参考图像。与基于主成分分析(PCA)、线性判别分析(LDA)等方法相比,投影是非线性的,依赖于参考图像的选择。此外,局部和全局处理紧密相连,各自的参数被联合学习。训练神经网络后,新的人脸图像可以通过比较各自的投影向量进行分类。实验表明,参考图像的选择会影响最终的识别性能,并且该方法在精度和鲁棒性方面优于线性投影方法。
{"title":"Face recognition using non-linear image reconstruction","authors":"S. Duffner, Christophe Garcia","doi":"10.1109/AVSS.2007.4425354","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425354","url":null,"abstract":"We present a face recognition technique based on a special type of convolutional neural network that is trained to extract characteristic features from face images and reconstruct the corresponding reference face images which are chosen beforehand for each individual to recognize. The reconstruction is realized by a so-called \"bottle-neck\" neural network that learns to project face images into a low-dimensional vector space and to reconstruct the respective reference images from the projected vectors. In contrast to methods based on the Principal Component Analysis (PCA), the Linear Discriminant Analysis (LDA) etc., the projection is non-linear and depends on the choice of the reference images. Moreover, local and global processing are closely interconnected and the respective parameters are conjointly learnt. Having trained the neural network, new face images can then be classified by comparing the respective projected vectors. We experimentally show that the choice of the reference images influences the final recognition performance and that this method outperforms linear projection methods in terms of precision and robustness.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128195805","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Distributed video surveillance using hardware-friendly sparse large margin classifiers 分布式视频监控使用硬件友好的稀疏大边界分类器
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425291
A. Kerhet, F. Leonardi, A. Boni, P. Lombardo, M. Magno, L. Benini
In contrast to video sensors which just "watch " the world, present-day research is aimed at developing intelligent devices able to interpret it locally. A number of such devices are available on the market, very powerful on the one hand, but requiring either connection to the power grid, or massive rechargeable batteries on the other. MicrelEye, the wireless video sensor node presented in this paper, targets a different design point: portability and a scanty power budget, while still providing a prominent level of intelligence, namely objects classification. To deal with such a challenging task, we propose and implement a new SVM-like hardware-oriented algorithm called ERSVM. The case study considered in this work is people detection. The obtained results suggest that the present technology allows for the design of simple intelligent video nodes capable of performing local classification tasks.
与仅仅“观察”世界的视频传感器不同,目前的研究旨在开发能够在本地解读世界的智能设备。市场上有很多这样的设备,一方面非常强大,但需要连接电网,或者另一方面需要大量的可充电电池。本文提出的无线视频传感器节点MicrelEye针对不同的设计点:便携性和低功耗预算,同时仍然提供突出的智能水平,即对象分类。为了处理这样一个具有挑战性的任务,我们提出并实现了一种新的类似svm的面向硬件的算法,称为ERSVM。在这项工作中考虑的案例研究是人的检测。所得结果表明,本技术允许设计能够执行局部分类任务的简单智能视频节点。
{"title":"Distributed video surveillance using hardware-friendly sparse large margin classifiers","authors":"A. Kerhet, F. Leonardi, A. Boni, P. Lombardo, M. Magno, L. Benini","doi":"10.1109/AVSS.2007.4425291","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425291","url":null,"abstract":"In contrast to video sensors which just \"watch \" the world, present-day research is aimed at developing intelligent devices able to interpret it locally. A number of such devices are available on the market, very powerful on the one hand, but requiring either connection to the power grid, or massive rechargeable batteries on the other. MicrelEye, the wireless video sensor node presented in this paper, targets a different design point: portability and a scanty power budget, while still providing a prominent level of intelligence, namely objects classification. To deal with such a challenging task, we propose and implement a new SVM-like hardware-oriented algorithm called ERSVM. The case study considered in this work is people detection. The obtained results suggest that the present technology allows for the design of simple intelligent video nodes capable of performing local classification tasks.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"91 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123572122","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 31
期刊
2007 IEEE Conference on Advanced Video and Signal Based Surveillance
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1