首页 > 最新文献

2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops最新文献

英文 中文
A dual-layer estimator architecture for long-term localization 用于长期定位的双层估计器体系结构
Anastasios I. Mourikis, S. Roumeliotis
In this paper, we present a localization algorithm for estimating the 3D position and orientation (pose) of a moving vehicle based on visual and inertial measurements. The main advantage of the proposed method is that it provides precise pose estimates at low computational cost. This is achieved by introducing a two-layer estimation architecture that processes measurements based on their information content. Inertial measurements and feature tracks between consecutive images are processed locally in the first layer (multi-state-constraint Kalman filter) providing estimates for the motion of the vehicle at a high rate. The second layer comprises a bundle adjustment iterative estimator that operates intermittently so as to (i) reduce the effect of the linearization errors, and (ii) update the state estimates every time an area is re-visited and features are re-detected (loop closure). Through this process reliable state estimates are available continuously, while the estimation errors remain bounded during long-term operation. The performance of the developed system is demonstrated in large-scale experiments, involving a vehicle localizing within an urban area.
在本文中,我们提出了一种基于视觉和惯性测量估计移动车辆的三维位置和方向(姿态)的定位算法。该方法的主要优点是能够以较低的计算成本提供精确的姿态估计。这是通过引入基于信息内容处理度量的两层评估体系结构来实现的。惯性测量和连续图像之间的特征轨迹在第一层(多状态约束卡尔曼滤波)进行局部处理,以高速估计车辆的运动。第二层包括一个间歇运行的束调整迭代估计器,以便(i)减少线性化误差的影响,(ii)每次重新访问一个区域和重新检测特征时更新状态估计(环路关闭)。通过该过程,可以连续获得可靠的状态估计,而在长期运行过程中,估计误差保持有界。该系统的性能在大规模实验中得到了验证,其中包括在城市区域内进行车辆定位。
{"title":"A dual-layer estimator architecture for long-term localization","authors":"Anastasios I. Mourikis, S. Roumeliotis","doi":"10.1109/CVPRW.2008.4563131","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563131","url":null,"abstract":"In this paper, we present a localization algorithm for estimating the 3D position and orientation (pose) of a moving vehicle based on visual and inertial measurements. The main advantage of the proposed method is that it provides precise pose estimates at low computational cost. This is achieved by introducing a two-layer estimation architecture that processes measurements based on their information content. Inertial measurements and feature tracks between consecutive images are processed locally in the first layer (multi-state-constraint Kalman filter) providing estimates for the motion of the vehicle at a high rate. The second layer comprises a bundle adjustment iterative estimator that operates intermittently so as to (i) reduce the effect of the linearization errors, and (ii) update the state estimates every time an area is re-visited and features are re-detected (loop closure). Through this process reliable state estimates are available continuously, while the estimation errors remain bounded during long-term operation. The performance of the developed system is demonstrated in large-scale experiments, involving a vehicle localizing within an urban area.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124425174","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 64
Fast gain-adaptive KLT tracking on the GPU GPU上的快速增益自适应KLT跟踪
C. Zach, D. Gallup, Jan-Michael Frahm
High-performance feature tracking from video input is a valuable tool in many computer vision techniques and mixed reality applications. This work presents a refined and substantially accelerated approach to KLT feature tracking performed on the GPU. Additionally, a global gain ratio between successive frames is estimated to compensate for changes in the camera exposure. The proposed approach achieves more than 200 frames per second on state-of-the art consumer GPUs for PAL (720 times 576) resolution data, and delivers real-time performance even on low-end mobile graphics processors.
视频输入的高性能特征跟踪在许多计算机视觉技术和混合现实应用中是一个有价值的工具。这项工作提出了一种在GPU上执行KLT特征跟踪的改进和实质上加速的方法。此外,估计连续帧之间的全局增益比以补偿相机曝光的变化。所提出的方法可以在最先进的消费级gpu上实现每秒200帧以上的PAL (720 × 576)分辨率数据,并且即使在低端移动图形处理器上也能提供实时性能。
{"title":"Fast gain-adaptive KLT tracking on the GPU","authors":"C. Zach, D. Gallup, Jan-Michael Frahm","doi":"10.1109/CVPRW.2008.4563089","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563089","url":null,"abstract":"High-performance feature tracking from video input is a valuable tool in many computer vision techniques and mixed reality applications. This work presents a refined and substantially accelerated approach to KLT feature tracking performed on the GPU. Additionally, a global gain ratio between successive frames is estimated to compensate for changes in the camera exposure. The proposed approach achieves more than 200 frames per second on state-of-the art consumer GPUs for PAL (720 times 576) resolution data, and delivers real-time performance even on low-end mobile graphics processors.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"67 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127970508","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 83
Incident light related distance error study and calibration of the PMD-range imaging camera pmd距离成像相机的入射光相关距离误差研究与标定
Jochen Radmer, Pol Moser Fuste, H. Schmidt, J. Krüger
For various applications, such as object recognition or tracking and especially when the object is partly occluded or articulated, 3D information is crucial for the robustness of the application. A recently developed sensor to acquire distance information is based on the Photo Mixer Device (PMD)for which a distance error based on different causes can be observed. This article presents an improved distance calibration approach for PMD-based distance sensoring which handles objects with different Lambertian reflectance properties. Within this scope the relation of the sources of distance errors were investigated. Where applicable they were isolated for relational studies with the actuating variables, i.e. integration time, amplitude and measured distance, as these are the only parameters available for the calibration. The calibration results of the proposed method excel the results of all other known methods. In particular with objects with unknown reflectance properties a significant reduction of the error is achieved.
对于各种应用程序,例如对象识别或跟踪,特别是当对象部分遮挡或铰接时,3D信息对于应用程序的鲁棒性至关重要。最近开发的一种用于获取距离信息的传感器是基于光混合装置(PMD)的,它可以观察到基于不同原因的距离误差。本文提出了一种改进的基于pmd的距离校准方法,用于处理具有不同朗伯反射率的物体。在此范围内,研究了距离误差源之间的关系。在适用的情况下,它们被隔离,用于与驱动变量的关系研究,即积分时间,振幅和测量距离,因为这些是唯一可用于校准的参数。该方法的校准结果优于所有其他已知方法的校准结果。特别是对于具有未知反射特性的物体,可以显著减少误差。
{"title":"Incident light related distance error study and calibration of the PMD-range imaging camera","authors":"Jochen Radmer, Pol Moser Fuste, H. Schmidt, J. Krüger","doi":"10.1109/CVPRW.2008.4563168","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563168","url":null,"abstract":"For various applications, such as object recognition or tracking and especially when the object is partly occluded or articulated, 3D information is crucial for the robustness of the application. A recently developed sensor to acquire distance information is based on the Photo Mixer Device (PMD)for which a distance error based on different causes can be observed. This article presents an improved distance calibration approach for PMD-based distance sensoring which handles objects with different Lambertian reflectance properties. Within this scope the relation of the sources of distance errors were investigated. Where applicable they were isolated for relational studies with the actuating variables, i.e. integration time, amplitude and measured distance, as these are the only parameters available for the calibration. The calibration results of the proposed method excel the results of all other known methods. In particular with objects with unknown reflectance properties a significant reduction of the error is achieved.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127983507","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 41
Variational registration of tensor-valued images 张量值图像的变分配准
S. Barbieri, M. Welk, J. Weickert
We present a variational framework for the registration of tensor-valued images. It is based on an energy functional with four terms: a data term based on a diffusion tensor constancy constraint, a compatibility term encoding the physical model linking domain deformations and tensor reorientation, and smoothness terms for deformation and tensor reorientation. Although the tensor deformation model employed here is designed with regard to diffusion tensor MRI data, the separation of data and compatibility term allows to adapt the model easily to different tensor deformation models. We minimise the energy functional with respect to both transformation fields by a multiscale gradient descent. Experiments demonstrate the viability and potential of this approach in the registration of tensor-valued images.
我们提出了一个张量值图像配准的变分框架。它基于一个包含四项的能量泛函:一个基于扩散张量常数约束的数据项,一个编码连接域变形和张量重定向的物理模型的兼容性项,以及一个用于变形和张量重定向的平滑项。虽然本文采用的张量变形模型是针对弥散张量MRI数据设计的,但由于数据与相容项的分离,使得模型可以很容易地适应不同的张量变形模型。我们通过多尺度梯度下降最小化了两个变换场的能量泛函。实验证明了该方法在张量值图像配准中的可行性和潜力。
{"title":"Variational registration of tensor-valued images","authors":"S. Barbieri, M. Welk, J. Weickert","doi":"10.1109/CVPRW.2008.4562964","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4562964","url":null,"abstract":"We present a variational framework for the registration of tensor-valued images. It is based on an energy functional with four terms: a data term based on a diffusion tensor constancy constraint, a compatibility term encoding the physical model linking domain deformations and tensor reorientation, and smoothness terms for deformation and tensor reorientation. Although the tensor deformation model employed here is designed with regard to diffusion tensor MRI data, the separation of data and compatibility term allows to adapt the model easily to different tensor deformation models. We minimise the energy functional with respect to both transformation fields by a multiscale gradient descent. Experiments demonstrate the viability and potential of this approach in the registration of tensor-valued images.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133617911","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Multiple cue integration in transductive confidence machines for head pose classification 基于多线索集成的头部姿态分类换能型置信度机器
V. Balasubramanian, S. Panchanathan, Shayok Chakraborty
An important facet of learning in an online setting is the confidence associated with a prediction on a given test data point. In an online learning scenario, it would be expected that the system can increase its confidence of prediction as training data increases. We present a statistical approach in this work to associate a confidence value with a predicted class label in an online learning scenario. Our work is based on the existing work on transductive confidence machines (TCM) [1], which provided a methodology to define a heuristic confidence measure. We applied this approach to the problem of head pose classification from face images, and extended the framework to compute a confidence value when multiple cues are extracted from images to perform classification. Our approach is based on combining the results of multiple hypotheses and obtaining an integrated p-value to validate a single test hypothesis. From our experiments on the widely accepted FERET database, we obtained results which corroborated the significance of confidence measures - particularly, in online learning approaches. We could infer from our results with transductive learning that using confidence measures in online learning could yield significant boosts in the prediction accuracy, which would be very useful in critical pattern recognition applications.
在线学习的一个重要方面是与给定测试数据点的预测相关的置信度。在在线学习场景中,可以期望系统随着训练数据的增加而增加其预测的置信度。在这项工作中,我们提出了一种统计方法,将置信度值与在线学习场景中的预测类标签相关联。我们的工作是基于现有的关于传导置信机(TCM)的工作[1],它提供了一种定义启发式置信度度量的方法。我们将该方法应用于人脸图像的头部姿态分类问题,并扩展了该框架,在从图像中提取多个线索进行分类时计算置信值。我们的方法是基于组合多个假设的结果并获得一个集成的p值来验证单个检验假设。从我们在广泛接受的FERET数据库上的实验中,我们获得的结果证实了信心措施的重要性,特别是在在线学习方法中。我们可以从转换学习的结果中推断,在在线学习中使用置信度度量可以显著提高预测精度,这在关键的模式识别应用中非常有用。
{"title":"Multiple cue integration in transductive confidence machines for head pose classification","authors":"V. Balasubramanian, S. Panchanathan, Shayok Chakraborty","doi":"10.1109/CVPRW.2008.4563070","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563070","url":null,"abstract":"An important facet of learning in an online setting is the confidence associated with a prediction on a given test data point. In an online learning scenario, it would be expected that the system can increase its confidence of prediction as training data increases. We present a statistical approach in this work to associate a confidence value with a predicted class label in an online learning scenario. Our work is based on the existing work on transductive confidence machines (TCM) [1], which provided a methodology to define a heuristic confidence measure. We applied this approach to the problem of head pose classification from face images, and extended the framework to compute a confidence value when multiple cues are extracted from images to perform classification. Our approach is based on combining the results of multiple hypotheses and obtaining an integrated p-value to validate a single test hypothesis. From our experiments on the widely accepted FERET database, we obtained results which corroborated the significance of confidence measures - particularly, in online learning approaches. We could infer from our results with transductive learning that using confidence measures in online learning could yield significant boosts in the prediction accuracy, which would be very useful in critical pattern recognition applications.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116951436","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Implementation of Advanced Encryption Standard for encryption and decryption of images and text on a GPU 高级加密标准在GPU上对图像和文本进行加密和解密的实现
Manoj Seshadrinathan, K. Dempski
In this paper, we propose a system for the complete implementation of the advanced encryption standard (AES) for encryption and decryption of images and text on a graphics processing unit. The GPU acts as a valuable co-processor that relieves the load off the CPU. In the decryption stage, we use a novel technique to display the decrypted images and text on the screen without bringing it onto CPU memory. We also present a system for encryption and decryption of hybrid map tiles generated from GIS data sets.
在本文中,我们提出了一种在图形处理单元上完全实现用于图像和文本加密和解密的高级加密标准(AES)的系统。GPU作为一个有价值的协处理器,减轻了CPU的负载。在解密阶段,我们使用一种新颖的技术在屏幕上显示解密后的图像和文本,而无需将其放入CPU内存。我们还提出了一个系统,用于加密和解密由GIS数据集生成的混合地图块。
{"title":"Implementation of Advanced Encryption Standard for encryption and decryption of images and text on a GPU","authors":"Manoj Seshadrinathan, K. Dempski","doi":"10.1109/CVPRW.2008.4563094","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563094","url":null,"abstract":"In this paper, we propose a system for the complete implementation of the advanced encryption standard (AES) for encryption and decryption of images and text on a graphics processing unit. The GPU acts as a valuable co-processor that relieves the load off the CPU. In the decryption stage, we use a novel technique to display the decrypted images and text on the screen without bringing it onto CPU memory. We also present a system for encryption and decryption of hybrid map tiles generated from GIS data sets.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134185389","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
New insights into the calibration of ToF-sensors tof传感器校准的新见解
Marvin Lindner, A. Kolb, T. Ringbeck
Time-of-flight (ToF) sensors have become an alternative to conventional distance sensing techniques like laser scanners or image based stereo. ToF sensors provide full range distance information at high frame-rates and thus have a significant impact onto current research in areas like online object recognition, collision prevention or scene reconstruction. However, ToF cameras like the photonic mixer device (PMD) still exhibit a number of challenges regarding static and dynamic effects, e.g. systematic distance errors and motion artefacts, respectively. Sensor calibration techniques reducing static system errors have been proposed and show promising results. However, current calibration techniques in general need a large set of reference data in order to determine the corresponding parameters for the calibration model. This paper introduces a new calibration approach which combines different demodulation techniques for the ToF- camera 's reference signal. Examples show, that the resulting combined demodulation technique yields improved distance values based on only two required reference data sets.
飞行时间(ToF)传感器已经成为传统距离传感技术(如激光扫描仪或基于图像的立体)的替代方案。ToF传感器以高帧率提供全范围距离信息,因此对当前在线物体识别、碰撞预防或场景重建等领域的研究产生了重大影响。然而,像光子混频器设备(PMD)这样的ToF相机仍然表现出一些关于静态和动态效果的挑战,例如系统距离误差和运动伪影。已经提出了减少静态系统误差的传感器校准技术,并显示出良好的效果。然而,目前的校准技术通常需要大量的参考数据来确定校准模型的相应参数。本文介绍了一种结合多种解调技术对ToF摄像机参考信号进行标定的新方法。实例表明,所得到的组合解调技术仅基于两个所需的参考数据集就能产生改进的距离值。
{"title":"New insights into the calibration of ToF-sensors","authors":"Marvin Lindner, A. Kolb, T. Ringbeck","doi":"10.1109/CVPRW.2008.4563172","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563172","url":null,"abstract":"Time-of-flight (ToF) sensors have become an alternative to conventional distance sensing techniques like laser scanners or image based stereo. ToF sensors provide full range distance information at high frame-rates and thus have a significant impact onto current research in areas like online object recognition, collision prevention or scene reconstruction. However, ToF cameras like the photonic mixer device (PMD) still exhibit a number of challenges regarding static and dynamic effects, e.g. systematic distance errors and motion artefacts, respectively. Sensor calibration techniques reducing static system errors have been proposed and show promising results. However, current calibration techniques in general need a large set of reference data in order to determine the corresponding parameters for the calibration model. This paper introduces a new calibration approach which combines different demodulation techniques for the ToF- camera 's reference signal. Examples show, that the resulting combined demodulation technique yields improved distance values based on only two required reference data sets.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"29 6","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"113933669","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 36
A probabilistic representation of LiDAR range data for efficient 3D object detection 用于有效三维目标检测的激光雷达距离数据的概率表示
Theodore C. Yapo, C. Stewart, R. Radke
We present a novel approach to 3D object detection in scenes scanned by LiDAR sensors, based on a probabilistic representation of free, occupied, and hidden space that extends the concept of occupancy grids from robot mapping algorithms. This scene representation naturally handles LiDAR sampling issues, can be used to fuse multiple LiDAR data sets, and captures the inherent uncertainty of the data due to occlusions and clutter. Using this model, we formulate a hypothesis testing methodology to determine the probability that given 3D objects are present in the scene. By propagating uncertainty in the original sample points, we are able to measure confidence in the detection results in a principled way. We demonstrate the approach in examples of detecting objects that are partially occluded by scene clutter such as camouflage netting.
我们提出了一种在激光雷达传感器扫描的场景中进行3D物体检测的新方法,该方法基于空闲空间、占用空间和隐藏空间的概率表示,扩展了机器人映射算法中占用网格的概念。这种场景表示自然地处理了LiDAR采样问题,可用于融合多个LiDAR数据集,并捕获由于遮挡和杂波导致的数据固有的不确定性。使用这个模型,我们制定了一个假设检验方法来确定给定的3D物体出现在场景中的概率。通过在原始样本点中传播不确定性,我们能够以原则性的方式测量检测结果的置信度。我们在检测被场景杂波(如伪装网)部分遮挡的物体的示例中演示了该方法。
{"title":"A probabilistic representation of LiDAR range data for efficient 3D object detection","authors":"Theodore C. Yapo, C. Stewart, R. Radke","doi":"10.1109/CVPRW.2008.4563033","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563033","url":null,"abstract":"We present a novel approach to 3D object detection in scenes scanned by LiDAR sensors, based on a probabilistic representation of free, occupied, and hidden space that extends the concept of occupancy grids from robot mapping algorithms. This scene representation naturally handles LiDAR sampling issues, can be used to fuse multiple LiDAR data sets, and captures the inherent uncertainty of the data due to occlusions and clutter. Using this model, we formulate a hypothesis testing methodology to determine the probability that given 3D objects are present in the scene. By propagating uncertainty in the original sample points, we are able to measure confidence in the detection results in a principled way. We demonstrate the approach in examples of detecting objects that are partially occluded by scene clutter such as camouflage netting.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123875887","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 29
Verifying liveness by multiple experts in face biometrics 由多名面部生物识别专家验证活体
K. Kollreider, H. Fronthaler, J. Bigün
Resisting spoofing attempts via photographs and video playbacks is a vital issue for the success of face biometrics. Yet, the ldquolivenessrdquo topic has only been partially studied in the past. In this paper we are suggesting a holistic liveness detection paradigm that collaborates with standard techniques in 2D face biometrics. The experiments show that many attacks are avertible via a combination of anti-spoofing measures. We have investigated the topic using real-time techniques and applied them to real-life spoofing scenarios in an indoor, yet uncontrolled environment.
抵制通过照片和视频回放的欺骗企图是面部生物识别技术成功的关键问题。然而,过去仅对ldolivenessrquo主题进行了部分研究。在本文中,我们提出了一种整体的活体检测范式,该范式与二维面部生物识别的标准技术相结合。实验表明,许多攻击可以通过反欺骗措施的组合来避免。我们使用实时技术研究了该主题,并将其应用于室内但不受控制的环境中的真实欺骗场景。
{"title":"Verifying liveness by multiple experts in face biometrics","authors":"K. Kollreider, H. Fronthaler, J. Bigün","doi":"10.1109/CVPRW.2008.4563115","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563115","url":null,"abstract":"Resisting spoofing attempts via photographs and video playbacks is a vital issue for the success of face biometrics. Yet, the ldquolivenessrdquo topic has only been partially studied in the past. In this paper we are suggesting a holistic liveness detection paradigm that collaborates with standard techniques in 2D face biometrics. The experiments show that many attacks are avertible via a combination of anti-spoofing measures. We have investigated the topic using real-time techniques and applied them to real-life spoofing scenarios in an indoor, yet uncontrolled environment.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127709336","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 131
Scalable classifiers for Internet vision tasks 用于互联网视觉任务的可扩展分类器
Tom Yeh, John J. Lee, Trevor Darrell
Object recognition systems designed for Internet applications typically need to adapt to userspsila needs in a flexible fashion and scale up to very large data sets. In this paper, we analyze the complexity of several multiclass SVM-based algorithms and highlight the computational bottleneck they suffer at test time: comparing the input image to every training image. We propose an algorithm that overcomes this bottleneck; it offers not only the efficiency of a simple nearest-neighbor classifier, by voting on class labels based on the k nearest neighbors quickly determined by a vocabulary tree, but also the recognition accuracy comparable to that of a complex SVM classifier, by incorporating SVM parameters into the voting scores incrementally accumulated from individual image features. Empirical results demonstrate that adjusting votes by relevant support vector weights can improve the recognition accuracy of a nearest-neighbor classifier without sacrificing speed. Compared to existing methods, our algorithm achieves a ten-fold speed increase while incurring an acceptable accuracy loss that can be easily offset by showing about two more labels in the result. The speed, scalability, and adaptability of our algorithm makes it suitable for Internet vision applications.
为Internet应用程序设计的对象识别系统通常需要以灵活的方式适应用户的需求,并扩展到非常大的数据集。在本文中,我们分析了几种基于svm的多类算法的复杂性,并强调了它们在测试时遇到的计算瓶颈:将输入图像与每个训练图像进行比较。我们提出了一种克服这一瓶颈的算法;它不仅提供了简单的最近邻分类器的效率,通过基于词汇树快速确定的k个最近邻对类标签进行投票,而且通过将SVM参数纳入从单个图像特征逐渐积累的投票分数中,其识别精度可与复杂的SVM分类器相媲美。实证结果表明,通过相关支持向量权重调整投票可以在不牺牲速度的情况下提高最近邻分类器的识别精度。与现有方法相比,我们的算法实现了十倍的速度提升,同时产生了可接受的精度损失,可以通过在结果中显示大约两个标签来轻松抵消。该算法的速度、可扩展性和适应性使其适合于互联网视觉应用。
{"title":"Scalable classifiers for Internet vision tasks","authors":"Tom Yeh, John J. Lee, Trevor Darrell","doi":"10.1109/CVPRW.2008.4562958","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4562958","url":null,"abstract":"Object recognition systems designed for Internet applications typically need to adapt to userspsila needs in a flexible fashion and scale up to very large data sets. In this paper, we analyze the complexity of several multiclass SVM-based algorithms and highlight the computational bottleneck they suffer at test time: comparing the input image to every training image. We propose an algorithm that overcomes this bottleneck; it offers not only the efficiency of a simple nearest-neighbor classifier, by voting on class labels based on the k nearest neighbors quickly determined by a vocabulary tree, but also the recognition accuracy comparable to that of a complex SVM classifier, by incorporating SVM parameters into the voting scores incrementally accumulated from individual image features. Empirical results demonstrate that adjusting votes by relevant support vector weights can improve the recognition accuracy of a nearest-neighbor classifier without sacrificing speed. Compared to existing methods, our algorithm achieves a ten-fold speed increase while incurring an acceptable accuracy loss that can be easily offset by showing about two more labels in the result. The speed, scalability, and adaptability of our algorithm makes it suitable for Internet vision applications.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127797960","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
期刊
2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1