首页 > 最新文献

2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops最新文献

英文 中文
Binocular dance pose recognition and body orientation estimation via multilinear analysis 基于多线性分析的双目舞蹈姿态识别与身体方位估计
Bo Peng, G. Qian
In this paper, we propose a novel approach to dance pose recognition and body orientation estimation using multilinear analysis. By performing tensor decomposition and projection using silhouette images obtained from wide base-line binocular cameras, low dimensional pose and body orientation coefficient vectors can be extracted. Different from traditional tensor-based recognition methods, the proposed approach takes the pose coefficient vector as features to train a family of support vector machines as pose classifiers. Using the body orientation coefficient vectors, a one-dimensional orientation manifold is learned and further used for the estimation of body orientation. Experiment results obtained using both synthetic and real image data showed the efficacy of the proposed approach, and that our approach outperformed the traditional tensor-based approach in the comparative test.
本文提出了一种基于多线性分析的舞蹈姿态识别和姿态估计方法。利用宽基线双目摄像机获得的轮廓图像进行张量分解和投影,提取低维姿态和身体方向系数矢量。与传统的基于张量的识别方法不同,该方法以位姿系数向量为特征,训练一组支持向量机作为位姿分类器。利用身体的方向系数向量,学习一维方向流形,并进一步用于身体的方向估计。合成和真实图像数据的实验结果均表明了该方法的有效性,并且在对比测试中优于传统的基于张量的方法。
{"title":"Binocular dance pose recognition and body orientation estimation via multilinear analysis","authors":"Bo Peng, G. Qian","doi":"10.1109/CVPRW.2008.4562970","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4562970","url":null,"abstract":"In this paper, we propose a novel approach to dance pose recognition and body orientation estimation using multilinear analysis. By performing tensor decomposition and projection using silhouette images obtained from wide base-line binocular cameras, low dimensional pose and body orientation coefficient vectors can be extracted. Different from traditional tensor-based recognition methods, the proposed approach takes the pose coefficient vector as features to train a family of support vector machines as pose classifiers. Using the body orientation coefficient vectors, a one-dimensional orientation manifold is learned and further used for the estimation of body orientation. Experiment results obtained using both synthetic and real image data showed the efficacy of the proposed approach, and that our approach outperformed the traditional tensor-based approach in the comparative test.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126884091","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 23
Circular generalized cylinder fitting for 3D reconstruction in endoscopic imaging based on MRF 基于MRF的内镜成像三维重建的圆形广义圆柱体拟合
Jin Zhou, Ananya Das, Feng Li, Baoxin Li
Endoscopy has become an established procedure for the diagnosis and therapy of various gastrointestinal (GI) ailments, and has also emerged as a commonly-used technique for minimally-invasive surgery. Most existing endoscopes are monocular, with stereo-endoscopy facing practical difficulties, preventing the physicians/surgeons from having a desired, realistic 3D view. Traditional monocular 3D reconstruction approaches (e.g., structure from motion) face extraordinary challenges for this application due to issues including noisy data, lack of textures supporting robust feature matching, nonrigidity of the objects, and glare artifacts from the imaging process, etc. In this paper, we propose a method to automatically reconstruct 3D structure from a monocular endoscopic video. Our approach attempts to address the above challenges by incorporating a circular generalized cylinder (CGC) model in 3D reconstruction. The CGC model is decomposed as a series of 3D circles. To reconstruct this model, we formulate the problem as one of maximum a posteriori estimation within a Markov random field framework, so as to ensure the smoothness constraints of the CGC model and to support robust search for the optimal solution, which is achieved by a two-stage heuristic search scheme. Both simulated and real data experiments demonstrate the effectiveness of the proposed approach.
内窥镜检查已成为各种胃肠道疾病的诊断和治疗的既定程序,也已成为一种常用的微创手术技术。大多数现有的内窥镜都是单目的,立体内窥镜面临着实际的困难,使医生/外科医生无法获得理想的、真实的3D视图。传统的单目3D重建方法(例如,从运动中获取结构)在这一应用中面临着巨大的挑战,原因包括数据噪声、缺乏支持鲁棒特征匹配的纹理、物体的非刚性以及成像过程中的眩光伪影等问题。在本文中,我们提出了一种从单眼内窥镜视频中自动重建三维结构的方法。我们的方法试图通过在三维重建中结合圆形广义圆柱体(CGC)模型来解决上述挑战。CGC模型被分解为一系列三维圆。为了重建该模型,我们将问题描述为马尔可夫随机场框架内的最大后验估计问题,以保证CGC模型的平滑性约束,并支持对最优解的鲁棒搜索,这是通过两阶段启发式搜索方案实现的。仿真实验和实际数据实验均证明了该方法的有效性。
{"title":"Circular generalized cylinder fitting for 3D reconstruction in endoscopic imaging based on MRF","authors":"Jin Zhou, Ananya Das, Feng Li, Baoxin Li","doi":"10.1109/CVPRW.2008.4563010","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563010","url":null,"abstract":"Endoscopy has become an established procedure for the diagnosis and therapy of various gastrointestinal (GI) ailments, and has also emerged as a commonly-used technique for minimally-invasive surgery. Most existing endoscopes are monocular, with stereo-endoscopy facing practical difficulties, preventing the physicians/surgeons from having a desired, realistic 3D view. Traditional monocular 3D reconstruction approaches (e.g., structure from motion) face extraordinary challenges for this application due to issues including noisy data, lack of textures supporting robust feature matching, nonrigidity of the objects, and glare artifacts from the imaging process, etc. In this paper, we propose a method to automatically reconstruct 3D structure from a monocular endoscopic video. Our approach attempts to address the above challenges by incorporating a circular generalized cylinder (CGC) model in 3D reconstruction. The CGC model is decomposed as a series of 3D circles. To reconstruct this model, we formulate the problem as one of maximum a posteriori estimation within a Markov random field framework, so as to ensure the smoothness constraints of the CGC model and to support robust search for the optimal solution, which is achieved by a two-stage heuristic search scheme. Both simulated and real data experiments demonstrate the effectiveness of the proposed approach.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"40 3","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"113976041","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 25
A GPU-based implementation of motion detection from a moving platform 基于gpu的移动平台运动检测实现
Qian Yu, G. Medioni
We describe a GPU-based implementation of motion detection from a moving platform. Motion detection from a moving platform is inherently difficult as the moving camera induces 2D motion field in the entire image. A step compensating for camera motion is required prior to estimating of the background model. Due to inevitable registration errors, the background model is estimated according to a sliding window of frames to avoid the case where erroneous registration influences the quality of the detection for the whole sequence. However, this approach involves several characteristics that put a heavy burden on real-time CPU implementation. We exploit GPU to achieve significant acceleration over standard CPU implementations. Our GPU-based implementation can build the background model and detect motion regions at around 18 fps on 320times240 videos that are captured for a moving camera.
我们描述了一个基于gpu的移动平台运动检测的实现。由于移动的摄像机会在整个图像中产生二维运动场,因此从移动平台进行运动检测本身就很困难。在估计背景模型之前,需要对相机运动进行步进补偿。由于不可避免的配准误差,背景模型采用帧间滑动窗口估计,避免配准错误影响整个序列的检测质量。然而,这种方法涉及到的几个特性给实时CPU实现带来了沉重的负担。我们利用GPU实现比标准CPU实现显著的加速。我们基于gpu的实现可以构建背景模型,并在为移动摄像机捕获的320times240视频中以大约18 fps的速度检测运动区域。
{"title":"A GPU-based implementation of motion detection from a moving platform","authors":"Qian Yu, G. Medioni","doi":"10.1109/CVPRW.2008.4563096","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563096","url":null,"abstract":"We describe a GPU-based implementation of motion detection from a moving platform. Motion detection from a moving platform is inherently difficult as the moving camera induces 2D motion field in the entire image. A step compensating for camera motion is required prior to estimating of the background model. Due to inevitable registration errors, the background model is estimated according to a sliding window of frames to avoid the case where erroneous registration influences the quality of the detection for the whole sequence. However, this approach involves several characteristics that put a heavy burden on real-time CPU implementation. We exploit GPU to achieve significant acceleration over standard CPU implementations. Our GPU-based implementation can build the background model and detect motion regions at around 18 fps on 320times240 videos that are captured for a moving camera.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"48 12","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114020286","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 40
Improving the selection and detection of visual landmarks through object tracking 通过目标跟踪改进视觉地标的选择和检测
P. Espinace, A. Soto
The unsupervised selection and posterior recognition of visual landmarks is a highly valuable perceptual capability for a mobile robot. Recently, we proposed a system that aims to achieve this capability by combining a bottom-up data driven approach with top-down feedback provided by high level semantic representations. The bottom-up approach is based on three main mechanisms: visual attention, area segmentation, and landmark characterization. The top-down feedback is based on two information sources: i) An estimation of the robot position that reduces the searching scope for potential matches with previously selected landmarks, ii) A set of weights that, according to the results of previous recognitions, controls the influence of different segmentation algorithms in the recognition of each landmark. In this paper we explore the benefits of extending our previous work by including a visual tracking step for each of the selected landmarks. Our intuition is that the inclusion of a tracking step can help to improve the model of each landmark by associating and selecting information from its most significant views. Furthermore, it can also help to avoid problems related to the selection of spurious landmarks. Our results confirm these intuitions by showing that the inclusion of the tracking step produces a significant increase in the recall rate for landmark recognition.
视觉标志的无监督选择和后验识别是移动机器人非常宝贵的感知能力。最近,我们提出了一个旨在通过将自底向上的数据驱动方法与由高级语义表示提供的自顶向下的反馈相结合来实现此功能的系统。自下而上的方法基于三个主要机制:视觉注意、区域分割和地标表征。自上而下的反馈基于两个信息源:i)对机器人位置的估计,减少与先前选择的地标的潜在匹配的搜索范围;ii)一组权重,根据先前的识别结果,控制不同分割算法在识别每个地标时的影响。在本文中,我们探讨了通过为每个选定的地标包括视觉跟踪步骤来扩展我们以前的工作的好处。我们的直觉是,包含跟踪步骤可以通过关联和选择来自其最重要视图的信息来帮助改进每个地标的模型。此外,它还可以帮助避免与选择虚假地标相关的问题。我们的研究结果证实了这些直觉,表明跟踪步骤的加入显著提高了地标识别的召回率。
{"title":"Improving the selection and detection of visual landmarks through object tracking","authors":"P. Espinace, A. Soto","doi":"10.1109/CVPRW.2008.4563133","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563133","url":null,"abstract":"The unsupervised selection and posterior recognition of visual landmarks is a highly valuable perceptual capability for a mobile robot. Recently, we proposed a system that aims to achieve this capability by combining a bottom-up data driven approach with top-down feedback provided by high level semantic representations. The bottom-up approach is based on three main mechanisms: visual attention, area segmentation, and landmark characterization. The top-down feedback is based on two information sources: i) An estimation of the robot position that reduces the searching scope for potential matches with previously selected landmarks, ii) A set of weights that, according to the results of previous recognitions, controls the influence of different segmentation algorithms in the recognition of each landmark. In this paper we explore the benefits of extending our previous work by including a visual tracking step for each of the selected landmarks. Our intuition is that the inclusion of a tracking step can help to improve the model of each landmark by associating and selecting information from its most significant views. Furthermore, it can also help to avoid problems related to the selection of spurious landmarks. Our results confirm these intuitions by showing that the inclusion of the tracking step produces a significant increase in the recall rate for landmark recognition.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"11 4","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120807545","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Design and calibration of a multi-view TOF sensor fusion system 多视点TOF传感器融合系统的设计与标定
Y. Kim, Derek Chan, C. Theobalt, S. Thrun
This paper describes the design and calibration of a system that enables simultaneous recording of dynamic scenes with multiple high-resolution video and low-resolution Swissranger time-of-flight (TOF) depth cameras. The system shall serve as a testbed for the development of new algorithms for high-quality multi-view dynamic scene reconstruction and 3D video. The paper also provides a detailed analysis of random and systematic depth camera noise which is important for reliable fusion of video and depth data. Finally, the paper describes how to compensate systematic depth errors and calibrate all dynamic depth and video data into a common frame.
本文描述了一个系统的设计和校准,该系统可以同时记录多个高分辨率视频和低分辨率Swissranger飞行时间(TOF)深度相机的动态场景。该系统将作为开发高质量多视角动态场景重建和3D视频新算法的试验台。本文还对随机和系统的深度摄像机噪声进行了详细的分析,这对视频和深度数据的可靠融合至关重要。最后,本文描述了如何补偿系统深度误差,并将所有动态深度和视频数据校准到一个公共帧中。
{"title":"Design and calibration of a multi-view TOF sensor fusion system","authors":"Y. Kim, Derek Chan, C. Theobalt, S. Thrun","doi":"10.1109/CVPRW.2008.4563160","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563160","url":null,"abstract":"This paper describes the design and calibration of a system that enables simultaneous recording of dynamic scenes with multiple high-resolution video and low-resolution Swissranger time-of-flight (TOF) depth cameras. The system shall serve as a testbed for the development of new algorithms for high-quality multi-view dynamic scene reconstruction and 3D video. The paper also provides a detailed analysis of random and systematic depth camera noise which is important for reliable fusion of video and depth data. Finally, the paper describes how to compensate systematic depth errors and calibrate all dynamic depth and video data into a common frame.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125811959","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 142
Investigating how and when perceptual organization cues improve boundary detection in natural images 研究感知组织线索如何以及何时改善自然图像的边界检测
Leandro A. Loss, G. Bebis, M. Nicolescu, A. Skurikhin
Boundary detection in natural images represents an important but also challenging problem in computer vision. Motivated by studies in psychophysics claiming that humans use multiple cues for segmentation, several promising methods have been proposed which perform boundary detection by optimally combining local image measurements such as color, texture, and brightness. Very interesting results have been reported by applying these methods on challenging datasets such as the Berkeley segmentation benchmark. Although combining different cues for boundary detection has been shown to outperform methods using a single cue, results can be further improved by integrating perceptual organization cues with the boundary detection process. The main goal of this study is to investigate how and when perceptual organization cues improve boundary detection in natural images. In this context, we investigate the idea of integrating with segmentation the iterative multi-scale tensor voting (IMSTV), a variant of tensor voting (TV) that performs perceptual grouping by analyzing information at multiple-scales and removing background clutter in an iterative fashion, preserving salient, organized structures. The key idea is to use IMSTV to post-process the boundary posterior probability (PB) map produced by segmentation algorithms. Detailed analysis of our experimental results reveals how and when perceptual organization cues are likely to improve or degrade boundary detection. In particular, we show that using perceptual grouping as a post-processing step improves boundary detection in 84% of the grayscale test images in the Berkeley segmentation dataset.
自然图像的边界检测是计算机视觉中一个重要而又具有挑战性的问题。由于心理物理学的研究表明人类使用多种线索进行分割,因此提出了几种有前途的方法,通过最佳地结合局部图像测量(如颜色、纹理和亮度)来进行边界检测。通过将这些方法应用于具有挑战性的数据集(如伯克利分割基准),已经报告了非常有趣的结果。虽然结合不同的线索进行边界检测已被证明优于使用单一线索的方法,但通过将感知组织线索与边界检测过程相结合,结果可以进一步改善。本研究的主要目的是研究感知组织线索如何以及何时改善自然图像的边界检测。在这种情况下,我们研究了将迭代多尺度张量投票(IMSTV)与分割相结合的想法,IMSTV是张量投票(TV)的一种变体,它通过在多尺度上分析信息并以迭代的方式去除背景杂波来执行感知分组,保留显著的、有组织的结构。其关键思想是利用IMSTV对分割算法产生的边界后验概率图进行后处理。对实验结果的详细分析揭示了感知组织线索如何以及何时可能改善或降低边界检测。特别是,我们表明,使用感知分组作为后处理步骤提高了伯克利分割数据集中84%的灰度测试图像的边界检测。
{"title":"Investigating how and when perceptual organization cues improve boundary detection in natural images","authors":"Leandro A. Loss, G. Bebis, M. Nicolescu, A. Skurikhin","doi":"10.1109/CVPRW.2008.4562974","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4562974","url":null,"abstract":"Boundary detection in natural images represents an important but also challenging problem in computer vision. Motivated by studies in psychophysics claiming that humans use multiple cues for segmentation, several promising methods have been proposed which perform boundary detection by optimally combining local image measurements such as color, texture, and brightness. Very interesting results have been reported by applying these methods on challenging datasets such as the Berkeley segmentation benchmark. Although combining different cues for boundary detection has been shown to outperform methods using a single cue, results can be further improved by integrating perceptual organization cues with the boundary detection process. The main goal of this study is to investigate how and when perceptual organization cues improve boundary detection in natural images. In this context, we investigate the idea of integrating with segmentation the iterative multi-scale tensor voting (IMSTV), a variant of tensor voting (TV) that performs perceptual grouping by analyzing information at multiple-scales and removing background clutter in an iterative fashion, preserving salient, organized structures. The key idea is to use IMSTV to post-process the boundary posterior probability (PB) map produced by segmentation algorithms. Detailed analysis of our experimental results reveals how and when perceptual organization cues are likely to improve or degrade boundary detection. In particular, we show that using perceptual grouping as a post-processing step improves boundary detection in 84% of the grayscale test images in the Berkeley segmentation dataset.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"54 9","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131470794","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
New insights into the calibration of ToF-sensors tof传感器校准的新见解
Marvin Lindner, A. Kolb, T. Ringbeck
Time-of-flight (ToF) sensors have become an alternative to conventional distance sensing techniques like laser scanners or image based stereo. ToF sensors provide full range distance information at high frame-rates and thus have a significant impact onto current research in areas like online object recognition, collision prevention or scene reconstruction. However, ToF cameras like the photonic mixer device (PMD) still exhibit a number of challenges regarding static and dynamic effects, e.g. systematic distance errors and motion artefacts, respectively. Sensor calibration techniques reducing static system errors have been proposed and show promising results. However, current calibration techniques in general need a large set of reference data in order to determine the corresponding parameters for the calibration model. This paper introduces a new calibration approach which combines different demodulation techniques for the ToF- camera 's reference signal. Examples show, that the resulting combined demodulation technique yields improved distance values based on only two required reference data sets.
飞行时间(ToF)传感器已经成为传统距离传感技术(如激光扫描仪或基于图像的立体)的替代方案。ToF传感器以高帧率提供全范围距离信息,因此对当前在线物体识别、碰撞预防或场景重建等领域的研究产生了重大影响。然而,像光子混频器设备(PMD)这样的ToF相机仍然表现出一些关于静态和动态效果的挑战,例如系统距离误差和运动伪影。已经提出了减少静态系统误差的传感器校准技术,并显示出良好的效果。然而,目前的校准技术通常需要大量的参考数据来确定校准模型的相应参数。本文介绍了一种结合多种解调技术对ToF摄像机参考信号进行标定的新方法。实例表明,所得到的组合解调技术仅基于两个所需的参考数据集就能产生改进的距离值。
{"title":"New insights into the calibration of ToF-sensors","authors":"Marvin Lindner, A. Kolb, T. Ringbeck","doi":"10.1109/CVPRW.2008.4563172","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563172","url":null,"abstract":"Time-of-flight (ToF) sensors have become an alternative to conventional distance sensing techniques like laser scanners or image based stereo. ToF sensors provide full range distance information at high frame-rates and thus have a significant impact onto current research in areas like online object recognition, collision prevention or scene reconstruction. However, ToF cameras like the photonic mixer device (PMD) still exhibit a number of challenges regarding static and dynamic effects, e.g. systematic distance errors and motion artefacts, respectively. Sensor calibration techniques reducing static system errors have been proposed and show promising results. However, current calibration techniques in general need a large set of reference data in order to determine the corresponding parameters for the calibration model. This paper introduces a new calibration approach which combines different demodulation techniques for the ToF- camera 's reference signal. Examples show, that the resulting combined demodulation technique yields improved distance values based on only two required reference data sets.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"29 6","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"113933669","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 36
A probabilistic representation of LiDAR range data for efficient 3D object detection 用于有效三维目标检测的激光雷达距离数据的概率表示
Theodore C. Yapo, C. Stewart, R. Radke
We present a novel approach to 3D object detection in scenes scanned by LiDAR sensors, based on a probabilistic representation of free, occupied, and hidden space that extends the concept of occupancy grids from robot mapping algorithms. This scene representation naturally handles LiDAR sampling issues, can be used to fuse multiple LiDAR data sets, and captures the inherent uncertainty of the data due to occlusions and clutter. Using this model, we formulate a hypothesis testing methodology to determine the probability that given 3D objects are present in the scene. By propagating uncertainty in the original sample points, we are able to measure confidence in the detection results in a principled way. We demonstrate the approach in examples of detecting objects that are partially occluded by scene clutter such as camouflage netting.
我们提出了一种在激光雷达传感器扫描的场景中进行3D物体检测的新方法,该方法基于空闲空间、占用空间和隐藏空间的概率表示,扩展了机器人映射算法中占用网格的概念。这种场景表示自然地处理了LiDAR采样问题,可用于融合多个LiDAR数据集,并捕获由于遮挡和杂波导致的数据固有的不确定性。使用这个模型,我们制定了一个假设检验方法来确定给定的3D物体出现在场景中的概率。通过在原始样本点中传播不确定性,我们能够以原则性的方式测量检测结果的置信度。我们在检测被场景杂波(如伪装网)部分遮挡的物体的示例中演示了该方法。
{"title":"A probabilistic representation of LiDAR range data for efficient 3D object detection","authors":"Theodore C. Yapo, C. Stewart, R. Radke","doi":"10.1109/CVPRW.2008.4563033","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563033","url":null,"abstract":"We present a novel approach to 3D object detection in scenes scanned by LiDAR sensors, based on a probabilistic representation of free, occupied, and hidden space that extends the concept of occupancy grids from robot mapping algorithms. This scene representation naturally handles LiDAR sampling issues, can be used to fuse multiple LiDAR data sets, and captures the inherent uncertainty of the data due to occlusions and clutter. Using this model, we formulate a hypothesis testing methodology to determine the probability that given 3D objects are present in the scene. By propagating uncertainty in the original sample points, we are able to measure confidence in the detection results in a principled way. We demonstrate the approach in examples of detecting objects that are partially occluded by scene clutter such as camouflage netting.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123875887","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 29
Verifying liveness by multiple experts in face biometrics 由多名面部生物识别专家验证活体
K. Kollreider, H. Fronthaler, J. Bigün
Resisting spoofing attempts via photographs and video playbacks is a vital issue for the success of face biometrics. Yet, the ldquolivenessrdquo topic has only been partially studied in the past. In this paper we are suggesting a holistic liveness detection paradigm that collaborates with standard techniques in 2D face biometrics. The experiments show that many attacks are avertible via a combination of anti-spoofing measures. We have investigated the topic using real-time techniques and applied them to real-life spoofing scenarios in an indoor, yet uncontrolled environment.
抵制通过照片和视频回放的欺骗企图是面部生物识别技术成功的关键问题。然而,过去仅对ldolivenessrquo主题进行了部分研究。在本文中,我们提出了一种整体的活体检测范式,该范式与二维面部生物识别的标准技术相结合。实验表明,许多攻击可以通过反欺骗措施的组合来避免。我们使用实时技术研究了该主题,并将其应用于室内但不受控制的环境中的真实欺骗场景。
{"title":"Verifying liveness by multiple experts in face biometrics","authors":"K. Kollreider, H. Fronthaler, J. Bigün","doi":"10.1109/CVPRW.2008.4563115","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563115","url":null,"abstract":"Resisting spoofing attempts via photographs and video playbacks is a vital issue for the success of face biometrics. Yet, the ldquolivenessrdquo topic has only been partially studied in the past. In this paper we are suggesting a holistic liveness detection paradigm that collaborates with standard techniques in 2D face biometrics. The experiments show that many attacks are avertible via a combination of anti-spoofing measures. We have investigated the topic using real-time techniques and applied them to real-life spoofing scenarios in an indoor, yet uncontrolled environment.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127709336","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 131
Scalable classifiers for Internet vision tasks 用于互联网视觉任务的可扩展分类器
Tom Yeh, John J. Lee, Trevor Darrell
Object recognition systems designed for Internet applications typically need to adapt to userspsila needs in a flexible fashion and scale up to very large data sets. In this paper, we analyze the complexity of several multiclass SVM-based algorithms and highlight the computational bottleneck they suffer at test time: comparing the input image to every training image. We propose an algorithm that overcomes this bottleneck; it offers not only the efficiency of a simple nearest-neighbor classifier, by voting on class labels based on the k nearest neighbors quickly determined by a vocabulary tree, but also the recognition accuracy comparable to that of a complex SVM classifier, by incorporating SVM parameters into the voting scores incrementally accumulated from individual image features. Empirical results demonstrate that adjusting votes by relevant support vector weights can improve the recognition accuracy of a nearest-neighbor classifier without sacrificing speed. Compared to existing methods, our algorithm achieves a ten-fold speed increase while incurring an acceptable accuracy loss that can be easily offset by showing about two more labels in the result. The speed, scalability, and adaptability of our algorithm makes it suitable for Internet vision applications.
为Internet应用程序设计的对象识别系统通常需要以灵活的方式适应用户的需求,并扩展到非常大的数据集。在本文中,我们分析了几种基于svm的多类算法的复杂性,并强调了它们在测试时遇到的计算瓶颈:将输入图像与每个训练图像进行比较。我们提出了一种克服这一瓶颈的算法;它不仅提供了简单的最近邻分类器的效率,通过基于词汇树快速确定的k个最近邻对类标签进行投票,而且通过将SVM参数纳入从单个图像特征逐渐积累的投票分数中,其识别精度可与复杂的SVM分类器相媲美。实证结果表明,通过相关支持向量权重调整投票可以在不牺牲速度的情况下提高最近邻分类器的识别精度。与现有方法相比,我们的算法实现了十倍的速度提升,同时产生了可接受的精度损失,可以通过在结果中显示大约两个标签来轻松抵消。该算法的速度、可扩展性和适应性使其适合于互联网视觉应用。
{"title":"Scalable classifiers for Internet vision tasks","authors":"Tom Yeh, John J. Lee, Trevor Darrell","doi":"10.1109/CVPRW.2008.4562958","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4562958","url":null,"abstract":"Object recognition systems designed for Internet applications typically need to adapt to userspsila needs in a flexible fashion and scale up to very large data sets. In this paper, we analyze the complexity of several multiclass SVM-based algorithms and highlight the computational bottleneck they suffer at test time: comparing the input image to every training image. We propose an algorithm that overcomes this bottleneck; it offers not only the efficiency of a simple nearest-neighbor classifier, by voting on class labels based on the k nearest neighbors quickly determined by a vocabulary tree, but also the recognition accuracy comparable to that of a complex SVM classifier, by incorporating SVM parameters into the voting scores incrementally accumulated from individual image features. Empirical results demonstrate that adjusting votes by relevant support vector weights can improve the recognition accuracy of a nearest-neighbor classifier without sacrificing speed. Compared to existing methods, our algorithm achieves a ten-fold speed increase while incurring an acceptable accuracy loss that can be easily offset by showing about two more labels in the result. The speed, scalability, and adaptability of our algorithm makes it suitable for Internet vision applications.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127797960","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
期刊
2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1