首页 > 最新文献

2012 Ninth Conference on Computer and Robot Vision最新文献

英文 中文
Framework for Natural Landmark-based Robot Localization 基于自然地标的机器人定位框架
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.25
Andrés Solís Montero, H. Sekkati, J. Lang, R. Laganière, J. James
In this paper we present a framework for vision-based robot localization using natural planar landmarks. Specifically, we demonstrate our framework with planar targets using Fern classifiers that have been shown to be robust against illumination changes, perspective distortion, motion blur, and occlusions. We add stratified sampling in the image plane to increase robustness of the localization scheme in cluttered environments and on-line checking for false detection of targets to decrease false positives. We use all matching points to improve pose estimation and an off-line target evaluation strategy to improve a priori map building. We report experiments demonstrating the accuracy and speed of localization. Our experiments entail synthetic and real data. Our framework and our improvements are however more general and the Fern classifier could be replaced by other techniques.
在本文中,我们提出了一个基于视觉的机器人定位框架,利用自然平面地标。具体来说,我们使用Fern分类器演示了我们的平面目标框架,该分类器已被证明对照明变化,透视失真,运动模糊和遮挡具有鲁棒性。我们在图像平面上增加分层采样,以提高定位方案在混乱环境下的鲁棒性,并在线检查目标的误检,以减少误报。我们使用所有匹配点来改进姿态估计,并使用离线目标评估策略来改进先验地图构建。我们报告的实验证明了定位的准确性和速度。我们的实验需要合成的和真实的数据。然而,我们的框架和改进是更通用的,Fern分类器可以被其他技术取代。
{"title":"Framework for Natural Landmark-based Robot Localization","authors":"Andrés Solís Montero, H. Sekkati, J. Lang, R. Laganière, J. James","doi":"10.1109/CRV.2012.25","DOIUrl":"https://doi.org/10.1109/CRV.2012.25","url":null,"abstract":"In this paper we present a framework for vision-based robot localization using natural planar landmarks. Specifically, we demonstrate our framework with planar targets using Fern classifiers that have been shown to be robust against illumination changes, perspective distortion, motion blur, and occlusions. We add stratified sampling in the image plane to increase robustness of the localization scheme in cluttered environments and on-line checking for false detection of targets to decrease false positives. We use all matching points to improve pose estimation and an off-line target evaluation strategy to improve a priori map building. We report experiments demonstrating the accuracy and speed of localization. Our experiments entail synthetic and real data. Our framework and our improvements are however more general and the Fern classifier could be replaced by other techniques.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129018550","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Coarse Head Pose Estimation using Image Abstraction 基于图像抽象的粗头部姿态估计
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.24
A. Puri, Hariprasad Kannan, P. Kalra
We present an algorithm to estimate the pose of a human head from a single image. It builds on the fact that only a limited set of cues are required to estimate human head pose and that most images contain far too many details than what are required for this task. Thus, non-photorealistic rendering is first used to eliminate irrelevant details from the picture and accentuate facial features critical to estimating head pose. The maximum likelihood pose range is then estimated by training a classifier on scaled down abstracted images. This algorithm covers a wide range of head orientations, can be used at various image resolutions, does not need personalized initialization, and is also relatively insensitive to illumination. Moreover, the facts that it performs competitively when compared with other state of the art methods and that it is fast enough to be used in real time systems make it a promising method for coarse head pose estimation.
我们提出了一种从单幅图像中估计人类头部姿态的算法。它建立在这样一个事实的基础上,即只需要一组有限的线索来估计人类的头部姿势,而且大多数图像包含的细节远远超过了这项任务所需的细节。因此,非真实感渲染首先用于消除图像中的不相关细节,并突出对估计头部姿势至关重要的面部特征。然后通过训练分类器在按比例缩小的抽象图像上估计最大似然姿态范围。该算法涵盖了广泛的头部方向,可以在各种图像分辨率下使用,不需要个性化初始化,并且对照明也相对不敏感。此外,与其他最先进的方法相比,它的性能具有竞争力,而且它的速度足够快,可以用于实时系统,这使它成为一种很有前途的粗略头部姿态估计方法。
{"title":"Coarse Head Pose Estimation using Image Abstraction","authors":"A. Puri, Hariprasad Kannan, P. Kalra","doi":"10.1109/CRV.2012.24","DOIUrl":"https://doi.org/10.1109/CRV.2012.24","url":null,"abstract":"We present an algorithm to estimate the pose of a human head from a single image. It builds on the fact that only a limited set of cues are required to estimate human head pose and that most images contain far too many details than what are required for this task. Thus, non-photorealistic rendering is first used to eliminate irrelevant details from the picture and accentuate facial features critical to estimating head pose. The maximum likelihood pose range is then estimated by training a classifier on scaled down abstracted images. This algorithm covers a wide range of head orientations, can be used at various image resolutions, does not need personalized initialization, and is also relatively insensitive to illumination. Moreover, the facts that it performs competitively when compared with other state of the art methods and that it is fast enough to be used in real time systems make it a promising method for coarse head pose estimation.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132506819","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
A Real Time Augmented Reality System Using GPU Acceleration 基于GPU加速的实时增强现实系统
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.21
David Chi Chung Tam, M. Fiala
Augmented Reality (AR) is an application of computer vision that is processor intensive and typically suffers from a trade-off between robust view alignment and real time performance. Real time AR that can function robustly in variable environments is a process difficult to achieve on a PC (personal computer) let alone on the mobile devices that will likely be where AR is adopted as a consumer application. Despite the availability of high quality feature matching algorithms such as SIFT, SURF and robust pose estimation algorithms such as EPNP, practical AR systems today rely on older methods such as Harris/KLT corners and template matching for performance reasons. SIFT-like algorithms are typically used only to initialize tracking by these methods. We demonstrate a practical system with real ime performance using only SURF without the need for tracking. We achieve this with extensive use of the Graphics Processing Unit (GPU) now prevalent in PC's. Due to mobile devices becoming equipped with GPU's we believe that this architecture will lead to practical robust AR.
增强现实(AR)是一种处理器密集型的计算机视觉应用,通常需要在鲁棒视图对齐和实时性能之间进行权衡。可以在可变环境中健壮地运行的实时增强现实是一个很难在PC(个人计算机)上实现的过程,更不用说在移动设备上实现了,移动设备可能是增强现实作为消费者应用程序采用的地方。尽管高质量的特征匹配算法(如SIFT, SURF)和鲁棒姿态估计算法(如EPNP)是可用的,但出于性能原因,目前的实际AR系统依赖于较旧的方法,如Harris/KLT角和模板匹配。类似sift的算法通常只用于通过这些方法初始化跟踪。我们演示了一个仅使用SURF而不需要跟踪的实时性能的实用系统。我们通过广泛使用图形处理单元(GPU)来实现这一目标。由于移动设备配备了GPU,我们相信这种架构将带来实用的强大增强现实。
{"title":"A Real Time Augmented Reality System Using GPU Acceleration","authors":"David Chi Chung Tam, M. Fiala","doi":"10.1109/CRV.2012.21","DOIUrl":"https://doi.org/10.1109/CRV.2012.21","url":null,"abstract":"Augmented Reality (AR) is an application of computer vision that is processor intensive and typically suffers from a trade-off between robust view alignment and real time performance. Real time AR that can function robustly in variable environments is a process difficult to achieve on a PC (personal computer) let alone on the mobile devices that will likely be where AR is adopted as a consumer application. Despite the availability of high quality feature matching algorithms such as SIFT, SURF and robust pose estimation algorithms such as EPNP, practical AR systems today rely on older methods such as Harris/KLT corners and template matching for performance reasons. SIFT-like algorithms are typically used only to initialize tracking by these methods. We demonstrate a practical system with real ime performance using only SURF without the need for tracking. We achieve this with extensive use of the Graphics Processing Unit (GPU) now prevalent in PC's. Due to mobile devices becoming equipped with GPU's we believe that this architecture will lead to practical robust AR.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"116 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133396600","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Perceptual Structure Distortion Ratio: An Image Quality Metric Based on Robust Measures of Complex Phase Order 感知结构失真率:基于复杂相位阶数鲁棒度量的图像质量度量
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.15
A. Wong
This paper investigates a novel perceptual driven approach to image quality assessment using complex wavelets called PSDR (Perceptual Structure Distortion Ratio). The measure is grounded in the modeling of the human vision system as a frequency-based processing system and that perceptual structural significance can be derived based on the concept of complex phase order. Built upon a robust complex phase order framework for measuring structural significance, preliminary results using test images under different types of distortions show that PSDR can be a promising direction for evaluating the quality of visual media.
本文研究了一种新的感知驱动的图像质量评估方法,该方法使用复杂小波,称为PSDR(感知结构失真比)。该方法基于将人类视觉系统建模为基于频率的处理系统,并基于复杂相位顺序的概念推导出感知结构意义。基于一个强大的复杂相序框架来测量结构显著性,使用不同类型失真的测试图像的初步结果表明,PSDR可以作为评估视觉媒体质量的一个有前途的方向。
{"title":"Perceptual Structure Distortion Ratio: An Image Quality Metric Based on Robust Measures of Complex Phase Order","authors":"A. Wong","doi":"10.1109/CRV.2012.15","DOIUrl":"https://doi.org/10.1109/CRV.2012.15","url":null,"abstract":"This paper investigates a novel perceptual driven approach to image quality assessment using complex wavelets called PSDR (Perceptual Structure Distortion Ratio). The measure is grounded in the modeling of the human vision system as a frequency-based processing system and that perceptual structural significance can be derived based on the concept of complex phase order. Built upon a robust complex phase order framework for measuring structural significance, preliminary results using test images under different types of distortions show that PSDR can be a promising direction for evaluating the quality of visual media.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134082916","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Robust Horizon Detection Using Segmentation for UAV Applications 基于分割的无人机鲁棒水平检测
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.52
Nasim Sepehri Boroujeni, S. A. Etemad, A. Whitehead
A critical step in navigation of unmanned aerial vehicles is the detection of the horizon line. This information can be used for adjusting flight parameters as well as obstacle avoidance. In this paper, a fast and robust technique for precise detection of the horizon path is proposed. The method is based on existence of a unique light field that occurs in imagery where the horizon is viewed. This light field exists in different scenes including sea-sky, soil-sky, and forest-sky horizon lines. Our proposed approach employs segmentation of the scene and subsequent analysis of the image segments for extraction of the mentioned field and thus the horizon path. Through various experiments carried out on our own dataset and that of another previously published paper, we illustrate the significance and accuracy of this technique for various types of terrains from water to ground, and even snow-covered ground. Finally, it is shown that robust performance and accuracy, speed, and extraction of the path as curves (as opposed to a straight line which is resulted from many other approaches) are the benefits of our method.
地平线探测是无人机导航的一个关键步骤。这些信息可以用于调整飞行参数以及避障。本文提出了一种快速、鲁棒的精确检测地平线路径的方法。这种方法是基于一个独特的光场的存在,这个光场出现在被观察到的地平线图像中。这个光场存在于不同的场景中,包括海-天、土-天和森林-天地平线。我们提出的方法采用场景分割和随后的图像片段分析,以提取提到的领域,从而提取地平线路径。通过对我们自己的数据集和另一篇先前发表的论文的数据集进行的各种实验,我们说明了该技术对从水到地面甚至积雪覆盖的各种类型的地形的重要性和准确性。最后,结果表明,鲁棒性和准确性、速度以及将路径提取为曲线(与许多其他方法产生的直线相反)是我们方法的优点。
{"title":"Robust Horizon Detection Using Segmentation for UAV Applications","authors":"Nasim Sepehri Boroujeni, S. A. Etemad, A. Whitehead","doi":"10.1109/CRV.2012.52","DOIUrl":"https://doi.org/10.1109/CRV.2012.52","url":null,"abstract":"A critical step in navigation of unmanned aerial vehicles is the detection of the horizon line. This information can be used for adjusting flight parameters as well as obstacle avoidance. In this paper, a fast and robust technique for precise detection of the horizon path is proposed. The method is based on existence of a unique light field that occurs in imagery where the horizon is viewed. This light field exists in different scenes including sea-sky, soil-sky, and forest-sky horizon lines. Our proposed approach employs segmentation of the scene and subsequent analysis of the image segments for extraction of the mentioned field and thus the horizon path. Through various experiments carried out on our own dataset and that of another previously published paper, we illustrate the significance and accuracy of this technique for various types of terrains from water to ground, and even snow-covered ground. Finally, it is shown that robust performance and accuracy, speed, and extraction of the path as curves (as opposed to a straight line which is resulted from many other approaches) are the benefits of our method.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124827160","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 47
A Learning Probabilistic Approach for Object Segmentation 目标分割的学习概率方法
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.19
Guillaume Larivière, M. S. Allili
This paper proposes a new method for figure-ground image segmentation based on a probabilistic learning approach of the object shape. Historically, segmentation is mostly defined as a data-driven bottom-up process, where pixels are grouped into regions/objects according to objective criteria, such as region homogeneity, etc. In particular, it aims at creating a partition of the image into contiguous, homogenous regions. In the proposed work, we propose to incorporate prior knowledge about the object shape and category to segment the object from the background. The segmentation process is composed of two parts. In the first part, object shape models are built using sets of object fragments. The second part starts by first segmenting an image into homogenous regions using the mean-shift algorithm. Then, several object hypotheses are tested and validated using the different object shape models as supporting information. As an output, our algorithm identifies the object category, position, as well as its optimal segmentation. Experimental results show the capacity of the approach to segment several object categories.
提出了一种基于物体形状概率学习的图像分割方法。从历史上看,分割大多被定义为数据驱动的自下而上的过程,其中像素根据客观标准(如区域均匀性等)分组到区域/对象中。特别是,它旨在将图像划分为连续的、同质的区域。在本文提出的工作中,我们建议结合关于物体形状和类别的先验知识来从背景中分割物体。分割过程由两部分组成。在第一部分中,使用对象碎片集构建对象形状模型。第二部分首先使用mean-shift算法将图像分割成均匀区域。然后,使用不同的物体形状模型作为支持信息,对几个物体假设进行了测试和验证。作为输出,我们的算法识别对象的类别,位置,以及其最佳分割。实验结果表明,该方法能够分割多个对象类别。
{"title":"A Learning Probabilistic Approach for Object Segmentation","authors":"Guillaume Larivière, M. S. Allili","doi":"10.1109/CRV.2012.19","DOIUrl":"https://doi.org/10.1109/CRV.2012.19","url":null,"abstract":"This paper proposes a new method for figure-ground image segmentation based on a probabilistic learning approach of the object shape. Historically, segmentation is mostly defined as a data-driven bottom-up process, where pixels are grouped into regions/objects according to objective criteria, such as region homogeneity, etc. In particular, it aims at creating a partition of the image into contiguous, homogenous regions. In the proposed work, we propose to incorporate prior knowledge about the object shape and category to segment the object from the background. The segmentation process is composed of two parts. In the first part, object shape models are built using sets of object fragments. The second part starts by first segmenting an image into homogenous regions using the mean-shift algorithm. Then, several object hypotheses are tested and validated using the different object shape models as supporting information. As an output, our algorithm identifies the object category, position, as well as its optimal segmentation. Experimental results show the capacity of the approach to segment several object categories.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115739005","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Specular-Reduced Imaging for Inspection of Machined Surfaces 用于加工表面检测的镜面降低成像
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.54
K. Sills, D. Capson, G. Bone
Specular surfaces pose difficulties for machine vision. In some applications, this may be further complicated by the presence of marks from a machining process. We propose a system that directly illuminates machined specular surfaces with a programmable array of high-power light-emitting diodes. A novel approach is described in which the angle of the incident light is varied over a series of images from which a specular-reduced median image is computed. A quality factor is used to quantitatively characterize the degree to which these specular-reduced median images approximate a diffusely lit image, and this quality factor is shown to depend linearly on the number of specular images used to produce the single specular-reduced median image. Defects such as porosity and scratches are shown to be identifiable in the specular-reduced median images of machined surfaces.
镜面给机器视觉带来了困难。在某些应用中,由于机加工过程中存在标记,这可能会进一步复杂化。我们提出了一个系统,直接照亮加工镜面与高功率发光二极管的可编程阵列。描述了一种新的方法,其中入射光的角度在一系列图像上变化,从这些图像中计算出镜面减少的中位数图像。质量因子用于定量表征这些镜面降低的中值图像近似漫射光图像的程度,并且该质量因子显示线性依赖于用于产生单个镜面降低的中值图像的镜面图像的数量。气孔和划痕等缺陷在加工表面的镜面降低的中位数图像中是可识别的。
{"title":"Specular-Reduced Imaging for Inspection of Machined Surfaces","authors":"K. Sills, D. Capson, G. Bone","doi":"10.1109/CRV.2012.54","DOIUrl":"https://doi.org/10.1109/CRV.2012.54","url":null,"abstract":"Specular surfaces pose difficulties for machine vision. In some applications, this may be further complicated by the presence of marks from a machining process. We propose a system that directly illuminates machined specular surfaces with a programmable array of high-power light-emitting diodes. A novel approach is described in which the angle of the incident light is varied over a series of images from which a specular-reduced median image is computed. A quality factor is used to quantitatively characterize the degree to which these specular-reduced median images approximate a diffusely lit image, and this quality factor is shown to depend linearly on the number of specular images used to produce the single specular-reduced median image. Defects such as porosity and scratches are shown to be identifiable in the specular-reduced median images of machined surfaces.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"119 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122083895","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Robust Body-Height Estimation for Applications in Automotive Industry 鲁棒体高估计在汽车工业中的应用
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.31
C. Scharfenberger, J. Zelek, David A Clausi
An automatic adjustment of the seat position according to the driver height significantly increases the level of comfort when entering a car. A camera attached to a vehicle can estimate the body heights of approaching drivers. However, absolute height estimation based on a single camera leads to several problems. Cost-sensitive cameras used in automotive industry provide low-resolution grayscale images, which make driver extraction in real-life parking scenarios difficult. Absolute height estimation also prerequisites a known camera position relative to a road surface, but this position is not available for any parking scenarios. Toward this, we first propose a background-based driver-extraction method that can operate on low-resolution grayscale images, and that is robust against shadows and illumination changes. Second, we derive a scheme for estimating the camera position relative to an unknown road surface using head and foot points of extracted persons. Our experimental results obtained from real-life video sequences show that the proposed schemes are highly suitable for robust driver extraction and height estimation in automotive industry.
根据驾驶员身高自动调整座椅位置,大大增加了进入汽车时的舒适度。安装在车辆上的摄像头可以估计接近司机的身高。然而,基于单个相机的绝对高度估计会导致几个问题。汽车行业中使用的成本敏感型摄像头提供的是低分辨率的灰度图像,这使得在现实停车场景中提取驾驶员信息变得困难。绝对高度估计也需要一个已知的相对于路面的摄像头位置,但这个位置不适用于任何停车场景。为此,我们首先提出了一种基于背景的驾驶员提取方法,该方法可以在低分辨率灰度图像上运行,并且对阴影和光照变化具有鲁棒性。其次,我们推导了一种利用提取的人的头和脚点来估计相机相对于未知路面的位置的方案。实际视频序列的实验结果表明,该方法非常适合于汽车行业的鲁棒驾驶员提取和高度估计。
{"title":"Robust Body-Height Estimation for Applications in Automotive Industry","authors":"C. Scharfenberger, J. Zelek, David A Clausi","doi":"10.1109/CRV.2012.31","DOIUrl":"https://doi.org/10.1109/CRV.2012.31","url":null,"abstract":"An automatic adjustment of the seat position according to the driver height significantly increases the level of comfort when entering a car. A camera attached to a vehicle can estimate the body heights of approaching drivers. However, absolute height estimation based on a single camera leads to several problems. Cost-sensitive cameras used in automotive industry provide low-resolution grayscale images, which make driver extraction in real-life parking scenarios difficult. Absolute height estimation also prerequisites a known camera position relative to a road surface, but this position is not available for any parking scenarios. Toward this, we first propose a background-based driver-extraction method that can operate on low-resolution grayscale images, and that is robust against shadows and illumination changes. Second, we derive a scheme for estimating the camera position relative to an unknown road surface using head and foot points of extracted persons. Our experimental results obtained from real-life video sequences show that the proposed schemes are highly suitable for robust driver extraction and height estimation in automotive industry.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117291036","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Planar Segmentation of RGBD Images Using Fast Linear Fitting and Markov Chain Monte Carlo 基于快速线性拟合和马尔科夫链蒙特卡罗的RGBD图像平面分割
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.12
Can Erdogan, Manohar Paluri, F. Dellaert
With the advent of affordable RGBD sensors such as the Kinect, the collection of depth and appearance information from a scene has become effortless. However, neither the correct noise model for these sensors, nor a principled methodology for extracting planar segmentations has been developed yet. In this work, we advance the state of art with the following contributions: we correctly model the Kinect sensor data by observing that the data has inherent noise only over the measured disparity values, we formulate plane fitting as a linear least-squares problem that allow us to quickly merge different segments, and we apply an advanced Markov Chain Monte Carlo (MCMC) method, generalized Swendsen-Wang sampling, to efficiently search the space of planar segmentations. We evaluate our plane fitting and surface reconstruction algorithms with simulated and real-world data.
随着诸如Kinect等价格合理的RGBD传感器的出现,从场景中收集深度和外观信息变得毫不费力。然而,这些传感器既没有正确的噪声模型,也没有提取平面分割的原则方法。在这项工作中,我们通过以下贡献推进了目前的技术水平:我们通过观察数据仅在测量的视差值上具有固有噪声来正确地建模Kinect传感器数据,我们将平面拟合定义为线性最小二乘问题,使我们能够快速合并不同的线段,并且我们应用了一种先进的马尔可夫链蒙特卡罗(MCMC)方法,广义Swendsen-Wang采样,以有效地搜索平面分割空间。我们用模拟和真实世界的数据来评估我们的平面拟合和表面重建算法。
{"title":"Planar Segmentation of RGBD Images Using Fast Linear Fitting and Markov Chain Monte Carlo","authors":"Can Erdogan, Manohar Paluri, F. Dellaert","doi":"10.1109/CRV.2012.12","DOIUrl":"https://doi.org/10.1109/CRV.2012.12","url":null,"abstract":"With the advent of affordable RGBD sensors such as the Kinect, the collection of depth and appearance information from a scene has become effortless. However, neither the correct noise model for these sensors, nor a principled methodology for extracting planar segmentations has been developed yet. In this work, we advance the state of art with the following contributions: we correctly model the Kinect sensor data by observing that the data has inherent noise only over the measured disparity values, we formulate plane fitting as a linear least-squares problem that allow us to quickly merge different segments, and we apply an advanced Markov Chain Monte Carlo (MCMC) method, generalized Swendsen-Wang sampling, to efficiently search the space of planar segmentations. We evaluate our plane fitting and surface reconstruction algorithms with simulated and real-world data.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116889898","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 49
A Virtual Vision Simulator for Camera Networks Research 面向摄像机网络的虚拟视觉模拟器研究
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.47
Wiktor Starzyk, Adam Domurad, F. Qureshi
Virtual Vision advocates developing visually and behaviorally realistic 3D synthetic environments to serve the needs of computer vision research. Virtual vision, especially, is well-suited for studying large-scale camera networks. A virtual vision simulator capable of generating "realistic" synthetic imagery from real-life scenes, involving pedestrians and other objects, is the sine qua non of carrying out virtual vision research. Here we develop a distributed, customizable virtual vision simulator capable of simulating pedestrian traffic in a variety of 3D environments. Virtual cameras deployed in this synthetic environment generate imagery using state-of-the-art computer graphics techniques, boasting realistic lighting effects, shadows, etc. The synthetic imagery is fed into a visual analysis pipeline that currently supports pedestrian detection and tracking. The results of this analysis can then be used for subsequent processing, such as camera control, coordination, and handoff. It is important to bear in mind that our visual analysis pipeline is designed to handle real world imagery without any modifications. Consequently, it closely mimics the performance of visual analysis routines that one might deploy on physical cameras. Our virtual vision simulator is realized as a collection of modules that communicate with each other over the network. Consequently, we can deploy our simulator over a network of computers, allowing us to simulate much larger camera networks and much more complex scenes then is otherwise possible.
虚拟视觉提倡开发视觉和行为逼真的三维合成环境,以满足计算机视觉研究的需要。尤其是虚拟视觉,非常适合研究大规模摄像机网络。虚拟视觉模拟器能够从包括行人和其他物体在内的现实场景中生成“逼真”的合成图像,这是开展虚拟视觉研究的必要条件。在这里,我们开发了一个分布式的、可定制的虚拟视觉模拟器,能够在各种3D环境中模拟行人交通。在这个合成环境中部署的虚拟摄像机使用最先进的计算机图形技术生成图像,具有逼真的灯光效果,阴影等。合成图像被输入到当前支持行人检测和跟踪的视觉分析管道中。分析的结果可用于后续处理,如相机控制、协调和切换。重要的是要记住,我们的视觉分析管道是为了处理真实世界的图像而设计的,没有任何修改。因此,它非常接近地模仿了可能部署在物理摄像机上的可视化分析例程的性能。我们的虚拟视觉模拟器是一个通过网络相互通信的模块集合。因此,我们可以在计算机网络上部署模拟器,使我们能够模拟更大的摄像机网络和更复杂的场景。
{"title":"A Virtual Vision Simulator for Camera Networks Research","authors":"Wiktor Starzyk, Adam Domurad, F. Qureshi","doi":"10.1109/CRV.2012.47","DOIUrl":"https://doi.org/10.1109/CRV.2012.47","url":null,"abstract":"Virtual Vision advocates developing visually and behaviorally realistic 3D synthetic environments to serve the needs of computer vision research. Virtual vision, especially, is well-suited for studying large-scale camera networks. A virtual vision simulator capable of generating \"realistic\" synthetic imagery from real-life scenes, involving pedestrians and other objects, is the sine qua non of carrying out virtual vision research. Here we develop a distributed, customizable virtual vision simulator capable of simulating pedestrian traffic in a variety of 3D environments. Virtual cameras deployed in this synthetic environment generate imagery using state-of-the-art computer graphics techniques, boasting realistic lighting effects, shadows, etc. The synthetic imagery is fed into a visual analysis pipeline that currently supports pedestrian detection and tracking. The results of this analysis can then be used for subsequent processing, such as camera control, coordination, and handoff. It is important to bear in mind that our visual analysis pipeline is designed to handle real world imagery without any modifications. Consequently, it closely mimics the performance of visual analysis routines that one might deploy on physical cameras. Our virtual vision simulator is realized as a collection of modules that communicate with each other over the network. Consequently, we can deploy our simulator over a network of computers, allowing us to simulate much larger camera networks and much more complex scenes then is otherwise possible.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125658967","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
期刊
2012 Ninth Conference on Computer and Robot Vision
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1