首页 > 最新文献

The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)最新文献

英文 中文
A Novel Clustering-Based Method for Adaptive Background Segmentation 一种新的基于聚类的自适应背景分割方法
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.5
S. Indupalli, M. Ali, B. Boufama
This paper presents a new histogram-based method for dynamic background modeling using a sequence of images extracted from video. In particular, a k-means clustering technique has been used to identify the foreground objects. Because of its shadow resistance and discriminative properties, we have used images in the HSV color space instead of the traditional RGB color space. The experimental results on real images are very encouraging as we were able to retrieve perfect backgrounds in simple scenes. In very complex scenes, the backgrounds we have obtained were very good. Furthermore, our method is very fast and could be used in real-time applications after optimization.
本文提出了一种基于直方图的动态背景建模新方法,该方法利用从视频中提取的图像序列进行动态背景建模。特别地,使用k-means聚类技术来识别前景对象。由于HSV具有抗阴影和判别的特性,我们使用HSV色彩空间中的图像来代替传统的RGB色彩空间。在真实图像上的实验结果非常令人鼓舞,因为我们能够在简单的场景中检索到完美的背景。在非常复杂的场景中,我们获得的背景非常好。此外,我们的方法速度非常快,经过优化后可以用于实时应用。
{"title":"A Novel Clustering-Based Method for Adaptive Background Segmentation","authors":"S. Indupalli, M. Ali, B. Boufama","doi":"10.1109/CRV.2006.5","DOIUrl":"https://doi.org/10.1109/CRV.2006.5","url":null,"abstract":"This paper presents a new histogram-based method for dynamic background modeling using a sequence of images extracted from video. In particular, a k-means clustering technique has been used to identify the foreground objects. Because of its shadow resistance and discriminative properties, we have used images in the HSV color space instead of the traditional RGB color space. The experimental results on real images are very encouraging as we were able to retrieve perfect backgrounds in simple scenes. In very complex scenes, the backgrounds we have obtained were very good. Furthermore, our method is very fast and could be used in real-time applications after optimization.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128754504","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Reflection Stereo - Novel Monocular Stereo using a Transparent Plate - 反射立体——利用透明板的新型单目立体
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.59
M. Shimizu, M. Okutomi
This paper proposes a simple and novel singlecamera depth estimation method using images reflected by a single transparent parallel planar plate. The transparent plate reflects and transmits the incident light on its surface. The transmitted light is then reflected on the rear-surface and is transmitted again to the air through the surface. These two light paths create an overlapped image that comprises two shifted images. The overlapped image is considered as a stereo image obtained from a narrow baseline stereo. The constraint of these stereo images is presented. The distance to the object can be derived by finding correspondences on the constraint lines using the autocorrelation function of the overlapped image. This paper presents experimental results obtained using an actual system with a transparent acrylic plate.
本文提出了一种简单、新颖的单相机深度估计方法,该方法利用单个透明平行平面反射的图像进行深度估计。透明板将入射光反射并透射到其表面。透射的光被反射到后表面,并通过表面再次透射到空气中。这两条光路产生了重叠的图像,包括两个移位的图像。将重叠图像视为由窄基线立体图像获得的立体图像。给出了这些立体图像的约束条件。利用重叠图像的自相关函数找到约束线上的对应关系,从而得到与目标的距离。本文介绍了在透明亚克力板的实际系统中获得的实验结果。
{"title":"Reflection Stereo - Novel Monocular Stereo using a Transparent Plate -","authors":"M. Shimizu, M. Okutomi","doi":"10.1109/CRV.2006.59","DOIUrl":"https://doi.org/10.1109/CRV.2006.59","url":null,"abstract":"This paper proposes a simple and novel singlecamera depth estimation method using images reflected by a single transparent parallel planar plate. The transparent plate reflects and transmits the incident light on its surface. The transmitted light is then reflected on the rear-surface and is transmitted again to the air through the surface. These two light paths create an overlapped image that comprises two shifted images. The overlapped image is considered as a stereo image obtained from a narrow baseline stereo. The constraint of these stereo images is presented. The distance to the object can be derived by finding correspondences on the constraint lines using the autocorrelation function of the overlapped image. This paper presents experimental results obtained using an actual system with a transparent acrylic plate.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126771841","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Image Inpainting and Segmentation using Hierarchical Level Set Method 基于层次水平集方法的图像绘制与分割
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.41
Xiaojun Du, D. Cho, T. D. Bui
Image inpainting is an artistic procedure to recover a damaged painting or picture. In this paper, we propose a novel approach for image inpainting. In this approach, the Mumford-Shah (MS) model and the level set method are employed to estimate image structure of the damaged region. This approach has been successfully used in image segmentation problem. Compared to some other inpainting methods, the MS model approach can detect and preserve edges in the inpainting areas. We propose in this paper a fast and efficient algorithm which can achieve both inpainting and segmentation. In previous works on the MS model, only one or two level set functions are used to segment an image. While this approach works well on some simple images, detailed edges cannot be detected on complicated images. Although multi-level set functions can be used to segment an image into many regions, the traditional approach causes extensive computations and the solutions depend on the location of the initial curves. Our proposed approach utilizes faster hierarchical level set method and can guarantee convergence independent of initial conditions. Because we can detect both the main structure and the detailed edges, the approach can preserve detailed edges in the inpainting area. Experimental results demonstrate the advantage of our method.
图像修复是修复受损绘画或图片的艺术过程。在本文中,我们提出了一种新的图像绘制方法。该方法采用Mumford-Shah (MS)模型和水平集方法对受损区域的图像结构进行估计。该方法已成功应用于图像分割问题。与其他图像补图方法相比,MS模型方法可以在补图区域检测和保持边缘。本文提出了一种快速有效的算法,可以同时实现图像的绘制和分割。在以往关于MS模型的工作中,只使用一个或两个水平集函数来分割图像。虽然这种方法在一些简单的图像上效果很好,但在复杂的图像上无法检测到详细的边缘。虽然多级集合函数可以将图像分割成多个区域,但传统的方法需要大量的计算,而且解依赖于初始曲线的位置。该方法采用了更快的层次水平集方法,并能保证不受初始条件影响的收敛性。由于该方法既可以检测到主体结构,又可以检测到细节边缘,因此可以在涂漆区域保留细节边缘。实验结果证明了该方法的优越性。
{"title":"Image Inpainting and Segmentation using Hierarchical Level Set Method","authors":"Xiaojun Du, D. Cho, T. D. Bui","doi":"10.1109/CRV.2006.41","DOIUrl":"https://doi.org/10.1109/CRV.2006.41","url":null,"abstract":"Image inpainting is an artistic procedure to recover a damaged painting or picture. In this paper, we propose a novel approach for image inpainting. In this approach, the Mumford-Shah (MS) model and the level set method are employed to estimate image structure of the damaged region. This approach has been successfully used in image segmentation problem. Compared to some other inpainting methods, the MS model approach can detect and preserve edges in the inpainting areas. We propose in this paper a fast and efficient algorithm which can achieve both inpainting and segmentation. In previous works on the MS model, only one or two level set functions are used to segment an image. While this approach works well on some simple images, detailed edges cannot be detected on complicated images. Although multi-level set functions can be used to segment an image into many regions, the traditional approach causes extensive computations and the solutions depend on the location of the initial curves. Our proposed approach utilizes faster hierarchical level set method and can guarantee convergence independent of initial conditions. Because we can detect both the main structure and the detailed edges, the approach can preserve detailed edges in the inpainting area. Experimental results demonstrate the advantage of our method.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"235 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123287255","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
The McGill Object Detection Suite 麦吉尔目标检测套件
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.75
Donovan H. Parks, M. Levine
Evaluation of object detection systems requires a set of test images with objects in heterogeneous scenes. Unfortunately, existing publicly available object databases provide few, if any, test images suitable for evaluating object detection systems. Here we present the McGill Object Detection Suite (MODS), a software package for creating test sets suitable for evaluating object detection systems. These test sets are created by superimposing objects from existing publicly available object databases onto heterogeneous backgrounds. The MODS is capable of creating test sets focusing on pose, scale, illumination, occlusion, or noise. This software package is being made publicly available to aid the computer vision community by providing standard test sets which will allow object detection systems to be systematically compared and characterized.
评估目标检测系统需要一组具有异构场景中目标的测试图像。不幸的是,现有的公开可用的对象数据库提供很少(如果有的话)适合评估对象检测系统的测试图像。在这里,我们介绍了McGill目标检测套件(MODS),一个用于创建适合评估目标检测系统的测试集的软件包。这些测试集是通过将现有的公共对象数据库中的对象叠加到异构背景上创建的。MODS能够创建专注于姿势,比例,照明,遮挡或噪声的测试集。这个软件包正在向公众开放,通过提供标准测试集来帮助计算机视觉社区,这些测试集将允许对目标检测系统进行系统的比较和表征。
{"title":"The McGill Object Detection Suite","authors":"Donovan H. Parks, M. Levine","doi":"10.1109/CRV.2006.75","DOIUrl":"https://doi.org/10.1109/CRV.2006.75","url":null,"abstract":"Evaluation of object detection systems requires a set of test images with objects in heterogeneous scenes. Unfortunately, existing publicly available object databases provide few, if any, test images suitable for evaluating object detection systems. Here we present the McGill Object Detection Suite (MODS), a software package for creating test sets suitable for evaluating object detection systems. These test sets are created by superimposing objects from existing publicly available object databases onto heterogeneous backgrounds. The MODS is capable of creating test sets focusing on pose, scale, illumination, occlusion, or noise. This software package is being made publicly available to aid the computer vision community by providing standard test sets which will allow object detection systems to be systematically compared and characterized.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"686 ","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120886960","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Building Local Safety Maps for a Wheelchair Robot using Vision and Lasers 利用视觉和激光为轮椅机器人构建局部安全地图
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.20
A. Murarka, Joseph Modayil, B. Kuipers
To be useful as a mobility assistant for a human driver, an intelligent robotic wheelchair must be able to distinguish between safe and hazardous regions in its immediate environment. We present a hybrid method using laser rangefinders and vision for building local 2D metrical maps that incorporate safety information (called local safety maps). Laser range-finders are used for localization and mapping of obstacles in the 2D laser plane, and vision is used for detection of hazards and other obstacles in 3D space. The hazards and obstacles identified by vision are projected into the travel plane of the robot and combined with the laser map to construct the local 2D safety map. The main contributions of this work are (i) the definition of a local 2D safety map, (ii) a hybrid method for building the safety map, and (iii) a method for removing noise from dense stereo data using motion.
为了成为人类驾驶员的移动助手,智能机器人轮椅必须能够区分其周围环境中的安全区域和危险区域。我们提出了一种使用激光测距仪和视觉的混合方法,用于构建包含安全信息的局部二维测量地图(称为局部安全地图)。激光测距仪用于二维激光平面上障碍物的定位和测绘,视觉用于三维空间中危险和其他障碍物的检测。将视觉识别的危险和障碍物投影到机器人的行走平面上,并与激光地图相结合,构建局部二维安全地图。这项工作的主要贡献是(i)局部二维安全地图的定义,(ii)构建安全地图的混合方法,以及(iii)使用运动从密集立体数据中去除噪声的方法。
{"title":"Building Local Safety Maps for a Wheelchair Robot using Vision and Lasers","authors":"A. Murarka, Joseph Modayil, B. Kuipers","doi":"10.1109/CRV.2006.20","DOIUrl":"https://doi.org/10.1109/CRV.2006.20","url":null,"abstract":"To be useful as a mobility assistant for a human driver, an intelligent robotic wheelchair must be able to distinguish between safe and hazardous regions in its immediate environment. We present a hybrid method using laser rangefinders and vision for building local 2D metrical maps that incorporate safety information (called local safety maps). Laser range-finders are used for localization and mapping of obstacles in the 2D laser plane, and vision is used for detection of hazards and other obstacles in 3D space. The hazards and obstacles identified by vision are projected into the travel plane of the robot and combined with the laser map to construct the local 2D safety map. The main contributions of this work are (i) the definition of a local 2D safety map, (ii) a hybrid method for building the safety map, and (iii) a method for removing noise from dense stereo data using motion.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133575278","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 43
An Edge Preserving Locally Adaptive Anti-aliasing Zooming Algorithm with Diffused Interpolation 一种带有扩散插值的保边局部自适应抗混叠缩放算法
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.8
Munib Arshad Chughtai, N. Khattak
In this paper the problem of producing an enlarged image from a given digital image is addressed (zooming). Different image interpolation techniques are used for image enlargement. During interpolation, preserving details and smoothing data at the same time for not introducing spurious artifacts (i.e. Aliasing) is difficult. A complete and a definitive solution to this problem is still an open issue. Although there are some well known methods in the market Parket [14], Sakamote [16], the paper proposes a method that considers discontinuities and luminance variations in a sequence of non linear iterations steps. All the pixels present near the edges are diffused into the edge in a way that aliasing is reduced to a greater extent. Hence the proposed method is completed in limited computational resources. The proposed method preserves edges and brings smoothness and at the same time controls the aliasing effect.
本文讨论了从给定的数字图像产生放大图像的问题(缩放)。不同的图像插值技术用于图像放大。在插值过程中,为了不引入虚假伪影(即混叠)而同时保留细节和平滑数据是困难的。对这一问题的全面和明确的解决办法仍然是一个悬而未决的问题。虽然市场上已有一些知名的方法Parket [14], Sakamote[16],但本文提出了一种考虑非线性迭代步骤序列不连续和亮度变化的方法。在边缘附近存在的所有像素都以一种更大程度上减少混叠的方式扩散到边缘。因此,该方法可以在有限的计算资源下完成。该方法既保留了图像的边缘,又保证了图像的平滑性,同时控制了混叠效应。
{"title":"An Edge Preserving Locally Adaptive Anti-aliasing Zooming Algorithm with Diffused Interpolation","authors":"Munib Arshad Chughtai, N. Khattak","doi":"10.1109/CRV.2006.8","DOIUrl":"https://doi.org/10.1109/CRV.2006.8","url":null,"abstract":"In this paper the problem of producing an enlarged image from a given digital image is addressed (zooming). Different image interpolation techniques are used for image enlargement. During interpolation, preserving details and smoothing data at the same time for not introducing spurious artifacts (i.e. Aliasing) is difficult. A complete and a definitive solution to this problem is still an open issue. Although there are some well known methods in the market Parket [14], Sakamote [16], the paper proposes a method that considers discontinuities and luminance variations in a sequence of non linear iterations steps. All the pixels present near the edges are diffused into the edge in a way that aliasing is reduced to a greater extent. Hence the proposed method is completed in limited computational resources. The proposed method preserves edges and brings smoothness and at the same time controls the aliasing effect.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130023872","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
3D Face Reconstruction from Stereo Video 从立体视频三维人脸重建
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.1
U. Park, Anil K. Jain
Face processing in video is receiving substantial attention due to its importance in many securityrelated applications. A video provides rich information about a face (multiple frames and temporal coherence) that can be utilized in conjunction with 3D face models, if available, to establish a subject’s identity. We propose a 3D face modeling method that reconstructs a user-specific model derived from a generic 3D face model and two video frames of the user. The user-specific 3D face model can be enrolled into the 3D face database at the enrollment stage to be used in later identification process. The reconstruction process can also be used for the probe data in recognition stage, where the reconstructed 3D face model using probe face is used to generate an optimal view and lighting for the recognition process. The advantage of utilizing reconstructed 3D face model is demonstrated by conducting face recognition experiments for 15 probe subjects against a gallery database containing 100 subjects.
视频中的人脸处理由于其在许多安全相关应用中的重要性而受到广泛关注。视频提供了关于人脸的丰富信息(多帧和时间相干性),如果可用,可以与3D人脸模型结合使用,以确定受试者的身份。我们提出了一种三维人脸建模方法,该方法从一个通用的三维人脸模型和用户的两个视频帧中重建一个用户特定的模型。用户特定的3D人脸模型可以在注册阶段注册到3D人脸数据库中,以供后期识别过程使用。该重建过程也可用于识别阶段的探针数据,其中使用探针面部重建的三维人脸模型用于生成识别过程的最佳视图和照明。通过对包含100个受试者的图库数据库进行15个探针受试者的人脸识别实验,证明了利用重建的三维人脸模型的优势。
{"title":"3D Face Reconstruction from Stereo Video","authors":"U. Park, Anil K. Jain","doi":"10.1109/CRV.2006.1","DOIUrl":"https://doi.org/10.1109/CRV.2006.1","url":null,"abstract":"Face processing in video is receiving substantial attention due to its importance in many securityrelated applications. A video provides rich information about a face (multiple frames and temporal coherence) that can be utilized in conjunction with 3D face models, if available, to establish a subject’s identity. We propose a 3D face modeling method that reconstructs a user-specific model derived from a generic 3D face model and two video frames of the user. The user-specific 3D face model can be enrolled into the 3D face database at the enrollment stage to be used in later identification process. The reconstruction process can also be used for the probe data in recognition stage, where the reconstructed 3D face model using probe face is used to generate an optimal view and lighting for the recognition process. The advantage of utilizing reconstructed 3D face model is demonstrated by conducting face recognition experiments for 15 probe subjects against a gallery database containing 100 subjects.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121999123","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Local Stereo Matching with Segmentation-based Outlier Rejection 基于分割的离群点抑制的局部立体匹配
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.49
M. Gerrits, P. Bekaert
We present a new window-based stereo matching algorithm which focuses on robust outlier rejection during aggregation. The main difficulty for window-based methods lies in determining the best window shape and size for each pixel. Working from the assumption that depth discontinuities occur at colour boundaries, we segment the reference image and consider all window pixels outside the image segment that contains the pixel under consideration as outliers and greatly reduce their weight in the aggregation process. We developed a variation on the recursive moving average implementation to keep processing times independent from window size. Together with a robust matching cost and the combination of the left and right disparity maps, this gives us a robust local algorithm that approximates the quality of global techniques without sacrificing the speed and simplicity of window-based aggregation.
提出了一种新的基于窗口的立体匹配算法,该算法的重点是在聚合过程中对异常值的鲁棒抑制。基于窗口的方法的主要困难在于确定每个像素的最佳窗口形状和大小。基于深度不连续发生在颜色边界的假设,我们对参考图像进行分割,并将包含所考虑像素的图像分割之外的所有窗口像素视为异常值,并在聚合过程中大大降低其权重。我们开发了一种递归移动平均实现的变体,以保持处理时间与窗口大小无关。结合鲁棒匹配成本和左右视差图的组合,这为我们提供了一个鲁棒的局部算法,它接近全局技术的质量,而不会牺牲基于窗口的聚合的速度和简单性。
{"title":"Local Stereo Matching with Segmentation-based Outlier Rejection","authors":"M. Gerrits, P. Bekaert","doi":"10.1109/CRV.2006.49","DOIUrl":"https://doi.org/10.1109/CRV.2006.49","url":null,"abstract":"We present a new window-based stereo matching algorithm which focuses on robust outlier rejection during aggregation. The main difficulty for window-based methods lies in determining the best window shape and size for each pixel. Working from the assumption that depth discontinuities occur at colour boundaries, we segment the reference image and consider all window pixels outside the image segment that contains the pixel under consideration as outliers and greatly reduce their weight in the aggregation process. We developed a variation on the recursive moving average implementation to keep processing times independent from window size. Together with a robust matching cost and the combination of the left and right disparity maps, this gives us a robust local algorithm that approximates the quality of global techniques without sacrificing the speed and simplicity of window-based aggregation.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"94 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124172528","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 130
A Pixel-Weighting Method for Discriminating Objects of Different Sizes in an Image Captured from a Single Camera 一种像素加权方法在单个相机捕获的图像中区分不同大小的目标
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.6
Mook-Kwang Park, Namsu Moon, Sang-Gyu Ryu, Jeongpyo Kong, Yongjin Lee, Wangjin Mun
A novel method of pixel-weighting is proposed to calculate the size of a detected object in an image captured using a single camera. The calculated object size does not vary significantly regardless of the location of the object in an image, which allows it to be effectively utilized in a vision-based surveillance sensing algorithm as a meaningful feature for discriminating human intruders from other objects. Experimental results show the feasibility of the proposed method.
提出了一种新的像素加权方法来计算单相机图像中被检测物体的大小。无论物体在图像中的位置如何,计算出的物体大小都不会发生显着变化,这使得它可以有效地用于基于视觉的监视传感算法,作为区分人类入侵者和其他物体的有意义的特征。实验结果表明了该方法的可行性。
{"title":"A Pixel-Weighting Method for Discriminating Objects of Different Sizes in an Image Captured from a Single Camera","authors":"Mook-Kwang Park, Namsu Moon, Sang-Gyu Ryu, Jeongpyo Kong, Yongjin Lee, Wangjin Mun","doi":"10.1109/CRV.2006.6","DOIUrl":"https://doi.org/10.1109/CRV.2006.6","url":null,"abstract":"A novel method of pixel-weighting is proposed to calculate the size of a detected object in an image captured using a single camera. The calculated object size does not vary significantly regardless of the location of the object in an image, which allows it to be effectively utilized in a vision-based surveillance sensing algorithm as a meaningful feature for discriminating human intruders from other objects. Experimental results show the feasibility of the proposed method.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132479162","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Autonomous Learning of Object Appearances using Colour Contour Frames 基于颜色轮廓框架的物体外观自主学习
Pub Date : 2006-06-07 DOI: 10.1109/CRV.2006.17
Per-Erik Forssén, A. Moe
In this paper we make use of the idea that a robot can autonomously discover objects and learn their appearances by poking and prodding at interesting parts of a scene. In order to make the resultant object recognition ability more robust, and discriminative, we replace earlier used colour histogram features with an invariant texture-patch method. The texture patches are extracted in a similarity invariant frame which is constructed from short colour contour segments. We demonstrate the robustness of our invariant frames with a repeatability test under general homography transformations of a planar scene. Through the repeatability test, we find that defining the frame using using ellipse segments instead of lines where this is appropriate improves repeatability. We also apply the developed features to autonomous learning of object appearances, and show how the learned objects can be recognised under out-of-plane rotation and scale changes.
在本文中,我们利用了机器人可以自主发现物体的想法,并通过戳戳场景中有趣的部分来学习它们的外观。为了提高目标识别的鲁棒性和区别性,我们用不变的纹理补丁方法代替了之前使用的颜色直方图特征。在由短彩色轮廓段构成的相似不变框架中提取纹理块。在平面场景的一般单应变换下,我们用可重复性检验证明了不变帧的鲁棒性。通过可重复性测试,我们发现使用椭圆段而不是线条来定义框架可以提高可重复性。我们还将开发的特征应用于物体外观的自主学习,并展示了如何在面外旋转和尺度变化的情况下识别学习到的物体。
{"title":"Autonomous Learning of Object Appearances using Colour Contour Frames","authors":"Per-Erik Forssén, A. Moe","doi":"10.1109/CRV.2006.17","DOIUrl":"https://doi.org/10.1109/CRV.2006.17","url":null,"abstract":"In this paper we make use of the idea that a robot can autonomously discover objects and learn their appearances by poking and prodding at interesting parts of a scene. In order to make the resultant object recognition ability more robust, and discriminative, we replace earlier used colour histogram features with an invariant texture-patch method. The texture patches are extracted in a similarity invariant frame which is constructed from short colour contour segments. We demonstrate the robustness of our invariant frames with a repeatability test under general homography transformations of a planar scene. Through the repeatability test, we find that defining the frame using using ellipse segments instead of lines where this is appropriate improves repeatability. We also apply the developed features to autonomous learning of object appearances, and show how the learned objects can be recognised under out-of-plane rotation and scale changes.","PeriodicalId":369170,"journal":{"name":"The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131128131","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
期刊
The 3rd Canadian Conference on Computer and Robot Vision (CRV'06)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1