首页 > 最新文献

Fourth Canadian Conference on Computer and Robot Vision (CRV '07)最新文献

英文 中文
Region detection and description for Object Category Recognition 目标类别识别的区域检测与描述
Pub Date : 2007-05-28 DOI: 10.1109/CRV.2007.55
E. F. Ersi, J. Zelek
The way images are decomposed and represented biases how well subsequent object learning and recognition methods will perform. We choose to initially represent the images by sets of local distinctive regions and their description vectors. We evaluate the problems of distinctive region detection and description in two separate stages, by first reviewing some of the state-of-the-art methods, and then discussing the methods we propose to use for object category recognition. In comparing the performance of our region detection-description technique for scale and rotation invariance with the performance of the other detection-description techniques, we find that our approach provides better results than existing methods, in the context of object category recognition. The evaluation consists of clustering similar descriptor regions and computing (1) the number of single measure clusters (measures intra-class sensitivity), (2) cluster precision clusters (measures how clusters are shared between different classes) and (3) the generalizability property of regions (measures matching to classes). Our technique, which is a variant on the Kadir-Brady saliency detector scored better and not worse than all the other methods evaluated.
图像分解和表示偏差的方式,以及后续对象学习和识别方法的表现。我们选择用一组局部特征区域及其描述向量来初始表示图像。我们在两个不同的阶段评估不同区域检测和描述的问题,首先回顾一些最先进的方法,然后讨论我们建议用于对象类别识别的方法。在将我们的区域检测描述技术在尺度和旋转不变性方面的性能与其他检测描述技术的性能进行比较时,我们发现我们的方法在对象类别识别方面比现有方法提供了更好的结果。评估包括聚类相似描述子区域和计算(1)单度量聚类的数量(衡量类内敏感性),(2)聚类精度聚类(衡量不同类之间如何共享聚类)和(3)区域的可泛化性(与类匹配的度量)。我们的技术是Kadir-Brady显着性检测器的一种变体,它的得分比所有其他评估方法都要高。
{"title":"Region detection and description for Object Category Recognition","authors":"E. F. Ersi, J. Zelek","doi":"10.1109/CRV.2007.55","DOIUrl":"https://doi.org/10.1109/CRV.2007.55","url":null,"abstract":"The way images are decomposed and represented biases how well subsequent object learning and recognition methods will perform. We choose to initially represent the images by sets of local distinctive regions and their description vectors. We evaluate the problems of distinctive region detection and description in two separate stages, by first reviewing some of the state-of-the-art methods, and then discussing the methods we propose to use for object category recognition. In comparing the performance of our region detection-description technique for scale and rotation invariance with the performance of the other detection-description techniques, we find that our approach provides better results than existing methods, in the context of object category recognition. The evaluation consists of clustering similar descriptor regions and computing (1) the number of single measure clusters (measures intra-class sensitivity), (2) cluster precision clusters (measures how clusters are shared between different classes) and (3) the generalizability property of regions (measures matching to classes). Our technique, which is a variant on the Kadir-Brady saliency detector scored better and not worse than all the other methods evaluated.","PeriodicalId":304254,"journal":{"name":"Fourth Canadian Conference on Computer and Robot Vision (CRV '07)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130250654","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Robust Subspace Position Measurement Using Localized Sub-Windows 基于局部子窗口的鲁棒子空间位置测量
Pub Date : 2007-05-28 DOI: 10.1109/CRV.2007.57
A. Smit, D. Schuurman
The use of localized principal component analysis is examined for visual position determination in the presence of varying degrees of occlusions. Occlusions lead to substantial position measurement errors when projecting images into eigenspace. One way to improve robustness to occlusions is to select small sub-windows so that if some sub-windows are occluded, others can still accurately identify position. The location of candidate sub-windows are predetermined from a set of training images by subtracting the average image from each and then selecting regions using an attention operator. Since attention operators can be computationally time-intensive, the location of all sub-windows are determined a-priori during the training phase. The sub-windows in each of the training images are then projected into eigenspace. Once the training phase is complete, the run-time execution can be performed efficiently since all the sub-windows have been preselected. Input images are classified by each sub-window; majority voting is then used to determine the position estimate. Various experiments are performed including linear and rotational motion, and the ego motion of a mobile robot. This technique is shown to provide greater position measurement accuracy in the presence of severe occlusions as compared to the projection of entire images.
局部主成分分析的使用检查了在不同程度的闭塞存在的视觉位置确定。当将图像投影到特征空间时,遮挡会导致大量的位置测量误差。提高遮挡鲁棒性的一种方法是选择较小的子窗口,以便在某些子窗口被遮挡的情况下,其他子窗口仍然可以准确地识别位置。候选子窗口的位置是通过从一组训练图像中减去每个图像的平均值,然后使用注意算子选择区域来确定的。由于注意算子的计算时间很长,所以所有子窗口的位置都是在训练阶段先验确定的。然后将每个训练图像中的子窗口投影到特征空间中。一旦训练阶段完成,运行时执行可以高效地执行,因为所有的子窗口都已经预选好了。输入图像按每个子窗口进行分类;然后使用多数投票来确定位置估计。进行了各种实验,包括直线运动和旋转运动,以及移动机器人的自我运动。与整个图像的投影相比,该技术在存在严重遮挡的情况下提供了更高的位置测量精度。
{"title":"Robust Subspace Position Measurement Using Localized Sub-Windows","authors":"A. Smit, D. Schuurman","doi":"10.1109/CRV.2007.57","DOIUrl":"https://doi.org/10.1109/CRV.2007.57","url":null,"abstract":"The use of localized principal component analysis is examined for visual position determination in the presence of varying degrees of occlusions. Occlusions lead to substantial position measurement errors when projecting images into eigenspace. One way to improve robustness to occlusions is to select small sub-windows so that if some sub-windows are occluded, others can still accurately identify position. The location of candidate sub-windows are predetermined from a set of training images by subtracting the average image from each and then selecting regions using an attention operator. Since attention operators can be computationally time-intensive, the location of all sub-windows are determined a-priori during the training phase. The sub-windows in each of the training images are then projected into eigenspace. Once the training phase is complete, the run-time execution can be performed efficiently since all the sub-windows have been preselected. Input images are classified by each sub-window; majority voting is then used to determine the position estimate. Various experiments are performed including linear and rotational motion, and the ego motion of a mobile robot. This technique is shown to provide greater position measurement accuracy in the presence of severe occlusions as compared to the projection of entire images.","PeriodicalId":304254,"journal":{"name":"Fourth Canadian Conference on Computer and Robot Vision (CRV '07)","volume":"44 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122149470","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Prototype No-Reference Video Quality System 一个原型无参考视频质量系统
Pub Date : 2007-05-28 DOI: 10.1109/CRV.2007.6
R. Dosselmann, X. Yang
This paper introduces a number of innovative no-reference algorithms to assess the perceived quality of realtime analog and digital television and video streams. A prototype system is developed to locate and measure the impact of three types of impairments that commonly impair television and video signals. Analog sequences are tested for the presence of random noise. In the case of digital signals, two fundamental types of errors are of interest. The first is the blocking artifact that is so pervasive among DCT-based compression schemes such as MPEG. The second category includes errors caused by random changes to the bit stream of a signal. Of the various forms that these distortions may take on, only those that appear as "colored blocks" are detected by this system. Ideas to address the remaining issues are discussed.
本文介绍了一些创新的无参考算法来评估实时模拟和数字电视和视频流的感知质量。开发了一种原型系统,用于定位和测量通常损害电视和视频信号的三种类型的损伤的影响。模拟序列测试随机噪声的存在。在数字信号的情况下,有两种基本类型的误差值得关注。首先是在基于dct的压缩方案(如MPEG)中普遍存在的阻塞伪影。第二类包括由信号的位流随机变化引起的错误。在这些扭曲可能呈现的各种形式中,只有那些呈现为“彩色块”的形式才会被该系统检测到。讨论了解决剩余问题的想法。
{"title":"A Prototype No-Reference Video Quality System","authors":"R. Dosselmann, X. Yang","doi":"10.1109/CRV.2007.6","DOIUrl":"https://doi.org/10.1109/CRV.2007.6","url":null,"abstract":"This paper introduces a number of innovative no-reference algorithms to assess the perceived quality of realtime analog and digital television and video streams. A prototype system is developed to locate and measure the impact of three types of impairments that commonly impair television and video signals. Analog sequences are tested for the presence of random noise. In the case of digital signals, two fundamental types of errors are of interest. The first is the blocking artifact that is so pervasive among DCT-based compression schemes such as MPEG. The second category includes errors caused by random changes to the bit stream of a signal. Of the various forms that these distortions may take on, only those that appear as \"colored blocks\" are detected by this system. Ideas to address the remaining issues are discussed.","PeriodicalId":304254,"journal":{"name":"Fourth Canadian Conference on Computer and Robot Vision (CRV '07)","volume":"426 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123146796","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Computing View-normalized Body Parts Trajectories 计算视图标准化的身体部位轨迹
Pub Date : 2007-05-28 DOI: 10.1109/CRV.2007.19
F. Jean, R. Bergevin, A. Albu
This paper proposes an approach to compute view normalized body part trajectories of pedestrians from monocular video sequences. The proposed approach first extracts the 2D trajectories of both feet and of the head from tracked silhouettes. On that basis, it segments the walking trajectory into piecewise linear segments. Finally, a normalization process is applied to head and feet trajectories over each obtained straight walking segment. View normalization makes head and feet trajectories appear as if seen from a fronto-parallel viewpoint. The latter is assumed to be optimal for gait modeling and recognition purposes. The proposed approach is fully automatic as it requires neither manual initialization nor camera calibration.
提出了一种从单目视频序列中计算行人的视图归一化身体部分轨迹的方法。该方法首先从跟踪轮廓中提取足部和头部的二维轨迹。在此基础上,将行走轨迹分段为分段线性段。最后,对每个获得的直线行走段的头和脚轨迹进行归一化处理。视图归一化使头部和脚的轨迹看起来就像从正面平行的视点看到的那样。后者被认为是最优的步态建模和识别目的。所提出的方法是全自动的,因为它既不需要手动初始化也不需要相机校准。
{"title":"Computing View-normalized Body Parts Trajectories","authors":"F. Jean, R. Bergevin, A. Albu","doi":"10.1109/CRV.2007.19","DOIUrl":"https://doi.org/10.1109/CRV.2007.19","url":null,"abstract":"This paper proposes an approach to compute view normalized body part trajectories of pedestrians from monocular video sequences. The proposed approach first extracts the 2D trajectories of both feet and of the head from tracked silhouettes. On that basis, it segments the walking trajectory into piecewise linear segments. Finally, a normalization process is applied to head and feet trajectories over each obtained straight walking segment. View normalization makes head and feet trajectories appear as if seen from a fronto-parallel viewpoint. The latter is assumed to be optimal for gait modeling and recognition purposes. The proposed approach is fully automatic as it requires neither manual initialization nor camera calibration.","PeriodicalId":304254,"journal":{"name":"Fourth Canadian Conference on Computer and Robot Vision (CRV '07)","volume":"84 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127534800","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Extending Graph-Cut to Continuous Value Domain Minimization 将图切割扩展到连续值域最小化
Pub Date : 2007-05-28 DOI: 10.1109/CRV.2007.29
M. Felsberg
In this paper we propose two methods for minimizing objective functions of discrete functions with continuous value domain. Many practical problems in the area of computer vision are continuous-valued, and discrete optimization methods of graph-cut type cannot be applied directly. This is different with the proposed methods. The first method is an add-on for multiple-label graph-cut. In the second one, binary graph-cut is firstly used to generate regions of support within different ranges of the signal. Secondly, a robust error minimization is approximated based on the previously determined regions. The advantages and properties of the new approaches are explained and visualized using synthetic test data. The methods are compared to ordinary multi-label graph-cut and robust smoothing for the application of disparity estimation. They show better quality of results compared to the other approaches and the second algorithm is significantly faster than multi-label graph-cut.
本文提出了具有连续值域的离散函数目标函数的两种最小化方法。计算机视觉领域的许多实际问题都是连续值问题,不能直接应用图切型的离散优化方法。这与提议的方法不同。第一种方法是用于多标签图切割的附加组件。在第二种方法中,首先使用二值图割在信号的不同范围内生成支持区域。其次,基于先前确定的区域进行鲁棒误差最小化近似。利用综合试验数据对新方法的优点和性能进行了说明和可视化。针对视差估计的应用,将该方法与普通的多标签图切割和鲁棒平滑进行了比较。与其他方法相比,它们显示出更好的结果质量,并且第二种算法明显比多标签图切割快。
{"title":"Extending Graph-Cut to Continuous Value Domain Minimization","authors":"M. Felsberg","doi":"10.1109/CRV.2007.29","DOIUrl":"https://doi.org/10.1109/CRV.2007.29","url":null,"abstract":"In this paper we propose two methods for minimizing objective functions of discrete functions with continuous value domain. Many practical problems in the area of computer vision are continuous-valued, and discrete optimization methods of graph-cut type cannot be applied directly. This is different with the proposed methods. The first method is an add-on for multiple-label graph-cut. In the second one, binary graph-cut is firstly used to generate regions of support within different ranges of the signal. Secondly, a robust error minimization is approximated based on the previously determined regions. The advantages and properties of the new approaches are explained and visualized using synthetic test data. The methods are compared to ordinary multi-label graph-cut and robust smoothing for the application of disparity estimation. They show better quality of results compared to the other approaches and the second algorithm is significantly faster than multi-label graph-cut.","PeriodicalId":304254,"journal":{"name":"Fourth Canadian Conference on Computer and Robot Vision (CRV '07)","volume":"43 5","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132879877","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
PFAAM An Active Appearance Model based Particle Filter for both Robust and Precise Tracking PFAAM是一种基于主动外观模型的粒子滤波算法,具有鲁棒性和精确性
Pub Date : 2007-05-28 DOI: 10.1109/CRV.2007.50
S. Fleck, M. Hoffmann, K. Hunter, A. Schilling
Model based tracking is one key component of many systems today, e.g. within video surveillance or human computer interfaces (HCI). Our approach consists of a combination of particle filters (PFs) and active appearance models (AAMs): the PFAAM. It combines the robustness of PFs with the precision of AAMs. Experimental results are given. PFAAM shows superior perfomance compared to both standard AAMs and PFs using AAMs as cues only, i.e. without using a local optimization loop.
基于模型的跟踪是当今许多系统的关键组成部分,例如在视频监控或人机接口(HCI)中。我们的方法由粒子滤波器(PFs)和主动外观模型(AAMs)的组合组成:PFAAM。它结合了PFs的鲁棒性和aam的精度。给出了实验结果。与标准aam和仅使用aam作为线索(即不使用局部优化循环)的PFAAM相比,PFAAM表现出卓越的性能。
{"title":"PFAAM An Active Appearance Model based Particle Filter for both Robust and Precise Tracking","authors":"S. Fleck, M. Hoffmann, K. Hunter, A. Schilling","doi":"10.1109/CRV.2007.50","DOIUrl":"https://doi.org/10.1109/CRV.2007.50","url":null,"abstract":"Model based tracking is one key component of many systems today, e.g. within video surveillance or human computer interfaces (HCI). Our approach consists of a combination of particle filters (PFs) and active appearance models (AAMs): the PFAAM. It combines the robustness of PFs with the precision of AAMs. Experimental results are given. PFAAM shows superior perfomance compared to both standard AAMs and PFs using AAMs as cues only, i.e. without using a local optimization loop.","PeriodicalId":304254,"journal":{"name":"Fourth Canadian Conference on Computer and Robot Vision (CRV '07)","volume":"127 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133635120","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
A Robust Video Foreground Segmentation by Using Generalized Gaussian Mixture Modeling 基于广义高斯混合建模的鲁棒视频前景分割
Pub Date : 2007-05-28 DOI: 10.1109/CRV.2007.7
M. S. Allili, N. Bouguila, D. Ziou
In this paper, we propose a robust video foreground modeling by using a finite mixture model of generalized Gaussian distributions (GDD). The model has a flexibility to model the video background in the presence of sudden illumination changes and shadows, allowing for an efficient foreground segmentation. In a first part of the present work, we propose a derivation of the online estimation of the parameters of the mixture of GDDS and we propose a Bayesian approach for the selection of the number of classes. In a second part, we show experiments of video foreground segmentation demonstrating the performance of the proposed model.
本文提出了一种基于广义高斯分布(GDD)的有限混合模型的鲁棒视频前景模型。该模型具有灵活性,可以在突然照明变化和阴影的情况下对视频背景进行建模,从而实现高效的前景分割。在本工作的第一部分,我们提出了GDDS混合参数的在线估计的推导,并提出了一种贝叶斯方法来选择类别的数量。在第二部分中,我们展示了视频前景分割的实验,证明了所提出模型的性能。
{"title":"A Robust Video Foreground Segmentation by Using Generalized Gaussian Mixture Modeling","authors":"M. S. Allili, N. Bouguila, D. Ziou","doi":"10.1109/CRV.2007.7","DOIUrl":"https://doi.org/10.1109/CRV.2007.7","url":null,"abstract":"In this paper, we propose a robust video foreground modeling by using a finite mixture model of generalized Gaussian distributions (GDD). The model has a flexibility to model the video background in the presence of sudden illumination changes and shadows, allowing for an efficient foreground segmentation. In a first part of the present work, we propose a derivation of the online estimation of the parameters of the mixture of GDDS and we propose a Bayesian approach for the selection of the number of classes. In a second part, we show experiments of video foreground segmentation demonstrating the performance of the proposed model.","PeriodicalId":304254,"journal":{"name":"Fourth Canadian Conference on Computer and Robot Vision (CRV '07)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131358917","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 85
Super-resolution based on interpolation and global sub pixel translation 基于插值和全局亚像素平移的超分辨率
Pub Date : 2007-05-28 DOI: 10.1109/CRV.2007.62
Kamel Mecheri, D. Ziou, F. Deschênes
In this paper we present a new class of reconstruction algorithms that are basically different from the traditional approaches. We deviate from the traditional technique which treats the pixels of the image as point samples. In this work, the pixels are treated as rectangular surface samples. It is in conformity with image formation process, in particular for CCD/CMOS sensors, which are a matrix of rectangular surfaces sensitive to the light. We show that results of better quality in terms of the measurements employed are obtained by formulating the reconstruction as a two-stage process: the restoration of image followed by the application of the point spread function (PSF) of the imaging sensor. By coupling the PSF with the reconstruction process, we satisfy a measure of accuracy that is based on the physical limitations of the sensor. Effective techniques for the restoration of image are derived to invert the effects of the PSF and estimate the original image. For the algorithm of restoration, we introduce a new method of interpolation implying a sequence of images, not necessarily a temporal sequence, shifted compared to an image of reference.
在本文中,我们提出了一种新的重建算法,它与传统的方法有本质的不同。我们偏离了将图像的像素作为点样本的传统技术。在这项工作中,像素被视为矩形表面样本。它符合图像形成过程,特别是对于CCD/CMOS传感器,它是一个对光敏感的矩形表面的矩阵。我们表明,就所采用的测量结果而言,通过将重建表述为两个阶段的过程获得了更好的质量:图像恢复,然后是成像传感器的点扩散函数(PSF)的应用。通过将PSF与重建过程耦合,我们满足了基于传感器物理限制的精度测量。提出了一种有效的图像恢复技术来反演PSF的影响并估计原始图像。对于恢复算法,我们引入了一种新的插值方法,这意味着与参考图像相比,图像序列(不一定是时间序列)发生了移位。
{"title":"Super-resolution based on interpolation and global sub pixel translation","authors":"Kamel Mecheri, D. Ziou, F. Deschênes","doi":"10.1109/CRV.2007.62","DOIUrl":"https://doi.org/10.1109/CRV.2007.62","url":null,"abstract":"In this paper we present a new class of reconstruction algorithms that are basically different from the traditional approaches. We deviate from the traditional technique which treats the pixels of the image as point samples. In this work, the pixels are treated as rectangular surface samples. It is in conformity with image formation process, in particular for CCD/CMOS sensors, which are a matrix of rectangular surfaces sensitive to the light. We show that results of better quality in terms of the measurements employed are obtained by formulating the reconstruction as a two-stage process: the restoration of image followed by the application of the point spread function (PSF) of the imaging sensor. By coupling the PSF with the reconstruction process, we satisfy a measure of accuracy that is based on the physical limitations of the sensor. Effective techniques for the restoration of image are derived to invert the effects of the PSF and estimate the original image. For the algorithm of restoration, we introduce a new method of interpolation implying a sequence of images, not necessarily a temporal sequence, shifted compared to an image of reference.","PeriodicalId":304254,"journal":{"name":"Fourth Canadian Conference on Computer and Robot Vision (CRV '07)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133530429","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Identification and Recognition of Objects in Color Stereo Images Using a Hierachial SOM 基于层次SOM的彩色立体图像中物体的识别
Pub Date : 2007-05-28 DOI: 10.1109/CRV.2007.39
G. Bertolini, S. Ramat
Identification and recognition of objects in digital images is a fundamental task in robotic vision. Here we propose an approach based on clustering of feature extracted from HSV color space and depth, using a hierarchical self organizing map (HSOM). Binocular images are first preprocessed using a watershed algorithm; adjacent regions are then merged based on HSV similarities. For each region we compute a six element features vector: median depth (computed as disparity), median H, S, V values, and the X and Y coordinates of its centroid. These are the input to the HSOM network which is allowed to learn on the first image of a sequence. The trained network is then used to segment other images of the same scene. If, on the new image, the same neuron responds to regions that belong to the same object, the object is considered as recognized. The technique achieves good results, recognizing up to 82% of the objects.
数字图像中物体的识别是机器人视觉的一项基本任务。本文提出了一种基于HSV颜色空间和深度提取特征聚类的方法,使用层次自组织映射(HSOM)。首先使用分水岭算法对双目图像进行预处理;然后根据HSV相似度合并相邻区域。对于每个区域,我们计算六个元素特征向量:中位数深度(以视差计算),中位数H, S, V值以及其质心的X和Y坐标。这些是HSOM网络的输入,它可以在序列的第一张图像上学习。然后使用训练好的网络来分割同一场景的其他图像。如果在新图像上,相同的神经元对属于同一物体的区域做出反应,则认为该物体已被识别。该技术取得了良好的效果,识别了高达82%的物体。
{"title":"Identification and Recognition of Objects in Color Stereo Images Using a Hierachial SOM","authors":"G. Bertolini, S. Ramat","doi":"10.1109/CRV.2007.39","DOIUrl":"https://doi.org/10.1109/CRV.2007.39","url":null,"abstract":"Identification and recognition of objects in digital images is a fundamental task in robotic vision. Here we propose an approach based on clustering of feature extracted from HSV color space and depth, using a hierarchical self organizing map (HSOM). Binocular images are first preprocessed using a watershed algorithm; adjacent regions are then merged based on HSV similarities. For each region we compute a six element features vector: median depth (computed as disparity), median H, S, V values, and the X and Y coordinates of its centroid. These are the input to the HSOM network which is allowed to learn on the first image of a sequence. The trained network is then used to segment other images of the same scene. If, on the new image, the same neuron responds to regions that belong to the same object, the object is considered as recognized. The technique achieves good results, recognizing up to 82% of the objects.","PeriodicalId":304254,"journal":{"name":"Fourth Canadian Conference on Computer and Robot Vision (CRV '07)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114115237","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Using a Raster Display for Photometric Stereo 使用光栅显示光度立体
Pub Date : 2007-05-28 DOI: 10.1109/CRV.2007.66
N. Funk, Herbert Yang
This paper presents a new controlled lighting apparatus which uses a raster display device as a light source. The setup has the advantage over other alternatives in that it is relatively inexpensive and uses commonly available components. The apparatus is studied through application to shape recovery using photometric stereo. Experiments on synthetic and real images demonstrate how the depth map of an object can be recovered using only a camera and a computer monitor.
本文介绍了一种以光栅显示器件为光源的可控照明装置。与其他替代方案相比,该设置具有优势,因为它相对便宜,并且使用常见的组件。通过在光度立体形状恢复中的应用,对该装置进行了研究。在合成图像和真实图像上的实验表明,仅使用相机和计算机显示器就可以恢复物体的深度图。
{"title":"Using a Raster Display for Photometric Stereo","authors":"N. Funk, Herbert Yang","doi":"10.1109/CRV.2007.66","DOIUrl":"https://doi.org/10.1109/CRV.2007.66","url":null,"abstract":"This paper presents a new controlled lighting apparatus which uses a raster display device as a light source. The setup has the advantage over other alternatives in that it is relatively inexpensive and uses commonly available components. The apparatus is studied through application to shape recovery using photometric stereo. Experiments on synthetic and real images demonstrate how the depth map of an object can be recovered using only a camera and a computer monitor.","PeriodicalId":304254,"journal":{"name":"Fourth Canadian Conference on Computer and Robot Vision (CRV '07)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128947996","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 34
期刊
Fourth Canadian Conference on Computer and Robot Vision (CRV '07)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1