首页 > 最新文献

Fourth Canadian Conference on Computer and Robot Vision (CRV '07)最新文献

英文 中文
Camera Sensor Model for Visual SLAM 视觉SLAM的相机传感器模型
Pub Date : 2007-05-28 DOI: 10.1109/CRV.2007.14
Jing Wu, Hong Zhang
In this paper, we present a technique for the construction of a camera sensor model for visual SLAM. The proposed method is an extension of the general camera calibration procedure and requires the camera to observe a planar checkerboard pattern shown at different orientations. By iteratively placing the pattern at different distances from the camera, we can find a relationship between the measurement noise covariance matrix and the range. We conclude that the error distribution of a camera sensor follows a Gaussian distribution, based on the Geary's test, and the magnitude of the error variance is linearly related to the range between the camera and the features being observed. Our sensor model can potentially benefit visual SLAM algorithms by varying its measurement noise covariance matrix with range.
本文提出了一种构建视觉SLAM相机传感器模型的技术。该方法是一般摄像机标定程序的扩展,要求摄像机观察在不同方向上显示的平面棋盘图案。通过在距离相机不同距离处迭代放置图案,我们可以找到测量噪声协方差矩阵与距离的关系。根据Geary的测试,我们得出结论,相机传感器的误差分布遵循高斯分布,误差方差的大小与相机与被观察特征之间的距离线性相关。我们的传感器模型可以通过改变其测量噪声协方差矩阵随距离的变化而潜在地有利于视觉SLAM算法。
{"title":"Camera Sensor Model for Visual SLAM","authors":"Jing Wu, Hong Zhang","doi":"10.1109/CRV.2007.14","DOIUrl":"https://doi.org/10.1109/CRV.2007.14","url":null,"abstract":"In this paper, we present a technique for the construction of a camera sensor model for visual SLAM. The proposed method is an extension of the general camera calibration procedure and requires the camera to observe a planar checkerboard pattern shown at different orientations. By iteratively placing the pattern at different distances from the camera, we can find a relationship between the measurement noise covariance matrix and the range. We conclude that the error distribution of a camera sensor follows a Gaussian distribution, based on the Geary's test, and the magnitude of the error variance is linearly related to the range between the camera and the features being observed. Our sensor model can potentially benefit visual SLAM algorithms by varying its measurement noise covariance matrix with range.","PeriodicalId":304254,"journal":{"name":"Fourth Canadian Conference on Computer and Robot Vision (CRV '07)","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116006858","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Finite Generalized Gaussian Mixture Modeling and Applications to Image and Video Foreground Segmentation 有限广义高斯混合建模及其在图像和视频前景分割中的应用
Pub Date : 2007-05-28 DOI: 10.1117/1.2898125
M. S. Allili, N. Bouguila, D. Ziou
In this paper, we propose a finite mixture model of generalized Gaussian distributions (GDD) for robust segmentation and data modeling in the presence of noise and outliers. The model has more flexibility to adapt the shape of data and less sensibility for over-fitting the number of classes than the Gaussian mixture. In a first part of the present work, we propose a derivation of the maximum-likelihood estimation of the parameters of the new mixture model and we propose an information-theory based approach for the selection of the number of classes. In a second part, we propose some applications relating to image, motion and foreground segmentation to measure the performance of the new model in image data modeling with comparison to the Gaussian mixture.
在本文中,我们提出了一种广义高斯分布(GDD)的有限混合模型,用于在存在噪声和异常值的情况下进行鲁棒分割和数据建模。与高斯混合模型相比,该模型具有更大的适应数据形状的灵活性和对类数过拟合的敏感性。在本工作的第一部分,我们提出了新混合模型参数的最大似然估计的推导,并提出了一种基于信息论的方法来选择类别的数量。在第二部分中,我们提出了一些与图像、运动和前景分割相关的应用,通过与高斯混合模型的比较,来衡量新模型在图像数据建模中的性能。
{"title":"Finite Generalized Gaussian Mixture Modeling and Applications to Image and Video Foreground Segmentation","authors":"M. S. Allili, N. Bouguila, D. Ziou","doi":"10.1117/1.2898125","DOIUrl":"https://doi.org/10.1117/1.2898125","url":null,"abstract":"In this paper, we propose a finite mixture model of generalized Gaussian distributions (GDD) for robust segmentation and data modeling in the presence of noise and outliers. The model has more flexibility to adapt the shape of data and less sensibility for over-fitting the number of classes than the Gaussian mixture. In a first part of the present work, we propose a derivation of the maximum-likelihood estimation of the parameters of the new mixture model and we propose an information-theory based approach for the selection of the number of classes. In a second part, we propose some applications relating to image, motion and foreground segmentation to measure the performance of the new model in image data modeling with comparison to the Gaussian mixture.","PeriodicalId":304254,"journal":{"name":"Fourth Canadian Conference on Computer and Robot Vision (CRV '07)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131109101","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 116
Non-Uniform Hierarchical Geo-consistency for Multi-baseline Stereo 多基线立体的非均匀分层地理一致性
Pub Date : 2007-05-28 DOI: 10.1109/CRV.2007.47
M. Drouin, Martin Trudeau, S. Roy
We propose a new and flexible hierarchical multibaseline stereo algorithm that features a non-uniform spatial decomposition of the disparity map. The visibility computation and refinement of the disparity map are integrated into a single iterative framework that does not add extra constraints to the cost function. This makes it possible to use a standard efficient stereo matcher during each iteration. The level of refinement is increased automatically where it is needed in order to preserve a good localization of boundaries. While two graph-theoretic stereo matchers are used in our experiments, our framework is general enough to be applied to many others. The validity of our framework is demonstrated using real imagery with ground truth.
我们提出了一种新的灵活的分层多基线立体算法,其特点是视差图的非均匀空间分解。视差图的可见性计算和细化集成到一个迭代框架中,不给代价函数增加额外的约束。这使得在每次迭代中使用标准的高效立体匹配器成为可能。为了保持边界的良好定位,在需要的地方自动增加精化水平。虽然在我们的实验中使用了两个图论立体匹配器,但我们的框架足够通用,可以应用于许多其他的。使用具有地面真值的真实图像证明了我们的框架的有效性。
{"title":"Non-Uniform Hierarchical Geo-consistency for Multi-baseline Stereo","authors":"M. Drouin, Martin Trudeau, S. Roy","doi":"10.1109/CRV.2007.47","DOIUrl":"https://doi.org/10.1109/CRV.2007.47","url":null,"abstract":"We propose a new and flexible hierarchical multibaseline stereo algorithm that features a non-uniform spatial decomposition of the disparity map. The visibility computation and refinement of the disparity map are integrated into a single iterative framework that does not add extra constraints to the cost function. This makes it possible to use a standard efficient stereo matcher during each iteration. The level of refinement is increased automatically where it is needed in order to preserve a good localization of boundaries. While two graph-theoretic stereo matchers are used in our experiments, our framework is general enough to be applied to many others. The validity of our framework is demonstrated using real imagery with ground truth.","PeriodicalId":304254,"journal":{"name":"Fourth Canadian Conference on Computer and Robot Vision (CRV '07)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128224469","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Efficient indexing for strongly similar subimage retrieval 强相似子图像检索的高效索引
Pub Date : 2007-05-28 DOI: 10.1109/CRV.2007.24
G. Roth, W. Scott
Strongly similar subimages contain different views of the same object. In subimage search, the user selects an image region and the retrieval system attempts to find matching subimages in an image database that are strongly similar. Solutions have been proposed using salient features or "interest points" that have associated descriptor vectors. However, searching large image databases by exhaustive comparison of interest point descriptors is not feasible. To solve this problem, we propose a novel off-line indexing scheme based on the most significant bits (MSBs) of these descriptors. On-line search uses this index file to limit the search to interest points whose descriptors have the same MSB value, a process up to three orders of magnitude faster than exhaustive search. It is also incremental, since the index file for a union of a group of images can be created by merging the index files of the individual image groups. The effectiveness of the approach is demonstrated experimentally on a variety of image databases.
非常相似的子图像包含同一对象的不同视图。在子图像搜索中,用户选择一个图像区域,检索系统尝试在图像数据库中找到匹配的强相似的子图像。解决方案已经提出使用显著特征或“兴趣点”,有相关的描述符向量。然而,通过兴趣点描述符的穷举比较来搜索大型图像数据库是不可行的。为了解决这个问题,我们提出了一种基于这些描述符的最高有效位(MSBs)的离线索引方案。在线搜索使用该索引文件将搜索限制在描述符具有相同MSB值的兴趣点上,这个过程比穷举搜索快三个数量级。它也是增量的,因为一组图像的联合索引文件可以通过合并单个图像组的索引文件来创建。在多种图像数据库上进行了实验,验证了该方法的有效性。
{"title":"Efficient indexing for strongly similar subimage retrieval","authors":"G. Roth, W. Scott","doi":"10.1109/CRV.2007.24","DOIUrl":"https://doi.org/10.1109/CRV.2007.24","url":null,"abstract":"Strongly similar subimages contain different views of the same object. In subimage search, the user selects an image region and the retrieval system attempts to find matching subimages in an image database that are strongly similar. Solutions have been proposed using salient features or \"interest points\" that have associated descriptor vectors. However, searching large image databases by exhaustive comparison of interest point descriptors is not feasible. To solve this problem, we propose a novel off-line indexing scheme based on the most significant bits (MSBs) of these descriptors. On-line search uses this index file to limit the search to interest points whose descriptors have the same MSB value, a process up to three orders of magnitude faster than exhaustive search. It is also incremental, since the index file for a union of a group of images can be created by merging the index files of the individual image groups. The effectiveness of the approach is demonstrated experimentally on a variety of image databases.","PeriodicalId":304254,"journal":{"name":"Fourth Canadian Conference on Computer and Robot Vision (CRV '07)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129392655","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Automatic Annotation of Humans in Surveillance Video 监控视频中人的自动标注
Pub Date : 2007-05-28 DOI: 10.1109/CRV.2007.12
D. Hansen, B. K. Mortensen, P. Duizer, Jens R. Andersen, T. Moeslund
In this paper we present a system for automatic annotation of humans passing a surveillance camera. Each human has 4 associated annotations: the primary color of the clothing, the height, and focus of attention. The annotation occurs after robust background subtraction based on a Codebook representation. The primary colors of the clothing are estimated by grouping similar pixels according to a body model. The height is estimated based on a 3D mapping using the head and feet. Lastly, the focus of attention is defined as the overall direction of the head, which is estimated using changes in intensity at four different positions. Results show successful detection and hence successful annotation for most test sequences.
本文介绍了一种对经过监控摄像机的人员进行自动标注的系统。每个人都有4个相关的注释:衣服的原色、身高和注意力的焦点。在基于Codebook表示的稳健背景减法之后发生注释。衣服的原色是通过根据身体模型分组相似的像素来估计的。高度是根据头部和脚的3D映射估计的。最后,注意力的焦点被定义为头部的整体方向,这是使用四个不同位置的强度变化来估计的。结果表明,大多数测试序列的检测和注释都是成功的。
{"title":"Automatic Annotation of Humans in Surveillance Video","authors":"D. Hansen, B. K. Mortensen, P. Duizer, Jens R. Andersen, T. Moeslund","doi":"10.1109/CRV.2007.12","DOIUrl":"https://doi.org/10.1109/CRV.2007.12","url":null,"abstract":"In this paper we present a system for automatic annotation of humans passing a surveillance camera. Each human has 4 associated annotations: the primary color of the clothing, the height, and focus of attention. The annotation occurs after robust background subtraction based on a Codebook representation. The primary colors of the clothing are estimated by grouping similar pixels according to a body model. The height is estimated based on a 3D mapping using the head and feet. Lastly, the focus of attention is defined as the overall direction of the head, which is estimated using changes in intensity at four different positions. Results show successful detection and hence successful annotation for most test sequences.","PeriodicalId":304254,"journal":{"name":"Fourth Canadian Conference on Computer and Robot Vision (CRV '07)","volume":"151 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122963662","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
Screen-Camera Calibration using a Spherical Mirror 使用球面镜校准屏幕摄像机
Pub Date : 2007-05-28 DOI: 10.1109/CRV.2007.59
Yannick Francken, C. Hermans, P. Bekaert
Developments in the consumer market have indicated that the average user of a personal computer is likely to also own a webcam. With the emergence of this new user group will come a new set of applications, which will require a user-friendly way to calibrate the position of the camera with respect to the location of the screen. This paper presents a fully automatic method to calibrate a screen-camera setup, using a single moving spherical mirror. Unlike other methods, our algorithm needs no user intervention other then moving around a spherical mirror. In addition, if the user provides the algorithm with the exact radius of the sphere in millimeters, the scale of the computed solution is uniquely defined.
消费市场的发展表明,个人电脑的普通用户可能也拥有一个网络摄像头。随着这一新的用户群体的出现,将会出现一套新的应用程序,这将需要一种用户友好的方式来校准相机相对于屏幕位置的位置。本文提出了一种利用单个运动球面反射镜对屏幕摄像机装置进行全自动标定的方法。与其他方法不同的是,我们的算法不需要用户干预,只需绕着球面镜移动即可。此外,如果用户为算法提供精确的球面半径(以毫米为单位),则计算解的尺度是唯一定义的。
{"title":"Screen-Camera Calibration using a Spherical Mirror","authors":"Yannick Francken, C. Hermans, P. Bekaert","doi":"10.1109/CRV.2007.59","DOIUrl":"https://doi.org/10.1109/CRV.2007.59","url":null,"abstract":"Developments in the consumer market have indicated that the average user of a personal computer is likely to also own a webcam. With the emergence of this new user group will come a new set of applications, which will require a user-friendly way to calibrate the position of the camera with respect to the location of the screen. This paper presents a fully automatic method to calibrate a screen-camera setup, using a single moving spherical mirror. Unlike other methods, our algorithm needs no user intervention other then moving around a spherical mirror. In addition, if the user provides the algorithm with the exact radius of the sphere in millimeters, the scale of the computed solution is uniquely defined.","PeriodicalId":304254,"journal":{"name":"Fourth Canadian Conference on Computer and Robot Vision (CRV '07)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122678002","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Fourier tags: Smoothly degradable fiducial markers for use in human-robot interaction 傅立叶标签:用于人机交互的平滑可降解基准标记
Pub Date : 2007-05-28 DOI: 10.1109/CRV.2007.34
Junaed Sattar, Eric Bourque, P. Giguère, G. Dudek
In this paper we introduce the Fourier tag, a synthetic fiducial marker used to visually encode information and provide controllable positioning. The Fourier tag is a synthetic target akin to a bar-code that specifies multi-bit information which can be efficiently and robustly detected in an image. Moreover, the Fourier tag has the beneficial property that the bit string it encodes has variable length as a function of the distance between the camera and the target. This follows from the fact that the effective resolution decreases as an effect of perspective. This paper introduces the Fourier tag, describes its design, and illustrates its properties experimentally.
本文介绍了一种合成基准标记——傅里叶标签,用于视觉编码信息和提供可控定位。傅里叶标签是一种类似于条形码的合成目标,它指定了可以在图像中有效和鲁棒地检测到的多比特信息。此外,傅里叶标签还有一个优点,即它所编码的比特串的长度随相机与目标之间的距离而变化。这是因为有效分辨率随着透视的影响而降低。本文介绍了傅里叶标签,描述了其设计,并对其特性进行了实验说明。
{"title":"Fourier tags: Smoothly degradable fiducial markers for use in human-robot interaction","authors":"Junaed Sattar, Eric Bourque, P. Giguère, G. Dudek","doi":"10.1109/CRV.2007.34","DOIUrl":"https://doi.org/10.1109/CRV.2007.34","url":null,"abstract":"In this paper we introduce the Fourier tag, a synthetic fiducial marker used to visually encode information and provide controllable positioning. The Fourier tag is a synthetic target akin to a bar-code that specifies multi-bit information which can be efficiently and robustly detected in an image. Moreover, the Fourier tag has the beneficial property that the bit string it encodes has variable length as a function of the distance between the camera and the target. This follows from the fact that the effective resolution decreases as an effect of perspective. This paper introduces the Fourier tag, describes its design, and illustrates its properties experimentally.","PeriodicalId":304254,"journal":{"name":"Fourth Canadian Conference on Computer and Robot Vision (CRV '07)","volume":"110 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132573991","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 66
Spatial Topology Graphs for Feature-Minimal Correspondence 最小特征对应的空间拓扑图
Pub Date : 2007-05-28 DOI: 10.1109/CRV.2007.60
Z. Tauber, Ze-Nian Li, M. S. Drew
Multiview image matching methods typically require feature point correspondences. We propose a novel spatial topology method that represents the space with a set of connected projective invariant features. Typically, isolated features, such as corners, cannot be matched reliably. Hence, limitations are imposed on viewpoint changes, or projective invariant descriptions are needed. The fundamental matrix is discovered using stochastic optimization requiring a large number of features. In contrast, our enhanced feature set models connectivity in space, forming a unique configuration that can be matched with few features and over large viewpoint changes. Our features are derived from edges, their curvatures, and neighborhood relationships. A probabilistic spatial topology graph models the space using these features and a second graph represents the neighborhood relationships. Probabilistic graph matching is used to find feature correspondences. Our results show robust feature detection and an average 80% discovery rate of feature matches.
多视图图像匹配方法通常需要特征点对应。我们提出了一种新的空间拓扑方法,用一组连通的射影不变特征来表示空间。通常,孤立的特征(如角)不能可靠地匹配。因此,对视点变化施加了限制,或者需要射影不变描述。基本矩阵是用需要大量特征的随机优化方法找到的。相比之下,我们增强的特征集在空间中建模连通性,形成一个独特的配置,可以与少数特征和大型视点变化相匹配。我们的特征来源于边缘、它们的曲率和邻域关系。概率空间拓扑图使用这些特征对空间进行建模,第二个图表示邻域关系。概率图匹配用于寻找特征对应。我们的结果显示了鲁棒的特征检测和平均80%的特征匹配发现率。
{"title":"Spatial Topology Graphs for Feature-Minimal Correspondence","authors":"Z. Tauber, Ze-Nian Li, M. S. Drew","doi":"10.1109/CRV.2007.60","DOIUrl":"https://doi.org/10.1109/CRV.2007.60","url":null,"abstract":"Multiview image matching methods typically require feature point correspondences. We propose a novel spatial topology method that represents the space with a set of connected projective invariant features. Typically, isolated features, such as corners, cannot be matched reliably. Hence, limitations are imposed on viewpoint changes, or projective invariant descriptions are needed. The fundamental matrix is discovered using stochastic optimization requiring a large number of features. In contrast, our enhanced feature set models connectivity in space, forming a unique configuration that can be matched with few features and over large viewpoint changes. Our features are derived from edges, their curvatures, and neighborhood relationships. A probabilistic spatial topology graph models the space using these features and a second graph represents the neighborhood relationships. Probabilistic graph matching is used to find feature correspondences. Our results show robust feature detection and an average 80% discovery rate of feature matches.","PeriodicalId":304254,"journal":{"name":"Fourth Canadian Conference on Computer and Robot Vision (CRV '07)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129216679","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Efficient Registration of 3D SPHARM Surfaces 三维SPHARM表面的有效配准
Pub Date : 2007-05-28 DOI: 10.1109/CRV.2007.26
Li Shen, Heng Huang, F. Makedon, A. Saykin
We present SHREC, an efficient algorithm for registration of 3D SPHARM (spherical harmonic) surfaces. SHREC follows the iterative closest point (ICP) registration strategy, and alternately improves the surface correspondence and adjusts the object pose. It establishes the surface correspondence by aligning the underlying SPHARM parameterization. It employs a rotational property of the harmonic expansion to accelerate its step for parameterization rotation. It uses a hierarchical icosahedron approach to sample the rotation space and searches for the best parameterization that matches the template. Our experimental results show that SHREC can not only create more accurate registration than previous methods but also do it efficiently. SHREC is a simple, efficient and general registration method, and has a great potential to be used in many shape modeling and analysis applications.
提出了一种三维球面谐波曲面配准的高效算法SHREC。SHREC采用迭代最近点(ICP)配准策略,交替改进表面对应性和调整目标位姿。它通过对齐底层SPHARM参数化来建立表面对应关系。它利用谐波展开的旋转特性来加速参数化旋转。它使用分层二十面体方法对旋转空间进行采样,并搜索与模板匹配的最佳参数化。实验结果表明,该方法不仅比以往的配准方法更准确,而且效率更高。SHREC是一种简单、高效、通用的配准方法,在许多形状建模和分析应用中具有很大的应用潜力。
{"title":"Efficient Registration of 3D SPHARM Surfaces","authors":"Li Shen, Heng Huang, F. Makedon, A. Saykin","doi":"10.1109/CRV.2007.26","DOIUrl":"https://doi.org/10.1109/CRV.2007.26","url":null,"abstract":"We present SHREC, an efficient algorithm for registration of 3D SPHARM (spherical harmonic) surfaces. SHREC follows the iterative closest point (ICP) registration strategy, and alternately improves the surface correspondence and adjusts the object pose. It establishes the surface correspondence by aligning the underlying SPHARM parameterization. It employs a rotational property of the harmonic expansion to accelerate its step for parameterization rotation. It uses a hierarchical icosahedron approach to sample the rotation space and searches for the best parameterization that matches the template. Our experimental results show that SHREC can not only create more accurate registration than previous methods but also do it efficiently. SHREC is a simple, efficient and general registration method, and has a great potential to be used in many shape modeling and analysis applications.","PeriodicalId":304254,"journal":{"name":"Fourth Canadian Conference on Computer and Robot Vision (CRV '07)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123385556","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 32
Real-time eye blink detection with GPU-based SIFT tracking 基于gpu的SIFT跟踪实时眨眼检测
Pub Date : 2007-05-28 DOI: 10.1109/CRV.2007.54
M. Lalonde, David Byrns, L. Gagnon, N. Teasdale, D. Laurendeau
This paper reports on the implementation of a GPU-based, real-time eye blink detector on very low contrast images acquired under near-infrared illumination. This detector is part of a multi-sensor data acquisition and analysis system for driver performance assessment and training. Eye blinks are detected inside regions of interest that are aligned with the subject's eyes at initialization. Alignment is maintained through time by tracking SIFT feature points that are used to estimate the affine transformation between the initial face pose and the pose in subsequent frames. The GPU implementation of the SIFT feature point extraction algorithm ensures real-time processing. An eye blink detection rate of 97% is obtained on a video dataset of 33,000 frames showing 237 blinks from 22 subjects.
本文报道了一种基于gpu的近红外低对比度图像实时眨眼检测器的实现。该检测器是用于驾驶员性能评估和培训的多传感器数据采集和分析系统的一部分。在初始化时,在与受试者眼睛对齐的感兴趣区域内检测到眨眼。通过跟踪SIFT特征点来保持对齐,这些特征点用于估计初始人脸姿态与后续帧中姿态之间的仿射变换。SIFT特征点提取算法的GPU实现保证了处理的实时性。在一个包含33,000帧的视频数据集上,从22个受试者中获得了237次眨眼,从而获得了97%的眨眼检测率。
{"title":"Real-time eye blink detection with GPU-based SIFT tracking","authors":"M. Lalonde, David Byrns, L. Gagnon, N. Teasdale, D. Laurendeau","doi":"10.1109/CRV.2007.54","DOIUrl":"https://doi.org/10.1109/CRV.2007.54","url":null,"abstract":"This paper reports on the implementation of a GPU-based, real-time eye blink detector on very low contrast images acquired under near-infrared illumination. This detector is part of a multi-sensor data acquisition and analysis system for driver performance assessment and training. Eye blinks are detected inside regions of interest that are aligned with the subject's eyes at initialization. Alignment is maintained through time by tracking SIFT feature points that are used to estimate the affine transformation between the initial face pose and the pose in subsequent frames. The GPU implementation of the SIFT feature point extraction algorithm ensures real-time processing. An eye blink detection rate of 97% is obtained on a video dataset of 33,000 frames showing 237 blinks from 22 subjects.","PeriodicalId":304254,"journal":{"name":"Fourth Canadian Conference on Computer and Robot Vision (CRV '07)","volume":"84 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122640448","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 133
期刊
Fourth Canadian Conference on Computer and Robot Vision (CRV '07)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1