首页 > 最新文献

2012 Ninth Conference on Computer and Robot Vision最新文献

英文 中文
Gaussian Process Gauss-Newton: Non-Parametric State Estimation 高斯过程高斯-牛顿:非参数状态估计
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.35
Chi Hay Tong, P. Furgale, T. Barfoot
In this paper, we present Gaussian Process Gauss-Newton (GPGN), an algorithm for non-parametric, continuous-time, nonlinear, batch state estimation. This work adapts the methods of Gaussian Process regression to the problem of batch state estimation by using the Gauss-Newton method. In particular, we formulate the estimation problem with a continuous-time state model, along with the more conventional discrete-time measurements. Our derivation utilizes a basis function approach, but through algebraic manipulations, returns to a non-parametric form by replacing the basis functions with covariance functions (i.e., the kernel trick). The algorithm is validated through hardware-based experiments utilizing the well-understood problem of 2D rover localization using a known map as an illustrative example, and is compared to the traditional discrete-time batch Gauss-Newton approach.
本文提出了一种非参数、连续时间、非线性、批处理状态估计的高斯过程高斯-牛顿(GPGN)算法。本文采用高斯-牛顿方法,将高斯过程回归方法应用于批量状态估计问题。特别地,我们用连续时间状态模型以及更传统的离散时间测量来表述估计问题。我们的推导利用基函数方法,但通过代数操作,通过用协方差函数(即核技巧)替换基函数返回到非参数形式。该算法通过基于硬件的实验进行了验证,利用众所周知的2D漫游车定位问题,以已知地图为例,并与传统的离散时间批处理高斯-牛顿方法进行了比较。
{"title":"Gaussian Process Gauss-Newton: Non-Parametric State Estimation","authors":"Chi Hay Tong, P. Furgale, T. Barfoot","doi":"10.1109/CRV.2012.35","DOIUrl":"https://doi.org/10.1109/CRV.2012.35","url":null,"abstract":"In this paper, we present Gaussian Process Gauss-Newton (GPGN), an algorithm for non-parametric, continuous-time, nonlinear, batch state estimation. This work adapts the methods of Gaussian Process regression to the problem of batch state estimation by using the Gauss-Newton method. In particular, we formulate the estimation problem with a continuous-time state model, along with the more conventional discrete-time measurements. Our derivation utilizes a basis function approach, but through algebraic manipulations, returns to a non-parametric form by replacing the basis functions with covariance functions (i.e., the kernel trick). The algorithm is validated through hardware-based experiments utilizing the well-understood problem of 2D rover localization using a known map as an illustrative example, and is compared to the traditional discrete-time batch Gauss-Newton approach.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125785136","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
Shape from Suggestive Contours Using 3D Priors 形状从暗示轮廓使用3D先验
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.38
S. Wuhrer, Chang Shu
This paper introduces an approach to predict the three-dimensional shape of an object belonging to a specific class of shapes shown in an input image. We use suggestive contour, a shape-suggesting image feature developed in computer graphics in the context of non-photorealistic rendering, to reconstruct 3D shapes. We learn a functional mapping from the shape space of suggestive contours to the space of 3D shapes and use this mapping to predict 3D shapes based on a single input image. We demonstrate that the method can be used to predict the shape of deformable objects and to predict the shape of human faces using synthetic experiments and experiments based on artist drawn sketches and photographs.
本文介绍了一种预测物体三维形状的方法,该物体属于输入图像中显示的特定形状类。我们使用暗示性轮廓(一种在非真实感渲染的背景下在计算机图形学中发展起来的暗示形状的图像特征)来重建3D形状。我们学习了从暗示轮廓的形状空间到3D形状空间的函数映射,并使用该映射来预测基于单个输入图像的3D形状。我们证明了该方法可以用于预测可变形物体的形状,并通过合成实验和基于艺术家绘制的草图和照片的实验来预测人脸的形状。
{"title":"Shape from Suggestive Contours Using 3D Priors","authors":"S. Wuhrer, Chang Shu","doi":"10.1109/CRV.2012.38","DOIUrl":"https://doi.org/10.1109/CRV.2012.38","url":null,"abstract":"This paper introduces an approach to predict the three-dimensional shape of an object belonging to a specific class of shapes shown in an input image. We use suggestive contour, a shape-suggesting image feature developed in computer graphics in the context of non-photorealistic rendering, to reconstruct 3D shapes. We learn a functional mapping from the shape space of suggestive contours to the space of 3D shapes and use this mapping to predict 3D shapes based on a single input image. We demonstrate that the method can be used to predict the shape of deformable objects and to predict the shape of human faces using synthetic experiments and experiments based on artist drawn sketches and photographs.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"119 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114287500","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Enhancing Exploration in Topological Worlds with Multiple Immovable Markers 利用多个不可移动标记增强拓扑世界的探索
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.49
Hui Wang, M. Jenkin, Patrick W. Dymond
The fundamental problem in robotic exploration and mapping of an unknown environment is answering the question 'have I been here before?', which is also known as the 'loop closing' problem. One approach to answering this problem in embedded topological worlds is to resort to the use of an external marking aid that can help the robot disambiguate places. This paper investigates the power of different marker-based aids in topological exploration. We describe enhanced versions of edge- and vertex-based marker algorithms and demonstrate algorithms with enhanced lower bounds in terms of number of markers and motions required in order to map an embedded topological environment.
机器人探索和绘制未知环境的基本问题是回答“我以前来过这里吗?”,这也被称为“闭环”问题。在嵌入式拓扑世界中回答这个问题的一种方法是求助于使用外部标记辅助工具,它可以帮助机器人消除位置的歧义。本文研究了不同的基于标记的辅助工具在拓扑探索中的作用。我们描述了基于边缘和顶点的标记算法的增强版本,并演示了在映射嵌入式拓扑环境所需的标记和运动数量方面具有增强下限的算法。
{"title":"Enhancing Exploration in Topological Worlds with Multiple Immovable Markers","authors":"Hui Wang, M. Jenkin, Patrick W. Dymond","doi":"10.1109/CRV.2012.49","DOIUrl":"https://doi.org/10.1109/CRV.2012.49","url":null,"abstract":"The fundamental problem in robotic exploration and mapping of an unknown environment is answering the question 'have I been here before?', which is also known as the 'loop closing' problem. One approach to answering this problem in embedded topological worlds is to resort to the use of an external marking aid that can help the robot disambiguate places. This paper investigates the power of different marker-based aids in topological exploration. We describe enhanced versions of edge- and vertex-based marker algorithms and demonstrate algorithms with enhanced lower bounds in terms of number of markers and motions required in order to map an embedded topological environment.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"93 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129463853","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Parallelizing a Face Detection and Tracking System for Multi-Core Processors 基于多核处理器的并行人脸检测与跟踪系统
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.45
A. Ranjan, S. Malik
This paper describes how to accelerate a real-world face detection and tracking system by taking advantage of the multiple processing cores that are present in most modern CPUs. This work makes three key contributions. The first is the presentation of a highly optimized serial face detection and tracking algorithm that uses motion estimation and local search windows to achieve fast processing rates. The second is redefining the face detection process based on a set of independent face scales that can be processed in parallel on separate CPU cores while also achieving a target processing rate. The third contribution is demonstrating how multiple cores can be used to accelerate the face tracking process which provides significant speed boosts when tracking a large number of faces simultaneously. Used in a real-world application, the parallel face detector and tracker yields a 50-70% speed boost over the serial version when tested on a commodity multi-core CPU.
本文描述了如何利用大多数现代cpu中存在的多个处理内核来加速现实世界中的人脸检测和跟踪系统。这项工作有三个关键贡献。首先,提出了一种高度优化的串行人脸检测和跟踪算法,该算法使用运动估计和局部搜索窗口来实现快速处理速率。第二是基于一组独立的人脸尺度重新定义人脸检测过程,这些人脸尺度可以在单独的CPU内核上并行处理,同时也达到目标处理速率。第三个贡献是展示了如何使用多个核心来加速人脸跟踪过程,这在同时跟踪大量人脸时提供了显着的速度提升。在实际应用中,当在商用多核CPU上测试时,并行人脸检测器和跟踪器的速度比串行版本提高了50-70%。
{"title":"Parallelizing a Face Detection and Tracking System for Multi-Core Processors","authors":"A. Ranjan, S. Malik","doi":"10.1109/CRV.2012.45","DOIUrl":"https://doi.org/10.1109/CRV.2012.45","url":null,"abstract":"This paper describes how to accelerate a real-world face detection and tracking system by taking advantage of the multiple processing cores that are present in most modern CPUs. This work makes three key contributions. The first is the presentation of a highly optimized serial face detection and tracking algorithm that uses motion estimation and local search windows to achieve fast processing rates. The second is redefining the face detection process based on a set of independent face scales that can be processed in parallel on separate CPU cores while also achieving a target processing rate. The third contribution is demonstrating how multiple cores can be used to accelerate the face tracking process which provides significant speed boosts when tracking a large number of faces simultaneously. Used in a real-world application, the parallel face detector and tracker yields a 50-70% speed boost over the serial version when tested on a commodity multi-core CPU.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129073183","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Large-Scale Tattoo Image Retrieval 大规模纹身图像检索
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.67
D. Manger
In current biometric-based identification systems, tattoos and other body modifications have shown to provide a useful source of information. Besides manual category label assignment, approaches utilizing state-of-the-art content-based image retrieval (CBIR) techniques have become increasingly popular. While local feature-based similarities of tattoo images achieve excellent retrieval accuracy, scalability to large image databases can be addressed with the popular bag-of-word model. In this paper, we show how recent advances in CBIR can be utilized to build up a large-scale tattoo image retrieval system. Compared to other systems, we chose a different approach to circumvent the loss of accuracy caused by the bag-of-word quantization. Its efficiency and effectiveness are shown in experiments with several tattoo databases of up to 330,000 images.
在目前的基于生物特征的识别系统中,纹身和其他身体修饰已被证明提供了有用的信息来源。除了手动分类标签分配之外,利用最先进的基于内容的图像检索(CBIR)技术的方法也越来越受欢迎。虽然基于局部特征的纹身图像相似度可以获得很好的检索精度,但流行的词袋模型可以解决大型图像数据库的可扩展性问题。在本文中,我们展示了如何利用CBIR的最新进展来建立一个大规模的纹身图像检索系统。与其他系统相比,我们选择了一种不同的方法来避免词袋量化引起的精度损失。它的效率和有效性在几个纹身数据库多达33万张图像的实验中得到了证明。
{"title":"Large-Scale Tattoo Image Retrieval","authors":"D. Manger","doi":"10.1109/CRV.2012.67","DOIUrl":"https://doi.org/10.1109/CRV.2012.67","url":null,"abstract":"In current biometric-based identification systems, tattoos and other body modifications have shown to provide a useful source of information. Besides manual category label assignment, approaches utilizing state-of-the-art content-based image retrieval (CBIR) techniques have become increasingly popular. While local feature-based similarities of tattoo images achieve excellent retrieval accuracy, scalability to large image databases can be addressed with the popular bag-of-word model. In this paper, we show how recent advances in CBIR can be utilized to build up a large-scale tattoo image retrieval system. Compared to other systems, we chose a different approach to circumvent the loss of accuracy caused by the bag-of-word quantization. Its efficiency and effectiveness are shown in experiments with several tattoo databases of up to 330,000 images.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115415498","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 34
Difference of Circles Feature Detector 圆差特征检测器
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.16
Abdullah Hojaij, Adel H. Fakih, A. Wong, J. Zelek
Feature detection is a crucial step in many Computer Vision applications such as matching, tracking, visual odometry and object recognition, etc. Detecting robust features that are persistent, rotation-invariant, and quickly calculated is a major problem in computer vision. Feature detectors using the difference of Gaussian (DoG) are computationally expensive, however, if the DoG is used with image sub sampling at higher orders, the detectors become fast but their feature localization becomes inaccurate. Detectors based on difference of octagons (DoO) or difference of stars (DoS) algorithm are fast and localize the features accurately, but they are not rotation-invariant. This paper introduces a novel technique for the difference of circles (DoC) algorithm, used for feature detection, that is perfectly rotation-invariant and has the potential of being very fast through using circular integral images. The performance of DoC algorithm is compared with the difference of stars algorithm presented by 'Willow Garage'. The experiments conducted concentrate on the rotation-invariance property of DoC.
特征检测是匹配、跟踪、视觉里程计和目标识别等计算机视觉应用的关键步骤。检测持久、旋转不变性和快速计算的鲁棒特征是计算机视觉中的主要问题。使用高斯差分法(DoG)的特征检测器计算量很大,但是,如果将DoG与高阶图像子采样一起使用,检测器速度很快,但特征定位不准确。基于八角形差(DoO)或星差(DoS)算法的检测器速度快,定位准确,但它们不是旋转不变性的。本文介绍了一种用于特征检测的圆差(DoC)算法的新技术,该算法具有完全的旋转不变性,并且通过使用圆形积分图像具有非常快的潜力。将DoC算法的性能与“柳树车库”提出的星差算法进行了比较。实验集中研究了DoC的旋转不变性。
{"title":"Difference of Circles Feature Detector","authors":"Abdullah Hojaij, Adel H. Fakih, A. Wong, J. Zelek","doi":"10.1109/CRV.2012.16","DOIUrl":"https://doi.org/10.1109/CRV.2012.16","url":null,"abstract":"Feature detection is a crucial step in many Computer Vision applications such as matching, tracking, visual odometry and object recognition, etc. Detecting robust features that are persistent, rotation-invariant, and quickly calculated is a major problem in computer vision. Feature detectors using the difference of Gaussian (DoG) are computationally expensive, however, if the DoG is used with image sub sampling at higher orders, the detectors become fast but their feature localization becomes inaccurate. Detectors based on difference of octagons (DoO) or difference of stars (DoS) algorithm are fast and localize the features accurately, but they are not rotation-invariant. This paper introduces a novel technique for the difference of circles (DoC) algorithm, used for feature detection, that is perfectly rotation-invariant and has the potential of being very fast through using circular integral images. The performance of DoC algorithm is compared with the difference of stars algorithm presented by 'Willow Garage'. The experiments conducted concentrate on the rotation-invariance property of DoC.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116128147","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A Variational Approach to Mapping and Localization 映射和定位的变分方法
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.72
A. Hogue, S. Khattak
A fundamental open problem in SLAM is the effective representation of the map in unknown, ambiguous, complex, dynamic environments. Representing such environments in a suitable manner is a complex task. Existing approaches to SLAM use map representations that store individual features (range measurements, image patches, or higher level semantic features) and their locations in the environment. The choice of how we represent the map produces limitations which in many ways are unfavourable for application in real-world scenarios. In this paper, we explore a new approach to SLAM that redefines sensing and robot motion as acts of deformation of a differentiable surface. Distance fields and level set methods are utilized to define a parallel to the components of the SLAM estimation process and an algorithm is developed and demonstrated. The variational framework developed is capable of representing complex dynamic scenes and spatially varying uncertainty for sensor and robot models.
SLAM的一个基本开放问题是如何在未知、模糊、复杂、动态的环境中有效地表示地图。以合适的方式表示这样的环境是一项复杂的任务。现有的SLAM方法使用存储单个特征(距离测量、图像补丁或更高级别语义特征)及其在环境中的位置的地图表示。我们如何表示地图的选择产生了许多限制,这些限制在许多方面不利于在现实场景中的应用。在本文中,我们探索了一种SLAM的新方法,该方法将传感和机器人运动重新定义为可微表面的变形行为。利用距离场和水平集方法来定义与SLAM估计过程并行的组件,并开发和演示了一种算法。所开发的变分框架能够表示传感器和机器人模型的复杂动态场景和空间变化的不确定性。
{"title":"A Variational Approach to Mapping and Localization","authors":"A. Hogue, S. Khattak","doi":"10.1109/CRV.2012.72","DOIUrl":"https://doi.org/10.1109/CRV.2012.72","url":null,"abstract":"A fundamental open problem in SLAM is the effective representation of the map in unknown, ambiguous, complex, dynamic environments. Representing such environments in a suitable manner is a complex task. Existing approaches to SLAM use map representations that store individual features (range measurements, image patches, or higher level semantic features) and their locations in the environment. The choice of how we represent the map produces limitations which in many ways are unfavourable for application in real-world scenarios. In this paper, we explore a new approach to SLAM that redefines sensing and robot motion as acts of deformation of a differentiable surface. Distance fields and level set methods are utilized to define a parallel to the components of the SLAM estimation process and an algorithm is developed and demonstrated. The variational framework developed is capable of representing complex dynamic scenes and spatially varying uncertainty for sensor and robot models.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"60 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122582019","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Learning Categorical Shape from Captioned Images 从标题图像中学习分类形状
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.37
T. S. Lee, S. Fidler, Alex Levinshtein, Sven J. Dickinson
Given a set of captioned images of cluttered scenes containing various objects in different positions and scales, we learn named contour models of object categories without relying on bounding box annotation. We extend a recent language-vision integration framework that finds spatial configurations of image features that co-occur with words in image captions. By substituting appearance features with local contour features, object categories are recognized by a contour model that grows along the object's boundary. Experiments on ETHZ are presented to show that 1) the extended framework is better able to learn named visual categories whose within-class variation is better captured by a shape model than an appearance model, and 2) typical object recognition methods fail when manually annotated bounding boxes are unavailable.
给定一组包含不同位置和尺度的各种物体的杂乱场景的字幕图像,我们学习物体类别的命名轮廓模型,而不依赖于边界框注释。我们扩展了最近的语言视觉集成框架,该框架发现图像特征的空间配置与图像标题中的单词共同出现。通过用局部轮廓特征代替外观特征,利用沿目标边界生长的轮廓模型识别目标类别。在ETHZ上进行的实验表明:1)扩展框架能够更好地学习命名的视觉类别,其类内变化被形状模型比外观模型更好地捕获;2)当手工标注的边界框不可用时,典型的对象识别方法会失败。
{"title":"Learning Categorical Shape from Captioned Images","authors":"T. S. Lee, S. Fidler, Alex Levinshtein, Sven J. Dickinson","doi":"10.1109/CRV.2012.37","DOIUrl":"https://doi.org/10.1109/CRV.2012.37","url":null,"abstract":"Given a set of captioned images of cluttered scenes containing various objects in different positions and scales, we learn named contour models of object categories without relying on bounding box annotation. We extend a recent language-vision integration framework that finds spatial configurations of image features that co-occur with words in image captions. By substituting appearance features with local contour features, object categories are recognized by a contour model that grows along the object's boundary. Experiments on ETHZ are presented to show that 1) the extended framework is better able to learn named visual categories whose within-class variation is better captured by a shape model than an appearance model, and 2) typical object recognition methods fail when manually annotated bounding boxes are unavailable.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"121 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131553446","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Accelerometer Localization in the View of a Stationary Camera 静止摄像机视图下加速度计的定位
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.22
Sebastian Stein, S. McKenna
This paper addresses the problem of localizing an accelerometer in the view of a stationary camera as a first step towards multi-model activity recognition. This problem is challenging as accelerometers are visually occluded, they measure proper acceleration including effects of gravity and their orientation is unknown and changes over time relative to camera viewpoint. Accelerometers are localized by matching acceleration estimated along visual point trajectories to accelerometer data. Trajectories are constructed from point feature tracking (KLT) and by grid sampling from a dense flow field. We also construct 3D trajectories with visual depth information. The similarity between accelerometer data and a trajectory is computed by counting the number of frames in which the norms of accelerations in both sequences exceed a threshold. For quantitative evaluation we collected a challenging dataset consisting of video and accelerometer data of a person preparing a mixed salad with accelerometer-equipped kitchen utensils. Trajectories from dense optical flow yielded a higher localization accuracy compared to point feature tracking.
本文解决了加速度计在静止相机视图下的定位问题,作为多模型活动识别的第一步。这个问题很有挑战性,因为加速度计在视觉上是被遮挡的,它们测量的是适当的加速度,包括重力的影响,它们的方向是未知的,并且相对于相机视点随时间而变化。通过将沿视觉点轨迹估计的加速度与加速度计数据匹配来定位加速度计。轨迹由点特征跟踪(KLT)和密集流场的网格采样构建。我们还利用视觉深度信息构建了三维轨迹。加速度计数据和轨迹之间的相似性是通过计算两个序列中加速度规范超过阈值的帧数来计算的。为了进行定量评估,我们收集了一个具有挑战性的数据集,其中包括一个人用配备加速度计的厨房用具准备混合沙拉的视频和加速度计数据。与点特征跟踪相比,密集光流轨迹的定位精度更高。
{"title":"Accelerometer Localization in the View of a Stationary Camera","authors":"Sebastian Stein, S. McKenna","doi":"10.1109/CRV.2012.22","DOIUrl":"https://doi.org/10.1109/CRV.2012.22","url":null,"abstract":"This paper addresses the problem of localizing an accelerometer in the view of a stationary camera as a first step towards multi-model activity recognition. This problem is challenging as accelerometers are visually occluded, they measure proper acceleration including effects of gravity and their orientation is unknown and changes over time relative to camera viewpoint. Accelerometers are localized by matching acceleration estimated along visual point trajectories to accelerometer data. Trajectories are constructed from point feature tracking (KLT) and by grid sampling from a dense flow field. We also construct 3D trajectories with visual depth information. The similarity between accelerometer data and a trajectory is computed by counting the number of frames in which the norms of accelerations in both sequences exceed a threshold. For quantitative evaluation we collected a challenging dataset consisting of video and accelerometer data of a person preparing a mixed salad with accelerometer-equipped kitchen utensils. Trajectories from dense optical flow yielded a higher localization accuracy compared to point feature tracking.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121753647","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Visual Place Categorization in Indoor Environments 室内环境中的视觉场所分类
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.66
E. F. Ersi, John K. Tsotsos
This paper addresses the problem of visual place categorization, which aims at augmenting different locations of the environment visited by an autonomous robot with information that relates them to human-understandable concepts. We formulate the problem of visual place categorization in terms of energy minimization. To label visual observations with place categories we present a global image representation that is invariant to common changes in dynamic environments and robust against intra-class variations. To satisfy temporal consistency, a general solution is presented that incorporates statistical cues, without being restricted by constant and small neighbourhood radii, or being dependent on the actual path followed by the robot. A set of experiments on publicly available databases demonstrates the advantages of the presented system and show a significant improvement over available methods.
本文解决了视觉位置分类的问题,其目的是用与人类可理解的概念相关的信息来增强自主机器人所访问的环境的不同位置。我们从能量最小化的角度来阐述视觉场所分类的问题。为了用地点类别标记视觉观察,我们提出了一种全局图像表示,该表示对动态环境中的常见变化不变,对类内变化具有鲁棒性。为了满足时间一致性,提出了一种包含统计线索的通用解决方案,不受恒定和小的邻域半径的限制,也不依赖于机器人所遵循的实际路径。一组公开可用数据库的实验证明了所提出的系统的优点,并显示出比现有方法的显著改进。
{"title":"Visual Place Categorization in Indoor Environments","authors":"E. F. Ersi, John K. Tsotsos","doi":"10.1109/CRV.2012.66","DOIUrl":"https://doi.org/10.1109/CRV.2012.66","url":null,"abstract":"This paper addresses the problem of visual place categorization, which aims at augmenting different locations of the environment visited by an autonomous robot with information that relates them to human-understandable concepts. We formulate the problem of visual place categorization in terms of energy minimization. To label visual observations with place categories we present a global image representation that is invariant to common changes in dynamic environments and robust against intra-class variations. To satisfy temporal consistency, a general solution is presented that incorporates statistical cues, without being restricted by constant and small neighbourhood radii, or being dependent on the actual path followed by the robot. A set of experiments on publicly available databases demonstrates the advantages of the presented system and show a significant improvement over available methods.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"63 3","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114014790","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2012 Ninth Conference on Computer and Robot Vision
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1