首页 > 最新文献

2014 Canadian Conference on Computer and Robot Vision最新文献

英文 中文
N-Gram Based Image Representation and Classification Using Perceptual Shape Features 基于感知形状特征的N-Gram图像表示与分类
Pub Date : 2014-05-06 DOI: 10.1109/CRV.2014.54
Albina Mukanova, Q. Gao, Gang Hu
Rapid growth of visual data processing and analysis applications, such as content based image retrieval, augmented reality, automated inspection and defect detection, medical image understanding, and remote sensing has made the problem of developing accurate and efficient image representation and classification methods one of the key research areas. This research proposes new higher-level perceptual shape features for image representation which are based on Gestalt principles of human vision. The concept of n-gram is adapted from text analysis as a grouping mechanism for coding global shape content of an image. The proposed perceptual shape features are translation, rotation, and scale invariant. Local shape features and n-gram grouping scheme are integrated together to create new Perceptual Shape Vocabulary (PSV). Different image representations based on PSVs with and without n-gram scheme are applied to image classification task using Support Vector Machine (SVM) classifier. The experimental evaluation results indicate that n-gram-based perceptual shape features can efficiently represent global shape information of an image, and augment the accuracy of image representation by low-level image features such as SIFT descriptors.
基于内容的图像检索、增强现实、自动检测和缺陷检测、医学图像理解和遥感等视觉数据处理和分析应用的快速增长,使得开发准确、高效的图像表示和分类方法成为关键研究领域之一。本研究提出了基于人类视觉格式塔原理的更高层次的图像感知形状特征。n-gram的概念改编自文本分析,作为编码图像全局形状内容的分组机制。提出的感知形状特征是平移、旋转和尺度不变的。将局部形状特征与n图分组方案相结合,生成新的感知形状词汇表。利用支持向量机(SVM)分类器,将基于带有和不带有n-gram方案的psv的不同图像表示应用于图像分类任务。实验评价结果表明,基于n-gram的感知形状特征可以有效地表示图像的全局形状信息,并通过SIFT描述子等低级图像特征增强图像表示的准确性。
{"title":"N-Gram Based Image Representation and Classification Using Perceptual Shape Features","authors":"Albina Mukanova, Q. Gao, Gang Hu","doi":"10.1109/CRV.2014.54","DOIUrl":"https://doi.org/10.1109/CRV.2014.54","url":null,"abstract":"Rapid growth of visual data processing and analysis applications, such as content based image retrieval, augmented reality, automated inspection and defect detection, medical image understanding, and remote sensing has made the problem of developing accurate and efficient image representation and classification methods one of the key research areas. This research proposes new higher-level perceptual shape features for image representation which are based on Gestalt principles of human vision. The concept of n-gram is adapted from text analysis as a grouping mechanism for coding global shape content of an image. The proposed perceptual shape features are translation, rotation, and scale invariant. Local shape features and n-gram grouping scheme are integrated together to create new Perceptual Shape Vocabulary (PSV). Different image representations based on PSVs with and without n-gram scheme are applied to image classification task using Support Vector Machine (SVM) classifier. The experimental evaluation results indicate that n-gram-based perceptual shape features can efficiently represent global shape information of an image, and augment the accuracy of image representation by low-level image features such as SIFT descriptors.","PeriodicalId":385422,"journal":{"name":"2014 Canadian Conference on Computer and Robot Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127376244","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Trajectory Estimation Using Relative Distances Extracted from Inter-image Homographies 基于图像间同形词提取相对距离的轨迹估计
Pub Date : 2014-05-06 DOI: 10.1109/CRV.2014.39
Mårten Wadenbäck, A. Heyden
The main idea of this paper is to use distances between camera positions to recover the trajectory of a mobile robot. We consider a mobile platform equipped with a single fixed camera using images of the floor and their associated inter-image homographies to find these distances. We show that under the assumptions that the camera is rigidly mounted with a constant tilt and travelling at a constant height above the floor, the distance between two camera positions may be expressed in terms of the condition number of the inter-image homography. Experiments are conducted on synthetic data to verify that the derived distance formula gives distances close to the true ones and is not too sensitive to noise. We also describe how the robot trajectory may be represented as a graph with edge lengths determined by the distances computed using the formula above, and present one possible method to construct this graph given some of these distances. The experiments show promising results.
本文的主要思想是利用相机位置之间的距离来恢复移动机器人的轨迹。我们考虑了一个移动平台,配备了一个固定的相机,使用地板的图像及其相关的图像间同形异构词来找到这些距离。我们证明了在摄像机固定安装并保持恒定倾斜并在地板上保持恒定高度的假设下,两个摄像机位置之间的距离可以用图像间单应性的条件数来表示。在合成数据上进行了实验,验证了导出的距离公式与真实距离接近,并且对噪声不太敏感。我们还描述了机器人轨迹如何被表示为一个图形,其边缘长度由使用上述公式计算的距离决定,并给出了一种可能的方法来构建这个图形给定这些距离中的一些。实验显示出令人满意的结果。
{"title":"Trajectory Estimation Using Relative Distances Extracted from Inter-image Homographies","authors":"Mårten Wadenbäck, A. Heyden","doi":"10.1109/CRV.2014.39","DOIUrl":"https://doi.org/10.1109/CRV.2014.39","url":null,"abstract":"The main idea of this paper is to use distances between camera positions to recover the trajectory of a mobile robot. We consider a mobile platform equipped with a single fixed camera using images of the floor and their associated inter-image homographies to find these distances. We show that under the assumptions that the camera is rigidly mounted with a constant tilt and travelling at a constant height above the floor, the distance between two camera positions may be expressed in terms of the condition number of the inter-image homography. Experiments are conducted on synthetic data to verify that the derived distance formula gives distances close to the true ones and is not too sensitive to noise. We also describe how the robot trajectory may be represented as a graph with edge lengths determined by the distances computed using the formula above, and present one possible method to construct this graph given some of these distances. The experiments show promising results.","PeriodicalId":385422,"journal":{"name":"2014 Canadian Conference on Computer and Robot Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128296578","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Decentralized Cooperative Localization for Heterogeneous Multi-robot System Using Split Covariance Intersection Filter 基于分割协方差交叉滤波的异构多机器人系统分散协同定位
Pub Date : 2014-05-06 DOI: 10.1109/CRV.2014.30
Thumeera R. Wanasinghe, G. Mann, R. Gosine
This study proposes the use of a split covariance intersection filter (Split-CIF) for decentralized multi-robot cooperative localization. In the proposed method each robot maintains a local extended Kalman filter to estimate its own pose in a pre-defined reference frame. When a robot receives pose information from neighbouring robots it employs a Split-CIF-based approach to fuse this received measurement with its local belief. For a team of N mobile robots, the processing and communication complexity of the proposed method is linear, O(N), with respect to the number of robots in the team. The proposed method does not demand for fully connected synchronous communication channels between robots and can work with any asynchronous and partially connected communication network. Additionally, the proposed method gives consistent state updates and is capable of handling independent and interdependent parts of the estimations separately. The numerical simulations presented validate the proposed algorithm. The simulation results demonstrate that the proposed algorithm is outperformed compared to single-robot localization algorithms and also demonstrate approximately the same estimation accuracy as a centralized cooperative localization approach but with reduced computational cost.
本研究提出了一种分割协方差交叉滤波器(split - cif)用于分散多机器人协同定位。在该方法中,每个机器人保持一个局部扩展卡尔曼滤波器来估计其在预定义参考系中的姿态。当一个机器人从邻近的机器人那里接收到姿势信息时,它采用一种基于split - cif的方法将接收到的测量结果与它的局部信念融合在一起。对于一个由N个移动机器人组成的团队,该方法的处理和通信复杂度与团队中的机器人数量呈线性关系,为O(N)。该方法不需要机器人之间的完全连接的同步通信通道,可以在任何异步和部分连接的通信网络中工作。此外,所提出的方法提供一致的状态更新,并能够分别处理独立和相互依赖的估计部分。仿真结果验证了该算法的有效性。仿真结果表明,该算法优于单机器人定位算法,且估计精度与集中式协同定位方法大致相同,但计算成本较低。
{"title":"Decentralized Cooperative Localization for Heterogeneous Multi-robot System Using Split Covariance Intersection Filter","authors":"Thumeera R. Wanasinghe, G. Mann, R. Gosine","doi":"10.1109/CRV.2014.30","DOIUrl":"https://doi.org/10.1109/CRV.2014.30","url":null,"abstract":"This study proposes the use of a split covariance intersection filter (Split-CIF) for decentralized multi-robot cooperative localization. In the proposed method each robot maintains a local extended Kalman filter to estimate its own pose in a pre-defined reference frame. When a robot receives pose information from neighbouring robots it employs a Split-CIF-based approach to fuse this received measurement with its local belief. For a team of N mobile robots, the processing and communication complexity of the proposed method is linear, O(N), with respect to the number of robots in the team. The proposed method does not demand for fully connected synchronous communication channels between robots and can work with any asynchronous and partially connected communication network. Additionally, the proposed method gives consistent state updates and is capable of handling independent and interdependent parts of the estimations separately. The numerical simulations presented validate the proposed algorithm. The simulation results demonstrate that the proposed algorithm is outperformed compared to single-robot localization algorithms and also demonstrate approximately the same estimation accuracy as a centralized cooperative localization approach but with reduced computational cost.","PeriodicalId":385422,"journal":{"name":"2014 Canadian Conference on Computer and Robot Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125897884","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 37
An Integrated Bud Detection and Localization System for Application in Greenhouse Automation 一种应用于温室自动化的综合花蕾检测与定位系统
Pub Date : 2014-05-06 DOI: 10.1109/CRV.2014.53
Cole Tarry, Patrick Wspanialy, M. Veres, M. Moussa
This paper presents an integrated system for chrysanthemum bud detection that can be used to automate labour intensive tasks in floriculture greenhouses. The system will detect buds and their 3D location in order to guide a robot arm to perform selective pruning tasks on each plant. The detection algorithm is based on using radial hough transform. Testing on several samples showed promising results.
本文介绍了一种菊花芽检测集成系统,该系统可用于花卉栽培温室劳动密集型任务的自动化。该系统将检测花蕾及其3D位置,以指导机械臂对每棵植物执行选择性修剪任务。该检测算法基于径向霍夫变换。对几个样本的测试显示出令人鼓舞的结果。
{"title":"An Integrated Bud Detection and Localization System for Application in Greenhouse Automation","authors":"Cole Tarry, Patrick Wspanialy, M. Veres, M. Moussa","doi":"10.1109/CRV.2014.53","DOIUrl":"https://doi.org/10.1109/CRV.2014.53","url":null,"abstract":"This paper presents an integrated system for chrysanthemum bud detection that can be used to automate labour intensive tasks in floriculture greenhouses. The system will detect buds and their 3D location in order to guide a robot arm to perform selective pruning tasks on each plant. The detection algorithm is based on using radial hough transform. Testing on several samples showed promising results.","PeriodicalId":385422,"journal":{"name":"2014 Canadian Conference on Computer and Robot Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114694095","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Asymmetric Rendezvous Search at Sea 海上非对称集合搜索
Pub Date : 2014-05-06 DOI: 10.1109/CRV.2014.31
Malika Meghjani, F. Shkurti, J. A. G. Higuera, A. Kalmbach, David Whitney, G. Dudek
In this paper we address the rendezvous problem between an autonomous underwater vehicle (AUV) and a passively floating drifter on the sea surface. The AUV's mission is to keep an estimate of the floating drifter's position while exploring the underwater environment and periodically attempting to rendezvous with it. We are interested in the case where the AUV loses track of the drifter, predicts its location and searches for it in the vicinity of the predicted location. We parameterize this search problem with respect to both the uncertainty in the drifter's position estimate and the ratio between the drifter and the AUV speeds. We examine two search strategies for the AUV, an inward spiral and an outward spiral. We derive conditions under which these patterns are guaranteed to find a drifter, and we empirically analyze them with respect to different parameters in simulation. In addition, we present results from field trials in which an AUV successfully found a drifter after periods of communication loss during which the robot was exploring.
本文研究了自主水下航行器(AUV)与海面被动浮漂器的交会问题。AUV的任务是在探索水下环境的同时保持对浮动漂浮物位置的估计,并定期尝试与它会合。我们感兴趣的是在这种情况下,AUV失去了漂浮物的轨迹,预测了它的位置,并在预测位置附近搜索它。我们根据漂浮器位置估计的不确定性和漂浮器与AUV速度的比值对搜索问题进行了参数化。我们研究了水下机器人的两种搜索策略,向内螺旋和向外螺旋。我们推导了这些模式保证找到漂移的条件,并在模拟中对不同的参数进行了经验分析。此外,我们还介绍了现场试验的结果,其中AUV在机器人探索期间通信丢失一段时间后成功地找到了漂浮物。
{"title":"Asymmetric Rendezvous Search at Sea","authors":"Malika Meghjani, F. Shkurti, J. A. G. Higuera, A. Kalmbach, David Whitney, G. Dudek","doi":"10.1109/CRV.2014.31","DOIUrl":"https://doi.org/10.1109/CRV.2014.31","url":null,"abstract":"In this paper we address the rendezvous problem between an autonomous underwater vehicle (AUV) and a passively floating drifter on the sea surface. The AUV's mission is to keep an estimate of the floating drifter's position while exploring the underwater environment and periodically attempting to rendezvous with it. We are interested in the case where the AUV loses track of the drifter, predicts its location and searches for it in the vicinity of the predicted location. We parameterize this search problem with respect to both the uncertainty in the drifter's position estimate and the ratio between the drifter and the AUV speeds. We examine two search strategies for the AUV, an inward spiral and an outward spiral. We derive conditions under which these patterns are guaranteed to find a drifter, and we empirically analyze them with respect to different parameters in simulation. In addition, we present results from field trials in which an AUV successfully found a drifter after periods of communication loss during which the robot was exploring.","PeriodicalId":385422,"journal":{"name":"2014 Canadian Conference on Computer and Robot Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122450622","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Autonomous Lecture Recording with a PTZ Camera While Complying with Cinematographic Rules 在遵守电影规则的情况下,使用PTZ相机自动录制讲座
Pub Date : 2014-05-06 DOI: 10.1109/CRV.2014.57
D. Hulens, T. Goedemé, Tom Rumes
Nowadays, many lectures and presentations are recorded and broadcasted for teleteaching applications. When no human camera crew is present, the most obvious choice is for static cameras. In order to enhance the viewing experience, more advanced systems automatically track and steer the camera towards the lecturer. In this paper we propose an even more advanced system that tracks the lecturer while taking cinematographic rules into account. On top of that, the lecturer can be filmed in different types of shots. Our system is able to detect and track the position of the lecturer, even with non-static backgrounds and in difficult illumination. We developed an action axis determination system, needed to apply cinematographic rules and to steer the Pan-Tilt-Zoom (PTZ)camera towards the lecturer.
如今,许多讲座和报告被录下来并广播,用于远程教学。当没有摄制组在场时,最明显的选择是静态摄像机。为了增强观看体验,更先进的系统会自动跟踪并引导摄像机朝向讲师。在本文中,我们提出了一个更先进的系统,跟踪讲师,同时考虑到电影规则。最重要的是,讲师可以拍摄不同类型的镜头。我们的系统能够检测和跟踪讲师的位置,即使在非静态背景和困难的照明下。我们开发了一个动作轴确定系统,需要应用电影规则,并将Pan-Tilt-Zoom (PTZ)摄像机转向讲师。
{"title":"Autonomous Lecture Recording with a PTZ Camera While Complying with Cinematographic Rules","authors":"D. Hulens, T. Goedemé, Tom Rumes","doi":"10.1109/CRV.2014.57","DOIUrl":"https://doi.org/10.1109/CRV.2014.57","url":null,"abstract":"Nowadays, many lectures and presentations are recorded and broadcasted for teleteaching applications. When no human camera crew is present, the most obvious choice is for static cameras. In order to enhance the viewing experience, more advanced systems automatically track and steer the camera towards the lecturer. In this paper we propose an even more advanced system that tracks the lecturer while taking cinematographic rules into account. On top of that, the lecturer can be filmed in different types of shots. Our system is able to detect and track the position of the lecturer, even with non-static backgrounds and in difficult illumination. We developed an action axis determination system, needed to apply cinematographic rules and to steer the Pan-Tilt-Zoom (PTZ)camera towards the lecturer.","PeriodicalId":385422,"journal":{"name":"2014 Canadian Conference on Computer and Robot Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128173240","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
MDS-based Multi-axial Dimensionality Reduction Model for Human Action Recognition 基于mds的人体动作识别多轴降维模型
Pub Date : 2014-05-06 DOI: 10.1109/CRV.2014.42
Redha Touati, M. Mignotte
In this paper, we present an original and efficient method of human action recognition in a video sequence. The proposed model is based on the generation and fusion of a set of prototypes generated from different view-points of the data cube of the video sequence. More precisely, each prototype is generated by using a multidimensional scaling (MDS) based nonlinear dimensionality reduction technique both along the temporal axis but also along the spatial axis (row and column) of the binary video sequence of 2D silhouettes. This strategy aims at modeling each human action in a low dimensional space, as a trajectory of points or a specific curve, for each viewpoint of the video cube in a complementary way. A simple K-NN classifier is then used to classify the prototype, for a given viewpoint, associated with each action to be recognized and then the fusion of the classification results for each viewpoint allow us to significantly improve the recognition rate performance. The experiments of our approach have been conducted on the publicly available Weizmann data-set and show the sensitivity of the proposed recognition system to each individual viewpoint and the efficiency of our multi-viewpoint based fusion approach compared to the best existing state-of-the-art human action recognition methods recently proposed in the literature.
本文提出了一种新颖有效的视频序列人体动作识别方法。该模型基于从视频序列的数据立方体的不同视点生成的一组原型的生成和融合。更准确地说,每个原型都是通过基于多维尺度(MDS)的非线性降维技术沿着二维轮廓的二进制视频序列的时间轴和空间轴(行和列)生成的。该策略旨在以互补的方式为视频立方体的每个视点建模低维空间中的每个人类动作,作为点的轨迹或特定曲线。然后使用一个简单的K-NN分类器对原型进行分类,对于给定的视点,与要识别的每个动作相关联,然后对每个视点的分类结果进行融合,使我们能够显着提高识别率性能。我们的方法已经在公开可用的Weizmann数据集上进行了实验,并显示了所提出的识别系统对每个单独视点的敏感性,以及与文献中最近提出的最佳现有最先进的人类行为识别方法相比,我们基于多视点的融合方法的效率。
{"title":"MDS-based Multi-axial Dimensionality Reduction Model for Human Action Recognition","authors":"Redha Touati, M. Mignotte","doi":"10.1109/CRV.2014.42","DOIUrl":"https://doi.org/10.1109/CRV.2014.42","url":null,"abstract":"In this paper, we present an original and efficient method of human action recognition in a video sequence. The proposed model is based on the generation and fusion of a set of prototypes generated from different view-points of the data cube of the video sequence. More precisely, each prototype is generated by using a multidimensional scaling (MDS) based nonlinear dimensionality reduction technique both along the temporal axis but also along the spatial axis (row and column) of the binary video sequence of 2D silhouettes. This strategy aims at modeling each human action in a low dimensional space, as a trajectory of points or a specific curve, for each viewpoint of the video cube in a complementary way. A simple K-NN classifier is then used to classify the prototype, for a given viewpoint, associated with each action to be recognized and then the fusion of the classification results for each viewpoint allow us to significantly improve the recognition rate performance. The experiments of our approach have been conducted on the publicly available Weizmann data-set and show the sensitivity of the proposed recognition system to each individual viewpoint and the efficiency of our multi-viewpoint based fusion approach compared to the best existing state-of-the-art human action recognition methods recently proposed in the literature.","PeriodicalId":385422,"journal":{"name":"2014 Canadian Conference on Computer and Robot Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114044697","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
Computer Vision-Based Identification of Individual Turtles Using Characteristic Patterns of Their Plastrons 基于计算机视觉的海龟个体特征识别
Pub Date : 2014-05-06 DOI: 10.1109/CRV.2014.35
T. Beugeling, A. Albu
The identification of pond turtles is important to scientists who monitor local populations, as it allows them to track the growth and health of subjects over their lifetime. Traditional non-invasive methods for turtle recognition involve the visual inspection of distinctive coloured patterns on their plastron. This visual inspection is time consuming and difficult to scale with a potential growth in the surveyed population. We propose an algorithm for automatic identification of individual turtles based on images of their plastron. Our approach uses a combination of image processing and neural networks. We perform a convexity-concavity analysis of the contours on the plastron. The output of this analysis is combined with additional region-based measurements to compute feature vectors that characterize individual turtles. These features are used to train a neural network. Our goal is to create a neural network which is able to query a database of images of turtles of known identity with an image of an unknown turtle, and which outputs the unknown turtle's identity. The paper provides a thorough experimental evaluation of the proposed approach. Results are promising and point towards future work in the area of standardized image acquisition and image denoising.
对监测当地种群的科学家来说,识别塘龟很重要,因为这使他们能够追踪它们一生中的生长和健康状况。传统的非侵入性海龟识别方法包括目视检查其板上独特的彩色图案。这种目视检查非常耗时,而且很难根据调查人口的潜在增长进行扩展。本文提出了一种基于龟体图像的龟个体自动识别算法。我们的方法结合了图像处理和神经网络。我们对平板上的轮廓进行了凹凸分析。该分析的输出与其他基于区域的测量相结合,以计算表征单个海龟的特征向量。这些特征被用来训练神经网络。我们的目标是创建一个神经网络,它能够用未知海龟的图像查询数据库中已知身份的海龟的图像,并输出未知海龟的身份。本文对所提出的方法进行了全面的实验评估。结果是有希望的,并指出了在标准化图像采集和图像去噪领域的未来工作。
{"title":"Computer Vision-Based Identification of Individual Turtles Using Characteristic Patterns of Their Plastrons","authors":"T. Beugeling, A. Albu","doi":"10.1109/CRV.2014.35","DOIUrl":"https://doi.org/10.1109/CRV.2014.35","url":null,"abstract":"The identification of pond turtles is important to scientists who monitor local populations, as it allows them to track the growth and health of subjects over their lifetime. Traditional non-invasive methods for turtle recognition involve the visual inspection of distinctive coloured patterns on their plastron. This visual inspection is time consuming and difficult to scale with a potential growth in the surveyed population. We propose an algorithm for automatic identification of individual turtles based on images of their plastron. Our approach uses a combination of image processing and neural networks. We perform a convexity-concavity analysis of the contours on the plastron. The output of this analysis is combined with additional region-based measurements to compute feature vectors that characterize individual turtles. These features are used to train a neural network. Our goal is to create a neural network which is able to query a database of images of turtles of known identity with an image of an unknown turtle, and which outputs the unknown turtle's identity. The paper provides a thorough experimental evaluation of the proposed approach. Results are promising and point towards future work in the area of standardized image acquisition and image denoising.","PeriodicalId":385422,"journal":{"name":"2014 Canadian Conference on Computer and Robot Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121589188","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Indoor Scene Recognition with a Visual Attention-Driven Spatial Pooling Strategy 基于视觉注意力驱动的空间池策略的室内场景识别
Pub Date : 2014-05-06 DOI: 10.1109/CRV.2014.43
Tarek Elguebaly, N. Bouguila
Scene recognition is an important research topic in robotics and computer vision. Even though scene recognition is a problem that has been studied in depth, indoor scene categorization has had a slow progress. Indoor scene recognition is a challenging problem due to the severe high intra-class variability, mainly due to the intrinsic variety of objects that may be present, and inter-class similarities of man-made indoor structures. Therefore, most scene recognition techniques that work well for outdoor scenes demonstrate low performance on indoor scenes. Thus, in this paper, we present a simple, yet effective method for indoor scene recognition. Our approach can be illustrated as follows. First, we extract dense SIFT descriptors. Then, we combine a saliency-driven perceptual pooling with a simple spatial pooling scheme. Once the spatial and the saliency-driven encoding have been determined, we use vector quantization to compute histograms of local features from each sub-region. Later, the histograms from all sub-regions are concatenated together to generate the final representation of the image. Finally, a model based mixture classifier, which uses mixture models to characterize class densities, is applied. In order to address the problem of modeling non-Gaussian data which are largely present in our final representation of images, we use the generalized Gaussian mixture (GGM) which can be a good alternative to the Gaussian thanks to its shape flexibility. The learning of the proposed statistical model is carried out using the rival penalized expectation-maximization (RPEM) algorithm which is able to perform model selection and parameter learning together in a single step. Furthermore, we take into account the feature selection problem by determining a set of relevant features for each data cluster, so that we can speed up the used learning algorithm and get rid of noisy, redundant, or uninformative feature. To validate the proposed method we test it on the MIT indoor scenes data set.
场景识别是机器人技术和计算机视觉领域的一个重要研究课题。尽管场景识别是一个已经深入研究的问题,但室内场景分类却进展缓慢。室内场景识别是一个具有挑战性的问题,主要是由于可能存在的物体的内在多样性和人工室内结构的类间相似性。因此,大多数场景识别技术在室外场景中表现良好,但在室内场景中表现不佳。因此,本文提出了一种简单而有效的室内场景识别方法。我们的方法可以说明如下。首先,提取密集SIFT描述子。然后,我们将显著性驱动的感知池与简单的空间池方案相结合。一旦确定了空间和显著性驱动的编码,我们使用矢量量化来计算每个子区域的局部特征直方图。然后,将所有子区域的直方图连接在一起以生成图像的最终表示。最后,提出了一种基于混合模型的混合分类器,该分类器利用混合模型来表征类密度。为了解决非高斯数据的建模问题,我们使用广义高斯混合(GGM),由于其形状的灵活性,它可以成为高斯的一个很好的替代品。采用竞争惩罚期望最大化(RPEM)算法对所提出的统计模型进行学习,该算法能够在一步中同时完成模型选择和参数学习。此外,我们通过为每个数据簇确定一组相关特征来考虑特征选择问题,以便我们可以加快使用的学习算法并去除噪声,冗余或无信息的特征。为了验证所提出的方法,我们在MIT室内场景数据集上进行了测试。
{"title":"Indoor Scene Recognition with a Visual Attention-Driven Spatial Pooling Strategy","authors":"Tarek Elguebaly, N. Bouguila","doi":"10.1109/CRV.2014.43","DOIUrl":"https://doi.org/10.1109/CRV.2014.43","url":null,"abstract":"Scene recognition is an important research topic in robotics and computer vision. Even though scene recognition is a problem that has been studied in depth, indoor scene categorization has had a slow progress. Indoor scene recognition is a challenging problem due to the severe high intra-class variability, mainly due to the intrinsic variety of objects that may be present, and inter-class similarities of man-made indoor structures. Therefore, most scene recognition techniques that work well for outdoor scenes demonstrate low performance on indoor scenes. Thus, in this paper, we present a simple, yet effective method for indoor scene recognition. Our approach can be illustrated as follows. First, we extract dense SIFT descriptors. Then, we combine a saliency-driven perceptual pooling with a simple spatial pooling scheme. Once the spatial and the saliency-driven encoding have been determined, we use vector quantization to compute histograms of local features from each sub-region. Later, the histograms from all sub-regions are concatenated together to generate the final representation of the image. Finally, a model based mixture classifier, which uses mixture models to characterize class densities, is applied. In order to address the problem of modeling non-Gaussian data which are largely present in our final representation of images, we use the generalized Gaussian mixture (GGM) which can be a good alternative to the Gaussian thanks to its shape flexibility. The learning of the proposed statistical model is carried out using the rival penalized expectation-maximization (RPEM) algorithm which is able to perform model selection and parameter learning together in a single step. Furthermore, we take into account the feature selection problem by determining a set of relevant features for each data cluster, so that we can speed up the used learning algorithm and get rid of noisy, redundant, or uninformative feature. To validate the proposed method we test it on the MIT indoor scenes data set.","PeriodicalId":385422,"journal":{"name":"2014 Canadian Conference on Computer and Robot Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-05-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128321471","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Speed Daemon: Experience-Based Mobile Robot Speed Scheduling 速度守护程序:基于经验的移动机器人速度调度
Pub Date : 2014-05-01 DOI: 10.1109/CRV.2014.16
C. Ostafew, Angela P. Schoellig, T. Barfoot, J. Collier
A time-optimal speed schedule results in a mobile robot driving along a planned path at or near the limits of the robot's capability. However, deriving models to predict the effect of increased speed can be very difficult. In this paper, we present a speed scheduler that uses previous experience, instead of complex models, to generate time-optimal speed schedules. The algorithm is designed for a vision-based, path-repeating mobile robot and uses experience to ensure reliable localization, low path-tracking errors, and realizable control inputs while maximizing the speed along the path. To our knowledge, this is the first speed scheduler to incorporate experience from previous path traversals in order to address system constraints. The proposed speed scheduler was tested in over 4 km of path traversals in outdoor terrain using a large Ackermann-steered robot travelling between 0.5 m/s and 2.0 m/s. The approach to speed scheduling is shown to generate fast speed schedules while remaining within the limits of the robot's capability.
时间最优速度调度导致移动机器人沿着计划路径行驶,达到或接近机器人能力的极限。然而,推导模型来预测速度增加的影响是非常困难的。在本文中,我们提出了一个速度调度程序,它使用以往的经验,而不是复杂的模型,以产生时间最优的速度调度。该算法设计用于基于视觉的路径重复移动机器人,并利用经验确保可靠的定位,低路径跟踪误差和可实现的控制输入,同时最大化沿路径的速度。据我们所知,这是第一个结合以前的路径遍历经验来解决系统约束的速度调度器。所提出的速度调度器在室外地形中进行了超过4公里的路径穿越测试,使用的是一个大型阿克曼操纵机器人,速度在0.5米/秒到2.0米/秒之间。速度调度方法可以在机器人的能力范围内生成快速的速度调度。
{"title":"Speed Daemon: Experience-Based Mobile Robot Speed Scheduling","authors":"C. Ostafew, Angela P. Schoellig, T. Barfoot, J. Collier","doi":"10.1109/CRV.2014.16","DOIUrl":"https://doi.org/10.1109/CRV.2014.16","url":null,"abstract":"A time-optimal speed schedule results in a mobile robot driving along a planned path at or near the limits of the robot's capability. However, deriving models to predict the effect of increased speed can be very difficult. In this paper, we present a speed scheduler that uses previous experience, instead of complex models, to generate time-optimal speed schedules. The algorithm is designed for a vision-based, path-repeating mobile robot and uses experience to ensure reliable localization, low path-tracking errors, and realizable control inputs while maximizing the speed along the path. To our knowledge, this is the first speed scheduler to incorporate experience from previous path traversals in order to address system constraints. The proposed speed scheduler was tested in over 4 km of path traversals in outdoor terrain using a large Ackermann-steered robot travelling between 0.5 m/s and 2.0 m/s. The approach to speed scheduling is shown to generate fast speed schedules while remaining within the limits of the robot's capability.","PeriodicalId":385422,"journal":{"name":"2014 Canadian Conference on Computer and Robot Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122686716","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
期刊
2014 Canadian Conference on Computer and Robot Vision
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1