首页 > 最新文献

2018 15th Conference on Computer and Robot Vision (CRV)最新文献

英文 中文
Systematic Street View Sampling: High Quality Annotation of Power Infrastructure in Rural Ontario 系统街景采样:安大略省农村电力基础设施的高质量注释
Pub Date : 2018-12-13 DOI: 10.1109/CRV.2018.00028
K. Dick, François Charih, Yasmina Souley Dosso, L. Russell, J. Green
Google Street View and the emergence of self-driving vehicles afford an unprecedented capacity to observe our planet. Fused with dramatic advances in artificial intelligence, the capability to extract patterns and meaning from those data streams heralds an era of insights into the physical world. In order to draw appropriate inferences about and between environments, the systematic selection of these data is necessary to create representative and unbiased samples. To this end, we introduce the Systematic Street View Sampler (S3) framework, enabling researchers to produce their own user-defined datasets of Street View imagery. We describe the algorithm and express its asymptotic complexity in relation to a new limiting computational resource (Google API Call Count). Using the Amazon Mechanical Turk distributed annotation environment, we demonstrate the utility of S3 in generating high quality representative datasets useful for machine vision applications. The S3 algorithm is open-source and available at github.com/CU-BIC/S3 along with the high quality dataset representing power infrastructure in rural regions of southern Ontario, Canada.
街景和自动驾驶汽车的出现提供了前所未有的能力来观察我们的星球。与人工智能的巨大进步相结合,从这些数据流中提取模式和意义的能力预示着一个洞察物理世界的时代。为了得出关于环境和环境之间的适当推论,系统地选择这些数据是必要的,以创建具有代表性和无偏的样本。为此,我们引入了系统街景采样器(S3)框架,使研究人员能够生成他们自己的用户定义的街景图像数据集。我们描述了该算法,并表示了它的渐近复杂性与一个新的限制计算资源(谷歌API调用计数)。使用Amazon Mechanical Turk分布式注释环境,我们演示了S3在生成机器视觉应用中有用的高质量代表性数据集方面的实用性。S3算法是开源的,可在github.com/CU-BIC/S3上获得,以及代表加拿大安大略省南部农村地区电力基础设施的高质量数据集。
{"title":"Systematic Street View Sampling: High Quality Annotation of Power Infrastructure in Rural Ontario","authors":"K. Dick, François Charih, Yasmina Souley Dosso, L. Russell, J. Green","doi":"10.1109/CRV.2018.00028","DOIUrl":"https://doi.org/10.1109/CRV.2018.00028","url":null,"abstract":"Google Street View and the emergence of self-driving vehicles afford an unprecedented capacity to observe our planet. Fused with dramatic advances in artificial intelligence, the capability to extract patterns and meaning from those data streams heralds an era of insights into the physical world. In order to draw appropriate inferences about and between environments, the systematic selection of these data is necessary to create representative and unbiased samples. To this end, we introduce the Systematic Street View Sampler (S3) framework, enabling researchers to produce their own user-defined datasets of Street View imagery. We describe the algorithm and express its asymptotic complexity in relation to a new limiting computational resource (Google API Call Count). Using the Amazon Mechanical Turk distributed annotation environment, we demonstrate the utility of S3 in generating high quality representative datasets useful for machine vision applications. The S3 algorithm is open-source and available at github.com/CU-BIC/S3 along with the high quality dataset representing power infrastructure in rural regions of southern Ontario, Canada.","PeriodicalId":281779,"journal":{"name":"2018 15th Conference on Computer and Robot Vision (CRV)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125521934","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Eye on the Sky: An Upward-Looking Monocular Teach-and-Repeat System for Indoor Environments 天空之眼:一种面向室内环境的单目教学与重复系统
Pub Date : 2018-05-08 DOI: 10.1109/CRV.2018.00056
N. Zhang, M. Warren, T. Barfoot
Visual Teach and Repeat (VT&R) allows a robotic vehicle to navigate autonomously along a network of paths in the presence of illumination and scene changes. Traditionally, the system uses a stereo camera as the primary sensor for triangulating visual landmarks and often operates in highly textured outdoor environments. In this paper, we modify the VT&R system to use a monocular pipeline under the same framework, but also target indoor operation as a demonstration of a low-cost VT&R solution for warehouse logistics in a visually difficult environment. Unlike previous monocular VT&R solutions, we make no assumptions about the nature of the scene (e.g., local ground planarity). This allows the system to be readily deployable on more vehicles in a wider range of environments. To test the system, and motivated by a warehouse navigation application, an upward pointing camera is mounted on a Clearpath Husky ground vehicle. We demonstrate the vehicle is able to navigate with a 99.6% autonomy rate using such a system during 1.1 kilometers of driving, with an average scene depth that varies from 8-16 meters. The cross-track deviation from the taught path is less than 0.5 meters over 90% of the path, reaching a maximum of 0.85 meters and an average of 0.26 meters.
视觉教学和重复(VT&R)允许机器人车辆在照明和场景变化的情况下沿着路径网络自主导航。传统上,该系统使用立体摄像机作为主要传感器来三角化视觉地标,通常在高度纹理的户外环境中运行。在本文中,我们修改了VT&R系统,在相同的框架下使用单目管道,但也针对室内操作,作为在视觉困难环境下仓库物流的低成本VT&R解决方案的演示。与以前的单目VT&R解决方案不同,我们没有对场景的性质(例如,局部地面平面度)进行假设。这使得该系统可以在更广泛的环境中部署在更多的车辆上。为了测试该系统,在仓库导航应用程序的激励下,在Clearpath Husky地面车辆上安装了一个向上指向的摄像头。我们证明,在1.1公里的行驶中,车辆能够使用该系统以99.6%的自主率导航,平均场景深度从8米到16米不等。90%以上的路径与教学路径的交叉偏差小于0.5米,最大偏差为0.85米,平均偏差为0.26米。
{"title":"Eye on the Sky: An Upward-Looking Monocular Teach-and-Repeat System for Indoor Environments","authors":"N. Zhang, M. Warren, T. Barfoot","doi":"10.1109/CRV.2018.00056","DOIUrl":"https://doi.org/10.1109/CRV.2018.00056","url":null,"abstract":"Visual Teach and Repeat (VT&R) allows a robotic vehicle to navigate autonomously along a network of paths in the presence of illumination and scene changes. Traditionally, the system uses a stereo camera as the primary sensor for triangulating visual landmarks and often operates in highly textured outdoor environments. In this paper, we modify the VT&R system to use a monocular pipeline under the same framework, but also target indoor operation as a demonstration of a low-cost VT&R solution for warehouse logistics in a visually difficult environment. Unlike previous monocular VT&R solutions, we make no assumptions about the nature of the scene (e.g., local ground planarity). This allows the system to be readily deployable on more vehicles in a wider range of environments. To test the system, and motivated by a warehouse navigation application, an upward pointing camera is mounted on a Clearpath Husky ground vehicle. We demonstrate the vehicle is able to navigate with a 99.6% autonomy rate using such a system during 1.1 kilometers of driving, with an average scene depth that varies from 8-16 meters. The cross-track deviation from the taught path is less than 0.5 meters over 90% of the path, reaching a maximum of 0.85 meters and an average of 0.26 meters.","PeriodicalId":281779,"journal":{"name":"2018 15th Conference on Computer and Robot Vision (CRV)","volume":"80 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134140536","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On the Robustness of Deep Learning Models to Universal Adversarial Attack 深度学习模型对通用对抗性攻击的鲁棒性研究
Pub Date : 2018-05-08 DOI: 10.1109/CRV.2018.00018
Rezaul Karim, Md. Amirul Islam, N. Mohammed, Neil D. B. Bruce
In recent years, there have been significant advances in deep learning applied to problems in high-level vision tasks (e.g. image classification, object detection, semantic segmentation etc.) which has been met with a great deal of success. State-of-the-art methods that have shown impressive results on recognition tasks typically share a common structure involving stage-wise encoding of the image, followed by a generic classifier. However, these architectures have been shown to be vulnerable to the adversarial perturbations which may undermine the security of the systems supported by deep neural nets. In this work, initially we present rigorous evaluation of adversarial attacks on recent deep learning models for two different high-level tasks (image classification and semantic segmentation). Then we propose a model and dataset independent approach to generate adversarial perturbation and also the transferability of perturbation across different datasets and tasks. Moreover, we analyze the effect of different network architectures which will aid future efforts in understanding and defending against adversarial perturbations. We perform comprehensive experiments on several standard image classification and segmentation datasets to demonstrate the effectiveness of our proposed approach.
近年来,深度学习在高级视觉任务(如图像分类、目标检测、语义分割等)中的应用取得了重大进展,并取得了很大的成功。在识别任务中显示出令人印象深刻的结果的最先进的方法通常具有一个共同的结构,包括图像的分阶段编码,然后是通用分类器。然而,这些架构已被证明容易受到对抗性扰动的影响,这可能会破坏深度神经网络支持的系统的安全性。在这项工作中,我们首先对两种不同高级任务(图像分类和语义分割)的最新深度学习模型的对抗性攻击进行了严格的评估。然后,我们提出了一种独立于模型和数据集的方法来产生对抗性扰动,以及扰动在不同数据集和任务之间的可转移性。此外,我们分析了不同网络架构的影响,这将有助于未来理解和防御对抗性扰动的努力。我们在几个标准图像分类和分割数据集上进行了全面的实验,以证明我们提出的方法的有效性。
{"title":"On the Robustness of Deep Learning Models to Universal Adversarial Attack","authors":"Rezaul Karim, Md. Amirul Islam, N. Mohammed, Neil D. B. Bruce","doi":"10.1109/CRV.2018.00018","DOIUrl":"https://doi.org/10.1109/CRV.2018.00018","url":null,"abstract":"In recent years, there have been significant advances in deep learning applied to problems in high-level vision tasks (e.g. image classification, object detection, semantic segmentation etc.) which has been met with a great deal of success. State-of-the-art methods that have shown impressive results on recognition tasks typically share a common structure involving stage-wise encoding of the image, followed by a generic classifier. However, these architectures have been shown to be vulnerable to the adversarial perturbations which may undermine the security of the systems supported by deep neural nets. In this work, initially we present rigorous evaluation of adversarial attacks on recent deep learning models for two different high-level tasks (image classification and semantic segmentation). Then we propose a model and dataset independent approach to generate adversarial perturbation and also the transferability of perturbation across different datasets and tasks. Moreover, we analyze the effect of different network architectures which will aid future efforts in understanding and defending against adversarial perturbations. We perform comprehensive experiments on several standard image classification and segmentation datasets to demonstrate the effectiveness of our proposed approach.","PeriodicalId":281779,"journal":{"name":"2018 15th Conference on Computer and Robot Vision (CRV)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131898478","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Online Model Adaptation for UAV Tracking with Convolutional Neural Network 基于卷积神经网络的无人机在线模型自适应跟踪
Pub Date : 2018-05-08 DOI: 10.1109/CRV.2018.00053
Zhuojin Sun, Yong Wang, R. Laganière
Unmanned aerial vehicle (UAV) tracking is a challenging problem and a core component of UAV applications. CNNs have shown impressive performance in computer vision applications, such as object detection, image classification and so on. In this work, a locally connected layer is employed in a CNN architecture to extract robust features. We also utilize focal loss function to focus training on hard examples. Our CNN is first pre-trained offline to learn robust features. The training data is classified according to the texture, color, size of the target and the background information properties. In a subsequent online tracking phase, this CNN is fine-tuned to adapt to the appearance changes of the tracked target. We applied this approach to the problem of UAV tracking and performed extensive experimental results on large scale benchmark datasets. Results obtained show that the proposed method performs favorably against the state-of-the-art trackers in terms of accuracy, robustness and efficiency.
无人机跟踪是一个具有挑战性的问题,也是无人机应用的核心组成部分。cnn在目标检测、图像分类等计算机视觉应用中表现出了令人印象深刻的性能。在这项工作中,在CNN架构中使用局部连接层来提取鲁棒特征。我们还利用焦点损失函数将训练集中在困难的例子上。我们的CNN首先进行离线预训练,以学习鲁棒特征。根据目标的纹理、颜色、大小和背景信息属性对训练数据进行分类。在随后的在线跟踪阶段,该CNN被微调以适应被跟踪目标的外观变化。我们将这种方法应用于无人机跟踪问题,并在大规模基准数据集上进行了广泛的实验结果。结果表明,该方法在精度、鲁棒性和效率方面优于现有的跟踪器。
{"title":"Online Model Adaptation for UAV Tracking with Convolutional Neural Network","authors":"Zhuojin Sun, Yong Wang, R. Laganière","doi":"10.1109/CRV.2018.00053","DOIUrl":"https://doi.org/10.1109/CRV.2018.00053","url":null,"abstract":"Unmanned aerial vehicle (UAV) tracking is a challenging problem and a core component of UAV applications. CNNs have shown impressive performance in computer vision applications, such as object detection, image classification and so on. In this work, a locally connected layer is employed in a CNN architecture to extract robust features. We also utilize focal loss function to focus training on hard examples. Our CNN is first pre-trained offline to learn robust features. The training data is classified according to the texture, color, size of the target and the background information properties. In a subsequent online tracking phase, this CNN is fine-tuned to adapt to the appearance changes of the tracked target. We applied this approach to the problem of UAV tracking and performed extensive experimental results on large scale benchmark datasets. Results obtained show that the proposed method performs favorably against the state-of-the-art trackers in terms of accuracy, robustness and efficiency.","PeriodicalId":281779,"journal":{"name":"2018 15th Conference on Computer and Robot Vision (CRV)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133752651","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
A Hierarchical Deep Architecture and Mini-batch Selection Method for Joint Traffic Sign and Light Detection 交通标志与信号灯联合检测的层次深度结构与小批量选择方法
Pub Date : 2018-05-08 DOI: 10.1109/CRV.2018.00024
Alex D. Pon, Oles Andrienko, Ali Harakeh, Steven L. Waslander
Traffic light and sign detectors on autonomous cars are integral for road scene perception. The literature is abundant with deep learning networks that detect either lights or signs, not both, which makes them unsuitable for real-life deployment due to the limited graphics processing unit (GPU) memory and power available on embedded systems. The root cause of this issue is that no public dataset contains both traffic light and sign labels, which leads to difficulties in developing a joint detection framework. We present a deep hierarchical architecture in conjunction with a mini-batch proposal selection mechanism that allows a network to detect both traffic lights and signs from training on separate traffic light and sign datasets. Our method solves the overlapping issue where instances from one dataset are not labelled in the other dataset. We are the first to present a network that performs joint detection on traffic lights and signs. We measure our network on the Tsinghua-Tencent 100K benchmark for traffic sign detection and the Bosch Small Traffic Lights benchmark for traffic light detection and show it outperforms the existing Bosch Small Traffic light state-of-the-art method. We focus on autonomous car deployment and show our network is more suitable than others because of its low memory footprint and real-time image processing time. Qualitative results can be viewed at https://youtu.be/ YmogPzBXOw.
自动驾驶汽车上的交通灯和标志探测器是道路场景感知不可或缺的一部分。文献中有大量深度学习网络只能检测灯光或标志,而不能同时检测两者,这使得它们不适合实际部署,因为嵌入式系统上的图形处理单元(GPU)内存和功率有限。这个问题的根本原因是没有公共数据集包含交通灯和标志标签,这导致开发联合检测框架的困难。我们提出了一个深度层次结构,结合一个小批量建议选择机制,允许网络从单独的交通灯和标志数据集的训练中检测交通灯和标志。我们的方法解决了一个数据集的实例在另一个数据集中没有标记的重叠问题。我们是第一个提出对交通信号灯和标志进行联合检测的网络。我们在清华-腾讯100K交通标志检测基准和博世小交通灯交通信号灯检测基准上测试了我们的网络,结果表明它优于现有的博世小交通灯最先进的方法。我们专注于自动驾驶汽车的部署,并证明我们的网络比其他网络更适合,因为它的内存占用少,实时图像处理时间长。定性结果可在https://youtu.be/ YmogPzBXOw上查看。
{"title":"A Hierarchical Deep Architecture and Mini-batch Selection Method for Joint Traffic Sign and Light Detection","authors":"Alex D. Pon, Oles Andrienko, Ali Harakeh, Steven L. Waslander","doi":"10.1109/CRV.2018.00024","DOIUrl":"https://doi.org/10.1109/CRV.2018.00024","url":null,"abstract":"Traffic light and sign detectors on autonomous cars are integral for road scene perception. The literature is abundant with deep learning networks that detect either lights or signs, not both, which makes them unsuitable for real-life deployment due to the limited graphics processing unit (GPU) memory and power available on embedded systems. The root cause of this issue is that no public dataset contains both traffic light and sign labels, which leads to difficulties in developing a joint detection framework. We present a deep hierarchical architecture in conjunction with a mini-batch proposal selection mechanism that allows a network to detect both traffic lights and signs from training on separate traffic light and sign datasets. Our method solves the overlapping issue where instances from one dataset are not labelled in the other dataset. We are the first to present a network that performs joint detection on traffic lights and signs. We measure our network on the Tsinghua-Tencent 100K benchmark for traffic sign detection and the Bosch Small Traffic Lights benchmark for traffic light detection and show it outperforms the existing Bosch Small Traffic light state-of-the-art method. We focus on autonomous car deployment and show our network is more suitable than others because of its low memory footprint and real-time image processing time. Qualitative results can be viewed at https://youtu.be/ YmogPzBXOw.","PeriodicalId":281779,"journal":{"name":"2018 15th Conference on Computer and Robot Vision (CRV)","volume":"101 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124608282","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 35
Counting Static Targets Using an Unmanned Aerial Vehicle On-the-Fly and Autonomously 基于无人机的静态目标动态自动计数
Pub Date : 2018-05-08 DOI: 10.1109/CRV.2018.00037
Dhruvil Darji, G. Vejarano
The counting of static targets on ground using an unmanned aerial vehicle (UAV) is proposed. To the best of our knowledge, this is the first paper to do such counting on-the-fly and autonomously. The flight path is programmed before take-off. The UAV captures images of the ground which are processed consecutively on-the-fly to count the number of targets along the flight path. Each image is processed using the proposed target-counting algorithm. First, targets' centers are detected in the current image, and second, the targets that were not covered in previous images are identified and counted. The performance of the algorithm depends on its ability to identify in the current image what targets were already counted in previous images, and this ability is affected by the limited accuracy of the UAV to stay on the flight path in the presence of wind. In the experimental evaluation, targets were distributed on ground on three different configurations: one line of targets along the flight path, parallel lines of targets at an angle with the flight path, and random. The accuracy of the target count was 96.0%, 88.9% and 91.9% respectively.
提出了利用无人机对地面静态目标进行计数的方法。据我们所知,这是第一篇即时自动计数的论文。飞行路线是在起飞前设定好的。无人机捕获地面图像,这些图像在飞行中连续处理,以计算飞行路径上的目标数量。使用所提出的目标计数算法对每个图像进行处理。首先在当前图像中检测目标的中心,然后对之前图像中未覆盖的目标进行识别和计数。该算法的性能取决于其识别当前图像中已经在先前图像中计算的目标的能力,并且这种能力受到无人机在存在风的情况下保持飞行路径的有限精度的影响。在实验评估中,地面目标按三种不同配置进行分布:目标沿航迹直线分布、目标与航迹成一定角度平行线分布、目标随机分布。目标计数正确率分别为96.0%、88.9%和91.9%。
{"title":"Counting Static Targets Using an Unmanned Aerial Vehicle On-the-Fly and Autonomously","authors":"Dhruvil Darji, G. Vejarano","doi":"10.1109/CRV.2018.00037","DOIUrl":"https://doi.org/10.1109/CRV.2018.00037","url":null,"abstract":"The counting of static targets on ground using an unmanned aerial vehicle (UAV) is proposed. To the best of our knowledge, this is the first paper to do such counting on-the-fly and autonomously. The flight path is programmed before take-off. The UAV captures images of the ground which are processed consecutively on-the-fly to count the number of targets along the flight path. Each image is processed using the proposed target-counting algorithm. First, targets' centers are detected in the current image, and second, the targets that were not covered in previous images are identified and counted. The performance of the algorithm depends on its ability to identify in the current image what targets were already counted in previous images, and this ability is affected by the limited accuracy of the UAV to stay on the flight path in the presence of wind. In the experimental evaluation, targets were distributed on ground on three different configurations: one line of targets along the flight path, parallel lines of targets at an angle with the flight path, and random. The accuracy of the target count was 96.0%, 88.9% and 91.9% respectively.","PeriodicalId":281779,"journal":{"name":"2018 15th Conference on Computer and Robot Vision (CRV)","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128315828","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Survey of Monocular SLAM Algorithms in Natural Environments 自然环境下单目SLAM算法研究进展
Pub Date : 2018-05-08 DOI: 10.1109/CRV.2018.00055
Georges Chahine, C. Pradalier
With the increased use of cameras in robotic applications, this paper presents quantitative and qualitative assessment of the most prominent monocular tracking and mapping algorithms in literature with a particular focus on natural environments. This paper is unique in both context and methodology since it quantifies the performance of the state-of-the-art in Visual Simultaneous Localization and Mapping methods in the specific context where images mostly include vegetation. Finally, we elaborate on the limitations of these algorithms and the challenges that they did not address or consider when working in the natural environment.
随着相机在机器人应用中的使用越来越多,本文对文献中最突出的单目跟踪和映射算法进行了定量和定性评估,特别关注自然环境。本文在背景和方法上都是独一无二的,因为它量化了图像主要包括植被的特定背景下视觉同步定位和地图绘制方法的最新性能。最后,我们详细阐述了这些算法的局限性以及它们在自然环境中工作时没有解决或考虑的挑战。
{"title":"Survey of Monocular SLAM Algorithms in Natural Environments","authors":"Georges Chahine, C. Pradalier","doi":"10.1109/CRV.2018.00055","DOIUrl":"https://doi.org/10.1109/CRV.2018.00055","url":null,"abstract":"With the increased use of cameras in robotic applications, this paper presents quantitative and qualitative assessment of the most prominent monocular tracking and mapping algorithms in literature with a particular focus on natural environments. This paper is unique in both context and methodology since it quantifies the performance of the state-of-the-art in Visual Simultaneous Localization and Mapping methods in the specific context where images mostly include vegetation. Finally, we elaborate on the limitations of these algorithms and the challenges that they did not address or consider when working in the natural environment.","PeriodicalId":281779,"journal":{"name":"2018 15th Conference on Computer and Robot Vision (CRV)","volume":"156 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123092573","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
De-noising of Lidar Point Clouds Corrupted by Snowfall 被降雪破坏的激光雷达点云去噪
Pub Date : 2018-05-08 DOI: 10.1109/CRV.2018.00043
Nicholas Charron, Stephen Phillips, Steven L. Waslander
A common problem in autonomous driving is designing a system that can operate in adverse weather conditions. Falling rain and snow tends to corrupt sensor measurements, particularly for lidar sensors. Surprisingly, very little research has been published on methods to de-noise point clouds which are collected by lidar in rainy or snowy weather conditions. In this paper, we present a method for removing snow noise by processing point clouds using a 3D outlier detection algorithm. Our method, the dynamic radius outlier removal filter, accounts for the variation in point cloud density with increasing distance from the sensor, with the goal of removing the noise caused by snow while retaining detail in environmental features (which is necessary for autonomous localization and navigation). The proposed method outperforms other noise-removal methods, including methods which operate on depth image representations of the lidar scans. We show on point clouds obtained while driving in falling snow that we can simultaneously obtain > 90% precision and recall, indicating that the proposed method is effective at removing snow, without removing environmental features.
自动驾驶的一个常见问题是设计一个可以在恶劣天气条件下运行的系统。降雨和降雪往往会破坏传感器测量,特别是激光雷达传感器。令人惊讶的是,关于在雨雪天气条件下激光雷达收集的点云去噪方法的研究很少发表。本文提出了一种利用三维离群点检测算法对点云进行去除雪噪声的方法。我们的方法,动态半径离群值去除过滤器,考虑到点云密度随距离传感器的增加而变化,其目标是去除雪引起的噪声,同时保留环境特征的细节(这对于自主定位和导航是必要的)。该方法优于其他降噪方法,包括对激光雷达扫描的深度图像表示进行操作的方法。在降雪天驾驶时获得的点云上,我们可以同时获得90%的准确率和召回率,这表明我们提出的方法在不去除环境特征的情况下可以有效地去除积雪。
{"title":"De-noising of Lidar Point Clouds Corrupted by Snowfall","authors":"Nicholas Charron, Stephen Phillips, Steven L. Waslander","doi":"10.1109/CRV.2018.00043","DOIUrl":"https://doi.org/10.1109/CRV.2018.00043","url":null,"abstract":"A common problem in autonomous driving is designing a system that can operate in adverse weather conditions. Falling rain and snow tends to corrupt sensor measurements, particularly for lidar sensors. Surprisingly, very little research has been published on methods to de-noise point clouds which are collected by lidar in rainy or snowy weather conditions. In this paper, we present a method for removing snow noise by processing point clouds using a 3D outlier detection algorithm. Our method, the dynamic radius outlier removal filter, accounts for the variation in point cloud density with increasing distance from the sensor, with the goal of removing the noise caused by snow while retaining detail in environmental features (which is necessary for autonomous localization and navigation). The proposed method outperforms other noise-removal methods, including methods which operate on depth image representations of the lidar scans. We show on point clouds obtained while driving in falling snow that we can simultaneously obtain > 90% precision and recall, indicating that the proposed method is effective at removing snow, without removing environmental features.","PeriodicalId":281779,"journal":{"name":"2018 15th Conference on Computer and Robot Vision (CRV)","volume":"310 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116758554","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 85
Automotive Semi-specular Surface Defect Detection System 汽车半镜面表面缺陷检测系统
Pub Date : 2018-05-08 DOI: 10.1109/CRV.2018.00047
Adarsh Tandiya, S. Akhtar, M. Moussa, Cole Tarry
In this work, a deflectometry based efficient semi-specular surface defect detection system is developed. This system, when integrated with a robotic arm, can detect defects on large and widely varying topological surfaces like a car bumper. A hybrid pipe line, that utilizes multi-threading, is designed to efficiently use resources and speedup the inspection process. Specific filters are also designed to eliminate spurious defects due to edges and acute curvature changes. The developed system was successful in consistently detecting various defects on large bumper parts with varying topology and color and can accommodate inherent ambient lighting and vibration issues.
本文研制了一种基于偏转法的高效半镜面表面缺陷检测系统。当该系统与机械臂集成时,可以检测像汽车保险杠这样的大而广泛变化的拓扑表面上的缺陷。采用多线程技术的混合式管线可以有效地利用资源,加快检测过程。特定的过滤器也被设计用来消除由于边缘和急剧曲率变化而产生的虚假缺陷。开发的系统成功地检测了具有不同拓扑结构和颜色的大型保险杠部件的各种缺陷,并且可以适应固有的环境光照和振动问题。
{"title":"Automotive Semi-specular Surface Defect Detection System","authors":"Adarsh Tandiya, S. Akhtar, M. Moussa, Cole Tarry","doi":"10.1109/CRV.2018.00047","DOIUrl":"https://doi.org/10.1109/CRV.2018.00047","url":null,"abstract":"In this work, a deflectometry based efficient semi-specular surface defect detection system is developed. This system, when integrated with a robotic arm, can detect defects on large and widely varying topological surfaces like a car bumper. A hybrid pipe line, that utilizes multi-threading, is designed to efficiently use resources and speedup the inspection process. Specific filters are also designed to eliminate spurious defects due to edges and acute curvature changes. The developed system was successful in consistently detecting various defects on large bumper parts with varying topology and color and can accommodate inherent ambient lighting and vibration issues.","PeriodicalId":281779,"journal":{"name":"2018 15th Conference on Computer and Robot Vision (CRV)","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120996873","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Gaze Selection for Enhanced Visual Odometry During Navigation 导航过程中增强视觉里程计的凝视选择
Pub Date : 2018-05-08 DOI: 10.1109/CRV.2018.00025
Travis Manderson, Andrew Holliday, G. Dudek
We present an approach to enhancing visual odometry and Simultaneous Localization and Mapping (SLAM) in the context of robot navigation by actively modulating the gaze direction to enhance the quality of the odometric estimates that are returned. We focus on two quality factors: i) stability of the visual features, and ii) consistency of the visual features with respect to robot motion and the associated correspondence between frames. We assume that local texture measures are associated with underlying scene content and thus with the quality of the visual features for the associated region of the scene. Based on this assumption, we train a machine-learning system to score different regions of an image based on their texture and then guide the robot's gaze toward high scoring image regions. Our work is targeted towards motion estimation and SLAM for small, lightweight, and autonomous air vehicles where computational resources are constrained in weight, size, and power. However, we believe that our work is also applicable to other types of robotic systems. Our experimental validation consists of simulations, constrained tests, and outdoor flight experiments on an unmanned aerial vehicle. We find that modulating gaze direction can improve localization accuracy by up to 62 percent.
我们提出了一种在机器人导航环境下通过主动调节凝视方向来增强视觉里程计和同步定位和映射(SLAM)的方法,以提高返回的里程估计的质量。我们关注两个质量因素:i)视觉特征的稳定性,以及ii)视觉特征与机器人运动和帧之间相关对应的一致性。我们假设局部纹理度量与底层场景内容相关,因此与场景相关区域的视觉特征质量相关。基于这一假设,我们训练了一个机器学习系统,根据图像的纹理对图像的不同区域进行评分,然后引导机器人的视线转向得分较高的图像区域。我们的工作是针对小型、轻量级和自主飞行器的运动估计和SLAM,这些飞行器的计算资源受到重量、尺寸和功率的限制。然而,我们相信我们的工作也适用于其他类型的机器人系统。我们的实验验证包括模拟、约束测试和无人机户外飞行实验。我们发现,调节凝视方向可以提高定位精度高达62%。
{"title":"Gaze Selection for Enhanced Visual Odometry During Navigation","authors":"Travis Manderson, Andrew Holliday, G. Dudek","doi":"10.1109/CRV.2018.00025","DOIUrl":"https://doi.org/10.1109/CRV.2018.00025","url":null,"abstract":"We present an approach to enhancing visual odometry and Simultaneous Localization and Mapping (SLAM) in the context of robot navigation by actively modulating the gaze direction to enhance the quality of the odometric estimates that are returned. We focus on two quality factors: i) stability of the visual features, and ii) consistency of the visual features with respect to robot motion and the associated correspondence between frames. We assume that local texture measures are associated with underlying scene content and thus with the quality of the visual features for the associated region of the scene. Based on this assumption, we train a machine-learning system to score different regions of an image based on their texture and then guide the robot's gaze toward high scoring image regions. Our work is targeted towards motion estimation and SLAM for small, lightweight, and autonomous air vehicles where computational resources are constrained in weight, size, and power. However, we believe that our work is also applicable to other types of robotic systems. Our experimental validation consists of simulations, constrained tests, and outdoor flight experiments on an unmanned aerial vehicle. We find that modulating gaze direction can improve localization accuracy by up to 62 percent.","PeriodicalId":281779,"journal":{"name":"2018 15th Conference on Computer and Robot Vision (CRV)","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-05-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122976278","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
期刊
2018 15th Conference on Computer and Robot Vision (CRV)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1