首页 > 最新文献

2017 14th Conference on Computer and Robot Vision (CRV)最新文献

英文 中文
Efficient Version-Space Reduction for Visual Tracking 有效的版本空间缩减视觉跟踪
Pub Date : 2017-04-02 DOI: 10.1109/CRV.2017.35
Kourosh Meshgi, Shigeyuki Oba, S. Ishii
Discriminative trackers, employ a classification approach to separate the target from its background. To cope with variations of the target shape and appearance, the classifier is updated online with different samples of the target and the background. Sample selection, labeling and updating the classifier is prone to various sources of errors that drift the tracker. We introduce the use of an efficient version space shrinking strategy to reduce the labeling errors and enhance its sampling strategy by measuring the uncertainty of the tracker about the samples. The proposed tracker, utilize an ensemble of classifiers that represents different hypotheses about the target, diversify them using boosting to provide a larger and more consistent coverage of the version-space and tune the classifiers' weights in voting. The proposed system adjusts the model update rate by promoting the co-training of the short-memory ensemble with a long-memory oracle. The proposed tracker outperformed state-of-the-art trackers on different sequences bearing various tracking challenges.
判别跟踪器采用分类方法将目标与其背景分离开来。为了应对目标形状和外观的变化,分类器使用目标和背景的不同样本在线更新。样本选择、标记和更新分类器容易受到各种错误来源的影响,从而使跟踪器漂移。我们引入了一种有效的版本空间收缩策略,通过测量跟踪器对样本的不确定性来减少标记误差并增强其采样策略。所提出的跟踪器,利用代表目标的不同假设的分类器的集合,使用提升来多样化它们,以提供更大、更一致的版本空间覆盖,并在投票中调整分类器的权重。该系统通过促进短记忆集合与长记忆集合的协同训练来调整模型更新速率。所提出的跟踪器在承载各种跟踪挑战的不同序列上优于最先进的跟踪器。
{"title":"Efficient Version-Space Reduction for Visual Tracking","authors":"Kourosh Meshgi, Shigeyuki Oba, S. Ishii","doi":"10.1109/CRV.2017.35","DOIUrl":"https://doi.org/10.1109/CRV.2017.35","url":null,"abstract":"Discriminative trackers, employ a classification approach to separate the target from its background. To cope with variations of the target shape and appearance, the classifier is updated online with different samples of the target and the background. Sample selection, labeling and updating the classifier is prone to various sources of errors that drift the tracker. We introduce the use of an efficient version space shrinking strategy to reduce the labeling errors and enhance its sampling strategy by measuring the uncertainty of the tracker about the samples. The proposed tracker, utilize an ensemble of classifiers that represents different hypotheses about the target, diversify them using boosting to provide a larger and more consistent coverage of the version-space and tune the classifiers' weights in voting. The proposed system adjusts the model update rate by promoting the co-training of the short-memory ensemble with a long-memory oracle. The proposed tracker outperformed state-of-the-art trackers on different sequences bearing various tracking challenges.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134036251","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Bootstrapping Labelled Dataset Construction for Cow Tracking and Behavior Analysis 用于奶牛跟踪和行为分析的自举标记数据集构建
Pub Date : 2017-03-30 DOI: 10.1109/CRV.2017.25
Aram Ter-Sarkisov, R. Ross, John D. Kelleher
This paper introduces a new approach to the long-term tracking of an object in a challenging environment. The object is a cow and the environment is an enclosure in a cowshed. Some of the key challenges in this domain are a cluttered background, low contrast and high similarity between moving objects – which greatly reduces the efficiency of most existing approaches, including those based on background subtraction. Our approach is split into object localization, instance segmentation, learning and tracking stages. Our solution is benchmarked against a range of semi-supervised object tracking algorithms and we show that the performance is strong and well suited to subsequent analysis. We present our solution as a first step towards broader tracking and behavior monitoring for cows in precision agriculture with the ultimate objective of early detection of lameness.
本文介绍了一种在复杂环境下对目标进行长期跟踪的新方法。对象是一头牛,环境是牛棚中的围栏。该领域的一些关键挑战是背景混乱,运动物体之间的低对比度和高相似性,这大大降低了大多数现有方法的效率,包括基于背景减法的方法。我们的方法分为对象定位、实例分割、学习和跟踪四个阶段。我们的解决方案针对一系列半监督对象跟踪算法进行了基准测试,我们表明性能很强,非常适合后续分析。我们提出了我们的解决方案,作为在精准农业中对奶牛进行更广泛的跟踪和行为监测的第一步,最终目标是早期发现跛行。
{"title":"Bootstrapping Labelled Dataset Construction for Cow Tracking and Behavior Analysis","authors":"Aram Ter-Sarkisov, R. Ross, John D. Kelleher","doi":"10.1109/CRV.2017.25","DOIUrl":"https://doi.org/10.1109/CRV.2017.25","url":null,"abstract":"This paper introduces a new approach to the long-term tracking of an object in a challenging environment. The object is a cow and the environment is an enclosure in a cowshed. Some of the key challenges in this domain are a cluttered background, low contrast and high similarity between moving objects – which greatly reduces the efficiency of most existing approaches, including those based on background subtraction. Our approach is split into object localization, instance segmentation, learning and tracking stages. Our solution is benchmarked against a range of semi-supervised object tracking algorithms and we show that the performance is strong and well suited to subsequent analysis. We present our solution as a first step towards broader tracking and behavior monitoring for cows in precision agriculture with the ultimate objective of early detection of lameness.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125171958","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Multi-path Region-Based Convolutional Neural Network for Accurate Detection of Unconstrained "Hard Faces" 基于多路径区域的卷积神经网络无约束“硬脸”精确检测
Pub Date : 2017-03-27 DOI: 10.1109/CRV.2017.20
Yuguang Liu, M. Levine
Large-scale variations still pose a challenge in unconstrained face detection. To the best of our knowledge, no current face detection algorithm can detect a face as large as 800 x 800 pixels while simultaneously detecting another one as small as 8 x 8 pixels within a single image with equally high accuracy. We propose a two-stage cascaded face detection framework, Multi-Path Region-based Convolutional Neural Network (MP-RCNN), that seamlessly combines a deep neural network with a classic learning strategy, to tackle this challenge. The first stage is a Multi-Path Region Proposal Network (MP-RPN) that proposes faces at three different scales. It simultaneously utilizes three parallel outputs of the convolutional feature maps to predict multi-scale candidate face regions. The "atrous" convolution trick (convolution with up-sampled filters) and a newly proposed sampling layer for "hard" examples are embedded in MP-RPN to further boost its performance. The second stage is a Boosted Forests classifier, which utilizes deep facial features pooled from inside the candidate face regions as well as deep contextual features pooled from a larger region surrounding the candidate face regions. This step is included to further remove hard negative samples. Experiments show that this approach achieves state-of-the-art face detection performance on the WIDER FACE dataset "hard" partition, outperforming the former best result by 9.6% for the Average Precision.
在无约束的人脸检测中,大规模变化仍然是一个挑战。据我们所知,目前还没有一种人脸检测算法能够在检测一张800 × 800像素大的人脸的同时,以同样高的精度检测一张8 × 8像素小的人脸。我们提出了一个两阶段级联的人脸检测框架,多路径基于区域的卷积神经网络(MP-RCNN),它将深度神经网络与经典学习策略无缝结合,以解决这一挑战。第一阶段是多路径区域建议网络(MP-RPN),该网络在三个不同的尺度上提出人脸。它同时利用卷积特征映射的三个并行输出来预测多尺度候选人脸区域。“atrous”卷积技巧(上采样滤波器的卷积)和新提出的“硬”样本采样层被嵌入MP-RPN中,以进一步提高其性能。第二阶段是增强森林分类器,它利用候选人脸区域内部的深度面部特征以及候选人脸区域周围更大区域的深度上下文特征。这一步包括进一步去除硬阴性样品。实验表明,该方法在更宽的face数据集“硬”分区上实现了最先进的人脸检测性能,平均精度比以前的最佳结果高出9.6%。
{"title":"Multi-path Region-Based Convolutional Neural Network for Accurate Detection of Unconstrained \"Hard Faces\"","authors":"Yuguang Liu, M. Levine","doi":"10.1109/CRV.2017.20","DOIUrl":"https://doi.org/10.1109/CRV.2017.20","url":null,"abstract":"Large-scale variations still pose a challenge in unconstrained face detection. To the best of our knowledge, no current face detection algorithm can detect a face as large as 800 x 800 pixels while simultaneously detecting another one as small as 8 x 8 pixels within a single image with equally high accuracy. We propose a two-stage cascaded face detection framework, Multi-Path Region-based Convolutional Neural Network (MP-RCNN), that seamlessly combines a deep neural network with a classic learning strategy, to tackle this challenge. The first stage is a Multi-Path Region Proposal Network (MP-RPN) that proposes faces at three different scales. It simultaneously utilizes three parallel outputs of the convolutional feature maps to predict multi-scale candidate face regions. The \"atrous\" convolution trick (convolution with up-sampled filters) and a newly proposed sampling layer for \"hard\" examples are embedded in MP-RPN to further boost its performance. The second stage is a Boosted Forests classifier, which utilizes deep facial features pooled from inside the candidate face regions as well as deep contextual features pooled from a larger region surrounding the candidate face regions. This step is included to further remove hard negative samples. Experiments show that this approach achieves state-of-the-art face detection performance on the WIDER FACE dataset \"hard\" partition, outperforming the former best result by 9.6% for the Average Precision.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"102 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-03-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134104678","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
4-DoF Tracking for Robot Fine Manipulation Tasks 机器人精细操作任务的四自由度跟踪
Pub Date : 2017-03-06 DOI: 10.1109/CRV.2017.41
Mennatullah Siam, Abhineet Singh, Camilo Perez, Martin Jägersand
This paper presents two visual trackers from the different paradigms of learning and registration based tracking and evaluates their application in image based visual servoing. They can track object motion with four degrees of freedom (DoF) which, as we will show here, is sufficient for many fine manipulation tasks. One of these trackers is a newly developed learning based tracker that relies on learning discriminative correlation filters while the other is a refinement of a recent 8 DoF RANSAC based tracker adapted with a new appearance model for tracking 4 DoF motion. Both trackers are shown to provide superior performance to several state of the art trackers on an existing dataset for manipulation tasks. Further, a new dataset with challenging sequences for fine manipulation tasks captured from robot mounted eye-in-hand (EIH) cameras is also presented. These sequences have a variety of challenges encountered during real tasks including jittery camera movement, motion blur, drastic scale changes and partial occlusions. Quantitative and qualitative results on these sequences are used to show that these two trackers are robust to failures while providing high precision that makes them suitable for such fine manipulation tasks.
本文介绍了基于学习的视觉跟踪和基于配准的视觉跟踪两种不同的模式,并评价了它们在基于图像的视觉伺服中的应用。它们可以用四个自由度(DoF)跟踪物体运动,正如我们将在这里展示的那样,这对于许多精细的操作任务来说已经足够了。其中一个跟踪器是新开发的基于学习的跟踪器,它依赖于学习判别相关滤波器,而另一个是对最近基于RANSAC的8 DoF跟踪器的改进,采用了新的外观模型来跟踪4 DoF运动。这两种跟踪器在操作任务的现有数据集上提供了优于几个最先进的跟踪器的性能。此外,还提出了一个具有挑战性序列的新数据集,用于从安装在机器人上的眼手相机(EIH)捕获的精细操作任务。这些序列在实际任务中遇到了各种各样的挑战,包括抖动的相机运动,运动模糊,剧烈的规模变化和部分遮挡。对这些序列的定量和定性结果表明,这两种跟踪器对故障具有鲁棒性,同时提供高精度,使其适用于此类精细操作任务。
{"title":"4-DoF Tracking for Robot Fine Manipulation Tasks","authors":"Mennatullah Siam, Abhineet Singh, Camilo Perez, Martin Jägersand","doi":"10.1109/CRV.2017.41","DOIUrl":"https://doi.org/10.1109/CRV.2017.41","url":null,"abstract":"This paper presents two visual trackers from the different paradigms of learning and registration based tracking and evaluates their application in image based visual servoing. They can track object motion with four degrees of freedom (DoF) which, as we will show here, is sufficient for many fine manipulation tasks. One of these trackers is a newly developed learning based tracker that relies on learning discriminative correlation filters while the other is a refinement of a recent 8 DoF RANSAC based tracker adapted with a new appearance model for tracking 4 DoF motion. Both trackers are shown to provide superior performance to several state of the art trackers on an existing dataset for manipulation tasks. Further, a new dataset with challenging sequences for fine manipulation tasks captured from robot mounted eye-in-hand (EIH) cameras is also presented. These sequences have a variety of challenges encountered during real tasks including jittery camera movement, motion blur, drastic scale changes and partial occlusions. Quantitative and qualitative results on these sequences are used to show that these two trackers are robust to failures while providing high precision that makes them suitable for such fine manipulation tasks.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-03-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128034957","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
The Ciona17 Dataset for Semantic Segmentation of Invasive Species in a Marine Aquaculture Environment 海洋水产养殖环境中入侵物种语义分割的Ciona17数据集
Pub Date : 2017-02-18 DOI: 10.5683/SP/NTUOK9
A. Galloway, Graham W. Taylor, Aaron Ramsay, M. Moussa
An original dataset for semantic segmentation, Ciona17, is introduced, which to the best of the authors' knowledge, is the first dataset of its kind with pixel-level annotations pertaining to invasive species in a marine environment. Diverse outdoor illumination, a range of object shapes, colour, and severe occlusion provide a significant real world challenge for the computer vision community. An accompanying ground-truthing tool for superpixel labeling, Truth and Crop, is also introduced. Finally, we provide a baseline using a variant of Fully Convolutional Networks, and report results in terms of the standard mean intersection over union (mIoU) metric.
介绍了语义分割的原始数据集Ciona17,据作者所知,这是同类数据集中第一个具有与海洋环境中入侵物种相关的像素级注释的数据集。不同的室外照明,各种物体形状,颜色和严重遮挡为计算机视觉社区提供了一个重大的现实世界挑战。此外,还介绍了一种用于超像素标记的地面真实工具Truth and Crop。最后,我们使用全卷积网络的一种变体提供了一个基线,并根据标准平均交联(mIoU)度量报告了结果。
{"title":"The Ciona17 Dataset for Semantic Segmentation of Invasive Species in a Marine Aquaculture Environment","authors":"A. Galloway, Graham W. Taylor, Aaron Ramsay, M. Moussa","doi":"10.5683/SP/NTUOK9","DOIUrl":"https://doi.org/10.5683/SP/NTUOK9","url":null,"abstract":"An original dataset for semantic segmentation, Ciona17, is introduced, which to the best of the authors' knowledge, is the first dataset of its kind with pixel-level annotations pertaining to invasive species in a marine environment. Diverse outdoor illumination, a range of object shapes, colour, and severe occlusion provide a significant real world challenge for the computer vision community. An accompanying ground-truthing tool for superpixel labeling, Truth and Crop, is also introduced. Finally, we provide a baseline using a variant of Fully Convolutional Networks, and report results in terms of the standard mean intersection over union (mIoU) metric.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-02-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131265682","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Building Damage Assessment Using Deep Learning and Ground-Level Image Data 使用深度学习和地面图像数据的建筑物损坏评估
Pub Date : 2017-01-20 DOI: 10.1109/CRV.2017.54
Karoon Rashedi Nia, Greg Mori
We propose a novel damage assessment deep model for buildings. Common damage assessment approaches utilize both pre-event and post-event data, which are not available in many cases. In this work, we focus on assessing damage to buildings using only post-disaster. We estimate severity of destruction via in a continuous fashion. Our model utilizes three different neural networks, one network for pre-processing the input data and two networks for extracting deep features from the input source. Combinations of these networks are distributed among three separate feature streams. A regressor summarizes the extracted features into a single continuous value denoting the destruction level. To evaluate the model, we collected a small dataset of ground-level image data of damaged buildings. Experimental results demonstrate that models taking advantage of hierarchical rich features outperform baseline methods.
提出了一种新的建筑物损伤评估深度模型。常见的损害评估方法利用事件发生前和事件发生后的数据,这在许多情况下是不可用的。在这项工作中,我们只关注灾后评估对建筑物的损害。我们以连续的方式估计破坏的严重程度。我们的模型使用三个不同的神经网络,一个网络用于预处理输入数据,两个网络用于从输入源提取深度特征。这些网络的组合分布在三个独立的特征流中。回归器将提取的特征总结为表示破坏程度的单个连续值。为了评估该模型,我们收集了一个小型的受损建筑物地面图像数据集。实验结果表明,利用层次丰富特征的模型优于基线方法。
{"title":"Building Damage Assessment Using Deep Learning and Ground-Level Image Data","authors":"Karoon Rashedi Nia, Greg Mori","doi":"10.1109/CRV.2017.54","DOIUrl":"https://doi.org/10.1109/CRV.2017.54","url":null,"abstract":"We propose a novel damage assessment deep model for buildings. Common damage assessment approaches utilize both pre-event and post-event data, which are not available in many cases. In this work, we focus on assessing damage to buildings using only post-disaster. We estimate severity of destruction via in a continuous fashion. Our model utilizes three different neural networks, one network for pre-processing the input data and two networks for extracting deep features from the input source. Combinations of these networks are distributed among three separate feature streams. A regressor summarizes the extracted features into a single continuous value denoting the destruction level. To evaluate the model, we collected a small dataset of ground-level image data of damaged buildings. Experimental results demonstrate that models taking advantage of hierarchical rich features outperform baseline methods.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114269584","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 49
Multiple Object Tracking with Kernelized Correlation Filters in Urban Mixed Traffic 基于核化相关滤波器的城市混合交通多目标跟踪
Pub Date : 2016-11-08 DOI: 10.1109/CRV.2017.18
Yuebin Yang, Guillaume-Alexandre Bilodeau
Recently, the Kernelized Correlation Filters tracker (KCF) achieved competitive performance and robustness in visual object tracking. On the other hand, visual trackers are not typically used in multiple object tracking. In this paper, we investigate how a robust visual tracker like KCF can improve multiple object tracking. Since KCF is a fast tracker, many KCF can be used in parallel and still result in fast tracking. We built a multiple object tracking system based on KCF and background subtraction. Background subtraction is applied to extract moving objects and get their scale and size in combination with KCF outputs, while KCF is used for data association and to handle fragmentation and occlusion problems. As a result, KCF and background subtraction help each other to take tracking decision at every frame. Sometimes KCF outputs are the most trustworthy (e.g. during occlusion), while in some other cases, it is the background subtraction outputs. To validate the effectiveness of our system, the algorithm was tested on four urban traffic videos from a standard dataset. Results show that our method is competitive with state-of-the-art trackers even if we use a much simpler data association step.
近年来,核化相关滤波器跟踪器在视觉目标跟踪中取得了较好的性能和鲁棒性。另一方面,视觉跟踪器通常不用于多目标跟踪。在本文中,我们研究了像KCF这样的鲁棒视觉跟踪器如何改善多目标跟踪。由于KCF是一个快速跟踪器,许多KCF可以并行使用,并且仍然导致快速跟踪。我们构建了一个基于KCF和背景相减的多目标跟踪系统。结合KCF输出,利用背景减法提取运动目标,得到运动目标的尺度和大小,利用KCF进行数据关联,处理碎片和遮挡问题。因此,KCF和背景减法在每一帧都能相互帮助做出跟踪决策。有时KCF输出是最可信的(例如在遮挡期间),而在其他一些情况下,它是背景减法输出。为了验证我们系统的有效性,我们对来自标准数据集的四个城市交通视频进行了测试。结果表明,即使我们使用更简单的数据关联步骤,我们的方法也与最先进的跟踪器相竞争。
{"title":"Multiple Object Tracking with Kernelized Correlation Filters in Urban Mixed Traffic","authors":"Yuebin Yang, Guillaume-Alexandre Bilodeau","doi":"10.1109/CRV.2017.18","DOIUrl":"https://doi.org/10.1109/CRV.2017.18","url":null,"abstract":"Recently, the Kernelized Correlation Filters tracker (KCF) achieved competitive performance and robustness in visual object tracking. On the other hand, visual trackers are not typically used in multiple object tracking. In this paper, we investigate how a robust visual tracker like KCF can improve multiple object tracking. Since KCF is a fast tracker, many KCF can be used in parallel and still result in fast tracking. We built a multiple object tracking system based on KCF and background subtraction. Background subtraction is applied to extract moving objects and get their scale and size in combination with KCF outputs, while KCF is used for data association and to handle fragmentation and occlusion problems. As a result, KCF and background subtraction help each other to take tracking decision at every frame. Sometimes KCF outputs are the most trustworthy (e.g. during occlusion), while in some other cases, it is the background subtraction outputs. To validate the effectiveness of our system, the algorithm was tested on four urban traffic videos from a standard dataset. Results show that our method is competitive with state-of-the-art trackers even if we use a much simpler data association step.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"108 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123650486","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 27
Unbiased Sparse Subspace Clustering by Selective Pursuit 基于选择性追踪的无偏稀疏子空间聚类
Pub Date : 2016-09-16 DOI: 10.1109/CRV.2017.28
H. Ackermann, B. Rosenhahn, M. Yang
Sparse subspace clustering (SSC) is an elegant approach for unsupervised segmentation if the data points of each cluster are located in linear subspaces. This model applies, for instance, in motion segmentation if some restrictions on the camera model hold. SSC requires that problems based on the l1-norm are solved to infer which points belong to the same subspace. If these unknown subspaces are well-separated this algorithm is guaranteed to succeed. The question how the distribution of points on the same subspace effects their clustering has received less attention. One case has been reported in which points of the same model are erroneously classified to belong to different subspaces. In this work, it will be theoretically shown when and why such spurious clusters occur. This claim is further substantiated by experimental evidence. Two algorithms based on the Dantzig selector and subspace selector are proposed to overcome this problem, and good results are reported.
稀疏子空间聚类(SSC)是一种很好的无监督分割方法,如果每个聚类的数据点都位于线性子空间中。这个模型适用于,例如,在运动分割中,如果相机模型的一些限制保持不变。SSC要求解决基于11范数的问题来推断哪些点属于同一子空间。如果这些未知子空间分离良好,则保证算法成功。关于点在同一子空间上的分布如何影响它们的聚类的问题很少受到关注。曾经报道过一种情况,同一模型的点被错误地分类为属于不同的子空间。在这项工作中,它将在理论上显示何时以及为什么这种虚假集群发生。实验证据进一步证实了这一说法。提出了基于Dantzig选择器和子空间选择器的两种算法来克服这一问题,并取得了良好的效果。
{"title":"Unbiased Sparse Subspace Clustering by Selective Pursuit","authors":"H. Ackermann, B. Rosenhahn, M. Yang","doi":"10.1109/CRV.2017.28","DOIUrl":"https://doi.org/10.1109/CRV.2017.28","url":null,"abstract":"Sparse subspace clustering (SSC) is an elegant approach for unsupervised segmentation if the data points of each cluster are located in linear subspaces. This model applies, for instance, in motion segmentation if some restrictions on the camera model hold. SSC requires that problems based on the l1-norm are solved to infer which points belong to the same subspace. If these unknown subspaces are well-separated this algorithm is guaranteed to succeed. The question how the distribution of points on the same subspace effects their clustering has received less attention. One case has been reported in which points of the same model are erroneously classified to belong to different subspaces. In this work, it will be theoretically shown when and why such spurious clusters occur. This claim is further substantiated by experimental evidence. Two algorithms based on the Dantzig selector and subspace selector are proposed to overcome this problem, and good results are reported.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"131 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124641080","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Autonomous Flying Cameraman with Embedded Person Detection and Tracking while Applying Cinematographic Rules 在应用电影规则的情况下,具有嵌入式人员检测和跟踪的自主飞行摄像机
Pub Date : 1900-01-01 DOI: 10.1109/CRV.2017.27
D. Hulens, T. Goedemé
Unmanned Aerial Vehicles (UAVs) enable numerous applications such as search and rescue operations, structural inspection of buildings, crop growth analysis in agriculture, performing 3D reconstruction and so on. For such applications, currently the UAV is steered manually. However, in this paper we aim to record semi-professional video footage (e.g. concerts, sport events) using fully autonomous UAVs. Evidently, this is challenging since we need to detect and track the actor on-board a UAV in real-time, while automatically – and smoothly – controlling the UAV based on these detections. For this, all four DOF (Degrees of freedom) are controlled in separate simultaneous control loops by our vision-based algorithms. Furthermore cinematographic rules need to be taken into account (e.g. the rule of thirds) which position the actor at the visually optimal location in the frame. We extensively validated our algorithms: each control loop and the overall final system is thoroughly evaluated with respect to both accuracy and control speed. We show that our system is able to efficiently control the UAV such that professional recordings are obtained.
无人驾驶飞行器(uav)可以实现许多应用,如搜索和救援行动,建筑物结构检查,农业作物生长分析,执行3D重建等。对于此类应用,目前无人机是手动操纵的。然而,在本文中,我们的目标是使用全自动无人机记录半专业视频片段(例如音乐会,体育赛事)。显然,这是具有挑战性的,因为我们需要实时检测和跟踪无人机上的行动者,同时基于这些检测自动平滑地控制无人机。为此,所有四个自由度(DOF)由我们基于视觉的算法在单独的同步控制回路中进行控制。此外,需要考虑电影规则(如三分法),将演员置于画面中视觉效果最佳的位置。我们广泛地验证了我们的算法:每个控制回路和整个最终系统都在精度和控制速度方面进行了彻底的评估。我们表明,我们的系统能够有效地控制无人机,从而获得专业的录音。
{"title":"Autonomous Flying Cameraman with Embedded Person Detection and Tracking while Applying Cinematographic Rules","authors":"D. Hulens, T. Goedemé","doi":"10.1109/CRV.2017.27","DOIUrl":"https://doi.org/10.1109/CRV.2017.27","url":null,"abstract":"Unmanned Aerial Vehicles (UAVs) enable numerous applications such as search and rescue operations, structural inspection of buildings, crop growth analysis in agriculture, performing 3D reconstruction and so on. For such applications, currently the UAV is steered manually. However, in this paper we aim to record semi-professional video footage (e.g. concerts, sport events) using fully autonomous UAVs. Evidently, this is challenging since we need to detect and track the actor on-board a UAV in real-time, while automatically – and smoothly – controlling the UAV based on these detections. For this, all four DOF (Degrees of freedom) are controlled in separate simultaneous control loops by our vision-based algorithms. Furthermore cinematographic rules need to be taken into account (e.g. the rule of thirds) which position the actor at the visually optimal location in the frame. We extensively validated our algorithms: each control loop and the overall final system is thoroughly evaluated with respect to both accuracy and control speed. We show that our system is able to efficiently control the UAV such that professional recordings are obtained.","PeriodicalId":308760,"journal":{"name":"2017 14th Conference on Computer and Robot Vision (CRV)","volume":"116 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128943058","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
期刊
2017 14th Conference on Computer and Robot Vision (CRV)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1