首页 > 最新文献

IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision最新文献

英文 中文
Composite Discriminant Factor analysis 复合判别因子分析
Vlad I. Morariu, Ejaz Ahmed, Venkataraman Santhanam, David Harwood, L. Davis
We propose a linear dimensionality reduction method, Composite Discriminant Factor (CDF) analysis, which searches for a discriminative but compact feature subspace that can be used as input to classifiers that suffer from problems such as multi-collinearity or the curse of dimensionality. The subspace selected by CDF maximizes the performance of the entire classification pipeline, and is chosen from a set of candidate subspaces that are each discriminative. Our method is based on Partial Least Squares (PLS) analysis, and can be viewed as a generalization of the PLS1 algorithm, designed to increase discrimination in classification tasks. We demonstrate our approach on the UCF50 action recognition dataset, two object detection datasets (INRIA pedestrians and vehicles from aerial imagery), and machine learning datasets from the UCI Machine Learning repository. Experimental results show that the proposed approach improves significantly in terms of accuracy over linear SVM, and also over PLS in terms of compactness and efficiency, while maintaining or improving accuracy.
我们提出了一种线性降维方法,即复合判别因子(CDF)分析,该方法搜索一个判别但紧凑的特征子空间,该子空间可用于遭受多重共线性或维数诅咒等问题的分类器的输入。CDF选择的子空间最大化了整个分类管道的性能,并且是从一组候选子空间中选择的,每个子空间都是有区别的。我们的方法基于偏最小二乘(PLS)分析,可以看作是PLS1算法的推广,旨在提高分类任务的辨别能力。我们在UCF50动作识别数据集、两个目标检测数据集(来自航空图像的INRIA行人和车辆)和来自UCI机器学习存储库的机器学习数据集上展示了我们的方法。实验结果表明,该方法在保持或提高精度的同时,在精度方面比线性支持向量机有显著提高,在紧凑性和效率方面也比PLS有显著提高。
{"title":"Composite Discriminant Factor analysis","authors":"Vlad I. Morariu, Ejaz Ahmed, Venkataraman Santhanam, David Harwood, L. Davis","doi":"10.1109/WACV.2014.6836052","DOIUrl":"https://doi.org/10.1109/WACV.2014.6836052","url":null,"abstract":"We propose a linear dimensionality reduction method, Composite Discriminant Factor (CDF) analysis, which searches for a discriminative but compact feature subspace that can be used as input to classifiers that suffer from problems such as multi-collinearity or the curse of dimensionality. The subspace selected by CDF maximizes the performance of the entire classification pipeline, and is chosen from a set of candidate subspaces that are each discriminative. Our method is based on Partial Least Squares (PLS) analysis, and can be viewed as a generalization of the PLS1 algorithm, designed to increase discrimination in classification tasks. We demonstrate our approach on the UCF50 action recognition dataset, two object detection datasets (INRIA pedestrians and vehicles from aerial imagery), and machine learning datasets from the UCI Machine Learning repository. Experimental results show that the proposed approach improves significantly in terms of accuracy over linear SVM, and also over PLS in terms of compactness and efficiency, while maintaining or improving accuracy.","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90376019","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Improving multiview face detection with multi-task deep convolutional neural networks 用多任务深度卷积神经网络改进多视图人脸检测
Cha Zhang, Zhengyou Zhang
Multiview face detection is a challenging problem due to dramatic appearance changes under various pose, illumination and expression conditions. In this paper, we present a multi-task deep learning scheme to enhance the detection performance. More specifically, we build a deep convolutional neural network that can simultaneously learn the face/nonface decision, the face pose estimation problem, and the facial landmark localization problem. We show that such a multi-task learning scheme can further improve the classifier's accuracy. On the challenging FDDB data set, our detector achieves over 3% improvement in detection rate at the same false positive rate compared with other state-of-the-art methods.
多视角人脸检测是一个具有挑战性的问题,因为在不同的姿势、光照和表情条件下,人脸的外观会发生巨大的变化。在本文中,我们提出了一种多任务深度学习方案来提高检测性能。更具体地说,我们构建了一个可以同时学习人脸/非人脸决策、人脸姿态估计问题和人脸地标定位问题的深度卷积神经网络。我们证明了这种多任务学习方案可以进一步提高分类器的准确率。在具有挑战性的FDDB数据集上,与其他最先进的方法相比,我们的检测器在相同的误报率下实现了超过3%的检测率提高。
{"title":"Improving multiview face detection with multi-task deep convolutional neural networks","authors":"Cha Zhang, Zhengyou Zhang","doi":"10.1109/WACV.2014.6835990","DOIUrl":"https://doi.org/10.1109/WACV.2014.6835990","url":null,"abstract":"Multiview face detection is a challenging problem due to dramatic appearance changes under various pose, illumination and expression conditions. In this paper, we present a multi-task deep learning scheme to enhance the detection performance. More specifically, we build a deep convolutional neural network that can simultaneously learn the face/nonface decision, the face pose estimation problem, and the facial landmark localization problem. We show that such a multi-task learning scheme can further improve the classifier's accuracy. On the challenging FDDB data set, our detector achieves over 3% improvement in detection rate at the same false positive rate compared with other state-of-the-art methods.","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73431647","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 203
Relative facial action unit detection 相对面部动作单元检测
M. Khademi, Louis-Philippe Morency
This paper presents a subject-independent facial action unit (AU) detection method by introducing the concept of relative AU detection, for scenarios where the neutral face is not provided. We propose a new classification objective function which analyzes the temporal neighborhood of the current frame to decide if the expression recently increased, decreased or showed no change. This approach is a significant change from the conventional absolute method which decides about AU classification using the current frame, without an explicit comparison with its neighboring frames. Our proposed method improves robustness to individual differences such as face scale and shape, age-related wrinkles, and transitions among expressions (e.g., lower intensity of expressions). Our experiments on three publicly available datasets (Extended Cohn-Kanade (CK+), Bosphorus, and DISFA databases) show significant improvement of our approach over conventional absolute techniques.
本文通过引入相对面部动作单元检测的概念,提出了一种与主体无关的面部动作单元(AU)检测方法,用于不提供中性面部的场景。我们提出了一种新的分类目标函数,该函数通过分析当前帧的时间邻域来判断表达式最近是否增加、减少或没有变化。该方法与传统的绝对分类方法相比是一个重大的变化,传统的绝对分类方法使用当前帧来决定AU分类,而不需要与相邻帧进行明确的比较。我们提出的方法提高了对个体差异的鲁棒性,如面部尺度和形状,与年龄相关的皱纹,以及表情之间的过渡(例如,表情强度较低)。我们在三个公开可用的数据集(Extended Cohn-Kanade (CK+), Bosphorus和DISFA数据库)上的实验表明,我们的方法比传统的绝对技术有了显著的改进。
{"title":"Relative facial action unit detection","authors":"M. Khademi, Louis-Philippe Morency","doi":"10.1109/WACV.2014.6835983","DOIUrl":"https://doi.org/10.1109/WACV.2014.6835983","url":null,"abstract":"This paper presents a subject-independent facial action unit (AU) detection method by introducing the concept of relative AU detection, for scenarios where the neutral face is not provided. We propose a new classification objective function which analyzes the temporal neighborhood of the current frame to decide if the expression recently increased, decreased or showed no change. This approach is a significant change from the conventional absolute method which decides about AU classification using the current frame, without an explicit comparison with its neighboring frames. Our proposed method improves robustness to individual differences such as face scale and shape, age-related wrinkles, and transitions among expressions (e.g., lower intensity of expressions). Our experiments on three publicly available datasets (Extended Cohn-Kanade (CK+), Bosphorus, and DISFA databases) show significant improvement of our approach over conventional absolute techniques.","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84428784","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Real-time 3D page tracking and book status recognition for high-speed book digitization based on adaptive capturing 基于自适应捕获的高速数字化图书实时三维页面跟踪与图书状态识别
Shohei Noguchi, M. Yamada, Yoshihiro Watanabe, M. Ishikawa
In this paper, we propose a new book digitization system that can obtain high-resolution document images while flipping the pages automatically. The distinctive feature of our system is the adaptive capturing that has a crucial role in achieving high speed and high resolution. This adaptive capturing requires observing the state of the flipped pages at high speed and with high accuracy. In order to meet this requirement, we newly propose a method of obtaining the 3D shape of the book, tracking each page, and evaluating the state. In addition, we explain the details of the proposed high-speed book digitization system. We also report some experiments conducted to verify the performance of the developed system.
在本文中,我们提出了一种新的图书数字化系统,可以在自动翻页的同时获得高分辨率的文档图像。本系统的特点是自适应捕获,它对实现高速度和高分辨率具有至关重要的作用。这种自适应捕获要求以高速度和高精度观察翻转页面的状态。为了满足这一要求,我们提出了一种获取图书三维形状、跟踪每一页、评估状态的方法。此外,我们还解释了所提出的高速图书数字化系统的细节。我们还报道了一些实验来验证所开发系统的性能。
{"title":"Real-time 3D page tracking and book status recognition for high-speed book digitization based on adaptive capturing","authors":"Shohei Noguchi, M. Yamada, Yoshihiro Watanabe, M. Ishikawa","doi":"10.1109/WACV.2014.6836108","DOIUrl":"https://doi.org/10.1109/WACV.2014.6836108","url":null,"abstract":"In this paper, we propose a new book digitization system that can obtain high-resolution document images while flipping the pages automatically. The distinctive feature of our system is the adaptive capturing that has a crucial role in achieving high speed and high resolution. This adaptive capturing requires observing the state of the flipped pages at high speed and with high accuracy. In order to meet this requirement, we newly propose a method of obtaining the 3D shape of the book, tracking each page, and evaluating the state. In addition, we explain the details of the proposed high-speed book digitization system. We also report some experiments conducted to verify the performance of the developed system.","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85222491","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
NRSfM using local rigidity NRSfM使用局部刚性
A. Rehan, Aamer Zaheer, Ijaz Akhter, Arfah Saeed, M. Usmani, Bilal Mahmood, Sohaib Khan
In this paper we show that typical nonrigid structure can often be approximated well as locally rigid sub-structures in time and space. Specifically, we assume that: 1) the structure can be approximated as rigid in a short local time window and 2) some point- pairs stay relatively rigid in space, maintaining a fixed distance between them during the sequence. First, we use the triangulation constraints in rigid SfM over a sliding time window to get an initial estimate of the nonrigid 3D structure. Then we automatically identify relatively rigid point-pairs in this structure, and use their length-constancy simultaneously with triangulation constraints to refine the structure estimate. Local factorization inherently handles small camera motion, short sequences and significant natural occlusions gracefully, performing better than nonrigid factorization methods. We show more stable and accurate results as compared to the state-of-the art on even short sequences starting from 15 frames only, containing camera rotations as small as 2° and up to 50% contiguous missing data.
本文证明了典型的非刚性结构在时间和空间上往往可以很好地近似为局部刚性子结构。具体来说,我们假设:1)结构在短的局部时间窗内可以近似为刚性;2)一些点对在空间上保持相对刚性,在序列中它们之间保持固定距离。首先,我们在滑动时间窗口上使用刚性SfM中的三角化约束来获得非刚性3D结构的初始估计。然后,我们自动识别出该结构中相对刚性的点对,并同时使用它们的长度常数与三角化约束来改进结构估计。局部分解固有地处理小摄像机运动、短序列和显著的自然遮挡,比非刚性分解方法表现得更好。我们展示了更稳定和准确的结果,相比于最先进的,甚至从15帧开始的短序列,包含相机旋转小至2°和高达50%的连续缺失数据。
{"title":"NRSfM using local rigidity","authors":"A. Rehan, Aamer Zaheer, Ijaz Akhter, Arfah Saeed, M. Usmani, Bilal Mahmood, Sohaib Khan","doi":"10.1109/WACV.2014.6836116","DOIUrl":"https://doi.org/10.1109/WACV.2014.6836116","url":null,"abstract":"In this paper we show that typical nonrigid structure can often be approximated well as locally rigid sub-structures in time and space. Specifically, we assume that: 1) the structure can be approximated as rigid in a short local time window and 2) some point- pairs stay relatively rigid in space, maintaining a fixed distance between them during the sequence. First, we use the triangulation constraints in rigid SfM over a sliding time window to get an initial estimate of the nonrigid 3D structure. Then we automatically identify relatively rigid point-pairs in this structure, and use their length-constancy simultaneously with triangulation constraints to refine the structure estimate. Local factorization inherently handles small camera motion, short sequences and significant natural occlusions gracefully, performing better than nonrigid factorization methods. We show more stable and accurate results as compared to the state-of-the art on even short sequences starting from 15 frames only, containing camera rotations as small as 2° and up to 50% contiguous missing data.","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81993638","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
Max residual classifier 最大残差分类器
H. Nguyen, Vishal M. Patel
We introduce a novel classifier, called max residual classifier (MRC), for learning a sparse representation jointly with a discriminative decision function. MRC seeks to maximize the differences between the residual errors of the wrong classes and the right one. This effectively leads to a more discriminative sparse representation and better classification accuracy. The optimization procedure is simple and efficient. Its objective function is closely related to the decision function of the residual classification strategy. Unlike existing methods for learning discriminative sparse representation that are restricted to a linear model, our approach is able to work with a non-linear model via the use of Mercer kernel. Experimental results show that MRC is able to capture meaningful and compact structures of data. Its performances compare favourably with the current state of the art on challenging benchmarks including rotated MNIST, Caltech-101, Caltech-256, and SHREC'11 non-rigid 3D shapes.
我们引入了一种新的分类器,称为最大残差分类器(MRC),用于与判别决策函数一起学习稀疏表示。MRC试图最大化错误类和正确类的残差之间的差异。这有效地导致了更具判别性的稀疏表示和更好的分类精度。优化过程简单、高效。其目标函数与残差分类策略的决策函数密切相关。与现有的仅限于线性模型的判别稀疏表示学习方法不同,我们的方法能够通过使用Mercer核来处理非线性模型。实验结果表明,MRC能够捕获有意义且紧凑的数据结构。在包括旋转MNIST、Caltech-101、Caltech-256和SHREC'11非刚性3D形状在内的具有挑战性的基准测试中,其性能与当前最先进的性能相媲美。
{"title":"Max residual classifier","authors":"H. Nguyen, Vishal M. Patel","doi":"10.1109/WACV.2014.6836050","DOIUrl":"https://doi.org/10.1109/WACV.2014.6836050","url":null,"abstract":"We introduce a novel classifier, called max residual classifier (MRC), for learning a sparse representation jointly with a discriminative decision function. MRC seeks to maximize the differences between the residual errors of the wrong classes and the right one. This effectively leads to a more discriminative sparse representation and better classification accuracy. The optimization procedure is simple and efficient. Its objective function is closely related to the decision function of the residual classification strategy. Unlike existing methods for learning discriminative sparse representation that are restricted to a linear model, our approach is able to work with a non-linear model via the use of Mercer kernel. Experimental results show that MRC is able to capture meaningful and compact structures of data. Its performances compare favourably with the current state of the art on challenging benchmarks including rotated MNIST, Caltech-101, Caltech-256, and SHREC'11 non-rigid 3D shapes.","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82576513","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Image segmentation of mesenchymal stem cells in diverse culturing conditions 不同培养条件下间充质干细胞的图像分割
M. J. Afridi, Chun Liu, C. Chan, S. Baek, Xiaoming Liu
Researchers in the areas of regenerative medicine and tissue engineering have great interests in understanding the relationship of different sets of culturing conditions and applied mechanical stimuli to the behavior of mesenchymal stem cells (MSCs). However, it is challenging to design a tool to perform automatic cell image analysis due to the diverse morphologies of MSCs. Therefore, as a primary step towards developing the tool, we propose a novel approach for accurate cell image segmentation. We collected three MSC datasets cultured on different surfaces and exposed to diverse mechanical stimuli. By analyzing existing approaches on our data, we choose to substantially extend binarization-based extraction of alignment score (BEAS) approach by extracting novel discriminating features and developing an adaptive threshold estimation model. Experimental results on our data shows our approach is superior to seven conventional techniques. We also define three quantitative measures to analyze the characteristics of images in our datasets. To the best of our knowledge, this is the first study that applied automatic segmentation to live MSC cultured on different surfaces with applied stimuli.
再生医学和组织工程领域的研究人员对了解不同培养条件和应用机械刺激对间充质干细胞(MSCs)行为的关系非常感兴趣。然而,由于MSCs的不同形态,设计一种工具来执行自动细胞图像分析是具有挑战性的。因此,作为开发该工具的第一步,我们提出了一种精确细胞图像分割的新方法。我们收集了三个在不同表面培养并暴露于不同机械刺激下的MSC数据集。通过分析现有的数据处理方法,我们选择通过提取新的判别特征和开发自适应阈值估计模型,大大扩展基于二值化的对齐分数提取(BEAS)方法。实验结果表明,该方法优于7种传统方法。我们还定义了三个定量度量来分析我们数据集中图像的特征。据我们所知,这是第一个将自动分割应用于在不同表面培养的活体MSC并施加刺激的研究。
{"title":"Image segmentation of mesenchymal stem cells in diverse culturing conditions","authors":"M. J. Afridi, Chun Liu, C. Chan, S. Baek, Xiaoming Liu","doi":"10.1109/WACV.2014.6836058","DOIUrl":"https://doi.org/10.1109/WACV.2014.6836058","url":null,"abstract":"Researchers in the areas of regenerative medicine and tissue engineering have great interests in understanding the relationship of different sets of culturing conditions and applied mechanical stimuli to the behavior of mesenchymal stem cells (MSCs). However, it is challenging to design a tool to perform automatic cell image analysis due to the diverse morphologies of MSCs. Therefore, as a primary step towards developing the tool, we propose a novel approach for accurate cell image segmentation. We collected three MSC datasets cultured on different surfaces and exposed to diverse mechanical stimuli. By analyzing existing approaches on our data, we choose to substantially extend binarization-based extraction of alignment score (BEAS) approach by extracting novel discriminating features and developing an adaptive threshold estimation model. Experimental results on our data shows our approach is superior to seven conventional techniques. We also define three quantitative measures to analyze the characteristics of images in our datasets. To the best of our knowledge, this is the first study that applied automatic segmentation to live MSC cultured on different surfaces with applied stimuli.","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89194629","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Video segmentation with joint object and trajectory labeling 结合目标和轨迹标记的视频分割
M. Yang, B. Rosenhahn
Unsupervised video object segmentation is a challenging problem because it involves a large amount of data and object appearance may significantly change over time. In this paper, we propose a bottom-up approach for the combination of object segmentation and motion segmentation using a novel graphical model, which is formulated as inference in a conditional random field (CRF) model. This model combines object labeling and trajectory clustering in a unified probabilistic framework. The CRF contains binary variables representing the class labels of image pixels as well as binary variables indicating the correctness of trajectory clustering, which integrates dense local interaction and sparse global constraint. An optimization scheme based on a coordinate ascent style procedure is proposed to solve the inference problem. We evaluate our proposed framework by comparing it to other video and motion segmentation algorithms. Our method achieves improved performance on state-of-the-art benchmark datasets.
无监督视频对象分割是一个具有挑战性的问题,因为它涉及大量的数据和对象的外观可能会随着时间的推移而发生显著变化。本文提出了一种自下而上的目标分割和运动分割相结合的方法,该方法使用一种新的图形模型,该模型被表述为条件随机场(CRF)模型中的推理。该模型将目标标记和轨迹聚类结合在一个统一的概率框架中。CRF包含表示图像像素类标号的二值变量和表示轨迹聚类正确性的二值变量,融合了密集的局部交互和稀疏的全局约束。提出了一种基于坐标上升式过程的优化方案来解决推理问题。我们通过比较其他视频和运动分割算法来评估我们提出的框架。我们的方法在最先进的基准数据集上实现了改进的性能。
{"title":"Video segmentation with joint object and trajectory labeling","authors":"M. Yang, B. Rosenhahn","doi":"10.1109/WACV.2014.6836017","DOIUrl":"https://doi.org/10.1109/WACV.2014.6836017","url":null,"abstract":"Unsupervised video object segmentation is a challenging problem because it involves a large amount of data and object appearance may significantly change over time. In this paper, we propose a bottom-up approach for the combination of object segmentation and motion segmentation using a novel graphical model, which is formulated as inference in a conditional random field (CRF) model. This model combines object labeling and trajectory clustering in a unified probabilistic framework. The CRF contains binary variables representing the class labels of image pixels as well as binary variables indicating the correctness of trajectory clustering, which integrates dense local interaction and sparse global constraint. An optimization scheme based on a coordinate ascent style procedure is proposed to solve the inference problem. We evaluate our proposed framework by comparing it to other video and motion segmentation algorithms. Our method achieves improved performance on state-of-the-art benchmark datasets.","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89947162","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Consensus-based matching and tracking of keypoints for object tracking 基于共识的目标跟踪关键点匹配与跟踪
G. Nebehay, R. Pflugfelder
We propose a novel keypoint-based method for long-term model-free object tracking in a combined matching-and-tracking framework. In order to localise the object in every frame, each keypoint casts votes for the object center. As erroneous keypoints are hard to avoid, we employ a novel consensus-based scheme for outlier detection in the voting behaviour. To make this approach computationally feasible, we propose not to employ an accumulator space for votes, but rather to cluster votes directly in the image space. By transforming votes based on the current keypoint constellation, we account for changes of the object in scale and rotation. In contrast to competing approaches, we refrain from updating the appearance information, thus avoiding the danger of making errors. The use of fast keypoint detectors and binary descriptors allows for our implementation to run in real-time. We demonstrate experimentally on a diverse dataset that is as large as 60 sequences that our method outperforms the state-of-the-art when high accuracy is required and visualise these results by employing a variant of success plots.
提出了一种基于关键点的无模型长期目标跟踪方法。为了在每一帧中定位对象,每个关键点对对象中心进行投票。由于错误的关键点难以避免,我们采用了一种新的基于共识的方案来检测投票行为中的异常值。为了使这种方法在计算上可行,我们建议不为投票使用累加器空间,而是直接在图像空间中对投票进行聚类。通过基于当前关键点星座的投票转换,我们考虑了物体在尺度和旋转上的变化。与竞争的方法相比,我们避免了更新外观信息,从而避免了出错的危险。使用快速关键点检测器和二进制描述符允许我们的实现实时运行。我们在多达60个序列的不同数据集上通过实验证明,当需要高精度时,我们的方法优于最先进的方法,并通过采用成功图的变体来可视化这些结果。
{"title":"Consensus-based matching and tracking of keypoints for object tracking","authors":"G. Nebehay, R. Pflugfelder","doi":"10.1109/WACV.2014.6836013","DOIUrl":"https://doi.org/10.1109/WACV.2014.6836013","url":null,"abstract":"We propose a novel keypoint-based method for long-term model-free object tracking in a combined matching-and-tracking framework. In order to localise the object in every frame, each keypoint casts votes for the object center. As erroneous keypoints are hard to avoid, we employ a novel consensus-based scheme for outlier detection in the voting behaviour. To make this approach computationally feasible, we propose not to employ an accumulator space for votes, but rather to cluster votes directly in the image space. By transforming votes based on the current keypoint constellation, we account for changes of the object in scale and rotation. In contrast to competing approaches, we refrain from updating the appearance information, thus avoiding the danger of making errors. The use of fast keypoint detectors and binary descriptors allows for our implementation to run in real-time. We demonstrate experimentally on a diverse dataset that is as large as 60 sequences that our method outperforms the state-of-the-art when high accuracy is required and visualise these results by employing a variant of success plots.","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88963142","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 166
System for semi-automated surveying of street-lighting poles from street-level panoramic images 从街道全景图像半自动测量街道灯杆的系统
L. Hazelhoff, Ivo M. Creusen, P. D. With
Accurate and up-to-date inventories of lighting poles are of interest to energy companies, beneficial for the transition to energy-efficient lighting and may contribute to a more adequate lighting of streets. This potentially improves social security and reduces crime and vandalism during nighttime. This paper describes a system for automated surveying of lighting poles from street-level panoramic images. The system consists of two independent detectors, focusing at the detection of the pole itself and at the detection of a specific lighting fixture type. Both follow the same approach, and start with detection of the feature of interest (pole or fixture) within the individual images, followed by a multi-view analysis to retrieve the real-world coordinates of the poles. Afterwards, the detection output of both algorithms is merged. Large-scale validations, covering about 135 km of road, show that over 91% of the lighting poles is found, while the precision remains above 50%. When applying this system in a semi-automated fashion, high-quality inventories can be created up to 5 times more efficiently compared to manually surveying all poles from the images.
准确和最新的灯杆库存是能源公司感兴趣的,有利于向节能照明过渡,并可能有助于更充分的街道照明。这可能会改善社会治安,减少夜间的犯罪和破坏行为。本文介绍了一种基于街道全景图像的灯杆自动测量系统。该系统由两个独立的检测器组成,专注于检测电线杆本身和检测特定的照明灯具类型。两者都采用相同的方法,从检测单个图像中的感兴趣的特征(极点或夹具)开始,然后进行多视图分析以检索极点的真实坐标。然后,将两种算法的检测输出进行合并。覆盖约135公里道路的大规模验证表明,超过91%的灯杆被发现,而精度保持在50%以上。当以半自动化的方式应用该系统时,与从图像中手动测量所有极点相比,创建高质量库存的效率可提高5倍。
{"title":"System for semi-automated surveying of street-lighting poles from street-level panoramic images","authors":"L. Hazelhoff, Ivo M. Creusen, P. D. With","doi":"10.1109/WACV.2014.6836109","DOIUrl":"https://doi.org/10.1109/WACV.2014.6836109","url":null,"abstract":"Accurate and up-to-date inventories of lighting poles are of interest to energy companies, beneficial for the transition to energy-efficient lighting and may contribute to a more adequate lighting of streets. This potentially improves social security and reduces crime and vandalism during nighttime. This paper describes a system for automated surveying of lighting poles from street-level panoramic images. The system consists of two independent detectors, focusing at the detection of the pole itself and at the detection of a specific lighting fixture type. Both follow the same approach, and start with detection of the feature of interest (pole or fixture) within the individual images, followed by a multi-view analysis to retrieve the real-world coordinates of the poles. Afterwards, the detection output of both algorithms is merged. Large-scale validations, covering about 135 km of road, show that over 91% of the lighting poles is found, while the precision remains above 50%. When applying this system in a semi-automated fashion, high-quality inventories can be created up to 5 times more efficiently compared to manually surveying all poles from the images.","PeriodicalId":73325,"journal":{"name":"IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86167045","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
期刊
IEEE Winter Conference on Applications of Computer Vision. IEEE Winter Conference on Applications of Computer Vision
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1