首页 > 最新文献

2011 International Conference on Computer Vision最新文献

英文 中文
Recursive MDL via graph cuts: Application to segmentation 通过图切割递归MDL:在分割中的应用
Pub Date : 2011-11-06 DOI: 10.1109/ICCV.2011.6126330
Lena Gorelick, Andrew Delong, O. Veksler, Yuri Boykov
We propose a novel patch-based image representation that is useful because it (1) inherently detects regions with repetitive structure at multiple scales and (2) yields a parameterless hierarchical segmentation. We describe an image by breaking it into coherent regions where each region is well-described (easily reconstructed) by repeatedly instantiating a patch using a set of simple transformations. In other words, a good segment is one that has sufficient repetition of some pattern, and a patch is useful if it contains a pattern that is repeated in the image.
我们提出了一种新的基于补丁的图像表示,它很有用,因为它(1)在多个尺度上固有地检测具有重复结构的区域,(2)产生无参数的分层分割。我们通过将图像分解成连贯的区域来描述图像,其中每个区域通过使用一组简单变换重复实例化补丁来很好地描述(容易重建)。换句话说,一个好的片段是一个对某些模式有足够重复的片段,而一个补丁是有用的,如果它包含一个在图像中重复的模式。
{"title":"Recursive MDL via graph cuts: Application to segmentation","authors":"Lena Gorelick, Andrew Delong, O. Veksler, Yuri Boykov","doi":"10.1109/ICCV.2011.6126330","DOIUrl":"https://doi.org/10.1109/ICCV.2011.6126330","url":null,"abstract":"We propose a novel patch-based image representation that is useful because it (1) inherently detects regions with repetitive structure at multiple scales and (2) yields a parameterless hierarchical segmentation. We describe an image by breaking it into coherent regions where each region is well-described (easily reconstructed) by repeatedly instantiating a patch using a set of simple transformations. In other words, a good segment is one that has sufficient repetition of some pattern, and a patch is useful if it contains a pattern that is repeated in the image.","PeriodicalId":6391,"journal":{"name":"2011 International Conference on Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2011-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90448657","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Unsupervised learning of a scene-specific coarse gaze estimator 场景特定的粗凝视估计器的无监督学习
Pub Date : 2011-11-06 DOI: 10.1109/ICCV.2011.6126516
Ben Benfold, I. Reid
We present a method to estimate the coarse gaze directions of people from surveillance data. Unlike previous work we aim to do this without recourse to a large hand-labelled corpus of training data. In contrast we propose a method for learning a classifier without any hand labelled data using only the output from an automatic tracking system. A Conditional Random Field is used to model the interactions between the head motion, walking direction, and appearance to recover the gaze directions and simultaneously train randomised decision tree classifiers. Experiments demonstrate performance exceeding that of conventionally trained classifiers on two large surveillance datasets.
提出了一种从监控数据中估计人的粗注视方向的方法。与以前的工作不同,我们的目标是在不依赖于大量手工标记的训练数据语料库的情况下做到这一点。相反,我们提出了一种学习分类器的方法,没有任何手工标记的数据,只使用自动跟踪系统的输出。使用条件随机场对头部运动、行走方向和外观之间的相互作用进行建模,以恢复凝视方向,同时训练随机决策树分类器。实验表明,在两个大型监控数据集上,分类器的性能优于常规训练的分类器。
{"title":"Unsupervised learning of a scene-specific coarse gaze estimator","authors":"Ben Benfold, I. Reid","doi":"10.1109/ICCV.2011.6126516","DOIUrl":"https://doi.org/10.1109/ICCV.2011.6126516","url":null,"abstract":"We present a method to estimate the coarse gaze directions of people from surveillance data. Unlike previous work we aim to do this without recourse to a large hand-labelled corpus of training data. In contrast we propose a method for learning a classifier without any hand labelled data using only the output from an automatic tracking system. A Conditional Random Field is used to model the interactions between the head motion, walking direction, and appearance to recover the gaze directions and simultaneously train randomised decision tree classifiers. Experiments demonstrate performance exceeding that of conventionally trained classifiers on two large surveillance datasets.","PeriodicalId":6391,"journal":{"name":"2011 International Conference on Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2011-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83935576","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 69
Discriminative figure-centric models for joint action localization and recognition 联合动作定位与识别的判别图形中心模型
Pub Date : 2011-11-06 DOI: 10.1109/ICCV.2011.6126472
Tian Lan, Yang Wang, Greg Mori
In this paper we develop an algorithm for action recognition and localization in videos. The algorithm uses a figure-centric visual word representation. Different from previous approaches it does not require reliable human detection and tracking as input. Instead, the person location is treated as a latent variable that is inferred simultaneously with action recognition. A spatial model for an action is learned in a discriminative fashion under a figure-centric representation. Temporal smoothness over video sequences is also enforced. We present results on the UCF-Sports dataset, verifying the effectiveness of our model in situations where detection and tracking of individuals is challenging.
本文提出了一种视频动作识别与定位算法。该算法使用以图形为中心的视觉单词表示。与以前的方法不同,它不需要可靠的人工检测和跟踪作为输入。相反,人的位置被视为与动作识别同时推断的潜在变量。动作的空间模型是在以图形为中心的表征下以判别方式学习的。视频序列的时间平滑性也被强制执行。我们在UCF-Sports数据集上展示了结果,验证了我们的模型在个体检测和跟踪具有挑战性的情况下的有效性。
{"title":"Discriminative figure-centric models for joint action localization and recognition","authors":"Tian Lan, Yang Wang, Greg Mori","doi":"10.1109/ICCV.2011.6126472","DOIUrl":"https://doi.org/10.1109/ICCV.2011.6126472","url":null,"abstract":"In this paper we develop an algorithm for action recognition and localization in videos. The algorithm uses a figure-centric visual word representation. Different from previous approaches it does not require reliable human detection and tracking as input. Instead, the person location is treated as a latent variable that is inferred simultaneously with action recognition. A spatial model for an action is learned in a discriminative fashion under a figure-centric representation. Temporal smoothness over video sequences is also enforced. We present results on the UCF-Sports dataset, verifying the effectiveness of our model in situations where detection and tracking of individuals is challenging.","PeriodicalId":6391,"journal":{"name":"2011 International Conference on Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2011-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"72901223","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 246
A joint learning framework for attribute models and object descriptions 一个用于属性模型和对象描述的联合学习框架
Pub Date : 2011-11-06 DOI: 10.1109/ICCV.2011.6126373
D. Mahajan, Sundararajan Sellamanickam, Vinod Nair
We present a new approach to learning attribute-based descriptions of objects. Unlike earlier works, we do not assume that the descriptions are hand-labeled. Instead, our approach jointly learns both the attribute classifiers and the descriptions from data. By incorporating class information into the attribute classifier learning, we get an attribute-level representation that generalizes well to both unseen examples of known classes and unseen classes. We consider two different settings, one with unlabeled images available for learning, and another without. The former corresponds to a novel transductive setting where the unlabeled images can come from new classes. Results from Animals with Attributes and a-Yahoo, a-Pascal benchmark datasets show that the learned representations give similar or even better accuracy than the hand-labeled descriptions.
我们提出了一种学习基于属性的对象描述的新方法。与早期的作品不同,我们不假设描述是手工标记的。相反,我们的方法从数据中联合学习属性分类器和描述。通过将类信息整合到属性分类器学习中,我们得到了一个属性级表示,它可以很好地泛化到已知类的未见示例和未见类。我们考虑了两种不同的设置,一种是没有标记的图像,另一种是没有标记的。前者对应于一种新的转换设置,其中未标记的图像可以来自新的类。来自Animals with Attributes和a-Yahoo, a-Pascal基准数据集的结果表明,与手工标记的描述相比,学习表征具有相似甚至更好的准确性。
{"title":"A joint learning framework for attribute models and object descriptions","authors":"D. Mahajan, Sundararajan Sellamanickam, Vinod Nair","doi":"10.1109/ICCV.2011.6126373","DOIUrl":"https://doi.org/10.1109/ICCV.2011.6126373","url":null,"abstract":"We present a new approach to learning attribute-based descriptions of objects. Unlike earlier works, we do not assume that the descriptions are hand-labeled. Instead, our approach jointly learns both the attribute classifiers and the descriptions from data. By incorporating class information into the attribute classifier learning, we get an attribute-level representation that generalizes well to both unseen examples of known classes and unseen classes. We consider two different settings, one with unlabeled images available for learning, and another without. The former corresponds to a novel transductive setting where the unlabeled images can come from new classes. Results from Animals with Attributes and a-Yahoo, a-Pascal benchmark datasets show that the learned representations give similar or even better accuracy than the hand-labeled descriptions.","PeriodicalId":6391,"journal":{"name":"2011 International Conference on Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2011-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77068419","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 92
Active clustering of document fragments using information derived from both images and catalogs 使用来自图像和目录的信息对文档片段进行主动聚类
Pub Date : 2011-11-06 DOI: 10.1109/ICCV.2011.6126428
Lior Wolf, Lior Litwak, N. Dershowitz, Roni Shweka, Y. Choueka
Many significant historical corpora contain leaves that are mixed up and no longer bound in their original state as multi-page documents. The reconstruction of old manuscripts from a mix of disjoint leaves can therefore be of paramount importance to historians and literary scholars. Previously, it was shown that visual similarity provides meaningful pair-wise similarities between handwritten leaves. Here, we go a step further and suggest a semiautomatic clustering tool that helps reconstruct the original documents. The proposed solution is based on a graphical model that makes inferences based on catalog information provided for each leaf as well as on the pairwise similarities of handwriting. Several novel active clustering techniques are explored, and the solution is applied to a significant part of the Cairo Genizah, where the problem of joining leaves remains unsolved even after a century of extensive study by hundreds of scholars.
许多重要的历史语料库包含混合的叶子,不再以多页文档的原始状态绑定。因此,对历史学家和文学学者来说,从脱节的树叶中重建旧手稿是至关重要的。之前的研究表明,视觉相似性提供了手写叶子之间有意义的成对相似性。在这里,我们更进一步,建议使用一种半自动聚类工具来帮助重建原始文档。所提出的解决方案基于图形模型,该模型根据为每个叶子提供的目录信息以及笔迹的两两相似性进行推断。研究人员探索了几种新颖的主动聚类技术,并将解决方案应用于Cairo Genizah的重要部分,在那里,即使经过数百名学者一个世纪的广泛研究,连接叶子的问题仍然没有解决。
{"title":"Active clustering of document fragments using information derived from both images and catalogs","authors":"Lior Wolf, Lior Litwak, N. Dershowitz, Roni Shweka, Y. Choueka","doi":"10.1109/ICCV.2011.6126428","DOIUrl":"https://doi.org/10.1109/ICCV.2011.6126428","url":null,"abstract":"Many significant historical corpora contain leaves that are mixed up and no longer bound in their original state as multi-page documents. The reconstruction of old manuscripts from a mix of disjoint leaves can therefore be of paramount importance to historians and literary scholars. Previously, it was shown that visual similarity provides meaningful pair-wise similarities between handwritten leaves. Here, we go a step further and suggest a semiautomatic clustering tool that helps reconstruct the original documents. The proposed solution is based on a graphical model that makes inferences based on catalog information provided for each leaf as well as on the pairwise similarities of handwriting. Several novel active clustering techniques are explored, and the solution is applied to a significant part of the Cairo Genizah, where the problem of joining leaves remains unsolved even after a century of extensive study by hundreds of scholars.","PeriodicalId":6391,"journal":{"name":"2011 International Conference on Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2011-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77122578","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Learning equivariant structured output SVM regressors 学习等变结构化输出SVM回归量
Pub Date : 2011-11-06 DOI: 10.1109/ICCV.2011.6126339
A. Vedaldi, Matthew B. Blaschko, Andrew Zisserman
Equivariance and invariance are often desired properties of a computer vision system. However, currently available strategies generally rely on virtual sampling, leaving open the question of how many samples are necessary, on the use of invariant feature representations, which can mistakenly discard information relevant to the vision task, or on the use of latent variable models, which result in non-convex training and expensive inference at test time. We propose here a generalization of structured output SVM regressors that can incorporate equivariance and invariance into a convex training procedure, enabling the incorporation of large families of transformations, while maintaining optimality and tractability. Importantly, test time inference does not require the estimation of latent variables, resulting in highly efficient objective functions. This results in a natural formulation for treating equivariance and invariance that is easily implemented as an adaptation of off-the-shelf optimization software, obviating the need for ad hoc sampling strategies. Theoretical results relating to vicinal risk, and experiments on challenging aerial car and pedestrian detection tasks show the effectiveness of the proposed solution.
等变性和不变性通常是计算机视觉系统所需要的特性。然而,目前可用的策略通常依赖于虚拟采样,留下了需要多少样本的问题;使用不变特征表示,这可能会错误地丢弃与视觉任务相关的信息;或者使用潜在变量模型,这导致非凸训练和昂贵的测试时推理。我们在这里提出了一种结构化输出SVM回归量的泛化,它可以将等方差和不变性纳入凸训练过程,从而可以在保持最优性和可追溯性的同时纳入大族变换。重要的是,测试时间推断不需要估计潜在变量,从而产生高效的目标函数。这就产生了一种处理等变性和不变性的自然公式,这种公式很容易作为现成优化软件的改编实现,从而避免了对特别采样策略的需要。与周边风险相关的理论结果以及具有挑战性的空中车辆和行人检测任务的实验结果表明了该方法的有效性。
{"title":"Learning equivariant structured output SVM regressors","authors":"A. Vedaldi, Matthew B. Blaschko, Andrew Zisserman","doi":"10.1109/ICCV.2011.6126339","DOIUrl":"https://doi.org/10.1109/ICCV.2011.6126339","url":null,"abstract":"Equivariance and invariance are often desired properties of a computer vision system. However, currently available strategies generally rely on virtual sampling, leaving open the question of how many samples are necessary, on the use of invariant feature representations, which can mistakenly discard information relevant to the vision task, or on the use of latent variable models, which result in non-convex training and expensive inference at test time. We propose here a generalization of structured output SVM regressors that can incorporate equivariance and invariance into a convex training procedure, enabling the incorporation of large families of transformations, while maintaining optimality and tractability. Importantly, test time inference does not require the estimation of latent variables, resulting in highly efficient objective functions. This results in a natural formulation for treating equivariance and invariance that is easily implemented as an adaptation of off-the-shelf optimization software, obviating the need for ad hoc sampling strategies. Theoretical results relating to vicinal risk, and experiments on challenging aerial car and pedestrian detection tasks show the effectiveness of the proposed solution.","PeriodicalId":6391,"journal":{"name":"2011 International Conference on Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2011-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85012249","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 33
Multi-class semi-supervised SVMs with Positiveness Exclusive Regularization 具有正排他正则化的多类半监督支持向量机
Pub Date : 2011-11-06 DOI: 10.1109/ICCV.2011.6126399
Xiaobai Liu, Xiao-Tong Yuan, Shuicheng Yan, Hai Jin
In this work, we address the problem of multi-class classification problem in semi-supervised setting. A regularized multi-task learning approach is presented to train multiple binary-class Semi-Supervised Support Vector Machines (S3VMs) using the one-vs-rest strategy within a joint framework. A novel type of regularization, namely Positiveness Exclusive Regularization (PER), is introduced to induce the following prior: if an unlabeled sample receives significant positive response from one of the classifiers, it is less likely for this sample to receive positive responses from the other classifiers. That is, we expect an exclusive relationship among different S3VMs for evaluating the same unlabeled sample. We propose to use an ℓ1,2-norm regularizer as an implementation of PER. The objective of our approach is to minimize an empirical risk regularized by a PER term and a manifold regularization term. An efficient Nesterov-type smoothing approximation based method is developed for optimization. Evaluations with comparisons are conducted on several benchmarks for visual classification to demonstrate the advantages of the proposed method.
在这项工作中,我们解决了半监督环境下的多类分类问题。提出了一种正则化多任务学习方法,在联合框架内使用1对1策略训练多个二元类半监督支持向量机(s3vm)。引入了一种新的正则化类型,即positive Exclusive regularization (PER),以诱导以下先验:如果未标记的样本从其中一个分类器接收到显著的正响应,则该样本从其他分类器接收到正响应的可能性较小。也就是说,我们期望在评估相同未标记样本的不同s3vm之间存在排他性关系。我们建议使用一个1,2-范数正则化器作为PER的实现。我们方法的目标是最小化由PER项和流形正则化项正则化的经验风险。提出了一种高效的基于nesterov型平滑近似的优化方法。对几种视觉分类基准进行了评价和比较,以证明所提出方法的优势。
{"title":"Multi-class semi-supervised SVMs with Positiveness Exclusive Regularization","authors":"Xiaobai Liu, Xiao-Tong Yuan, Shuicheng Yan, Hai Jin","doi":"10.1109/ICCV.2011.6126399","DOIUrl":"https://doi.org/10.1109/ICCV.2011.6126399","url":null,"abstract":"In this work, we address the problem of multi-class classification problem in semi-supervised setting. A regularized multi-task learning approach is presented to train multiple binary-class Semi-Supervised Support Vector Machines (S3VMs) using the one-vs-rest strategy within a joint framework. A novel type of regularization, namely Positiveness Exclusive Regularization (PER), is introduced to induce the following prior: if an unlabeled sample receives significant positive response from one of the classifiers, it is less likely for this sample to receive positive responses from the other classifiers. That is, we expect an exclusive relationship among different S3VMs for evaluating the same unlabeled sample. We propose to use an ℓ1,2-norm regularizer as an implementation of PER. The objective of our approach is to minimize an empirical risk regularized by a PER term and a manifold regularization term. An efficient Nesterov-type smoothing approximation based method is developed for optimization. Evaluations with comparisons are conducted on several benchmarks for visual classification to demonstrate the advantages of the proposed method.","PeriodicalId":6391,"journal":{"name":"2011 International Conference on Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2011-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87352335","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Linear stereo matching 线性立体匹配
Pub Date : 2011-11-06 DOI: 10.1109/ICCV.2011.6126434
Leonardo De-Maeztu, S. Mattoccia, A. Villanueva, R. Cabeza
Recent local stereo matching algorithms based on an adaptive-weight strategy achieve accuracy similar to global approaches. One of the major problems of these algorithms is that they are computationally expensive and this complexity increases proportionally to the window size. This paper proposes a novel cost aggregation step with complexity independent of the window size (i.e. O(1)) that outperforms state-of-the-art O(1) methods. Moreover, compared to other O(1) approaches, our method does not rely on integral histograms enabling aggregation using colour images instead of grayscale ones. Finally, to improve the results of the proposed algorithm a disparity refinement pipeline is also proposed. The overall algorithm produces results comparable to those of state-of-the-art stereo matching algorithms.
近年来基于自适应权重策略的局部立体匹配算法达到了与全局匹配方法相似的精度。这些算法的一个主要问题是它们的计算成本很高,而且这种复杂性随着窗口大小的增加而成比例地增加。本文提出了一种新的成本聚合步骤,其复杂性与窗口大小无关(即O(1)),优于最先进的O(1)方法。此外,与其他O(1)方法相比,我们的方法不依赖于积分直方图,可以使用彩色图像而不是灰度图像进行聚合。最后,为了改进算法的结果,还提出了视差细化管道。整体算法产生的结果可与最先进的立体匹配算法相媲美。
{"title":"Linear stereo matching","authors":"Leonardo De-Maeztu, S. Mattoccia, A. Villanueva, R. Cabeza","doi":"10.1109/ICCV.2011.6126434","DOIUrl":"https://doi.org/10.1109/ICCV.2011.6126434","url":null,"abstract":"Recent local stereo matching algorithms based on an adaptive-weight strategy achieve accuracy similar to global approaches. One of the major problems of these algorithms is that they are computationally expensive and this complexity increases proportionally to the window size. This paper proposes a novel cost aggregation step with complexity independent of the window size (i.e. O(1)) that outperforms state-of-the-art O(1) methods. Moreover, compared to other O(1) approaches, our method does not rely on integral histograms enabling aggregation using colour images instead of grayscale ones. Finally, to improve the results of the proposed algorithm a disparity refinement pipeline is also proposed. The overall algorithm produces results comparable to those of state-of-the-art stereo matching algorithms.","PeriodicalId":6391,"journal":{"name":"2011 International Conference on Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2011-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84108777","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 102
Robust topological features for deformation invariant image matching 形变不变图像匹配的鲁棒拓扑特征
Pub Date : 2011-11-06 DOI: 10.1109/ICCV.2011.6126538
E. Lobaton, Ramanarayan Vasudevan, R. Alterovitz, R. Bajcsy
Local photometric descriptors are a crucial low level component of numerous computer vision algorithms. In practice, these descriptors are constructed to be invariant to a class of transformations. However, the development of a descriptor that is simultaneously robust to noise and invariant under general deformation has proven difficult. In this paper, we introduce the Topological-Attributed Relational Graph (T-ARG), a new local photometric descriptor constructed from homology that is provably invariant to locally bounded deformation. This new robust topological descriptor is backed by a formal mathematical framework. We apply T-ARG to a set of benchmark images to evaluate its performance. Results indicate that T-ARG significantly outperforms traditional descriptors for noisy, deforming images.
局部光度描述符是众多计算机视觉算法中至关重要的底层组成部分。在实践中,这些描述符被构造为对于一类转换是不变的。然而,开发一种同时对噪声具有鲁棒性和在一般变形下不变性的描述子已被证明是困难的。本文引入了拓扑属性关系图(T-ARG),这是一种由同调构造的新的局部光度描述子,它对局部有界变形是可证明不变的。这种新的鲁棒拓扑描述符由形式化的数学框架支持。我们将T-ARG应用于一组基准图像来评估其性能。结果表明,T-ARG显著优于传统的描述符对噪声,变形图像。
{"title":"Robust topological features for deformation invariant image matching","authors":"E. Lobaton, Ramanarayan Vasudevan, R. Alterovitz, R. Bajcsy","doi":"10.1109/ICCV.2011.6126538","DOIUrl":"https://doi.org/10.1109/ICCV.2011.6126538","url":null,"abstract":"Local photometric descriptors are a crucial low level component of numerous computer vision algorithms. In practice, these descriptors are constructed to be invariant to a class of transformations. However, the development of a descriptor that is simultaneously robust to noise and invariant under general deformation has proven difficult. In this paper, we introduce the Topological-Attributed Relational Graph (T-ARG), a new local photometric descriptor constructed from homology that is provably invariant to locally bounded deformation. This new robust topological descriptor is backed by a formal mathematical framework. We apply T-ARG to a set of benchmark images to evaluate its performance. Results indicate that T-ARG significantly outperforms traditional descriptors for noisy, deforming images.","PeriodicalId":6391,"journal":{"name":"2011 International Conference on Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2011-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82827172","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Kinecting the dots: Particle based scene flow from depth sensors 连接点:基于深度传感器的粒子场景流
Pub Date : 2011-11-06 DOI: 10.1109/ICCV.2011.6126509
Simon Hadfield, R. Bowden
The motion field of a scene can be used for object segmentation and to provide features for classification tasks like action recognition. Scene flow is the full 3D motion field of the scene, and is more difficult to estimate than it's 2D counterpart, optical flow. Current approaches use a smoothness cost for regularisation, which tends to over-smooth at object boundaries. This paper presents a novel formulation for scene flow estimation, a collection of moving points in 3D space, modelled using a particle filter that supports multiple hypotheses and does not oversmooth the motion field. In addition, this paper is the first to address scene flow estimation, while making use of modern depth sensors and monocular appearance images, rather than traditional multi-viewpoint rigs. The algorithm is applied to an existing scene flow dataset, where it achieves comparable results to approaches utilising multiple views, while taking a fraction of the time.
场景的运动场可用于对象分割,并为动作识别等分类任务提供特征。场景流是场景的全3D运动场,比2D光流更难估计。当前的方法使用平滑代价进行正则化,这往往在对象边界处过于平滑。本文提出了一种场景流估计的新公式,即3D空间中移动点的集合,使用支持多个假设且不会过度平滑运动场的粒子滤波器建模。此外,本文首次解决了场景流估计问题,同时利用现代深度传感器和单目外观图像,而不是传统的多视点平台。该算法应用于现有的场景流数据集,在那里它实现了与使用多个视图的方法相当的结果,同时花费了一小部分时间。
{"title":"Kinecting the dots: Particle based scene flow from depth sensors","authors":"Simon Hadfield, R. Bowden","doi":"10.1109/ICCV.2011.6126509","DOIUrl":"https://doi.org/10.1109/ICCV.2011.6126509","url":null,"abstract":"The motion field of a scene can be used for object segmentation and to provide features for classification tasks like action recognition. Scene flow is the full 3D motion field of the scene, and is more difficult to estimate than it's 2D counterpart, optical flow. Current approaches use a smoothness cost for regularisation, which tends to over-smooth at object boundaries. This paper presents a novel formulation for scene flow estimation, a collection of moving points in 3D space, modelled using a particle filter that supports multiple hypotheses and does not oversmooth the motion field. In addition, this paper is the first to address scene flow estimation, while making use of modern depth sensors and monocular appearance images, rather than traditional multi-viewpoint rigs. The algorithm is applied to an existing scene flow dataset, where it achieves comparable results to approaches utilising multiple views, while taking a fraction of the time.","PeriodicalId":6391,"journal":{"name":"2011 International Conference on Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2011-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83243789","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 97
期刊
2011 International Conference on Computer Vision
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1