首页 > 最新文献

2009 IEEE Conference on Computer Vision and Pattern Recognition最新文献

英文 中文
SIFT-Rank: Ordinal description for invariant feature correspondence SIFT-Rank:不变特征对应的序数描述
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206849
M. Toews, W. Wells
This paper investigates ordinal image description for invariant feature correspondence. Ordinal description is a meta-technique which considers image measurements in terms of their ranks in a sorted array, instead of the measurement values themselves. Rank-ordering normalizes descriptors in a manner invariant under monotonic deformations of the underlying image measurements, and therefore serves as a simple, non-parametric substitute for ad hoc scaling and thresholding techniques currently used. Ordinal description is particularly well-suited for invariant features, as the high dimensionality of state-of-the-art descriptors permits a large number of unique rank-orderings, and the computationally complex step of sorting is only required once after geometrical normalization. Correspondence trials based on a benchmark data set show that in general, rank-ordered SIFT (SIFT-rank) descriptors outperform other state-of-the-art descriptors in terms of precision-recall, including standard SIFT and GLOH.
研究了基于不变特征对应的有序图像描述。序数描述是一种元技术,它根据图像在排序数组中的排名来考虑图像测量值,而不是测量值本身。秩排序在底层图像测量的单调变形下以一种不变的方式规范化描述符,因此可以作为当前使用的特别缩放和阈值技术的简单,非参数替代。有序描述特别适合于不变量特征,因为最先进的描述符的高维性允许大量唯一的秩排序,并且在几何归一化之后只需要一次计算复杂的排序步骤。基于基准数据集的对应试验表明,一般来说,秩序SIFT (SIFT-rank)描述符在精确召回率方面优于其他最先进的描述符,包括标准SIFT和GLOH。
{"title":"SIFT-Rank: Ordinal description for invariant feature correspondence","authors":"M. Toews, W. Wells","doi":"10.1109/CVPR.2009.5206849","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206849","url":null,"abstract":"This paper investigates ordinal image description for invariant feature correspondence. Ordinal description is a meta-technique which considers image measurements in terms of their ranks in a sorted array, instead of the measurement values themselves. Rank-ordering normalizes descriptors in a manner invariant under monotonic deformations of the underlying image measurements, and therefore serves as a simple, non-parametric substitute for ad hoc scaling and thresholding techniques currently used. Ordinal description is particularly well-suited for invariant features, as the high dimensionality of state-of-the-art descriptors permits a large number of unique rank-orderings, and the computationally complex step of sorting is only required once after geometrical normalization. Correspondence trials based on a benchmark data set show that in general, rank-ordered SIFT (SIFT-rank) descriptors outperform other state-of-the-art descriptors in terms of precision-recall, including standard SIFT and GLOH.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130719068","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 61
Nonparametric discriminant HMM and application to facial expression recognition 非参数判别HMM及其在面部表情识别中的应用
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206509
Lifeng Shang, Kwok-Ping Chan
This paper presents a nonparametric discriminant HMM and applies it to facial expression recognition. In the proposed HMM, we introduce an effective nonparametric output probability estimation method to increase the discrimination ability at both hidden state level and class level. The proposed method uses a nonparametric adaptive kernel to utilize information from all classes and improve the discrimination at class level. The discrimination between hidden states is increased by defining membership coefficients which associate each reference vector with hidden states. The adaption of such coefficients is obtained by the expectation maximization (EM) method. Furthermore, we present a general formula for the estimation of output probability, which provides a way to develop new HMMs. Finally, we evaluate the performance of the proposed method on the CMU expression database and compare it with other nonparametric HMMs.
提出了一种非参数判别HMM,并将其应用于面部表情识别。在隐马尔可夫模型中,我们引入了一种有效的非参数输出概率估计方法,提高了隐马尔可夫模型在隐藏状态和类别层面的识别能力。该方法利用非参数自适应核来利用所有类的信息,提高了类水平的识别能力。通过定义将每个参考向量与隐藏状态关联起来的隶属系数来增加隐藏状态之间的区分。利用期望最大化方法对这些系数进行自适应。此外,我们还给出了输出概率估计的一般公式,为开发新的hmm提供了一条途径。最后,我们在CMU表达式数据库上评估了该方法的性能,并将其与其他非参数hmm进行了比较。
{"title":"Nonparametric discriminant HMM and application to facial expression recognition","authors":"Lifeng Shang, Kwok-Ping Chan","doi":"10.1109/CVPR.2009.5206509","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206509","url":null,"abstract":"This paper presents a nonparametric discriminant HMM and applies it to facial expression recognition. In the proposed HMM, we introduce an effective nonparametric output probability estimation method to increase the discrimination ability at both hidden state level and class level. The proposed method uses a nonparametric adaptive kernel to utilize information from all classes and improve the discrimination at class level. The discrimination between hidden states is increased by defining membership coefficients which associate each reference vector with hidden states. The adaption of such coefficients is obtained by the expectation maximization (EM) method. Furthermore, we present a general formula for the estimation of output probability, which provides a way to develop new HMMs. Finally, we evaluate the performance of the proposed method on the CMU expression database and compare it with other nonparametric HMMs.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130968026","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 67
Geometric and probabilistic image dissimilarity measures for common field of view detection 通用视场检测的几何和概率图像不相似度量
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206810
Marcel Brückner, Ferid Bajramovic, Joachim Denzler
Detecting image pairs with a common field of view is an important prerequisite for many computer vision tasks. Typically, common local features are used as a criterion for identifying such image pairs. This approach, however, requires a reliable method for matching features, which is generally a very difficult problem, especially in situations with a wide baseline or ambiguities in the scene. We propose two new approaches for the common field of view problem. The first one is still based on feature matching. Instead of requiring a very low false positive rate for the feature matching, however, geometric constraints are used to assess matches which may contain many false positives. The second approach completely avoids hard matching of features by evaluating the entropy of correspondence probabilities. We perform quantitative experiments on three different hand labeled scenes with varying difficulty. In moderately difficult situations with a medium baseline and few ambiguities in the scene, our proposed methods give similarly good results to the classical matching based method. On the most challenging scene having a wide baseline and many ambiguities, the performance of the classical method deteriorates, while ours are much less affected and still produce good results. Hence, our methods show the best overall performance in a combined evaluation.
检测具有共同视场的图像对是许多计算机视觉任务的重要前提。通常,使用共同的局部特征作为识别此类图像对的标准。然而,这种方法需要一种可靠的方法来匹配特征,这通常是一个非常困难的问题,特别是在场景中有宽基线或模糊的情况下。针对共同视场问题,我们提出了两种新的解决方法。第一种方法仍然是基于特征匹配。而不是要求一个非常低的假阳性率的特征匹配,然而,使用几何约束来评估匹配可能包含许多假阳性。第二种方法通过评估对应概率的熵完全避免了特征的硬匹配。我们以不同的难度对三种不同的手标记场景进行了定量实验。在中等难度的场景中,我们提出的方法与经典的基于匹配的方法具有相似的效果。在最具挑战性的具有宽基线和许多模糊性的场景中,经典方法的性能会下降,而我们的方法受到的影响要小得多,并且仍然可以产生良好的结果。因此,我们的方法在综合评估中显示出最佳的整体性能。
{"title":"Geometric and probabilistic image dissimilarity measures for common field of view detection","authors":"Marcel Brückner, Ferid Bajramovic, Joachim Denzler","doi":"10.1109/CVPR.2009.5206810","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206810","url":null,"abstract":"Detecting image pairs with a common field of view is an important prerequisite for many computer vision tasks. Typically, common local features are used as a criterion for identifying such image pairs. This approach, however, requires a reliable method for matching features, which is generally a very difficult problem, especially in situations with a wide baseline or ambiguities in the scene. We propose two new approaches for the common field of view problem. The first one is still based on feature matching. Instead of requiring a very low false positive rate for the feature matching, however, geometric constraints are used to assess matches which may contain many false positives. The second approach completely avoids hard matching of features by evaluating the entropy of correspondence probabilities. We perform quantitative experiments on three different hand labeled scenes with varying difficulty. In moderately difficult situations with a medium baseline and few ambiguities in the scene, our proposed methods give similarly good results to the classical matching based method. On the most challenging scene having a wide baseline and many ambiguities, the performance of the classical method deteriorates, while ours are much less affected and still produce good results. Hence, our methods show the best overall performance in a combined evaluation.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133324071","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Learning signs from subtitles: A weakly supervised approach to sign language recognition 从字幕中学习手语:一个弱监督的手语识别方法
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206647
H. Cooper, R. Bowden
This paper introduces a fully automated, unsupervised method to recognise sign from subtitles. It does this by using data mining to align correspondences in sections of videos. Based on head and hand tracking, a novel temporally constrained adaptation of a priori mining is used to extract similar regions of video, with the aid of a proposed contextual negative selection method. These regions are refined in the temporal domain to isolate the occurrences of similar signs in each example. The system is shown to automatically identify and segment signs from standard news broadcasts containing a variety of topics.
本文介绍了一种全自动、无监督的字幕符号识别方法。它通过使用数据挖掘来对齐视频部分中的对应关系来实现这一点。在头部和手部跟踪的基础上,使用一种新的先验挖掘的时间约束适应方法,在提出的上下文否定选择方法的帮助下提取视频的相似区域。这些区域在时间域中进行细化,以隔离每个示例中出现的类似符号。该系统被证明可以自动识别和分割包含各种主题的标准新闻广播的标志。
{"title":"Learning signs from subtitles: A weakly supervised approach to sign language recognition","authors":"H. Cooper, R. Bowden","doi":"10.1109/CVPR.2009.5206647","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206647","url":null,"abstract":"This paper introduces a fully automated, unsupervised method to recognise sign from subtitles. It does this by using data mining to align correspondences in sections of videos. Based on head and hand tracking, a novel temporally constrained adaptation of a priori mining is used to extract similar regions of video, with the aid of a proposed contextual negative selection method. These regions are refined in the temporal domain to isolate the occurrences of similar signs in each example. The system is shown to automatically identify and segment signs from standard news broadcasts containing a variety of topics.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132144116","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 76
Relighting objects from image collections 重新照亮图像集合中的对象
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206753
Tom Haber, Christian Fuchs, P. Bekaert, H. Seidel, M. Goesele, H. Lensch
We present an approach for recovering the reflectance of a static scene with known geometry from a collection of images taken under distant, unknown illumination. In contrast to previous work, we allow the illumination to vary between the images, which greatly increases the applicability of the approach. Using an all-frequency relighting framework based on wavelets, we are able to simultaneously estimate the per-image incident illumination and the per-surface point reflectance. The wavelet framework allows for incorporating various reflection models. We demonstrate the quality of our results for synthetic test cases as well as for several datasets captured under laboratory conditions. Combined with multi-view stereo reconstruction, we are even able to recover the geometry and reflectance of a scene solely using images collected from the Internet.
我们提出了一种方法,用于恢复一个静态场景的反射率与已知的几何形状从图像的集合下拍摄的遥远,未知的照明。与以前的工作相比,我们允许图像之间的照明变化,这大大增加了该方法的适用性。利用基于小波的全频率重照框架,我们能够同时估计每幅图像的入射照度和每表面的点反射率。小波框架允许合并各种反射模型。我们演示了合成测试用例以及在实验室条件下捕获的几个数据集的结果质量。结合多视点立体重建,我们甚至可以仅使用从互联网上收集的图像来恢复场景的几何形状和反射率。
{"title":"Relighting objects from image collections","authors":"Tom Haber, Christian Fuchs, P. Bekaert, H. Seidel, M. Goesele, H. Lensch","doi":"10.1109/CVPR.2009.5206753","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206753","url":null,"abstract":"We present an approach for recovering the reflectance of a static scene with known geometry from a collection of images taken under distant, unknown illumination. In contrast to previous work, we allow the illumination to vary between the images, which greatly increases the applicability of the approach. Using an all-frequency relighting framework based on wavelets, we are able to simultaneously estimate the per-image incident illumination and the per-surface point reflectance. The wavelet framework allows for incorporating various reflection models. We demonstrate the quality of our results for synthetic test cases as well as for several datasets captured under laboratory conditions. Combined with multi-view stereo reconstruction, we are even able to recover the geometry and reflectance of a scene solely using images collected from the Internet.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132450800","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 108
Robust multi-class transductive learning with graphs 基于图的鲁棒多类转换学习
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206871
W. Liu, Shih-Fu Chang
Graph-based methods form a main category of semi-supervised learning, offering flexibility and easy implementation in many applications. However, the performance of these methods is often sensitive to the construction of a neighborhood graph, which is non-trivial for many real-world problems. In this paper, we propose a novel framework that builds on learning the graph given labeled and unlabeled data. The paper has two major contributions. Firstly, we use a nonparametric algorithm to learn the entire adjacency matrix of a symmetry-favored k-NN graph, assuming that the matrix is doubly stochastic. The nonparametric algorithm makes the constructed graph highly robust to noisy samples and capable of approximating underlying submanifolds or clusters. Secondly, to address multi-class semi-supervised classification, we formulate a constrained label propagation problem on the learned graph by incorporating class priors, leading to a simple closed-form solution. Experimental results on both synthetic and real-world datasets show that our approach is significantly better than the state-of-the-art graph-based semi-supervised learning algorithms in terms of accuracy and robustness.
基于图的方法构成了半监督学习的主要类别,在许多应用中提供了灵活性和易于实现。然而,这些方法的性能往往对邻域图的构造很敏感,这对于许多现实问题来说是非平凡的。在本文中,我们提出了一个新的框架,该框架建立在学习给定标记和未标记数据的图的基础上。这篇论文有两个主要贡献。首先,我们使用一种非参数算法来学习对称偏好k-NN图的整个邻接矩阵,假设矩阵是双随机的。非参数算法使构造的图对有噪声的样本具有很强的鲁棒性,并且能够逼近底层的子流形或聚类。其次,为了解决多类半监督分类问题,我们通过结合类先验,在学习图上形成一个约束标签传播问题,从而得到一个简单的闭形式解。在合成数据集和真实数据集上的实验结果表明,我们的方法在准确性和鲁棒性方面明显优于最先进的基于图的半监督学习算法。
{"title":"Robust multi-class transductive learning with graphs","authors":"W. Liu, Shih-Fu Chang","doi":"10.1109/CVPR.2009.5206871","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206871","url":null,"abstract":"Graph-based methods form a main category of semi-supervised learning, offering flexibility and easy implementation in many applications. However, the performance of these methods is often sensitive to the construction of a neighborhood graph, which is non-trivial for many real-world problems. In this paper, we propose a novel framework that builds on learning the graph given labeled and unlabeled data. The paper has two major contributions. Firstly, we use a nonparametric algorithm to learn the entire adjacency matrix of a symmetry-favored k-NN graph, assuming that the matrix is doubly stochastic. The nonparametric algorithm makes the constructed graph highly robust to noisy samples and capable of approximating underlying submanifolds or clusters. Secondly, to address multi-class semi-supervised classification, we formulate a constrained label propagation problem on the learned graph by incorporating class priors, leading to a simple closed-form solution. Experimental results on both synthetic and real-world datasets show that our approach is significantly better than the state-of-the-art graph-based semi-supervised learning algorithms in terms of accuracy and robustness.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132751541","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 166
Pictorial structures revisited: People detection and articulated pose estimation 重新审视图像结构:人物检测和关节姿态估计
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206754
Mykhaylo Andriluka, S. Roth, B. Schiele
Non-rigid object detection and articulated pose estimation are two related and challenging problems in computer vision. Numerous models have been proposed over the years and often address different special cases, such as pedestrian detection or upper body pose estimation in TV footage. This paper shows that such specialization may not be necessary, and proposes a generic approach based on the pictorial structures framework. We show that the right selection of components for both appearance and spatial modeling is crucial for general applicability and overall performance of the model. The appearance of body parts is modeled using densely sampled shape context descriptors and discriminatively trained AdaBoost classifiers. Furthermore, we interpret the normalized margin of each classifier as likelihood in a generative model. Non-Gaussian relationships between parts are represented as Gaussians in the coordinate system of the joint between parts. The marginal posterior of each part is inferred using belief propagation. We demonstrate that such a model is equally suitable for both detection and pose estimation tasks, outperforming the state of the art on three recently proposed datasets.
非刚性目标检测和关节姿态估计是计算机视觉中两个相关且具有挑战性的问题。多年来,已经提出了许多模型,并且通常针对不同的特殊情况,例如行人检测或电视镜头中的上半身姿势估计。本文表明,这种专业化可能是不必要的,并提出了一种基于图形结构框架的通用方法。我们表明,正确选择用于外观和空间建模的组件对于模型的一般适用性和整体性能至关重要。使用密集采样的形状上下文描述符和判别训练的AdaBoost分类器对身体部位的外观进行建模。此外,我们将每个分类器的归一化边缘解释为生成模型中的可能性。部件间的非高斯关系在部件间关节的坐标系中表示为高斯关系。每个部分的边际后验通过信念传播来推断。我们证明了这样的模型同样适用于检测和姿态估计任务,在最近提出的三个数据集上表现优于最先进的状态。
{"title":"Pictorial structures revisited: People detection and articulated pose estimation","authors":"Mykhaylo Andriluka, S. Roth, B. Schiele","doi":"10.1109/CVPR.2009.5206754","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206754","url":null,"abstract":"Non-rigid object detection and articulated pose estimation are two related and challenging problems in computer vision. Numerous models have been proposed over the years and often address different special cases, such as pedestrian detection or upper body pose estimation in TV footage. This paper shows that such specialization may not be necessary, and proposes a generic approach based on the pictorial structures framework. We show that the right selection of components for both appearance and spatial modeling is crucial for general applicability and overall performance of the model. The appearance of body parts is modeled using densely sampled shape context descriptors and discriminatively trained AdaBoost classifiers. Furthermore, we interpret the normalized margin of each classifier as likelihood in a generative model. Non-Gaussian relationships between parts are represented as Gaussians in the coordinate system of the joint between parts. The marginal posterior of each part is inferred using belief propagation. We demonstrate that such a model is equally suitable for both detection and pose estimation tasks, outperforming the state of the art on three recently proposed datasets.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131462900","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 877
Interval HSV: Extracting ink annotations 间隔HSV:提取墨水注释
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206554
John C. Femiani, A. Razdan
The HSV color space is an intuitive way to reason about color, but the nonlinear relationship to RGB coordinates complicates histogram analysis of colors in HSV. We present novel Interval-HSV formulas to identify a range in HSV for each RGB interval. We show the usefulness by introducing a parameter-free and completely automatic technique to extract both colored and black ink annotations from faded backgrounds such as digitized aerial photographs, maps, or printed-text documents. We discuss the characteristics of ink mixing in the HSV color space and discover a single feature, the upper limit of the saturation-interval, to extract ink even when it is achromatic. We form robust Interval-HV histograms in order to identify the number and colors of inks in the image.
HSV颜色空间是一种直观的推理颜色的方法,但是与RGB坐标的非线性关系使HSV中颜色的直方图分析变得复杂。我们提出了新的区间-HSV公式来确定每个RGB区间的HSV范围。我们通过引入无参数和完全自动化的技术来从褪色的背景(如数字化航空照片,地图或印刷文本文档)中提取彩色和黑色墨水注释来展示其实用性。我们讨论了HSV色彩空间中油墨混合的特性,并发现了一个单一的特征,即饱和度区间的上限,即使在消色差的情况下也可以提取油墨。为了识别图像中墨水的数量和颜色,我们形成了鲁棒的Interval-HV直方图。
{"title":"Interval HSV: Extracting ink annotations","authors":"John C. Femiani, A. Razdan","doi":"10.1109/CVPR.2009.5206554","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206554","url":null,"abstract":"The HSV color space is an intuitive way to reason about color, but the nonlinear relationship to RGB coordinates complicates histogram analysis of colors in HSV. We present novel Interval-HSV formulas to identify a range in HSV for each RGB interval. We show the usefulness by introducing a parameter-free and completely automatic technique to extract both colored and black ink annotations from faded backgrounds such as digitized aerial photographs, maps, or printed-text documents. We discuss the characteristics of ink mixing in the HSV color space and discover a single feature, the upper limit of the saturation-interval, to extract ink even when it is achromatic. We form robust Interval-HV histograms in order to identify the number and colors of inks in the image.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128813840","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Beyond the graphs: Semi-parametric semi-supervised discriminant analysis 图外:半参数半监督判别分析
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206675
Fei Wang, Xin Wang, Ta-Hsin Li
Linear discriminant analysis (LDA) is a popular feature extraction method that has aroused considerable interests in computer vision and pattern recognition fields. The projection vectors of LDA is usually achieved by maximizing the between-class scatter and simultaneously minimizing the within-class scatter of the data set. However, in practice, there is usually a lack of sufficient labeled data, which makes the estimated projection direction inaccurate. To address the above limitations, in this paper, we propose a novel semi-supervised discriminant analysis approach. Unlike traditional graph based methods, our algorithm incorporates the geometric information revealed by both labeled and unlabeled data points in a semi-parametric way. Specifically, the final projections of the data points will contain two parts: a discriminant part learned by traditional LDA (or KDA) on the labeled points and a geometrical part learned by kernel PCA on the whole data set. Therefore we call our algorithm semi-parametric semi-supervised discriminant analysis (SSDA). Experimental results on face recognition and image retrieval tasks are presented to show the effectiveness of our method.
线性判别分析(LDA)是一种流行的特征提取方法,在计算机视觉和模式识别领域引起了广泛的关注。LDA的投影向量通常是通过最大化类间散点同时最小化数据集的类内散点来实现的。然而,在实践中,通常缺乏足够的标记数据,这使得估计的投影方向不准确。为了解决上述局限性,本文提出了一种新颖的半监督判别分析方法。与传统的基于图的方法不同,我们的算法以半参数的方式结合了标记和未标记数据点所揭示的几何信息。具体来说,数据点的最终投影将包含两部分:传统LDA(或KDA)对标记点学习的判别部分和核主成分分析对整个数据集学习的几何部分。因此,我们称该算法为半参数半监督判别分析(SSDA)。在人脸识别和图像检索任务上的实验结果表明了该方法的有效性。
{"title":"Beyond the graphs: Semi-parametric semi-supervised discriminant analysis","authors":"Fei Wang, Xin Wang, Ta-Hsin Li","doi":"10.1109/CVPR.2009.5206675","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206675","url":null,"abstract":"Linear discriminant analysis (LDA) is a popular feature extraction method that has aroused considerable interests in computer vision and pattern recognition fields. The projection vectors of LDA is usually achieved by maximizing the between-class scatter and simultaneously minimizing the within-class scatter of the data set. However, in practice, there is usually a lack of sufficient labeled data, which makes the estimated projection direction inaccurate. To address the above limitations, in this paper, we propose a novel semi-supervised discriminant analysis approach. Unlike traditional graph based methods, our algorithm incorporates the geometric information revealed by both labeled and unlabeled data points in a semi-parametric way. Specifically, the final projections of the data points will contain two parts: a discriminant part learned by traditional LDA (or KDA) on the labeled points and a geometrical part learned by kernel PCA on the whole data set. Therefore we call our algorithm semi-parametric semi-supervised discriminant analysis (SSDA). Experimental results on face recognition and image retrieval tasks are presented to show the effectiveness of our method.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127650339","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Fourier analysis and Gabor filtering for texture analysis and local reconstruction of general shapes 傅里叶分析和Gabor滤波用于纹理分析和一般形状的局部重建
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206591
Fabio Galasso, Joan Lasenby
Since the pioneering work of Gibson in 1950, Shape-From-Texture has been considered by researchers as a hard problem, mainly due to restrictive assumptions which often limit its applicability. We assume a very general stochastic homogeneity and perspective camera model, for both deterministic and stochastic textures. A multi-scale distortion is efficiently estimated with a previously presented method based on Fourier analysis and Gabor filters. The novel 3D reconstruction method that we propose applies to general shapes, and includes non-developable and extensive surfaces. Our algorithm is accurate, robust and compares favorably to the present state of the art of Shape-From-Texture. Results show its application to non-invasively study shape changes with laid-on textures, while rendering and re-texturing of cloth is suggested for future work.
自1950年Gibson的开创性工作以来,形状-从纹理一直被研究人员认为是一个难题,主要是因为限制性的假设往往限制了它的适用性。我们假设一个非常一般的随机均匀性和透视相机模型,对于确定性和随机纹理。先前提出的基于傅里叶分析和Gabor滤波器的方法可以有效地估计多尺度失真。我们提出的新的三维重建方法适用于一般形状,包括不可展开和广泛的表面。我们的算法准确,鲁棒性好,与目前的形状-纹理技术相比具有优势。结果表明,该方法可用于无创地研究布面纹理的形状变化,并为今后的工作提供了一些建议。
{"title":"Fourier analysis and Gabor filtering for texture analysis and local reconstruction of general shapes","authors":"Fabio Galasso, Joan Lasenby","doi":"10.1109/CVPR.2009.5206591","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206591","url":null,"abstract":"Since the pioneering work of Gibson in 1950, Shape-From-Texture has been considered by researchers as a hard problem, mainly due to restrictive assumptions which often limit its applicability. We assume a very general stochastic homogeneity and perspective camera model, for both deterministic and stochastic textures. A multi-scale distortion is efficiently estimated with a previously presented method based on Fourier analysis and Gabor filters. The novel 3D reconstruction method that we propose applies to general shapes, and includes non-developable and extensive surfaces. Our algorithm is accurate, robust and compares favorably to the present state of the art of Shape-From-Texture. Results show its application to non-invasively study shape changes with laid-on textures, while rendering and re-texturing of cloth is suggested for future work.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127816939","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
期刊
2009 IEEE Conference on Computer Vision and Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1