首页 > 最新文献

2013 IEEE International Conference on Computer Vision最新文献

英文 中文
Constant Time Weighted Median Filtering for Stereo Matching and Beyond 常时间加权中值滤波用于立体匹配及以后
Pub Date : 2013-12-01 DOI: 10.1109/ICCV.2013.13
Ziyang Ma, Kaiming He, Yichen Wei, Jian Sun, E. Wu
Despite the continuous advances in local stereo matching for years, most efforts are on developing robust cost computation and aggregation methods. Little attention has been seriously paid to the disparity refinement. In this work, we study weighted median filtering for disparity refinement. We discover that with this refinement, even the simple box filter aggregation achieves comparable accuracy with various sophisticated aggregation methods (with the same refinement). This is due to the nice weighted median filtering properties of removing outlier error while respecting edges/structures. This reveals that the previously overlooked refinement can be at least as crucial as aggregation. We also develop the first constant time algorithm for the previously time-consuming weighted median filter. This makes the simple combination ``box aggregation + weighted median'' an attractive solution in practice for both speed and accuracy. As a byproduct, the fast weighted median filtering unleashes its potential in other applications that were hampered by high complexities. We show its superiority in various applications such as depth up sampling, clip-art JPEG artifact removal, and image stylization.
尽管近年来在局部立体匹配方面不断取得进展,但大多数努力都集中在开发鲁棒的成本计算和聚合方法上。对于视差的细化,很少有人重视。在这项工作中,我们研究了视差细化的加权中值滤波。我们发现,通过这种细化,即使是简单的框式过滤器聚合也可以与各种复杂的聚合方法(具有相同的细化)达到相当的精度。这是由于良好的加权中值滤波特性,在尊重边缘/结构的同时去除离群值误差。这表明之前被忽略的细化至少可以和聚合一样重要。我们还为以前耗时的加权中值滤波器开发了第一个常数时间算法。这使得简单的组合“框聚合+加权中值”在速度和准确性方面都是一个有吸引力的解决方案。作为副产品,快速加权中值滤波在其他被高度复杂性阻碍的应用中释放了它的潜力。我们展示了它在各种应用中的优势,如深度向上采样,剪贴艺术JPEG伪影去除和图像风格化。
{"title":"Constant Time Weighted Median Filtering for Stereo Matching and Beyond","authors":"Ziyang Ma, Kaiming He, Yichen Wei, Jian Sun, E. Wu","doi":"10.1109/ICCV.2013.13","DOIUrl":"https://doi.org/10.1109/ICCV.2013.13","url":null,"abstract":"Despite the continuous advances in local stereo matching for years, most efforts are on developing robust cost computation and aggregation methods. Little attention has been seriously paid to the disparity refinement. In this work, we study weighted median filtering for disparity refinement. We discover that with this refinement, even the simple box filter aggregation achieves comparable accuracy with various sophisticated aggregation methods (with the same refinement). This is due to the nice weighted median filtering properties of removing outlier error while respecting edges/structures. This reveals that the previously overlooked refinement can be at least as crucial as aggregation. We also develop the first constant time algorithm for the previously time-consuming weighted median filter. This makes the simple combination ``box aggregation + weighted median'' an attractive solution in practice for both speed and accuracy. As a byproduct, the fast weighted median filtering unleashes its potential in other applications that were hampered by high complexities. We show its superiority in various applications such as depth up sampling, clip-art JPEG artifact removal, and image stylization.","PeriodicalId":6351,"journal":{"name":"2013 IEEE International Conference on Computer Vision","volume":"83 1","pages":"49-56"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75015966","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 284
Fluttering Pattern Generation Using Modified Legendre Sequence for Coded Exposure Imaging 基于改进Legendre序列的编码曝光成像中颤振模式生成
Pub Date : 2013-12-01 DOI: 10.1109/ICCV.2013.128
Hae-Gon Jeon, Joon-Young Lee, Yudeog Han, Seon Joo Kim, In-So Kweon
Finding a good binary sequence is critical in determining the performance of the coded exposure imaging, but previous methods mostly rely on a random search for finding the binary codes, which could easily fail to find good long sequences due to the exponentially growing search space. In this paper, we present a new computationally efficient algorithm for generating the binary sequence, which is especially well suited for longer sequences. We show that the concept of the low autocorrelation binary sequence that has been well exploited in the information theory community can be applied for generating the fluttering patterns of the shutter, propose a new measure of a good binary sequence, and present a new algorithm by modifying the Legendre sequence for the coded exposure imaging. Experiments using both synthetic and real data show that our new algorithm consistently generates better binary sequences for the coded exposure problem, yielding better deblurring and resolution enhancement results compared to the previous methods for generating the binary codes.
寻找一个好的二进制序列是决定编码曝光成像性能的关键,但以往的方法大多依靠随机搜索来寻找二进制代码,由于搜索空间呈指数级增长,很容易找不到好的长序列。在本文中,我们提出了一种新的计算效率高的二进制序列生成算法,它特别适合于较长的序列。本文研究了信息论领域中低自相关二值序列的概念,提出了一种新的二值序列度量方法,并通过对Legendre序列的修改,提出了一种新的编码曝光成像算法。使用合成数据和真实数据进行的实验表明,与以前的生成二进制代码的方法相比,我们的新算法始终能够生成更好的编码曝光问题的二进制序列,产生更好的去模糊和分辨率增强结果。
{"title":"Fluttering Pattern Generation Using Modified Legendre Sequence for Coded Exposure Imaging","authors":"Hae-Gon Jeon, Joon-Young Lee, Yudeog Han, Seon Joo Kim, In-So Kweon","doi":"10.1109/ICCV.2013.128","DOIUrl":"https://doi.org/10.1109/ICCV.2013.128","url":null,"abstract":"Finding a good binary sequence is critical in determining the performance of the coded exposure imaging, but previous methods mostly rely on a random search for finding the binary codes, which could easily fail to find good long sequences due to the exponentially growing search space. In this paper, we present a new computationally efficient algorithm for generating the binary sequence, which is especially well suited for longer sequences. We show that the concept of the low autocorrelation binary sequence that has been well exploited in the information theory community can be applied for generating the fluttering patterns of the shutter, propose a new measure of a good binary sequence, and present a new algorithm by modifying the Legendre sequence for the coded exposure imaging. Experiments using both synthetic and real data show that our new algorithm consistently generates better binary sequences for the coded exposure problem, yielding better deblurring and resolution enhancement results compared to the previous methods for generating the binary codes.","PeriodicalId":6351,"journal":{"name":"2013 IEEE International Conference on Computer Vision","volume":"13 1","pages":"1001-1008"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75213966","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Enhanced Continuous Tabu Search for Parameter Estimation in Multiview Geometry 多视图几何参数估计的增强连续禁忌搜索
Pub Date : 2013-12-01 DOI: 10.1109/ICCV.2013.402
Guoqing Zhou, Qing Wang
Optimization using the L_infty norm has been becoming an effective way to solve parameter estimation problems in multiview geometry. But the computational cost increases rapidly with the size of measurement data. Although some strategies have been presented to improve the efficiency of L_infty optimization, it is still an open issue. In the paper, we propose a novel approach under the framework of enhanced continuous tabu search (ECTS) for generic parameter estimation in multiview geometry. ECTS is an optimization method in the domain of artificial intelligence, which has an interesting ability of covering a wide solution space by promoting the search far away from current solution and consecutively decreasing the possibility of trapping in the local minima. Taking the triangulation as an example, we propose the corresponding ways in the key steps of ECTS, diversification and intensification. We also present theoretical proof to guarantee the global convergence of search with probability one. Experimental results have validated that the ECTS based approach can obtain global optimum efficiently, especially for large scale dimension of parameter. Potentially, the novel ECTS based algorithm can be applied in many applications of multiview geometry.
利用l_inty范数进行优化已成为解决多视点几何参数估计问题的有效方法。但随着测量数据量的增加,计算成本迅速增加。虽然已经提出了一些策略来提高l_inty优化的效率,但它仍然是一个开放的问题。本文提出了一种基于增强连续禁忌搜索(ECTS)框架的多视图几何通用参数估计方法。ECTS是人工智能领域的一种优化方法,它具有一种有趣的能力,即通过促进远离当前解的搜索,并不断降低陷入局部极小值的可能性,从而覆盖广泛的解空间。并以三角网为例,在ECTS、多样化和集约化的关键环节提出了相应的对策。并给出了保证搜索全局收敛概率为1的理论证明。实验结果表明,基于ECTS的方法能够有效地获得全局最优,特别是对于大尺度的参数。这种新的基于ECTS的算法可以应用于多视图几何的许多应用中。
{"title":"Enhanced Continuous Tabu Search for Parameter Estimation in Multiview Geometry","authors":"Guoqing Zhou, Qing Wang","doi":"10.1109/ICCV.2013.402","DOIUrl":"https://doi.org/10.1109/ICCV.2013.402","url":null,"abstract":"Optimization using the L_infty norm has been becoming an effective way to solve parameter estimation problems in multiview geometry. But the computational cost increases rapidly with the size of measurement data. Although some strategies have been presented to improve the efficiency of L_infty optimization, it is still an open issue. In the paper, we propose a novel approach under the framework of enhanced continuous tabu search (ECTS) for generic parameter estimation in multiview geometry. ECTS is an optimization method in the domain of artificial intelligence, which has an interesting ability of covering a wide solution space by promoting the search far away from current solution and consecutively decreasing the possibility of trapping in the local minima. Taking the triangulation as an example, we propose the corresponding ways in the key steps of ECTS, diversification and intensification. We also present theoretical proof to guarantee the global convergence of search with probability one. Experimental results have validated that the ECTS based approach can obtain global optimum efficiently, especially for large scale dimension of parameter. Potentially, the novel ECTS based algorithm can be applied in many applications of multiview geometry.","PeriodicalId":6351,"journal":{"name":"2013 IEEE International Conference on Computer Vision","volume":"52 1","pages":"3240-3247"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74174954","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Hierarchical Part Matching for Fine-Grained Visual Categorization 面向细粒度视觉分类的分层部件匹配
Pub Date : 2013-12-01 DOI: 10.1109/ICCV.2013.206
Lingxi Xie, Q. Tian, Richang Hong, Shuicheng Yan, Bo Zhang
As a special topic in computer vision, fine-grained visual categorization (FGVC) has been attracting growing attention these years. Different with traditional image classification tasks in which objects have large inter-class variation, the visual concepts in the fine-grained datasets, such as hundreds of bird species, often have very similar semantics. Due to the large inter-class similarity, it is very difficult to classify the objects without locating really discriminative features, therefore it becomes more important for the algorithm to make full use of the part information in order to train a robust model. In this paper, we propose a powerful flowchart named Hierarchical Part Matching (HPM) to cope with fine-grained classification tasks. We extend the Bag-of-Features (BoF) model by introducing several novel modules to integrate into image representation, including foreground inference and segmentation, Hierarchical Structure Learning (HSL), and Geometric Phrase Pooling (GPP). We verify in experiments that our algorithm achieves the state-of-the-art classification accuracy in the Caltech-UCSD-Birds-200-2011 dataset by making full use of the ground-truth part annotations.
细粒度视觉分类(FGVC)作为计算机视觉领域的一个特殊研究课题,近年来受到越来越多的关注。与传统图像分类任务中对象具有较大的类间差异不同,细粒度数据集(如数百种鸟类)中的视觉概念通常具有非常相似的语义。由于类间相似性较大,如果不找到真正的判别特征,则很难对目标进行分类,因此充分利用零件信息以训练出鲁棒模型就显得尤为重要。在本文中,我们提出了一个强大的流程图,称为层次匹配(HPM)来处理细粒度的分类任务。我们扩展了特征袋(BoF)模型,引入了几个新的模块集成到图像表示中,包括前景推断和分割、层次结构学习(HSL)和几何短语池(GPP)。通过实验验证,我们的算法充分利用了ground-truth部分注释,在Caltech-UCSD-Birds-200-2011数据集中达到了最先进的分类精度。
{"title":"Hierarchical Part Matching for Fine-Grained Visual Categorization","authors":"Lingxi Xie, Q. Tian, Richang Hong, Shuicheng Yan, Bo Zhang","doi":"10.1109/ICCV.2013.206","DOIUrl":"https://doi.org/10.1109/ICCV.2013.206","url":null,"abstract":"As a special topic in computer vision, fine-grained visual categorization (FGVC) has been attracting growing attention these years. Different with traditional image classification tasks in which objects have large inter-class variation, the visual concepts in the fine-grained datasets, such as hundreds of bird species, often have very similar semantics. Due to the large inter-class similarity, it is very difficult to classify the objects without locating really discriminative features, therefore it becomes more important for the algorithm to make full use of the part information in order to train a robust model. In this paper, we propose a powerful flowchart named Hierarchical Part Matching (HPM) to cope with fine-grained classification tasks. We extend the Bag-of-Features (BoF) model by introducing several novel modules to integrate into image representation, including foreground inference and segmentation, Hierarchical Structure Learning (HSL), and Geometric Phrase Pooling (GPP). We verify in experiments that our algorithm achieves the state-of-the-art classification accuracy in the Caltech-UCSD-Birds-200-2011 dataset by making full use of the ground-truth part annotations.","PeriodicalId":6351,"journal":{"name":"2013 IEEE International Conference on Computer Vision","volume":"50 1","pages":"1641-1648"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74702930","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 137
Optimization Problems for Fast AAM Fitting in-the-Wild 野外快速AAM拟合的优化问题
Pub Date : 2013-12-01 DOI: 10.1109/ICCV.2013.79
Georgios Tzimiropoulos, M. Pantic
We describe a very simple framework for deriving the most-well known optimization problems in Active Appearance Models (AAMs), and most importantly for providing efficient solutions. Our formulation results in two optimization problems for fast and exact AAM fitting, and one new algorithm which has the important advantage of being applicable to 3D. We show that the dominant cost for both forward and inverse algorithms is a few times mN which is the cost of projecting an image onto the appearance subspace. This makes both algorithms not only computationally realizable but also very attractive speed-wise for most current systems. Because exact AAM fitting is no longer computationally prohibitive, we trained AAMs in-the-wild with the goal of investigating whether AAMs benefit from such a training process. Our results show that although we did not use sophisticated shape priors, robust features or robust norms for improving performance, AAMs perform notably well and in some cases comparably with current state-of-the-art methods. We provide Matlab source code for training, fitting and reproducing the results presented in this paper at http://ibug.doc.ic.ac.uk/resources.
我们描述了一个非常简单的框架,用于推导活动外观模型(AAMs)中最著名的优化问题,最重要的是提供有效的解决方案。我们的公式解决了两个快速精确的AAM拟合优化问题,以及一个新的算法,该算法具有适用于三维的重要优势。我们证明了正向和逆算法的主要成本是mN的几倍,这是将图像投影到外观子空间的成本。这使得这两种算法不仅在计算上可实现,而且在速度方面对大多数当前系统都非常有吸引力。由于精确的AAM拟合不再是计算上的禁忌,我们在野外训练AAM,目的是调查AAM是否从这样的训练过程中受益。我们的研究结果表明,尽管我们没有使用复杂的形状先验、稳健的特征或稳健的规范来提高性能,但AAMs的表现非常好,在某些情况下可以与当前最先进的方法相媲美。我们在http://ibug.doc.ic.ac.uk/resources上提供了用于训练、拟合和再现本文中给出的结果的Matlab源代码。
{"title":"Optimization Problems for Fast AAM Fitting in-the-Wild","authors":"Georgios Tzimiropoulos, M. Pantic","doi":"10.1109/ICCV.2013.79","DOIUrl":"https://doi.org/10.1109/ICCV.2013.79","url":null,"abstract":"We describe a very simple framework for deriving the most-well known optimization problems in Active Appearance Models (AAMs), and most importantly for providing efficient solutions. Our formulation results in two optimization problems for fast and exact AAM fitting, and one new algorithm which has the important advantage of being applicable to 3D. We show that the dominant cost for both forward and inverse algorithms is a few times mN which is the cost of projecting an image onto the appearance subspace. This makes both algorithms not only computationally realizable but also very attractive speed-wise for most current systems. Because exact AAM fitting is no longer computationally prohibitive, we trained AAMs in-the-wild with the goal of investigating whether AAMs benefit from such a training process. Our results show that although we did not use sophisticated shape priors, robust features or robust norms for improving performance, AAMs perform notably well and in some cases comparably with current state-of-the-art methods. We provide Matlab source code for training, fitting and reproducing the results presented in this paper at http://ibug.doc.ic.ac.uk/resources.","PeriodicalId":6351,"journal":{"name":"2013 IEEE International Conference on Computer Vision","volume":"21 1","pages":"593-600"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75254306","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 250
Shufflets: Shared Mid-level Parts for Fast Object Detection Shufflets:用于快速对象检测的共享中级部件
Pub Date : 2013-12-01 DOI: 10.1109/ICCV.2013.176
Iasonas Kokkinos
We present a method to identify and exploit structures that are shared across different object categories, by using sparse coding to learn a shared basis for the 'part' and 'root' templates of Deformable Part Models (DPMs).Our first contribution consists in using Shift-Invariant Sparse Coding (SISC) to learn mid-level elements that can translate during coding. This results in systematically better approximations than those attained using standard sparse coding. To emphasize that the learned mid-level structures are shiftable we call them shufflets.Our second contribution consists in using the resulting score to construct probabilistic upper bounds to the exact template scores, instead of taking them 'at face value' as is common in current works. We integrate shufflets in Dual- Tree Branch-and-Bound and cascade-DPMs and demonstrate that we can achieve a substantial acceleration, with practically no loss in performance.
我们提出了一种方法,通过使用稀疏编码来学习可变形零件模型(dpm)的“部分”和“根”模板的共享基础,来识别和利用跨不同对象类别共享的结构。我们的第一个贡献是使用平移不变稀疏编码(SISC)来学习可以在编码过程中翻译的中级元素。这比使用标准稀疏编码获得的近似结果系统地更好。为了强调习得的中级结构是可移动的,我们称它们为shufflet。我们的第二个贡献在于使用结果分数来构建精确模板分数的概率上限,而不是像当前工作中常见的那样“表面价值”。我们在双树分支绑定和级联dpm中集成了shufflets,并证明了我们可以在几乎没有性能损失的情况下实现实质性的加速。
{"title":"Shufflets: Shared Mid-level Parts for Fast Object Detection","authors":"Iasonas Kokkinos","doi":"10.1109/ICCV.2013.176","DOIUrl":"https://doi.org/10.1109/ICCV.2013.176","url":null,"abstract":"We present a method to identify and exploit structures that are shared across different object categories, by using sparse coding to learn a shared basis for the 'part' and 'root' templates of Deformable Part Models (DPMs).Our first contribution consists in using Shift-Invariant Sparse Coding (SISC) to learn mid-level elements that can translate during coding. This results in systematically better approximations than those attained using standard sparse coding. To emphasize that the learned mid-level structures are shiftable we call them shufflets.Our second contribution consists in using the resulting score to construct probabilistic upper bounds to the exact template scores, instead of taking them 'at face value' as is common in current works. We integrate shufflets in Dual- Tree Branch-and-Bound and cascade-DPMs and demonstrate that we can achieve a substantial acceleration, with practically no loss in performance.","PeriodicalId":6351,"journal":{"name":"2013 IEEE International Conference on Computer Vision","volume":"22 1","pages":"1393-1400"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"72823109","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
Linear Sequence Discriminant Analysis: A Model-Based Dimensionality Reduction Method for Vector Sequences 线性序列判别分析:基于模型的向量序列降维方法
Pub Date : 2013-12-01 DOI: 10.1109/ICCV.2013.115
Bing Su, Xiaoqing Ding
Dimensionality reduction for vectors in sequences is challenging since labels are attached to sequences as a whole. This paper presents a model-based dimensionality reduction method for vector sequences, namely linear sequence discriminant analysis (LSDA), which attempts to find a subspace in which sequences of the same class are projected together while those of different classes are projected as far as possible. For each sequence class, an HMM is built from states of which statistics are extracted. Means of these states are linked in order to form a mean sequence, and the variance of the sequence class is defined as the sum of all variances of component states. LSDA then learns a transformation by maximizing the separability between sequence classes and at the same time minimizing the within-sequence class scatter. DTW distance between mean sequences is used to measure the separability between sequence classes. We show that the optimization problem can be approximately transformed into an eigen decomposition problem. LDA can be seen as a special case of LSDA by considering non-sequential vectors as sequences of length one. The effectiveness of the proposed LSDA is demonstrated on two individual sequence datasets from UCI machine learning repository as well as two concatenate sequence datasets: APTI Arabic printed text database and IFN/ENIT Arabic handwriting database.
序列中向量的降维具有挑战性,因为标签是作为一个整体附加在序列上的。本文提出了一种基于模型的向量序列降维方法,即线性序列判别分析(LSDA),它试图找到一个子空间,在该子空间中,同一类序列被投影在一起,而不同类的序列被尽可能地投影。对于每个序列类,根据提取统计信息的状态构建HMM。将这些状态的均值联系起来,形成均值序列,序列类的方差定义为各分量状态方差的总和。然后LSDA通过最大化序列类之间的可分离性,同时最小化序列类内的分散来学习转换。平均序列之间的DTW距离用于度量序列类之间的可分性。我们证明了优化问题可以近似地转化为特征分解问题。LDA可以看作是LDA的一种特殊情况,它将非顺序向量视为长度为1的序列。在来自UCI机器学习存储库的两个单独的序列数据集以及两个连接的序列数据集:APTI阿拉伯印刷文本数据库和IFN/ENIT阿拉伯手写数据库上,证明了所提出的LSDA的有效性。
{"title":"Linear Sequence Discriminant Analysis: A Model-Based Dimensionality Reduction Method for Vector Sequences","authors":"Bing Su, Xiaoqing Ding","doi":"10.1109/ICCV.2013.115","DOIUrl":"https://doi.org/10.1109/ICCV.2013.115","url":null,"abstract":"Dimensionality reduction for vectors in sequences is challenging since labels are attached to sequences as a whole. This paper presents a model-based dimensionality reduction method for vector sequences, namely linear sequence discriminant analysis (LSDA), which attempts to find a subspace in which sequences of the same class are projected together while those of different classes are projected as far as possible. For each sequence class, an HMM is built from states of which statistics are extracted. Means of these states are linked in order to form a mean sequence, and the variance of the sequence class is defined as the sum of all variances of component states. LSDA then learns a transformation by maximizing the separability between sequence classes and at the same time minimizing the within-sequence class scatter. DTW distance between mean sequences is used to measure the separability between sequence classes. We show that the optimization problem can be approximately transformed into an eigen decomposition problem. LDA can be seen as a special case of LSDA by considering non-sequential vectors as sequences of length one. The effectiveness of the proposed LSDA is demonstrated on two individual sequence datasets from UCI machine learning repository as well as two concatenate sequence datasets: APTI Arabic printed text database and IFN/ENIT Arabic handwriting database.","PeriodicalId":6351,"journal":{"name":"2013 IEEE International Conference on Computer Vision","volume":"49 1","pages":"889-896"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77171732","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Prime Object Proposals with Randomized Prim's Algorithm 随机Prim算法的素数目标建议
Pub Date : 2013-12-01 DOI: 10.1109/ICCV.2013.315
Santiago Manén, M. Guillaumin, L. Gool
Generic object detection is the challenging task of proposing windows that localize all the objects in an image, regardless of their classes. Such detectors have recently been shown to benefit many applications such as speeding-up class-specific object detection, weakly supervised learning of object detectors and object discovery. In this paper, we introduce a novel and very efficient method for generic object detection based on a randomized version of Prim's algorithm. Using the connectivity graph of an image's super pixels, with weights modelling the probability that neighbouring super pixels belong to the same object, the algorithm generates random partial spanning trees with large expected sum of edge weights. Object localizations are proposed as bounding-boxes of those partial trees. Our method has several benefits compared to the state-of-the-art. Thanks to the efficiency of Prim's algorithm, it samples proposals very quickly: 1000 proposals are obtained in about 0.7s. With proposals bound to super pixel boundaries yet diversified by randomization, it yields very high detection rates and windows that tightly fit objects. In extensive experiments on the challenging PASCAL VOC 2007 and 2012 and SUN2012 benchmark datasets, we show that our method improves over state-of-the-art competitors for a wide range of evaluation scenarios.
通用对象检测是一项具有挑战性的任务,它提出的窗口可以定位图像中的所有对象,而不考虑它们的类别。这种检测器最近被证明有益于许多应用,例如加速特定类的对象检测、对象检测器的弱监督学习和对象发现。本文提出了一种基于随机化Prim算法的通用目标检测方法。该算法利用图像超像素的连通性图,利用权重建模相邻超像素属于同一对象的概率,生成具有较大期望边权和的随机部分生成树。目标定位被提出为这些部分树的边界框。与最先进的方法相比,我们的方法有几个好处。由于Prim算法的效率,它对提议进行采样的速度非常快,大约0.7s就能得到1000个提议。由于建议绑定到超像素边界,但通过随机化多样化,它产生非常高的检测率和紧密适合对象的窗口。在具有挑战性的PASCAL VOC 2007和2012以及SUN2012基准数据集的广泛实验中,我们表明,我们的方法在广泛的评估场景中优于最先进的竞争对手。
{"title":"Prime Object Proposals with Randomized Prim's Algorithm","authors":"Santiago Manén, M. Guillaumin, L. Gool","doi":"10.1109/ICCV.2013.315","DOIUrl":"https://doi.org/10.1109/ICCV.2013.315","url":null,"abstract":"Generic object detection is the challenging task of proposing windows that localize all the objects in an image, regardless of their classes. Such detectors have recently been shown to benefit many applications such as speeding-up class-specific object detection, weakly supervised learning of object detectors and object discovery. In this paper, we introduce a novel and very efficient method for generic object detection based on a randomized version of Prim's algorithm. Using the connectivity graph of an image's super pixels, with weights modelling the probability that neighbouring super pixels belong to the same object, the algorithm generates random partial spanning trees with large expected sum of edge weights. Object localizations are proposed as bounding-boxes of those partial trees. Our method has several benefits compared to the state-of-the-art. Thanks to the efficiency of Prim's algorithm, it samples proposals very quickly: 1000 proposals are obtained in about 0.7s. With proposals bound to super pixel boundaries yet diversified by randomization, it yields very high detection rates and windows that tightly fit objects. In extensive experiments on the challenging PASCAL VOC 2007 and 2012 and SUN2012 benchmark datasets, we show that our method improves over state-of-the-art competitors for a wide range of evaluation scenarios.","PeriodicalId":6351,"journal":{"name":"2013 IEEE International Conference on Computer Vision","volume":"34 1","pages":"2536-2543"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77601515","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 303
Automatic Kronecker Product Model Based Detection of Repeated Patterns in 2D Urban Images 基于Kronecker产品模型的二维城市图像重复模式自动检测
Pub Date : 2013-12-01 DOI: 10.1109/ICCV.2013.57
Juan Liu, E. Psarakis, I. Stamos
Repeated patterns (such as windows, tiles, balconies and doors) are prominent and significant features in urban scenes. Therefore, detection of these repeated patterns becomes very important for city scene analysis. This paper attacks the problem of repeated patterns detection in a precise, efficient and automatic way, by combining traditional feature extraction followed by a Kronecker product low-rank modeling approach. Our method is tailored for 2D images of building facades. We have developed algorithms for automatic selection of a representative texture within facade images using vanishing points and Harris corners. After rectifying the input images, we describe novel algorithms that extract repeated patterns by using Kronecker product based modeling that is based on a solid theoretical foundation. Our approach is unique and has not ever been used for facade analysis. We have tested our algorithms in a large set of images.
重复的图案(如窗户、瓷砖、阳台和门)是城市场景中突出而重要的特征。因此,这些重复模式的检测对于城市场景分析就变得非常重要。本文将传统的特征提取与Kronecker积低秩建模相结合,以精确、高效、自动化的方式解决了重复模式检测问题。我们的方法是为建筑立面的二维图像量身定制的。我们已经开发了使用消失点和哈里斯角在立面图像中自动选择代表性纹理的算法。在对输入图像进行校正后,我们描述了基于坚实理论基础的基于Kronecker积的建模来提取重复模式的新算法。我们的方法是独一无二的,从未被用于立面分析。我们已经在大量图像中测试了我们的算法。
{"title":"Automatic Kronecker Product Model Based Detection of Repeated Patterns in 2D Urban Images","authors":"Juan Liu, E. Psarakis, I. Stamos","doi":"10.1109/ICCV.2013.57","DOIUrl":"https://doi.org/10.1109/ICCV.2013.57","url":null,"abstract":"Repeated patterns (such as windows, tiles, balconies and doors) are prominent and significant features in urban scenes. Therefore, detection of these repeated patterns becomes very important for city scene analysis. This paper attacks the problem of repeated patterns detection in a precise, efficient and automatic way, by combining traditional feature extraction followed by a Kronecker product low-rank modeling approach. Our method is tailored for 2D images of building facades. We have developed algorithms for automatic selection of a representative texture within facade images using vanishing points and Harris corners. After rectifying the input images, we describe novel algorithms that extract repeated patterns by using Kronecker product based modeling that is based on a solid theoretical foundation. Our approach is unique and has not ever been used for facade analysis. We have tested our algorithms in a large set of images.","PeriodicalId":6351,"journal":{"name":"2013 IEEE International Conference on Computer Vision","volume":"85 1","pages":"401-408"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76563391","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Hierarchical Data-Driven Descent for Efficient Optimal Deformation Estimation 有效的最优变形估计的分层数据驱动下降
Pub Date : 2013-12-01 DOI: 10.1109/ICCV.2013.284
Yuandong Tian, S. Narasimhan
Real-world surfaces such as clothing, water and human body deform in complex ways. The image distortions observed are high-dimensional and non-linear, making it hard to estimate these deformations accurately. The recent data-driven descent approach applies Nearest Neighbor estimators iteratively on a particular distribution of training samples to obtain a globally optimal and dense deformation field between a template and a distorted image. In this work, we develop a hierarchical structure for the Nearest Neighbor estimators, each of which can have only a local image support. We demonstrate in both theory and practice that this algorithm has several advantages over the non-hierarchical version: it guarantees global optimality with significantly fewer training samples, is several orders faster, provides a metric to decide whether a given image is ``hard'' (or ``easy'') requiring more (or less) samples, and can handle more complex scenes that include both global motion and local deformation. The proposed algorithm successfully tracks a broad range of non-rigid scenes including water, clothing, and medical images, and compares favorably against several other deformation estimation and tracking approaches that do not provide optimality guarantees.
现实世界的表面,如衣服、水和人体,都以复杂的方式变形。观察到的图像畸变是高维和非线性的,很难准确估计这些变形。最近的数据驱动下降方法在特定的训练样本分布上迭代地应用最近邻估计器来获得模板和变形图像之间的全局最优和密集的变形场。在这项工作中,我们为最近邻估计器开发了一个分层结构,每个最近邻估计器只能有一个局部图像支持。我们在理论和实践中都证明了这种算法比非分层版本有几个优点:它保证了用更少的训练样本实现全局最优,速度快几个数量级,提供了一个指标来决定给定图像是需要更多(或更少)样本的“难”(或“容易”),并且可以处理更复杂的场景,包括全局运动和局部变形。该算法成功地跟踪了广泛的非刚性场景,包括水、衣服和医学图像,并且与其他几种不提供最优性保证的变形估计和跟踪方法相比具有优势。
{"title":"Hierarchical Data-Driven Descent for Efficient Optimal Deformation Estimation","authors":"Yuandong Tian, S. Narasimhan","doi":"10.1109/ICCV.2013.284","DOIUrl":"https://doi.org/10.1109/ICCV.2013.284","url":null,"abstract":"Real-world surfaces such as clothing, water and human body deform in complex ways. The image distortions observed are high-dimensional and non-linear, making it hard to estimate these deformations accurately. The recent data-driven descent approach applies Nearest Neighbor estimators iteratively on a particular distribution of training samples to obtain a globally optimal and dense deformation field between a template and a distorted image. In this work, we develop a hierarchical structure for the Nearest Neighbor estimators, each of which can have only a local image support. We demonstrate in both theory and practice that this algorithm has several advantages over the non-hierarchical version: it guarantees global optimality with significantly fewer training samples, is several orders faster, provides a metric to decide whether a given image is ``hard'' (or ``easy'') requiring more (or less) samples, and can handle more complex scenes that include both global motion and local deformation. The proposed algorithm successfully tracks a broad range of non-rigid scenes including water, clothing, and medical images, and compares favorably against several other deformation estimation and tracking approaches that do not provide optimality guarantees.","PeriodicalId":6351,"journal":{"name":"2013 IEEE International Conference on Computer Vision","volume":"22 1","pages":"2288-2295"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74622269","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
期刊
2013 IEEE International Conference on Computer Vision
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1