首页 > 最新文献

The First Asian Conference on Pattern Recognition最新文献

英文 中文
A vision-based walking motion parameters capturing system 基于视觉的步行运动参数捕获系统
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166625
W. Lin, Chung-Lin Huang, Shih-Chung Hsu, Hung-Wei Lin, Hau-Wei Wang
The markerless vision-based human motion parameters capturing has been widely applied for human-machine interface. However, it faces two problems: the high-dimensional parameter estimation and the self-occlusion. Here, we propose a 3-D human model with structural, kinematic, and temporal constraints to track a walking human object in any viewing direction. Our method modifies the Annealed Particle Filter (APF) by applying the pre-trained spatial correlation map and the temporal constraint to estimate the motion parameters of a walking human object. In the experiments, we demonstrate that the proposed method requires less computation time and generates more accurate results.
基于无标记视觉的人体运动参数捕获在人机界面中得到了广泛的应用。然而,它面临着两个问题:高维参数估计和自遮挡。在这里,我们提出了一个具有结构、运动学和时间约束的3-D人体模型,可以在任何观看方向上跟踪行走的人体物体。该方法对退火粒子滤波器(退火粒子滤波器)进行了改进,利用预训练的空间相关图和时间约束来估计人体行走物体的运动参数。实验结果表明,该方法计算时间短,计算结果准确。
{"title":"A vision-based walking motion parameters capturing system","authors":"W. Lin, Chung-Lin Huang, Shih-Chung Hsu, Hung-Wei Lin, Hau-Wei Wang","doi":"10.1109/ACPR.2011.6166625","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166625","url":null,"abstract":"The markerless vision-based human motion parameters capturing has been widely applied for human-machine interface. However, it faces two problems: the high-dimensional parameter estimation and the self-occlusion. Here, we propose a 3-D human model with structural, kinematic, and temporal constraints to track a walking human object in any viewing direction. Our method modifies the Annealed Particle Filter (APF) by applying the pre-trained spatial correlation map and the temporal constraint to estimate the motion parameters of a walking human object. In the experiments, we demonstrate that the proposed method requires less computation time and generates more accurate results.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127282219","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Building facade interpretation exploiting repetition and mixed templates 建筑立面解释利用重复和混合模板
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166611
Gang Zeng, Rui Gan, H. Zha
Like many natural and other man-made objects, buildings contain repeating elements. The repetition is an important cue for most applications, and can be partial, approximate or both. This paper presents a robust and accurate building facade interpretation algorithm that processes a single input image and efficiently discovers and extracts the repeating elements (e.g. windows) without any prior knowledge about their shape, intensity or structure. The method is based on locally registering certain key regions in pairs and using these matches to accumulate evidence for averaged templates. These templates are determined via the graph-theoretical concept of minimum spanning tree (MST) and via mutual information (MI). Based on the templates, the repeating elements are finally extracted from the input image. Real scene examples demonstrate the ability of the proposed algorithm to capture important high-level information about the structure of a building facade, which in turn can support further processing operations, including compression, segmentation, editing and reconstruction.
像许多自然和其他人造物体一样,建筑包含重复的元素。重复是大多数应用程序的重要提示,可以是部分的,近似的或两者兼而有之。本文提出了一种鲁棒且精确的建筑立面解释算法,该算法处理单个输入图像,并有效地发现和提取重复元素(例如窗户),而无需事先了解其形状,强度或结构。该方法基于局部注册某些关键区域,并利用这些匹配来积累平均模板的证据。这些模板是通过最小生成树(MST)和互信息(MI)的图论概念确定的。基于模板,最后从输入图像中提取重复元素。真实场景示例证明了该算法能够捕获有关建筑立面结构的重要高层信息,进而支持进一步的处理操作,包括压缩、分割、编辑和重建。
{"title":"Building facade interpretation exploiting repetition and mixed templates","authors":"Gang Zeng, Rui Gan, H. Zha","doi":"10.1109/ACPR.2011.6166611","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166611","url":null,"abstract":"Like many natural and other man-made objects, buildings contain repeating elements. The repetition is an important cue for most applications, and can be partial, approximate or both. This paper presents a robust and accurate building facade interpretation algorithm that processes a single input image and efficiently discovers and extracts the repeating elements (e.g. windows) without any prior knowledge about their shape, intensity or structure. The method is based on locally registering certain key regions in pairs and using these matches to accumulate evidence for averaged templates. These templates are determined via the graph-theoretical concept of minimum spanning tree (MST) and via mutual information (MI). Based on the templates, the repeating elements are finally extracted from the input image. Real scene examples demonstrate the ability of the proposed algorithm to capture important high-level information about the structure of a building facade, which in turn can support further processing operations, including compression, segmentation, editing and reconstruction.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"28 7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130329518","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Class-imbalance learning based discriminant analysis 基于阶级失衡学习的判别分析
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166659
Xiaoyuan Jing, Chao Lan, Min Li, Yong-Fang Yao, D. Zhang, Jing-yu Yang
Feature extraction is an important research topic in the field of pattern recognition. The class-specific idea tends to recast a traditional multi-class feature extraction and recognition task into several binary class problems, and therefore inevitably class imbalance problem, where the minority class is the specific class, and the majority class consists of all the other classes. However, discriminative information from binary class problems is usually limited, and imbalanced data may have negative effect on the recognition performance. For solving these problems, in this paper, we propose two novel approaches to learn discriminant features from imbalanced data, named class-balanced discrimination (CBD) and orthogonal CBD (OCBD). For a specific class, we select a reduced counterpart class whose data are nearest to the data of specific class, and further divide them into smaller subsets, each of which has the same size as the specific class, to achieve balance. Then, each subset is combined with the minority class, and linear discriminant analysis (LDA) is performed on them to extract discriminative vectors. To further remove redundant information, we impose orthogonal constraint on the extracted discriminant vectors among correlated classes. Experimental results on three public image databases demonstrate that the proposed approaches outperform several related image feature extraction and recognition methods.
特征提取是模式识别领域的一个重要研究课题。特定类的思想倾向于将传统的多类特征提取和识别任务改造为几个二元类问题,从而不可避免地出现类不平衡问题,即少数类是特定类,而多数类是由所有其他类组成。然而,从二分类问题中得到的判别信息通常是有限的,而且数据的不平衡可能会对识别性能产生负面影响。为了解决这些问题,本文提出了两种从不平衡数据中学习判别特征的新方法,即类平衡判别(class-balanced discrimination, CBD)和正交判别(orthogonal CBD, OCBD)。对于特定类,我们选择数据最接近特定类数据的约简对应类,并将其进一步划分为更小的子集,每个子集与特定类具有相同的大小,以达到平衡。然后,将每个子集与少数类结合,对其进行线性判别分析(LDA),提取判别向量;为了进一步去除冗余信息,我们对提取的相关类之间的判别向量施加正交约束。在三个公共图像数据库上的实验结果表明,该方法优于几种相关的图像特征提取和识别方法。
{"title":"Class-imbalance learning based discriminant analysis","authors":"Xiaoyuan Jing, Chao Lan, Min Li, Yong-Fang Yao, D. Zhang, Jing-yu Yang","doi":"10.1109/ACPR.2011.6166659","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166659","url":null,"abstract":"Feature extraction is an important research topic in the field of pattern recognition. The class-specific idea tends to recast a traditional multi-class feature extraction and recognition task into several binary class problems, and therefore inevitably class imbalance problem, where the minority class is the specific class, and the majority class consists of all the other classes. However, discriminative information from binary class problems is usually limited, and imbalanced data may have negative effect on the recognition performance. For solving these problems, in this paper, we propose two novel approaches to learn discriminant features from imbalanced data, named class-balanced discrimination (CBD) and orthogonal CBD (OCBD). For a specific class, we select a reduced counterpart class whose data are nearest to the data of specific class, and further divide them into smaller subsets, each of which has the same size as the specific class, to achieve balance. Then, each subset is combined with the minority class, and linear discriminant analysis (LDA) is performed on them to extract discriminative vectors. To further remove redundant information, we impose orthogonal constraint on the extracted discriminant vectors among correlated classes. Experimental results on three public image databases demonstrate that the proposed approaches outperform several related image feature extraction and recognition methods.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"19 12","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131658159","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Multi-view moving objects classification via transfer learning 基于迁移学习的多视角运动物体分类
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166551
Jianyun Liu, Yunhong Wang, Zhaoxiang Zhang, Yi Mo
Moving objects classification in traffic scene videos is a hot topic in recent years. It has significant meaning to intelligent traffic system by classifying moving traffic objects into pedestrians, motor vehicles, non-motor vehicles etc.. Traditional machine learning approaches make the assumption that source scene objects and target scene objects share same distributions, which does not hold for most occasions. Under this circumstance, large amount of manual labeling for target scene data is needed, which is time and labor consuming. In this paper, we introduce TrAdaBoost, a transfer learning algorithm, to bridge the gap between source and target scene. During training procedure, TrAdaBoost makes full use of the source scene data that is most similar to the target scene data so that only small number of labeled target scene data could help improve the performance significantly. The features used for classification are Histogram of Oriented Gradient features of the appearance based instances. The experiment results show the outstanding performance of the transfer learning method comparing with traditional machine learning algorithm.
交通场景视频中的运动目标分类是近年来研究的热点问题。将运动交通对象划分为行人、机动车、非机动车等,对智能交通系统具有重要意义。传统的机器学习方法假设源场景对象和目标场景对象共享相同的分布,这在大多数情况下是不成立的。在这种情况下,需要对目标场景数据进行大量的人工标注,耗时耗力。在本文中,我们引入了一种迁移学习算法TrAdaBoost来弥合源场景和目标场景之间的差距。在训练过程中,TrAdaBoost充分利用与目标场景数据最相似的源场景数据,只需要少量标记的目标场景数据就能显著提高性能。用于分类的特征是基于外观实例的定向梯度特征直方图。实验结果表明,与传统的机器学习算法相比,迁移学习方法具有优异的性能。
{"title":"Multi-view moving objects classification via transfer learning","authors":"Jianyun Liu, Yunhong Wang, Zhaoxiang Zhang, Yi Mo","doi":"10.1109/ACPR.2011.6166551","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166551","url":null,"abstract":"Moving objects classification in traffic scene videos is a hot topic in recent years. It has significant meaning to intelligent traffic system by classifying moving traffic objects into pedestrians, motor vehicles, non-motor vehicles etc.. Traditional machine learning approaches make the assumption that source scene objects and target scene objects share same distributions, which does not hold for most occasions. Under this circumstance, large amount of manual labeling for target scene data is needed, which is time and labor consuming. In this paper, we introduce TrAdaBoost, a transfer learning algorithm, to bridge the gap between source and target scene. During training procedure, TrAdaBoost makes full use of the source scene data that is most similar to the target scene data so that only small number of labeled target scene data could help improve the performance significantly. The features used for classification are Histogram of Oriented Gradient features of the appearance based instances. The experiment results show the outstanding performance of the transfer learning method comparing with traditional machine learning algorithm.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"194 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129279338","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Target-oriented shape modeling with structure constraint for image segmentation 基于结构约束的面向目标形状建模图像分割
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166707
Wuxia Zhang, Yuan Yuan, Xuelong Li, Pingkun Yan
Image segmentation plays a critical role in medical imaging applications, whereas it is still a challenging problem due to the complex shapes and complicated texture of structures in medical images. Model based methods have been widely used for medical image segmentation as a priori knowledge can be incorporated. Accurate shape prior estimation is one of the major factors affecting the accuracy of model based segmentation methods. This paper proposes a novel statistical shape modeling method, which aims to estimate target-oriented shape prior by applying the constraint from the intrinsic structure of the training shape set. The proposed shape modeling method is incorporated into a deformable model based framework for image segmentation. The experimental results showed that the proposed method can achieve more accurate segmentation compared with other existing methods.
图像分割在医学成像应用中起着至关重要的作用,但由于医学图像结构的形状和纹理复杂,图像分割仍然是一个具有挑战性的问题。基于模型的方法在医学图像分割中得到了广泛的应用,因为它可以纳入先验知识。准确的形状先验估计是影响基于模型的分割方法精度的主要因素之一。本文提出了一种新的统计形状建模方法,该方法利用训练形状集的内在结构约束来预估目标导向形状。将所提出的形状建模方法整合到基于可变形模型的图像分割框架中。实验结果表明,与现有的分割方法相比,该方法可以实现更精确的分割。
{"title":"Target-oriented shape modeling with structure constraint for image segmentation","authors":"Wuxia Zhang, Yuan Yuan, Xuelong Li, Pingkun Yan","doi":"10.1109/ACPR.2011.6166707","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166707","url":null,"abstract":"Image segmentation plays a critical role in medical imaging applications, whereas it is still a challenging problem due to the complex shapes and complicated texture of structures in medical images. Model based methods have been widely used for medical image segmentation as a priori knowledge can be incorporated. Accurate shape prior estimation is one of the major factors affecting the accuracy of model based segmentation methods. This paper proposes a novel statistical shape modeling method, which aims to estimate target-oriented shape prior by applying the constraint from the intrinsic structure of the training shape set. The proposed shape modeling method is incorporated into a deformable model based framework for image segmentation. The experimental results showed that the proposed method can achieve more accurate segmentation compared with other existing methods.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131345821","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning from error: A two-level combined model for image classification 从错误中学习:图像分类的两级组合模型
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166669
Mingyang Jiang, Chunxiao Li, Zirui Deng, Jufu Feng, Liwei Wang
We propose an error learning model for image classification. Motivated by the observation that classifiers trained using local grid regions of the images are often biased, i.e., contain many classification error, we present a two-level combined model to learn useful classification information from these errors, based on Bayes rule. We give theoretical analysis and explanation to show that this error learning model is effective to correct the classification errors made by the local region classifiers. We conduct extensive experiments on benchmark image classification datasets, promising results are obtained.
提出了一种用于图像分类的错误学习模型。由于观察到使用图像的局部网格区域训练的分类器经常有偏差,即包含许多分类错误,我们提出了一个基于贝叶斯规则的两级组合模型,从这些错误中学习有用的分类信息。理论分析和解释表明,该错误学习模型能够有效地修正局部区域分类器的分类错误。我们在基准图像分类数据集上进行了大量的实验,得到了令人满意的结果。
{"title":"Learning from error: A two-level combined model for image classification","authors":"Mingyang Jiang, Chunxiao Li, Zirui Deng, Jufu Feng, Liwei Wang","doi":"10.1109/ACPR.2011.6166669","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166669","url":null,"abstract":"We propose an error learning model for image classification. Motivated by the observation that classifiers trained using local grid regions of the images are often biased, i.e., contain many classification error, we present a two-level combined model to learn useful classification information from these errors, based on Bayes rule. We give theoretical analysis and explanation to show that this error learning model is effective to correct the classification errors made by the local region classifiers. We conduct extensive experiments on benchmark image classification datasets, promising results are obtained.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115013739","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An efficient self-learning people counting system 一个有效的自学计数系统
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166686
Jingwen Li, Lei Huang, Chang-ping Liu
People counting is a challenging task and has attracted much attention in the area of video surveillance. In this paper, we present an efficient self-learning people counting system which can count the exact number of people in a region of interest. This system based on bag-of-features model can effectively detect the pedestrians some of which are usually treated as background because they are static or move slowly. The system can also select pedestrian and non-pedestrian samples automatically and update the classifier in real-time to make it more suitable for certain specific scene. Experimental results on a practical public dataset named CASIA Pedestrian Counting Dataset show that the proposed people counting system is robust and accurate.
在视频监控领域,人员统计是一项具有挑战性的工作,一直备受关注。在本文中,我们提出了一种高效的自学习计数系统,它可以准确地计数出感兴趣区域内的人数。该系统基于特征袋模型,可以有效地检测行人,其中一些行人由于静止或移动缓慢而被视为背景。系统还可以自动选择行人和非行人样本,并实时更新分类器,使其更适合特定场景。在CASIA行人计数数据集上的实验结果表明,该算法具有较好的鲁棒性和准确性。
{"title":"An efficient self-learning people counting system","authors":"Jingwen Li, Lei Huang, Chang-ping Liu","doi":"10.1109/ACPR.2011.6166686","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166686","url":null,"abstract":"People counting is a challenging task and has attracted much attention in the area of video surveillance. In this paper, we present an efficient self-learning people counting system which can count the exact number of people in a region of interest. This system based on bag-of-features model can effectively detect the pedestrians some of which are usually treated as background because they are static or move slowly. The system can also select pedestrian and non-pedestrian samples automatically and update the classifier in real-time to make it more suitable for certain specific scene. Experimental results on a practical public dataset named CASIA Pedestrian Counting Dataset show that the proposed people counting system is robust and accurate.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115646972","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Elliptical symmetric distribution based maximal margin classification for hyperspectral imagery 基于椭圆对称分布的高光谱图像最大边界分类
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166571
Lin He, Z. Yu, Z. Gu, Yuanqing Li
It has been verified that hyperspectral data is statistically characterized by elliptical symmetric distribution. Accordingly, we introduce the ellipsoidal discriminant boundaries and present an elliptical symmetric distribution based maximal margin (ESD-MM) classifier for hypespectral classification. In this method, the characteristic of elliptical symmetric distribution (ESD) of hyperspectral data is combined with the maximal margin rule. This strategy enables the ESD-MM classifier to achieve good performance, especially when follows dimensionality reduction. Experimental results on real Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) data demonstrated that ESD-MM classifier has better performance than commonly used Bayes classifier, Fisher linear discriminant (FLD) and linear support vector machine (SVM).
验证了高光谱数据在统计上具有椭圆对称分布的特征。在此基础上,引入椭球体判别边界,提出了一种基于椭圆对称分布的最大边界分类器(ESD-MM)。该方法将高光谱数据的椭圆对称分布(ESD)特性与最大边界规则相结合。该策略使ESD-MM分类器能够获得良好的性能,特别是在进行降维时。在机载可见/红外成像光谱仪(AVIRIS)的真实数据上进行的实验结果表明,静电散射- mm分类器的分类性能优于常用的贝叶斯分类器、Fisher线性判别器(FLD)和线性支持向量机(SVM)。
{"title":"Elliptical symmetric distribution based maximal margin classification for hyperspectral imagery","authors":"Lin He, Z. Yu, Z. Gu, Yuanqing Li","doi":"10.1109/ACPR.2011.6166571","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166571","url":null,"abstract":"It has been verified that hyperspectral data is statistically characterized by elliptical symmetric distribution. Accordingly, we introduce the ellipsoidal discriminant boundaries and present an elliptical symmetric distribution based maximal margin (ESD-MM) classifier for hypespectral classification. In this method, the characteristic of elliptical symmetric distribution (ESD) of hyperspectral data is combined with the maximal margin rule. This strategy enables the ESD-MM classifier to achieve good performance, especially when follows dimensionality reduction. Experimental results on real Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) data demonstrated that ESD-MM classifier has better performance than commonly used Bayes classifier, Fisher linear discriminant (FLD) and linear support vector machine (SVM).","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116076639","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Interclass visual similarity based visual vocabulary learning 基于班级间视觉相似性的视觉词汇学习
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166597
Guangming Chang, Chunfen Yuan, Weiming Hu
Visual vocabulary is now widely used in many video analysis tasks, such as event detection, video retrieval and video classification. In most approaches the vocabularies are solely based on statistics of visual features and generated by clustering. Little attention has been paid to the interclass similarity among different events or actions. In this paper, we present a novel approach to mine the interclass visual similarity statistically and then use it to supervise the generation of visual vocabulary. We construct a measurement of interclass similarity, embed the similarity to the Euclidean distance and use the refined distance to generate visual vocabulary iteratively. The experiments in Weizmann and KTH datasets show that our approach outperforms the traditional vocabulary based approach by about 5%.
视觉词汇在视频事件检测、视频检索、视频分类等视频分析任务中得到了广泛的应用。在大多数方法中,词汇表仅基于视觉特征的统计并通过聚类生成。人们很少关注不同事件或行为之间的类间相似性。本文提出了一种统计挖掘类间视觉相似度的新方法,并用它来监督视觉词汇的生成。我们构建了类间相似度度量,将相似度嵌入到欧几里得距离中,并使用改进的距离迭代生成视觉词汇。在Weizmann和KTH数据集上的实验表明,我们的方法比传统的基于词汇的方法性能高出约5%。
{"title":"Interclass visual similarity based visual vocabulary learning","authors":"Guangming Chang, Chunfen Yuan, Weiming Hu","doi":"10.1109/ACPR.2011.6166597","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166597","url":null,"abstract":"Visual vocabulary is now widely used in many video analysis tasks, such as event detection, video retrieval and video classification. In most approaches the vocabularies are solely based on statistics of visual features and generated by clustering. Little attention has been paid to the interclass similarity among different events or actions. In this paper, we present a novel approach to mine the interclass visual similarity statistically and then use it to supervise the generation of visual vocabulary. We construct a measurement of interclass similarity, embed the similarity to the Euclidean distance and use the refined distance to generate visual vocabulary iteratively. The experiments in Weizmann and KTH datasets show that our approach outperforms the traditional vocabulary based approach by about 5%.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"87 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116086711","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Saliency based natural image understanding 基于自然图像理解的显著性
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166648
Qingshan Li, Yue Zhou, Lei Xu
This paper presents a novel method for natural image understanding. We improved the effect of saliency detection for the purpose of image segmentation at first. Then Graph cuts are used to find global optimal segmentation of N-dimensional image. After that, we adopt the scheme of supervised learning to classify the scene type of the image. The main advantages of our method are that: Firstly we revised the existed sparse saliency model to better suit for image segmentation, Secondly we propose a new color modeling method during the process of GrabCut segmentation. Finally we extract object-level top down information and low-level image cues together to distinguish the type of images. Experiments show that our proposed scheme can obtain comparable performance to other approaches.
提出了一种新的自然图像理解方法。我们首先改进了显著性检测的效果,以达到图像分割的目的。然后利用图切算法对n维图像进行全局最优分割。然后,我们采用监督学习的方案对图像的场景类型进行分类。该方法的主要优点是:首先,我们改进了现有的稀疏显著性模型,使其更适合图像分割;其次,我们在GrabCut分割过程中提出了一种新的颜色建模方法。最后,我们提取对象级自顶向下信息和底层图像线索,以区分图像的类型。实验结果表明,该方法可以获得与其他方法相当的性能。
{"title":"Saliency based natural image understanding","authors":"Qingshan Li, Yue Zhou, Lei Xu","doi":"10.1109/ACPR.2011.6166648","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166648","url":null,"abstract":"This paper presents a novel method for natural image understanding. We improved the effect of saliency detection for the purpose of image segmentation at first. Then Graph cuts are used to find global optimal segmentation of N-dimensional image. After that, we adopt the scheme of supervised learning to classify the scene type of the image. The main advantages of our method are that: Firstly we revised the existed sparse saliency model to better suit for image segmentation, Secondly we propose a new color modeling method during the process of GrabCut segmentation. Finally we extract object-level top down information and low-level image cues together to distinguish the type of images. Experiments show that our proposed scheme can obtain comparable performance to other approaches.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"256 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122502641","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
The First Asian Conference on Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1