首页 > 最新文献

The First Asian Conference on Pattern Recognition最新文献

英文 中文
A vision-based walking motion parameters capturing system 基于视觉的步行运动参数捕获系统
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166625
W. Lin, Chung-Lin Huang, Shih-Chung Hsu, Hung-Wei Lin, Hau-Wei Wang
The markerless vision-based human motion parameters capturing has been widely applied for human-machine interface. However, it faces two problems: the high-dimensional parameter estimation and the self-occlusion. Here, we propose a 3-D human model with structural, kinematic, and temporal constraints to track a walking human object in any viewing direction. Our method modifies the Annealed Particle Filter (APF) by applying the pre-trained spatial correlation map and the temporal constraint to estimate the motion parameters of a walking human object. In the experiments, we demonstrate that the proposed method requires less computation time and generates more accurate results.
基于无标记视觉的人体运动参数捕获在人机界面中得到了广泛的应用。然而,它面临着两个问题:高维参数估计和自遮挡。在这里,我们提出了一个具有结构、运动学和时间约束的3-D人体模型,可以在任何观看方向上跟踪行走的人体物体。该方法对退火粒子滤波器(退火粒子滤波器)进行了改进,利用预训练的空间相关图和时间约束来估计人体行走物体的运动参数。实验结果表明,该方法计算时间短,计算结果准确。
{"title":"A vision-based walking motion parameters capturing system","authors":"W. Lin, Chung-Lin Huang, Shih-Chung Hsu, Hung-Wei Lin, Hau-Wei Wang","doi":"10.1109/ACPR.2011.6166625","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166625","url":null,"abstract":"The markerless vision-based human motion parameters capturing has been widely applied for human-machine interface. However, it faces two problems: the high-dimensional parameter estimation and the self-occlusion. Here, we propose a 3-D human model with structural, kinematic, and temporal constraints to track a walking human object in any viewing direction. Our method modifies the Annealed Particle Filter (APF) by applying the pre-trained spatial correlation map and the temporal constraint to estimate the motion parameters of a walking human object. In the experiments, we demonstrate that the proposed method requires less computation time and generates more accurate results.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127282219","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Building facade interpretation exploiting repetition and mixed templates 建筑立面解释利用重复和混合模板
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166611
Gang Zeng, Rui Gan, H. Zha
Like many natural and other man-made objects, buildings contain repeating elements. The repetition is an important cue for most applications, and can be partial, approximate or both. This paper presents a robust and accurate building facade interpretation algorithm that processes a single input image and efficiently discovers and extracts the repeating elements (e.g. windows) without any prior knowledge about their shape, intensity or structure. The method is based on locally registering certain key regions in pairs and using these matches to accumulate evidence for averaged templates. These templates are determined via the graph-theoretical concept of minimum spanning tree (MST) and via mutual information (MI). Based on the templates, the repeating elements are finally extracted from the input image. Real scene examples demonstrate the ability of the proposed algorithm to capture important high-level information about the structure of a building facade, which in turn can support further processing operations, including compression, segmentation, editing and reconstruction.
像许多自然和其他人造物体一样,建筑包含重复的元素。重复是大多数应用程序的重要提示,可以是部分的,近似的或两者兼而有之。本文提出了一种鲁棒且精确的建筑立面解释算法,该算法处理单个输入图像,并有效地发现和提取重复元素(例如窗户),而无需事先了解其形状,强度或结构。该方法基于局部注册某些关键区域,并利用这些匹配来积累平均模板的证据。这些模板是通过最小生成树(MST)和互信息(MI)的图论概念确定的。基于模板,最后从输入图像中提取重复元素。真实场景示例证明了该算法能够捕获有关建筑立面结构的重要高层信息,进而支持进一步的处理操作,包括压缩、分割、编辑和重建。
{"title":"Building facade interpretation exploiting repetition and mixed templates","authors":"Gang Zeng, Rui Gan, H. Zha","doi":"10.1109/ACPR.2011.6166611","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166611","url":null,"abstract":"Like many natural and other man-made objects, buildings contain repeating elements. The repetition is an important cue for most applications, and can be partial, approximate or both. This paper presents a robust and accurate building facade interpretation algorithm that processes a single input image and efficiently discovers and extracts the repeating elements (e.g. windows) without any prior knowledge about their shape, intensity or structure. The method is based on locally registering certain key regions in pairs and using these matches to accumulate evidence for averaged templates. These templates are determined via the graph-theoretical concept of minimum spanning tree (MST) and via mutual information (MI). Based on the templates, the repeating elements are finally extracted from the input image. Real scene examples demonstrate the ability of the proposed algorithm to capture important high-level information about the structure of a building facade, which in turn can support further processing operations, including compression, segmentation, editing and reconstruction.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"28 7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130329518","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Class-imbalance learning based discriminant analysis 基于阶级失衡学习的判别分析
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166659
Xiaoyuan Jing, Chao Lan, Min Li, Yong-Fang Yao, D. Zhang, Jing-yu Yang
Feature extraction is an important research topic in the field of pattern recognition. The class-specific idea tends to recast a traditional multi-class feature extraction and recognition task into several binary class problems, and therefore inevitably class imbalance problem, where the minority class is the specific class, and the majority class consists of all the other classes. However, discriminative information from binary class problems is usually limited, and imbalanced data may have negative effect on the recognition performance. For solving these problems, in this paper, we propose two novel approaches to learn discriminant features from imbalanced data, named class-balanced discrimination (CBD) and orthogonal CBD (OCBD). For a specific class, we select a reduced counterpart class whose data are nearest to the data of specific class, and further divide them into smaller subsets, each of which has the same size as the specific class, to achieve balance. Then, each subset is combined with the minority class, and linear discriminant analysis (LDA) is performed on them to extract discriminative vectors. To further remove redundant information, we impose orthogonal constraint on the extracted discriminant vectors among correlated classes. Experimental results on three public image databases demonstrate that the proposed approaches outperform several related image feature extraction and recognition methods.
特征提取是模式识别领域的一个重要研究课题。特定类的思想倾向于将传统的多类特征提取和识别任务改造为几个二元类问题,从而不可避免地出现类不平衡问题,即少数类是特定类,而多数类是由所有其他类组成。然而,从二分类问题中得到的判别信息通常是有限的,而且数据的不平衡可能会对识别性能产生负面影响。为了解决这些问题,本文提出了两种从不平衡数据中学习判别特征的新方法,即类平衡判别(class-balanced discrimination, CBD)和正交判别(orthogonal CBD, OCBD)。对于特定类,我们选择数据最接近特定类数据的约简对应类,并将其进一步划分为更小的子集,每个子集与特定类具有相同的大小,以达到平衡。然后,将每个子集与少数类结合,对其进行线性判别分析(LDA),提取判别向量;为了进一步去除冗余信息,我们对提取的相关类之间的判别向量施加正交约束。在三个公共图像数据库上的实验结果表明,该方法优于几种相关的图像特征提取和识别方法。
{"title":"Class-imbalance learning based discriminant analysis","authors":"Xiaoyuan Jing, Chao Lan, Min Li, Yong-Fang Yao, D. Zhang, Jing-yu Yang","doi":"10.1109/ACPR.2011.6166659","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166659","url":null,"abstract":"Feature extraction is an important research topic in the field of pattern recognition. The class-specific idea tends to recast a traditional multi-class feature extraction and recognition task into several binary class problems, and therefore inevitably class imbalance problem, where the minority class is the specific class, and the majority class consists of all the other classes. However, discriminative information from binary class problems is usually limited, and imbalanced data may have negative effect on the recognition performance. For solving these problems, in this paper, we propose two novel approaches to learn discriminant features from imbalanced data, named class-balanced discrimination (CBD) and orthogonal CBD (OCBD). For a specific class, we select a reduced counterpart class whose data are nearest to the data of specific class, and further divide them into smaller subsets, each of which has the same size as the specific class, to achieve balance. Then, each subset is combined with the minority class, and linear discriminant analysis (LDA) is performed on them to extract discriminative vectors. To further remove redundant information, we impose orthogonal constraint on the extracted discriminant vectors among correlated classes. Experimental results on three public image databases demonstrate that the proposed approaches outperform several related image feature extraction and recognition methods.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"19 12","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131658159","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Multi-view moving objects classification via transfer learning 基于迁移学习的多视角运动物体分类
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166551
Jianyun Liu, Yunhong Wang, Zhaoxiang Zhang, Yi Mo
Moving objects classification in traffic scene videos is a hot topic in recent years. It has significant meaning to intelligent traffic system by classifying moving traffic objects into pedestrians, motor vehicles, non-motor vehicles etc.. Traditional machine learning approaches make the assumption that source scene objects and target scene objects share same distributions, which does not hold for most occasions. Under this circumstance, large amount of manual labeling for target scene data is needed, which is time and labor consuming. In this paper, we introduce TrAdaBoost, a transfer learning algorithm, to bridge the gap between source and target scene. During training procedure, TrAdaBoost makes full use of the source scene data that is most similar to the target scene data so that only small number of labeled target scene data could help improve the performance significantly. The features used for classification are Histogram of Oriented Gradient features of the appearance based instances. The experiment results show the outstanding performance of the transfer learning method comparing with traditional machine learning algorithm.
交通场景视频中的运动目标分类是近年来研究的热点问题。将运动交通对象划分为行人、机动车、非机动车等,对智能交通系统具有重要意义。传统的机器学习方法假设源场景对象和目标场景对象共享相同的分布,这在大多数情况下是不成立的。在这种情况下,需要对目标场景数据进行大量的人工标注,耗时耗力。在本文中,我们引入了一种迁移学习算法TrAdaBoost来弥合源场景和目标场景之间的差距。在训练过程中,TrAdaBoost充分利用与目标场景数据最相似的源场景数据,只需要少量标记的目标场景数据就能显著提高性能。用于分类的特征是基于外观实例的定向梯度特征直方图。实验结果表明,与传统的机器学习算法相比,迁移学习方法具有优异的性能。
{"title":"Multi-view moving objects classification via transfer learning","authors":"Jianyun Liu, Yunhong Wang, Zhaoxiang Zhang, Yi Mo","doi":"10.1109/ACPR.2011.6166551","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166551","url":null,"abstract":"Moving objects classification in traffic scene videos is a hot topic in recent years. It has significant meaning to intelligent traffic system by classifying moving traffic objects into pedestrians, motor vehicles, non-motor vehicles etc.. Traditional machine learning approaches make the assumption that source scene objects and target scene objects share same distributions, which does not hold for most occasions. Under this circumstance, large amount of manual labeling for target scene data is needed, which is time and labor consuming. In this paper, we introduce TrAdaBoost, a transfer learning algorithm, to bridge the gap between source and target scene. During training procedure, TrAdaBoost makes full use of the source scene data that is most similar to the target scene data so that only small number of labeled target scene data could help improve the performance significantly. The features used for classification are Histogram of Oriented Gradient features of the appearance based instances. The experiment results show the outstanding performance of the transfer learning method comparing with traditional machine learning algorithm.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"194 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129279338","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Target-oriented shape modeling with structure constraint for image segmentation 基于结构约束的面向目标形状建模图像分割
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166707
Wuxia Zhang, Yuan Yuan, Xuelong Li, Pingkun Yan
Image segmentation plays a critical role in medical imaging applications, whereas it is still a challenging problem due to the complex shapes and complicated texture of structures in medical images. Model based methods have been widely used for medical image segmentation as a priori knowledge can be incorporated. Accurate shape prior estimation is one of the major factors affecting the accuracy of model based segmentation methods. This paper proposes a novel statistical shape modeling method, which aims to estimate target-oriented shape prior by applying the constraint from the intrinsic structure of the training shape set. The proposed shape modeling method is incorporated into a deformable model based framework for image segmentation. The experimental results showed that the proposed method can achieve more accurate segmentation compared with other existing methods.
图像分割在医学成像应用中起着至关重要的作用,但由于医学图像结构的形状和纹理复杂,图像分割仍然是一个具有挑战性的问题。基于模型的方法在医学图像分割中得到了广泛的应用,因为它可以纳入先验知识。准确的形状先验估计是影响基于模型的分割方法精度的主要因素之一。本文提出了一种新的统计形状建模方法,该方法利用训练形状集的内在结构约束来预估目标导向形状。将所提出的形状建模方法整合到基于可变形模型的图像分割框架中。实验结果表明,与现有的分割方法相比,该方法可以实现更精确的分割。
{"title":"Target-oriented shape modeling with structure constraint for image segmentation","authors":"Wuxia Zhang, Yuan Yuan, Xuelong Li, Pingkun Yan","doi":"10.1109/ACPR.2011.6166707","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166707","url":null,"abstract":"Image segmentation plays a critical role in medical imaging applications, whereas it is still a challenging problem due to the complex shapes and complicated texture of structures in medical images. Model based methods have been widely used for medical image segmentation as a priori knowledge can be incorporated. Accurate shape prior estimation is one of the major factors affecting the accuracy of model based segmentation methods. This paper proposes a novel statistical shape modeling method, which aims to estimate target-oriented shape prior by applying the constraint from the intrinsic structure of the training shape set. The proposed shape modeling method is incorporated into a deformable model based framework for image segmentation. The experimental results showed that the proposed method can achieve more accurate segmentation compared with other existing methods.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131345821","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Quadratic-chi similarity metric learning for histogram feature 直方图特征的二次chi相似度度量学习
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166698
Xinyuan Cai, Baihua Xiao, Chunheng Wang, Rongguo Zhang
Histogram features, such as SIFT, HOG, LBP et al, are widely used in modern computer vision algorithms. According to [18], chi-square distance is an effective measure for comparing histogram features. In this paper, we propose a new method, named the Quadric-chi similarity metric learning (QCSML) for histogram features. The main contribution of this paper is that we propose a new metric learning method based on chi-square distance, in contrast with traditional Mahalanobis distance metric learning methods. The use of quadric-chi similarity in our method leads to an effective learning algorithm. Our method is tested on SIFT features for face identification, and compared with the state-of-art metric learning method (LDML) on the benchmark dataset, the Labeled Faces in the Wild (LFW). Experimental results show that our method can achieve clear performance gains over LDML.
直方图特征在现代计算机视觉算法中得到了广泛的应用,如SIFT、HOG、LBP等。根据[18],卡方距离是比较直方图特征的有效度量。在本文中,我们提出了一种新的直方图特征的相似度度量学习(QCSML)方法。本文的主要贡献在于,与传统的马氏距离度量学习方法相比,我们提出了一种新的基于卡方距离的度量学习方法。在我们的方法中使用二次chi相似度导致了一个有效的学习算法。我们的方法在SIFT特征上进行了人脸识别测试,并在基准数据集Labeled Faces in the Wild (LFW)上与最先进的度量学习方法(LDML)进行了比较。实验结果表明,与LDML相比,我们的方法可以获得明显的性能提升。
{"title":"Quadratic-chi similarity metric learning for histogram feature","authors":"Xinyuan Cai, Baihua Xiao, Chunheng Wang, Rongguo Zhang","doi":"10.1109/ACPR.2011.6166698","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166698","url":null,"abstract":"Histogram features, such as SIFT, HOG, LBP et al, are widely used in modern computer vision algorithms. According to [18], chi-square distance is an effective measure for comparing histogram features. In this paper, we propose a new method, named the Quadric-chi similarity metric learning (QCSML) for histogram features. The main contribution of this paper is that we propose a new metric learning method based on chi-square distance, in contrast with traditional Mahalanobis distance metric learning methods. The use of quadric-chi similarity in our method leads to an effective learning algorithm. Our method is tested on SIFT features for face identification, and compared with the state-of-art metric learning method (LDML) on the benchmark dataset, the Labeled Faces in the Wild (LFW). Experimental results show that our method can achieve clear performance gains over LDML.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124818472","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Face recognition with continuous occlusion using partially iteratively reweighted sparse coding 基于部分迭代重加权稀疏编码的连续遮挡人脸识别
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166617
Xiao-Xin Li, D. Dai, Xiao-Fei Zhang, Chuan-Xian Ren
Partially occluded faces are common in automatic face recognition in the real world. Existing methods, such as sparse error correction with Markov random fields, correntropy-based sparse representation and robust sparse coding, are all based on error correction, which relies on the perfect reconstruction of the occluded facial image and limits their recognition rates especially when the occluded regions are large. It helps to enhance recognition rates if we can detect the occluded portions and exclude them from further classification. Based on a magnitude order measure, we propose an innovative effective occlusion detection algorithm, called Partially Iteratively Reweighted Sparse Coding (PIRSC). Compared to the state-of-the-art methods, our PIRSC based classifier greatly improve the face recognition rate especially when the occlusion percentage is large.
部分遮挡人脸是现实世界中人脸自动识别中常见的现象。现有的基于马尔可夫随机场的稀疏纠错、基于相关熵的稀疏表示和鲁棒稀疏编码等方法都是基于纠错的,这依赖于对被遮挡的面部图像的完美重建,并且限制了它们的识别率,特别是当被遮挡区域较大时。如果我们能够检测出遮挡部分并将其排除在进一步分类之外,将有助于提高识别率。基于数量级度量,我们提出了一种创新的有效遮挡检测算法,称为部分迭代重加权稀疏编码(PIRSC)。与现有的分类器相比,我们的基于PIRSC的分类器大大提高了人脸识别率,特别是在遮挡百分比较大的情况下。
{"title":"Face recognition with continuous occlusion using partially iteratively reweighted sparse coding","authors":"Xiao-Xin Li, D. Dai, Xiao-Fei Zhang, Chuan-Xian Ren","doi":"10.1109/ACPR.2011.6166617","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166617","url":null,"abstract":"Partially occluded faces are common in automatic face recognition in the real world. Existing methods, such as sparse error correction with Markov random fields, correntropy-based sparse representation and robust sparse coding, are all based on error correction, which relies on the perfect reconstruction of the occluded facial image and limits their recognition rates especially when the occluded regions are large. It helps to enhance recognition rates if we can detect the occluded portions and exclude them from further classification. Based on a magnitude order measure, we propose an innovative effective occlusion detection algorithm, called Partially Iteratively Reweighted Sparse Coding (PIRSC). Compared to the state-of-the-art methods, our PIRSC based classifier greatly improve the face recognition rate especially when the occlusion percentage is large.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122093256","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Human centric object detection in highly crowded scenes 高度拥挤场景中以人为中心的目标检测
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166674
Genquan Duan, H. Ai, Takayoshi Yamashita, S. Lao
In this paper, we propose to detect human centric objects, including face, head shoulder, upper body, left body, right body and whole body, which can provide essential information to locate humans in highly crowed scenes. In the literature, the approaches to detect multi-class objects are either taking each class independently to learn and apply its classifier successively or taking all classes as a whole to learn individual classifier based on sharing features and to detect by step-by-step dividing. Different from these works, we consider two issues, one is the similarities and discriminations of different classes and the other is the semantic relations among them. Our main idea is to predict class labels quickly using a Salient Patch Model (SPM) first, and then do detection accurately using detectors of predicted classes in which a Semantic Relation Model (SRM) is proposed to capture relations among classes for efficient inferences. SPM and SRM are designed for these two issues respectively. Experiments on challenging real-world datasets demonstrate that our proposed approach can achieve significant performance improvements.
本文提出对以人为中心的物体进行检测,包括人脸、头肩、上半身、左身体、右身体和全身,可以为在高度拥挤的场景中定位人类提供必要的信息。在文献中,多类物体检测的方法是:将每一类单独学习并依次应用其分类器,或将所有类作为一个整体,根据共享特征学习单个分类器,分步进行检测。与这些作品不同的是,我们考虑了两个问题,一个是不同类别之间的相似和区别,另一个是它们之间的语义关系。我们的主要思想是首先使用显著补丁模型(SPM)快速预测类标签,然后使用预测类的检测器进行准确检测,其中提出了语义关系模型(SRM)来捕获类之间的关系以进行有效推断。SPM和SRM分别针对这两个问题设计。在具有挑战性的真实数据集上的实验表明,我们提出的方法可以实现显着的性能改进。
{"title":"Human centric object detection in highly crowded scenes","authors":"Genquan Duan, H. Ai, Takayoshi Yamashita, S. Lao","doi":"10.1109/ACPR.2011.6166674","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166674","url":null,"abstract":"In this paper, we propose to detect human centric objects, including face, head shoulder, upper body, left body, right body and whole body, which can provide essential information to locate humans in highly crowed scenes. In the literature, the approaches to detect multi-class objects are either taking each class independently to learn and apply its classifier successively or taking all classes as a whole to learn individual classifier based on sharing features and to detect by step-by-step dividing. Different from these works, we consider two issues, one is the similarities and discriminations of different classes and the other is the semantic relations among them. Our main idea is to predict class labels quickly using a Salient Patch Model (SPM) first, and then do detection accurately using detectors of predicted classes in which a Semantic Relation Model (SRM) is proposed to capture relations among classes for efficient inferences. SPM and SRM are designed for these two issues respectively. Experiments on challenging real-world datasets demonstrate that our proposed approach can achieve significant performance improvements.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"113 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124747145","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning from error: A two-level combined model for image classification 从错误中学习:图像分类的两级组合模型
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166669
Mingyang Jiang, Chunxiao Li, Zirui Deng, Jufu Feng, Liwei Wang
We propose an error learning model for image classification. Motivated by the observation that classifiers trained using local grid regions of the images are often biased, i.e., contain many classification error, we present a two-level combined model to learn useful classification information from these errors, based on Bayes rule. We give theoretical analysis and explanation to show that this error learning model is effective to correct the classification errors made by the local region classifiers. We conduct extensive experiments on benchmark image classification datasets, promising results are obtained.
提出了一种用于图像分类的错误学习模型。由于观察到使用图像的局部网格区域训练的分类器经常有偏差,即包含许多分类错误,我们提出了一个基于贝叶斯规则的两级组合模型,从这些错误中学习有用的分类信息。理论分析和解释表明,该错误学习模型能够有效地修正局部区域分类器的分类错误。我们在基准图像分类数据集上进行了大量的实验,得到了令人满意的结果。
{"title":"Learning from error: A two-level combined model for image classification","authors":"Mingyang Jiang, Chunxiao Li, Zirui Deng, Jufu Feng, Liwei Wang","doi":"10.1109/ACPR.2011.6166669","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166669","url":null,"abstract":"We propose an error learning model for image classification. Motivated by the observation that classifiers trained using local grid regions of the images are often biased, i.e., contain many classification error, we present a two-level combined model to learn useful classification information from these errors, based on Bayes rule. We give theoretical analysis and explanation to show that this error learning model is effective to correct the classification errors made by the local region classifiers. We conduct extensive experiments on benchmark image classification datasets, promising results are obtained.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115013739","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An efficient self-learning people counting system 一个有效的自学计数系统
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166686
Jingwen Li, Lei Huang, Chang-ping Liu
People counting is a challenging task and has attracted much attention in the area of video surveillance. In this paper, we present an efficient self-learning people counting system which can count the exact number of people in a region of interest. This system based on bag-of-features model can effectively detect the pedestrians some of which are usually treated as background because they are static or move slowly. The system can also select pedestrian and non-pedestrian samples automatically and update the classifier in real-time to make it more suitable for certain specific scene. Experimental results on a practical public dataset named CASIA Pedestrian Counting Dataset show that the proposed people counting system is robust and accurate.
在视频监控领域,人员统计是一项具有挑战性的工作,一直备受关注。在本文中,我们提出了一种高效的自学习计数系统,它可以准确地计数出感兴趣区域内的人数。该系统基于特征袋模型,可以有效地检测行人,其中一些行人由于静止或移动缓慢而被视为背景。系统还可以自动选择行人和非行人样本,并实时更新分类器,使其更适合特定场景。在CASIA行人计数数据集上的实验结果表明,该算法具有较好的鲁棒性和准确性。
{"title":"An efficient self-learning people counting system","authors":"Jingwen Li, Lei Huang, Chang-ping Liu","doi":"10.1109/ACPR.2011.6166686","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166686","url":null,"abstract":"People counting is a challenging task and has attracted much attention in the area of video surveillance. In this paper, we present an efficient self-learning people counting system which can count the exact number of people in a region of interest. This system based on bag-of-features model can effectively detect the pedestrians some of which are usually treated as background because they are static or move slowly. The system can also select pedestrian and non-pedestrian samples automatically and update the classifier in real-time to make it more suitable for certain specific scene. Experimental results on a practical public dataset named CASIA Pedestrian Counting Dataset show that the proposed people counting system is robust and accurate.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115646972","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
期刊
The First Asian Conference on Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1