首页 > 最新文献

The First Asian Conference on Pattern Recognition最新文献

英文 中文
Estimation of street crossing intention from a pedestrian's posture on a sidewalk using multiple image frames 利用多帧图像估计行人在人行道上的姿势过马路的意图
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166694
Ryusuke Furuhashi, K. Yamada
Pedestrian protection is an active area in the research field of advanced driver assistance systems. A pedestrian who intends to cross the road is more critical to the driver than one who has no intention. This paper proposes a method for estimating the crossing intention of a pedestrian on a sidewalk from the pedestrian's posture and change in posture in multiple frames during a short period of video images. We evaluate the method in an indoor simulated environment and a real, outdoor environment and demonstrate the results and performance.
行人保护是先进驾驶辅助系统研究的一个活跃领域。一个想过马路的行人对司机来说比一个没有过马路意图的行人更重要。本文提出了一种基于短时间视频图像中多帧行人的姿势和姿势变化来估计人行道上行人过马路意图的方法。我们在室内模拟环境和真实的室外环境中对该方法进行了评估,并展示了结果和性能。
{"title":"Estimation of street crossing intention from a pedestrian's posture on a sidewalk using multiple image frames","authors":"Ryusuke Furuhashi, K. Yamada","doi":"10.1109/ACPR.2011.6166694","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166694","url":null,"abstract":"Pedestrian protection is an active area in the research field of advanced driver assistance systems. A pedestrian who intends to cross the road is more critical to the driver than one who has no intention. This paper proposes a method for estimating the crossing intention of a pedestrian on a sidewalk from the pedestrian's posture and change in posture in multiple frames during a short period of video images. We evaluate the method in an indoor simulated environment and a real, outdoor environment and demonstrate the results and performance.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128578456","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Feature extraction using class-oriented regression embedding 基于类回归嵌入的特征提取
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166639
Yi Chen, Zhong Jin
Based on linear regression techniques, we present a new supervised learning algorithm called Class-oriented Regression Embedding (CRE) for feature extraction. By minimizing the intra-class reconstruction error, CRE finds a low-dimensional subspace in which samples can be best represented as a combination of their intra-class samples. This characteristic can significantly strengthen the performance of the newly proposed classifier called linear regression-based classification (LRC). The experimental results on the extended-YALE Face Database B (YaleB) and CENPARMI handwritten numeral database show the effectiveness and robustness of CRE plus LRC.
基于线性回归技术,提出了一种新的监督学习算法——面向类的回归嵌入(CRE)。通过最小化类内重构误差,CRE找到了一个低维子空间,在这个子空间中,样本可以最好地表示为它们的类内样本的组合。这个特征可以显著增强新提出的基于线性回归的分类器(LRC)的性能。在扩展的耶鲁人脸数据库B (YaleB)和CENPARMI手写数字数据库上的实验结果表明,CRE + LRC的有效性和鲁棒性。
{"title":"Feature extraction using class-oriented regression embedding","authors":"Yi Chen, Zhong Jin","doi":"10.1109/ACPR.2011.6166639","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166639","url":null,"abstract":"Based on linear regression techniques, we present a new supervised learning algorithm called Class-oriented Regression Embedding (CRE) for feature extraction. By minimizing the intra-class reconstruction error, CRE finds a low-dimensional subspace in which samples can be best represented as a combination of their intra-class samples. This characteristic can significantly strengthen the performance of the newly proposed classifier called linear regression-based classification (LRC). The experimental results on the extended-YALE Face Database B (YaleB) and CENPARMI handwritten numeral database show the effectiveness and robustness of CRE plus LRC.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134427104","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A volumetric spin-off EGI for registration of volume datasets 用于注册体积数据集的体积衍生EGI
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166603
Chun Dong, Timothy S Newman
A spin-off of the Extended Gaussian Image [1] (EGI) registration technique to volumetric datasets is presented. This spin-off technique directly allows recovery of the rotation (and indirectly may allow recovery of the translation) transformations that aligns one volumetric dataset to another. An extension of the basic EGI's orientation histogram to volumetric datasets is also described. Using this histogram, a volume gradient orientation histogram, enables the registration (i.e., aligning) of two instances of one subject. The spin-off technique can be useful for fully automated registration without extraction of higher level features or markers. Results on multiple types of datasets are also reported.
提出了一种扩展高斯图像[1](EGI)配准技术在体积数据集上的应用。这种派生技术直接允许恢复旋转(并间接允许恢复平移)转换,使一个体积数据集与另一个体积数据集对齐。将基本EGI的方向直方图扩展到体积数据集也进行了描述。使用这个直方图,即体积梯度方向直方图,可以对一个主题的两个实例进行配准(即对齐)。衍生技术可以用于完全自动化注册,而无需提取更高级别的特征或标记。对多种类型数据集的结果也进行了报道。
{"title":"A volumetric spin-off EGI for registration of volume datasets","authors":"Chun Dong, Timothy S Newman","doi":"10.1109/ACPR.2011.6166603","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166603","url":null,"abstract":"A spin-off of the Extended Gaussian Image [1] (EGI) registration technique to volumetric datasets is presented. This spin-off technique directly allows recovery of the rotation (and indirectly may allow recovery of the translation) transformations that aligns one volumetric dataset to another. An extension of the basic EGI's orientation histogram to volumetric datasets is also described. Using this histogram, a volume gradient orientation histogram, enables the registration (i.e., aligning) of two instances of one subject. The spin-off technique can be useful for fully automated registration without extraction of higher level features or markers. Results on multiple types of datasets are also reported.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132774749","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Streaming audio classification in Smart Home environments 智能家居环境中的流媒体音频分类
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166676
W. Liao, Jingye Wen, J. Kuo
In this research, we develop and integrate methods for real-time streaming audio classification based on psychoacoustic models of hearing as well as techniques in pattern recognition. Specifically, a framework for auditory event detection and signal description by means of computer vision approach has been designed to enable real-time processing and classification of audio signals present in home environments. Local binary patterns are employed to describe the extracted sound blobs in the spectrogram. Experimental results show that the proposed approach is quite effective, achieving an overall recognition rate of 80–90% for 8 types of audio input. The performance degrades only slightly in the presence of noise and other interferences.
在本研究中,我们开发并整合了基于听觉心理声学模型和模式识别技术的实时流音频分类方法。具体而言,通过计算机视觉方法设计了听觉事件检测和信号描述框架,以实现对家庭环境中存在的音频信号的实时处理和分类。用局部二值模式描述声谱图中提取的声团。实验结果表明,该方法是非常有效的,对8种音频输入的总体识别率达到80-90%。在存在噪声和其他干扰的情况下,性能只会略有下降。
{"title":"Streaming audio classification in Smart Home environments","authors":"W. Liao, Jingye Wen, J. Kuo","doi":"10.1109/ACPR.2011.6166676","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166676","url":null,"abstract":"In this research, we develop and integrate methods for real-time streaming audio classification based on psychoacoustic models of hearing as well as techniques in pattern recognition. Specifically, a framework for auditory event detection and signal description by means of computer vision approach has been designed to enable real-time processing and classification of audio signals present in home environments. Local binary patterns are employed to describe the extracted sound blobs in the spectrogram. Experimental results show that the proposed approach is quite effective, achieving an overall recognition rate of 80–90% for 8 types of audio input. The performance degrades only slightly in the presence of noise and other interferences.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132946708","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
View-invariant action recognition in surveillance videos 监控视频中的视不变动作识别
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166671
Fang Zhang, Yunhong Wang, Zhaoxiang Zhang
Recently, human action recognition has been a popular and important topic in computer vision. However, except some conventional problems such as noise, low resolution etc., view-invariant recognition is one of the most challenging problems. In this paper, we focus on solve multi-view action recognition from surveillance video. To detect moving objects from complicated backgrounds, this paper employs improved Gaussian mixed model, which uses K-means clustering to initialize the model and it gets better motion detection results for surveillance videos. We demonstrate the silhouette representation “Envelope Shape” can solve the viewpoint problem in surveillance videos. The experiment results demonstrate that our human action recognition system is fast and efficient on CASIA activity analysis database.
近年来,人体动作识别已成为计算机视觉领域的一个热点和重要课题。然而,除了一些传统的问题,如噪声、低分辨率等,视点不变识别是最具挑战性的问题之一。本文主要解决监控视频中的多视点动作识别问题。为了检测复杂背景下的运动物体,本文采用改进的高斯混合模型,该模型使用K-means聚类对模型进行初始化,得到了较好的监控视频运动检测结果。我们证明了轮廓表示“包络形状”可以解决监控视频中的视点问题。实验结果表明,我们的人体动作识别系统在CASIA活动分析数据库上是快速有效的。
{"title":"View-invariant action recognition in surveillance videos","authors":"Fang Zhang, Yunhong Wang, Zhaoxiang Zhang","doi":"10.1109/ACPR.2011.6166671","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166671","url":null,"abstract":"Recently, human action recognition has been a popular and important topic in computer vision. However, except some conventional problems such as noise, low resolution etc., view-invariant recognition is one of the most challenging problems. In this paper, we focus on solve multi-view action recognition from surveillance video. To detect moving objects from complicated backgrounds, this paper employs improved Gaussian mixed model, which uses K-means clustering to initialize the model and it gets better motion detection results for surveillance videos. We demonstrate the silhouette representation “Envelope Shape” can solve the viewpoint problem in surveillance videos. The experiment results demonstrate that our human action recognition system is fast and efficient on CASIA activity analysis database.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129372814","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Binary mask estimation for voiced speech segregation using Bayesian method 基于贝叶斯方法的语音分离二值掩码估计
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6305053
Shan Liang, Wenju Liu
The ideal binary mask (IBM) estimation has been set as the computational goal of Computational auditory scene analysis (CASA). A lot of effort has been made in the IBM estimation via statistical learning method. The current Bayesian methods usually estimate the mask value of each time-frequency (T-F) unit independently with only local auditory features. In this paper, we propose a new Bayesian approach. First, a set of pitch-based auditory features are summarized to exploit the inherent characteristics of the reliable and unreliable time-frequency (T-F) units. A rough estimation is obtained according to Maximum Likelihood (ML) rule. Then, we propose a prior model which is derived from onset/offset segmentation to improve the estimation. Finally, an efficient Markov Chain Monte Carlo (MCMC) procedure is applied to approach the maximum a posterior (MAP) estimation. Proposed method is evaluated on Cooke's 100 mixtures and compared with previous model. Experiments show that our method performs better.
将理想二值掩码估计作为计算听觉场景分析(CASA)的计算目标。IBM通过统计学习方法进行了大量的研究。目前的贝叶斯方法通常只使用局部听觉特征独立估计每个时频单元的掩模值。本文提出了一种新的贝叶斯方法。首先,总结了一组基于音高的听觉特征,以挖掘可靠和不可靠时频(T-F)单元的固有特征。根据最大似然(ML)规则得到一个粗略的估计。然后,我们提出了一种基于起始/偏移分割的先验模型来改进估计。最后,应用一种有效的马尔可夫链蒙特卡罗(MCMC)方法逼近最大后验估计(MAP)。在Cooke's 100混合物上对该方法进行了评价,并与已有模型进行了比较。实验表明,该方法具有较好的性能。
{"title":"Binary mask estimation for voiced speech segregation using Bayesian method","authors":"Shan Liang, Wenju Liu","doi":"10.1109/ACPR.2011.6305053","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6305053","url":null,"abstract":"The ideal binary mask (IBM) estimation has been set as the computational goal of Computational auditory scene analysis (CASA). A lot of effort has been made in the IBM estimation via statistical learning method. The current Bayesian methods usually estimate the mask value of each time-frequency (T-F) unit independently with only local auditory features. In this paper, we propose a new Bayesian approach. First, a set of pitch-based auditory features are summarized to exploit the inherent characteristics of the reliable and unreliable time-frequency (T-F) units. A rough estimation is obtained according to Maximum Likelihood (ML) rule. Then, we propose a prior model which is derived from onset/offset segmentation to improve the estimation. Finally, an efficient Markov Chain Monte Carlo (MCMC) procedure is applied to approach the maximum a posterior (MAP) estimation. Proposed method is evaluated on Cooke's 100 mixtures and compared with previous model. Experiments show that our method performs better.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125865647","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A parameter independent line fitting method 一种参数无关的直线拟合方法
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166585
D. Prasad, Hiok Chai Quek, M. Leung, Siu-Yeung Cho
We prove that when a line is approximated using digital line, the error in the slope of the digital line has a definite upper bound and is strongly dependent on the two pixels chosen for defining the digital line. Thus, an analytical expression of the maximum deviation of the pixels from the digital line can be derived. Using this, the conventional line fitting methods that use maximum tolerable deviation as the optimization goal can be made control-parameter independent. This error bound can be used to make the most recent and sophisticated line fitting methods parameter independent and more robust to digitization noises. In our knowledge, this is the first line fitting method completely devoid of any control parameter. Such control-parameter independent line fitting algorithm retains the characteristics of the digital curve with sufficient reliability and precision and provides good dimensionality reduction in representing the digital curves. Extensive results have been generated for 9 datasets comprising of about a hundred thousand images. The proposed method shows robust and repeatable performance across all the datasets with low standard deviation in the performance.
我们证明了当用数字线近似一条线时,数字线斜率的误差有一个确定的上限,并且强烈依赖于选择用于定义数字线的两个像素。因此,可以推导出像素与数字线的最大偏差的解析表达式。利用这种方法,可以使以最大容许偏差为优化目标的传统线拟合方法与控制参数无关。这种误差界限可以使最新和最复杂的线拟合方法与参数无关,并且对数字化噪声具有更强的鲁棒性。据我们所知,这是第一个完全没有任何控制参数的直线拟合方法。这种与控制参数无关的线拟合算法保留了数字曲线的特征,具有足够的可靠性和精度,并且在表示数字曲线时具有良好的降维性。广泛的结果已经产生了9个数据集,包括大约10万张图像。该方法在所有数据集上都具有鲁棒性和可重复性,且性能标准差低。
{"title":"A parameter independent line fitting method","authors":"D. Prasad, Hiok Chai Quek, M. Leung, Siu-Yeung Cho","doi":"10.1109/ACPR.2011.6166585","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166585","url":null,"abstract":"We prove that when a line is approximated using digital line, the error in the slope of the digital line has a definite upper bound and is strongly dependent on the two pixels chosen for defining the digital line. Thus, an analytical expression of the maximum deviation of the pixels from the digital line can be derived. Using this, the conventional line fitting methods that use maximum tolerable deviation as the optimization goal can be made control-parameter independent. This error bound can be used to make the most recent and sophisticated line fitting methods parameter independent and more robust to digitization noises. In our knowledge, this is the first line fitting method completely devoid of any control parameter. Such control-parameter independent line fitting algorithm retains the characteristics of the digital curve with sufficient reliability and precision and provides good dimensionality reduction in representing the digital curves. Extensive results have been generated for 9 datasets comprising of about a hundred thousand images. The proposed method shows robust and repeatable performance across all the datasets with low standard deviation in the performance.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130045659","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 39
People counting using ellipse detection and forward/backward tracing 人计数使用椭圆检测和向前/向后跟踪
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166629
Chung-Lin Huang, Shih-Chung Hsu, I-Chung Tsao, Ben-Syuan Huang, Hau-Wei Wang, Hung-Wei Lin
There are two different people counting methods: (1) counting people across a detecting line in certain time duration and (2) estimating the total number of people in some region at certain time instance. This paper presents a new approach to count the number of people crossing a line of interest (LOI). First, the foreground object silhouettes are extracted described as blobs. Second, we generate the blob linkage based on the one-to-one or one-to-many correspondence between the blobs in every two consecutive frames. Third, we label the number of objects in each blob by applying the ellipse detection technique. However, the occlusion problem jeopardizes the labeling process. Here, we use forward/backward tracing to re-label the number of objects in the occluded blob. In the experiments, we illustrate the effectiveness of our method.
有两种不同的人数计数方法:(1)在某一时间段内对穿过检测线的人数进行计数;(2)估计某一时间段某一区域的总人数。本文提出了一种计算越过兴趣线人数的新方法。首先,提取前景目标轮廓,描述为blobs。其次,我们基于每两个连续帧中blob之间的一对一或一对多对应关系生成blob链接。第三,利用椭圆检测技术对每个blob中的目标数量进行标记。然而,遮挡问题危及标记过程。在这里,我们使用向前/向后跟踪来重新标记遮挡blob中的对象数量。在实验中,我们证明了该方法的有效性。
{"title":"People counting using ellipse detection and forward/backward tracing","authors":"Chung-Lin Huang, Shih-Chung Hsu, I-Chung Tsao, Ben-Syuan Huang, Hau-Wei Wang, Hung-Wei Lin","doi":"10.1109/ACPR.2011.6166629","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166629","url":null,"abstract":"There are two different people counting methods: (1) counting people across a detecting line in certain time duration and (2) estimating the total number of people in some region at certain time instance. This paper presents a new approach to count the number of people crossing a line of interest (LOI). First, the foreground object silhouettes are extracted described as blobs. Second, we generate the blob linkage based on the one-to-one or one-to-many correspondence between the blobs in every two consecutive frames. Third, we label the number of objects in each blob by applying the ellipse detection technique. However, the occlusion problem jeopardizes the labeling process. Here, we use forward/backward tracing to re-label the number of objects in the occluded blob. In the experiments, we illustrate the effectiveness of our method.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122555131","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Hierarchical video object segmentation 分层视频对象分割
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166705
Junliang Xing, H. Ai, S. Lao
In this paper, we propose a general video object segmentation framework which views object segmentation from a unified Bayesian perspective and optimizes the MAP formulated problem in a progressive manner. Based on object detection and tracking results, a three-level hierarchical video object segmentation approach is presented. At the first level, an offline learned segmentor is applied to each object tracking result of current frame to get a coarse segmentation. At the second level, the coarse segmentation is updated into an intermediate segmentation by a temporal model which propagates the fine segmentation of previous frame to current frame based on a discriminative feature points voting process. At the third level, the intermediate segmentation is refined by an iterative procedure which uses online collected color-and-shape information to get the final result. We apply the approach to pedestrian segmentation on many challenging datasets that demonstrates its effectiveness.
本文提出了一种通用的视频目标分割框架,该框架从统一的贝叶斯角度看待目标分割,并逐步优化MAP公式问题。基于目标检测和跟踪结果,提出了一种三级分层视频目标分割方法。首先,对当前帧的每个目标跟踪结果应用离线学习的分割器进行粗分割;第二层是基于判别特征点投票过程将前一帧的精细分割传播到当前帧的时间模型将粗分割更新为中间分割;在第三层,使用在线收集的颜色和形状信息,通过迭代过程对中间分割进行细化,得到最终结果。我们将该方法应用于许多具有挑战性的数据集上的行人分割,证明了其有效性。
{"title":"Hierarchical video object segmentation","authors":"Junliang Xing, H. Ai, S. Lao","doi":"10.1109/ACPR.2011.6166705","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166705","url":null,"abstract":"In this paper, we propose a general video object segmentation framework which views object segmentation from a unified Bayesian perspective and optimizes the MAP formulated problem in a progressive manner. Based on object detection and tracking results, a three-level hierarchical video object segmentation approach is presented. At the first level, an offline learned segmentor is applied to each object tracking result of current frame to get a coarse segmentation. At the second level, the coarse segmentation is updated into an intermediate segmentation by a temporal model which propagates the fine segmentation of previous frame to current frame based on a discriminative feature points voting process. At the third level, the intermediate segmentation is refined by an iterative procedure which uses online collected color-and-shape information to get the final result. We apply the approach to pedestrian segmentation on many challenging datasets that demonstrates its effectiveness.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"3489 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127510064","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Face recognition using the Weber Local Descriptor 使用韦伯局部描述符的人脸识别
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166675
Dayi Gong, Shutao Li, Yin Xiang
This paper presents a face recognition method using the Weber Local Descriptor (WLD) feature. The WLD consists of differential excitation component and orientation component, which contains abundant local texture information. In our method, we firstly divide face images into a set of sub-regions and extract their WLD features respectively. We introduce the Sobel descriptor to obtain the orientation component. Then each of sub-regions of probe image is recognized by nearest neighborhood method and the results are fused in decision level through voting to yield the final recognition result. The experimental results over ORL and Yale face database verify the effectiveness of our method.
提出了一种基于韦伯局部描述子(WLD)特征的人脸识别方法。WLD由差分激励分量和方向分量组成,其中包含丰富的局部纹理信息。在我们的方法中,我们首先将人脸图像分成一组子区域,并分别提取它们的WLD特征。我们引入Sobel描述符来获得方向分量。然后用最近邻法对探测图像的各个子区域进行识别,并在决策层通过投票将结果融合,得到最终的识别结果。在ORL和耶鲁人脸数据库上的实验结果验证了该方法的有效性。
{"title":"Face recognition using the Weber Local Descriptor","authors":"Dayi Gong, Shutao Li, Yin Xiang","doi":"10.1109/ACPR.2011.6166675","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166675","url":null,"abstract":"This paper presents a face recognition method using the Weber Local Descriptor (WLD) feature. The WLD consists of differential excitation component and orientation component, which contains abundant local texture information. In our method, we firstly divide face images into a set of sub-regions and extract their WLD features respectively. We introduce the Sobel descriptor to obtain the orientation component. Then each of sub-regions of probe image is recognized by nearest neighborhood method and the results are fused in decision level through voting to yield the final recognition result. The experimental results over ORL and Yale face database verify the effectiveness of our method.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125400951","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
期刊
The First Asian Conference on Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1