首页 > 最新文献

2010 20th International Conference on Pattern Recognition最新文献

英文 中文
Bounding-Box Based Segmentation with Single Min-cut Using Distant Pixel Similarity 基于边界盒的单最小分割方法
Pub Date : 2010-10-07 DOI: 10.1109/ICPR.2010.1074
V. Pham, Keita Takahashi, T. Naemura
This paper addresses the problem of interactive image segmentation with a user-supplied object bounding box. The underlying problem is the classification of pixels into foreground and background, where only background information is provided with sample pixels. Many approaches treat appearance models as an unknown variable and optimize the segmentation and appearance alternatively, in an expectation maximization manner. In this paper, we describe a novel approach to this problem: the objective function is expressed purely in terms of the unknown segmentation and can be optimized using only one minimum cut calculation. We aim to optimize the trade-off of making the foreground layer as large as possible while keeping the similarity between the foreground and background layers as small as possible. This similarity is formulated using the similarities of distant pixel pairs. We evaluated our algorithm on the GrabCut dataset and demonstrated that high-quality segmentations were attained at a fast calculation speed.
本文解决了使用用户提供的对象边界框进行交互式图像分割的问题。潜在的问题是将像素分类为前景和背景,其中只有背景信息提供了样本像素。许多方法将外观模型作为一个未知变量,并以期望最大化的方式交替优化分割和外观。在本文中,我们描述了一种新的方法来解决这个问题:目标函数纯粹用未知分割来表示,并且只需要一次最小切割计算就可以优化。我们的目标是优化权衡,使前景层尽可能大,同时保持前景层和背景层之间的相似性尽可能小。这种相似性是使用远像素对的相似性来表示的。我们在GrabCut数据集上评估了我们的算法,并证明了在快速的计算速度下获得了高质量的分割。
{"title":"Bounding-Box Based Segmentation with Single Min-cut Using Distant Pixel Similarity","authors":"V. Pham, Keita Takahashi, T. Naemura","doi":"10.1109/ICPR.2010.1074","DOIUrl":"https://doi.org/10.1109/ICPR.2010.1074","url":null,"abstract":"This paper addresses the problem of interactive image segmentation with a user-supplied object bounding box. The underlying problem is the classification of pixels into foreground and background, where only background information is provided with sample pixels. Many approaches treat appearance models as an unknown variable and optimize the segmentation and appearance alternatively, in an expectation maximization manner. In this paper, we describe a novel approach to this problem: the objective function is expressed purely in terms of the unknown segmentation and can be optimized using only one minimum cut calculation. We aim to optimize the trade-off of making the foreground layer as large as possible while keeping the similarity between the foreground and background layers as small as possible. This similarity is formulated using the similarities of distant pixel pairs. We evaluated our algorithm on the GrabCut dataset and demonstrated that high-quality segmentations were attained at a fast calculation speed.","PeriodicalId":309591,"journal":{"name":"2010 20th International Conference on Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2010-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129501070","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Underwater Mine Classification with Imperfect Labels 标签不完善的水下矿井分类
Pub Date : 2010-10-07 DOI: 10.1109/ICPR.2010.1011
David P. Williams
A new algorithm for performing classification with imperfectly labeled data is presented. The proposed approach is motivated by the insight that the average prediction of a group of sufficiently informed people is often more accurate than the prediction of any one supposed expert. This idea that the "wisdom of crowds" can outperform a single expert is implemented by drawing sets of labels as samples from a Bernoulli distribution with a specified labeling error rate. Additionally, ideas from multiple imputation are exploited to provide a principled way for determining an appropriate number of label sampling rounds to consider. The approach is demonstrated in the context of an underwater mine classification application on real synthetic aperture sonar data collected at sea, with promising results.
提出了一种对不完全标记数据进行分类的新算法。提出这种方法的动机是一种洞察力,即一群充分知情的人的平均预测往往比任何一个所谓的专家的预测更准确。这种“群体智慧”可以胜过单个专家的想法是通过从具有指定标签错误率的伯努利分布中绘制标签集作为样本来实现的。此外,利用多重输入的思想,为确定要考虑的适当标签采样轮数提供了一种原则性的方法。最后,将该方法应用于海上真实合成孔径声呐数据进行水雷分类,取得了良好的效果。
{"title":"Underwater Mine Classification with Imperfect Labels","authors":"David P. Williams","doi":"10.1109/ICPR.2010.1011","DOIUrl":"https://doi.org/10.1109/ICPR.2010.1011","url":null,"abstract":"A new algorithm for performing classification with imperfectly labeled data is presented. The proposed approach is motivated by the insight that the average prediction of a group of sufficiently informed people is often more accurate than the prediction of any one supposed expert. This idea that the \"wisdom of crowds\" can outperform a single expert is implemented by drawing sets of labels as samples from a Bernoulli distribution with a specified labeling error rate. Additionally, ideas from multiple imputation are exploited to provide a principled way for determining an appropriate number of label sampling rounds to consider. The approach is demonstrated in the context of an underwater mine classification application on real synthetic aperture sonar data collected at sea, with promising results.","PeriodicalId":309591,"journal":{"name":"2010 20th International Conference on Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2010-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115016755","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Stereo-Based Multi-person Tracking Using Overlapping Silhouette Templates 基于立体的多人跟踪使用重叠剪影模板
Pub Date : 2010-10-07 DOI: 10.1109/ICPR.2010.1046
Junji Satake, J. Miura
This paper describes a stereo-based person tracking method for a person following robot. Many previous works on person tracking use laser range finders which can provide very accurate range measurements. Stereo-based systems have also been popular, but most of them are not used for controlling a real robot. We previously developed a tracking method which uses depth templates of person shape applied to a dense depth image. The method, however, sometimes failed when complex occlusions occurred. In this paper, we propose an accurate, stable tracking method using overlapping silhouette templates which consider how persons overlap in the image. Experimental results show the effectiveness of the proposed method.
针对人跟踪机器人,提出了一种基于立体的人跟踪方法。许多以前的工作人员跟踪使用激光测距仪,可以提供非常精确的距离测量。基于立体声的系统也很受欢迎,但其中大多数并不用于控制真正的机器人。我们之前开发了一种跟踪方法,该方法将人物形状的深度模板应用于密集深度图像。然而,当发生复杂的闭塞时,该方法有时会失败。在本文中,我们提出了一种精确的、稳定的跟踪方法,使用重叠轮廓模板,考虑人物在图像中的重叠情况。实验结果表明了该方法的有效性。
{"title":"Stereo-Based Multi-person Tracking Using Overlapping Silhouette Templates","authors":"Junji Satake, J. Miura","doi":"10.1109/ICPR.2010.1046","DOIUrl":"https://doi.org/10.1109/ICPR.2010.1046","url":null,"abstract":"This paper describes a stereo-based person tracking method for a person following robot. Many previous works on person tracking use laser range finders which can provide very accurate range measurements. Stereo-based systems have also been popular, but most of them are not used for controlling a real robot. We previously developed a tracking method which uses depth templates of person shape applied to a dense depth image. The method, however, sometimes failed when complex occlusions occurred. In this paper, we propose an accurate, stable tracking method using overlapping silhouette templates which consider how persons overlap in the image. Experimental results show the effectiveness of the proposed method.","PeriodicalId":309591,"journal":{"name":"2010 20th International Conference on Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2010-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123653598","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Detection of Shapes in 2D Point Clouds Generated from Images 由图像生成的二维点云的形状检测
Pub Date : 2010-10-07 DOI: 10.1109/ICPR.2010.647
Jingyong Su, Zhiqiang Zhu, Anuj Srivastava, F. Huffer
We present a novel statistical framework for detecting pre-determined shape classes in 2D cluttered point clouds, which are in turn extracted from images. In this model based approach, we use a 1D Poisson process for sampling points on shapes, a 2D Poisson process for points from background clutter, and an additive Gaussian model for noise. Combining these with a past stochastic model on shapes of continuous 2D contours, and optimization over unknown pose and scale, we develop a generalized likelihood ratio test for shape detection. We demonstrate the efficiency of this method and its robustness to clutter using both simulated and real data.
我们提出了一种新的统计框架,用于检测从图像中提取的二维杂乱点云中预先确定的形状类。在这种基于模型的方法中,我们使用一维泊松过程对形状上的点进行采样,使用二维泊松过程对背景杂波中的点进行采样,并使用加性高斯模型对噪声进行采样。结合过去连续二维轮廓形状的随机模型,以及未知姿态和比例的优化,我们开发了一种用于形状检测的广义似然比检验。通过仿真和实际数据验证了该方法的有效性和对杂波的鲁棒性。
{"title":"Detection of Shapes in 2D Point Clouds Generated from Images","authors":"Jingyong Su, Zhiqiang Zhu, Anuj Srivastava, F. Huffer","doi":"10.1109/ICPR.2010.647","DOIUrl":"https://doi.org/10.1109/ICPR.2010.647","url":null,"abstract":"We present a novel statistical framework for detecting pre-determined shape classes in 2D cluttered point clouds, which are in turn extracted from images. In this model based approach, we use a 1D Poisson process for sampling points on shapes, a 2D Poisson process for points from background clutter, and an additive Gaussian model for noise. Combining these with a past stochastic model on shapes of continuous 2D contours, and optimization over unknown pose and scale, we develop a generalized likelihood ratio test for shape detection. We demonstrate the efficiency of this method and its robustness to clutter using both simulated and real data.","PeriodicalId":309591,"journal":{"name":"2010 20th International Conference on Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2010-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129129580","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Colour Constant Image Sharpening 彩色恒定图像锐化
Pub Date : 2010-10-07 DOI: 10.1109/ICPR.2010.1104
A. Alsam
In this paper, we introduce a new sharpening method which guarantees colour constancy and resolves the problem of equiluminance colours. The algorithm is similar to unsharp masking in that the gradients are calculated at different scales by blurring the original with a variable size kernel. The main difference is in the blurring stage where we calculate the average of an n times n neighborhood by projecting each colour vector onto the space of the center pixel before averaging. Thus starting with the center pixel we define a projection matrix onto the space of that vector. Each neighboring colour is then projected onto the center and the result is summed up. The projection step results in an average vector which shares the direction of the original center pixel. The difference between the center pixel and the average is by definition a vector which is scalar away from the center pixel. Thus adding the average to the center pixel is guaranteed not to result in colour shifts. This projection step is also shown to remedy the problem of equiluminance colours and can be used for $m$-dimensional data. Finally, the results indicate that the new sharpening method results in better sharpening than that achieved using unsharp masking with noticeably less halos around strong edges. The latter aspect of the algorithm is believed to be due to the asymmetric nature of the projection step.
本文介绍了一种新的锐化方法,既保证了图像的色彩稳定性,又解决了图像的亮度问题。该算法类似于非锐化掩蔽,通过变大小核模糊原始图像,在不同尺度上计算梯度。主要的区别是在模糊阶段,我们通过在平均之前将每个颜色向量投影到中心像素的空间来计算n乘以n邻域的平均值。因此,从中心像素开始,我们在该向量的空间上定义一个投影矩阵。然后将每个相邻的颜色投影到中心,并将结果汇总。投影步骤产生一个平均向量,该向量共享原始中心像素的方向。根据定义,中心像素和平均值之间的差是距离中心像素的标量向量。因此,将平均值添加到中心像素保证不会导致颜色偏移。这个投影步骤也被证明可以纠正亮度颜色的问题,并且可以用于$m$维数据。最后,结果表明,新的锐化方法比使用非锐化遮罩的锐化效果更好,在强边缘周围的光晕明显减少。该算法的后一个方面被认为是由于投影步骤的不对称性质。
{"title":"Colour Constant Image Sharpening","authors":"A. Alsam","doi":"10.1109/ICPR.2010.1104","DOIUrl":"https://doi.org/10.1109/ICPR.2010.1104","url":null,"abstract":"In this paper, we introduce a new sharpening method which guarantees colour constancy and resolves the problem of equiluminance colours. The algorithm is similar to unsharp masking in that the gradients are calculated at different scales by blurring the original with a variable size kernel. The main difference is in the blurring stage where we calculate the average of an n times n neighborhood by projecting each colour vector onto the space of the center pixel before averaging. Thus starting with the center pixel we define a projection matrix onto the space of that vector. Each neighboring colour is then projected onto the center and the result is summed up. The projection step results in an average vector which shares the direction of the original center pixel. The difference between the center pixel and the average is by definition a vector which is scalar away from the center pixel. Thus adding the average to the center pixel is guaranteed not to result in colour shifts. This projection step is also shown to remedy the problem of equiluminance colours and can be used for $m$-dimensional data. Finally, the results indicate that the new sharpening method results in better sharpening than that achieved using unsharp masking with noticeably less halos around strong edges. The latter aspect of the algorithm is believed to be due to the asymmetric nature of the projection step.","PeriodicalId":309591,"journal":{"name":"2010 20th International Conference on Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2010-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125944653","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Implicit Feature-Based Alignment System for Radiotherapy 基于隐式特征的放射治疗对准系统
Pub Date : 2010-10-07 DOI: 10.1109/ICPR.2010.559
Ryoichi Yamakoshi, Kousuke Hirasawa, H. Okuda, H. Kage, K. Sumi, H. Sakamoto, Yuri Ivanov, Toshihiro Yanou, D. Suga, Masao Murakami
In this paper we present a robust alignment algorithm for correcting the effects of out-of-plane rotation to be used for automatic alignment of the Computed Tomography (CT) volumes and the generally low quality fluoroscopic images for radiotherapy applications. Analyzing not only in-plane but also out-of-plane rotation effects on the Dignitary Reconstructed Radiograph (DRR) images, we develop simple alignment algorithm that extracts a set of implicit features from DRR. Using these SIFT-based features, we align DRRs with the fluoroscopic images of the patient and evaluate the alignment accuracy. We compare our approach with traditional techniques based on gradient-based operators and show that our algorithm performs faster while in most cases delivering higher accuracy.
在本文中,我们提出了一种鲁棒对准算法,用于校正面外旋转的影响,用于计算机断层扫描(CT)体积和放射治疗应用中一般低质量的透视图像的自动对准。分析了显像重构x线图像的面内和面外旋转效应,提出了一种简单的对齐算法,从DRR图像中提取一组隐式特征。利用这些基于sift的特征,我们将DRRs与患者的透视图像对齐并评估对齐精度。我们将我们的方法与基于梯度算子的传统技术进行了比较,结果表明我们的算法执行速度更快,在大多数情况下提供更高的精度。
{"title":"Implicit Feature-Based Alignment System for Radiotherapy","authors":"Ryoichi Yamakoshi, Kousuke Hirasawa, H. Okuda, H. Kage, K. Sumi, H. Sakamoto, Yuri Ivanov, Toshihiro Yanou, D. Suga, Masao Murakami","doi":"10.1109/ICPR.2010.559","DOIUrl":"https://doi.org/10.1109/ICPR.2010.559","url":null,"abstract":"In this paper we present a robust alignment algorithm for correcting the effects of out-of-plane rotation to be used for automatic alignment of the Computed Tomography (CT) volumes and the generally low quality fluoroscopic images for radiotherapy applications. Analyzing not only in-plane but also out-of-plane rotation effects on the Dignitary Reconstructed Radiograph (DRR) images, we develop simple alignment algorithm that extracts a set of implicit features from DRR. Using these SIFT-based features, we align DRRs with the fluoroscopic images of the patient and evaluate the alignment accuracy. We compare our approach with traditional techniques based on gradient-based operators and show that our algorithm performs faster while in most cases delivering higher accuracy.","PeriodicalId":309591,"journal":{"name":"2010 20th International Conference on Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2010-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130394398","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Kernel Domain Description with Incomplete Data: Using Instance-Specific Margins to Avoid Imputation 不完整数据下的核域描述:使用实例特定边界避免插值
Pub Date : 2010-10-07 DOI: 10.1109/ICPR.2010.716
Adam Gripton, W. Lu
We present a method of performing kernel space domain description of a dataset with incomplete entries without the need for imputation, allowing kernel features of a class of data with missing features to be rigorously described. This addresses the problem that absent data completion is usually required before kernel classifiers, such as support vector domain description (SVDD), can be applied; equally, few existing techniques for incomplete data adequately address the issue of kernel spaces. Our method, which we call instance-specific domain description (ISDD), uses a parametrisation framework to compute minimal kernelised distances between data points with missing features through a series of optimisation runs, allowing evaluation of the kernel distance while avoiding subjective completions of missing data. We compare results of our method against those achieved by SVDD applied to an imputed dataset, using synthetic and experimental datasets where feature absence has a non-trivial structure. We show that our methods can achieve tighter sphere bounds when applied to linear and quadratic kernels.
我们提出了一种对具有不完整条目的数据集执行核空间域描述而不需要插入的方法,允许对一类具有缺失特征的数据的核特征进行严格描述。这解决了在应用支持向量域描述(SVDD)等内核分类器之前通常需要缺失数据补全的问题;同样,针对不完整数据的现有技术很少能充分解决内核空间的问题。我们的方法,我们称之为实例特定域描述(ISDD),使用参数化框架通过一系列优化运行来计算具有缺失特征的数据点之间的最小核化距离,允许评估核距离,同时避免主观完成缺失数据。我们将我们的方法的结果与应用于输入数据集的SVDD获得的结果进行比较,使用合成和实验数据集,其中特征缺失具有非平凡结构。我们证明,当应用于线性和二次核时,我们的方法可以获得更紧的球界。
{"title":"Kernel Domain Description with Incomplete Data: Using Instance-Specific Margins to Avoid Imputation","authors":"Adam Gripton, W. Lu","doi":"10.1109/ICPR.2010.716","DOIUrl":"https://doi.org/10.1109/ICPR.2010.716","url":null,"abstract":"We present a method of performing kernel space domain description of a dataset with incomplete entries without the need for imputation, allowing kernel features of a class of data with missing features to be rigorously described. This addresses the problem that absent data completion is usually required before kernel classifiers, such as support vector domain description (SVDD), can be applied; equally, few existing techniques for incomplete data adequately address the issue of kernel spaces. Our method, which we call instance-specific domain description (ISDD), uses a parametrisation framework to compute minimal kernelised distances between data points with missing features through a series of optimisation runs, allowing evaluation of the kernel distance while avoiding subjective completions of missing data. We compare results of our method against those achieved by SVDD applied to an imputed dataset, using synthetic and experimental datasets where feature absence has a non-trivial structure. We show that our methods can achieve tighter sphere bounds when applied to linear and quadratic kernels.","PeriodicalId":309591,"journal":{"name":"2010 20th International Conference on Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2010-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132186748","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Scribe Identification in Medieval English Manuscripts 中世纪英语手稿中的抄写员鉴定
Pub Date : 2010-10-07 DOI: 10.1109/ICPR.2010.463
Tara Gilliam, Richard C. Wilson, J. A. Clark
In this paper we present work on automated scribe identification on a new Middle-English manuscript dataset from around the 14th -- 15th century. We discuss the image and textual problems encountered in processing historical documents, and demonstrate the effect of accounting for manuscript style on the writer identification rate. The grapheme codebook method is used to achieve a Top-1 classification accuracy of up to 77% with a modification to the distance measure. The performance of the Sparse Multinomial Logistic Regression classifier is compared against five k-nn classifiers. We also consider classification against the principal components and propose a method for visualising the principal component vectors in terms of the original grapheme features.
在本文中,我们介绍了在大约14 - 15世纪的一个新的中古英语手稿数据集上的自动抄写员识别工作。我们讨论了历史文献处理中遇到的图像和文本问题,并论证了手稿风格核算对作者识别率的影响。通过对距离度量的修改,使用字形码本方法实现了高达77%的Top-1分类准确率。将稀疏多项式逻辑回归分类器的性能与五种k-nn分类器进行了比较。我们还考虑了针对主成分的分类,并提出了一种根据原始字素特征可视化主成分向量的方法。
{"title":"Scribe Identification in Medieval English Manuscripts","authors":"Tara Gilliam, Richard C. Wilson, J. A. Clark","doi":"10.1109/ICPR.2010.463","DOIUrl":"https://doi.org/10.1109/ICPR.2010.463","url":null,"abstract":"In this paper we present work on automated scribe identification on a new Middle-English manuscript dataset from around the 14th -- 15th century. We discuss the image and textual problems encountered in processing historical documents, and demonstrate the effect of accounting for manuscript style on the writer identification rate. The grapheme codebook method is used to achieve a Top-1 classification accuracy of up to 77% with a modification to the distance measure. The performance of the Sparse Multinomial Logistic Regression classifier is compared against five k-nn classifiers. We also consider classification against the principal components and propose a method for visualising the principal component vectors in terms of the original grapheme features.","PeriodicalId":309591,"journal":{"name":"2010 20th International Conference on Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2010-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115163694","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Human Body Parts Tracking Using Sequential Markov Random Fields 基于顺序马尔可夫随机场的人体部位跟踪
Pub Date : 2010-10-07 DOI: 10.1109/ICPR.2010.1158
Xiao-Qin Cao, Jia Zeng, Zhi-Qiang Liu
Automatically tracking human body parts is a difficult problem because of background clutters, missing body parts, and the high degrees of freedoms and complex kinematics of the articulated human body. This paper presents the sequential Markov random fields (SMRFs) for tracking and labeling moving human body parts automatically by learning the spatio-temporal structures of human motions in the setting of occlusions and clutters. We employ a hybrid strategy, where the temporal dependencies between two successive human poses are described by the sequential Monte Carlo method, and the spatial relationships between body parts in a pose is described by the Markov random fields. Efficient inference and learning algorithms are developed based on the relaxation labeling. Experimental results show that the SMRF can effectively track human body parts in natural scenes.
由于背景杂乱、缺少人体部件以及关节人体的高自由度和复杂的运动特性,人体部位的自动跟踪是一个难题。本文提出了一种序列马尔可夫随机场(SMRFs),通过学习人体运动在遮挡和杂乱环境下的时空结构,实现对人体运动部位的自动跟踪和标记。我们采用了一种混合策略,其中两个连续人体姿势之间的时间依赖关系由顺序蒙特卡罗方法描述,而一个姿势中身体部位之间的空间关系由马尔可夫随机场描述。基于松弛标记开发了高效的推理和学习算法。实验结果表明,该方法可以有效地跟踪自然场景中的人体部位。
{"title":"Human Body Parts Tracking Using Sequential Markov Random Fields","authors":"Xiao-Qin Cao, Jia Zeng, Zhi-Qiang Liu","doi":"10.1109/ICPR.2010.1158","DOIUrl":"https://doi.org/10.1109/ICPR.2010.1158","url":null,"abstract":"Automatically tracking human body parts is a difficult problem because of background clutters, missing body parts, and the high degrees of freedoms and complex kinematics of the articulated human body. This paper presents the sequential Markov random fields (SMRFs) for tracking and labeling moving human body parts automatically by learning the spatio-temporal structures of human motions in the setting of occlusions and clutters. We employ a hybrid strategy, where the temporal dependencies between two successive human poses are described by the sequential Monte Carlo method, and the spatial relationships between body parts in a pose is described by the Markov random fields. Efficient inference and learning algorithms are developed based on the relaxation labeling. Experimental results show that the SMRF can effectively track human body parts in natural scenes.","PeriodicalId":309591,"journal":{"name":"2010 20th International Conference on Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2010-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117173316","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Word Clustering Using PLSA Enhanced with Long Distance Bigrams 基于PLSA的长距离双元词聚类
Pub Date : 2010-10-07 DOI: 10.1109/ICPR.2010.1027
Bassiou Nikoletta, Kotropoulos Constantine
Probabilistic latent semantic analysis is enhanced with long distance bigram models in order to improve word clustering. The long distance bigram probabilities and the interpolated long distance bigram probabilities at varying distances within a context capture different aspects of contextual information. In addition, the baseline bigram, which incorporates trigger-pairs for various histories, is tested in the same framework. The experimental results collected on publicly available corpora (CISI, Cran field, Medline, and NPL) demonstrate the superiority of the long distance bigrams over the baseline bigrams as well as the superiority of the interpolated long distance bigrams against the long distance bigrams and the baseline bigram with trigger-pairs in yielding more compact clusters containing less outliers.
为了提高聚类能力,本文利用长距离双元图模型增强了概率潜在语义分析。长距离重图概率和在上下文中不同距离上插值的长距离重图概率捕获上下文信息的不同方面。此外,在同一框架中测试了包含各种历史记录的触发器对的基线双元图。在公开可用的语料库(CISI, Cran field, Medline和NPL)上收集的实验结果表明,长距离双元图优于基线双元图,而插入的长距离双元图优于长距离双元图和具有触发对的基线双元图,可以产生包含更少异常值的更紧凑的聚类。
{"title":"Word Clustering Using PLSA Enhanced with Long Distance Bigrams","authors":"Bassiou Nikoletta, Kotropoulos Constantine","doi":"10.1109/ICPR.2010.1027","DOIUrl":"https://doi.org/10.1109/ICPR.2010.1027","url":null,"abstract":"Probabilistic latent semantic analysis is enhanced with long distance bigram models in order to improve word clustering. The long distance bigram probabilities and the interpolated long distance bigram probabilities at varying distances within a context capture different aspects of contextual information. In addition, the baseline bigram, which incorporates trigger-pairs for various histories, is tested in the same framework. The experimental results collected on publicly available corpora (CISI, Cran field, Medline, and NPL) demonstrate the superiority of the long distance bigrams over the baseline bigrams as well as the superiority of the interpolated long distance bigrams against the long distance bigrams and the baseline bigram with trigger-pairs in yielding more compact clusters containing less outliers.","PeriodicalId":309591,"journal":{"name":"2010 20th International Conference on Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2010-10-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121670661","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
期刊
2010 20th International Conference on Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1