首页 > 最新文献

2016 23rd International Conference on Pattern Recognition (ICPR)最新文献

英文 中文
A computational approach to relative aesthetics 相对美学的计算方法
Pub Date : 2017-04-05 DOI: 10.1109/ICPR.2016.7900003
Vijetha Gattupalli, P. S. Chandakkar, Baoxin Li
Computational visual aesthetics has recently become an active research area. Existing state-of-art methods formulate this as a binary classification task where a given image is predicted to be beautiful or not. In many applications such as image retrieval and enhancement, it is more important to rank images based on their aesthetic quality instead of binary-categorizing them. Furthermore, in such applications, it may be possible that all images belong to the same category. Hence determining the aesthetic ranking of the images is more appropriate. To this end, we formulate a novel problem of ranking images with respect to their aesthetic quality. We construct a new dataset of image pairs with relative labels by carefully selecting images from the popular AVA dataset. Unlike in aesthetics classification, there is no single threshold which would determine the ranking order of the images across our entire dataset. We propose a deep neural network based approach that is trained on image pairs by incorporating principles from relative learning. Results show that such relative training procedure allows our network to rank the images with a higher accuracy than a state-of-art network trained on the same set of images using binary labels.
计算视觉美学近年来成为一个活跃的研究领域。现有的最先进的方法将其表述为一个二元分类任务,其中给定的图像被预测为美丽或不美丽。在图像检索和增强等许多应用中,基于图像的审美质量对图像进行排序比基于图像的二值分类更为重要。此外,在这种应用中,可能所有图像都属于同一类别。因此确定图像的审美排序是比较合适的。为此,我们提出了一个新的问题,即根据图像的审美质量对其进行排名。我们通过从流行的AVA数据集中仔细选择图像来构建具有相对标签的图像对的新数据集。与美学分类不同,没有单一的阈值来决定整个数据集中图像的排名顺序。我们提出了一种基于深度神经网络的方法,该方法通过结合相对学习原理对图像对进行训练。结果表明,这种相对训练过程使我们的网络能够以比使用二值标签在同一组图像上训练的最先进的网络更高的精度对图像进行排名。
{"title":"A computational approach to relative aesthetics","authors":"Vijetha Gattupalli, P. S. Chandakkar, Baoxin Li","doi":"10.1109/ICPR.2016.7900003","DOIUrl":"https://doi.org/10.1109/ICPR.2016.7900003","url":null,"abstract":"Computational visual aesthetics has recently become an active research area. Existing state-of-art methods formulate this as a binary classification task where a given image is predicted to be beautiful or not. In many applications such as image retrieval and enhancement, it is more important to rank images based on their aesthetic quality instead of binary-categorizing them. Furthermore, in such applications, it may be possible that all images belong to the same category. Hence determining the aesthetic ranking of the images is more appropriate. To this end, we formulate a novel problem of ranking images with respect to their aesthetic quality. We construct a new dataset of image pairs with relative labels by carefully selecting images from the popular AVA dataset. Unlike in aesthetics classification, there is no single threshold which would determine the ranking order of the images across our entire dataset. We propose a deep neural network based approach that is trained on image pairs by incorporating principles from relative learning. Results show that such relative training procedure allows our network to rank the images with a higher accuracy than a state-of-art network trained on the same set of images using binary labels.","PeriodicalId":151180,"journal":{"name":"2016 23rd International Conference on Pattern Recognition (ICPR)","volume":"137 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-04-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131475939","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Clustering for point pattern data 点模式数据的聚类
Pub Date : 2017-02-08 DOI: 10.1109/ICPR.2016.7900123
Nhat-Quang Tran, B. Vo, Dinh Q. Phung, B. Vo
Clustering is one of the most common unsupervised learning tasks in machine learning and data mining. Clustering algorithms have been used in a plethora of applications across several scientific fields. However, there has been limited research in the clustering of point patterns - sets or multi-sets of unordered elements - that are found in numerous applications and data sources. In this paper, we propose two approaches for clustering point patterns. The first is a non-parametric method based on novel distances for sets. The second is a model-based approach, formulated via random finite set theory, and solved by the Expectation-Maximization algorithm. Numerical experiments show that the proposed methods perform well on both simulated and real data.
聚类是机器学习和数据挖掘中最常见的无监督学习任务之一。聚类算法已经在多个科学领域的大量应用中使用。然而,对于存在于众多应用和数据源中的点模式(无序元素的集合或多集合)的聚类研究非常有限。在本文中,我们提出了两种聚类点模式的方法。第一种是基于新距离的非参数方法。第二种是基于模型的方法,通过随机有限集理论制定,并通过期望最大化算法解决。数值实验表明,所提出的方法在模拟数据和实际数据上都有良好的效果。
{"title":"Clustering for point pattern data","authors":"Nhat-Quang Tran, B. Vo, Dinh Q. Phung, B. Vo","doi":"10.1109/ICPR.2016.7900123","DOIUrl":"https://doi.org/10.1109/ICPR.2016.7900123","url":null,"abstract":"Clustering is one of the most common unsupervised learning tasks in machine learning and data mining. Clustering algorithms have been used in a plethora of applications across several scientific fields. However, there has been limited research in the clustering of point patterns - sets or multi-sets of unordered elements - that are found in numerous applications and data sources. In this paper, we propose two approaches for clustering point patterns. The first is a non-parametric method based on novel distances for sets. The second is a model-based approach, formulated via random finite set theory, and solved by the Expectation-Maximization algorithm. Numerical experiments show that the proposed methods perform well on both simulated and real data.","PeriodicalId":151180,"journal":{"name":"2016 23rd International Conference on Pattern Recognition (ICPR)","volume":"64 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-02-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126284493","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Using Convolutional 3D Neural Networks for User-independent continuous gesture recognition 基于卷积三维神经网络的用户独立连续手势识别
Pub Date : 2016-12-04 DOI: 10.1109/ICPR.2016.7899606
Necati Cihan Camgöz, Simon Hadfield, Oscar Koller, R. Bowden
In this paper, we propose using 3D Convolutional Neural Networks for large scale user-independent continuous gesture recognition. We have trained an end-to-end deep network for continuous gesture recognition (jointly learning both the feature representation and the classifier). The network performs three-dimensional (i.e. space-time) convolutions to extract features related to both the appearance and motion from volumes of color frames. Space-time invariance of the extracted features is encoded via pooling layers. The earlier stages of the network are partially initialized using the work of Tran et al. before being adapted to the task of gesture recognition. An earlier version of the proposed method, which was trained for 11,250 iterations, was submitted to ChaLearn 2016 Continuous Gesture Recognition Challenge and ranked 2nd with the Mean Jaccard Index Score of 0.269235. When the proposed method was further trained for 28,750 iterations, it achieved state-of-the-art performance on the same dataset, yielding a 0.314779 Mean Jaccard Index Score.
在本文中,我们提出使用三维卷积神经网络进行大规模的独立于用户的连续手势识别。我们已经训练了一个端到端的深度网络,用于连续的手势识别(共同学习特征表示和分类器)。该网络执行三维(即时空)卷积,从彩色帧的体积中提取与外观和运动相关的特征。通过池化层对提取的特征进行时空不变性编码。在适应手势识别任务之前,网络的早期阶段使用Tran等人的工作进行部分初始化。该方法的早期版本经过了11,250次迭代的训练,提交给了ChaLearn 2016年连续手势识别挑战赛,并以0.269235的平均Jaccard指数得分排名第二。当提出的方法被进一步训练28,750次迭代时,它在相同的数据集上达到了最先进的性能,产生0.314779的平均Jaccard指数得分。
{"title":"Using Convolutional 3D Neural Networks for User-independent continuous gesture recognition","authors":"Necati Cihan Camgöz, Simon Hadfield, Oscar Koller, R. Bowden","doi":"10.1109/ICPR.2016.7899606","DOIUrl":"https://doi.org/10.1109/ICPR.2016.7899606","url":null,"abstract":"In this paper, we propose using 3D Convolutional Neural Networks for large scale user-independent continuous gesture recognition. We have trained an end-to-end deep network for continuous gesture recognition (jointly learning both the feature representation and the classifier). The network performs three-dimensional (i.e. space-time) convolutions to extract features related to both the appearance and motion from volumes of color frames. Space-time invariance of the extracted features is encoded via pooling layers. The earlier stages of the network are partially initialized using the work of Tran et al. before being adapted to the task of gesture recognition. An earlier version of the proposed method, which was trained for 11,250 iterations, was submitted to ChaLearn 2016 Continuous Gesture Recognition Challenge and ranked 2nd with the Mean Jaccard Index Score of 0.269235. When the proposed method was further trained for 28,750 iterations, it achieved state-of-the-art performance on the same dataset, yielding a 0.314779 Mean Jaccard Index Score.","PeriodicalId":151180,"journal":{"name":"2016 23rd International Conference on Pattern Recognition (ICPR)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126612111","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 87
Novel generative model for facial expressions based on statistical shape analysis of landmarks trajectories 基于地标轨迹统计形状分析的面部表情生成模型
Pub Date : 2016-12-04 DOI: 10.1109/ICPR.2016.7899760
P. Desrosiers, M. Daoudi, M. Devanne
We propose a novel geometric framework for analyzing spontaneous facial expressions, with the specific goal of comparing, matching, and averaging the shapes of landmarks trajectories. Here we represent facial expressions by the motion of the landmarks across the time. The trajectories are represented by curves. We use elastic shape analysis of these curves to develop a Riemannian framework for analyzing shapes of these trajectories. In terms of empirical evaluation, our results on two databases: UvA-NEMO and Cohn-Kanade CK+ are very promising. From a theoretical perspective, this framework allows formal statistical inferences, such as generation of facial expressions.
我们提出了一种新的几何框架来分析自发的面部表情,其具体目标是比较、匹配和平均地标轨迹的形状。在这里,我们通过地标在时间上的运动来表示面部表情。轨迹用曲线表示。我们使用这些曲线的弹性形状分析来开发一个黎曼框架来分析这些轨迹的形状。在实证评价方面,我们在UvA-NEMO和Cohn-Kanade CK+两个数据库上的结果非常有前景。从理论的角度来看,这个框架允许正式的统计推断,比如面部表情的生成。
{"title":"Novel generative model for facial expressions based on statistical shape analysis of landmarks trajectories","authors":"P. Desrosiers, M. Daoudi, M. Devanne","doi":"10.1109/ICPR.2016.7899760","DOIUrl":"https://doi.org/10.1109/ICPR.2016.7899760","url":null,"abstract":"We propose a novel geometric framework for analyzing spontaneous facial expressions, with the specific goal of comparing, matching, and averaging the shapes of landmarks trajectories. Here we represent facial expressions by the motion of the landmarks across the time. The trajectories are represented by curves. We use elastic shape analysis of these curves to develop a Riemannian framework for analyzing shapes of these trajectories. In terms of empirical evaluation, our results on two databases: UvA-NEMO and Cohn-Kanade CK+ are very promising. From a theoretical perspective, this framework allows formal statistical inferences, such as generation of facial expressions.","PeriodicalId":151180,"journal":{"name":"2016 23rd International Conference on Pattern Recognition (ICPR)","volume":"49 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129499919","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Exploiting social and mobility patterns for friendship prediction in location-based social networks 利用基于位置的社交网络中的社交和移动模式进行友谊预测
Pub Date : 2016-12-04 DOI: 10.1109/ICPR.2016.7900016
J. Valverde-Rebaza, M. Roche, P. Poncelet, A. Lopes
Link prediction is a “hot topic” in network analysis and has been largely used for friendship recommendation in social networks. With the increased use of location-based services, it is possible to improve the accuracy of link prediction methods by using the mobility of users. The majority of the link prediction methods focus on the importance of location for their visitors, disregarding the strength of relationships existing between these visitors. We, therefore, propose three new methods for friendship prediction by combining, efficiently, social and mobility patterns of users in location-based social networks (LBSNs). Experiments conducted on real-world datasets demonstrate that our proposals achieve a competitive performance with methods from the literature and, in most of the cases, outperform them. Moreover, our proposals use less computational resources by reducing considerably the number of irrelevant predictions, making the link prediction task more efficient and applicable for real world applications.
链接预测是网络分析中的一个“热门话题”,在社交网络中被大量用于好友推荐。随着基于位置的服务使用的增加,利用用户的移动性来提高链接预测方法的准确性是可能的。大多数链接预测方法关注的是位置对访问者的重要性,而忽略了这些访问者之间存在的关系的强度。因此,我们提出了三种新的友谊预测方法,通过有效地结合基于位置的社交网络(LBSNs)中用户的社交和移动模式。在真实世界数据集上进行的实验表明,我们的建议与文献中的方法实现了竞争性能,并且在大多数情况下优于它们。此外,我们的建议通过大大减少不相关预测的数量使用更少的计算资源,使链路预测任务更有效,更适用于现实世界的应用。
{"title":"Exploiting social and mobility patterns for friendship prediction in location-based social networks","authors":"J. Valverde-Rebaza, M. Roche, P. Poncelet, A. Lopes","doi":"10.1109/ICPR.2016.7900016","DOIUrl":"https://doi.org/10.1109/ICPR.2016.7900016","url":null,"abstract":"Link prediction is a “hot topic” in network analysis and has been largely used for friendship recommendation in social networks. With the increased use of location-based services, it is possible to improve the accuracy of link prediction methods by using the mobility of users. The majority of the link prediction methods focus on the importance of location for their visitors, disregarding the strength of relationships existing between these visitors. We, therefore, propose three new methods for friendship prediction by combining, efficiently, social and mobility patterns of users in location-based social networks (LBSNs). Experiments conducted on real-world datasets demonstrate that our proposals achieve a competitive performance with methods from the literature and, in most of the cases, outperform them. Moreover, our proposals use less computational resources by reducing considerably the number of irrelevant predictions, making the link prediction task more efficient and applicable for real world applications.","PeriodicalId":151180,"journal":{"name":"2016 23rd International Conference on Pattern Recognition (ICPR)","volume":"143 8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129484956","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Comparative study of descriptors with dense key points 密集点描述符的比较研究
Pub Date : 2016-12-04 DOI: 10.1109/ICPR.2016.7899928
H. Chatoux, F. Lecellier, C. Fernandez-Maloigne
A great deal of features detectors and descriptors are proposed every years for several computer vision applications. In this paper, we concentrate on dense detector applied to different descriptors. Eight descriptors are compared, three from gradient based family (SIFT, SURF, DAISY), others from binary category (BRIEF, ORB, BRISK, FREAK and LATCH). These descriptors are created and defined with certain invariance properties. We want to verify their invariances with various geometric and photometric transformations, varying one at a time. Deformations are computed from an original image. Descriptors are tested on five transformations: scale, rotation, viewpoint, illumination plus reflection. Overall, descriptors display the right invariances. This paper's objective is to establish a reproducible protocol to test descriptors invariances.
每年都有大量的特征检测器和描述符被提出,用于各种计算机视觉应用。本文主要研究了不同描述符下的密集检测器。比较了八个描述符,三个来自基于梯度的族(SIFT, SURF, DAISY),其他来自二元类(BRIEF, ORB, BRISK, FREAK和LATCH)。这些描述符是用某些不变性属性创建和定义的。我们想通过各种几何和光度变换来验证它们的不变性,每次变化一个。变形是从原始图像计算的。描述符在五种变换下进行测试:比例、旋转、视点、光照和反射。总的来说,描述符显示了正确的不变性。本文的目的是建立一个可重复的协议来测试描述符的不变性。
{"title":"Comparative study of descriptors with dense key points","authors":"H. Chatoux, F. Lecellier, C. Fernandez-Maloigne","doi":"10.1109/ICPR.2016.7899928","DOIUrl":"https://doi.org/10.1109/ICPR.2016.7899928","url":null,"abstract":"A great deal of features detectors and descriptors are proposed every years for several computer vision applications. In this paper, we concentrate on dense detector applied to different descriptors. Eight descriptors are compared, three from gradient based family (SIFT, SURF, DAISY), others from binary category (BRIEF, ORB, BRISK, FREAK and LATCH). These descriptors are created and defined with certain invariance properties. We want to verify their invariances with various geometric and photometric transformations, varying one at a time. Deformations are computed from an original image. Descriptors are tested on five transformations: scale, rotation, viewpoint, illumination plus reflection. Overall, descriptors display the right invariances. This paper's objective is to establish a reproducible protocol to test descriptors invariances.","PeriodicalId":151180,"journal":{"name":"2016 23rd International Conference on Pattern Recognition (ICPR)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121954066","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
Remote photoplethysmography based on implicit living skin tissue segmentation 基于隐式活体皮肤组织分割的远程光容积脉搏波
Pub Date : 2016-12-04 DOI: 10.1109/ICPR.2016.7899660
Serge Bobbia, Y. Benezeth, Julien Dubois
Region of interest selection is an essential part for remote photoplethysmography (rPPG) algorithms. Most of the time, face detection provided by a supervised learning of physical appearance features coupled with skin detection is used for region of interest selection. However, both methods have several limitations and we propose to implicitly select living skin tissue via their particular pulsatility feature. The input video stream is decomposed into several temporal superpixels from which pulse signals are extracted. Pulsatility measure for each temporal superpixel is then used to merge pulse traces and estimate the photoplethysmogram signal. This allows to select skin tissue and furthermore to favor areas where the pulse trace is more predominant. Experimental results showed that our method perform better than state of the art algorithms without any critical face or skin detection.
兴趣区选择是远程光容积脉搏波(rPPG)算法的重要组成部分。大多数情况下,通过对身体外观特征的监督学习和皮肤检测提供的面部检测用于兴趣区域的选择。然而,这两种方法都有一些局限性,我们建议通过其特定的脉动特征隐式选择活体皮肤组织。输入视频流被分解为多个时间超像素,从中提取脉冲信号。然后使用每个时间超像素的脉搏测量来合并脉冲轨迹并估计光容积脉搏图信号。这允许选择皮肤组织,进一步有利于脉冲痕迹更占优势的区域。实验结果表明,在没有任何关键的面部或皮肤检测的情况下,我们的方法比最先进的算法表现得更好。
{"title":"Remote photoplethysmography based on implicit living skin tissue segmentation","authors":"Serge Bobbia, Y. Benezeth, Julien Dubois","doi":"10.1109/ICPR.2016.7899660","DOIUrl":"https://doi.org/10.1109/ICPR.2016.7899660","url":null,"abstract":"Region of interest selection is an essential part for remote photoplethysmography (rPPG) algorithms. Most of the time, face detection provided by a supervised learning of physical appearance features coupled with skin detection is used for region of interest selection. However, both methods have several limitations and we propose to implicitly select living skin tissue via their particular pulsatility feature. The input video stream is decomposed into several temporal superpixels from which pulse signals are extracted. Pulsatility measure for each temporal superpixel is then used to merge pulse traces and estimate the photoplethysmogram signal. This allows to select skin tissue and furthermore to favor areas where the pulse trace is more predominant. Experimental results showed that our method perform better than state of the art algorithms without any critical face or skin detection.","PeriodicalId":151180,"journal":{"name":"2016 23rd International Conference on Pattern Recognition (ICPR)","volume":"2009 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127334850","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 23
Effective surface normals based action recognition in depth images 基于深度图像表面法线的有效动作识别
Pub Date : 2016-12-04 DOI: 10.1109/ICPR.2016.7899736
X. Nguyen, T. Nguyen, F. Charpillet
In this paper, we propose a new local descriptor for action recognition in depth images. The proposed descriptor relies on surface normals in 4D space of depth, time, spatial coordinates and higher-order partial derivatives of depth values along spatial coordinates. In order to classify actions, we follow the traditional Bag-of-words (BoW) approach, and propose two encoding methods termed Multi-Scale Fisher Vector (MSFV) and Temporal Sparse Coding based Fisher Vector Coding (TSCFVC) to form global representations of depth sequences. The high-dimensional action descriptors resulted from the two encoding methods are fed to a linear SVM for efficient action classification. Our proposed methods are evaluated on two public benchmark datasets, MSRAction3D and MSRGesture3D. The experimental result shows the effectiveness of the proposed methods on both the datasets.
本文提出了一种新的局部描述符用于深度图像的动作识别。该描述符依赖于深度、时间、空间坐标和深度值沿空间坐标的高阶偏导数在四维空间中的表面法线。为了对动作进行分类,我们遵循传统的词袋(BoW)方法,提出了多尺度Fisher向量(MSFV)和基于时间稀疏编码的Fisher向量编码(TSCFVC)两种编码方法来形成深度序列的全局表示。将两种编码方法产生的高维动作描述符输入到线性支持向量机中进行有效的动作分类。我们提出的方法在两个公共基准数据集MSRAction3D和MSRGesture3D上进行了评估。实验结果表明了所提方法在两种数据集上的有效性。
{"title":"Effective surface normals based action recognition in depth images","authors":"X. Nguyen, T. Nguyen, F. Charpillet","doi":"10.1109/ICPR.2016.7899736","DOIUrl":"https://doi.org/10.1109/ICPR.2016.7899736","url":null,"abstract":"In this paper, we propose a new local descriptor for action recognition in depth images. The proposed descriptor relies on surface normals in 4D space of depth, time, spatial coordinates and higher-order partial derivatives of depth values along spatial coordinates. In order to classify actions, we follow the traditional Bag-of-words (BoW) approach, and propose two encoding methods termed Multi-Scale Fisher Vector (MSFV) and Temporal Sparse Coding based Fisher Vector Coding (TSCFVC) to form global representations of depth sequences. The high-dimensional action descriptors resulted from the two encoding methods are fed to a linear SVM for efficient action classification. Our proposed methods are evaluated on two public benchmark datasets, MSRAction3D and MSRGesture3D. The experimental result shows the effectiveness of the proposed methods on both the datasets.","PeriodicalId":151180,"journal":{"name":"2016 23rd International Conference on Pattern Recognition (ICPR)","volume":"57 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121786348","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A 2D shape structure for decomposition and part similarity 二维形状结构的分解和零件相似度
Pub Date : 2016-12-04 DOI: 10.1109/ICPR.2016.7900130
Kathryn Leonard, Géraldine Morin, S. Hahmann, A. Carlier
This paper presents a multilevel analysis of 2D shapes and uses it to find similarities between the different parts of a shape. Such an analysis is important for many applications such as shape comparison, editing, and compression. Our robust and stable method decomposes a shape into parts, determines a parts hierarchy, and measures similarity between parts based on a salience measure on the medial axis, the Weighted Extended Distance Function, providing a multi-resolution partition of the shape that is stable across scale and articulation. Comparison with an extensive user study on the MPEG-7 database demonstrates that our geometric results are consistent with user perception.
本文提出了一种二维形状的多层次分析方法,并用它来寻找形状不同部分之间的相似性。这种分析对于形状比较、编辑和压缩等许多应用程序都很重要。我们的鲁棒和稳定的方法将形状分解成零件,确定零件层次结构,并基于中间轴上的显著性度量(加权扩展距离函数)来测量零件之间的相似性,从而提供一个跨尺度和关节稳定的形状的多分辨率分区。与MPEG-7数据库上广泛的用户研究比较表明,我们的几何结果与用户感知一致。
{"title":"A 2D shape structure for decomposition and part similarity","authors":"Kathryn Leonard, Géraldine Morin, S. Hahmann, A. Carlier","doi":"10.1109/ICPR.2016.7900130","DOIUrl":"https://doi.org/10.1109/ICPR.2016.7900130","url":null,"abstract":"This paper presents a multilevel analysis of 2D shapes and uses it to find similarities between the different parts of a shape. Such an analysis is important for many applications such as shape comparison, editing, and compression. Our robust and stable method decomposes a shape into parts, determines a parts hierarchy, and measures similarity between parts based on a salience measure on the medial axis, the Weighted Extended Distance Function, providing a multi-resolution partition of the shape that is stable across scale and articulation. Comparison with an extensive user study on the MPEG-7 database demonstrates that our geometric results are consistent with user perception.","PeriodicalId":151180,"journal":{"name":"2016 23rd International Conference on Pattern Recognition (ICPR)","volume":"49 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125241837","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
SCALP: Superpixels with Contour Adherence using Linear Path 头皮:超像素与轮廓依附性使用线性路径
Pub Date : 2016-12-04 DOI: 10.1109/ICPR.2016.7899991
Rémi Giraud, Vinh-Thong Ta, N. Papadakis
Superpixel decomposition methods are generally used as a pre-processing step to speed up image processing tasks. They group the pixels of an image into homogeneous regions while trying to respect existing contours. For all state-of-the-art superpixel decomposition methods, a trade-off is made between 1) computational time, 2) adherence to image contours and 3) regularity and compactness of the decomposition. In this paper, we propose a fast method to compute Superpixels with Contour Adherence using Linear Path (SCALP) in an iterative clustering framework. The distance computed when trying to associate a pixel to a superpixel during the clustering is enhanced by considering the linear path to the superpixel barycenter. The proposed framework produces regular and compact superpixels that adhere to the image contours. We provide a detailed evaluation of SCALP on the standard Berkeley Segmentation Dataset. The obtained results outperform state-of-the-art methods in terms of standard superpixel and contour detection metrics.
超像素分解方法通常用作预处理步骤,以加快图像处理任务。他们将图像的像素分组到均匀的区域,同时试图尊重现有的轮廓。对于所有最先进的超像素分解方法,都要在以下方面做出权衡:1)计算时间;2)对图像轮廓的依从性;3)分解的规律性和紧凑性。本文提出了一种基于迭代聚类框架的线性路径(头皮)快速计算具有轮廓依附性的超像素的方法。通过考虑到超像素质心的线性路径,在聚类过程中尝试将像素与超像素关联时计算的距离得到增强。所提出的框架产生符合图像轮廓的规则和紧凑的超像素。我们在标准的伯克利分割数据集上对头皮进行了详细的评估。所获得的结果在标准超像素和轮廓检测指标方面优于最先进的方法。
{"title":"SCALP: Superpixels with Contour Adherence using Linear Path","authors":"Rémi Giraud, Vinh-Thong Ta, N. Papadakis","doi":"10.1109/ICPR.2016.7899991","DOIUrl":"https://doi.org/10.1109/ICPR.2016.7899991","url":null,"abstract":"Superpixel decomposition methods are generally used as a pre-processing step to speed up image processing tasks. They group the pixels of an image into homogeneous regions while trying to respect existing contours. For all state-of-the-art superpixel decomposition methods, a trade-off is made between 1) computational time, 2) adherence to image contours and 3) regularity and compactness of the decomposition. In this paper, we propose a fast method to compute Superpixels with Contour Adherence using Linear Path (SCALP) in an iterative clustering framework. The distance computed when trying to associate a pixel to a superpixel during the clustering is enhanced by considering the linear path to the superpixel barycenter. The proposed framework produces regular and compact superpixels that adhere to the image contours. We provide a detailed evaluation of SCALP on the standard Berkeley Segmentation Dataset. The obtained results outperform state-of-the-art methods in terms of standard superpixel and contour detection metrics.","PeriodicalId":151180,"journal":{"name":"2016 23rd International Conference on Pattern Recognition (ICPR)","volume":"73 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133578007","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
期刊
2016 23rd International Conference on Pattern Recognition (ICPR)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1