首页 > 最新文献

2013 2nd IAPR Asian Conference on Pattern Recognition最新文献

英文 中文
Estimating the Structure of Rooms from a Single Fisheye Image 从单张鱼眼图像估计房间结构
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.148
Hanchao Jia, Shigang Li
Most existing approaches to indoor scene understanding formulate the problem based on the pinhole camera geometry. Unfortunately, these approaches cannot be utilized well for an omni directional image. In this paper, we focus on the problem of estimating the spatial layout of rooms from a single fisheye image. Considering the wide field of view of fisheye cameras, we introduce a structure symmetrical rule which describes geometric constraints. A method is given to estimate and recover the preliminary spatial layout of room only from a collection of line segments extracted from a fisheye image. Then, an orientation map of structure is generated. Finally, we refine the spatial layout to obtain the main structure. The experiments demonstrate that our approach based on geometric reasoning can be used to estimate the structure of indoor scene from a single fisheye image.
大多数现有的室内场景理解方法都是基于针孔摄像机的几何形状来解决问题的。不幸的是,这些方法不能很好地用于全向图像。本文主要研究从单张鱼眼图像中估计房间空间布局的问题。考虑到鱼眼相机的大视场,我们引入了描述几何约束的结构对称规则。给出了一种仅从鱼眼图像中提取线段集合来估计和恢复房间初步空间布局的方法。然后,生成结构的方位图。最后,对空间布局进行细化,得到主体结构。实验表明,基于几何推理的方法可以从单幅鱼眼图像中估计出室内场景的结构。
{"title":"Estimating the Structure of Rooms from a Single Fisheye Image","authors":"Hanchao Jia, Shigang Li","doi":"10.1109/ACPR.2013.148","DOIUrl":"https://doi.org/10.1109/ACPR.2013.148","url":null,"abstract":"Most existing approaches to indoor scene understanding formulate the problem based on the pinhole camera geometry. Unfortunately, these approaches cannot be utilized well for an omni directional image. In this paper, we focus on the problem of estimating the spatial layout of rooms from a single fisheye image. Considering the wide field of view of fisheye cameras, we introduce a structure symmetrical rule which describes geometric constraints. A method is given to estimate and recover the preliminary spatial layout of room only from a collection of line segments extracted from a fisheye image. Then, an orientation map of structure is generated. Finally, we refine the spatial layout to obtain the main structure. The experiments demonstrate that our approach based on geometric reasoning can be used to estimate the structure of indoor scene from a single fisheye image.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"36 3","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120857889","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Text Image Classifier Using Image-Wise Annotation 使用图像智能注释的文本图像分类器
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.160
N. Chiba
A text image classifier that requires only image-wise annotation is proposed. Although text detection methods using classifiers have been investigated, they require character-wise annotation by human operators, which is the most time-consuming phase when constructing a text detection system. The proposed classifier uses image-wise annotation whether the image contains text or not, which requires much less effort by an operator than that of character-wise annotation. From this annotation, the classifier estimates likelihood of detecting text-character candidates in an image as well as the threshold value for the system to determine if the image contains text based on prior probabilities. Experiments using real images showed the effectiveness of the proposed text image classifier.
提出了一种只需要图像注释的文本图像分类器。虽然已经研究了使用分类器的文本检测方法,但它们需要人工操作员进行逐字符注释,这是构建文本检测系统时最耗时的阶段。无论图像是否包含文本,所提出的分类器都使用图像智能注释,这比字符智能注释需要的操作要少得多。从这个注释中,分类器估计在图像中检测文本字符候选的可能性,以及系统根据先验概率确定图像是否包含文本的阈值。使用真实图像的实验证明了本文提出的文本图像分类器的有效性。
{"title":"Text Image Classifier Using Image-Wise Annotation","authors":"N. Chiba","doi":"10.1109/ACPR.2013.160","DOIUrl":"https://doi.org/10.1109/ACPR.2013.160","url":null,"abstract":"A text image classifier that requires only image-wise annotation is proposed. Although text detection methods using classifiers have been investigated, they require character-wise annotation by human operators, which is the most time-consuming phase when constructing a text detection system. The proposed classifier uses image-wise annotation whether the image contains text or not, which requires much less effort by an operator than that of character-wise annotation. From this annotation, the classifier estimates likelihood of detecting text-character candidates in an image as well as the threshold value for the system to determine if the image contains text based on prior probabilities. Experiments using real images showed the effectiveness of the proposed text image classifier.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115182698","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Head and Upper Body Pose Estimation in Team Sport Videos 团队运动录像中头部和上身姿势的估计
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.177
Masaki Hayashi, Taiki Yamamoto, Y. Aoki, Kyoko Oshima, Masamoto Tanabiki
We propose a head and upper body pose estimation method in low-resolution team sports videos such as for American Football or Hockey, where all players wear helmets and often lean forward. Compared to the pedestrian cases in surveillance videos, head pose estimation technique for team sports videos has to deal with various types of activities (poses) and image scales according to the position of the player in the field. Using both the pelvis aligned player tracker and the head tracker, our system tracks the player's pelvis and head positions, which results in estimation of player's 2D spine. Then, we estimate the head and upper body orientations independently with random decision forest classifiers learned from a dataset including multiple-scale images. Integrating upper body direction and 2D spine pose, we also estimate the 3D spine pose of the player. Experiments show our method can estimate head and upper body pose accurately for sports players with intensive movement even without any temporal filtering techniques by focusing on the upper body region.
我们提出了一种低分辨率团队运动视频(如美式橄榄球或曲棍球)中头部和上身姿势估计方法,在这些视频中,所有球员都戴着头盔,经常身体前倾。与监控视频中的行人案例相比,团队运动视频的头部姿态估计技术需要根据球员在场地中的位置处理各种类型的活动(姿态)和图像尺度。使用骨盆对齐的玩家追踪器和头部追踪器,我们的系统可以追踪玩家的骨盆和头部位置,从而估算出玩家的2D脊柱。然后,我们使用从包含多尺度图像的数据集中学习的随机决策森林分类器独立估计头部和上半身的方向。结合上身方向和2D脊柱姿态,我们还可以估算出玩家的3D脊柱姿态。实验表明,该方法可以在不使用任何时间过滤技术的情况下,对运动强度较大的运动员进行头部和上身姿势的准确估计。
{"title":"Head and Upper Body Pose Estimation in Team Sport Videos","authors":"Masaki Hayashi, Taiki Yamamoto, Y. Aoki, Kyoko Oshima, Masamoto Tanabiki","doi":"10.1109/ACPR.2013.177","DOIUrl":"https://doi.org/10.1109/ACPR.2013.177","url":null,"abstract":"We propose a head and upper body pose estimation method in low-resolution team sports videos such as for American Football or Hockey, where all players wear helmets and often lean forward. Compared to the pedestrian cases in surveillance videos, head pose estimation technique for team sports videos has to deal with various types of activities (poses) and image scales according to the position of the player in the field. Using both the pelvis aligned player tracker and the head tracker, our system tracks the player's pelvis and head positions, which results in estimation of player's 2D spine. Then, we estimate the head and upper body orientations independently with random decision forest classifiers learned from a dataset including multiple-scale images. Integrating upper body direction and 2D spine pose, we also estimate the 3D spine pose of the player. Experiments show our method can estimate head and upper body pose accurately for sports players with intensive movement even without any temporal filtering techniques by focusing on the upper body region.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115992019","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Sclera Recognition - A Survey 巩膜识别-一项调查
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.168
Abhijit Das, U. Pal, M. Blumenstein, M. A. Ferrer-Ballester
This paper presents a survey on sclera-based biometric recognition. Among the various biometric methods, sclera is one of the novel and promising biometric techniques. The sclera, a white region of connective tissue and blood vessels, surrounds the iris. A survey of the techniques available in the area of sclera biometrics will be of great assistance to researchers, and hence a comprehensive effort is made in this article to discuss the advancements reported in this regard during the past few decades. As a limited number of publications are found in the literature, an attempt is made in this paper to increase awareness of this area so that the topic gains popularity and interest among researchers. In this survey, a brief introduction is given initially about the sclera biometric, which is subsequently followed by background concepts, various pre-processing techniques, feature extraction and finally classification techniques associated with the sclera biometric. Benchmarking databases are very important for any pattern recognition related research, so the databases related with this work is also discussed. Finally, our observations, future scope and existing difficulties, which are unsolved in sclera biometrics, are discussed. We hope that this survey will serve to focus more researcher attention towards the emerging sclera biometric.
本文介绍了基于巩膜的生物特征识别的研究进展。在众多的生物识别方法中,巩膜是一种新颖而有发展前途的生物识别技术。巩膜是由结缔组织和血管组成的白色区域,包围着虹膜。对巩膜生物识别技术的调查将对研究人员有很大的帮助,因此本文将全面讨论过去几十年来在这方面的进展。由于在文献中发现的出版物数量有限,因此本文试图提高对该领域的认识,从而使该主题受到研究人员的欢迎和兴趣。本文首先对巩膜生物识别技术进行了简要介绍,然后介绍了背景概念、各种预处理技术、特征提取以及与巩膜生物识别技术相关的分类技术。基准数据库对于任何模式识别相关的研究都是非常重要的,因此本文还讨论了与此工作相关的数据库。最后,讨论了我们的观察结果,未来的范围和目前巩膜生物识别尚未解决的困难。我们希望这项调查将有助于把更多的研究人员的注意力集中到新兴的巩膜生物识别。
{"title":"Sclera Recognition - A Survey","authors":"Abhijit Das, U. Pal, M. Blumenstein, M. A. Ferrer-Ballester","doi":"10.1109/ACPR.2013.168","DOIUrl":"https://doi.org/10.1109/ACPR.2013.168","url":null,"abstract":"This paper presents a survey on sclera-based biometric recognition. Among the various biometric methods, sclera is one of the novel and promising biometric techniques. The sclera, a white region of connective tissue and blood vessels, surrounds the iris. A survey of the techniques available in the area of sclera biometrics will be of great assistance to researchers, and hence a comprehensive effort is made in this article to discuss the advancements reported in this regard during the past few decades. As a limited number of publications are found in the literature, an attempt is made in this paper to increase awareness of this area so that the topic gains popularity and interest among researchers. In this survey, a brief introduction is given initially about the sclera biometric, which is subsequently followed by background concepts, various pre-processing techniques, feature extraction and finally classification techniques associated with the sclera biometric. Benchmarking databases are very important for any pattern recognition related research, so the databases related with this work is also discussed. Finally, our observations, future scope and existing difficulties, which are unsolved in sclera biometrics, are discussed. We hope that this survey will serve to focus more researcher attention towards the emerging sclera biometric.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"54 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114159378","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 55
A New Multispectral Method for Face Liveness Detection 一种新的多光谱人脸活力检测方法
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.169
Yueyang Wang, X. Hao, Yali Hou, Changqing Guo
A face recognition system can be deceived by photos, mimic masks, mannequins and etc. And with the advances in the 3D printing technology, a more robust face liveness detection method is needed. In this paper, a gradient-based multispectral method has been proposed for face liveness detection. Based on two spectral bands, the developed method is tested for the classification of genuine faces and common disguised faces. A true positive rate of 96.7% and a true negative rate of 97% have been achieved. The performance of the method is also tested when face rotation occurs. The contributions of this paper are: First, a gradient-based multispectral method has been proposed. Except for the reflectance of the skin regions, the reflectance of other distinctive regions in a face are also considered in the developed method. Second, the method is tested based on a dataset with both planar photos and 3D mannequins and masks. The performance on different face orientations is also discussed.
人脸识别系统可以被照片、模拟面具、人体模型等欺骗。随着3D打印技术的进步,需要一种更加鲁棒的人脸活体检测方法。本文提出了一种基于梯度的多光谱人脸活体检测方法。基于两个光谱波段,对该方法进行了真实人脸和常见伪装人脸的分类试验。真阳性率为96.7%,真阴性率为97%。当人脸发生旋转时,测试了该方法的性能。本文的主要贡献有:首先,提出了一种基于梯度的多光谱方法。该方法除了考虑皮肤区域的反射率外,还考虑了面部其他特征区域的反射率。其次,基于平面照片和三维人体模型和面具的数据集对该方法进行了测试。讨论了在不同面向下的性能。
{"title":"A New Multispectral Method for Face Liveness Detection","authors":"Yueyang Wang, X. Hao, Yali Hou, Changqing Guo","doi":"10.1109/ACPR.2013.169","DOIUrl":"https://doi.org/10.1109/ACPR.2013.169","url":null,"abstract":"A face recognition system can be deceived by photos, mimic masks, mannequins and etc. And with the advances in the 3D printing technology, a more robust face liveness detection method is needed. In this paper, a gradient-based multispectral method has been proposed for face liveness detection. Based on two spectral bands, the developed method is tested for the classification of genuine faces and common disguised faces. A true positive rate of 96.7% and a true negative rate of 97% have been achieved. The performance of the method is also tested when face rotation occurs. The contributions of this paper are: First, a gradient-based multispectral method has been proposed. Except for the reflectance of the skin regions, the reflectance of other distinctive regions in a face are also considered in the developed method. Second, the method is tested based on a dataset with both planar photos and 3D mannequins and masks. The performance on different face orientations is also discussed.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114438674","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 28
Edge Guided High Order Image Smoothing 边缘引导的高阶图像平滑
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.47
Haoxing Wang, Longquan Dai, Xiaopeng Zhang
Edge-preserving smoothing has recently emerged as a valuable tool for a variety of applications in computer graphics and image processing. Edge-preserving smoothing using first order smoothness prior in the regularization term under optimization framework tends to bias the smoothing result forward the constant image. Although using high order smoothness prior can alleviate this problem, it tends to obtain the over-smoothed result. In this paper, we present an effective and practical image editing method which can sharply preserve the salient edges and at the same time smooths the continuous regions using high order smoothness prior to achieve the smoothing results different from the first order smoothness prior. Finally, we demonstrate the effectiveness of our method in the context of image denoising, image abstraction and image enhancement.
在计算机图形学和图像处理的各种应用中,边缘保持平滑已成为一种有价值的工具。在优化框架下,在正则化项中使用一阶平滑先验的保边平滑会使平滑结果向常数图像偏移。虽然使用高阶平滑先验可以缓解这一问题,但往往会得到过度平滑的结果。本文提出了一种有效实用的图像编辑方法,该方法在保留显著边缘的同时,利用高阶平滑先验对连续区域进行平滑,从而获得不同于一阶平滑先验的平滑效果。最后,我们在图像去噪、图像抽象和图像增强方面验证了该方法的有效性。
{"title":"Edge Guided High Order Image Smoothing","authors":"Haoxing Wang, Longquan Dai, Xiaopeng Zhang","doi":"10.1109/ACPR.2013.47","DOIUrl":"https://doi.org/10.1109/ACPR.2013.47","url":null,"abstract":"Edge-preserving smoothing has recently emerged as a valuable tool for a variety of applications in computer graphics and image processing. Edge-preserving smoothing using first order smoothness prior in the regularization term under optimization framework tends to bias the smoothing result forward the constant image. Although using high order smoothness prior can alleviate this problem, it tends to obtain the over-smoothed result. In this paper, we present an effective and practical image editing method which can sharply preserve the salient edges and at the same time smooths the continuous regions using high order smoothness prior to achieve the smoothing results different from the first order smoothness prior. Finally, we demonstrate the effectiveness of our method in the context of image denoising, image abstraction and image enhancement.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"61 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121983328","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
3-D Recovery of a Non-rigid Object from a Single Camera View Employing Multiple Coordinates Representation 采用多坐标表示的单摄像机视图非刚性物体的三维恢复
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.174
Shota Ishikawa, J. Tan, Hyoungseop Kim, S. Ishikawa
This paper proposes a novel technique for 3-D recovery of a non-rigid object, such as a human in motion, from a single camera view. To achieve the 3-D recovery, the proposed technique performs segmentation of an object under deformation into respective parts which are all regarded as rigid objects. For high accuracy segmentation, multi-stage learning and local subspace affinity are employed for the segmentation. Each part recovers its 3-D shape by applying the factorization method to it. Obviously the deformed portion containing twist or stretch motion cannot recover the 3-D shape by this procedure. The idea of the present paper is to recover such deformed portion by averaging the 3-D locations of a point on the portion described by the coordinates of respective parts. The experiments employing a synthetic non-rigid object and real human motion data show effectiveness of the proposed technique.
本文提出了一种新的非刚性物体(如运动中的人体)的单镜头三维恢复技术。为了实现三维恢复,该技术将变形物体分割成各自的部分,这些部分都被视为刚性物体。为了提高分割精度,采用了多阶段学习和局部子空间亲和的分割方法。每个部件通过因式分解方法恢复其三维形状。显然,包含扭转或拉伸运动的变形部分无法通过该方法恢复三维形状。本文的思想是通过对由各部分坐标描述的部分上的点的三维位置进行平均来恢复这些变形部分。利用合成非刚体和真实人体运动数据进行的实验表明了该方法的有效性。
{"title":"3-D Recovery of a Non-rigid Object from a Single Camera View Employing Multiple Coordinates Representation","authors":"Shota Ishikawa, J. Tan, Hyoungseop Kim, S. Ishikawa","doi":"10.1109/ACPR.2013.174","DOIUrl":"https://doi.org/10.1109/ACPR.2013.174","url":null,"abstract":"This paper proposes a novel technique for 3-D recovery of a non-rigid object, such as a human in motion, from a single camera view. To achieve the 3-D recovery, the proposed technique performs segmentation of an object under deformation into respective parts which are all regarded as rigid objects. For high accuracy segmentation, multi-stage learning and local subspace affinity are employed for the segmentation. Each part recovers its 3-D shape by applying the factorization method to it. Obviously the deformed portion containing twist or stretch motion cannot recover the 3-D shape by this procedure. The idea of the present paper is to recover such deformed portion by averaging the 3-D locations of a point on the portion described by the coordinates of respective parts. The experiments employing a synthetic non-rigid object and real human motion data show effectiveness of the proposed technique.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122167396","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
An Efficient Approach to Web Near-Duplicate Image Detection 一种有效的Web近重复图像检测方法
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.101
Jun Li, Shan Zhou, Junliang Xing, Changyin Sun, Weiming Hu
This paper presents an improved bag-of-words (BoW) framework for detecting near-duplicates of images on the Web and makes three main contributions. Firstly, based on the SIFT feature descriptors, Locality-constrained Linear Coding (LLC) with the spatial pyramid is introduced to encode features. Secondly, a weighted Chi-square distance metric is proposed to compare two histograms, with an inverted indexing scheme for fast similarity evaluation. Thirdly, a 6K dataset consisting of eight categories of objects, which can also be applicable to image retrieval and classification, is built and will be made available to the public in the future. We verify our technique on two benchmarks: our 6K dataset and the publicly available University of Kentucky Benchmark (UKB). The promising experimental results demonstrate the effectiveness and efficiency of our approach for Web Near-Duplicate Image Detection (Web-NDID), which outperforms several state-of-the-art methods.
本文提出了一种改进的词袋(BoW)框架,用于检测网络上图像的近重复,并做出了三个主要贡献。首先,在SIFT特征描述符的基础上,引入空间金字塔的位置约束线性编码(LLC)对特征进行编码;其次,提出了加权卡方距离度量来比较两个直方图,并采用倒排索引方案进行快速相似度评估。第三,构建一个由8类物体组成的6K数据集,该数据集也可用于图像检索和分类,并将在未来向公众开放。我们在两个基准测试上验证了我们的技术:我们的6K数据集和公开可用的肯塔基大学基准测试(UKB)。有希望的实验结果证明了我们的方法在网络近重复图像检测(Web- ndid)中的有效性和效率,它优于几种最先进的方法。
{"title":"An Efficient Approach to Web Near-Duplicate Image Detection","authors":"Jun Li, Shan Zhou, Junliang Xing, Changyin Sun, Weiming Hu","doi":"10.1109/ACPR.2013.101","DOIUrl":"https://doi.org/10.1109/ACPR.2013.101","url":null,"abstract":"This paper presents an improved bag-of-words (BoW) framework for detecting near-duplicates of images on the Web and makes three main contributions. Firstly, based on the SIFT feature descriptors, Locality-constrained Linear Coding (LLC) with the spatial pyramid is introduced to encode features. Secondly, a weighted Chi-square distance metric is proposed to compare two histograms, with an inverted indexing scheme for fast similarity evaluation. Thirdly, a 6K dataset consisting of eight categories of objects, which can also be applicable to image retrieval and classification, is built and will be made available to the public in the future. We verify our technique on two benchmarks: our 6K dataset and the publicly available University of Kentucky Benchmark (UKB). The promising experimental results demonstrate the effectiveness and efficiency of our approach for Web Near-Duplicate Image Detection (Web-NDID), which outperforms several state-of-the-art methods.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116814579","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Large-Scale Face Recognition on Smart Devices 智能设备上的大规模人脸识别
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.189
Jian-jun Hao, Yusuke Morishita, Toshinori Hosoi, K. Sakurai, Hitoshi Imaoka, Takao Imaizumi, Hideki Irisawa
Most of highly accurate face recognition methods are not suitable for real-time requirement in smart devices which have computational limitations. In this demonstration, we exhibit a face recognition application, in which only essential facial features from images are used for personal identification. In the algorithm used in this application, the face feature size is dramatically compressed into 512 bytes per face in spite of high recognition rate, a false rejection rate of 1.6% at false acceptance rate of 0.1% on identification photos. Consequently, computational cost for face matching is reduced dramatically and the system achieves 1.16 million times matching/second in dual-core 1.5GHz ARM processor. The demonstration on the smart device shows a high recognition performance and the feasibility for diverse applications.
大多数高精度的人脸识别方法都不适合智能设备的实时性要求,因为智能设备的计算能力有限。在本演示中,我们展示了一个人脸识别应用程序,其中仅使用图像中的基本面部特征进行个人识别。在本应用程序中使用的算法中,人脸特征大小被显著压缩为每张人脸512字节,尽管识别率很高,识别照片的错误拒绝率为1.6%,错误接受率为0.1%。从而大大降低了人脸匹配的计算成本,在双核1.5GHz ARM处理器上实现了116万次/秒的匹配。在智能设备上的演示表明,该方法具有较高的识别性能和多种应用的可行性。
{"title":"Large-Scale Face Recognition on Smart Devices","authors":"Jian-jun Hao, Yusuke Morishita, Toshinori Hosoi, K. Sakurai, Hitoshi Imaoka, Takao Imaizumi, Hideki Irisawa","doi":"10.1109/ACPR.2013.189","DOIUrl":"https://doi.org/10.1109/ACPR.2013.189","url":null,"abstract":"Most of highly accurate face recognition methods are not suitable for real-time requirement in smart devices which have computational limitations. In this demonstration, we exhibit a face recognition application, in which only essential facial features from images are used for personal identification. In the algorithm used in this application, the face feature size is dramatically compressed into 512 bytes per face in spite of high recognition rate, a false rejection rate of 1.6% at false acceptance rate of 0.1% on identification photos. Consequently, computational cost for face matching is reduced dramatically and the system achieves 1.16 million times matching/second in dual-core 1.5GHz ARM processor. The demonstration on the smart device shows a high recognition performance and the feasibility for diverse applications.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128443289","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Image Segmentation by Bilayer Superpixel Grouping 基于双层超像素分组的图像分割
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.62
M. Yang
The task of image segmentation is to group image pixels into visually meaningful objects. It has long been a challenging problem in computer vision and image processing. In this paper we address the segmentation as a super pixel grouping problem. We propose a novel graph-based segmentation framework which is able to integrate different cues from bilayer super pixels simultaneously. The key idea is that segmentation is formulated as grouping a subset of super pixels that partitions a bilayer graph over super pixels, with graph edges encoding super pixel similarity. We first construct a bipartite graph incorporating super pixel cue and long-range cue. Furthermore, mid-range cue is also incorporated in a hybrid graph model. Segmentation is solved by spectral clustering. Our approach is fully automatic, bottom-up, and unsupervised. We evaluate our proposed framework by comparing it to other generic segmentation approaches on the state-of-the-art benchmark database.
图像分割的任务是将图像像素分组为视觉上有意义的对象。长期以来,它一直是计算机视觉和图像处理领域的一个难题。在本文中,我们将分割作为一个超像素分组问题来解决。我们提出了一种新的基于图的分割框架,该框架能够同时整合来自双层超像素的不同线索。关键思想是,分割被表述为对超级像素的子集进行分组,该子集在超级像素上划分双层图,图边编码超级像素相似性。我们首先构造了一个包含超像素线索和远程线索的二部图。此外,还在混合图模型中加入了中程线索。采用谱聚类方法解决分割问题。我们的方法是全自动的,自下而上的,无监督的。我们通过将其与最先进的基准数据库上的其他通用分割方法进行比较来评估我们提出的框架。
{"title":"Image Segmentation by Bilayer Superpixel Grouping","authors":"M. Yang","doi":"10.1109/ACPR.2013.62","DOIUrl":"https://doi.org/10.1109/ACPR.2013.62","url":null,"abstract":"The task of image segmentation is to group image pixels into visually meaningful objects. It has long been a challenging problem in computer vision and image processing. In this paper we address the segmentation as a super pixel grouping problem. We propose a novel graph-based segmentation framework which is able to integrate different cues from bilayer super pixels simultaneously. The key idea is that segmentation is formulated as grouping a subset of super pixels that partitions a bilayer graph over super pixels, with graph edges encoding super pixel similarity. We first construct a bipartite graph incorporating super pixel cue and long-range cue. Furthermore, mid-range cue is also incorporated in a hybrid graph model. Segmentation is solved by spectral clustering. Our approach is fully automatic, bottom-up, and unsupervised. We evaluate our proposed framework by comparing it to other generic segmentation approaches on the state-of-the-art benchmark database.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129232430","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2013 2nd IAPR Asian Conference on Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1