首页 > 最新文献

2013 2nd IAPR Asian Conference on Pattern Recognition最新文献

英文 中文
Melanoma Classification Using Dermoscopy Imaging and Ensemble Learning 使用皮肤镜成像和集成学习进行黑色素瘤分类
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.102
G. Schaefer, B. Krawczyk, M. E. Celebi, H. Iyatomi
Malignant melanoma, the deadliest form of skin cancer, is one of the most rapidly increasing cancers in the world. Early diagnosis is crucial, since if detected early, it can be cured through a simple excision. In this paper, we present an effective approach to melanoma classification from dermoscopic images of skin lesions. First, we perform automatic border detection to delineate the lesion from the background skin. Shape features are then extracted from this border, while colour and texture features are obtained based on a division of the image into clinically significant regions. The derived features are then used in a pattern classification stage for which we employ a dedicated ensemble learning approach to address the class imbalance in the training data. Our classifier committee trains individual classifiers on balanced subspaces, removes redundant predictors based on a diversity measure and combines the remaining classifiers using a neural network fuser. Experimental results on a large dataset of dermoscopic skin lesion images show our approach to work well, to provide both high sensitivity and specificity, and the use of our classifier ensemble to lead to statistically better recognition performance.
恶性黑色素瘤是最致命的皮肤癌,也是世界上增长最快的癌症之一。早期诊断是至关重要的,因为如果早期发现,它可以通过简单的切除治愈。在本文中,我们提出了一种有效的方法来分类黑色素瘤从皮肤镜图像的皮肤病变。首先,我们进行自动边界检测,从背景皮肤中勾画病灶。然后从该边界提取形状特征,同时根据将图像划分为临床重要区域来获得颜色和纹理特征。然后在模式分类阶段使用衍生的特征,为此我们采用专用的集成学习方法来解决训练数据中的类不平衡。我们的分类器委员会在平衡子空间上训练单个分类器,基于多样性度量去除冗余预测器,并使用神经网络融合器组合剩余的分类器。在大型皮肤镜皮肤病变图像数据集上的实验结果表明,我们的方法工作良好,提供了高灵敏度和特异性,并且使用我们的分类器集合可以获得统计上更好的识别性能。
{"title":"Melanoma Classification Using Dermoscopy Imaging and Ensemble Learning","authors":"G. Schaefer, B. Krawczyk, M. E. Celebi, H. Iyatomi","doi":"10.1109/ACPR.2013.102","DOIUrl":"https://doi.org/10.1109/ACPR.2013.102","url":null,"abstract":"Malignant melanoma, the deadliest form of skin cancer, is one of the most rapidly increasing cancers in the world. Early diagnosis is crucial, since if detected early, it can be cured through a simple excision. In this paper, we present an effective approach to melanoma classification from dermoscopic images of skin lesions. First, we perform automatic border detection to delineate the lesion from the background skin. Shape features are then extracted from this border, while colour and texture features are obtained based on a division of the image into clinically significant regions. The derived features are then used in a pattern classification stage for which we employ a dedicated ensemble learning approach to address the class imbalance in the training data. Our classifier committee trains individual classifiers on balanced subspaces, removes redundant predictors based on a diversity measure and combines the remaining classifiers using a neural network fuser. Experimental results on a large dataset of dermoscopic skin lesion images show our approach to work well, to provide both high sensitivity and specificity, and the use of our classifier ensemble to lead to statistically better recognition performance.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122930802","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Magic Mirror: An Intelligent Fashion Recommendation System 魔镜:智能时尚推荐系统
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.212
Si Liu, Luoqi Liu, Shuicheng Yan
This paper mainly introduces the techniques required for a future system, called Magic Mirror. Imagine when you wake up in the morning and prepare for the coming day, the Magic Mirror will automatically recommend to you the most appropriate styles of hair, makeup, and dressing, according to the events and activities on your calendar, with which it is linked, so that you can present yourself on these occasions with elegant and suitable appearance. The work shall focus on the mathematical models for these tasks, particularly on how to model the relations between low-level human body features, middle-level facial/body attributes, and high-level recommendations. Being automatic and intelligent are the two main characteristics of the system, and this work shall show two prototype sub-systems related with the whole Magic Mirror system.
本文主要介绍了未来系统“魔镜”所需的技术。想象一下,当你早上醒来,为即将到来的一天做准备时,魔镜会根据你日历上的事件和活动,自动为你推荐最合适的发型、妆容和着装,让你在这些场合以优雅、合适的形象呈现自己。这些工作将集中在这些任务的数学模型上,特别是如何对低级人体特征、中级面部/身体属性和高级推荐之间的关系进行建模。自动化和智能化是系统的两个主要特点,本作品将展示与整个魔镜系统相关的两个原型子系统。
{"title":"Magic Mirror: An Intelligent Fashion Recommendation System","authors":"Si Liu, Luoqi Liu, Shuicheng Yan","doi":"10.1109/ACPR.2013.212","DOIUrl":"https://doi.org/10.1109/ACPR.2013.212","url":null,"abstract":"This paper mainly introduces the techniques required for a future system, called Magic Mirror. Imagine when you wake up in the morning and prepare for the coming day, the Magic Mirror will automatically recommend to you the most appropriate styles of hair, makeup, and dressing, according to the events and activities on your calendar, with which it is linked, so that you can present yourself on these occasions with elegant and suitable appearance. The work shall focus on the mathematical models for these tasks, particularly on how to model the relations between low-level human body features, middle-level facial/body attributes, and high-level recommendations. Being automatic and intelligent are the two main characteristics of the system, and this work shall show two prototype sub-systems related with the whole Magic Mirror system.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122289356","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Nuclear Norm Based 2DPCA 基于核规范的2DPCA
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.10
Fanlong Zhang, J. Qian, Jian Yang
This paper presents a novel method, namely nuclear norm based 2DPCA (N-2DPCA), for image feature extraction. Unlike the conventional 2DPCA, N-2DPCA uses a nuclear norm based reconstruction error criterion. The criterion is minimized by converting the nuclear norm based optimization problem into a series of F-norm based optimization problems. N-2DPCA is applied to face recognition and is evaluated using the Extended Yale B and CMU PIE databases. Experimental results demonstrate that our method is more effective and robust than PCA, 2DPCA and L1-Norm based 2DPCA.
提出了一种基于核范数的2DPCA (N-2DPCA)图像特征提取方法。与传统的2DPCA不同,N-2DPCA使用基于核范数的重构误差准则。通过将基于核范数的优化问题转化为一系列基于f范数的优化问题,使准则最小化。N-2DPCA应用于人脸识别,并使用扩展耶鲁B和CMU PIE数据库进行评估。实验结果表明,该方法比PCA、2DPCA和基于L1-Norm的2DPCA具有更好的鲁棒性和有效性。
{"title":"Nuclear Norm Based 2DPCA","authors":"Fanlong Zhang, J. Qian, Jian Yang","doi":"10.1109/ACPR.2013.10","DOIUrl":"https://doi.org/10.1109/ACPR.2013.10","url":null,"abstract":"This paper presents a novel method, namely nuclear norm based 2DPCA (N-2DPCA), for image feature extraction. Unlike the conventional 2DPCA, N-2DPCA uses a nuclear norm based reconstruction error criterion. The criterion is minimized by converting the nuclear norm based optimization problem into a series of F-norm based optimization problems. N-2DPCA is applied to face recognition and is evaluated using the Extended Yale B and CMU PIE databases. Experimental results demonstrate that our method is more effective and robust than PCA, 2DPCA and L1-Norm based 2DPCA.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125464308","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
How Do Facial Expressions Contribute to Age Prediction? 面部表情如何有助于年龄预测?
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.161
Yingmei Piao, Mineichi Kudo
Human age estimation based on facial images has many potential applications in practice. However, the current age estimation techniques are not matured. Most studies focus only on neutral faces, that is, expressionless faces. Several expressions such as happy expression, may help to improve the prediction accuracy. Recently, some works reported that expressions could badly impact on the accuracy. In this paper, we investigated the degree of facial expression impact on age prediction subjectively and objectively. It was revealed that expressions do not contribute for age prediction so much.
基于人脸图像的人类年龄估计在实践中有许多潜在的应用。然而,目前的年龄估计技术还不成熟。大多数研究只关注中性面孔,即没有表情的面孔。一些表情,如快乐的表情,可能有助于提高预测的准确性。最近,一些作品报道了表达会严重影响准确性。本文从主观上和客观上考察了面部表情对年龄预测的影响程度。结果表明,表情对年龄预测的贡献并不大。
{"title":"How Do Facial Expressions Contribute to Age Prediction?","authors":"Yingmei Piao, Mineichi Kudo","doi":"10.1109/ACPR.2013.161","DOIUrl":"https://doi.org/10.1109/ACPR.2013.161","url":null,"abstract":"Human age estimation based on facial images has many potential applications in practice. However, the current age estimation techniques are not matured. Most studies focus only on neutral faces, that is, expressionless faces. Several expressions such as happy expression, may help to improve the prediction accuracy. Recently, some works reported that expressions could badly impact on the accuracy. In this paper, we investigated the degree of facial expression impact on age prediction subjectively and objectively. It was revealed that expressions do not contribute for age prediction so much.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121454415","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Deformed and Touched Characters Recognition 变形和触摸字符识别
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.193
Tadashi Hyuga, H. Wada, Tomoyoshi Aizawa, Yoshihisa Ijiri, M. Kawade
In this demonstration, we will show our Optical Character Recognition(OCR) technique. Character deformation and touching problems often occur during high-speed printing process in the machine vision industry. As a result, it is difficult for OCR system to segment and recognize characters properly. To solve these problems, we propose a novel OCR technique which is robust against deformation and touching. It splits regions of characters simply and excessively, recognizes all segments and merged regions, and obtains optimal segments using graph theory.
在这个演示中,我们将展示我们的光学字符识别(OCR)技术。在机器视觉行业中,高速印刷过程中经常出现字符变形和触摸问题。因此,OCR系统难以对字符进行正确的分割和识别。为了解决这些问题,我们提出了一种新的抗变形和抗触摸的OCR技术。该算法对字符区域进行简单而过度的分割,对所有分割区域和合并区域进行识别,并利用图论得到最优分割区域。
{"title":"Deformed and Touched Characters Recognition","authors":"Tadashi Hyuga, H. Wada, Tomoyoshi Aizawa, Yoshihisa Ijiri, M. Kawade","doi":"10.1109/ACPR.2013.193","DOIUrl":"https://doi.org/10.1109/ACPR.2013.193","url":null,"abstract":"In this demonstration, we will show our Optical Character Recognition(OCR) technique. Character deformation and touching problems often occur during high-speed printing process in the machine vision industry. As a result, it is difficult for OCR system to segment and recognize characters properly. To solve these problems, we propose a novel OCR technique which is robust against deformation and touching. It splits regions of characters simply and excessively, recognizes all segments and merged regions, and obtains optimal segments using graph theory.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127876425","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Multi-layered Background Modeling for Complex Environment Surveillance 复杂环境监测的多层背景建模
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.83
S. Yoshinaga, Atsushi Shimada, H. Nagahara, R. Taniguchi, Kouichiro Kajitani, Takeshi Naito
Many background models have been proposed to adapt to "illumination changes" and "dynamic changes" such as swaying motion of tree branches. However, the problem of background maintenance in complex environment, where foreground objects pass in front of stationary objects which cease moving, is still far from being completely solved. To address this problem, we propose a framework for multi-layered background modeling, in which we conserve the background models for stationary objects hierarchically in addition to the one for the initial background. To realize this framework, we also propose a spatio-temporal background model based on the similarity in the intensity changes among pixels. Experimental results on complex scenes, such as a bus stop and an intersection, show that our proposed method can adapt to both appearances and disappearances of stationary objects thanks to the multi-layered background modeling framework.
为了适应“光照变化”和“动态变化”(如树枝的摇摆运动),人们提出了许多背景模型。然而,在前景物体经过静止物体而静止物体停止运动的复杂环境下,背景维护问题还远未完全解决。为了解决这一问题,我们提出了一种多层背景建模框架,在该框架中,除了初始背景模型外,我们还分层保留静止物体的背景模型。为了实现这一框架,我们还提出了基于像素间强度变化相似性的时空背景模型。在公交车站和十字路口等复杂场景下的实验结果表明,基于多层背景建模框架的方法能够适应静止物体的出现和消失。
{"title":"Multi-layered Background Modeling for Complex Environment Surveillance","authors":"S. Yoshinaga, Atsushi Shimada, H. Nagahara, R. Taniguchi, Kouichiro Kajitani, Takeshi Naito","doi":"10.1109/ACPR.2013.83","DOIUrl":"https://doi.org/10.1109/ACPR.2013.83","url":null,"abstract":"Many background models have been proposed to adapt to \"illumination changes\" and \"dynamic changes\" such as swaying motion of tree branches. However, the problem of background maintenance in complex environment, where foreground objects pass in front of stationary objects which cease moving, is still far from being completely solved. To address this problem, we propose a framework for multi-layered background modeling, in which we conserve the background models for stationary objects hierarchically in addition to the one for the initial background. To realize this framework, we also propose a spatio-temporal background model based on the similarity in the intensity changes among pixels. Experimental results on complex scenes, such as a bus stop and an intersection, show that our proposed method can adapt to both appearances and disappearances of stationary objects thanks to the multi-layered background modeling framework.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127767480","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Towards Robust Gait Recognition 稳健的步态识别
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.211
Yasushi Makihara
Gait recognition is a method of biometric person authentication from his/her unconscious walking manner. Unlike the other biometrics such as DNA, fingerprint, vein, and iris, the gait can be recognized even at a distance from a camera without subjects' cooperation, and hence it is expected to be applied to many fields: criminal investigation, forensic science, and surveillance. However, the absence of the subjects' cooperation may sometimes induces large intra-subject variations of the gait due to the changes of viewpoints, walking directions, speeds, clothes, and shoes. We therefore develop methods of robust gait recognition with (1) an appearance-based view transformation model, (2) a kinematics-based speed transformation model. Moreover, CCTV footages are often stored as low frame-rate videos due to limitation of communication bandwidth and storage size, which makes it much more difficult to observe a continuous gait motion and hence significantly degrades the gait recognition performance. We therefore solve this problem with (3) a technique of periodic temporal super resolution from a low frame-rate video. We show the efficiency of the proposed methods with our constructed gait databases.
步态识别是一种根据人的无意识行走方式对人进行生物识别的方法。与DNA、指纹、静脉、虹膜等其他生物识别技术不同,即使在距离摄像机很远的地方也能识别出步态,无需受试者的配合,因此有望在刑事调查、法医学、监视等领域得到应用。然而,缺乏受试者的合作有时会引起受试者内部由于视点、行走方向、速度、衣服和鞋子的变化而产生的步态变化。因此,我们开发了鲁棒步态识别方法,包括:(1)基于外观的视图转换模型,(2)基于运动学的速度转换模型。此外,由于通信带宽和存储容量的限制,CCTV视频通常以低帧率的视频形式存储,这使得观察连续的步态运动变得更加困难,从而大大降低了步态识别的性能。因此,我们用(3)低帧率视频的周期性时间超分辨率技术来解决这个问题。我们用所构建的步态数据库证明了所提方法的有效性。
{"title":"Towards Robust Gait Recognition","authors":"Yasushi Makihara","doi":"10.1109/ACPR.2013.211","DOIUrl":"https://doi.org/10.1109/ACPR.2013.211","url":null,"abstract":"Gait recognition is a method of biometric person authentication from his/her unconscious walking manner. Unlike the other biometrics such as DNA, fingerprint, vein, and iris, the gait can be recognized even at a distance from a camera without subjects' cooperation, and hence it is expected to be applied to many fields: criminal investigation, forensic science, and surveillance. However, the absence of the subjects' cooperation may sometimes induces large intra-subject variations of the gait due to the changes of viewpoints, walking directions, speeds, clothes, and shoes. We therefore develop methods of robust gait recognition with (1) an appearance-based view transformation model, (2) a kinematics-based speed transformation model. Moreover, CCTV footages are often stored as low frame-rate videos due to limitation of communication bandwidth and storage size, which makes it much more difficult to observe a continuous gait motion and hence significantly degrades the gait recognition performance. We therefore solve this problem with (3) a technique of periodic temporal super resolution from a low frame-rate video. We show the efficiency of the proposed methods with our constructed gait databases.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129056148","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Compacting Large and Loose Communities 压缩大型和松散的社区
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.137
V. Chandrashekar, Shailesh Kumar, C. V. Jawahar
Detecting compact overlapping communities in large networks is an important pattern recognition problem with applications in many domains. Most community detection algorithms trade-off between community sizes, their compactness and the scalability of finding communities. Clique Percolation Method (CPM) and Local Fitness Maximization (LFM) are two prominent and commonly used overlapping community detection methods that scale with large networks. However, significant number of communities found by them are large, noisy, and loose. In this paper, we propose a general algorithm that takes such large and loose communities generated by any method and refines them into compact communities in a systematic fashion. We define a new measure of community-ness based on eigenvector centrality, identify loose communities using this measure and propose an algorithm for partitioning such loose communities into compact communities. We refine the communities found by CPM and LFM using our method and show their effectiveness compared to the original communities in a recommendation engine task.
大型网络中紧密重叠社区的检测是一个重要的模式识别问题,在许多领域都有应用。大多数社区检测算法在社区大小、紧凑性和发现社区的可扩展性之间进行权衡。Clique peration Method (CPM)和Local Fitness Maximization (LFM)是两种重要且常用的大型网络重叠社区检测方法。然而,他们发现的大量社区规模大、嘈杂、松散。在本文中,我们提出了一种通用算法,该算法可以将任何方法生成的如此大且松散的社区以系统的方式精炼成紧凑的社区。我们定义了一种基于特征向量中心性的新的社区度量,利用该度量来识别松散社区,并提出了一种将松散社区划分为紧密社区的算法。我们使用我们的方法对CPM和LFM发现的社区进行了细化,并在推荐引擎任务中与原始社区进行了比较,展示了它们的有效性。
{"title":"Compacting Large and Loose Communities","authors":"V. Chandrashekar, Shailesh Kumar, C. V. Jawahar","doi":"10.1109/ACPR.2013.137","DOIUrl":"https://doi.org/10.1109/ACPR.2013.137","url":null,"abstract":"Detecting compact overlapping communities in large networks is an important pattern recognition problem with applications in many domains. Most community detection algorithms trade-off between community sizes, their compactness and the scalability of finding communities. Clique Percolation Method (CPM) and Local Fitness Maximization (LFM) are two prominent and commonly used overlapping community detection methods that scale with large networks. However, significant number of communities found by them are large, noisy, and loose. In this paper, we propose a general algorithm that takes such large and loose communities generated by any method and refines them into compact communities in a systematic fashion. We define a new measure of community-ness based on eigenvector centrality, identify loose communities using this measure and propose an algorithm for partitioning such loose communities into compact communities. We refine the communities found by CPM and LFM using our method and show their effectiveness compared to the original communities in a recommendation engine task.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129125574","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Rapid Mobile Object Recognition Using Fisher Vector 使用Fisher矢量快速移动目标识别
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.39
Yoshiyuki Kawano, Keiji Yanai
We propose a real-time object recognition method for a smart phone, which consists of light-weight local features, Fisher Vector and linear SVM. As light local descriptors, we adopt a HOG Patch descriptor and a Color Patch descriptor, and sample them from an image densely. Then we encode them with Fisher Vector representation, which can save the number of visual words greatly. As a classifier, we use a liner SVM the computational cost of which is very low. In the experiments, we have achieved the 79.2% classification rate for the top 5 category candidates for a 100-category food dataset. It outperformed the results using a conventional bag-of-features representation with a chi-square-RBF-kernel-based SVM. Moreover, the processing time of food recognition takes only 0.065 seconds, which is four times as faster as the existing work.
提出了一种基于轻量级局部特征、Fisher向量和线性支持向量机的智能手机实时目标识别方法。作为轻量级局部描述符,我们采用HOG Patch描述符和Color Patch描述符,并对它们进行密集采样。然后用Fisher向量表示对其进行编码,可以大大节省视觉词的数量。作为分类器,我们使用线性支持向量机,其计算成本非常低。在实验中,我们对100类食品数据集的前5类候选分类达到了79.2%的分类率。它优于使用传统的特征袋表示和基于卡方rbf核的支持向量机的结果。此外,食物识别的处理时间仅为0.065秒,比现有工作快了4倍。
{"title":"Rapid Mobile Object Recognition Using Fisher Vector","authors":"Yoshiyuki Kawano, Keiji Yanai","doi":"10.1109/ACPR.2013.39","DOIUrl":"https://doi.org/10.1109/ACPR.2013.39","url":null,"abstract":"We propose a real-time object recognition method for a smart phone, which consists of light-weight local features, Fisher Vector and linear SVM. As light local descriptors, we adopt a HOG Patch descriptor and a Color Patch descriptor, and sample them from an image densely. Then we encode them with Fisher Vector representation, which can save the number of visual words greatly. As a classifier, we use a liner SVM the computational cost of which is very low. In the experiments, we have achieved the 79.2% classification rate for the top 5 category candidates for a 100-category food dataset. It outperformed the results using a conventional bag-of-features representation with a chi-square-RBF-kernel-based SVM. Moreover, the processing time of food recognition takes only 0.065 seconds, which is four times as faster as the existing work.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"63 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114897434","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
A Maximum Correlation Feature Descriptor for Heterogeneous Face Recognition 异构人脸识别的最大相关特征描述符
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.12
Dihong Gong, J. Zheng
Heterogeneous Face Recognition (HFR) refers to matching probe face images to a gallery of face images taken from alternate imaging modality, for example matching near infrared (NIR) face images to photographs. Matching heterogeneous face images has important practical applications such as surveillance and forensics, which is yet a challenging problem in face recognition community due to the large within-class discrepancy incurred from modality differences. In this paper, a novel feature descriptor is proposed in which the features of both gallery and probe face images are extracted with an adaptive feature descriptor which can maximize the correlation of the encoded face images between the modalities, so as to reduce the within-class variations at the feature extraction stage. The effectiveness of the proposed approach is demonstrated on the scenario of matching NIR face images to photographs based on a very large dataset consists of 2800 different persons.
异质人脸识别(HFR)是指将探针人脸图像与从其他成像方式获取的人脸图像进行匹配,例如将近红外人脸图像与照片进行匹配。异构人脸图像的匹配在监控和取证等领域有着重要的实际应用,但由于模态差异导致的类内差异较大,这是人脸识别领域的一个难题。本文提出了一种新的特征描述符,通过自适应特征描述符提取图库和探测图像的特征,使编码后的图像在模态之间的相关性最大化,从而减少特征提取阶段的类内变化。在基于2800个不同人的大型数据集的近红外人脸图像与照片匹配的场景中,证明了所提出方法的有效性。
{"title":"A Maximum Correlation Feature Descriptor for Heterogeneous Face Recognition","authors":"Dihong Gong, J. Zheng","doi":"10.1109/ACPR.2013.12","DOIUrl":"https://doi.org/10.1109/ACPR.2013.12","url":null,"abstract":"Heterogeneous Face Recognition (HFR) refers to matching probe face images to a gallery of face images taken from alternate imaging modality, for example matching near infrared (NIR) face images to photographs. Matching heterogeneous face images has important practical applications such as surveillance and forensics, which is yet a challenging problem in face recognition community due to the large within-class discrepancy incurred from modality differences. In this paper, a novel feature descriptor is proposed in which the features of both gallery and probe face images are extracted with an adaptive feature descriptor which can maximize the correlation of the encoded face images between the modalities, so as to reduce the within-class variations at the feature extraction stage. The effectiveness of the proposed approach is demonstrated on the scenario of matching NIR face images to photographs based on a very large dataset consists of 2800 different persons.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133790830","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
期刊
2013 2nd IAPR Asian Conference on Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1