首页 > 最新文献

The First Asian Conference on Pattern Recognition最新文献

英文 中文
Online codebook reweighting using pairwise constraints for image classification 利用两两约束进行图像分类的在线码本重加权
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166560
Xin Zhao, Weiqiang Ren, Kaiqi Huang, T. Tan
Bag-of-words (BoW) model is widely used for image classification. Recently, the framework of sparse coding and max pooling proved an effective approach for image classification. Max pooling adopts a winner-take-all strategy. Thus, it can be regarded as a codebook weighting process. The results of this process are the weights of the associated codebook. However, there are high intra-class variations and strong background clutters in many image classification tasks. The weights obtained by max pooling only have limited information. This paper presents a codebook reweighting algorithm using pairwise constraints to improve the performance of sparse coding and max pooling framework. Pairwise constraints are the natural way of encoding the relationships between pairs of images. Therefore, the reweighted codebook is more effective to describe the relevance between pairs of images. An efficient online learning algorithm is presented based on passive-aggressive training strategy. We compare our method with other state-of-the-art methods on Graz-01 & 02 datasets. Experimental results illustrate the effectiveness and efficiency of our method for image classification.
词袋模型(Bag-of-words, BoW)被广泛用于图像分类。近年来,稀疏编码和最大池化框架被证明是一种有效的图像分类方法。最大池采用赢者通吃的策略。因此,它可以看作是一个码本加权过程。此过程的结果是相关码本的权重。然而,在许多图像分类任务中,存在着较大的类内变异和较强的背景杂波。最大池化获得的权值信息有限。为了提高稀疏编码和最大池化框架的性能,提出了一种利用成对约束的码本重加权算法。成对约束是对图像对之间的关系进行编码的自然方式。因此,重新加权的码本可以更有效地描述图像对之间的相关性。提出了一种基于被动攻击训练策略的高效在线学习算法。我们将我们的方法与其他最先进的方法在grazi -01和02数据集上进行比较。实验结果证明了该方法的有效性和高效性。
{"title":"Online codebook reweighting using pairwise constraints for image classification","authors":"Xin Zhao, Weiqiang Ren, Kaiqi Huang, T. Tan","doi":"10.1109/ACPR.2011.6166560","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166560","url":null,"abstract":"Bag-of-words (BoW) model is widely used for image classification. Recently, the framework of sparse coding and max pooling proved an effective approach for image classification. Max pooling adopts a winner-take-all strategy. Thus, it can be regarded as a codebook weighting process. The results of this process are the weights of the associated codebook. However, there are high intra-class variations and strong background clutters in many image classification tasks. The weights obtained by max pooling only have limited information. This paper presents a codebook reweighting algorithm using pairwise constraints to improve the performance of sparse coding and max pooling framework. Pairwise constraints are the natural way of encoding the relationships between pairs of images. Therefore, the reweighted codebook is more effective to describe the relevance between pairs of images. An efficient online learning algorithm is presented based on passive-aggressive training strategy. We compare our method with other state-of-the-art methods on Graz-01 & 02 datasets. Experimental results illustrate the effectiveness and efficiency of our method for image classification.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"70 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128948879","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Silhouette extraction based on time-series statistical modeling and k-means clustering 基于时间序列统计建模和k-means聚类的轮廓提取
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166672
A. Hamad, N. Tsumura
This paper proposes a simple and a robust method to detect and extract the silhouettes from a video sequence of a static camera based on background subtraction technique. The proposed method analyse the pixel history as a time series observations. A robust technique to detect motion based on kernel density estimation is presented. Two consecutive stages of the k-means clustering algorithm are utilized to identify the most reliable background regions and decrease false positives. Pixel and object based updating mechanism is presented to cope with challenges like gradual and sudden illumination changes, ghost appearance, and non-stationary background objects. Experimental results show the efficiency and the robustness of the proposed method to detect and extract silhouettes for outdoor and indoor environments.
本文提出了一种简单、鲁棒的基于背景减法的静态摄像机视频序列剪影检测与提取方法。该方法将像素历史作为时间序列观测来分析。提出了一种基于核密度估计的鲁棒运动检测技术。利用连续两个阶段的k-means聚类算法来识别最可靠的背景区域并减少误报。提出了基于像素和对象的更新机制,以应对渐变和突然的光照变化、鬼影现象和非静止背景物体等挑战。实验结果表明了该方法在室外和室内环境下轮廓检测和提取的有效性和鲁棒性。
{"title":"Silhouette extraction based on time-series statistical modeling and k-means clustering","authors":"A. Hamad, N. Tsumura","doi":"10.1109/ACPR.2011.6166672","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166672","url":null,"abstract":"This paper proposes a simple and a robust method to detect and extract the silhouettes from a video sequence of a static camera based on background subtraction technique. The proposed method analyse the pixel history as a time series observations. A robust technique to detect motion based on kernel density estimation is presented. Two consecutive stages of the k-means clustering algorithm are utilized to identify the most reliable background regions and decrease false positives. Pixel and object based updating mechanism is presented to cope with challenges like gradual and sudden illumination changes, ghost appearance, and non-stationary background objects. Experimental results show the efficiency and the robustness of the proposed method to detect and extract silhouettes for outdoor and indoor environments.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"83 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126272285","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Contextual Constrained Independent Component Analysis based foreground detection for indoor surveillance 基于上下文约束独立分量分析的室内监控前景检测
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166656
Zhong Zhang, Baihua Xiao, Chunheng Wang, Wen Zhou, Shuang Liu
Recently, Independent Component Analysis based foreground detection has been proposed for indoor surveillance applications where the foreground tends to move slowly or remain still. Yet such a method often causes discrete segmented foreground objects. In this paper, we propose a novel foreground detection method named Contextual Constrained Independent Component Analysis (CCICA) to tackle this problem. In our method, the contextual constraints are explicitly added to the optimization objective function, which indicate the similarity relationship among neighboring pixels. In this way, the obtained de-mixing matrix can produce the complete foreground compared with the previous ICA model. In addition, our method performs robust to the indoor illumination changes and features a high processing speed. Two sets of image sequences involving room lights switching on/of and door opening/closing are tested. The experimental results clearly demonstrate an improvement over the basic ICA model and the image difference method.
最近,基于独立分量分析的前景检测被提出用于前景移动缓慢或保持静止的室内监控应用。然而,这种方法往往导致离散的分割前景对象。为了解决这一问题,本文提出了一种新的前景检测方法——上下文约束独立分量分析(CCICA)。在我们的方法中,上下文约束被显式地添加到优化目标函数中,这表明了相邻像素之间的相似关系。这样,得到的去混矩阵与之前的ICA模型相比,可以得到完整的前景。此外,该方法对室内光照变化具有鲁棒性,处理速度快。测试了两组图像序列,包括房间灯的开/关和门的开/关。实验结果清楚地表明,该方法比基本ICA模型和图像差分方法有了改进。
{"title":"Contextual Constrained Independent Component Analysis based foreground detection for indoor surveillance","authors":"Zhong Zhang, Baihua Xiao, Chunheng Wang, Wen Zhou, Shuang Liu","doi":"10.1109/ACPR.2011.6166656","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166656","url":null,"abstract":"Recently, Independent Component Analysis based foreground detection has been proposed for indoor surveillance applications where the foreground tends to move slowly or remain still. Yet such a method often causes discrete segmented foreground objects. In this paper, we propose a novel foreground detection method named Contextual Constrained Independent Component Analysis (CCICA) to tackle this problem. In our method, the contextual constraints are explicitly added to the optimization objective function, which indicate the similarity relationship among neighboring pixels. In this way, the obtained de-mixing matrix can produce the complete foreground compared with the previous ICA model. In addition, our method performs robust to the indoor illumination changes and features a high processing speed. Two sets of image sequences involving room lights switching on/of and door opening/closing are tested. The experimental results clearly demonstrate an improvement over the basic ICA model and the image difference method.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114332398","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Designing and selecting features for MR image segmentation 磁共振图像分割特征的设计与选择
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166535
Meijuan Yang, Yuan Yuan, Xuelong Li, Pingkun Yan
Deformable models have obtained considerable success in medical image segmentation, due to its ability of capturing the shape variation of the target structure. Boundary feature is used to guide contour deformation, which plays an decisive part in deformable model based segmentation. However, it is still a challenging task to obtain a distinctive image feature to describe the boundaries, since boundaries are not necessarily in accordance with edges or ridges. Another challenge is to infer the shape for the given image appearance. In this paper, the anatomical structures from MR images are aimed to be segmented. First, a new normal vector feature profile (NVFP) is employed to describe the local image appearance of a contour point formed by a series of modified SIFT local descriptors along the normal direction of that point. Second, the shape of the target structure is inferred by matching two image appearances of the test image and learned image appearance. A new match function is designed to incorporate the new NVFP to deformable models. During the optimization procedure of the segmentation algorithm, the nearest neighbor approach is used to compute the displacement of each contour point to guide the global shape deformation. Experimental results on prostate and bladder MR images show that the proposed method has a better performance than the previous method.
可变形模型由于能够捕捉目标结构的形状变化,在医学图像分割中取得了相当大的成功。利用边界特征指导轮廓变形,在基于变形模型的分割中起着决定性的作用。然而,由于边界不一定与边缘或脊一致,因此如何获得鲜明的图像特征来描述边界仍然是一项具有挑战性的任务。另一个挑战是推断给定图像外观的形状。本文的目的是对磁共振图像中的解剖结构进行分割。首先,利用一种新的法向量特征轮廓(NVFP)来描述由一系列改进的SIFT局部描述子沿轮廓点的法线方向形成的轮廓点的局部图像外观;其次,通过匹配测试图像和学习图像的两个图像外观来推断目标结构的形状;设计了一个新的匹配函数,将新的NVFP结合到可变形模型中。在分割算法的优化过程中,采用最近邻法计算每个轮廓点的位移,以指导全局形状变形。在前列腺和膀胱MR图像上的实验结果表明,该方法具有较好的性能。
{"title":"Designing and selecting features for MR image segmentation","authors":"Meijuan Yang, Yuan Yuan, Xuelong Li, Pingkun Yan","doi":"10.1109/ACPR.2011.6166535","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166535","url":null,"abstract":"Deformable models have obtained considerable success in medical image segmentation, due to its ability of capturing the shape variation of the target structure. Boundary feature is used to guide contour deformation, which plays an decisive part in deformable model based segmentation. However, it is still a challenging task to obtain a distinctive image feature to describe the boundaries, since boundaries are not necessarily in accordance with edges or ridges. Another challenge is to infer the shape for the given image appearance. In this paper, the anatomical structures from MR images are aimed to be segmented. First, a new normal vector feature profile (NVFP) is employed to describe the local image appearance of a contour point formed by a series of modified SIFT local descriptors along the normal direction of that point. Second, the shape of the target structure is inferred by matching two image appearances of the test image and learned image appearance. A new match function is designed to incorporate the new NVFP to deformable models. During the optimization procedure of the segmentation algorithm, the nearest neighbor approach is used to compute the displacement of each contour point to guide the global shape deformation. Experimental results on prostate and bladder MR images show that the proposed method has a better performance than the previous method.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"114 26","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"113946008","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-view vehicle detection and tracking in crossroads 十字路口多视点车辆检测与跟踪
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166688
Liwei Liu, Junliang Xing, H. Ai
Multi-view vehicle detection and tracking in crossroads is of fundamental importance in traffic surveillance yet still remains a very challenging task. The view changes of different vehicles and their occlusions in crossroads are two main difficulties that often fail many existing methods. To handle these difficulties, we propose a new method for multi-view vehicle detection and tracking that innovates mainly on two aspects: the two-stage view selection and the dual-layer occlusion handling. For the two-stage view selection, a Multi-Modal Particle Filter (MMPF) is proposed to track vehicles in explicit view, i.e. frontal (rear) view or side view. In the second stage, for the vehicles in inexplicit views, i.e. intermediate views between frontal and side view, spatial-temporal analysis is employed to further decide their views so as to maintain the consistence of view transition. For the dual-layer occlusion handling, a cluster based dedicated vehicle model for partial occlusion and a backward retracking procedure for full occlusion are integrated complementarily to deal with occlusion problems. The two-stage view selection is efficient for fusing multiple detectors, while the dual-layer occlusion handling improves tracking performance effectively. Extensive experiments under different weather conditions, including snowy, sunny and cloudy, demonstrate the effectiveness and efficiency of our method.
交叉口多视点车辆检测与跟踪是交通监控的基础,但也是一项非常具有挑战性的任务。在十字路口,不同车辆的视野变化和遮挡是许多现有方法难以解决的两个主要问题。针对这些困难,本文提出了一种新的多视角车辆检测与跟踪方法,主要从两阶段视角选择和双层遮挡处理两个方面进行了创新。对于两阶段的视图选择,提出了一种多模态粒子滤波器(MMPF)来跟踪显视图(即前(后)视图或侧视图)中的车辆。第二阶段,对于处于不明确视角(即介于正视图和侧视图之间的中间视角)的车辆,通过时空分析进一步确定其视角,以保持视角转换的一致性。在双层遮挡处理中,将基于聚类的局部遮挡专用车辆模型与完全遮挡的向后回溯过程相结合,互补处理遮挡问题。两阶段的视图选择对于融合多个检测器是有效的,而双层遮挡处理有效地提高了跟踪性能。在不同的天气条件下,包括下雪、晴天和阴天,大量的实验证明了我们的方法的有效性和效率。
{"title":"Multi-view vehicle detection and tracking in crossroads","authors":"Liwei Liu, Junliang Xing, H. Ai","doi":"10.1109/ACPR.2011.6166688","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166688","url":null,"abstract":"Multi-view vehicle detection and tracking in crossroads is of fundamental importance in traffic surveillance yet still remains a very challenging task. The view changes of different vehicles and their occlusions in crossroads are two main difficulties that often fail many existing methods. To handle these difficulties, we propose a new method for multi-view vehicle detection and tracking that innovates mainly on two aspects: the two-stage view selection and the dual-layer occlusion handling. For the two-stage view selection, a Multi-Modal Particle Filter (MMPF) is proposed to track vehicles in explicit view, i.e. frontal (rear) view or side view. In the second stage, for the vehicles in inexplicit views, i.e. intermediate views between frontal and side view, spatial-temporal analysis is employed to further decide their views so as to maintain the consistence of view transition. For the dual-layer occlusion handling, a cluster based dedicated vehicle model for partial occlusion and a backward retracking procedure for full occlusion are integrated complementarily to deal with occlusion problems. The two-stage view selection is efficient for fusing multiple detectors, while the dual-layer occlusion handling improves tracking performance effectively. Extensive experiments under different weather conditions, including snowy, sunny and cloudy, demonstrate the effectiveness and efficiency of our method.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"24 6","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114033755","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Emergency light detection in tunnel environment: An efficient method 隧道环境中应急光探测的一种有效方法
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166528
Zhipeng Wang, M. Kagesawa, Shintaro Ono, A. Banno, K. Ikeuchi
Automobile navigation in tunnel environment is challenging. GPS sensors and ordinary cameras can't function effectively. For navigation, infrared cameras are installed on top of our experimental vehicle, and here we propose an efficient object detection method to detect emergency lights from the collected data in tunnel environment. The proposed method firstly detects keypoints by setting thresholds for intensity of uniformly sampled points. Each keypoint is then verified by the appearance of its surrounding sub-image. After clustering the keypoints which satisfy the verification, the method verifies the keypoint clusters by their appearance and temporal information. Though the later steps are time-consuming, they deal with very few instances. And this improves the efficiency of the method, while not losing effectiveness of the appearance and temporal information. Thus the method gives promising results in real time. Detection performance and efficiency are verified by experiments carried on challenging real data.
隧道环境下的汽车导航具有一定的挑战性。GPS传感器和普通摄像头无法有效发挥作用。在导航方面,我们的实验车辆顶部安装了红外摄像机,在这里我们提出了一种高效的目标检测方法,从采集的数据中检测隧道环境中的应急灯。该方法首先通过对均匀采样点的强度设置阈值来检测关键点。然后通过其周围子图像的外观来验证每个关键点。该方法将满足验证条件的关键点聚类后,根据关键点的外观和时间信息对关键点聚类进行验证。尽管后面的步骤很耗时,但它们处理的实例很少。在不损失外观信息和时间信息有效性的前提下,提高了方法的效率。因此,该方法的实时性较好。在具有挑战性的真实数据上进行了实验,验证了检测的性能和效率。
{"title":"Emergency light detection in tunnel environment: An efficient method","authors":"Zhipeng Wang, M. Kagesawa, Shintaro Ono, A. Banno, K. Ikeuchi","doi":"10.1109/ACPR.2011.6166528","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166528","url":null,"abstract":"Automobile navigation in tunnel environment is challenging. GPS sensors and ordinary cameras can't function effectively. For navigation, infrared cameras are installed on top of our experimental vehicle, and here we propose an efficient object detection method to detect emergency lights from the collected data in tunnel environment. The proposed method firstly detects keypoints by setting thresholds for intensity of uniformly sampled points. Each keypoint is then verified by the appearance of its surrounding sub-image. After clustering the keypoints which satisfy the verification, the method verifies the keypoint clusters by their appearance and temporal information. Though the later steps are time-consuming, they deal with very few instances. And this improves the efficiency of the method, while not losing effectiveness of the appearance and temporal information. Thus the method gives promising results in real time. Detection performance and efficiency are verified by experiments carried on challenging real data.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122887984","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Real-time lip reading system for fixed phrase and its combination 固定短语及其组合的实时唇读系统
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166595
T. Saitoh
This paper presents a high-level real-time lip reading system that can recognize both fixed phrase and its combination. Lip reading provides an important means for realizing a communication support interface for speech handicaps. The approach is based on the Viola-Jones face detector, lip extraction based on active appearance model, automatic utterance section detector, and phrase classifier using DP matching. Our system has two unique ideas; combined phrase recognition, and remove false recognition phrase from a target list. These ideas are simple but efficient in practical use. We demonstrated our system with five subjects and confirmed the usefulness.
本文提出了一种高水平的实时唇读系统,可以同时识别固定短语及其组合。唇读是实现语言障碍交流支持界面的重要手段。该方法基于Viola-Jones人脸检测器、基于主动外观模型的唇形提取、自动话语片段检测器和基于DP匹配的短语分类器。我们的系统有两个独特的想法;组合短语识别,并从目标列表中删除错误的识别短语。这些想法简单,但在实际应用中很有效。我们用5个受试者演示了我们的系统,并证实了它的实用性。
{"title":"Real-time lip reading system for fixed phrase and its combination","authors":"T. Saitoh","doi":"10.1109/ACPR.2011.6166595","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166595","url":null,"abstract":"This paper presents a high-level real-time lip reading system that can recognize both fixed phrase and its combination. Lip reading provides an important means for realizing a communication support interface for speech handicaps. The approach is based on the Viola-Jones face detector, lip extraction based on active appearance model, automatic utterance section detector, and phrase classifier using DP matching. Our system has two unique ideas; combined phrase recognition, and remove false recognition phrase from a target list. These ideas are simple but efficient in practical use. We demonstrated our system with five subjects and confirmed the usefulness.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128762869","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Image mosaic using Log-polar binning 使用对数极分割的图像拼接
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166693
Le Thanh Hoan, Youngjae Chun, Kyoungsu Oh
Image mosaic is a large image assembled from many smaller tiles which one tile itself is an actual image. In this research, we introduce an efficient method to make image mosaic. Our method is based on Log-polar mapping which enables us to detect the color and shape change. We also successfully make an image mosaic version by exploiting GPU power. Our algorithm is simple, easy to implement, gives better result than conventional method and can be improved to higher precision.
图像拼接是由许多较小的瓷砖拼接而成的大图像,其中一个瓷砖本身就是一个实际的图像。本文介绍了一种有效的图像拼接方法。我们的方法是基于对数极坐标映射,使我们能够检测颜色和形状的变化。我们还成功地利用GPU的能力制作了图像拼接版本。该算法简单,易于实现,结果优于传统方法,可提高精度。
{"title":"Image mosaic using Log-polar binning","authors":"Le Thanh Hoan, Youngjae Chun, Kyoungsu Oh","doi":"10.1109/ACPR.2011.6166693","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166693","url":null,"abstract":"Image mosaic is a large image assembled from many smaller tiles which one tile itself is an actual image. In this research, we introduce an efficient method to make image mosaic. Our method is based on Log-polar mapping which enables us to detect the color and shape change. We also successfully make an image mosaic version by exploiting GPU power. Our algorithm is simple, easy to implement, gives better result than conventional method and can be improved to higher precision.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129394249","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Improvements to facial contour detection by hierarchical fitting and regression 基于层次拟合和回归的人脸轮廓检测改进
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166689
Atsushi Irie, M. Takagiwa, Kozo Moriyama, Takayoshi Yamashita
There are many methods based on shape and texture models for detecting eye and mouth contour points from facial images. They reduce the false positive rate by utilizing a global model and adapting it for a given face. Changes to facial expressions are coupled with changes to the shapes of eyes and mouth, and a global facial model in itself cannot be adapted to all human facial expressions. Therefore, a hierarchical model fitting approach has been developed, whereby the global fitting captures the facial shape using the global model and the local fitting captures the each facial parts using these local models. This can detect facial contours with high accuracy for expressions to which the global model cannot be adapted.
基于形状和纹理模型的眼、口轮廓点检测方法有很多。他们利用一个全局模型,并根据给定的人脸进行调整,从而降低了误报率。面部表情的变化伴随着眼睛和嘴的形状的变化,一个全局的面部模型本身不能适应所有的人类面部表情。因此,开发了一种分层模型拟合方法,其中全局拟合使用全局模型捕获面部形状,局部拟合使用这些局部模型捕获每个面部部分。对于全局模型无法适应的表情,这种方法可以高精度地检测出面部轮廓。
{"title":"Improvements to facial contour detection by hierarchical fitting and regression","authors":"Atsushi Irie, M. Takagiwa, Kozo Moriyama, Takayoshi Yamashita","doi":"10.1109/ACPR.2011.6166689","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166689","url":null,"abstract":"There are many methods based on shape and texture models for detecting eye and mouth contour points from facial images. They reduce the false positive rate by utilizing a global model and adapting it for a given face. Changes to facial expressions are coupled with changes to the shapes of eyes and mouth, and a global facial model in itself cannot be adapted to all human facial expressions. Therefore, a hierarchical model fitting approach has been developed, whereby the global fitting captures the facial shape using the global model and the local fitting captures the each facial parts using these local models. This can detect facial contours with high accuracy for expressions to which the global model cannot be adapted.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126902199","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
Decision fusion for block linear regression classification based on confidence index 基于置信度指标的块线性回归分类决策融合
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166708
Yi-fei Xu, He-lei Wu
We consider the problem of recognizing human faces with varying expression and illumination, and a novel confidence index based block linear regression classification method is proposed. Our approach divides images into blocks, and each block is identified using the linear regression classifier separately. We develop a confidence index model to measure the recognition confidence of each block, and the final decision is achieved by aggregating individual results with the designed Bayesian decision fusion algorithm. The performances of our approach and conventional algorithms are evaluated under conditions of varying expression and illumination using benchmark databases, improvements demonstrate the proposed approach is robustness to both expression and illumination variations.
针对不同表情和光照的人脸识别问题,提出了一种基于置信度指数的块线性回归分类方法。我们的方法将图像分成块,每个块分别使用线性回归分类器进行识别。我们建立了一个置信度指数模型来衡量每个块的识别置信度,并利用所设计的贝叶斯决策融合算法对单个结果进行汇总,从而得到最终的决策。在不同的表达和光照条件下,使用基准数据库对我们的方法和传统算法的性能进行了评估,改进表明我们的方法对表达和光照变化都具有鲁棒性。
{"title":"Decision fusion for block linear regression classification based on confidence index","authors":"Yi-fei Xu, He-lei Wu","doi":"10.1109/ACPR.2011.6166708","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166708","url":null,"abstract":"We consider the problem of recognizing human faces with varying expression and illumination, and a novel confidence index based block linear regression classification method is proposed. Our approach divides images into blocks, and each block is identified using the linear regression classifier separately. We develop a confidence index model to measure the recognition confidence of each block, and the final decision is achieved by aggregating individual results with the designed Bayesian decision fusion algorithm. The performances of our approach and conventional algorithms are evaluated under conditions of varying expression and illumination using benchmark databases, improvements demonstrate the proposed approach is robustness to both expression and illumination variations.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125790024","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
The First Asian Conference on Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1