首页 > 最新文献

The First Asian Conference on Pattern Recognition最新文献

英文 中文
Classification based character segmentation guided by Fast-Hessian-Affine regions 基于快速黑森仿射区域的分类字符分割
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166546
Takahiro Ota, T. Wada
This paper presents a method of fast and accurate character localization for OCR (Optical Character Reader). We already proposed an acceleration framework of arbitrary classifiers, classifier molding, for real-time verification of characters printed by Industrial Ink Jet Printer (IIJP). In this framework, the behavior of accurate but slow character classifier is learnt by linear regression tree. The resulted classifier is up to 1,500 times faster than the original one but is not fast enough for real-time pyramidal scan of VGA images, which is necessary for scale-free character recognition. For solving this problem, we also proposed CCS (Classification based Character Segmentation). This method finds character arrangement that maximizes the sum of the likelihood of character regions assuming that all characters are horizontally aligned with almost regular intervals. This assumption is not always true even for the characters printed by IIJP. For solving this problem, we extended the idea of CCS to arbitrary located characters. Our method first generates character-region candidates based on local elliptical regions, named Fast-Hessian-Affine regions, and finds most likely character arrangement. Through experiments, we confirmed that our method quickly and accurately recognizes non-uniformly arranged characters.
提出了一种快速准确的OCR (Optical character Reader)字符定位方法。我们已经提出了一个任意分类器的加速框架,分类器成型,用于工业喷墨打印机(IIJP)打印的字符的实时验证。在该框架中,通过线性回归树学习准确但速度慢的字符分类器的行为。所得到的分类器比原来的分类器快了1500倍,但对于VGA图像的实时金字塔扫描来说还不够快,这是无比例字符识别所必需的。为了解决这个问题,我们还提出了基于分类的字符分割(CCS)。该方法找到的字符排列,最大限度地提高字符区域的可能性的总和,假设所有字符水平对齐几乎有规则的间隔。即使对于IIJP打印的字符,这个假设也不总是正确的。为了解决这个问题,我们将CCS的理念扩展到任意位置的字符。该方法首先基于局部椭圆区域(Fast-Hessian-Affine region)生成候选字符区域,并找到最可能的字符排列。实验结果表明,该方法能够快速准确地识别非均匀排列字符。
{"title":"Classification based character segmentation guided by Fast-Hessian-Affine regions","authors":"Takahiro Ota, T. Wada","doi":"10.1109/ACPR.2011.6166546","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166546","url":null,"abstract":"This paper presents a method of fast and accurate character localization for OCR (Optical Character Reader). We already proposed an acceleration framework of arbitrary classifiers, classifier molding, for real-time verification of characters printed by Industrial Ink Jet Printer (IIJP). In this framework, the behavior of accurate but slow character classifier is learnt by linear regression tree. The resulted classifier is up to 1,500 times faster than the original one but is not fast enough for real-time pyramidal scan of VGA images, which is necessary for scale-free character recognition. For solving this problem, we also proposed CCS (Classification based Character Segmentation). This method finds character arrangement that maximizes the sum of the likelihood of character regions assuming that all characters are horizontally aligned with almost regular intervals. This assumption is not always true even for the characters printed by IIJP. For solving this problem, we extended the idea of CCS to arbitrary located characters. Our method first generates character-region candidates based on local elliptical regions, named Fast-Hessian-Affine regions, and finds most likely character arrangement. Through experiments, we confirmed that our method quickly and accurately recognizes non-uniformly arranged characters.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"140 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122404767","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Online codebook reweighting using pairwise constraints for image classification 利用两两约束进行图像分类的在线码本重加权
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166560
Xin Zhao, Weiqiang Ren, Kaiqi Huang, T. Tan
Bag-of-words (BoW) model is widely used for image classification. Recently, the framework of sparse coding and max pooling proved an effective approach for image classification. Max pooling adopts a winner-take-all strategy. Thus, it can be regarded as a codebook weighting process. The results of this process are the weights of the associated codebook. However, there are high intra-class variations and strong background clutters in many image classification tasks. The weights obtained by max pooling only have limited information. This paper presents a codebook reweighting algorithm using pairwise constraints to improve the performance of sparse coding and max pooling framework. Pairwise constraints are the natural way of encoding the relationships between pairs of images. Therefore, the reweighted codebook is more effective to describe the relevance between pairs of images. An efficient online learning algorithm is presented based on passive-aggressive training strategy. We compare our method with other state-of-the-art methods on Graz-01 & 02 datasets. Experimental results illustrate the effectiveness and efficiency of our method for image classification.
词袋模型(Bag-of-words, BoW)被广泛用于图像分类。近年来,稀疏编码和最大池化框架被证明是一种有效的图像分类方法。最大池采用赢者通吃的策略。因此,它可以看作是一个码本加权过程。此过程的结果是相关码本的权重。然而,在许多图像分类任务中,存在着较大的类内变异和较强的背景杂波。最大池化获得的权值信息有限。为了提高稀疏编码和最大池化框架的性能,提出了一种利用成对约束的码本重加权算法。成对约束是对图像对之间的关系进行编码的自然方式。因此,重新加权的码本可以更有效地描述图像对之间的相关性。提出了一种基于被动攻击训练策略的高效在线学习算法。我们将我们的方法与其他最先进的方法在grazi -01和02数据集上进行比较。实验结果证明了该方法的有效性和高效性。
{"title":"Online codebook reweighting using pairwise constraints for image classification","authors":"Xin Zhao, Weiqiang Ren, Kaiqi Huang, T. Tan","doi":"10.1109/ACPR.2011.6166560","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166560","url":null,"abstract":"Bag-of-words (BoW) model is widely used for image classification. Recently, the framework of sparse coding and max pooling proved an effective approach for image classification. Max pooling adopts a winner-take-all strategy. Thus, it can be regarded as a codebook weighting process. The results of this process are the weights of the associated codebook. However, there are high intra-class variations and strong background clutters in many image classification tasks. The weights obtained by max pooling only have limited information. This paper presents a codebook reweighting algorithm using pairwise constraints to improve the performance of sparse coding and max pooling framework. Pairwise constraints are the natural way of encoding the relationships between pairs of images. Therefore, the reweighted codebook is more effective to describe the relevance between pairs of images. An efficient online learning algorithm is presented based on passive-aggressive training strategy. We compare our method with other state-of-the-art methods on Graz-01 & 02 datasets. Experimental results illustrate the effectiveness and efficiency of our method for image classification.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"70 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128948879","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A new approach of color image quantization based on Normalized Cut algorithm 基于归一化切割算法的彩色图像量化新方法
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166589
Jin Zhang, Yonghong Song, Yuanlin Zhang, Xiaobing Wang
This paper presents a novel color quantization method based on Normalized Cut clustering algorithm, in order to generate a quantized image with the minimum loss of information and the maximum compression ratio, which benefits the storage and transmission of the color image. This new method uses a deformed Median Cut algorithm as a coarse partition of color pixels in the RGB color space, and then take the average color of each partition as the representative color of a node to construct a condensed graph. By employing the Normalized Cut clustering algorithm, we could get the palette with defined color number, and then reconstruct the quantized image. Experiments on common used test images demonstrate that our method is very competitive with state-of-the-art color quantization methods in terms of image quality, compression ratio and computation time.
提出了一种基于归一化切聚类算法的彩色量化方法,以最小的信息损失和最大的压缩比生成量化图像,有利于彩色图像的存储和传输。该方法采用一种变形中值切割算法对RGB颜色空间中的颜色像素进行粗分割,然后以每个分割的平均颜色作为节点的代表颜色来构造一个凝聚图。采用归一化切聚类算法,得到具有定义颜色数的调色板,然后重建量化后的图像。在常用的测试图像上进行的实验表明,我们的方法在图像质量、压缩比和计算时间方面与目前最先进的颜色量化方法相比具有很强的竞争力。
{"title":"A new approach of color image quantization based on Normalized Cut algorithm","authors":"Jin Zhang, Yonghong Song, Yuanlin Zhang, Xiaobing Wang","doi":"10.1109/ACPR.2011.6166589","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166589","url":null,"abstract":"This paper presents a novel color quantization method based on Normalized Cut clustering algorithm, in order to generate a quantized image with the minimum loss of information and the maximum compression ratio, which benefits the storage and transmission of the color image. This new method uses a deformed Median Cut algorithm as a coarse partition of color pixels in the RGB color space, and then take the average color of each partition as the representative color of a node to construct a condensed graph. By employing the Normalized Cut clustering algorithm, we could get the palette with defined color number, and then reconstruct the quantized image. Experiments on common used test images demonstrate that our method is very competitive with state-of-the-art color quantization methods in terms of image quality, compression ratio and computation time.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134223912","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Designing and selecting features for MR image segmentation 磁共振图像分割特征的设计与选择
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166535
Meijuan Yang, Yuan Yuan, Xuelong Li, Pingkun Yan
Deformable models have obtained considerable success in medical image segmentation, due to its ability of capturing the shape variation of the target structure. Boundary feature is used to guide contour deformation, which plays an decisive part in deformable model based segmentation. However, it is still a challenging task to obtain a distinctive image feature to describe the boundaries, since boundaries are not necessarily in accordance with edges or ridges. Another challenge is to infer the shape for the given image appearance. In this paper, the anatomical structures from MR images are aimed to be segmented. First, a new normal vector feature profile (NVFP) is employed to describe the local image appearance of a contour point formed by a series of modified SIFT local descriptors along the normal direction of that point. Second, the shape of the target structure is inferred by matching two image appearances of the test image and learned image appearance. A new match function is designed to incorporate the new NVFP to deformable models. During the optimization procedure of the segmentation algorithm, the nearest neighbor approach is used to compute the displacement of each contour point to guide the global shape deformation. Experimental results on prostate and bladder MR images show that the proposed method has a better performance than the previous method.
可变形模型由于能够捕捉目标结构的形状变化,在医学图像分割中取得了相当大的成功。利用边界特征指导轮廓变形,在基于变形模型的分割中起着决定性的作用。然而,由于边界不一定与边缘或脊一致,因此如何获得鲜明的图像特征来描述边界仍然是一项具有挑战性的任务。另一个挑战是推断给定图像外观的形状。本文的目的是对磁共振图像中的解剖结构进行分割。首先,利用一种新的法向量特征轮廓(NVFP)来描述由一系列改进的SIFT局部描述子沿轮廓点的法线方向形成的轮廓点的局部图像外观;其次,通过匹配测试图像和学习图像的两个图像外观来推断目标结构的形状;设计了一个新的匹配函数,将新的NVFP结合到可变形模型中。在分割算法的优化过程中,采用最近邻法计算每个轮廓点的位移,以指导全局形状变形。在前列腺和膀胱MR图像上的实验结果表明,该方法具有较好的性能。
{"title":"Designing and selecting features for MR image segmentation","authors":"Meijuan Yang, Yuan Yuan, Xuelong Li, Pingkun Yan","doi":"10.1109/ACPR.2011.6166535","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166535","url":null,"abstract":"Deformable models have obtained considerable success in medical image segmentation, due to its ability of capturing the shape variation of the target structure. Boundary feature is used to guide contour deformation, which plays an decisive part in deformable model based segmentation. However, it is still a challenging task to obtain a distinctive image feature to describe the boundaries, since boundaries are not necessarily in accordance with edges or ridges. Another challenge is to infer the shape for the given image appearance. In this paper, the anatomical structures from MR images are aimed to be segmented. First, a new normal vector feature profile (NVFP) is employed to describe the local image appearance of a contour point formed by a series of modified SIFT local descriptors along the normal direction of that point. Second, the shape of the target structure is inferred by matching two image appearances of the test image and learned image appearance. A new match function is designed to incorporate the new NVFP to deformable models. During the optimization procedure of the segmentation algorithm, the nearest neighbor approach is used to compute the displacement of each contour point to guide the global shape deformation. Experimental results on prostate and bladder MR images show that the proposed method has a better performance than the previous method.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"114 26","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"113946008","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-view vehicle detection and tracking in crossroads 十字路口多视点车辆检测与跟踪
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166688
Liwei Liu, Junliang Xing, H. Ai
Multi-view vehicle detection and tracking in crossroads is of fundamental importance in traffic surveillance yet still remains a very challenging task. The view changes of different vehicles and their occlusions in crossroads are two main difficulties that often fail many existing methods. To handle these difficulties, we propose a new method for multi-view vehicle detection and tracking that innovates mainly on two aspects: the two-stage view selection and the dual-layer occlusion handling. For the two-stage view selection, a Multi-Modal Particle Filter (MMPF) is proposed to track vehicles in explicit view, i.e. frontal (rear) view or side view. In the second stage, for the vehicles in inexplicit views, i.e. intermediate views between frontal and side view, spatial-temporal analysis is employed to further decide their views so as to maintain the consistence of view transition. For the dual-layer occlusion handling, a cluster based dedicated vehicle model for partial occlusion and a backward retracking procedure for full occlusion are integrated complementarily to deal with occlusion problems. The two-stage view selection is efficient for fusing multiple detectors, while the dual-layer occlusion handling improves tracking performance effectively. Extensive experiments under different weather conditions, including snowy, sunny and cloudy, demonstrate the effectiveness and efficiency of our method.
交叉口多视点车辆检测与跟踪是交通监控的基础,但也是一项非常具有挑战性的任务。在十字路口,不同车辆的视野变化和遮挡是许多现有方法难以解决的两个主要问题。针对这些困难,本文提出了一种新的多视角车辆检测与跟踪方法,主要从两阶段视角选择和双层遮挡处理两个方面进行了创新。对于两阶段的视图选择,提出了一种多模态粒子滤波器(MMPF)来跟踪显视图(即前(后)视图或侧视图)中的车辆。第二阶段,对于处于不明确视角(即介于正视图和侧视图之间的中间视角)的车辆,通过时空分析进一步确定其视角,以保持视角转换的一致性。在双层遮挡处理中,将基于聚类的局部遮挡专用车辆模型与完全遮挡的向后回溯过程相结合,互补处理遮挡问题。两阶段的视图选择对于融合多个检测器是有效的,而双层遮挡处理有效地提高了跟踪性能。在不同的天气条件下,包括下雪、晴天和阴天,大量的实验证明了我们的方法的有效性和效率。
{"title":"Multi-view vehicle detection and tracking in crossroads","authors":"Liwei Liu, Junliang Xing, H. Ai","doi":"10.1109/ACPR.2011.6166688","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166688","url":null,"abstract":"Multi-view vehicle detection and tracking in crossroads is of fundamental importance in traffic surveillance yet still remains a very challenging task. The view changes of different vehicles and their occlusions in crossroads are two main difficulties that often fail many existing methods. To handle these difficulties, we propose a new method for multi-view vehicle detection and tracking that innovates mainly on two aspects: the two-stage view selection and the dual-layer occlusion handling. For the two-stage view selection, a Multi-Modal Particle Filter (MMPF) is proposed to track vehicles in explicit view, i.e. frontal (rear) view or side view. In the second stage, for the vehicles in inexplicit views, i.e. intermediate views between frontal and side view, spatial-temporal analysis is employed to further decide their views so as to maintain the consistence of view transition. For the dual-layer occlusion handling, a cluster based dedicated vehicle model for partial occlusion and a backward retracking procedure for full occlusion are integrated complementarily to deal with occlusion problems. The two-stage view selection is efficient for fusing multiple detectors, while the dual-layer occlusion handling improves tracking performance effectively. Extensive experiments under different weather conditions, including snowy, sunny and cloudy, demonstrate the effectiveness and efficiency of our method.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"24 6","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114033755","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Emergency light detection in tunnel environment: An efficient method 隧道环境中应急光探测的一种有效方法
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166528
Zhipeng Wang, M. Kagesawa, Shintaro Ono, A. Banno, K. Ikeuchi
Automobile navigation in tunnel environment is challenging. GPS sensors and ordinary cameras can't function effectively. For navigation, infrared cameras are installed on top of our experimental vehicle, and here we propose an efficient object detection method to detect emergency lights from the collected data in tunnel environment. The proposed method firstly detects keypoints by setting thresholds for intensity of uniformly sampled points. Each keypoint is then verified by the appearance of its surrounding sub-image. After clustering the keypoints which satisfy the verification, the method verifies the keypoint clusters by their appearance and temporal information. Though the later steps are time-consuming, they deal with very few instances. And this improves the efficiency of the method, while not losing effectiveness of the appearance and temporal information. Thus the method gives promising results in real time. Detection performance and efficiency are verified by experiments carried on challenging real data.
隧道环境下的汽车导航具有一定的挑战性。GPS传感器和普通摄像头无法有效发挥作用。在导航方面,我们的实验车辆顶部安装了红外摄像机,在这里我们提出了一种高效的目标检测方法,从采集的数据中检测隧道环境中的应急灯。该方法首先通过对均匀采样点的强度设置阈值来检测关键点。然后通过其周围子图像的外观来验证每个关键点。该方法将满足验证条件的关键点聚类后,根据关键点的外观和时间信息对关键点聚类进行验证。尽管后面的步骤很耗时,但它们处理的实例很少。在不损失外观信息和时间信息有效性的前提下,提高了方法的效率。因此,该方法的实时性较好。在具有挑战性的真实数据上进行了实验,验证了检测的性能和效率。
{"title":"Emergency light detection in tunnel environment: An efficient method","authors":"Zhipeng Wang, M. Kagesawa, Shintaro Ono, A. Banno, K. Ikeuchi","doi":"10.1109/ACPR.2011.6166528","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166528","url":null,"abstract":"Automobile navigation in tunnel environment is challenging. GPS sensors and ordinary cameras can't function effectively. For navigation, infrared cameras are installed on top of our experimental vehicle, and here we propose an efficient object detection method to detect emergency lights from the collected data in tunnel environment. The proposed method firstly detects keypoints by setting thresholds for intensity of uniformly sampled points. Each keypoint is then verified by the appearance of its surrounding sub-image. After clustering the keypoints which satisfy the verification, the method verifies the keypoint clusters by their appearance and temporal information. Though the later steps are time-consuming, they deal with very few instances. And this improves the efficiency of the method, while not losing effectiveness of the appearance and temporal information. Thus the method gives promising results in real time. Detection performance and efficiency are verified by experiments carried on challenging real data.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122887984","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Real-time lip reading system for fixed phrase and its combination 固定短语及其组合的实时唇读系统
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166595
T. Saitoh
This paper presents a high-level real-time lip reading system that can recognize both fixed phrase and its combination. Lip reading provides an important means for realizing a communication support interface for speech handicaps. The approach is based on the Viola-Jones face detector, lip extraction based on active appearance model, automatic utterance section detector, and phrase classifier using DP matching. Our system has two unique ideas; combined phrase recognition, and remove false recognition phrase from a target list. These ideas are simple but efficient in practical use. We demonstrated our system with five subjects and confirmed the usefulness.
本文提出了一种高水平的实时唇读系统,可以同时识别固定短语及其组合。唇读是实现语言障碍交流支持界面的重要手段。该方法基于Viola-Jones人脸检测器、基于主动外观模型的唇形提取、自动话语片段检测器和基于DP匹配的短语分类器。我们的系统有两个独特的想法;组合短语识别,并从目标列表中删除错误的识别短语。这些想法简单,但在实际应用中很有效。我们用5个受试者演示了我们的系统,并证实了它的实用性。
{"title":"Real-time lip reading system for fixed phrase and its combination","authors":"T. Saitoh","doi":"10.1109/ACPR.2011.6166595","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166595","url":null,"abstract":"This paper presents a high-level real-time lip reading system that can recognize both fixed phrase and its combination. Lip reading provides an important means for realizing a communication support interface for speech handicaps. The approach is based on the Viola-Jones face detector, lip extraction based on active appearance model, automatic utterance section detector, and phrase classifier using DP matching. Our system has two unique ideas; combined phrase recognition, and remove false recognition phrase from a target list. These ideas are simple but efficient in practical use. We demonstrated our system with five subjects and confirmed the usefulness.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128762869","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Image mosaic using Log-polar binning 使用对数极分割的图像拼接
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166693
Le Thanh Hoan, Youngjae Chun, Kyoungsu Oh
Image mosaic is a large image assembled from many smaller tiles which one tile itself is an actual image. In this research, we introduce an efficient method to make image mosaic. Our method is based on Log-polar mapping which enables us to detect the color and shape change. We also successfully make an image mosaic version by exploiting GPU power. Our algorithm is simple, easy to implement, gives better result than conventional method and can be improved to higher precision.
图像拼接是由许多较小的瓷砖拼接而成的大图像,其中一个瓷砖本身就是一个实际的图像。本文介绍了一种有效的图像拼接方法。我们的方法是基于对数极坐标映射,使我们能够检测颜色和形状的变化。我们还成功地利用GPU的能力制作了图像拼接版本。该算法简单,易于实现,结果优于传统方法,可提高精度。
{"title":"Image mosaic using Log-polar binning","authors":"Le Thanh Hoan, Youngjae Chun, Kyoungsu Oh","doi":"10.1109/ACPR.2011.6166693","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166693","url":null,"abstract":"Image mosaic is a large image assembled from many smaller tiles which one tile itself is an actual image. In this research, we introduce an efficient method to make image mosaic. Our method is based on Log-polar mapping which enables us to detect the color and shape change. We also successfully make an image mosaic version by exploiting GPU power. Our algorithm is simple, easy to implement, gives better result than conventional method and can be improved to higher precision.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129394249","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Improvements to facial contour detection by hierarchical fitting and regression 基于层次拟合和回归的人脸轮廓检测改进
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166689
Atsushi Irie, M. Takagiwa, Kozo Moriyama, Takayoshi Yamashita
There are many methods based on shape and texture models for detecting eye and mouth contour points from facial images. They reduce the false positive rate by utilizing a global model and adapting it for a given face. Changes to facial expressions are coupled with changes to the shapes of eyes and mouth, and a global facial model in itself cannot be adapted to all human facial expressions. Therefore, a hierarchical model fitting approach has been developed, whereby the global fitting captures the facial shape using the global model and the local fitting captures the each facial parts using these local models. This can detect facial contours with high accuracy for expressions to which the global model cannot be adapted.
基于形状和纹理模型的眼、口轮廓点检测方法有很多。他们利用一个全局模型,并根据给定的人脸进行调整,从而降低了误报率。面部表情的变化伴随着眼睛和嘴的形状的变化,一个全局的面部模型本身不能适应所有的人类面部表情。因此,开发了一种分层模型拟合方法,其中全局拟合使用全局模型捕获面部形状,局部拟合使用这些局部模型捕获每个面部部分。对于全局模型无法适应的表情,这种方法可以高精度地检测出面部轮廓。
{"title":"Improvements to facial contour detection by hierarchical fitting and regression","authors":"Atsushi Irie, M. Takagiwa, Kozo Moriyama, Takayoshi Yamashita","doi":"10.1109/ACPR.2011.6166689","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166689","url":null,"abstract":"There are many methods based on shape and texture models for detecting eye and mouth contour points from facial images. They reduce the false positive rate by utilizing a global model and adapting it for a given face. Changes to facial expressions are coupled with changes to the shapes of eyes and mouth, and a global facial model in itself cannot be adapted to all human facial expressions. Therefore, a hierarchical model fitting approach has been developed, whereby the global fitting captures the facial shape using the global model and the local fitting captures the each facial parts using these local models. This can detect facial contours with high accuracy for expressions to which the global model cannot be adapted.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126902199","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
Decision fusion for block linear regression classification based on confidence index 基于置信度指标的块线性回归分类决策融合
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166708
Yi-fei Xu, He-lei Wu
We consider the problem of recognizing human faces with varying expression and illumination, and a novel confidence index based block linear regression classification method is proposed. Our approach divides images into blocks, and each block is identified using the linear regression classifier separately. We develop a confidence index model to measure the recognition confidence of each block, and the final decision is achieved by aggregating individual results with the designed Bayesian decision fusion algorithm. The performances of our approach and conventional algorithms are evaluated under conditions of varying expression and illumination using benchmark databases, improvements demonstrate the proposed approach is robustness to both expression and illumination variations.
针对不同表情和光照的人脸识别问题,提出了一种基于置信度指数的块线性回归分类方法。我们的方法将图像分成块,每个块分别使用线性回归分类器进行识别。我们建立了一个置信度指数模型来衡量每个块的识别置信度,并利用所设计的贝叶斯决策融合算法对单个结果进行汇总,从而得到最终的决策。在不同的表达和光照条件下,使用基准数据库对我们的方法和传统算法的性能进行了评估,改进表明我们的方法对表达和光照变化都具有鲁棒性。
{"title":"Decision fusion for block linear regression classification based on confidence index","authors":"Yi-fei Xu, He-lei Wu","doi":"10.1109/ACPR.2011.6166708","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166708","url":null,"abstract":"We consider the problem of recognizing human faces with varying expression and illumination, and a novel confidence index based block linear regression classification method is proposed. Our approach divides images into blocks, and each block is identified using the linear regression classifier separately. We develop a confidence index model to measure the recognition confidence of each block, and the final decision is achieved by aggregating individual results with the designed Bayesian decision fusion algorithm. The performances of our approach and conventional algorithms are evaluated under conditions of varying expression and illumination using benchmark databases, improvements demonstrate the proposed approach is robustness to both expression and illumination variations.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125790024","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
The First Asian Conference on Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1