首页 > 最新文献

The First Asian Conference on Pattern Recognition最新文献

英文 中文
Wavelet-domain image shrinkage using variance field diffusion 利用方差场扩散的小波域图像收缩
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166660
Zhenyu Liu, Jing Tian, Li Chen, Yongtao Wang
Wavelet shrinkage is an image denoising technique based on the concept of thresholding the wavelet coefficients. The key challenge of wavelet shrinkage is to find an appropriate threshold value, which is typically controlled by the signal variance. To tackle this challenge, a new image shrinkage approach is proposed in this paper by using a variance field diffusion, which can provide more accurate variance estimation. Experimental results are provided to demonstrate the superior performance of the proposed approach.
小波收缩是一种基于小波系数阈值化的图像去噪技术。小波压缩的关键挑战是找到一个合适的阈值,该阈值通常由信号方差控制。为了解决这一问题,本文提出了一种新的图像收缩方法,该方法使用方差场扩散,可以提供更准确的方差估计。实验结果证明了该方法的优越性。
{"title":"Wavelet-domain image shrinkage using variance field diffusion","authors":"Zhenyu Liu, Jing Tian, Li Chen, Yongtao Wang","doi":"10.1109/ACPR.2011.6166660","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166660","url":null,"abstract":"Wavelet shrinkage is an image denoising technique based on the concept of thresholding the wavelet coefficients. The key challenge of wavelet shrinkage is to find an appropriate threshold value, which is typically controlled by the signal variance. To tackle this challenge, a new image shrinkage approach is proposed in this paper by using a variance field diffusion, which can provide more accurate variance estimation. Experimental results are provided to demonstrate the superior performance of the proposed approach.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114696561","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sparse Representations, Compressive Sensing and dictionaries for pattern recognition 稀疏表示、压缩感知和模式识别字典
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166711
Vishal M. Patel, R. Chellappa
In recent years, the theories of Compressive Sensing (CS), Sparse Representation (SR) and Dictionary Learning (DL) have emerged as powerful tools for efficiently processing data in non-traditional ways. An area of promise for these theories is object recognition. In this paper, we review the role of SR, CS and DL for object recognition. Algorithms to perform object recognition using these theories are reviewed. An important aspect in object recognition is feature extraction. Recent works in SR and CS have shown that if sparsity in the recognition problem is properly harnessed then the choice of features is less critical. What becomes critical, however, is the number of features and the sparsity of representation. This issue is discussed in detail.
近年来,压缩感知(CS)、稀疏表示(SR)和字典学习(DL)等理论已成为以非传统方式高效处理数据的强大工具。这些理论的一个有希望的领域是物体识别。本文综述了SR、CS和DL在物体识别中的作用。算法执行对象识别使用这些理论进行了审查。特征提取是目标识别的一个重要方面。最近在SR和CS方面的工作表明,如果识别问题中的稀疏性得到适当利用,那么特征的选择就不那么重要了。然而,至关重要的是特征的数量和表示的稀疏性。对这个问题进行了详细的讨论。
{"title":"Sparse Representations, Compressive Sensing and dictionaries for pattern recognition","authors":"Vishal M. Patel, R. Chellappa","doi":"10.1109/ACPR.2011.6166711","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166711","url":null,"abstract":"In recent years, the theories of Compressive Sensing (CS), Sparse Representation (SR) and Dictionary Learning (DL) have emerged as powerful tools for efficiently processing data in non-traditional ways. An area of promise for these theories is object recognition. In this paper, we review the role of SR, CS and DL for object recognition. Algorithms to perform object recognition using these theories are reviewed. An important aspect in object recognition is feature extraction. Recent works in SR and CS have shown that if sparsity in the recognition problem is properly harnessed then the choice of features is less critical. What becomes critical, however, is the number of features and the sparsity of representation. This issue is discussed in detail.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"127 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124006093","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 79
Automatic generation of training samples and a learning method based on advanced MILBoost for human detection 训练样本的自动生成及基于高级MILBoost的人体检测学习方法
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166556
Yuji Yamauchi, H. Fujiyoshi
Statistical learning methods for human detection require large quantities of training samples and thus suffer from high sample collection costs. Their detection performance is also liable to be lower when the training samples are collected in a different environment than the one in which the detection system must operate. In this paper we propose a generative learning method that uses the automatic generation of training samples from 3D models together with an advanced MILBoost learning algorithm. In this study, we use a three-dimensional human model to automatically generate positive samples for learning specialized to specific scenes. Negative training samples are collected by random automatic extraction from video stream, but some of these samples may be collected with incorrect labeling. When a classifier is trained by statistical learning using incorrectly labeled training samples, detection performance is impaired. Therefore, in this study an improved version of MILBoost is used to perform generative learning which is immune to the adverse effects of incorrectly labeled samples among the training samples. In evaluation, we found that a classifier trained using training samples generated from a 3D human model was capable of better detection performance than a classifier trained using training samples extracted by hand. The proposed method can also mitigate the degradation of detection performance when there are image of people mixed in with the negative samples used for learning.
用于人体检测的统计学习方法需要大量的训练样本,因此样本收集成本高。当训练样本在不同于检测系统必须运行的环境中收集时,它们的检测性能也容易降低。在本文中,我们提出了一种生成式学习方法,该方法使用3D模型自动生成训练样本以及先进的MILBoost学习算法。在本研究中,我们使用三维人体模型自动生成用于特定场景学习的正样本。从视频流中随机自动提取负训练样本,但其中一些样本可能会被错误的标记所收集。当分类器使用不正确标记的训练样本进行统计学习训练时,检测性能会受到损害。因此,在本研究中,使用改进版本的MILBoost进行生成式学习,该学习不受训练样本中错误标记样本的不利影响。在评估中,我们发现使用从3D人体模型生成的训练样本训练的分类器比使用手工提取的训练样本训练的分类器具有更好的检测性能。该方法还可以缓解学习用的负样本中混入人的图像对检测性能的影响。
{"title":"Automatic generation of training samples and a learning method based on advanced MILBoost for human detection","authors":"Yuji Yamauchi, H. Fujiyoshi","doi":"10.1109/ACPR.2011.6166556","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166556","url":null,"abstract":"Statistical learning methods for human detection require large quantities of training samples and thus suffer from high sample collection costs. Their detection performance is also liable to be lower when the training samples are collected in a different environment than the one in which the detection system must operate. In this paper we propose a generative learning method that uses the automatic generation of training samples from 3D models together with an advanced MILBoost learning algorithm. In this study, we use a three-dimensional human model to automatically generate positive samples for learning specialized to specific scenes. Negative training samples are collected by random automatic extraction from video stream, but some of these samples may be collected with incorrect labeling. When a classifier is trained by statistical learning using incorrectly labeled training samples, detection performance is impaired. Therefore, in this study an improved version of MILBoost is used to perform generative learning which is immune to the adverse effects of incorrectly labeled samples among the training samples. In evaluation, we found that a classifier trained using training samples generated from a 3D human model was capable of better detection performance than a classifier trained using training samples extracted by hand. The proposed method can also mitigate the degradation of detection performance when there are image of people mixed in with the negative samples used for learning.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116943553","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Hierarchical orthogonal matching pursuit for face recognition 人脸识别的层次正交匹配追踪
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166530
Huaping Liu, F. Sun
This paper tries to exploit the joint group intrinsics in face recognition problem by using sparse representation with multiple features. We claim that different feature vectors of one test face image share the same sparsity pattern at the higher group level, but not necessarily at the lower (inside the group) level. This means that they share the same active groups, but not necessarily the same active set. To this end, a hierarchical orthogonal matching pursuit algorithm is developed. The basic idea of this approach is straightforward: At each iteration step, we first select the best group which is shared by different features, then we select the best atoms (within this group) for each feature. This algorithm is very efficient and shows good performance in standard face recognition dataset.
本文试图利用多特征稀疏表示来开发人脸识别问题中的联合群特征。我们声称,一个测试人脸图像的不同特征向量在较高的组水平上共享相同的稀疏模式,但不一定在较低的(组内)水平上共享相同的稀疏模式。这意味着它们共享相同的活动组,但不一定是相同的活动集。为此,提出了一种分层正交匹配追踪算法。这种方法的基本思想很简单:在每个迭代步骤中,我们首先选择由不同特征共享的最佳组,然后为每个特征选择(在该组内的)最佳原子。该算法在标准人脸识别数据集上显示出良好的性能。
{"title":"Hierarchical orthogonal matching pursuit for face recognition","authors":"Huaping Liu, F. Sun","doi":"10.1109/ACPR.2011.6166530","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166530","url":null,"abstract":"This paper tries to exploit the joint group intrinsics in face recognition problem by using sparse representation with multiple features. We claim that different feature vectors of one test face image share the same sparsity pattern at the higher group level, but not necessarily at the lower (inside the group) level. This means that they share the same active groups, but not necessarily the same active set. To this end, a hierarchical orthogonal matching pursuit algorithm is developed. The basic idea of this approach is straightforward: At each iteration step, we first select the best group which is shared by different features, then we select the best atoms (within this group) for each feature. This algorithm is very efficient and shows good performance in standard face recognition dataset.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122565467","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Universal no reference image quality assessment metrics based on local dependency 通用的无参考图像质量评价指标基于局部依赖性
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166657
Fei Gao, Xinbo Gao, D. Tao, Xuelong Li, Lihuo He, Wen Lu
No reference image quality assessment (NR-IQA) is to evaluate image quality blindly without the ground truth. Most of the emerging NR-IQA algorithms are only effective for some specific distortion. Universal metrics that can work for various categories of distortions have hardly been explored, and the algorithms available are not fully adequate in performance. In this paper, we study the local dependency (LD) characteristic of natural images, and propose two universal NR-IQA metrics: LD global scheme (LD-GS) and LD two-step scheme (LD-TS). We claim that the local dependency characteristic among wavelet coefficients is disturbed by various distortion processes, and the disturbances are strongly correlated to image qualities. Experimental results on LIVE database II demonstrate that both the proposed metrics are highly consistent with the human perception and outpace the state-of-the-art NR-IQA indexes and some full reference quality indicators for diverse distortions and across the entire database.
无参考图像质量评价(NR-IQA)是在不了解实际情况的情况下对图像质量进行盲目评价。大多数新兴的NR-IQA算法仅对某些特定的失真有效。通用的指标,可以工作的各种类别的扭曲几乎没有被探索,和算法可用的不是完全足够的性能。本文研究了自然图像的局部依赖性(LD)特征,提出了两个通用的NR-IQA度量:LD全局方案(LD- gs)和LD两步方案(LD- ts)。我们认为小波系数之间的局部依赖特性会受到各种失真过程的干扰,并且这些干扰与图像质量有很强的相关性。在LIVE数据库II上的实验结果表明,所提出的两个指标与人类感知高度一致,并且超过了最先进的NR-IQA指标和一些完整的参考质量指标,适用于各种失真和整个数据库。
{"title":"Universal no reference image quality assessment metrics based on local dependency","authors":"Fei Gao, Xinbo Gao, D. Tao, Xuelong Li, Lihuo He, Wen Lu","doi":"10.1109/ACPR.2011.6166657","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166657","url":null,"abstract":"No reference image quality assessment (NR-IQA) is to evaluate image quality blindly without the ground truth. Most of the emerging NR-IQA algorithms are only effective for some specific distortion. Universal metrics that can work for various categories of distortions have hardly been explored, and the algorithms available are not fully adequate in performance. In this paper, we study the local dependency (LD) characteristic of natural images, and propose two universal NR-IQA metrics: LD global scheme (LD-GS) and LD two-step scheme (LD-TS). We claim that the local dependency characteristic among wavelet coefficients is disturbed by various distortion processes, and the disturbances are strongly correlated to image qualities. Experimental results on LIVE database II demonstrate that both the proposed metrics are highly consistent with the human perception and outpace the state-of-the-art NR-IQA indexes and some full reference quality indicators for diverse distortions and across the entire database.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123331150","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Robust moving object segmentation with two-stage optimization 基于两阶段优化的鲁棒运动目标分割
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166695
Jianwei Ding, Xin Zhao, Kaiqi Huang, T. Tan
Inspired by interactive segmentation algorithms, we propose an online and unsupervised technique to extract moving objects from videos captured by stationary cameras. Our method consists of two main optimization steps, from local optimal extraction to global optimal segmentation. In the first stage, reliable foreground and background pixels are extracted from input image by modeling distributions of foreground and background with color and motion cues. These reliable pixels provide hard constraints for the next step of segmentation. Then global optimal segmentation of moving object is implemented by graph cuts in the second stage. Experimental results on several challenging videos demonstrate the effectiveness and robustness of the proposed approach.
受交互式分割算法的启发,我们提出了一种在线和无监督的技术,从固定摄像机捕获的视频中提取运动物体。我们的方法包括两个主要的优化步骤,从局部最优提取到全局最优分割。在第一阶段,通过使用颜色和运动线索对前景和背景分布进行建模,从输入图像中提取可靠的前景和背景像素。这些可靠的像素为下一步的分割提供了严格的约束。第二阶段通过图割实现运动目标的全局最优分割。在几个具有挑战性的视频上的实验结果证明了该方法的有效性和鲁棒性。
{"title":"Robust moving object segmentation with two-stage optimization","authors":"Jianwei Ding, Xin Zhao, Kaiqi Huang, T. Tan","doi":"10.1109/ACPR.2011.6166695","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166695","url":null,"abstract":"Inspired by interactive segmentation algorithms, we propose an online and unsupervised technique to extract moving objects from videos captured by stationary cameras. Our method consists of two main optimization steps, from local optimal extraction to global optimal segmentation. In the first stage, reliable foreground and background pixels are extracted from input image by modeling distributions of foreground and background with color and motion cues. These reliable pixels provide hard constraints for the next step of segmentation. Then global optimal segmentation of moving object is implemented by graph cuts in the second stage. Experimental results on several challenging videos demonstrate the effectiveness and robustness of the proposed approach.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123592048","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Saliency model based head pose estimation by sparse optical flow 基于显著性模型的稀疏光流头姿估计
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166668
Tao Xu, Chao Wang, Yunhong Wang, Zhaoxiang Zhang
Head pose plays an important role in Human-Computer interaction, and its estimation is a challenge problem compared to face detection and recognition in computer vision. In this paper, a novel and efficient method is proposed to estimate head pose in real-time video sequences. A saliency model based segmentation method is used not only to extract feature points of face, but also to update and rectify the location of feature points when missing happened. This step also gives a benchmark for vector generation in pose estimation. In subsequent frames feature points will be tracked by sparse optical flow method and head pose can be determined from vectors generated by feature points between successive frames. Via a voting scheme, these vectors with angle and length can give a robust estimation of the head pose. Compared with other methods, annotated training data set and training procedure is not essential in our method. Initialization and re-initialization can be done automatically and are robust for profile head pose. Experimental results show an efficient and robust estimation of the head pose.
头部姿态在人机交互中起着重要的作用,与人脸检测和识别相比,头部姿态的估计是计算机视觉中一个具有挑战性的问题。本文提出了一种新的、高效的实时视频序列头部姿态估计方法。基于显著性模型的分割方法不仅可以提取人脸的特征点,而且可以在特征点缺失时更新和校正特征点的位置。这一步也为姿态估计中的矢量生成提供了一个基准。在后续的帧中,使用稀疏光流方法跟踪特征点,并根据连续帧之间特征点生成的向量确定头部姿态。通过一种投票方案,这些带有角度和长度的向量可以给出头部姿态的鲁棒估计。与其他方法相比,我们的方法不需要标注训练数据集和训练过程。初始化和重新初始化可以自动完成,并且对轮廓头姿态具有鲁棒性。实验结果表明,该方法对头部姿态进行了有效的鲁棒估计。
{"title":"Saliency model based head pose estimation by sparse optical flow","authors":"Tao Xu, Chao Wang, Yunhong Wang, Zhaoxiang Zhang","doi":"10.1109/ACPR.2011.6166668","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166668","url":null,"abstract":"Head pose plays an important role in Human-Computer interaction, and its estimation is a challenge problem compared to face detection and recognition in computer vision. In this paper, a novel and efficient method is proposed to estimate head pose in real-time video sequences. A saliency model based segmentation method is used not only to extract feature points of face, but also to update and rectify the location of feature points when missing happened. This step also gives a benchmark for vector generation in pose estimation. In subsequent frames feature points will be tracked by sparse optical flow method and head pose can be determined from vectors generated by feature points between successive frames. Via a voting scheme, these vectors with angle and length can give a robust estimation of the head pose. Compared with other methods, annotated training data set and training procedure is not essential in our method. Initialization and re-initialization can be done automatically and are robust for profile head pose. Experimental results show an efficient and robust estimation of the head pose.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128992857","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
An efficient coarse-to-fine scheme for text detection in videos 一种高效的视频文本检测粗到精方案
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166605
Liuan Wang, Lin-Lin Huang, Yang Wu
To achieve fast and accurate text detection from videos, we propose an efficient coarse-to-fine scheme comprising three stages: key frame extraction, candidate text line detection and fine text detection. Key frames, which are assumed to carry texts, are extracted based on multi-threshold difference of color histogram (MDCH). From the key frames, candidate text lines are detected by morphological operations and connected component analysis. Sliding window classification is performed on the candidate text lines so as to detect refined text lines. We use two types of features: histogram of gradients (HOG) and local assembled binary (LAB), and two classifiers: Real Adaboost and polynomial neural network (PNN), for improving the classification accuracy. The effectiveness of the proposed method has been demonstrated by the experiment results on a large video dataset. Also, the benefits of key frame extraction and combining multiple features and classifiers have been justified.
为了实现快速准确的视频文本检测,我们提出了一种高效的从粗到精的方案,包括三个阶段:关键帧提取、候选文本行检测和精细文本检测。基于多阈值颜色直方图差分(MDCH)提取假定携带文本的关键帧。从关键帧中,通过形态学操作和连接成分分析检测候选文本行。对候选文本行进行滑动窗口分类,以检测精细文本行。为了提高分类精度,我们使用了梯度直方图(HOG)和局部组合二值(LAB)两种特征,以及Real Adaboost和多项式神经网络(PNN)两种分类器。在大型视频数据集上的实验结果验证了该方法的有效性。此外,关键帧提取和组合多个特征和分类器的好处也得到了证明。
{"title":"An efficient coarse-to-fine scheme for text detection in videos","authors":"Liuan Wang, Lin-Lin Huang, Yang Wu","doi":"10.1109/ACPR.2011.6166605","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166605","url":null,"abstract":"To achieve fast and accurate text detection from videos, we propose an efficient coarse-to-fine scheme comprising three stages: key frame extraction, candidate text line detection and fine text detection. Key frames, which are assumed to carry texts, are extracted based on multi-threshold difference of color histogram (MDCH). From the key frames, candidate text lines are detected by morphological operations and connected component analysis. Sliding window classification is performed on the candidate text lines so as to detect refined text lines. We use two types of features: histogram of gradients (HOG) and local assembled binary (LAB), and two classifiers: Real Adaboost and polynomial neural network (PNN), for improving the classification accuracy. The effectiveness of the proposed method has been demonstrated by the experiment results on a large video dataset. Also, the benefits of key frame extraction and combining multiple features and classifiers have been justified.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126904256","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Current trend in natural disaster warning systems based on computer vision techniques 基于计算机视觉技术的自然灾害预警系统的发展趋势
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166591
ByoungChul Ko, Joon-Young Kwak, June-Hyeok Hong, J. Nam
In this paper, a review of vision-based natural disaster warning methods is presented. Because natural disaster warning is receiving a lot of attention in recent research, a comprehensive review of various disaster-warning techniques developed in recent years is needed. This paper surveys recent studies on warning systems four different types of natural disaster, i.e., wildfire smoke and flame detection, water level detection for flood prevention, and coastal zone monitoring, using computer vision and pattern-recognition techniques. Finally, we conclude with some thoughts about future research directions.
本文对基于视觉的自然灾害预警方法进行了综述。由于自然灾害预警在近年来的研究中受到了广泛的关注,因此有必要对近年来发展起来的各种灾害预警技术进行全面的回顾。本文综述了近年来利用计算机视觉和模式识别技术在野火烟雾和火焰探测、防洪水位探测和海岸带监测等四种不同类型自然灾害预警系统方面的研究进展。最后,对今后的研究方向进行了展望。
{"title":"Current trend in natural disaster warning systems based on computer vision techniques","authors":"ByoungChul Ko, Joon-Young Kwak, June-Hyeok Hong, J. Nam","doi":"10.1109/ACPR.2011.6166591","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166591","url":null,"abstract":"In this paper, a review of vision-based natural disaster warning methods is presented. Because natural disaster warning is receiving a lot of attention in recent research, a comprehensive review of various disaster-warning techniques developed in recent years is needed. This paper surveys recent studies on warning systems four different types of natural disaster, i.e., wildfire smoke and flame detection, water level detection for flood prevention, and coastal zone monitoring, using computer vision and pattern-recognition techniques. Finally, we conclude with some thoughts about future research directions.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125999631","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Automatic image cropping using sparse coding 使用稀疏编码自动图像裁剪
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166623
Jieying She, Duo-Chao Wang, Mingli Song
Image cropping is a technique to help people improve their taken photos' quality by discarding unnecessary parts of a photo. In this paper, we propose a new approach to crop the photo for better composition through learning the structure. Firstly, we classify photos into different categories. Then we extract the graph-based visual saliency map of these photos, based on which we build a dictionary for each categories. Finally, by solving the sparse coding problem of each input photo based on the dictionary, we find a cropped region that can be best decoded by this dictionary. The experimental results demonstrate that our technique is applicable to a wide range of photos and produce more agreeable resulting photos.
图像裁剪是一种通过去除照片中不必要的部分来帮助人们提高照片质量的技术。在本文中,我们提出了一种新的方法,通过学习结构来裁剪照片以获得更好的构图。首先,我们把照片分成不同的类别。然后,我们提取这些照片的基于图形的视觉显著性图,并在此基础上为每个类别构建字典。最后,通过基于字典对每张输入照片进行稀疏编码,找到一个裁剪后的区域,该区域可以被该字典进行最佳解码。实验结果表明,我们的技术适用于更广泛的照片,并产生更令人满意的结果照片。
{"title":"Automatic image cropping using sparse coding","authors":"Jieying She, Duo-Chao Wang, Mingli Song","doi":"10.1109/ACPR.2011.6166623","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166623","url":null,"abstract":"Image cropping is a technique to help people improve their taken photos' quality by discarding unnecessary parts of a photo. In this paper, we propose a new approach to crop the photo for better composition through learning the structure. Firstly, we classify photos into different categories. Then we extract the graph-based visual saliency map of these photos, based on which we build a dictionary for each categories. Finally, by solving the sparse coding problem of each input photo based on the dictionary, we find a cropped region that can be best decoded by this dictionary. The experimental results demonstrate that our technique is applicable to a wide range of photos and produce more agreeable resulting photos.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"73 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127220822","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
期刊
The First Asian Conference on Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1