首页 > 最新文献

2013 2nd IAPR Asian Conference on Pattern Recognition最新文献

英文 中文
A Robust and Efficient Minutia-Based Fingerprint Matching Algorithm 一种鲁棒高效的基于细节的指纹匹配算法
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.43
Wen Wen, Zhi Qi, Zhi Li, Junhao Zhang, Yuguang Gong, Peng Cao
In this paper, we propose a novel robust and efficient minutia-based fingerprint matching algorithm. There are two key contributions. First, we apply a set of global level minutia dependent features, i.e., the qualities that measure the reliabilities of the extracted minutiae and the area of overlapping regions between the query and template images of fingerprints. The implementation of these easy-to-get minutia dependent features presents coherence to the well-accepted fingerprint template standards. Besides, the reasonable combination of them results in the robustness to poor quality fingerprint images. Second, we implement a hierarchical recognition strategy, which applies a procedure of global matching that refines the local matching decision towards a genuine result over the entire images. Other than the much improved accuracy, our algorithm also promotes the efficiency, because compared with other state-of-the-art matching approaches, it does not make use of any time-consuming operations or any complex feature structures. The experimental results demonstrate the proposed method exhibits an excellent accuracy that exceeds the performances of well-known minutia based matchers. Meanwhile, the proposed algorithm presents potentials to serve a real-time fingerprint recognition system.
本文提出了一种鲁棒高效的基于细节的指纹匹配算法。有两个关键贡献。首先,我们应用了一组全局级别的细节依赖特征,即衡量提取细节的可靠性和指纹查询图像与模板图像之间重叠区域面积的质量。这些易于获得的细节依赖特征的实现与公认的指纹模板标准具有一致性。此外,它们的合理组合使其对质量较差的指纹图像具有较强的鲁棒性。其次,我们实现了一种分层识别策略,该策略应用全局匹配过程,将局部匹配决策细化到整个图像的真实结果。与其他最先进的匹配方法相比,我们的算法不仅提高了精度,而且提高了效率,因为它不使用任何耗时的操作或任何复杂的特征结构。实验结果表明,所提出的方法具有优异的精度,超过了已知的基于细节的匹配器的性能。同时,该算法具有服务于实时指纹识别系统的潜力。
{"title":"A Robust and Efficient Minutia-Based Fingerprint Matching Algorithm","authors":"Wen Wen, Zhi Qi, Zhi Li, Junhao Zhang, Yuguang Gong, Peng Cao","doi":"10.1109/ACPR.2013.43","DOIUrl":"https://doi.org/10.1109/ACPR.2013.43","url":null,"abstract":"In this paper, we propose a novel robust and efficient minutia-based fingerprint matching algorithm. There are two key contributions. First, we apply a set of global level minutia dependent features, i.e., the qualities that measure the reliabilities of the extracted minutiae and the area of overlapping regions between the query and template images of fingerprints. The implementation of these easy-to-get minutia dependent features presents coherence to the well-accepted fingerprint template standards. Besides, the reasonable combination of them results in the robustness to poor quality fingerprint images. Second, we implement a hierarchical recognition strategy, which applies a procedure of global matching that refines the local matching decision towards a genuine result over the entire images. Other than the much improved accuracy, our algorithm also promotes the efficiency, because compared with other state-of-the-art matching approaches, it does not make use of any time-consuming operations or any complex feature structures. The experimental results demonstrate the proposed method exhibits an excellent accuracy that exceeds the performances of well-known minutia based matchers. Meanwhile, the proposed algorithm presents potentials to serve a real-time fingerprint recognition system.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"48 82","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114088056","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Mammary Gland Tumor Detection in Cats Using Ant Colony Optimisation 基于蚁群优化的猫乳腺肿瘤检测
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.173
Hossam M. Moftah, Mohammad Ibrahim, A. Hassanien, G. Schaefer
Mammary gland tumors are among the most common tumors in cats. Over 85 percent of mammary tumors in cats are malignant and they tend to grow and metastasize quickly to different organs like lungs and lymph nodes. Similar to breast tumors in humans, they start as a small lump in a mammary gland and then grow and increase in size unless detected and treated. In this paper, we present an approach to detect broadenoma mammary gland tumors in cats using ant colony optimisation. Image features can then be extracted from the segmented image regions. To evaluate the performance of our presented approach, 25 microscopical images were taken from tissue slides of broadenomas from three cat cases. The experimental results obtained confirm that the effectiveness and performance of the proposed system is high.
乳腺肿瘤是猫体内最常见的肿瘤之一。超过85%的猫乳腺肿瘤是恶性的,它们倾向于迅速生长和转移到不同的器官,如肺和淋巴结。与人类的乳腺肿瘤类似,它们一开始只是乳腺中的一个小肿块,如果不及时发现和治疗,就会逐渐增大。在本文中,我们提出了一种方法来检测宽腺瘤乳腺肿瘤的猫使用蚁群优化。然后可以从分割的图像区域中提取图像特征。为了评估我们所提出的方法的性能,我们从三个猫病例的阔腺瘤的组织载玻片上取下了25张显微图像。实验结果表明,该系统具有较高的有效性和性能。
{"title":"Mammary Gland Tumor Detection in Cats Using Ant Colony Optimisation","authors":"Hossam M. Moftah, Mohammad Ibrahim, A. Hassanien, G. Schaefer","doi":"10.1109/ACPR.2013.173","DOIUrl":"https://doi.org/10.1109/ACPR.2013.173","url":null,"abstract":"Mammary gland tumors are among the most common tumors in cats. Over 85 percent of mammary tumors in cats are malignant and they tend to grow and metastasize quickly to different organs like lungs and lymph nodes. Similar to breast tumors in humans, they start as a small lump in a mammary gland and then grow and increase in size unless detected and treated. In this paper, we present an approach to detect broadenoma mammary gland tumors in cats using ant colony optimisation. Image features can then be extracted from the segmented image regions. To evaluate the performance of our presented approach, 25 microscopical images were taken from tissue slides of broadenomas from three cat cases. The experimental results obtained confirm that the effectiveness and performance of the proposed system is high.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121567544","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Bag-of-Words Against Nearest-Neighbor Search for Visual Object Retrieval 针对最近邻搜索的词袋视觉对象检索
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.56
Cai-Zhi Zhu, Xiaoping Zhou, S. Satoh
We compare the Bag-of-Words (BoW) framework with the Approximate Nearest-Neighbor (ANN) based system in the context of visual object retrieval. This comparison is motivated by the implicit connection between these two methods: generally speaking, the BoW framework can be regarded as a quantization-guided ANN voting system. The value of establishing such comparison lies in: first, by comparing with other quantization-free ANN system, the performance loss caused by the quantization error in the BoW framework can be estimated quantitatively. Second, this comparison completely inspects the pros and cons of both ANN and BoW methods, thus to facilitate new algorithm design. In this study, by taking an independent dataset as the reference to validate matches, we design an ANN voting system that outperforms all other methods. Comprehensive and computationally intensive experiments are conducted on two Oxford datasets and two TrecVid instance search datasets, and the new state-of-the-art is achieved.
在视觉对象检索的背景下,我们比较了词袋(BoW)框架和基于近似最近邻(ANN)的系统。这种比较的动机是这两种方法之间的隐式联系:一般来说,BoW框架可以被看作是一个量化引导的ANN投票系统。建立这种比较的价值在于:首先,通过与其他无量化的人工神经网络系统进行比较,可以定量地估计BoW框架中量化误差造成的性能损失。其次,这种比较全面地考察了ANN和BoW方法的优缺点,从而便于新的算法设计。在本研究中,我们以一个独立的数据集作为验证匹配的参考,设计了一个优于所有其他方法的ANN投票系统。在两个Oxford数据集和两个trevid实例搜索数据集上进行了全面和计算密集型的实验,并实现了新的状态。
{"title":"Bag-of-Words Against Nearest-Neighbor Search for Visual Object Retrieval","authors":"Cai-Zhi Zhu, Xiaoping Zhou, S. Satoh","doi":"10.1109/ACPR.2013.56","DOIUrl":"https://doi.org/10.1109/ACPR.2013.56","url":null,"abstract":"We compare the Bag-of-Words (BoW) framework with the Approximate Nearest-Neighbor (ANN) based system in the context of visual object retrieval. This comparison is motivated by the implicit connection between these two methods: generally speaking, the BoW framework can be regarded as a quantization-guided ANN voting system. The value of establishing such comparison lies in: first, by comparing with other quantization-free ANN system, the performance loss caused by the quantization error in the BoW framework can be estimated quantitatively. Second, this comparison completely inspects the pros and cons of both ANN and BoW methods, thus to facilitate new algorithm design. In this study, by taking an independent dataset as the reference to validate matches, we design an ANN voting system that outperforms all other methods. Comprehensive and computationally intensive experiments are conducted on two Oxford datasets and two TrecVid instance search datasets, and the new state-of-the-art is achieved.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121644210","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Automatic Segmentation and Classification of Liver Abnormalities Using Fractal Dimension 基于分形维数的肝脏异常自动分割与分类
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.172
A. Anter, A. Hassanien, G. Schaefer
Abnormalities in the liver include masses which can be benign or malignant. Due to the presence of these abnormalities, the regularity of the liver structure is altered, which changes its fractal dimension. In this paper, we present a computer aided diagnostic system for classifying liver abnormalities from abdominal CT images using fractal dimension features. We integrate different methods for liver segmentation and abnormality classification and propose an attempt that combines different techniques in order to compensate their individual weaknesses and to exploit their strengths. Classification is based on fractal dimension, with six different features being employed for extracted regions of interest. Experimental results confirm that our approach is robust, fast and able to effectively detect the presence of abnormalities in the liver.
肝脏异常包括肿块,可为良性或恶性。由于这些异常的存在,肝脏结构的规律性被改变,从而改变了其分形维数。本文提出一种利用分形维数特征对腹部CT图像进行肝脏异常分类的计算机辅助诊断系统。我们将不同的肝脏分割和异常分类方法进行整合,提出一种结合不同技术的尝试,以弥补各自的不足,发挥各自的优势。分类是基于分形维数,与六个不同的特征被用于提取感兴趣的区域。实验结果证实,我们的方法是稳健的,快速的,能够有效地检测肝脏异常的存在。
{"title":"Automatic Segmentation and Classification of Liver Abnormalities Using Fractal Dimension","authors":"A. Anter, A. Hassanien, G. Schaefer","doi":"10.1109/ACPR.2013.172","DOIUrl":"https://doi.org/10.1109/ACPR.2013.172","url":null,"abstract":"Abnormalities in the liver include masses which can be benign or malignant. Due to the presence of these abnormalities, the regularity of the liver structure is altered, which changes its fractal dimension. In this paper, we present a computer aided diagnostic system for classifying liver abnormalities from abdominal CT images using fractal dimension features. We integrate different methods for liver segmentation and abnormality classification and propose an attempt that combines different techniques in order to compensate their individual weaknesses and to exploit their strengths. Classification is based on fractal dimension, with six different features being employed for extracted regions of interest. Experimental results confirm that our approach is robust, fast and able to effectively detect the presence of abnormalities in the liver.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114694295","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Aircraft Detection by Deep Belief Nets 基于深度信念网的飞机检测
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.5
Xueyun Chen, Shiming Xiang, Cheng-Lin Liu, Chunhong Pan
Aircraft detection is a difficult task in high-resolution remote sensing images, due to the variable sizes, colors, orientations and complex backgrounds. In this paper, an effective aircraft detection method is proposed which exactly locates the object by outputting its geometric center, orientation, position. To reduce the influence of background, multi-images including gradient image and gray thresholding images of the object were input to a Deep Belief Net (DBN), which was pre-trained first to learn features and later fine-tuned by back-propagation to yield a robust detector. Experimental results show that DBNs can detecte the tiny blurred aircrafts correctly in many difficult airport images, DBNs outperform the traditional Feature Classifier methods in robustness and accuracy, and the multi-images help improve the detection precision of DBN than using only single-image.
在高分辨率遥感图像中,由于图像的大小、颜色、方向和背景复杂,飞机检测是一项艰巨的任务。本文提出了一种有效的飞机检测方法,通过输出目标的几何中心、方位、位置来精确定位目标。为了减少背景的影响,将包括梯度图像和灰度阈值图像在内的多幅图像输入到深度信念网络(DBN)中,该网络首先进行预训练以学习特征,然后通过反向传播进行微调以产生鲁棒检测器。实验结果表明,DBN能够在多幅难度较大的机场图像中正确检测出微小的模糊飞机,在鲁棒性和准确率上都优于传统的特征分类器方法,多幅图像比单幅图像更有助于提高DBN的检测精度。
{"title":"Aircraft Detection by Deep Belief Nets","authors":"Xueyun Chen, Shiming Xiang, Cheng-Lin Liu, Chunhong Pan","doi":"10.1109/ACPR.2013.5","DOIUrl":"https://doi.org/10.1109/ACPR.2013.5","url":null,"abstract":"Aircraft detection is a difficult task in high-resolution remote sensing images, due to the variable sizes, colors, orientations and complex backgrounds. In this paper, an effective aircraft detection method is proposed which exactly locates the object by outputting its geometric center, orientation, position. To reduce the influence of background, multi-images including gradient image and gray thresholding images of the object were input to a Deep Belief Net (DBN), which was pre-trained first to learn features and later fine-tuned by back-propagation to yield a robust detector. Experimental results show that DBNs can detecte the tiny blurred aircrafts correctly in many difficult airport images, DBNs outperform the traditional Feature Classifier methods in robustness and accuracy, and the multi-images help improve the detection precision of DBN than using only single-image.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132237081","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 57
Emotional Speech Recognition Using Acoustic Models of Decomposed Component Words 基于分解成分词声学模型的情绪语音识别
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.13
Vivatchai Kaveeta, K. Patanukhom
This paper presents a novel approach for emotional speech recognition. Instead of using a full length of speech for classification, the proposed method decomposes speech signals into component words, groups the words into segments and generates an acoustic model for each segment by using features such as audio power, MFCC, log attack time, spectrum spread and segment duration. Based on the proposed segment-based classification, unknown speech signals can be recognized into sequences of segment emotions. Emotion profiles (EPs) are extracted from the emotion sequences. Finally, speech emotion can be determined by using EP as features. Experiments are conducted by using 6,810 training samples and 722 test samples which are composed of eight emotional classes from IEMOCAP database. In comparison with a conventional method, the proposed method can improve recognition rate from 46.81% to 58.59% in eight emotion classification and from 60.18% to 71.25% in four emotion classification.
提出了一种新的情感语音识别方法。该方法不使用完整的语音长度进行分类,而是将语音信号分解为成分词,将词分组为段,并利用音频功率、MFCC、日志攻击时间、频谱扩展和段持续时间等特征为每个段生成声学模型。基于所提出的基于片段的分类方法,可以将未知语音信号识别为片段情绪序列。从情感序列中提取情感轮廓(EPs)。最后,利用EP作为特征来确定语音情绪。实验采用IEMOCAP数据库中的6810个训练样本和722个测试样本组成的8个情绪类。与传统方法相比,该方法在8种情绪分类中将识别率从46.81%提高到58.59%,在4种情绪分类中将识别率从60.18%提高到71.25%。
{"title":"Emotional Speech Recognition Using Acoustic Models of Decomposed Component Words","authors":"Vivatchai Kaveeta, K. Patanukhom","doi":"10.1109/ACPR.2013.13","DOIUrl":"https://doi.org/10.1109/ACPR.2013.13","url":null,"abstract":"This paper presents a novel approach for emotional speech recognition. Instead of using a full length of speech for classification, the proposed method decomposes speech signals into component words, groups the words into segments and generates an acoustic model for each segment by using features such as audio power, MFCC, log attack time, spectrum spread and segment duration. Based on the proposed segment-based classification, unknown speech signals can be recognized into sequences of segment emotions. Emotion profiles (EPs) are extracted from the emotion sequences. Finally, speech emotion can be determined by using EP as features. Experiments are conducted by using 6,810 training samples and 722 test samples which are composed of eight emotional classes from IEMOCAP database. In comparison with a conventional method, the proposed method can improve recognition rate from 46.81% to 58.59% in eight emotion classification and from 60.18% to 71.25% in four emotion classification.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131752889","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Gaze Estimation in Children's Peer-Play Scenarios 儿童同伴游戏情境中的凝视估计
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.178
Dingrui Duan, Lu Tian, J. Cui, Li Wang, H. Zha, H. Aghajan
Gaze is a powerful cue for children's social behavior analysis. In this paper, a novel method is proposed to estimate children's gaze orientation in the experimental data of developmental psychology based on head pose estimation. In consideration of the possible errors of head pose estimation results, temporal information and potential targets are both introduced to improve the results of gaze estimation. At last, this method is evaluated by a dataset of children's peer-play scenarios and the results show that this method has a good performance. According to the experimental valuation and analysis, in a certain peer-play scenario, potential targets are powerful spatial cues for children's gaze estimation and temporal information also provides some cues to improve the estimation results.
凝视是儿童社会行为分析的有力线索。本文提出了一种基于头部姿态估计的发展心理学实验数据中儿童注视方向估计方法。考虑到头部姿态估计结果可能存在的误差,引入时间信息和潜在目标信息对注视估计结果进行改进。最后,通过一个儿童同伴游戏场景数据集对该方法进行了评价,结果表明该方法具有良好的性能。根据实验评估和分析,在一定的同伴游戏场景下,潜在目标是儿童注视估计的强大空间线索,而时间信息也为提高儿童注视估计结果提供了一些线索。
{"title":"Gaze Estimation in Children's Peer-Play Scenarios","authors":"Dingrui Duan, Lu Tian, J. Cui, Li Wang, H. Zha, H. Aghajan","doi":"10.1109/ACPR.2013.178","DOIUrl":"https://doi.org/10.1109/ACPR.2013.178","url":null,"abstract":"Gaze is a powerful cue for children's social behavior analysis. In this paper, a novel method is proposed to estimate children's gaze orientation in the experimental data of developmental psychology based on head pose estimation. In consideration of the possible errors of head pose estimation results, temporal information and potential targets are both introduced to improve the results of gaze estimation. At last, this method is evaluated by a dataset of children's peer-play scenarios and the results show that this method has a good performance. According to the experimental valuation and analysis, in a certain peer-play scenario, potential targets are powerful spatial cues for children's gaze estimation and temporal information also provides some cues to improve the estimation results.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130887038","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Transparent Text Detection and Background Recovery 透明文本检测和背景恢复
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.57
Xinhao Liu, N. Chiba
We propose two methods for detecting transparent text in images and recovering the background behind the text. Although text detection in natural scenes is an active research area, most current methods are focused on non-transparent text. To detect transparent text, we developed an adaptive edge detection method for edge-based text detection that can accurately detect text even under low contrast, which is common among transparent text images and a method for recovering the original background content behind the detected transparent text. Experiments using real images show the effectiveness of the proposed methods.
我们提出了两种检测图像中透明文本和恢复文本背后背景的方法。虽然自然场景中的文本检测是一个活跃的研究领域,但目前大多数方法都集中在非透明文本上。为了检测透明文本,我们开发了一种基于边缘的文本检测的自适应边缘检测方法,即使在低对比度下也能准确检测出透明文本图像中常见的文本,以及一种恢复被检测透明文本背后原始背景内容的方法。实际图像实验表明了所提方法的有效性。
{"title":"Transparent Text Detection and Background Recovery","authors":"Xinhao Liu, N. Chiba","doi":"10.1109/ACPR.2013.57","DOIUrl":"https://doi.org/10.1109/ACPR.2013.57","url":null,"abstract":"We propose two methods for detecting transparent text in images and recovering the background behind the text. Although text detection in natural scenes is an active research area, most current methods are focused on non-transparent text. To detect transparent text, we developed an adaptive edge detection method for edge-based text detection that can accurately detect text even under low contrast, which is common among transparent text images and a method for recovering the original background content behind the detected transparent text. Experiments using real images show the effectiveness of the proposed methods.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"77 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133562602","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Direct Ego-Motion Estimation Using Normal Flows 使用正常流的直接自我运动估计
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.130
Ding Yuan, Miao Liu, Hong Zhang
In this paper we present a novel method that estimates the motion parameters of a monocular camera, which is under unconstrained movement. Different from the traditional works which tackle the problem by establishing motion correspondences, or by calculating optical flows within the image sequence, the proposed method estimates the motion parameters directly by using the information of spatio-temporal gradient of the image intensity. Hence, our method requires no specific assumptions about the captured scene, like it is smooth almost everywhere or it must contain distinct features etc. We have tested the methods on both synthetic image data and real image sequences. Experimental results show that the developed methods are effective in determining the camera motion parameters.
本文提出了一种估计无约束运动单目摄像机运动参数的新方法。与传统的通过建立运动对应关系或计算图像序列内的光流来解决问题不同,该方法直接利用图像强度的时空梯度信息来估计运动参数。因此,我们的方法不需要对捕获的场景进行特定的假设,比如它几乎到处都是光滑的,或者它必须包含明显的特征等。我们已经在合成图像数据和真实图像序列上进行了测试。实验结果表明,所提出的方法在确定摄像机运动参数方面是有效的。
{"title":"Direct Ego-Motion Estimation Using Normal Flows","authors":"Ding Yuan, Miao Liu, Hong Zhang","doi":"10.1109/ACPR.2013.130","DOIUrl":"https://doi.org/10.1109/ACPR.2013.130","url":null,"abstract":"In this paper we present a novel method that estimates the motion parameters of a monocular camera, which is under unconstrained movement. Different from the traditional works which tackle the problem by establishing motion correspondences, or by calculating optical flows within the image sequence, the proposed method estimates the motion parameters directly by using the information of spatio-temporal gradient of the image intensity. Hence, our method requires no specific assumptions about the captured scene, like it is smooth almost everywhere or it must contain distinct features etc. We have tested the methods on both synthetic image data and real image sequences. Experimental results show that the developed methods are effective in determining the camera motion parameters.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133092510","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Real-Time Foreground Segmentation from Moving Camera Based on Case-Based Trajectory Classification 基于案例轨迹分类的运动摄像机实时前景分割
Pub Date : 2013-11-05 DOI: 10.1109/ACPR.2013.146
Yosuke Nonaka, Atsushi Shimada, H. Nagahara, R. Taniguchi
Recently, several methods for foreground segmentation from moving camera have been proposed. A trajectory-based method is one of typical approaches to segment video frames into foreground and background regions. The method obtains long term trajectories from entire of video frame and segments them by learning pixel or motion based object features. However, it often needs large amount of computational cost and memory resource to maintain trajectories. We present a trajectory-based method which aims for real-time foreground segmentation from moving camera. Unlike conventional methods, we use trajectories which are sparsely obtained from two successive video frames. In addition, our method enables using spatio-temporal feature of trajectories by introducing case-based approach to improve detection results. We compare our method with previous approaches and show results on challenging video sequences.
近年来,人们提出了几种运动相机前景分割的方法。基于轨迹的方法是将视频帧分割为前景和背景区域的典型方法之一。该方法从整个视频帧中获取长期轨迹,并通过学习基于像素或运动的目标特征对其进行分割。然而,它通常需要大量的计算成本和内存资源来维持轨迹。提出了一种基于运动轨迹的前景实时分割方法。与传统方法不同,我们使用从两个连续视频帧稀疏获得的轨迹。此外,我们的方法通过引入基于案例的方法来提高检测结果,从而利用轨迹的时空特征。我们将我们的方法与以前的方法进行比较,并在具有挑战性的视频序列上显示结果。
{"title":"Real-Time Foreground Segmentation from Moving Camera Based on Case-Based Trajectory Classification","authors":"Yosuke Nonaka, Atsushi Shimada, H. Nagahara, R. Taniguchi","doi":"10.1109/ACPR.2013.146","DOIUrl":"https://doi.org/10.1109/ACPR.2013.146","url":null,"abstract":"Recently, several methods for foreground segmentation from moving camera have been proposed. A trajectory-based method is one of typical approaches to segment video frames into foreground and background regions. The method obtains long term trajectories from entire of video frame and segments them by learning pixel or motion based object features. However, it often needs large amount of computational cost and memory resource to maintain trajectories. We present a trajectory-based method which aims for real-time foreground segmentation from moving camera. Unlike conventional methods, we use trajectories which are sparsely obtained from two successive video frames. In addition, our method enables using spatio-temporal feature of trajectories by introducing case-based approach to improve detection results. We compare our method with previous approaches and show results on challenging video sequences.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131696696","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
期刊
2013 2nd IAPR Asian Conference on Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1