首页 > 最新文献

2013 Fourth National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics (NCVPRIPG)最新文献

英文 中文
Stochastic resonance aided robust techniques for segmentation of medical ultrasound images 随机共振辅助鲁棒医学超声图像分割技术
J. V. Sagar, C. Bhagvati
The existence of stochastic resonance has been demonstrated in physical, biological and geological systems for boosting weak signals to make them detectable. Narrow regions, small features and low-contrast or subtle edges, in noisy images, correspond to such weak signals. In this paper, the occurrence and exploitation of stochastic resonance in the detection, extraction and analysis of such features is demonstrated both mathematically and empirically. The mathematical results are confirmed by simulation studies. Finally, results on medical ultrasound images demonstrate that several subtle features lost by the application of robust techniques such as mean shift filter are recovered by stochastic resonance. These results reconfirm the mathematical and simulation findings.
随机共振的存在已经在物理、生物和地质系统中得到证明,它可以增强弱信号,使其可被探测到。在有噪声的图像中,狭窄的区域、小的特征和低对比度或微妙的边缘对应于这样的弱信号。本文从数学和经验两方面论证了随机共振在这些特征的检测、提取和分析中的发生和利用。仿真研究证实了数学结果。最后,对医学超声图像的研究结果表明,随机共振可以恢复由于平均移位滤波器等鲁棒技术而丢失的一些细微特征。这些结果再次证实了数学和模拟结果。
{"title":"Stochastic resonance aided robust techniques for segmentation of medical ultrasound images","authors":"J. V. Sagar, C. Bhagvati","doi":"10.1109/NCVPRIPG.2013.6776209","DOIUrl":"https://doi.org/10.1109/NCVPRIPG.2013.6776209","url":null,"abstract":"The existence of stochastic resonance has been demonstrated in physical, biological and geological systems for boosting weak signals to make them detectable. Narrow regions, small features and low-contrast or subtle edges, in noisy images, correspond to such weak signals. In this paper, the occurrence and exploitation of stochastic resonance in the detection, extraction and analysis of such features is demonstrated both mathematically and empirically. The mathematical results are confirmed by simulation studies. Finally, results on medical ultrasound images demonstrate that several subtle features lost by the application of robust techniques such as mean shift filter are recovered by stochastic resonance. These results reconfirm the mathematical and simulation findings.","PeriodicalId":436402,"journal":{"name":"2013 Fourth National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics (NCVPRIPG)","volume":"49 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115751374","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fast area of contact computation for collision detection of a deformable object using FEM 基于有限元法的可变形物体碰撞检测的快速接触面积计算
P. Shrivastava, Sukhendu Das
In case of detection and analysis of deformation in collision scenarios, using a method based on area of contact instead of a point of contact generates numerically stable impulse forces. Area of contact improves the stability of control algorithms, but it is often associated with high computational cost. In this paper, we alleviate this problem by proposing a novel algorithm for collision detection of a deformable mesh against rigid structures. We reuse the data structures maintained for elastic force computations in the FEM, for the purpose of collision detection. Parallel constructs on GPU using reduced model make the simulations interactive even for meshes with thousands of elements. Since we don't maintain any additional complex structure for keeping track of the deformable body at each iteration, we significantly reduce the usage of GPU memory bandwidth. Efficiency of our method is illustrated by reporting high culling efficiency on various tests.
在碰撞变形的检测和分析中,使用基于接触面积而不是接触点的方法可以产生数值稳定的冲力。接触面积提高了控制算法的稳定性,但往往伴随着较高的计算成本。在本文中,我们通过提出一种新的算法来缓解这个问题,该算法用于变形网格与刚性结构的碰撞检测。为了碰撞检测的目的,我们重用了FEM中弹性力计算的数据结构。采用简化模型在GPU上并行构建,使仿真即使对具有数千个单元的网格也具有交互性。因为我们不需要在每次迭代中维护任何额外的复杂结构来跟踪可变形的物体,所以我们大大减少了GPU内存带宽的使用。通过报告各种测试的高剔除效率来说明我们的方法的效率。
{"title":"Fast area of contact computation for collision detection of a deformable object using FEM","authors":"P. Shrivastava, Sukhendu Das","doi":"10.1109/NCVPRIPG.2013.6776197","DOIUrl":"https://doi.org/10.1109/NCVPRIPG.2013.6776197","url":null,"abstract":"In case of detection and analysis of deformation in collision scenarios, using a method based on area of contact instead of a point of contact generates numerically stable impulse forces. Area of contact improves the stability of control algorithms, but it is often associated with high computational cost. In this paper, we alleviate this problem by proposing a novel algorithm for collision detection of a deformable mesh against rigid structures. We reuse the data structures maintained for elastic force computations in the FEM, for the purpose of collision detection. Parallel constructs on GPU using reduced model make the simulations interactive even for meshes with thousands of elements. Since we don't maintain any additional complex structure for keeping track of the deformable body at each iteration, we significantly reduce the usage of GPU memory bandwidth. Efficiency of our method is illustrated by reporting high culling efficiency on various tests.","PeriodicalId":436402,"journal":{"name":"2013 Fourth National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics (NCVPRIPG)","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123559768","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
OD-Match: PatchMatch based Optic Disk detection OD-Match:基于PatchMatch的光盘检测
S. Ramakanth, R. Venkatesh Babu
Approximate Nearest-Neighbour Field has been an area of interest in recent research for a wide variety of topics in graphics and multimedia community. Medical image processing is a relatively unaffected field by these developments in ANNF computations, brought about by various extremely efficient algorithms like PatchMatch. In this paper, we use Generalized PatchMatch for Optic Disk detection, in retinal images, and show that by making use of efficient ANNF computations we are able to generate results with 98% accuracy with an average time of 0.5 sec. This is significantly faster than conventional Optic Disk detection methods, which average at 95-97% accuracy with 3-5 sec average computation time.
在图形和多媒体社区的各种主题中,近似近邻域一直是最近研究的一个感兴趣的领域。医学图像处理是一个相对不受神经神经网络计算发展影响的领域,这些发展是由各种极其高效的算法(如PatchMatch)带来的。在本文中,我们在视网膜图像中使用广义PatchMatch进行视盘检测,并表明通过使用高效的ANNF计算,我们能够在平均0.5秒的时间内生成准确率为98%的结果。这比传统的视盘检测方法要快得多,传统的视盘检测方法平均准确率为95-97%,平均计算时间为3-5秒。
{"title":"OD-Match: PatchMatch based Optic Disk detection","authors":"S. Ramakanth, R. Venkatesh Babu","doi":"10.1109/NCVPRIPG.2013.6776206","DOIUrl":"https://doi.org/10.1109/NCVPRIPG.2013.6776206","url":null,"abstract":"Approximate Nearest-Neighbour Field has been an area of interest in recent research for a wide variety of topics in graphics and multimedia community. Medical image processing is a relatively unaffected field by these developments in ANNF computations, brought about by various extremely efficient algorithms like PatchMatch. In this paper, we use Generalized PatchMatch for Optic Disk detection, in retinal images, and show that by making use of efficient ANNF computations we are able to generate results with 98% accuracy with an average time of 0.5 sec. This is significantly faster than conventional Optic Disk detection methods, which average at 95-97% accuracy with 3-5 sec average computation time.","PeriodicalId":436402,"journal":{"name":"2013 Fourth National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics (NCVPRIPG)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126763682","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Symmetry based 3D reconstruction of repeated cylinders 基于对称的重复圆柱三维重建
Adersh Miglani, Sumantra Dutta Roy, S. Chaudhury, J. B. Srivastava
First, we describe how 360°-rotational symmetry may be used for three dimensional reconstruction of repeated cylinders from a single perspective image. In our experiments, we consider translational and affine repetition of cylinders with vertical and random orientations. Later, we create a virtual camera configuration for retrieving pose and location of repeated cylinders. The combination of 360°-rotational symmetry and camera center is used to identify two orthogonal planes called axis plane and orthogonal axis plane. These two planes are the basis for the proposed reconstruction framework and virtual camera configuration. Furthermore, we discuss possible extension of our method in vision tasks based on motion analysis.
首先,我们描述了360°旋转对称如何用于从单一视角图像重复圆柱体的三维重建。在我们的实验中,我们考虑了垂直和随机方向的柱体的平移和仿射重复。随后,我们创建了一个虚拟摄像机配置,用于检索重复圆柱体的姿态和位置。利用360°旋转对称和相机中心的结合来识别两个正交平面,称为轴平面和正交轴平面。这两个平面是所提出的重建框架和虚拟摄像机配置的基础。此外,我们还讨论了该方法在基于运动分析的视觉任务中的可能扩展。
{"title":"Symmetry based 3D reconstruction of repeated cylinders","authors":"Adersh Miglani, Sumantra Dutta Roy, S. Chaudhury, J. B. Srivastava","doi":"10.1109/NCVPRIPG.2013.6776266","DOIUrl":"https://doi.org/10.1109/NCVPRIPG.2013.6776266","url":null,"abstract":"First, we describe how 360°-rotational symmetry may be used for three dimensional reconstruction of repeated cylinders from a single perspective image. In our experiments, we consider translational and affine repetition of cylinders with vertical and random orientations. Later, we create a virtual camera configuration for retrieving pose and location of repeated cylinders. The combination of 360°-rotational symmetry and camera center is used to identify two orthogonal planes called axis plane and orthogonal axis plane. These two planes are the basis for the proposed reconstruction framework and virtual camera configuration. Furthermore, we discuss possible extension of our method in vision tasks based on motion analysis.","PeriodicalId":436402,"journal":{"name":"2013 Fourth National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics (NCVPRIPG)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121815271","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
A novel framework for multi-focus image fusion 一种新的多焦点图像融合框架
G. Bhatnagar, Q.M. Jonathan Wu
One of the foremost requisite for human perception and computer vision task is to get an image with all objects in focus. The image fusion process, as one of the solutions, allows getting a clear fused image from several images acquired with different focus levels of a scene. In this paper, a novel framework for multi-focus image fusion is proposed, which is computationally simple since it realizes only in the spatial domain. The proposed framework is based on the fractal dimensions of the images into the fusion process. The extensive experiments on different multi-focus image sets demonstrate that it is consistently superior to the conventional image fusion methods in terms of visual and quantitative evaluations.
人类感知和计算机视觉任务的首要条件之一是获得所有物体聚焦的图像。图像融合处理作为一种解决方案,可以从一个场景的不同焦点级别获取的多幅图像中获得清晰的融合图像。本文提出了一种新的多焦点图像融合框架,该框架只在空间域中实现,计算简单。所提出的框架是基于分形维数对图像进行融合处理。在不同多焦点图像集上的大量实验表明,该方法在视觉和定量评价方面始终优于传统的图像融合方法。
{"title":"A novel framework for multi-focus image fusion","authors":"G. Bhatnagar, Q.M. Jonathan Wu","doi":"10.1109/NCVPRIPG.2013.6776177","DOIUrl":"https://doi.org/10.1109/NCVPRIPG.2013.6776177","url":null,"abstract":"One of the foremost requisite for human perception and computer vision task is to get an image with all objects in focus. The image fusion process, as one of the solutions, allows getting a clear fused image from several images acquired with different focus levels of a scene. In this paper, a novel framework for multi-focus image fusion is proposed, which is computationally simple since it realizes only in the spatial domain. The proposed framework is based on the fractal dimensions of the images into the fusion process. The extensive experiments on different multi-focus image sets demonstrate that it is consistently superior to the conventional image fusion methods in terms of visual and quantitative evaluations.","PeriodicalId":436402,"journal":{"name":"2013 Fourth National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics (NCVPRIPG)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117094666","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Perceptual video hashing based on the Achlioptas's random projections 基于Achlioptas随机投影的感知视频散列
R. Sandeep, P. Bora
A perceptual video hashing function maps the perceptual content of a video into a fixed-length binary string called the perceptual hash. Perceptual hashing is a promising solution to the content-identification and the content-authentication problems. The projections of image and video data onto a subspace have been exploited in the literature to get a compact hash function. We propose a new perceptual video hashing algorithm based on the Achlioptas's random projections. Simulation results show that the proposed perceptual hash function is robust to common signal and image processing attacks.
感知视频哈希函数将视频的感知内容映射到一个固定长度的二进制字符串,称为感知哈希。感知哈希是解决内容识别和内容认证问题的一种很有前途的方法。图像和视频数据在子空间上的投影在文献中被用来得到一个紧哈希函数。提出了一种基于Achlioptas随机投影的感知视频哈希算法。仿真结果表明,所提出的感知哈希函数对常见的信号和图像处理攻击具有较强的鲁棒性。
{"title":"Perceptual video hashing based on the Achlioptas's random projections","authors":"R. Sandeep, P. Bora","doi":"10.1109/NCVPRIPG.2013.6776252","DOIUrl":"https://doi.org/10.1109/NCVPRIPG.2013.6776252","url":null,"abstract":"A perceptual video hashing function maps the perceptual content of a video into a fixed-length binary string called the perceptual hash. Perceptual hashing is a promising solution to the content-identification and the content-authentication problems. The projections of image and video data onto a subspace have been exploited in the literature to get a compact hash function. We propose a new perceptual video hashing algorithm based on the Achlioptas's random projections. Simulation results show that the proposed perceptual hash function is robust to common signal and image processing attacks.","PeriodicalId":436402,"journal":{"name":"2013 Fourth National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics (NCVPRIPG)","volume":"44 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124769019","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Source color error analysis for robust separation of reflection components 反射分量鲁棒分离的源色误差分析
S. Biswas, K. Shafique
In this paper, we address the problem of separating the diffuse and specular reflection components of complex textured surfaces from a single color image. Unlike most previous approaches that assume accurate knowledge of illumination source color for this task, we analyze errors in source color information to perform robust separation. The analysis leads to a simple, efficient and robust algorithm to estimate the diffuse and specular components using the estimated source color. The algorithm is completely automatic and does not need explicit color segmentation or color boundary detection as required by many existing methods. Results on complex textured images show the effectiveness of the proposed algorithm for robust reflection component separation.
在本文中,我们解决了从单色图像中分离复杂纹理表面的漫反射和镜面反射组件的问题。与之前大多数假设光源颜色准确的方法不同,我们分析光源颜色信息中的误差来执行鲁棒分离。通过分析得出一种简单、高效和鲁棒的算法,利用估计的光源颜色来估计漫反射和反射分量。该算法是完全自动化的,不需要像许多现有方法那样需要明确的颜色分割或颜色边界检测。在复杂纹理图像上的实验结果表明了该算法对鲁棒反射分量分离的有效性。
{"title":"Source color error analysis for robust separation of reflection components","authors":"S. Biswas, K. Shafique","doi":"10.1109/NCVPRIPG.2013.6776253","DOIUrl":"https://doi.org/10.1109/NCVPRIPG.2013.6776253","url":null,"abstract":"In this paper, we address the problem of separating the diffuse and specular reflection components of complex textured surfaces from a single color image. Unlike most previous approaches that assume accurate knowledge of illumination source color for this task, we analyze errors in source color information to perform robust separation. The analysis leads to a simple, efficient and robust algorithm to estimate the diffuse and specular components using the estimated source color. The algorithm is completely automatic and does not need explicit color segmentation or color boundary detection as required by many existing methods. Results on complex textured images show the effectiveness of the proposed algorithm for robust reflection component separation.","PeriodicalId":436402,"journal":{"name":"2013 Fourth National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics (NCVPRIPG)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121193460","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Real time anomaly detection in H.264 compressed videos 实时异常检测在H.264压缩视频
Sovan Biswas, R. Venkatesh Babu
Real time anomaly detection is the need of the hour for any security applications. In this paper, we have proposed a real-time anomaly detection algorithm by utilizing cues from the motion vectors in H.264/AVC compressed domain. The discussed work is principally motivated by the observation that motion vectors (MVs) exhibit different characteristics during anomaly. We have observed that H.264 motion vector magnitude contains relevant information which can be used to model the usual behavior (UB) effectively. This is subsequently extended to detect abnormality/anomaly based on the probability of occurrence of a behavior. Additionally, we have suggested a hierarchical approach through Motion Pyramid for High Resolution videos to further increase the detection rate. The proposed algorithm has performed extremely well on UMN and Peds Anomaly Detection Video datasets, with a detection speed of >150 and 65-75 frames per sec in respective datasets resulting in more than 200× speedup along with comparable accuracy to pixel domain state-of-the-art algorithms.
实时异常检测是任何安全应用程序都需要的。在本文中,我们提出了一种利用H.264/AVC压缩域中的运动矢量线索的实时异常检测算法。讨论的工作主要是由于观察到运动向量(mv)在异常期间表现出不同的特征。我们已经观察到H.264运动矢量幅度包含了可以用来有效地模拟通常行为(UB)的相关信息。这随后扩展到基于行为发生的概率来检测异常/异常。此外,我们还提出了一种通过运动金字塔对高分辨率视频进行分层的方法,以进一步提高检测率。所提出的算法在UMN和Peds异常检测视频数据集上表现非常好,在各自的数据集上的检测速度为bb0 150和65-75帧/秒,导致超过200倍的加速以及与像素域最先进算法相当的精度。
{"title":"Real time anomaly detection in H.264 compressed videos","authors":"Sovan Biswas, R. Venkatesh Babu","doi":"10.1109/NCVPRIPG.2013.6776164","DOIUrl":"https://doi.org/10.1109/NCVPRIPG.2013.6776164","url":null,"abstract":"Real time anomaly detection is the need of the hour for any security applications. In this paper, we have proposed a real-time anomaly detection algorithm by utilizing cues from the motion vectors in H.264/AVC compressed domain. The discussed work is principally motivated by the observation that motion vectors (MVs) exhibit different characteristics during anomaly. We have observed that H.264 motion vector magnitude contains relevant information which can be used to model the usual behavior (UB) effectively. This is subsequently extended to detect abnormality/anomaly based on the probability of occurrence of a behavior. Additionally, we have suggested a hierarchical approach through Motion Pyramid for High Resolution videos to further increase the detection rate. The proposed algorithm has performed extremely well on UMN and Peds Anomaly Detection Video datasets, with a detection speed of >150 and 65-75 frames per sec in respective datasets resulting in more than 200× speedup along with comparable accuracy to pixel domain state-of-the-art algorithms.","PeriodicalId":436402,"journal":{"name":"2013 Fourth National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics (NCVPRIPG)","volume":"70 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126793925","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 50
Script independent detection of bold words in multi font-size documents 在多种字体大小的文档中,独立于脚本的粗体字检测
P. Saikrishna, A. Ramakrishnan
A script independent, font-size independent scheme is proposed for detecting bold words in printed pages. In OCR applications such as minor modifications of an existing printed form, it is desirable to reproduce the font size and characteristics such as bold, and italics in the OCR recognized document. In this morphological opening based detection of bold (MOBDoB) method, the binarized image is segmented into sub-images with uniform font sizes, using the word height information. Rough estimation of the stroke widths of characters in each sub-image is obtained from the density. Each sub-image is then opened with a square structuring element of size determined by the respective stroke width. The union of all the opened sub-images is used to determine the locations of the bold words. Extracting all such words from the binarized image gives the final image. A minimum of 98 % of bold words were detected from a total of 65 Tamil, Kannada and English pages and the false alarm rate is less than 0.4 %.
提出了一种与文字无关、字体大小无关的打印页面粗体字检测方案。在OCR应用程序中,例如对现有打印表单进行微小修改,需要在OCR识别的文档中重现字体大小和特征,例如粗体和斜体。在基于形态学开度的黑体检测(MOBDoB)方法中,利用词高信息将二值化后的图像分割成具有统一字体大小的子图像。从密度中得到各子图像中字符笔画宽度的粗略估计。然后用正方形结构元素打开每个子图像,其大小由各自的笔画宽度决定。所有打开的子图像的并集用于确定加粗单词的位置。从二值化后的图像中提取所有这些词就得到了最终的图像。从总共65个泰米尔语、卡纳达语和英语页面中,至少检测出98%的粗体单词,误报率低于0.4%。
{"title":"Script independent detection of bold words in multi font-size documents","authors":"P. Saikrishna, A. Ramakrishnan","doi":"10.1109/NCVPRIPG.2013.6776180","DOIUrl":"https://doi.org/10.1109/NCVPRIPG.2013.6776180","url":null,"abstract":"A script independent, font-size independent scheme is proposed for detecting bold words in printed pages. In OCR applications such as minor modifications of an existing printed form, it is desirable to reproduce the font size and characteristics such as bold, and italics in the OCR recognized document. In this morphological opening based detection of bold (MOBDoB) method, the binarized image is segmented into sub-images with uniform font sizes, using the word height information. Rough estimation of the stroke widths of characters in each sub-image is obtained from the density. Each sub-image is then opened with a square structuring element of size determined by the respective stroke width. The union of all the opened sub-images is used to determine the locations of the bold words. Extracting all such words from the binarized image gives the final image. A minimum of 98 % of bold words were detected from a total of 65 Tamil, Kannada and English pages and the false alarm rate is less than 0.4 %.","PeriodicalId":436402,"journal":{"name":"2013 Fourth National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics (NCVPRIPG)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125984775","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Indian Movie Face Database: A benchmark for face recognition under wide variations 印度电影人脸数据库:广泛变化下的人脸识别基准
S. Setty, M. Husain, Parisa Beham, Jyothi Gudavalli, Menaka Kandasamy, R. Vaddi, V. Hemadri, J C Karure, Raja Raju, B. Rajan, Vijay Kumar, C V Jawahar
Recognizing human faces in the wild is emerging as a critically important, and technically challenging computer vision problem. With a few notable exceptions, most previous works in the last several decades have focused on recognizing faces captured in a laboratory setting. However, with the introduction of databases such as LFW and Pubfigs, face recognition community is gradually shifting its focus on much more challenging unconstrained settings. Since its introduction, LFW verification benchmark is getting a lot of attention with various researchers contributing towards state-of-the-results. To further boost the unconstrained face recognition research, we introduce a more challenging Indian Movie Face Database (IMFDB) that has much more variability compared to LFW and Pubfigs. The database consists of 34512 faces of 100 known actors collected from approximately 103 Indian movies. Unlike LFW and Pubfigs which used face detectors to automatically detect the faces from the web collection, faces in IMFDB are detected manually from all the movies. Manual selection of faces from movies resulted in high degree of variability (in scale, pose, expression, illumination, age, occlusion, makeup) which one could ever see in natural world. IMFDB is the first face database that provides a detailed annotation in terms of age, pose, gender, expression, amount of occlusion, for each face which may help other face related applications.
在野外识别人脸正在成为一个极其重要的、技术上具有挑战性的计算机视觉问题。除了少数值得注意的例外,在过去的几十年里,大多数先前的工作都集中在识别在实验室环境中捕捉到的人脸。然而,随着LFW和Pubfigs等数据库的引入,人脸识别社区正逐渐将重点转向更具挑战性的无约束环境。自引入以来,LFW验证基准得到了许多研究人员的关注,他们对状态-结果做出了贡献。为了进一步推动无约束人脸识别研究,我们引入了一个更具挑战性的印度电影人脸数据库(IMFDB),与LFW和Pubfigs相比,它具有更多的可变性。该数据库包括从大约103部印度电影中收集的100位已知演员的34512张脸。与LFW和Pubfigs使用人脸检测器自动从网络集合中检测人脸不同,IMFDB中的人脸是从所有电影中手动检测的。从电影中手动选择人脸导致了高度的可变性(在规模,姿势,表情,照明,年龄,遮挡,化妆),这在自然界中是可以看到的。IMFDB是第一个为每张脸提供年龄、姿势、性别、表情、遮挡量等详细注释的人脸数据库,这可能有助于其他与人脸相关的应用。
{"title":"Indian Movie Face Database: A benchmark for face recognition under wide variations","authors":"S. Setty, M. Husain, Parisa Beham, Jyothi Gudavalli, Menaka Kandasamy, R. Vaddi, V. Hemadri, J C Karure, Raja Raju, B. Rajan, Vijay Kumar, C V Jawahar","doi":"10.1109/NCVPRIPG.2013.6776225","DOIUrl":"https://doi.org/10.1109/NCVPRIPG.2013.6776225","url":null,"abstract":"Recognizing human faces in the wild is emerging as a critically important, and technically challenging computer vision problem. With a few notable exceptions, most previous works in the last several decades have focused on recognizing faces captured in a laboratory setting. However, with the introduction of databases such as LFW and Pubfigs, face recognition community is gradually shifting its focus on much more challenging unconstrained settings. Since its introduction, LFW verification benchmark is getting a lot of attention with various researchers contributing towards state-of-the-results. To further boost the unconstrained face recognition research, we introduce a more challenging Indian Movie Face Database (IMFDB) that has much more variability compared to LFW and Pubfigs. The database consists of 34512 faces of 100 known actors collected from approximately 103 Indian movies. Unlike LFW and Pubfigs which used face detectors to automatically detect the faces from the web collection, faces in IMFDB are detected manually from all the movies. Manual selection of faces from movies resulted in high degree of variability (in scale, pose, expression, illumination, age, occlusion, makeup) which one could ever see in natural world. IMFDB is the first face database that provides a detailed annotation in terms of age, pose, gender, expression, amount of occlusion, for each face which may help other face related applications.","PeriodicalId":436402,"journal":{"name":"2013 Fourth National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics (NCVPRIPG)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130751789","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 88
期刊
2013 Fourth National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics (NCVPRIPG)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1