首页 > 最新文献

33rd Applied Imagery Pattern Recognition Workshop (AIPR'04)最新文献

英文 中文
Toward view-invariant representations of object structure learned using object constancy cues in natural movies 利用自然电影中物体恒常性线索学习物体结构的视点不变表征
Pub Date : 2004-10-13 DOI: 10.1109/AIPR.2004.47
J. Colombe
An approach to learning view-invariant object representations was explored based on the learning of legal or naturalistic view transformations in time, learned from the statistical properties of natural movies. A simple cell layer responded to localized oriented image structure, and a complex cell layer learned to respond to those subsets of simple cells with the strongest tendencies to trade off activity with each other in response to movement of the visual stimulus. Tradeoffs between simple cells were strongest in response to same-orientation translation, and fell off rapidly with changes in orientation. The local complex cell responses thus became insensitive to typical object motion, evidenced by broadening of response to stimulus phase, while remaining sensitive to local object form. The model makes predictions about synaptic learning rules in complex cells, and mechanisms of successive view-invariance in the primate ventral stream.
从自然电影的统计特性中学习合法或自然的视图变换,探索了一种学习视图不变对象表示的方法。一个简单的细胞层对局部定向的图像结构做出反应,而一个复杂的细胞层学会了对那些简单细胞子集做出反应,这些细胞子集在对视觉刺激的运动做出反应时,具有最强的相互权衡活动的倾向。简单细胞之间的权衡对相同方向的翻译反应最强,并且随着方向的变化而迅速下降。局部复杂细胞反应对典型物体运动变得不敏感,表现为对刺激阶段的反应扩大,而对局部物体形态保持敏感。该模型预测了复杂细胞的突触学习规则,以及灵长类动物腹侧流连续视点不变性的机制。
{"title":"Toward view-invariant representations of object structure learned using object constancy cues in natural movies","authors":"J. Colombe","doi":"10.1109/AIPR.2004.47","DOIUrl":"https://doi.org/10.1109/AIPR.2004.47","url":null,"abstract":"An approach to learning view-invariant object representations was explored based on the learning of legal or naturalistic view transformations in time, learned from the statistical properties of natural movies. A simple cell layer responded to localized oriented image structure, and a complex cell layer learned to respond to those subsets of simple cells with the strongest tendencies to trade off activity with each other in response to movement of the visual stimulus. Tradeoffs between simple cells were strongest in response to same-orientation translation, and fell off rapidly with changes in orientation. The local complex cell responses thus became insensitive to typical object motion, evidenced by broadening of response to stimulus phase, while remaining sensitive to local object form. The model makes predictions about synaptic learning rules in complex cells, and mechanisms of successive view-invariance in the primate ventral stream.","PeriodicalId":120814,"journal":{"name":"33rd Applied Imagery Pattern Recognition Workshop (AIPR'04)","volume":"111 1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116650632","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A nonlinear technique for enhancement of color images: an architectural perspective for real-time applications 彩色图像增强的非线性技术:实时应用的架构视角
Pub Date : 2004-10-13 DOI: 10.1109/AIPR.2004.6
H. T. Ngo, Li Tao, V. Asari
In this paper, an efficient hardware design for a nonlinear technique for enhancement of color images is presented. The enhancement technique works very effectively for images captured under extremely dark environment as well as non-uniform lighting environment where 'bright" regions are kept unaffected and 'dark' objects in 'bright' background. For efficient implementation of the nonlinear technique on a targeted FPGA board, estimation techniques for logarithm and inverse logarithm are introduced. The estimation method helps to reduce the computational time and FPGA resources significantly compared to conventional implementations of computational intensive operations such as logarithm. The enhancement technique is further analyzed and rearranged into hardware algorithmic steps to better suit the high performance implementation. A number of parallel functional modules are designed to operate simultaneously to optimally utilize the operation-level parallelism available in the technique. Sequential operations are partitioned into well-balance workload stages of a pipelined system based on the inter-data-dependency of the algorithmic steps to better utilize the resources in a FPGA such as on-chip RAM and logic-blocks. The image enhancement system is designed to target the high-performance for real time color image enhancement with minimum 25 frames per second.
本文提出了一种用于彩色图像非线性增强的有效硬件设计方法。增强技术非常有效地捕获图像在极端黑暗的环境,以及不均匀的照明环境中,“明亮”的区域保持不受影响,“黑暗”的对象在“明亮”的背景。为了在目标FPGA板上有效地实现非线性技术,介绍了对数和逆对数估计技术。与传统的计算密集型运算(如对数)相比,该估计方法有助于显著减少计算时间和FPGA资源。进一步分析了增强技术,并将其重新编排为硬件算法步骤,以更好地适应高性能实现。许多并行功能模块被设计为同时运行,以最佳地利用该技术中可用的操作级并行性。基于算法步骤之间的数据依赖性,顺序操作被划分为流水线系统中平衡良好的工作负载阶段,以更好地利用FPGA中的资源,如片上RAM和逻辑块。该图像增强系统旨在以每秒至少25帧的高性能实时彩色图像增强为目标。
{"title":"A nonlinear technique for enhancement of color images: an architectural perspective for real-time applications","authors":"H. T. Ngo, Li Tao, V. Asari","doi":"10.1109/AIPR.2004.6","DOIUrl":"https://doi.org/10.1109/AIPR.2004.6","url":null,"abstract":"In this paper, an efficient hardware design for a nonlinear technique for enhancement of color images is presented. The enhancement technique works very effectively for images captured under extremely dark environment as well as non-uniform lighting environment where 'bright\" regions are kept unaffected and 'dark' objects in 'bright' background. For efficient implementation of the nonlinear technique on a targeted FPGA board, estimation techniques for logarithm and inverse logarithm are introduced. The estimation method helps to reduce the computational time and FPGA resources significantly compared to conventional implementations of computational intensive operations such as logarithm. The enhancement technique is further analyzed and rearranged into hardware algorithmic steps to better suit the high performance implementation. A number of parallel functional modules are designed to operate simultaneously to optimally utilize the operation-level parallelism available in the technique. Sequential operations are partitioned into well-balance workload stages of a pipelined system based on the inter-data-dependency of the algorithmic steps to better utilize the resources in a FPGA such as on-chip RAM and logic-blocks. The image enhancement system is designed to target the high-performance for real time color image enhancement with minimum 25 frames per second.","PeriodicalId":120814,"journal":{"name":"33rd Applied Imagery Pattern Recognition Workshop (AIPR'04)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125364104","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A fuzzy find matching tool for image text analysis 图像文本分析的模糊查找匹配工具
Pub Date : 2004-10-13 DOI: 10.1109/AIPR.2004.2
S. Berkovich, M. Inayatullah
In this paper, we present a study based on the effect of using a fuzzy find matching technique with the objective of increasing the precision of textual information retrieval for image analysis while allowing for mismatches. This technique is very helpful for searching those areas of interest where chances of misspelling are more likely, for example retrieving text information from an image. The technique we propose can be used as an embedded component or a post-processing tool for image analysis resulting in a faster retrieval of corresponding information from images with given keywords. In the case of no hits, our technique outcome would suggest few close matches that can be further searched in keywords to get a closer match with improved speed of searching.
在本文中,我们提出了一项基于使用模糊查找匹配技术的效果的研究,目的是在允许不匹配的情况下提高图像分析文本信息检索的精度。这种技术对于搜索那些更有可能出现拼写错误的领域非常有帮助,例如从图像中检索文本信息。我们提出的技术可以用作图像分析的嵌入式组件或后处理工具,从而更快地从具有给定关键字的图像中检索相应信息。在没有命中的情况下,我们的技术结果将建议很少的接近匹配,可以在关键字中进一步搜索以获得更接近的匹配,并提高搜索速度。
{"title":"A fuzzy find matching tool for image text analysis","authors":"S. Berkovich, M. Inayatullah","doi":"10.1109/AIPR.2004.2","DOIUrl":"https://doi.org/10.1109/AIPR.2004.2","url":null,"abstract":"In this paper, we present a study based on the effect of using a fuzzy find matching technique with the objective of increasing the precision of textual information retrieval for image analysis while allowing for mismatches. This technique is very helpful for searching those areas of interest where chances of misspelling are more likely, for example retrieving text information from an image. The technique we propose can be used as an embedded component or a post-processing tool for image analysis resulting in a faster retrieval of corresponding information from images with given keywords. In the case of no hits, our technique outcome would suggest few close matches that can be further searched in keywords to get a closer match with improved speed of searching.","PeriodicalId":120814,"journal":{"name":"33rd Applied Imagery Pattern Recognition Workshop (AIPR'04)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121912492","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Approach to target detection based on relevant metric for scoring performance 基于相关评分指标的目标检测方法
Pub Date : 2004-10-13 DOI: 10.1109/AIPR.2004.14
J. Theiler, N. Harvey, N. David, J. Irvine
Improved target detection, reduced false alarm rates, and enhanced timeliness are critical to meeting the requirements of current and future military missions. We present a new approach to target detection, based on a suite of image processing and exploitation tools developed under the intelligent searching of images and signals (ISIS) program at Los Alamos National Laboratory. Performance assessment of these algorithms relies on a new metric for scoring target detection that is relevant to the analyst's needs. An object-based loss function is defined by the degree to which the automated processing focuses the analyst's attention on the true targets and avoids false positives. For target detection techniques that produce a pixel-by-pixel classification (and thereby produce not just an identification of the target, but a segmentation as well), standard scoring rules are not appropriate because they unduly penalize partial detections. From a practical standpoint, it is not necessary to identify every single pixel that is on the target; all that is required is that the processing draw the analyst's attention to the target. By employing this scoring metric directly into the target detection algorithm, improved performance in this more practical context can be obtained.
改进目标探测、降低误报率和增强及时性对于满足当前和未来军事任务的要求至关重要。我们提出了一种新的目标检测方法,该方法基于洛斯阿拉莫斯国家实验室在图像和信号智能搜索(ISIS)计划下开发的一套图像处理和利用工具。这些算法的性能评估依赖于与分析人员的需求相关的目标检测评分的新度量。基于对象的损失函数是由自动化处理将分析人员的注意力集中在真实目标上并避免误报的程度来定义的。对于产生逐像素分类的目标检测技术(因此不仅产生目标的识别,而且还产生分割),标准评分规则是不合适的,因为它们过度地惩罚了部分检测。从实用的角度来看,没有必要识别目标上的每一个像素;所需要做的就是把分析人员的注意力吸引到目标上。通过将该评分指标直接应用到目标检测算法中,可以在这种更实际的情况下获得更好的性能。
{"title":"Approach to target detection based on relevant metric for scoring performance","authors":"J. Theiler, N. Harvey, N. David, J. Irvine","doi":"10.1109/AIPR.2004.14","DOIUrl":"https://doi.org/10.1109/AIPR.2004.14","url":null,"abstract":"Improved target detection, reduced false alarm rates, and enhanced timeliness are critical to meeting the requirements of current and future military missions. We present a new approach to target detection, based on a suite of image processing and exploitation tools developed under the intelligent searching of images and signals (ISIS) program at Los Alamos National Laboratory. Performance assessment of these algorithms relies on a new metric for scoring target detection that is relevant to the analyst's needs. An object-based loss function is defined by the degree to which the automated processing focuses the analyst's attention on the true targets and avoids false positives. For target detection techniques that produce a pixel-by-pixel classification (and thereby produce not just an identification of the target, but a segmentation as well), standard scoring rules are not appropriate because they unduly penalize partial detections. From a practical standpoint, it is not necessary to identify every single pixel that is on the target; all that is required is that the processing draw the analyst's attention to the target. By employing this scoring metric directly into the target detection algorithm, improved performance in this more practical context can be obtained.","PeriodicalId":120814,"journal":{"name":"33rd Applied Imagery Pattern Recognition Workshop (AIPR'04)","volume":"76 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132819389","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Fusion of intensity, texture, and color in video tracking based on mutual information 基于互信息的视频跟踪中强度、纹理和颜色的融合
Pub Date : 2004-10-13 DOI: 10.1109/AIPR.2004.26
J. Mundy, Chung-Fu Chang
Next-generation reconnaissance systems (NGRS) offer dynamic tasking of a menu of sensor modalities such as video, multi/hyper-spectral and polarization data. A key issue is how best to exploit these modes in time critical scenarios such as target tracking and event detection. It is essential to be able to represent diverse sensor content in a unified measurement space so that the contribution of each modality can be evaluated in terms of its contribution to the exploitation task. In this paper, mutual information is used to represent the content of individual sensor channels. A series of experiments on video tracking have been carried out to demonstrate the effectiveness of mutual information as a fusion framework. These experiments quantify the relative information content of intensity, color, and polarization image channels.
下一代侦察系统(NGRS)提供一系列传感器模式的动态任务,如视频、多/高光谱和偏振数据。一个关键问题是如何在时间关键的场景(如目标跟踪和事件检测)中最好地利用这些模式。必须能够在统一的测量空间中表示不同的传感器内容,以便可以根据其对开发任务的贡献来评估每种模式的贡献。在本文中,互信息被用来表示单个传感器通道的内容。通过一系列的视频跟踪实验,验证了互信息融合框架的有效性。这些实验量化了强度、颜色和偏振图像通道的相对信息含量。
{"title":"Fusion of intensity, texture, and color in video tracking based on mutual information","authors":"J. Mundy, Chung-Fu Chang","doi":"10.1109/AIPR.2004.26","DOIUrl":"https://doi.org/10.1109/AIPR.2004.26","url":null,"abstract":"Next-generation reconnaissance systems (NGRS) offer dynamic tasking of a menu of sensor modalities such as video, multi/hyper-spectral and polarization data. A key issue is how best to exploit these modes in time critical scenarios such as target tracking and event detection. It is essential to be able to represent diverse sensor content in a unified measurement space so that the contribution of each modality can be evaluated in terms of its contribution to the exploitation task. In this paper, mutual information is used to represent the content of individual sensor channels. A series of experiments on video tracking have been carried out to demonstrate the effectiveness of mutual information as a fusion framework. These experiments quantify the relative information content of intensity, color, and polarization image channels.","PeriodicalId":120814,"journal":{"name":"33rd Applied Imagery Pattern Recognition Workshop (AIPR'04)","volume":"65 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133266178","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
Comparison of non-parametric methods for assessing classifier performance in terms of ROC parameters 根据ROC参数评估分类器性能的非参数方法的比较
Pub Date : 2004-10-13 DOI: 10.1109/AIPR.2004.18
W. Yousef, R. F. Wagner, M. Loew
The most common metric to assess a classifier's performance is the classification error rate, or the probability of misclassification (PMC). Receiver operating characteristic (ROC) analysis is a more general way to measure the performance. Some metrics that summarize the ROC curve are the two normal-deviate-axes parameters, i.e., a and b, and the area under the curve (AUC). The parameters "a" and "b" represent the intercept and slope, respectively, for the ROC curve if plotted on normal-deviate-axes scale. AUC represents the average of the classifier TPF over FPF resulting from considering different threshold values. In the present work, we used Monte-Carlo simulations to compare different bootstrap-based estimators, e.g., leave-one-out, .632, and .632+ bootstraps, to estimate the AUC. The results show the comparable performance of the different estimators in terms of RMS, while the .632+ is the least biased.
评估分类器性能的最常见指标是分类错误率,或错误分类的概率(PMC)。接受者工作特征(ROC)分析是衡量绩效的一种更通用的方法。总结ROC曲线的一些指标是两个正态偏离轴参数,即a和b,以及曲线下面积(AUC)。参数“a”和“b”分别表示ROC曲线的截距和斜率,如果在正态偏差轴尺度上绘制。AUC表示由于考虑不同阈值而产生的分类器TPF与FPF的平均值。在目前的工作中,我们使用蒙特卡罗模拟来比较不同的基于bootstrap的估计器,例如,left - 1 out, .632和.632+ bootstrap,以估计AUC。结果表明,不同的估计器在RMS方面具有可比性,而0.632 +是偏差最小的。
{"title":"Comparison of non-parametric methods for assessing classifier performance in terms of ROC parameters","authors":"W. Yousef, R. F. Wagner, M. Loew","doi":"10.1109/AIPR.2004.18","DOIUrl":"https://doi.org/10.1109/AIPR.2004.18","url":null,"abstract":"The most common metric to assess a classifier's performance is the classification error rate, or the probability of misclassification (PMC). Receiver operating characteristic (ROC) analysis is a more general way to measure the performance. Some metrics that summarize the ROC curve are the two normal-deviate-axes parameters, i.e., a and b, and the area under the curve (AUC). The parameters \"a\" and \"b\" represent the intercept and slope, respectively, for the ROC curve if plotted on normal-deviate-axes scale. AUC represents the average of the classifier TPF over FPF resulting from considering different threshold values. In the present work, we used Monte-Carlo simulations to compare different bootstrap-based estimators, e.g., leave-one-out, .632, and .632+ bootstraps, to estimate the AUC. The results show the comparable performance of the different estimators in terms of RMS, while the .632+ is the least biased.","PeriodicalId":120814,"journal":{"name":"33rd Applied Imagery Pattern Recognition Workshop (AIPR'04)","volume":"449 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116189182","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 41
Mountain clustering on nonuniform grids 非均匀网格上的山聚类
Pub Date : 2004-10-13 DOI: 10.1109/AIPR.2004.31
J. T. Rickard, R. Yager, W. Miller
We describe an improvement on the mountain method (MM) of clustering originally proposed by Yager and Filev. The new technique employs a data-driven, hierarchical partitioning of the data set to be clustered, using a "p-tree" algorithm. The centroids of data subsets in the terminal nodes of the p-tree are the set of candidate cluster centers to which the iterative candidate cluster center selection process of MM is applied. As the data dimension and/or the number of uniform grid lines used in the original MM increases, our approach requires exponentially fewer cluster centers to be evaluated by the MM selection algorithm. Sample data sets illustrate the performance of this new technique.
我们描述了对Yager和Filev最初提出的山法(MM)聚类的改进。新技术采用数据驱动的分层划分,使用“p树”算法对数据集进行聚类。p-tree终端节点的数据子集的质心是应用MM迭代候选聚类中心选择过程的候选聚类中心集合。随着原始MM中使用的数据维度和/或均匀网格线数量的增加,我们的方法需要MM选择算法评估的聚类中心呈指数级减少。示例数据集说明了这种新技术的性能。
{"title":"Mountain clustering on nonuniform grids","authors":"J. T. Rickard, R. Yager, W. Miller","doi":"10.1109/AIPR.2004.31","DOIUrl":"https://doi.org/10.1109/AIPR.2004.31","url":null,"abstract":"We describe an improvement on the mountain method (MM) of clustering originally proposed by Yager and Filev. The new technique employs a data-driven, hierarchical partitioning of the data set to be clustered, using a \"p-tree\" algorithm. The centroids of data subsets in the terminal nodes of the p-tree are the set of candidate cluster centers to which the iterative candidate cluster center selection process of MM is applied. As the data dimension and/or the number of uniform grid lines used in the original MM increases, our approach requires exponentially fewer cluster centers to be evaluated by the MM selection algorithm. Sample data sets illustrate the performance of this new technique.","PeriodicalId":120814,"journal":{"name":"33rd Applied Imagery Pattern Recognition Workshop (AIPR'04)","volume":"176 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129885192","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Face recognition by capturing eye illumination spot 通过捕捉眼睛照明点进行人脸识别
Pub Date : 2004-10-13 DOI: 10.1109/AIPR.2004.24
N. Singh
Here a method is presented which recognizes a face by capturing a unique illumination spot created in the centre of the eye i.e. the cornea. This illumination spot is always there whenever we are in light or in front of a light source. This algorithm works by running an edge detector on a sample face. The edge detector produces a dark spot in place of the illumination spot and a semicircular arc in place of the beginning of the cornea. This arc and the dark spot act as a unique template for capturing eye in the image and following this we extract other features i.e. eyebrows, lip line and chin line. Following this we calculate the vertical distance of each feature from every other feature and use these distance parameters to recognize faces.
这里提出了一种方法,该方法通过捕捉在眼睛中心(即角膜)创建的独特照明点来识别人脸。每当我们在灯光下或在光源前时,这个照明点总是存在的。该算法通过在样本面部上运行边缘检测器来工作。所述边缘检测器产生代替所述照明点的暗点和代替所述角膜起点的半圆弧。这个弧线和暗点作为一个独特的模板来捕捉图像中的眼睛,然后我们提取其他特征,如眉毛,唇线和下巴线。接下来,我们计算每个特征与其他特征之间的垂直距离,并使用这些距离参数来识别人脸。
{"title":"Face recognition by capturing eye illumination spot","authors":"N. Singh","doi":"10.1109/AIPR.2004.24","DOIUrl":"https://doi.org/10.1109/AIPR.2004.24","url":null,"abstract":"Here a method is presented which recognizes a face by capturing a unique illumination spot created in the centre of the eye i.e. the cornea. This illumination spot is always there whenever we are in light or in front of a light source. This algorithm works by running an edge detector on a sample face. The edge detector produces a dark spot in place of the illumination spot and a semicircular arc in place of the beginning of the cornea. This arc and the dark spot act as a unique template for capturing eye in the image and following this we extract other features i.e. eyebrows, lip line and chin line. Following this we calculate the vertical distance of each feature from every other feature and use these distance parameters to recognize faces.","PeriodicalId":120814,"journal":{"name":"33rd Applied Imagery Pattern Recognition Workshop (AIPR'04)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129064911","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Recognition of middle age Persian characters using a set of invariant moments 用一组不变矩识别中年波斯语字符
Pub Date : 2004-10-13 DOI: 10.1109/AIPR.2004.39
S. Alirezaee, H. Aghaeinia, M. Ahmadi, K. Faez
In this paper, recognition of ancient middle Persian documents is studied. Our major attention has been focused on feature extraction and classification. A set of invariant moments has been selected as the features and the minimum mean distance (three versions of which that is called MMD1, MMD2, MMD3), KNN and Parzen as the classifier. Preprocessing is also considered in this paper which allows, the effects of under sampling (resolution pyramids), smoothing, and thinning be investigated. The algorithm has been tested not only on the original and smoothed images but also on the skeletonized and under sampled version of the text under test. The results show an acceptable recognition rate with the selected features with the proposed processing for the middle age Persian. The best-achieved classification rates are 95% and 90.5% for smoothed and original character images respectively. It was interesting to note that KNN and MMD2 classifiers yielded better recognition rate.
本文主要研究中古波斯文献的识别问题。我们的主要注意力集中在特征提取和分类上。选取一组不变矩作为特征,最小平均距离(MMD1、MMD2、MMD3三个版本)、KNN和Parzen作为分类器。本文还考虑了预处理,研究了欠采样(分辨率金字塔)、平滑和细化的影响。该算法不仅在原始图像和平滑图像上进行了测试,而且在被测文本的骨架化和欠采样版本上进行了测试。结果表明,所选择的特征与所提出的处理方法对中年波斯人具有可接受的识别率。对平滑图像和原始图像的分类率分别为95%和90.5%。有趣的是,KNN和MMD2分类器产生了更好的识别率。
{"title":"Recognition of middle age Persian characters using a set of invariant moments","authors":"S. Alirezaee, H. Aghaeinia, M. Ahmadi, K. Faez","doi":"10.1109/AIPR.2004.39","DOIUrl":"https://doi.org/10.1109/AIPR.2004.39","url":null,"abstract":"In this paper, recognition of ancient middle Persian documents is studied. Our major attention has been focused on feature extraction and classification. A set of invariant moments has been selected as the features and the minimum mean distance (three versions of which that is called MMD1, MMD2, MMD3), KNN and Parzen as the classifier. Preprocessing is also considered in this paper which allows, the effects of under sampling (resolution pyramids), smoothing, and thinning be investigated. The algorithm has been tested not only on the original and smoothed images but also on the skeletonized and under sampled version of the text under test. The results show an acceptable recognition rate with the selected features with the proposed processing for the middle age Persian. The best-achieved classification rates are 95% and 90.5% for smoothed and original character images respectively. It was interesting to note that KNN and MMD2 classifiers yielded better recognition rate.","PeriodicalId":120814,"journal":{"name":"33rd Applied Imagery Pattern Recognition Workshop (AIPR'04)","volume":"89 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124605979","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Online handwritten circuit recognition on a tablet PC 平板电脑上的在线手写电路识别
Pub Date : 2004-10-13 DOI: 10.1109/AIPR.2004.35
O. Ejofodomi, Shani Ross, A. Jendoubi, M. Chouikha, J. Zeng
Tablet PC-based engineering software can be used as an effective teaching tool for core engineering courses such as electronics, signal and systems, and digital systems. Wireless connection between tablet PCs of students and the teaching professor will substantially improve students' involvement during the course. Circuit drawing is an important task especially in undergraduate courses such as electronics and digital systems. Most existing software tools for circuit drawing use a toolbox where symbols for all circuit components are prepared and ready for pick up. A user has to go through a number of layers of menus each time he/she wants to use a circuit symbol. To improve human computer interaction, we have developed an online recognition system on a tablet PC using C# for the handwritten circuit and its components. The system can recognize and redraw many circuits and their components such as resistors, capacitors, ground and various voltage power supplies, which are drawn with a stylus pen on a tablet PC. We present details of our approach and preliminary results of an experimental system.
基于平板电脑的工程软件可以作为电子学、信号与系统、数字系统等核心工程课程的有效教学工具。学生的平板电脑与授课教授之间的无线连接将大大提高学生在课程中的参与度。电路绘制是一项重要的任务,特别是在电子和数字系统等本科课程中。大多数现有的电路绘图软件工具使用一个工具箱,其中所有电路元件的符号都准备好了,随时可以拿起。用户每次想要使用电路符号时,都要经过几层菜单。为了提高人机交互性,我们在平板电脑上使用c#开发了一个手写电路及其组件的在线识别系统。该系统可以识别和重绘电阻、电容、接地、各种电压电源等多种电路及其元件,在平板电脑上用触控笔绘制。我们详细介绍了我们的方法和实验系统的初步结果。
{"title":"Online handwritten circuit recognition on a tablet PC","authors":"O. Ejofodomi, Shani Ross, A. Jendoubi, M. Chouikha, J. Zeng","doi":"10.1109/AIPR.2004.35","DOIUrl":"https://doi.org/10.1109/AIPR.2004.35","url":null,"abstract":"Tablet PC-based engineering software can be used as an effective teaching tool for core engineering courses such as electronics, signal and systems, and digital systems. Wireless connection between tablet PCs of students and the teaching professor will substantially improve students' involvement during the course. Circuit drawing is an important task especially in undergraduate courses such as electronics and digital systems. Most existing software tools for circuit drawing use a toolbox where symbols for all circuit components are prepared and ready for pick up. A user has to go through a number of layers of menus each time he/she wants to use a circuit symbol. To improve human computer interaction, we have developed an online recognition system on a tablet PC using C# for the handwritten circuit and its components. The system can recognize and redraw many circuits and their components such as resistors, capacitors, ground and various voltage power supplies, which are drawn with a stylus pen on a tablet PC. We present details of our approach and preliminary results of an experimental system.","PeriodicalId":120814,"journal":{"name":"33rd Applied Imagery Pattern Recognition Workshop (AIPR'04)","volume":"60 6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127384532","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
期刊
33rd Applied Imagery Pattern Recognition Workshop (AIPR'04)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1