首页 > 最新文献

33rd Applied Imagery Pattern Recognition Workshop (AIPR'04)最新文献

英文 中文
Assessing the performance of an automated video ground truthing application 评估自动视频地面真实性应用程序的性能
Pub Date : 2004-10-13 DOI: 10.1109/AIPR.2004.15
Scott K. Ralph, J. Irvine, M. R. Stevens, M. Snorrason, D. Gwilt
Present methods of quantifying the performance of ATR algorithms involves the use of large video datasets that must be truthed by hand, frame-by-frame, requiring vast amounts of time. To reduce this cost, we have developed an application that significantly reduces the cost by only requiring the operator to grade a relatively sparse number of data "keyframes". A correlation-based template matching algorithm computes the best position, orientation and scale when interpolating between keyframes. We demonstrate the performance of the automated truthing application, and compare the results to those of a series of human operator test subjects. The START-generated truth is shown to be very close to the mean truth data given by the human operators. Additionally the savings in labor producing the results is also demonstrated.
目前量化ATR算法性能的方法涉及到使用大型视频数据集,这些数据集必须手工逐帧处理,需要大量的时间。为了降低这一成本,我们开发了一个应用程序,该应用程序只要求操作员对相对稀疏的数据“关键帧”进行分级,从而大大降低了成本。在关键帧之间插值时,基于关联的模板匹配算法计算最佳位置、方向和比例。我们演示了自动真相应用程序的性能,并将结果与一系列人类操作员测试对象的结果进行了比较。start生成的真值非常接近人类操作员给出的平均真值数据。此外,还证明了产生结果的劳动力节省。
{"title":"Assessing the performance of an automated video ground truthing application","authors":"Scott K. Ralph, J. Irvine, M. R. Stevens, M. Snorrason, D. Gwilt","doi":"10.1109/AIPR.2004.15","DOIUrl":"https://doi.org/10.1109/AIPR.2004.15","url":null,"abstract":"Present methods of quantifying the performance of ATR algorithms involves the use of large video datasets that must be truthed by hand, frame-by-frame, requiring vast amounts of time. To reduce this cost, we have developed an application that significantly reduces the cost by only requiring the operator to grade a relatively sparse number of data \"keyframes\". A correlation-based template matching algorithm computes the best position, orientation and scale when interpolating between keyframes. We demonstrate the performance of the automated truthing application, and compare the results to those of a series of human operator test subjects. The START-generated truth is shown to be very close to the mean truth data given by the human operators. Additionally the savings in labor producing the results is also demonstrated.","PeriodicalId":120814,"journal":{"name":"33rd Applied Imagery Pattern Recognition Workshop (AIPR'04)","volume":"89 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127191990","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Illumination invariant faces 光照不变面
Pub Date : 2004-10-13 DOI: 10.1109/AIPR.2004.27
Rajkiran Gottumukkal, V. Asari
We create a model of joint color changes in face images due to lighting variations. This is done by observing how colors of an individual's face with fixed pose and expression are mapped to new colors under different lighting conditions. One of the challenges we are dealing with in this work is that the scenes are not constant for different lighting. Hence we cannot observe the joint color changes of the scenes. However all the scenes have a human subject with approximately frontal pose, so we use the color changes observed on a human subjects face to learn the color mapping. The joint color mappings are represented in a low dimensional subspace obtained using singular value decomposition (SVD). Using these maps the detected face from a new image can be transformed to appear as if taken under canonical lighting condition.
我们创建了一个面部图像中由于光照变化而引起的关节颜色变化的模型。这是通过观察一个人在不同的光照条件下,固定姿势和表情的面部颜色如何映射到新的颜色来完成的。我们在这项工作中面临的挑战之一是,不同的照明场景不是恒定的。因此,我们无法观察到场景的联合颜色变化。然而,所有的场景都有一个近似正面姿势的人类主体,所以我们使用在人类主体脸上观察到的颜色变化来学习颜色映射。利用奇异值分解(SVD)在低维子空间中表示联合颜色映射。使用这些地图,从新图像中检测到的人脸可以转换为在标准照明条件下拍摄的样子。
{"title":"Illumination invariant faces","authors":"Rajkiran Gottumukkal, V. Asari","doi":"10.1109/AIPR.2004.27","DOIUrl":"https://doi.org/10.1109/AIPR.2004.27","url":null,"abstract":"We create a model of joint color changes in face images due to lighting variations. This is done by observing how colors of an individual's face with fixed pose and expression are mapped to new colors under different lighting conditions. One of the challenges we are dealing with in this work is that the scenes are not constant for different lighting. Hence we cannot observe the joint color changes of the scenes. However all the scenes have a human subject with approximately frontal pose, so we use the color changes observed on a human subjects face to learn the color mapping. The joint color mappings are represented in a low dimensional subspace obtained using singular value decomposition (SVD). Using these maps the detected face from a new image can be transformed to appear as if taken under canonical lighting condition.","PeriodicalId":120814,"journal":{"name":"33rd Applied Imagery Pattern Recognition Workshop (AIPR'04)","volume":"143 7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129470020","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Advanced algorithms for autonomous hyperspectral change detection 自主高光谱变化检测的先进算法
Pub Date : 2004-10-13 DOI: 10.1109/AIPR.2004.10
A. Schaum, A. Stocker
Persistent ISR (intelligence surveillance and reconnaissance) has proven its value as a tactic for national defense. This activity can collect, in particular, information necessary for executing an important concept of operations: wide-area autonomous change detection over long time intervals. Here we describe the remarkable potential of hyperspectral remote sensing systems for enabling such missions, using either visible or thermal infrared wavelengths. First we describe blind change detection, in which no target knowledge is assumed. Targets that have moved can nevertheless be distinguished from naturally occurring background radiometric changes through the use of multivariate statistics informed by simple physics. Detection relies on the ability of hyperspectral algorithms to predict certain conserved properties of background spectral patterns over long time intervals. We also describe a method of mitigating the most worrisome practical engineering difficulty in pixel-level change detection, image misregistration. This has led, in turn, to a method of estimating spectral signature evolution using multiple-scene statistics. Finally, we present a signature-based detection technique that fuses two discrimination mechanisms: use of some prior knowledge of target spectra, and the fact that a change has occurred.
持续ISR(情报监视和侦察)已经证明了其作为国防战术的价值。该活动可以特别收集执行一个重要操作概念所必需的信息:长时间间隔的广域自主变更检测。在这里,我们描述了使用可见光或热红外波长的高光谱遥感系统实现此类任务的显着潜力。首先,我们描述了盲变化检测,其中不假设目标知识。然而,通过使用由简单物理学提供信息的多元统计,可以将移动的目标与自然发生的背景辐射变化区分开来。检测依赖于高光谱算法在长时间间隔内预测背景光谱模式的某些保守属性的能力。我们还描述了一种方法,以减轻像素级变化检测中最令人担忧的实际工程困难,即图像错配。这反过来又导致了一种使用多场景统计估计光谱特征演变的方法。最后,我们提出了一种基于特征的检测技术,该技术融合了两种识别机制:利用目标光谱的一些先验知识和已经发生变化的事实。
{"title":"Advanced algorithms for autonomous hyperspectral change detection","authors":"A. Schaum, A. Stocker","doi":"10.1109/AIPR.2004.10","DOIUrl":"https://doi.org/10.1109/AIPR.2004.10","url":null,"abstract":"Persistent ISR (intelligence surveillance and reconnaissance) has proven its value as a tactic for national defense. This activity can collect, in particular, information necessary for executing an important concept of operations: wide-area autonomous change detection over long time intervals. Here we describe the remarkable potential of hyperspectral remote sensing systems for enabling such missions, using either visible or thermal infrared wavelengths. First we describe blind change detection, in which no target knowledge is assumed. Targets that have moved can nevertheless be distinguished from naturally occurring background radiometric changes through the use of multivariate statistics informed by simple physics. Detection relies on the ability of hyperspectral algorithms to predict certain conserved properties of background spectral patterns over long time intervals. We also describe a method of mitigating the most worrisome practical engineering difficulty in pixel-level change detection, image misregistration. This has led, in turn, to a method of estimating spectral signature evolution using multiple-scene statistics. Finally, we present a signature-based detection technique that fuses two discrimination mechanisms: use of some prior knowledge of target spectra, and the fact that a change has occurred.","PeriodicalId":120814,"journal":{"name":"33rd Applied Imagery Pattern Recognition Workshop (AIPR'04)","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116987824","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 30
An efficient selected feature set for the middle age Persian character recognition 中年波斯语字符识别的有效选择特征集
Pub Date : 2004-10-13 DOI: 10.1109/AIPR.2004.12
S. Alirezaee, H. Aghaeinia, M. Ahmadi, K. Faez
In this paper, a morphological based method for recognition of handwritten middle Persian characters is presented. After pre-processing and noise cancellation, morphological erosion operator with many structure elements is applied. The structure elements are with variable length lines at directions 0, 45, 90, 135 degrees. A five element feature set has been defined so: (1) relative energy of eroded version with respect to the original image energy (REL/spl I.bar/ENG),(2) displacement of the center of mass (CM/spl I.bar//spl I.bar/DIS), (3) minimum eigenvalue (EIG/spl I.bar/MIN), (4) maximum eigenvalue (EIG/spl I.bar/MAX) and (5) its direction (EIG-DIR). These features are used to design a feedforward neural network with one hidden layer. The best classification error is about 2.39% (97.61% recognition rate), and is achieved with 150 neurons for the hidden layer.
本文提出了一种基于形态学的中波斯手写体汉字识别方法。经过预处理和去噪后,采用多结构元形态侵蚀算子。结构元件在0、45、90、135度方向上具有可变长度线。定义了一个五元素特征集:(1)侵蚀版相对于原始图像能量的相对能量(REL/spl I.bar/ENG),(2)质心位移(CM/spl I.bar//spl I.bar/DIS),(3)最小特征值(EIG/spl I.bar/MIN),(4)最大特征值(EIG/spl I.bar/MAX)和(5)其方向(EIG- dir)。利用这些特征设计了一个具有一个隐藏层的前馈神经网络。在隐藏层使用150个神经元时,分类误差达到2.39%(识别率为97.61%)。
{"title":"An efficient selected feature set for the middle age Persian character recognition","authors":"S. Alirezaee, H. Aghaeinia, M. Ahmadi, K. Faez","doi":"10.1109/AIPR.2004.12","DOIUrl":"https://doi.org/10.1109/AIPR.2004.12","url":null,"abstract":"In this paper, a morphological based method for recognition of handwritten middle Persian characters is presented. After pre-processing and noise cancellation, morphological erosion operator with many structure elements is applied. The structure elements are with variable length lines at directions 0, 45, 90, 135 degrees. A five element feature set has been defined so: (1) relative energy of eroded version with respect to the original image energy (REL/spl I.bar/ENG),(2) displacement of the center of mass (CM/spl I.bar//spl I.bar/DIS), (3) minimum eigenvalue (EIG/spl I.bar/MIN), (4) maximum eigenvalue (EIG/spl I.bar/MAX) and (5) its direction (EIG-DIR). These features are used to design a feedforward neural network with one hidden layer. The best classification error is about 2.39% (97.61% recognition rate), and is achieved with 150 neurons for the hidden layer.","PeriodicalId":120814,"journal":{"name":"33rd Applied Imagery Pattern Recognition Workshop (AIPR'04)","volume":"64 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116175316","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Unsupervised fuzzy-membership estimation of terms in semantic and syntactic lexical classes 语义和句法词汇类中术语的无监督模糊隶属度估计
Pub Date : 2004-10-13 DOI: 10.1109/AIPR.2004.48
David Portnoy, P. Bock
The objective of this research is to discover fuzzy semantic and syntactic relationships among English words at various levels of abstraction without using any other sources of semantic or syntactical reference information (e.g. dictionaries, lexicons, grammar rules, etc...) An agglomerative clustering algorithm is applied to the co-occurrence space formed by subsets of target words and training words the output of which is a set of semantic or syntactic classes. The fuzzy-relationships (membership coefficients) between test words and the semantic and syntactic classes are estimated by the non-negative least-squares solution to the system of linear equations defined. Experiments using raw text in 218 unrelated novels have yielded promising results. It is expected that larger and/or more narrowly focused training sets would yield even better and more diverse results.
本研究的目的是在不使用任何其他语义或句法参考信息来源(如字典、词汇、语法规则等)的情况下,在不同抽象层次上发现英语单词之间的模糊语义和句法关系。将聚类算法应用于目标词和训练词子集组成的共现空间,其输出是一组语义或句法类。测试词与语义类和句法类之间的模糊关系(隶属系数)由定义的线性方程组的非负最小二乘解估计。利用218篇不相关小说的原始文本进行的实验取得了可喜的结果。预计更大和/或更狭窄的训练集将产生更好和更多样化的结果。
{"title":"Unsupervised fuzzy-membership estimation of terms in semantic and syntactic lexical classes","authors":"David Portnoy, P. Bock","doi":"10.1109/AIPR.2004.48","DOIUrl":"https://doi.org/10.1109/AIPR.2004.48","url":null,"abstract":"The objective of this research is to discover fuzzy semantic and syntactic relationships among English words at various levels of abstraction without using any other sources of semantic or syntactical reference information (e.g. dictionaries, lexicons, grammar rules, etc...) An agglomerative clustering algorithm is applied to the co-occurrence space formed by subsets of target words and training words the output of which is a set of semantic or syntactic classes. The fuzzy-relationships (membership coefficients) between test words and the semantic and syntactic classes are estimated by the non-negative least-squares solution to the system of linear equations defined. Experiments using raw text in 218 unrelated novels have yielded promising results. It is expected that larger and/or more narrowly focused training sets would yield even better and more diverse results.","PeriodicalId":120814,"journal":{"name":"33rd Applied Imagery Pattern Recognition Workshop (AIPR'04)","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123944118","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Texture discrimination and classification using pulse images 基于脉冲图像的纹理识别与分类
Pub Date : 2004-10-13 DOI: 10.1109/AIPR.2004.44
Guisong Wang, J. Kinser
Neural models of the mammalian visual cortex have been used in digital image processing for many years. The neural models have demonstrated a robust talent for extracting image segments that are inherent in the image. This paper explores the use of a simple neural model for the extraction of texture information. The neural firing patterns in the model are dependent upon the input texture and by examining the neural patterns it is possible to classify the texture in the input image. Two applications presented here are the comparison of performance to other texture analysis techniques using two standard databases and the classification of texture regions in a medical image.
哺乳动物视觉皮层的神经模型已在数字图像处理中应用多年。神经模型在提取图像中固有的图像片段方面表现出强大的才能。本文探讨了一种简单的神经模型用于纹理信息的提取。模型中的神经放电模式依赖于输入的纹理,通过检查神经模式,可以对输入图像中的纹理进行分类。本文介绍的两个应用是使用两个标准数据库与其他纹理分析技术的性能比较以及医学图像中纹理区域的分类。
{"title":"Texture discrimination and classification using pulse images","authors":"Guisong Wang, J. Kinser","doi":"10.1109/AIPR.2004.44","DOIUrl":"https://doi.org/10.1109/AIPR.2004.44","url":null,"abstract":"Neural models of the mammalian visual cortex have been used in digital image processing for many years. The neural models have demonstrated a robust talent for extracting image segments that are inherent in the image. This paper explores the use of a simple neural model for the extraction of texture information. The neural firing patterns in the model are dependent upon the input texture and by examining the neural patterns it is possible to classify the texture in the input image. Two applications presented here are the comparison of performance to other texture analysis techniques using two standard databases and the classification of texture regions in a medical image.","PeriodicalId":120814,"journal":{"name":"33rd Applied Imagery Pattern Recognition Workshop (AIPR'04)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132374436","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Neurally-based algorithms for image processing 基于神经的图像处理算法
Pub Date : 2004-10-13 DOI: 10.1109/AIPR.2004.34
Mark Flynn, H. Abarbanel, Garrett T. Kenyon
One of the more difficult problems in image processing is segmentation. The human brain has an ability that is unmatched by any current technology for breaking down the world into distributed features and reconstructing them into distinct objects. Neurons encode information both in the number of spikes fired in a given time period, which indicates the strength with which a given local feature is present, and in the temporal code or relative timing of the spike, indicating whether the individual features are part of the same or different objects. Neurons that respond to contiguous stimuli produce synchronous oscillations, while those that are not fire independently. Thus, neural synchrony could be used as a tag for each pixel in an image indicating to which object it belongs. We have developed a simulation based on the primary visual cortex. We found that neurons that respond to the same object oscillate synchronously while those that respond to different objects fire independently.
图像分割是图像处理中比较困难的问题之一。人类的大脑有一种能力,这是目前任何技术都无法比拟的,它可以将世界分解成分布的特征,并将它们重建成不同的物体。神经元对信息进行编码,包括在给定时间段内触发的峰值数量,这表明给定局部特征存在的强度,以及脉冲的时间编码或相对时间,表明单个特征是相同还是不同物体的一部分。对连续刺激作出反应的神经元产生同步振荡,而那些不独立放电的神经元产生同步振荡。因此,神经同步可以用作图像中每个像素的标签,表明它属于哪个对象。我们开发了一个基于初级视觉皮层的模拟系统。我们发现,对同一物体作出反应的神经元同步振荡,而对不同物体作出反应的神经元则独立振荡。
{"title":"Neurally-based algorithms for image processing","authors":"Mark Flynn, H. Abarbanel, Garrett T. Kenyon","doi":"10.1109/AIPR.2004.34","DOIUrl":"https://doi.org/10.1109/AIPR.2004.34","url":null,"abstract":"One of the more difficult problems in image processing is segmentation. The human brain has an ability that is unmatched by any current technology for breaking down the world into distributed features and reconstructing them into distinct objects. Neurons encode information both in the number of spikes fired in a given time period, which indicates the strength with which a given local feature is present, and in the temporal code or relative timing of the spike, indicating whether the individual features are part of the same or different objects. Neurons that respond to contiguous stimuli produce synchronous oscillations, while those that are not fire independently. Thus, neural synchrony could be used as a tag for each pixel in an image indicating to which object it belongs. We have developed a simulation based on the primary visual cortex. We found that neurons that respond to the same object oscillate synchronously while those that respond to different objects fire independently.","PeriodicalId":120814,"journal":{"name":"33rd Applied Imagery Pattern Recognition Workshop (AIPR'04)","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115609615","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Designing templates for cellular neural networks using particle swarm optimization 基于粒子群优化的细胞神经网络模板设计
Pub Date : 2004-10-13 DOI: 10.1109/AIPR.2004.21
H. Firpi, E. Goodman
Designing or learning of templates for cellular neural networks constitutes one of the crucial research problems of this paradigm. In this work, we present the use of a particle swarm optimizer, a global search algorithm, to design a template set for a CNN. A brief overview of the algorithms and methods is given. Design of popular templates is performed using the search algorithm described.
细胞神经网络模板的设计或学习是这一范式的关键研究问题之一。在这项工作中,我们提出了使用粒子群优化器,一种全局搜索算法,为CNN设计模板集。给出了算法和方法的简要概述。使用所描述的搜索算法对常用模板进行设计。
{"title":"Designing templates for cellular neural networks using particle swarm optimization","authors":"H. Firpi, E. Goodman","doi":"10.1109/AIPR.2004.21","DOIUrl":"https://doi.org/10.1109/AIPR.2004.21","url":null,"abstract":"Designing or learning of templates for cellular neural networks constitutes one of the crucial research problems of this paradigm. In this work, we present the use of a particle swarm optimizer, a global search algorithm, to design a template set for a CNN. A brief overview of the algorithms and methods is given. Design of popular templates is performed using the search algorithm described.","PeriodicalId":120814,"journal":{"name":"33rd Applied Imagery Pattern Recognition Workshop (AIPR'04)","volume":"129 2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124587977","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Biologically inspired approaches to automated feature extraction and target recognition 生物学启发的自动特征提取和目标识别方法
Pub Date : 2004-10-13 DOI: 10.1109/AIPR.2004.17
G. Carpenter, S. Martens, E. Mingolla, Ogi J. Ogas, C. Gaddam
Ongoing research at Boston University has produced computational models of biological vision and learning that embody a growing corpus of scientific data and predictions. Vision models perform long-range grouping and figure/ground segmentation, and memory models create attentionally controlled recognition codes that intrinsically combine bottom-up activation and top-down learned expectations. These two streams of research form the foundation of novel dynamically integrated systems for image understanding. Simulations using multispectral images illustrate road completion across occlusions in a cluttered scene and information fusion from input labels that are simultaneously inconsistent and correct. The CNS Vision and Technology Labs (cns.bu.edu/visionlab and cns.bu.edu/iechlab) are further integrating science and technology through analysis, testing, and development of cognitive and neural models for large-scale applications, complemented by software specification and code distribution.
波士顿大学正在进行的研究已经产生了生物视觉和学习的计算模型,这些模型体现了越来越多的科学数据和预测。视觉模型执行远程分组和图形/地面分割,记忆模型创建注意力控制的识别代码,本质上结合了自下而上的激活和自上而下的学习期望。这两方面的研究构成了图像理解动态集成系统的基础。使用多光谱图像的模拟演示了在混乱场景中跨越遮挡的道路完成以及同时不一致和正确的输入标签的信息融合。CNS视觉和技术实验室(cns.bu.edu/visionlab和cns.bu.edu/iechlab)通过分析、测试和开发大规模应用的认知和神经模型,并辅以软件规范和代码分发,进一步整合科学和技术。
{"title":"Biologically inspired approaches to automated feature extraction and target recognition","authors":"G. Carpenter, S. Martens, E. Mingolla, Ogi J. Ogas, C. Gaddam","doi":"10.1109/AIPR.2004.17","DOIUrl":"https://doi.org/10.1109/AIPR.2004.17","url":null,"abstract":"Ongoing research at Boston University has produced computational models of biological vision and learning that embody a growing corpus of scientific data and predictions. Vision models perform long-range grouping and figure/ground segmentation, and memory models create attentionally controlled recognition codes that intrinsically combine bottom-up activation and top-down learned expectations. These two streams of research form the foundation of novel dynamically integrated systems for image understanding. Simulations using multispectral images illustrate road completion across occlusions in a cluttered scene and information fusion from input labels that are simultaneously inconsistent and correct. The CNS Vision and Technology Labs (cns.bu.edu/visionlab and cns.bu.edu/iechlab) are further integrating science and technology through analysis, testing, and development of cognitive and neural models for large-scale applications, complemented by software specification and code distribution.","PeriodicalId":120814,"journal":{"name":"33rd Applied Imagery Pattern Recognition Workshop (AIPR'04)","volume":"88 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128933850","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
System capabilities, requirements and design of the GDL gunfire detection and location system GDL炮火探测和定位系统的系统能力、需求和设计
Pub Date : 2004-10-13 DOI: 10.1109/AIPR.2004.42
J. Price, Carlos Maraviglia, William Seisler, E. Williams, M. Pauli
Using image processing techniques, the Gunfire Detection and Location (GDL) system detects gunfire and aims a suite of imagers at the muzzle flash point of origin. This detection and location function is critical for force and perimeter defense in densely populated areas as well as difficult operating environments such as a remote desert. This paper defines requirements of the GDL project. The GDL system is the result of research into using real-time image processing of mid-wave infrared imagery to detect gunfire and pin point its origin. Varieties of modern imagers are made available over the spectrum to aid an operator in assessing a detected signal. By using optical and acoustical methods, a design effort was launched to yield five vehicle based platforms. The hardware and algorithm used to implement the five basic functions is discussed in this paper. Issues such as component reliability, thermal issues, camera sensitivity operated during the daytime and nighttime, and optical design and bore sighting had to be united into a system designed to operate in the desert and powered from a high mobility multi-purpose wheeled vehicle (HMMWV). The design, construction and testing was conducted in nine months. The project has yielded a system architecture that will be further tested and refined in the next phase of this project. Experiences with the development phase of GDL and future directions are described in this paper.
使用图像处理技术,炮火探测和定位(GDL)系统探测炮火并将一套成像仪对准枪口的起始闪点。这种探测和定位功能对于人口稠密地区以及偏远沙漠等困难操作环境中的部队和外围防御至关重要。本文定义了GDL项目的需求。GDL系统是利用中波红外图像的实时图像处理来探测炮火并确定其起源的研究结果。各种现代成像仪在频谱上可用,以帮助操作员评估检测到的信号。通过使用光学和声学方法,设计工作开始产生五种基于车辆的平台。本文讨论了实现这五个基本功能的硬件和算法。组件可靠性、热问题、白天和夜间操作的相机灵敏度、光学设计和镗床瞄准等问题必须统一到一个系统中,该系统设计用于在沙漠中运行,并由高机动性多用途轮式车辆(HMMWV)提供动力。设计、建造和测试耗时9个月。该项目已经产生了一个系统架构,将在该项目的下一阶段进一步测试和完善。本文介绍了GDL发展阶段的经验和未来的发展方向。
{"title":"System capabilities, requirements and design of the GDL gunfire detection and location system","authors":"J. Price, Carlos Maraviglia, William Seisler, E. Williams, M. Pauli","doi":"10.1109/AIPR.2004.42","DOIUrl":"https://doi.org/10.1109/AIPR.2004.42","url":null,"abstract":"Using image processing techniques, the Gunfire Detection and Location (GDL) system detects gunfire and aims a suite of imagers at the muzzle flash point of origin. This detection and location function is critical for force and perimeter defense in densely populated areas as well as difficult operating environments such as a remote desert. This paper defines requirements of the GDL project. The GDL system is the result of research into using real-time image processing of mid-wave infrared imagery to detect gunfire and pin point its origin. Varieties of modern imagers are made available over the spectrum to aid an operator in assessing a detected signal. By using optical and acoustical methods, a design effort was launched to yield five vehicle based platforms. The hardware and algorithm used to implement the five basic functions is discussed in this paper. Issues such as component reliability, thermal issues, camera sensitivity operated during the daytime and nighttime, and optical design and bore sighting had to be united into a system designed to operate in the desert and powered from a high mobility multi-purpose wheeled vehicle (HMMWV). The design, construction and testing was conducted in nine months. The project has yielded a system architecture that will be further tested and refined in the next phase of this project. Experiences with the development phase of GDL and future directions are described in this paper.","PeriodicalId":120814,"journal":{"name":"33rd Applied Imagery Pattern Recognition Workshop (AIPR'04)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2004-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129071902","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
期刊
33rd Applied Imagery Pattern Recognition Workshop (AIPR'04)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1