首页 > 最新文献

2009 IEEE Applied Imagery Pattern Recognition Workshop (AIPR 2009)最新文献

英文 中文
Objective performance evaluation of a moving object super-resolution system 运动目标超分辨系统的客观性能评价
Pub Date : 2009-10-01 DOI: 10.1109/AIPR.2009.5466315
J. Laflen, C. Greco, G. Brooksby, E. Barrett
We present evaluation of the performance of moving object super-resolution (MOSR) through objective image quality metrics. MOSR systems require detection, tracking, and local sub-pixel registration of objects of interest, prior to superresolution. Nevertheless, MOSR can provide additional information otherwise undetected in raw video. We measure the extent of this benefit through the following objective image quality metrics: (1) Modulation Transfer Function (MTF), (2) Subjective Quality Factor (SQF), (3) Image Quality from the Natural Scene (MITRE IQM), and (4) minimum resolvable Rayleigh distance (RD). We also study the impact of non-ideal factors, such as image noise, frame-to-frame jitter, and object rotation, upon this performance. To study these factors, we generated controlled sequences of synthetic images of targets moving against a random field. The targets exemplified aspects of the objective metrics, containing either horizontal, vertical, or circular sinusoidal gratings, or a field of impulses separated by varying distances. High-resolution sequences were rendered and then appropriately filtered assuming a circular aperture and square, filled collector prior to decimation. A fully implemented MOSR system was used to generate super-resolved images of the moving targets. The MTF, SQF, IQM, and RD measures were acquired from each of the high, low, and super-resolved image sequences, and indicate the objective benefit of super-resolution. To contrast with MOSR, the low-resolution sequences were also up-sampled in the Fourier domain, and the objective measures were collected for these Fourier up-sampled sequences, as well. Our study consisted of over 800 different sequences, representing various combinations of non-ideal factors.
我们通过客观图像质量指标来评价运动目标超分辨率(MOSR)的性能。在超分辨率之前,MOSR系统需要对感兴趣的目标进行检测、跟踪和局部亚像素配准。然而,MOSR可以提供额外的信息,否则未检测到的原始视频。我们通过以下客观图像质量指标来衡量这种好处的程度:(1)调制传递函数(MTF),(2)主观质量因子(SQF),(3)自然场景图像质量(MITRE IQM),以及(4)最小可分辨瑞利距离(RD)。我们还研究了非理想因素,如图像噪声、帧间抖动和对象旋转对该性能的影响。为了研究这些因素,我们生成了目标在随机场运动的受控合成图像序列。目标举例说明了客观度量的各个方面,包括水平、垂直或圆形正弦光栅,或由不同距离分隔的脉冲场。高分辨率序列被渲染,然后适当地过滤,假设在抽取之前有一个圆形孔径和方形填充收集器。利用完全实现的MOSR系统生成运动目标的超分辨图像。从高分辨率、低分辨率和超分辨率图像序列中分别获得MTF、SQF、IQM和RD度量,并表明了超分辨率的客观效益。为了与MOSR相比,低分辨率序列也在傅里叶域中进行上采样,并收集这些傅里叶上采样序列的客观度量。我们的研究包括800多个不同的序列,代表了各种非理想因素的组合。
{"title":"Objective performance evaluation of a moving object super-resolution system","authors":"J. Laflen, C. Greco, G. Brooksby, E. Barrett","doi":"10.1109/AIPR.2009.5466315","DOIUrl":"https://doi.org/10.1109/AIPR.2009.5466315","url":null,"abstract":"We present evaluation of the performance of moving object super-resolution (MOSR) through objective image quality metrics. MOSR systems require detection, tracking, and local sub-pixel registration of objects of interest, prior to superresolution. Nevertheless, MOSR can provide additional information otherwise undetected in raw video. We measure the extent of this benefit through the following objective image quality metrics: (1) Modulation Transfer Function (MTF), (2) Subjective Quality Factor (SQF), (3) Image Quality from the Natural Scene (MITRE IQM), and (4) minimum resolvable Rayleigh distance (RD). We also study the impact of non-ideal factors, such as image noise, frame-to-frame jitter, and object rotation, upon this performance. To study these factors, we generated controlled sequences of synthetic images of targets moving against a random field. The targets exemplified aspects of the objective metrics, containing either horizontal, vertical, or circular sinusoidal gratings, or a field of impulses separated by varying distances. High-resolution sequences were rendered and then appropriately filtered assuming a circular aperture and square, filled collector prior to decimation. A fully implemented MOSR system was used to generate super-resolved images of the moving targets. The MTF, SQF, IQM, and RD measures were acquired from each of the high, low, and super-resolved image sequences, and indicate the objective benefit of super-resolution. To contrast with MOSR, the low-resolution sequences were also up-sampled in the Fourier domain, and the objective measures were collected for these Fourier up-sampled sequences, as well. Our study consisted of over 800 different sequences, representing various combinations of non-ideal factors.","PeriodicalId":266025,"journal":{"name":"2009 IEEE Applied Imagery Pattern Recognition Workshop (AIPR 2009)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125332920","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Automated 3D object identification using Bayesian networks 使用贝叶斯网络的自动三维物体识别
Pub Date : 2009-10-01 DOI: 10.1109/AIPR.2009.5466289
Prudhvi K. Gurram, E. Saber, F. Sahin, H. Rhody
3D object reconstruction from images involves two important parts: object identification and object modeling. Human beings are very adept at automatically identifying different objects in a scene due to the extensive training they receive over their lifetimes. Similarly, machines need to be trained to perform this task. At present, automated 3D object identification process from aerial video imagery encounters various problems due to uncertainties in data. The first problem is setting the input parameters of segmentation algorithm for accurate identification of the homogeneous surfaces in the scene. The second problem is deterministic inference used on the features extracted from these homogeneous surfaces or segments to identify different objects such as buildings, and trees. These problems would result in the 3D models being overfitted to a particular data set as a result of which they would fail when applied to other data sets. In this paper, an algorithm for using probabilistic inference to determine input segmentation parameters and to identify 3D objects from aerial video imagery is described. Bayesian networks are used to perform the probabilistic inference. In order to improve the accuracy of the identification process, information from Lidar data is fused with the visual imagery in a Bayesian network. The imagery is generated using the DIRSIG (Digital Imaging and Remote Sensing Image Generation) model at RIT. The parameters of the airborne sensor such as focal length, detector size, average flying height and the external parameters such as solar zenith angle can be simulated using this tool. The results show a significant improvement in the accuracy of object identification when Lidar data is fused with visual imagery compared to that when visual imagery is used alone.
从图像中重建三维物体包括两个重要部分:物体识别和物体建模。人类非常擅长自动识别场景中的不同物体,这是由于他们一生中接受了广泛的训练。同样,机器也需要经过训练才能完成这项任务。目前,基于航拍视频图像的三维目标自动识别过程中,由于数据的不确定性存在各种问题。第一个问题是设置分割算法的输入参数,以准确识别场景中的均匀曲面。第二个问题是对从这些同质表面或片段中提取的特征进行确定性推理,以识别不同的物体,如建筑物和树木。这些问题将导致3D模型过度拟合到特定的数据集,从而导致它们在应用于其他数据集时失败。本文描述了一种利用概率推理确定输入分割参数并从航拍视频图像中识别三维目标的算法。采用贝叶斯网络进行概率推理。为了提高识别过程的准确性,将激光雷达数据信息与视觉图像融合在贝叶斯网络中。图像是使用RIT的DIRSIG(数字成像和遥感图像生成)模型生成的。利用该工具可以模拟机载传感器的焦距、探测器尺寸、平均飞行高度等参数以及太阳天顶角等外部参数。结果表明,与单独使用视觉图像相比,激光雷达数据与视觉图像融合后的目标识别精度有显著提高。
{"title":"Automated 3D object identification using Bayesian networks","authors":"Prudhvi K. Gurram, E. Saber, F. Sahin, H. Rhody","doi":"10.1109/AIPR.2009.5466289","DOIUrl":"https://doi.org/10.1109/AIPR.2009.5466289","url":null,"abstract":"3D object reconstruction from images involves two important parts: object identification and object modeling. Human beings are very adept at automatically identifying different objects in a scene due to the extensive training they receive over their lifetimes. Similarly, machines need to be trained to perform this task. At present, automated 3D object identification process from aerial video imagery encounters various problems due to uncertainties in data. The first problem is setting the input parameters of segmentation algorithm for accurate identification of the homogeneous surfaces in the scene. The second problem is deterministic inference used on the features extracted from these homogeneous surfaces or segments to identify different objects such as buildings, and trees. These problems would result in the 3D models being overfitted to a particular data set as a result of which they would fail when applied to other data sets. In this paper, an algorithm for using probabilistic inference to determine input segmentation parameters and to identify 3D objects from aerial video imagery is described. Bayesian networks are used to perform the probabilistic inference. In order to improve the accuracy of the identification process, information from Lidar data is fused with the visual imagery in a Bayesian network. The imagery is generated using the DIRSIG (Digital Imaging and Remote Sensing Image Generation) model at RIT. The parameters of the airborne sensor such as focal length, detector size, average flying height and the external parameters such as solar zenith angle can be simulated using this tool. The results show a significant improvement in the accuracy of object identification when Lidar data is fused with visual imagery compared to that when visual imagery is used alone.","PeriodicalId":266025,"journal":{"name":"2009 IEEE Applied Imagery Pattern Recognition Workshop (AIPR 2009)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124434216","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fuzzy rule based unsupervised approach for salient gene extraction 基于模糊规则的非监督显著性基因提取方法
Pub Date : 2009-10-01 DOI: 10.1109/AIPR.2009.5466302
N. Verma, Payal Gupta, P. Agrawal, Yan Cui
This paper presents a novel fuzzy rule based gene ranking algorithm for extracting salient genes from a large set of microarray data which helps us to reduce computational efforts towards model building process. The proposed algorithm is an unsupervised approach and does not require class information for gene ranking and Microarray data has been used to form a set of robust fuzzy rule base which helps us to find salient genes based on its average relevance with already formed fuzzy rules in rule base. Fuzzy rule based ranking has been carried out to select salient genes based on their average firing strength in order of high relevancy and only top ranked genes are utilized to classify normal and cancerous tissues for a carcinoma dataset [1]. Result validate the effectiveness of our gene ranking method as for the same no. of genes, our ranking scheme helps to improve the classifier performance by selecting better salient genes.
本文提出了一种新的基于模糊规则的基因排序算法,用于从大量的微阵列数据中提取显著基因,这有助于减少模型构建过程的计算量。该算法是一种无监督的方法,不需要分类信息来进行基因排序,并使用微阵列数据形成一套鲁棒模糊规则库,根据其与规则库中已形成的模糊规则的平均相关性来帮助我们找到显著基因。基于模糊规则的排序,根据它们的平均发射强度,按照相关度高的顺序选择显著基因,只有排名靠前的基因才被用来对癌数据集的正常组织和癌组织进行分类[1]。结果验证了所提出的基因排序方法对于相同编号的基因排序方法的有效性。对于基因,我们的排序方案通过选择更好的显著基因来提高分类器的性能。
{"title":"Fuzzy rule based unsupervised approach for salient gene extraction","authors":"N. Verma, Payal Gupta, P. Agrawal, Yan Cui","doi":"10.1109/AIPR.2009.5466302","DOIUrl":"https://doi.org/10.1109/AIPR.2009.5466302","url":null,"abstract":"This paper presents a novel fuzzy rule based gene ranking algorithm for extracting salient genes from a large set of microarray data which helps us to reduce computational efforts towards model building process. The proposed algorithm is an unsupervised approach and does not require class information for gene ranking and Microarray data has been used to form a set of robust fuzzy rule base which helps us to find salient genes based on its average relevance with already formed fuzzy rules in rule base. Fuzzy rule based ranking has been carried out to select salient genes based on their average firing strength in order of high relevancy and only top ranked genes are utilized to classify normal and cancerous tissues for a carcinoma dataset [1]. Result validate the effectiveness of our gene ranking method as for the same no. of genes, our ranking scheme helps to improve the classifier performance by selecting better salient genes.","PeriodicalId":266025,"journal":{"name":"2009 IEEE Applied Imagery Pattern Recognition Workshop (AIPR 2009)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127248721","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Kalman filter based video background estimation 基于卡尔曼滤波的视频背景估计
Pub Date : 2009-10-01 DOI: 10.1109/AIPR.2009.5466306
J. Scott, M. Pusateri, Duane C. Cornish
Transferring responsibility for object tracking in a video scene to computer vision rather than human operators has the appeal that the computer will remain vigilant under all circumstances while operator attention can wane. However, when operating at their peak performance, human operators often outperform computer vision because of their ability to adapt to changes in the scene. While many tracking algorithms are available, background subtraction, where a background image is subtracted from the current frame to isolate the foreground objects in a scene, remains a well proven and popular technique. Under some circumstances, a background image can be obtained manually when no foreground objects are present. In the case of persistent surveillance outdoors, the background has a time evolution due to diurnal changes, weather, and seasonal changes. Such changes render a fixed background scene inadequate. We present a method for estimating the background of a scene utilizing a Kalman filter approach. Our method applies a one-dimensional Kalman filter to each pixel of the camera array to track the pixel intensity. We designed the algorithm to track the background intensity of a scene assuming that the camera view is relatively stationary and that the time evolution of the background occurs much slower than the time evolution of relevant foreground events. This allows the background subtraction algorithm to adapt automatically to changes in the scene. The algorithm is a two step process of mean intensity update and standard deviation update. These updates are derived from standard Kalman filter equations. Our algorithm also allows objects to transition between the background and foreground as appropriate by modeling the input standard deviation. For example, a car entering a parking lot surveillance camera field of view would initially be included in the foreground. However, once parked, it will eventually transition to the background. We present results validating our algorithm's ability to estimate backgrounds in a variety of scenes. We demonstrate the application of our method to track objects using simple frame detection with no temporal coherency.
将视频场景中目标跟踪的责任转移给计算机视觉而不是人类操作员具有吸引力,即计算机将在任何情况下保持警惕,而操作员的注意力可能会减弱。然而,当操作达到最佳性能时,由于人类操作员适应场景变化的能力,他们的表现往往优于计算机视觉。虽然有许多可用的跟踪算法,但背景减法,即从当前帧中减去背景图像以隔离场景中的前景物体,仍然是一种经过验证和流行的技术。在某些情况下,当前景对象不存在时,可以手动获得背景图像。在户外持续监测的情况下,由于昼夜变化、天气和季节变化,背景具有时间演变。这样的变化使得固定的背景场景显得不够。我们提出了一种利用卡尔曼滤波方法估计场景背景的方法。我们的方法对相机阵列的每个像素应用一维卡尔曼滤波来跟踪像素强度。我们设计了一种算法来跟踪场景的背景强度,假设相机视图相对静止,背景的时间演变比相关前景事件的时间演变慢得多。这使得背景减法算法能够自动适应场景的变化。该算法分为平均强度更新和标准差更新两步。这些更新是由标准卡尔曼滤波方程导出的。我们的算法还允许对象在背景和前景之间适当地通过建模输入标准偏差过渡。例如,一辆进入停车场的汽车,监控摄像头的视野最初将包括在前景中。然而,一旦停车,它最终会过渡到背景。我们展示的结果验证了我们的算法在各种场景中估计背景的能力。我们演示了我们的方法在使用无时间相干的简单帧检测来跟踪对象的应用。
{"title":"Kalman filter based video background estimation","authors":"J. Scott, M. Pusateri, Duane C. Cornish","doi":"10.1109/AIPR.2009.5466306","DOIUrl":"https://doi.org/10.1109/AIPR.2009.5466306","url":null,"abstract":"Transferring responsibility for object tracking in a video scene to computer vision rather than human operators has the appeal that the computer will remain vigilant under all circumstances while operator attention can wane. However, when operating at their peak performance, human operators often outperform computer vision because of their ability to adapt to changes in the scene. While many tracking algorithms are available, background subtraction, where a background image is subtracted from the current frame to isolate the foreground objects in a scene, remains a well proven and popular technique. Under some circumstances, a background image can be obtained manually when no foreground objects are present. In the case of persistent surveillance outdoors, the background has a time evolution due to diurnal changes, weather, and seasonal changes. Such changes render a fixed background scene inadequate. We present a method for estimating the background of a scene utilizing a Kalman filter approach. Our method applies a one-dimensional Kalman filter to each pixel of the camera array to track the pixel intensity. We designed the algorithm to track the background intensity of a scene assuming that the camera view is relatively stationary and that the time evolution of the background occurs much slower than the time evolution of relevant foreground events. This allows the background subtraction algorithm to adapt automatically to changes in the scene. The algorithm is a two step process of mean intensity update and standard deviation update. These updates are derived from standard Kalman filter equations. Our algorithm also allows objects to transition between the background and foreground as appropriate by modeling the input standard deviation. For example, a car entering a parking lot surveillance camera field of view would initially be included in the foreground. However, once parked, it will eventually transition to the background. We present results validating our algorithm's ability to estimate backgrounds in a variety of scenes. We demonstrate the application of our method to track objects using simple frame detection with no temporal coherency.","PeriodicalId":266025,"journal":{"name":"2009 IEEE Applied Imagery Pattern Recognition Workshop (AIPR 2009)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115852862","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 46
Detection and recognition of 3D targets in panchromatic gray scale imagery using a biologically-inspired algorithm 利用生物启发算法检测和识别全色灰度图像中的三维目标
Pub Date : 2009-10-01 DOI: 10.1109/AIPR.2009.5466310
Patricia Murphy, Pedro A. Rodriguez, Sean R. Martin
A three-dimensional (3D) target detection and recognition algorithm, using the biologically-inspired MapSeeking Circuit (MSC), is implemented to efficiently solve the typical template matching problem in computer vision. Given a 3D template model of a vehicle, this prototype locates the vehicle in a two-dimensional (2D) panchromatic image and determines its pose (i.e. viewing azimuth, elevation, scale, and in-plane rotation). In our implementation, we introduce a detection stage followed by the spawning of multiple MSC processes in parallel to classify and determine the pose of the detection candidates. Our implementation increases the speed of detection and allows efficient classification when multiple targets are present in the same image. We present promising results after applying our algorithm to challenging real world test imagery.
为了有效解决计算机视觉中典型的模板匹配问题,提出了一种基于生物寻图电路(MSC)的三维目标检测与识别算法。给定车辆的3D模板模型,该原型将车辆定位在二维(2D)全色图像中,并确定其姿态(即观看方位角、仰角、比例和平面内旋转)。在我们的实现中,我们引入了一个检测阶段,随后并行生成多个MSC进程,以分类和确定检测候选对象的姿态。我们的实现提高了检测速度,并允许在同一图像中存在多个目标时进行有效分类。在将我们的算法应用于具有挑战性的真实世界测试图像后,我们展示了有希望的结果。
{"title":"Detection and recognition of 3D targets in panchromatic gray scale imagery using a biologically-inspired algorithm","authors":"Patricia Murphy, Pedro A. Rodriguez, Sean R. Martin","doi":"10.1109/AIPR.2009.5466310","DOIUrl":"https://doi.org/10.1109/AIPR.2009.5466310","url":null,"abstract":"A three-dimensional (3D) target detection and recognition algorithm, using the biologically-inspired MapSeeking Circuit (MSC), is implemented to efficiently solve the typical template matching problem in computer vision. Given a 3D template model of a vehicle, this prototype locates the vehicle in a two-dimensional (2D) panchromatic image and determines its pose (i.e. viewing azimuth, elevation, scale, and in-plane rotation). In our implementation, we introduce a detection stage followed by the spawning of multiple MSC processes in parallel to classify and determine the pose of the detection candidates. Our implementation increases the speed of detection and allows efficient classification when multiple targets are present in the same image. We present promising results after applying our algorithm to challenging real world test imagery.","PeriodicalId":266025,"journal":{"name":"2009 IEEE Applied Imagery Pattern Recognition Workshop (AIPR 2009)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122604012","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Technical maturity evaluations for sensor fusion technologies 传感器融合技术成熟度评价
Pub Date : 2009-10-01 DOI: 10.1109/AIPR.2009.5466319
Mike Engle, S. Sarkani, T. Mazzuchi
The National Geospatial-Intelligence Agency (NGA) routinely works with commercial and academic partners to develop and refine technologies needed to meet the evolving imagery-based intelligence problems of the intelligence community (IC). There is an existing Research and Development entity within the NGA which includes the systems engineering framework required to incorporate, develop and transition applicable technologies for use by analysts. In order to better understand where work may fall within this framework, it is necessary to identify the inherent technical maturity of the research in question. Technology Readiness Levels (TRLs), which were originally developed by NASA and are used by the DOD for most development and procurement programs, are used by NGA as a quick indication of both technical maturity and inherent risk (technical, schedule, cost or transition). This paper discusses the different GEOINT-focused performance evaluations pertinent to the TRLs then provides a brief introduction to an applicable multi-sensor data fusion framework.
美国国家地理空间情报局(NGA)定期与商业和学术伙伴合作,开发和完善所需的技术,以满足情报界(IC)不断发展的基于图像的情报问题。在NGA中有一个现有的研究和发展实体,它包括系统工程框架,需要合并、开发和转换适用于分析人员使用的技术。为了更好地理解工作在这个框架中的位置,有必要确定所讨论的研究的固有技术成熟度。技术就绪水平(trl)最初由NASA开发,并被国防部用于大多数开发和采购计划,NGA将其用作技术成熟度和固有风险(技术、进度、成本或过渡)的快速指示。本文讨论了与trl相关的不同的以geoint为中心的性能评估,然后简要介绍了适用的多传感器数据融合框架。
{"title":"Technical maturity evaluations for sensor fusion technologies","authors":"Mike Engle, S. Sarkani, T. Mazzuchi","doi":"10.1109/AIPR.2009.5466319","DOIUrl":"https://doi.org/10.1109/AIPR.2009.5466319","url":null,"abstract":"The National Geospatial-Intelligence Agency (NGA) routinely works with commercial and academic partners to develop and refine technologies needed to meet the evolving imagery-based intelligence problems of the intelligence community (IC). There is an existing Research and Development entity within the NGA which includes the systems engineering framework required to incorporate, develop and transition applicable technologies for use by analysts. In order to better understand where work may fall within this framework, it is necessary to identify the inherent technical maturity of the research in question. Technology Readiness Levels (TRLs), which were originally developed by NASA and are used by the DOD for most development and procurement programs, are used by NGA as a quick indication of both technical maturity and inherent risk (technical, schedule, cost or transition). This paper discusses the different GEOINT-focused performance evaluations pertinent to the TRLs then provides a brief introduction to an applicable multi-sensor data fusion framework.","PeriodicalId":266025,"journal":{"name":"2009 IEEE Applied Imagery Pattern Recognition Workshop (AIPR 2009)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128857867","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
3D shape retrieval by visual parts similarity 基于视觉零件相似度的三维形状检索
Pub Date : 2009-10-01 DOI: 10.1109/AIPR.2009.5466316
A. Godil, A. I. Wagan, S. Bres, Xiaolan Li
In this paper we propose a novel algorithm for 3D shape searching based on the visual similarity by cutting the object into parts. This method rectify some of the shortcomings of the visual similarity based methods, so that it can better account for objects with deformation, articulation, concave areas, and parts of the object not visible because of self occlusion. As the first step, the 3D objects are partitioned into a number of parts by using cutting planes or by mesh segmentation. Then a number of silhouettes from different directions are rendered of those parts. Then Zernike moments are applied on the silhouettes to generate shape descriptors. The distance measure is based on minimizing the distance among all the combinations of shape descriptors and then these distances are used for similarity based searching.
本文提出了一种基于视觉相似性的三维形状搜索算法。该方法纠正了基于视觉相似度方法的一些缺点,使其能够更好地解释具有变形、衔接、凹区域以及物体因自身遮挡而不可见的部分。首先,通过切割平面或网格分割将三维物体分割成多个部分。然后从不同的方向渲染这些部分的一些剪影。然后在轮廓上应用泽尼克矩来生成形状描述符。距离度量是基于最小化所有形状描述符组合之间的距离,然后将这些距离用于基于相似性的搜索。
{"title":"3D shape retrieval by visual parts similarity","authors":"A. Godil, A. I. Wagan, S. Bres, Xiaolan Li","doi":"10.1109/AIPR.2009.5466316","DOIUrl":"https://doi.org/10.1109/AIPR.2009.5466316","url":null,"abstract":"In this paper we propose a novel algorithm for 3D shape searching based on the visual similarity by cutting the object into parts. This method rectify some of the shortcomings of the visual similarity based methods, so that it can better account for objects with deformation, articulation, concave areas, and parts of the object not visible because of self occlusion. As the first step, the 3D objects are partitioned into a number of parts by using cutting planes or by mesh segmentation. Then a number of silhouettes from different directions are rendered of those parts. Then Zernike moments are applied on the silhouettes to generate shape descriptors. The distance measure is based on minimizing the distance among all the combinations of shape descriptors and then these distances are used for similarity based searching.","PeriodicalId":266025,"journal":{"name":"2009 IEEE Applied Imagery Pattern Recognition Workshop (AIPR 2009)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115754763","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Spatial-spectral cross correlation for reliable multispectral image registration 可靠的多光谱图像配准的空间光谱相互关
Pub Date : 2009-10-01 DOI: 10.1109/AIPR.2009.5466291
Zhengwei Yang, Guangrong Shen, Wei Wang, Zhenhua Qian, Ying Ke
This paper presents a normalized spatial-spectral cross correlation method for multispectral image registration. This method generalized correlation coefficients defined in a spatial domain or a spectral domain into a spatial-spectral domain. This novel spatial-spectral signature based method significantly increases the discrimination of the correlation coefficient for a given template window size, increases the registration reliability, robustness and accuracy, as compared with the classic normalized spatial cross correlation method. It is invariant to the dynamic range and robust to the noise yet it is straightforward with minimum preprocessing required. The experimental results show that the normalized spatial-spectral cross correlation method is superior to the traditional normalized spatial cross correlation method in effective registering multispectral images. However, the experimental results also show that only those statistically highly independent spectral bands are helpful for enhancing the robustness and reliability of the NSSCC multispectral image registration. Specifically, it is found that the near infrared band together with visual bands will gives the best registration results.
提出了一种用于多光谱图像配准的归一化空间-光谱相互关方法。该方法将在空间域或谱域定义的相关系数推广到空间-谱域。与传统的归一化空间互相关方法相比,该方法在给定模板窗口大小的情况下显著提高了相关系数的判别能力,提高了配准的可靠性、鲁棒性和准确性。该方法对动态范围具有不变性,对噪声具有鲁棒性,且简单易行,所需的预处理最少。实验结果表明,归一化空间-光谱互关方法在多光谱图像配准方面优于传统的归一化空间互关方法。然而,实验结果也表明,只有统计上高度独立的光谱带才有助于提高NSSCC多光谱图像配准的鲁棒性和可靠性。具体来说,发现近红外波段与可见光波段结合可以得到最好的配准效果。
{"title":"Spatial-spectral cross correlation for reliable multispectral image registration","authors":"Zhengwei Yang, Guangrong Shen, Wei Wang, Zhenhua Qian, Ying Ke","doi":"10.1109/AIPR.2009.5466291","DOIUrl":"https://doi.org/10.1109/AIPR.2009.5466291","url":null,"abstract":"This paper presents a normalized spatial-spectral cross correlation method for multispectral image registration. This method generalized correlation coefficients defined in a spatial domain or a spectral domain into a spatial-spectral domain. This novel spatial-spectral signature based method significantly increases the discrimination of the correlation coefficient for a given template window size, increases the registration reliability, robustness and accuracy, as compared with the classic normalized spatial cross correlation method. It is invariant to the dynamic range and robust to the noise yet it is straightforward with minimum preprocessing required. The experimental results show that the normalized spatial-spectral cross correlation method is superior to the traditional normalized spatial cross correlation method in effective registering multispectral images. However, the experimental results also show that only those statistically highly independent spectral bands are helpful for enhancing the robustness and reliability of the NSSCC multispectral image registration. Specifically, it is found that the near infrared band together with visual bands will gives the best registration results.","PeriodicalId":266025,"journal":{"name":"2009 IEEE Applied Imagery Pattern Recognition Workshop (AIPR 2009)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133412750","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Large-scale functional models of visual cortex for remote sensing 遥感视觉皮层大尺度功能模型
Pub Date : 2009-10-01 DOI: 10.1109/AIPR.2009.5466323
S. Brumby, Garrett T. Kenyon, Will Landecker, Craig Rasmussen, S. Swaminarayan, L. Bettencourt
Neuroscience has revealed many properties of neurons and of the functional organization of visual cortex that are believed to be essential to human vision, but are missing in standard artificial neural networks. Equally important may be the sheer scale of visual cortex requiring ~1 petaflop of computation, while the scale of human visual experience greatly exceeds standard computer vision datasets: the retina delivers ~1 petapixel/year to the brain, driving learning at many levels of the cortical system. We describe work at Los Alamos National Laboratory (LANL) to develop large-scale functional models of visual cortex on LANL's Roadrunner petaflop supercomputer. An initial run of a simple region V1 code achieved 1.144 petaflops during trials at the IBM facility in Poughkeepsie, NY (June 2008). Here, we present criteria for assessing when a set of learned local representations is ¿complete¿ along with general criteria for assessing computer vision models based on their projected scaling behavior. Finally, we extend one class of biologically-inspired learning models to problems of remote sensing imagery.
神经科学揭示了神经元和视觉皮层功能组织的许多特性,这些特性被认为对人类视觉至关重要,但在标准的人工神经网络中却缺失了。同样重要的可能是视觉皮层的绝对规模需要1千万亿次的计算,而人类视觉经验的规模大大超过了标准的计算机视觉数据集:视网膜每年向大脑传递1千万亿次,推动皮层系统的许多层面的学习。我们描述了在洛斯阿拉莫斯国家实验室(LANL)在LANL的Roadrunner petaflop超级计算机上开发大规模视觉皮层功能模型的工作。2008年6月,在纽约波基普西的IBM设施中,一个简单的区域V1代码的初步运行达到了每秒1.144千万亿次。在这里,我们提出了评估一组学习到的局部表示何时“完整”的标准,以及基于其投影缩放行为评估计算机视觉模型的一般标准。最后,我们将一类受生物启发的学习模型扩展到遥感图像问题。
{"title":"Large-scale functional models of visual cortex for remote sensing","authors":"S. Brumby, Garrett T. Kenyon, Will Landecker, Craig Rasmussen, S. Swaminarayan, L. Bettencourt","doi":"10.1109/AIPR.2009.5466323","DOIUrl":"https://doi.org/10.1109/AIPR.2009.5466323","url":null,"abstract":"Neuroscience has revealed many properties of neurons and of the functional organization of visual cortex that are believed to be essential to human vision, but are missing in standard artificial neural networks. Equally important may be the sheer scale of visual cortex requiring ~1 petaflop of computation, while the scale of human visual experience greatly exceeds standard computer vision datasets: the retina delivers ~1 petapixel/year to the brain, driving learning at many levels of the cortical system. We describe work at Los Alamos National Laboratory (LANL) to develop large-scale functional models of visual cortex on LANL's Roadrunner petaflop supercomputer. An initial run of a simple region V1 code achieved 1.144 petaflops during trials at the IBM facility in Poughkeepsie, NY (June 2008). Here, we present criteria for assessing when a set of learned local representations is ¿complete¿ along with general criteria for assessing computer vision models based on their projected scaling behavior. Finally, we extend one class of biologically-inspired learning models to problems of remote sensing imagery.","PeriodicalId":266025,"journal":{"name":"2009 IEEE Applied Imagery Pattern Recognition Workshop (AIPR 2009)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115339265","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
Persistence and tracking: Putting vehicles and trajectories in context 持久性和跟踪:将车辆和轨迹置于上下文中
Pub Date : 2009-10-01 DOI: 10.1109/AIPR.2009.5466307
Robert Pless, M. Dixon, Nathan Jacobs, P. Baker, Nicholas L. Cassimatis, Derek P. Brock, R. Hartley, Dennis Perzanowski
City-scale tracking of all objects visible in a camera network or aerial video surveillance is an important tool in surveillance and traffic monitoring. We propose a framework for human guided tracking based on explicitly considering the context surrounding the urban multi-vehicle tracking problem. This framework is based on a standard (but state of the art) probabilistic tracking model. Our contribution is to explicitly detail where human annotation of the scene (e.g. “this is a lane”), a track (e.g. “this track is bad”), or a pair of tracks (e.g. “these two tracks are confused”) can be naturally integrated within the probabilistic tracking framework. For an early prototype system, we offer results and examples from a dense urban traffic camera network tracking, querying data with thousands of vehicles over 30 minutes.
在城市范围内对所有可见物体进行跟踪的摄像机网络或航空视频监控是监控和交通监控的重要工具。在明确考虑城市多车跟踪问题背景的基础上,提出了一种人类引导跟踪框架。这个框架是基于一个标准的(但是最先进的)概率跟踪模型。我们的贡献是明确地详细说明人类对场景的注释(例如,“这是一条车道”)、一条轨道(例如,“这条轨道很糟糕”)或一对轨道(例如,“这两条轨道混淆了”)可以自然地集成在概率跟踪框架中。对于一个早期的原型系统,我们提供了一个密集的城市交通摄像头网络跟踪的结果和示例,在30分钟内查询了数千辆汽车的数据。
{"title":"Persistence and tracking: Putting vehicles and trajectories in context","authors":"Robert Pless, M. Dixon, Nathan Jacobs, P. Baker, Nicholas L. Cassimatis, Derek P. Brock, R. Hartley, Dennis Perzanowski","doi":"10.1109/AIPR.2009.5466307","DOIUrl":"https://doi.org/10.1109/AIPR.2009.5466307","url":null,"abstract":"City-scale tracking of all objects visible in a camera network or aerial video surveillance is an important tool in surveillance and traffic monitoring. We propose a framework for human guided tracking based on explicitly considering the context surrounding the urban multi-vehicle tracking problem. This framework is based on a standard (but state of the art) probabilistic tracking model. Our contribution is to explicitly detail where human annotation of the scene (e.g. “this is a lane”), a track (e.g. “this track is bad”), or a pair of tracks (e.g. “these two tracks are confused”) can be naturally integrated within the probabilistic tracking framework. For an early prototype system, we offer results and examples from a dense urban traffic camera network tracking, querying data with thousands of vehicles over 30 minutes.","PeriodicalId":266025,"journal":{"name":"2009 IEEE Applied Imagery Pattern Recognition Workshop (AIPR 2009)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133631716","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
期刊
2009 IEEE Applied Imagery Pattern Recognition Workshop (AIPR 2009)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1