首页 > 最新文献

2010 IEEE 39th Applied Imagery Pattern Recognition Workshop (AIPR)最新文献

英文 中文
A combinatorial method for tracing objects using semantics of their shape 一种利用物体形状语义跟踪物体的组合方法
Pub Date : 2010-10-01 DOI: 10.1109/AIPR.2010.5759716
C. Diegert
We present a shape-first approach to finding automobiles and trucks in overhead images and include results from our analysis of an image from the Overhead Imaging Research Dataset [1]. For the OIRDS, our shape-first approach traces candidate vehicle outlines by exploiting knowledge about an overhead image of a vehicle: a vehicle's outline fits into a rectangle, this rectangle is sized to allow vehicles to use local roads, and rectangles from two different vehicles are disjoint. Our shape-first approach can efficiently process high-resolution overhead imaging over wide areas to provide tips and cues for human analysts, or for subsequent automatic processing using machine learning or other analysis based on color, tone, pattern, texture, size, and/or location (shape first). In fact, computationally-intensive complex structural, syntactic, and statistical analysis may be possible when a shape-first work flow sends a list of specific tips and cues down a processing pipeline rather than sending the whole of wide area imaging information. This data flow may fit well when bandwidth is limited between computers delivering ad hoc image exploitation and an imaging sensor. As expected, our early computational experiments find that the shape-first processing stage appears to reliably detect rectangular shapes from vehicles. More intriguing is that our computational experiments with six-inch GSD OIRDS benchmark images show that the shape-first stage can be efficient, and that candidate vehicle locations corresponding to features that do not include vehicles are unlikely to trigger tips and cues. We found that stopping with just the shape-first list of candidate vehicle locations, and then solving a weighted, maximal independent vertex set problem to resolve conflicts among candidate vehicle locations, often correctly traces the vehicles in an OIRDS scene.
我们提出了一种形状优先的方法来在架空图像中寻找汽车和卡车,并包括我们对架空成像研究数据集[1]中的图像的分析结果。对于OIRDS,我们的形状优先方法通过利用有关车辆头顶图像的知识来跟踪候选车辆轮廓:车辆的轮廓适合矩形,该矩形的大小允许车辆使用本地道路,来自两辆不同车辆的矩形是不相交的。我们的形状优先方法可以有效地处理大范围的高分辨率头顶成像,为人类分析师提供提示和线索,或者使用机器学习或基于颜色、色调、图案、纹理、大小和/或位置(形状优先)的其他分析进行后续自动处理。事实上,当形状优先的工作流程将特定提示和线索的列表发送到处理管道中,而不是发送整个广域成像信息时,计算密集型复杂结构、语法和统计分析是可能的。这种数据流可能很适合带宽有限的计算机之间提供特别的图像开发和成像传感器。正如预期的那样,我们的早期计算实验发现,形状优先处理阶段似乎可以可靠地检测到车辆的矩形形状。更有趣的是,我们对6英寸GSD OIRDS基准图像进行的计算实验表明,形状第一阶段是有效的,与不包括车辆的特征相对应的候选车辆位置不太可能触发提示和线索。我们发现,仅停止候选车辆位置的形状优先列表,然后求解加权的最大独立顶点集问题来解决候选车辆位置之间的冲突,通常可以正确地跟踪OIRDS场景中的车辆。
{"title":"A combinatorial method for tracing objects using semantics of their shape","authors":"C. Diegert","doi":"10.1109/AIPR.2010.5759716","DOIUrl":"https://doi.org/10.1109/AIPR.2010.5759716","url":null,"abstract":"We present a shape-first approach to finding automobiles and trucks in overhead images and include results from our analysis of an image from the Overhead Imaging Research Dataset [1]. For the OIRDS, our shape-first approach traces candidate vehicle outlines by exploiting knowledge about an overhead image of a vehicle: a vehicle's outline fits into a rectangle, this rectangle is sized to allow vehicles to use local roads, and rectangles from two different vehicles are disjoint. Our shape-first approach can efficiently process high-resolution overhead imaging over wide areas to provide tips and cues for human analysts, or for subsequent automatic processing using machine learning or other analysis based on color, tone, pattern, texture, size, and/or location (shape first). In fact, computationally-intensive complex structural, syntactic, and statistical analysis may be possible when a shape-first work flow sends a list of specific tips and cues down a processing pipeline rather than sending the whole of wide area imaging information. This data flow may fit well when bandwidth is limited between computers delivering ad hoc image exploitation and an imaging sensor. As expected, our early computational experiments find that the shape-first processing stage appears to reliably detect rectangular shapes from vehicles. More intriguing is that our computational experiments with six-inch GSD OIRDS benchmark images show that the shape-first stage can be efficient, and that candidate vehicle locations corresponding to features that do not include vehicles are unlikely to trigger tips and cues. We found that stopping with just the shape-first list of candidate vehicle locations, and then solving a weighted, maximal independent vertex set problem to resolve conflicts among candidate vehicle locations, often correctly traces the vehicles in an OIRDS scene.","PeriodicalId":128378,"journal":{"name":"2010 IEEE 39th Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128839726","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Object identification in dynamic environment using sensor fusion 基于传感器融合的动态环境目标识别
Pub Date : 2010-10-01 DOI: 10.1109/AIPR.2010.5759682
K. S. Nagla, M. Uddin, Dilbag Singh, Rajeev Kumar
Multisensor data fusion is highly applicable in robotics applications because the relationships among objects and events changes due to the change in orientation of robot, snag in sensory information, sensor range and environmental conditions etc. High level and low level image processing in machine vision are widely involved to investigate object identification in complex application. Due to the limitations of vision technology still it is difficult to identify the objects in certain environments. A new technique of object identification using sonar sensor fusion has been proposed. This paper explains the computational account of the data fusion using Bayesian and neural network to recognize the shape of object in the dynamic environment.
多传感器数据融合在机器人应用中具有很高的应用价值,因为物体和事件之间的关系会因机器人方向的变化、感知信息的中断、传感器范围和环境条件等而发生变化。机器视觉中的高级和低级图像处理被广泛应用于研究复杂应用中的目标识别。由于视觉技术的限制,在某些环境中识别物体仍然存在困难。提出了一种基于声纳传感器融合的目标识别新技术。本文阐述了利用贝叶斯和神经网络进行动态环境中物体形状识别的数据融合的计算过程。
{"title":"Object identification in dynamic environment using sensor fusion","authors":"K. S. Nagla, M. Uddin, Dilbag Singh, Rajeev Kumar","doi":"10.1109/AIPR.2010.5759682","DOIUrl":"https://doi.org/10.1109/AIPR.2010.5759682","url":null,"abstract":"Multisensor data fusion is highly applicable in robotics applications because the relationships among objects and events changes due to the change in orientation of robot, snag in sensory information, sensor range and environmental conditions etc. High level and low level image processing in machine vision are widely involved to investigate object identification in complex application. Due to the limitations of vision technology still it is difficult to identify the objects in certain environments. A new technique of object identification using sonar sensor fusion has been proposed. This paper explains the computational account of the data fusion using Bayesian and neural network to recognize the shape of object in the dynamic environment.","PeriodicalId":128378,"journal":{"name":"2010 IEEE 39th Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"81 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126416323","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Rock image segmentation using watershed with shape markers 带形状标记的分水岭岩石图像分割
Pub Date : 2010-10-01 DOI: 10.1109/AIPR.2010.5759719
A. Amankwah, C. Aldrich
We propose a method for the creation of object markers used in watershed segmentation of rock images. First, we use adaptive thresholding to segment the rock image since rock particles local background is often different from surrounding particle regions. Object markers are then extracted using the compactness of objects and adaptive morphological reconstruction. The choice of the feature compactness is motivated by the fact that crushed rocks tend to have rounded shapes. Experimental results after comparing the segmented images show that the performance of our algorithm is superior to most standard methods of watershed segmentation. We also show that the proposed algorithm was more robust in the estimation of fines in rock samples than the traditional methods.
我们提出了一种用于岩石图像分水岭分割的目标标记的创建方法。首先,由于岩石颗粒的局部背景通常与周围颗粒区域不同,我们采用自适应阈值分割方法对岩石图像进行分割。然后利用对象的紧密度和自适应形态重建提取对象标记。选择紧凑性的原因是碎石往往呈圆形。实验结果表明,该算法的分割性能优于大多数标准的分水岭分割方法。我们还表明,与传统方法相比,该算法在估计岩石样品中的细粒方面具有更强的鲁棒性。
{"title":"Rock image segmentation using watershed with shape markers","authors":"A. Amankwah, C. Aldrich","doi":"10.1109/AIPR.2010.5759719","DOIUrl":"https://doi.org/10.1109/AIPR.2010.5759719","url":null,"abstract":"We propose a method for the creation of object markers used in watershed segmentation of rock images. First, we use adaptive thresholding to segment the rock image since rock particles local background is often different from surrounding particle regions. Object markers are then extracted using the compactness of objects and adaptive morphological reconstruction. The choice of the feature compactness is motivated by the fact that crushed rocks tend to have rounded shapes. Experimental results after comparing the segmented images show that the performance of our algorithm is superior to most standard methods of watershed segmentation. We also show that the proposed algorithm was more robust in the estimation of fines in rock samples than the traditional methods.","PeriodicalId":128378,"journal":{"name":"2010 IEEE 39th Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115252359","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
A probabilistic framework for unsupervised evaluation and ranking of image segmentations 一种用于图像分割的无监督评价和排序的概率框架
Pub Date : 2010-10-01 DOI: 10.1109/AIPR.2010.5759690
M. Jaber, S. R. Vantaram, E. Saber
In this paper, a Bayesian Network (BN) framework for unsupervised evaluation of image segmentation quality is proposed. This image understanding algorithm utilizes a set of given Segmentation Maps (SMs) ranging from under-segmented to over-segmented results for a target image, to identify the semantically meaningful ones and rank the SMs according to their applicability in image processing and computer vision systems. Images acquired from the Berkeley segmentation dataset along with their corresponding SMs are used to train and test the proposed algorithm. Low-level local and global image features are employed to define an optimal BN structure and to estimate the inference between its nodes. Furthermore, given several SMs of a test image, the optimal BN is utilized to estimate the probability that a given map is the most favorable segmentation for that image. The algorithm is evaluated on a separate set of images (none of which are included in the training set) wherein the ranked SMs (according to their probabilities of being acceptable segmentation as estimated by the proposed algorithm) are compared to the ground-truth maps generated by human observers. The Normalized Probabilistic Rand (NPR) index is used as an objective metric to quantify our algorithm's performance. The proposed algorithm is designed to serve as a pre-processing module in various bottom-up image processing frameworks such as content-based image retrieval and region-of-interest detection.
本文提出了一种用于图像分割质量无监督评价的贝叶斯网络框架。这种图像理解算法利用一组给定的分割图(Segmentation Maps, SMs),从目标图像的未分割到过度分割的结果,来识别语义上有意义的分割图,并根据它们在图像处理和计算机视觉系统中的适用性对这些分割图进行排序。从Berkeley分割数据集中获取的图像及其相应的SMs用于训练和测试所提出的算法。使用低层次的局部和全局图像特征来定义最优的BN结构并估计其节点之间的推断。此外,给定测试图像的几个SMs,使用最优BN来估计给定映射是该图像最有利分割的概率。该算法在一组单独的图像上进行评估(这些图像都不包括在训练集中),其中排名的SMs(根据所提出的算法估计的可接受分割的概率)与人类观察者生成的真实地图进行比较。采用归一化概率兰德(NPR)指标作为客观指标来量化算法的性能。该算法被设计为各种自下而上的图像处理框架(如基于内容的图像检索和兴趣区域检测)的预处理模块。
{"title":"A probabilistic framework for unsupervised evaluation and ranking of image segmentations","authors":"M. Jaber, S. R. Vantaram, E. Saber","doi":"10.1109/AIPR.2010.5759690","DOIUrl":"https://doi.org/10.1109/AIPR.2010.5759690","url":null,"abstract":"In this paper, a Bayesian Network (BN) framework for unsupervised evaluation of image segmentation quality is proposed. This image understanding algorithm utilizes a set of given Segmentation Maps (SMs) ranging from under-segmented to over-segmented results for a target image, to identify the semantically meaningful ones and rank the SMs according to their applicability in image processing and computer vision systems. Images acquired from the Berkeley segmentation dataset along with their corresponding SMs are used to train and test the proposed algorithm. Low-level local and global image features are employed to define an optimal BN structure and to estimate the inference between its nodes. Furthermore, given several SMs of a test image, the optimal BN is utilized to estimate the probability that a given map is the most favorable segmentation for that image. The algorithm is evaluated on a separate set of images (none of which are included in the training set) wherein the ranked SMs (according to their probabilities of being acceptable segmentation as estimated by the proposed algorithm) are compared to the ground-truth maps generated by human observers. The Normalized Probabilistic Rand (NPR) index is used as an objective metric to quantify our algorithm's performance. The proposed algorithm is designed to serve as a pre-processing module in various bottom-up image processing frameworks such as content-based image retrieval and region-of-interest detection.","PeriodicalId":128378,"journal":{"name":"2010 IEEE 39th Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122921052","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Use of remote sensing to screen earthen levees 利用遥感技术筛选土堤
Pub Date : 2010-10-01 DOI: 10.1109/AIPR.2010.5759704
J. Aanstoos, K. Hasan, C. O'Hara, S. Prasad, Lalitha Dabbiru, Majid Mahrooghy, R. Nóbrega, Matthew A. Lee, B. Shrestha
Multi-polarized L-band Synthetic Aperture Radar is investigated for its potential to screen earthen levees for weak points. Various feature detection and classification algorithms are tested for this application, including both radiometric and textural methods such as grey-level co-occurrence matrix and wavelet features.
研究了多极化l波段合成孔径雷达对土堤薄弱点的屏蔽潜力。各种特征检测和分类算法进行了测试,包括辐射和纹理方法,如灰度共现矩阵和小波特征。
{"title":"Use of remote sensing to screen earthen levees","authors":"J. Aanstoos, K. Hasan, C. O'Hara, S. Prasad, Lalitha Dabbiru, Majid Mahrooghy, R. Nóbrega, Matthew A. Lee, B. Shrestha","doi":"10.1109/AIPR.2010.5759704","DOIUrl":"https://doi.org/10.1109/AIPR.2010.5759704","url":null,"abstract":"Multi-polarized L-band Synthetic Aperture Radar is investigated for its potential to screen earthen levees for weak points. Various feature detection and classification algorithms are tested for this application, including both radiometric and textural methods such as grey-level co-occurrence matrix and wavelet features.","PeriodicalId":128378,"journal":{"name":"2010 IEEE 39th Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"72 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126345594","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
Material detection with a CCD polarization imager 用CCD偏振成像仪进行材料检测
Pub Date : 2010-10-01 DOI: 10.1109/AIPR.2010.5759710
V. Gruev, Rob Perkins, Timothy York
We present a novel polarization image sensor by monolithically integrating aluminum nanowire optical filters with CCD imaging array. The CCD polarization image sensor is composed of 1000 by 1000 imaging elements with 7.4μm pixel pitch. The image sensor has a dynamic range of 65dB and signal-to-noise ratio of 60dB. The CCD array is covered with an array of pixel-pitch matched nanowire polarization filters with four different orientations offset by 45°. Raw polarization data is presented to a DSP board at 40 frames per second, where degree and angle of polarization is computed. The final polarization results are presented in false color representation. The imaging sensor is used to detect the index of refraction of several flat surfaces.
提出了一种将铝纳米线滤光片与CCD成像阵列单片集成的新型偏振图像传感器。CCD偏振图像传感器由1000 × 1000个成像单元组成,像素间距为7.4μm。图像传感器动态范围为65dB,信噪比为60dB。CCD阵列上覆盖了一组像素间距匹配的纳米线偏振滤波器,四个不同的方向偏移45°。将原始偏振数据以每秒40帧的速度传送到DSP板上,计算偏振度和偏振角。最后的偏振结果用假色表示。该成像传感器用于检测几种平面的折射率。
{"title":"Material detection with a CCD polarization imager","authors":"V. Gruev, Rob Perkins, Timothy York","doi":"10.1109/AIPR.2010.5759710","DOIUrl":"https://doi.org/10.1109/AIPR.2010.5759710","url":null,"abstract":"We present a novel polarization image sensor by monolithically integrating aluminum nanowire optical filters with CCD imaging array. The CCD polarization image sensor is composed of 1000 by 1000 imaging elements with 7.4μm pixel pitch. The image sensor has a dynamic range of 65dB and signal-to-noise ratio of 60dB. The CCD array is covered with an array of pixel-pitch matched nanowire polarization filters with four different orientations offset by 45°. Raw polarization data is presented to a DSP board at 40 frames per second, where degree and angle of polarization is computed. The final polarization results are presented in false color representation. The imaging sensor is used to detect the index of refraction of several flat surfaces.","PeriodicalId":128378,"journal":{"name":"2010 IEEE 39th Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125798785","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A system and method for auto-correction of first order lens distortion 一种一阶透镜畸变自动校正系统及方法
Pub Date : 2010-10-01 DOI: 10.1109/AIPR.2010.5759696
Jonathan Fry, M. Pusateri
In multispectral imaging systems, correction for lens distortion is required to allow pixel by pixel fusion techniques to be applied. While correction of optical aberration can be extended to higher order terms, for many systems, a first order correction is sufficient to achieve desired results. In producing a multispectral imaging system in production quantities, the process of producing the corrections needs to be largely automated as each lens will require its own corrections. We discuss an auto-correction and bench sighting method application to a dual band imaging system. In principle, we wish to image a dual band target and completely determine the lens distortion parameters for the given optics. We begin with a scale-preserving, radial, first-order lens distortion model; this model allows the horizontal field of view to be determined independently of the distortion. It has the benefits of simple parameterization and the ability to correct mild to moderate distortion that may be expected of production optics. The correction process starts with imaging a dual band target. A feature extraction algorithm is applied to the imagery from both bands to generate a large number of correlated feature points. Using the feature points, we derive an over-determined system of equations; the solution to this system yields the distortion parameters for the lens. Using these parameters, an interpolation map can be generated unique to the lenses involved. The interpolation map is used in real-time to correct the distortion while preserving the horizontal field of view constraint on the system.
在多光谱成像系统中,需要对透镜畸变进行校正,以便应用逐像素融合技术。虽然光学像差的校正可以扩展到高阶项,但对于许多系统,一阶校正足以达到预期的结果。在批量生产多光谱成像系统时,产生校正的过程需要在很大程度上自动化,因为每个镜头都需要自己的校正。讨论了一种自动校正和台架瞄准方法在双波段成像系统中的应用。原则上,我们希望对双波段目标成像,并完全确定给定光学器件的透镜畸变参数。我们从一个保持尺度的径向一阶透镜畸变模型开始;该模型允许水平视场的确定独立于畸变。它具有简单的参数化和纠正生产光学器件可能出现的轻度到中度畸变的能力。校正过程从对双波段目标成像开始。对两个波段的图像采用特征提取算法,生成大量相关的特征点。利用特征点,导出了一个过定方程组;对该系统的求解得到了透镜的畸变参数。使用这些参数,可以为所涉及的镜头生成唯一的插值图。在保持系统水平视场约束的前提下,利用插值图实时校正畸变。
{"title":"A system and method for auto-correction of first order lens distortion","authors":"Jonathan Fry, M. Pusateri","doi":"10.1109/AIPR.2010.5759696","DOIUrl":"https://doi.org/10.1109/AIPR.2010.5759696","url":null,"abstract":"In multispectral imaging systems, correction for lens distortion is required to allow pixel by pixel fusion techniques to be applied. While correction of optical aberration can be extended to higher order terms, for many systems, a first order correction is sufficient to achieve desired results. In producing a multispectral imaging system in production quantities, the process of producing the corrections needs to be largely automated as each lens will require its own corrections. We discuss an auto-correction and bench sighting method application to a dual band imaging system. In principle, we wish to image a dual band target and completely determine the lens distortion parameters for the given optics. We begin with a scale-preserving, radial, first-order lens distortion model; this model allows the horizontal field of view to be determined independently of the distortion. It has the benefits of simple parameterization and the ability to correct mild to moderate distortion that may be expected of production optics. The correction process starts with imaging a dual band target. A feature extraction algorithm is applied to the imagery from both bands to generate a large number of correlated feature points. Using the feature points, we derive an over-determined system of equations; the solution to this system yields the distortion parameters for the lens. Using these parameters, an interpolation map can be generated unique to the lenses involved. The interpolation map is used in real-time to correct the distortion while preserving the horizontal field of view constraint on the system.","PeriodicalId":128378,"journal":{"name":"2010 IEEE 39th Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126875181","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Gray-level co-occurrence matrices as features in edge enhanced images 灰度共现矩阵作为边缘增强图像的特征
Pub Date : 2010-10-01 DOI: 10.1109/AIPR.2010.5759705
Peter J. Costianes, Joseph B. Plock
In 1973, Haralick, Shanmugam, and Dinstein published a paper in the IEEE Transactions on Systems, Man, and Cybernetics which proposed using Gray-Level Cooccurrence Matrices (GLCM) as a basis to define 2-D texture1. Over 14 different texture measures were defined using GLCM. In images with n × n grey levels, the size of the GLCM would be n × n which, for large n such as n=256, put a large computational load on the process and was also best suited for pixel distributions that were rather stochastic in nature. Such features as entropy, variance, correlation, etc. were proposed using the GLCM. When attempting to provide feature measures for man-made targets, most of the information contained in the target is contained by its edge distribution. Previous approaches form an edge outline of the target and then use some techniques such as Fourier descriptors to represent the target. However, in this case, extra steps need to be taken in order to assure that the edge outline is continuous or gaps in the outline somehow are dealt with when creating the Fourier coefficients for the feature vector. This paper presents an approach using GLCM where the gray scale image is put through an edge enhancement using any one of several edge operators. The resultant image is a binary image. For each point in the edge image, a 2×2 GLCM is created by placing an n × n window centered around the point and using the n2 neighboring points to build the GLCM's. This window should be sufficiently large to enclose the target of interest and the GLCM created provides the elements needed to define the features for the edge enhanced target. All software was created in Matlab2 using Matlab functions.
1973年,Haralick、Shanmugam和Dinstein在《IEEE系统、人与控制论学报》(IEEE Transactions on Systems, Man, and Cybernetics)上发表了一篇论文,提出使用灰度协同矩阵(GLCM)作为定义二维纹理的基础1。使用GLCM定义了超过14种不同的纹理测量。在灰度为n × n的图像中,GLCM的大小为n × n,对于较大的n(如n=256),会给该过程带来很大的计算负荷,并且也最适合于本质上相当随机的像素分布。利用GLCM提出了熵、方差、相关性等特征。在试图为人造目标提供特征度量时,目标中包含的大部分信息都包含在目标的边缘分布中。以前的方法先形成目标的边缘轮廓,然后使用傅立叶描述子等技术来表示目标。然而,在这种情况下,需要采取额外的步骤,以确保边缘轮廓是连续的,或者在为特征向量创建傅里叶系数时以某种方式处理轮廓中的间隙。本文提出了一种使用GLCM的方法,其中灰度图像使用几种边缘算子中的任何一种进行边缘增强。得到的图像是二值图像。对于边缘图像中的每个点,通过在该点周围放置一个n × n的窗口并使用n2个相邻点构建GLCM来创建2×2 GLCM。该窗口应该足够大,以包含感兴趣的目标,并且创建的GLCM提供了定义边缘增强目标的特征所需的元素。所有软件都是在Matlab2中使用Matlab函数创建的。
{"title":"Gray-level co-occurrence matrices as features in edge enhanced images","authors":"Peter J. Costianes, Joseph B. Plock","doi":"10.1109/AIPR.2010.5759705","DOIUrl":"https://doi.org/10.1109/AIPR.2010.5759705","url":null,"abstract":"In 1973, Haralick, Shanmugam, and Dinstein published a paper in the IEEE Transactions on Systems, Man, and Cybernetics which proposed using Gray-Level Cooccurrence Matrices (GLCM) as a basis to define 2-D texture1. Over 14 different texture measures were defined using GLCM. In images with n × n grey levels, the size of the GLCM would be n × n which, for large n such as n=256, put a large computational load on the process and was also best suited for pixel distributions that were rather stochastic in nature. Such features as entropy, variance, correlation, etc. were proposed using the GLCM. When attempting to provide feature measures for man-made targets, most of the information contained in the target is contained by its edge distribution. Previous approaches form an edge outline of the target and then use some techniques such as Fourier descriptors to represent the target. However, in this case, extra steps need to be taken in order to assure that the edge outline is continuous or gaps in the outline somehow are dealt with when creating the Fourier coefficients for the feature vector. This paper presents an approach using GLCM where the gray scale image is put through an edge enhancement using any one of several edge operators. The resultant image is a binary image. For each point in the edge image, a 2×2 GLCM is created by placing an n × n window centered around the point and using the n2 neighboring points to build the GLCM's. This window should be sufficiently large to enclose the target of interest and the GLCM created provides the elements needed to define the features for the edge enhanced target. All software was created in Matlab2 using Matlab functions.","PeriodicalId":128378,"journal":{"name":"2010 IEEE 39th Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"207 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115162296","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Head pose estimation from images using Canonical Correlation Analysis 基于典型相关分析的图像头部姿态估计
Pub Date : 2010-10-01 DOI: 10.1109/AIPR.2010.5759706
J. Foytik, V. Asari, M. Youssef, R. Tompkins
Head pose estimation, though a trivial task for the human visual system, remains a challenging problem for computer vision systems. The task requires identifying the modes of image variance that directly pertain to pose changes, while generalizing across face identity and mitigating other image variances. Conventional methods such as Principal Component Analysis (PCA) fail to identify the true relationship between the observed space and the pose variable, while supervised methods such as Linear Discriminant Analysis (LDA) neglect the continuous nature of pose variation and take a discrete multi-class approach. We present a method for estimating head pose using Canonical Correlation Analysis (CCA), where pose variation is regarded as a continuous variable and is represented by a manifold in feature space. The proposed technique directly identifies the underlying dimension that maximizes correlation between the observed image and pose variable. It is shown to increase estimation accuracy and provide a more compact image representation that better captures pose features. Additionally, an enhanced version of the system is proposed that utilizes Gabor filters for providing pose sensitive input to the correlation based system. The preprocessed input serves to increase the overall accuracy of the pose estimation system. The accuracy of the techniques is evaluated using the Pointing '04 and CUbiC FacePix(30) pose varying face databases and is shown to produce a lower estimation error when compared to both PCA and LDA based methods.
头部姿态估计对于人类视觉系统来说是一个微不足道的任务,但对于计算机视觉系统来说仍然是一个具有挑战性的问题。该任务需要识别与姿态变化直接相关的图像方差模式,同时推广面部身份并减轻其他图像方差。传统的主成分分析(PCA)等方法无法识别观测空间与位姿变量之间的真实关系,而线性判别分析(LDA)等监督方法忽略了位姿变化的连续性,采用离散的多类方法。我们提出了一种使用典型相关分析(CCA)估计头部姿态的方法,其中姿态变化被视为连续变量,并由特征空间中的流形表示。提出的技术直接识别潜在的维度,最大限度地提高了观察到的图像和姿态变量之间的相关性。它被证明可以提高估计精度,并提供更紧凑的图像表示,更好地捕捉姿态特征。此外,提出了系统的增强版本,利用Gabor滤波器为基于相关性的系统提供姿态敏感输入。预处理后的输入有助于提高姿态估计系统的整体精度。使用Pointing '04和CUbiC FacePix(30)对不同人脸数据库的技术精度进行了评估,与基于PCA和LDA的方法相比,这些技术产生的估计误差更低。
{"title":"Head pose estimation from images using Canonical Correlation Analysis","authors":"J. Foytik, V. Asari, M. Youssef, R. Tompkins","doi":"10.1109/AIPR.2010.5759706","DOIUrl":"https://doi.org/10.1109/AIPR.2010.5759706","url":null,"abstract":"Head pose estimation, though a trivial task for the human visual system, remains a challenging problem for computer vision systems. The task requires identifying the modes of image variance that directly pertain to pose changes, while generalizing across face identity and mitigating other image variances. Conventional methods such as Principal Component Analysis (PCA) fail to identify the true relationship between the observed space and the pose variable, while supervised methods such as Linear Discriminant Analysis (LDA) neglect the continuous nature of pose variation and take a discrete multi-class approach. We present a method for estimating head pose using Canonical Correlation Analysis (CCA), where pose variation is regarded as a continuous variable and is represented by a manifold in feature space. The proposed technique directly identifies the underlying dimension that maximizes correlation between the observed image and pose variable. It is shown to increase estimation accuracy and provide a more compact image representation that better captures pose features. Additionally, an enhanced version of the system is proposed that utilizes Gabor filters for providing pose sensitive input to the correlation based system. The preprocessed input serves to increase the overall accuracy of the pose estimation system. The accuracy of the techniques is evaluated using the Pointing '04 and CUbiC FacePix(30) pose varying face databases and is shown to produce a lower estimation error when compared to both PCA and LDA based methods.","PeriodicalId":128378,"journal":{"name":"2010 IEEE 39th Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"68 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123657198","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Intelligent management of multiple sensors for enhanced situational awareness 智能管理多个传感器,增强态势感知能力
Pub Date : 2010-10-01 DOI: 10.1109/AIPR.2010.5759715
Eric D. Nelson, J. Irvine
Wide area motion imagery (WAMI) offers the promise of persistent surveillance over large regions. However, the combination of lower frame rate and coarser spatial resolution found in most WAMI systems can limit the ability to track multiple targets. One way to address this limitation is to employ the wide-area sensor in concert with one or more high resolution sensors. We have developed a capability called Sensor Management for Adaptive Reconnaissance and Tracking (SMART), for tasking an arbitrary number of high-fidelity assets, working with the WAMI sensor to maximize situational awareness based on a prevailing set of conditions and target priorities. We present a simulation framework for exploring performance of various sensor management strategies and present the findings from an initial set of experiments.
广域运动图像(WAMI)提供了在大范围内持续监控的希望。然而,在大多数WAMI系统中,较低的帧率和较粗的空间分辨率的结合会限制跟踪多个目标的能力。解决这一限制的一种方法是将广域传感器与一个或多个高分辨率传感器结合使用。我们已经开发了一种称为自适应侦察和跟踪传感器管理(SMART)的能力,用于分配任意数量的高保真资产,与WAMI传感器一起工作,根据当前条件和目标优先级最大化态势感知。我们提出了一个模拟框架,用于探索各种传感器管理策略的性能,并提出了一组初始实验的结果。
{"title":"Intelligent management of multiple sensors for enhanced situational awareness","authors":"Eric D. Nelson, J. Irvine","doi":"10.1109/AIPR.2010.5759715","DOIUrl":"https://doi.org/10.1109/AIPR.2010.5759715","url":null,"abstract":"Wide area motion imagery (WAMI) offers the promise of persistent surveillance over large regions. However, the combination of lower frame rate and coarser spatial resolution found in most WAMI systems can limit the ability to track multiple targets. One way to address this limitation is to employ the wide-area sensor in concert with one or more high resolution sensors. We have developed a capability called Sensor Management for Adaptive Reconnaissance and Tracking (SMART), for tasking an arbitrary number of high-fidelity assets, working with the WAMI sensor to maximize situational awareness based on a prevailing set of conditions and target priorities. We present a simulation framework for exploring performance of various sensor management strategies and present the findings from an initial set of experiments.","PeriodicalId":128378,"journal":{"name":"2010 IEEE 39th Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"66 1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123524259","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2010 IEEE 39th Applied Imagery Pattern Recognition Workshop (AIPR)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1