首页 > 最新文献

Proceedings IEEE Workshop on Computer Vision Beyond the Visible Spectrum: Methods and Applications (CVBVS'99)最新文献

英文 中文
Quasi-invariants for recognition of articulated and non-standard objects in SAR images SAR图像中铰接和非标准目标识别的准不变量
G. Jones, B. Bhanu
Using SAR scattering center locations and magnitudes as features, invariances with articulation (i.e., turret rotation for the ZSU 23/4 gun and T72 tank), with configuration variants (e.g. fuel barrels, searchlights, etc.) and with a depression angle change are shown for real SAR images obtained from the MSTAR public data. This location and magnitude quasi-invariance forms a basis for an innovative SAR recognition engine that successfully identifies real articulated and non-standard configuration vehicles based on non-articulated, standard recognition models. Identification performance results are given as confusion matrices and ROC curves for articulated objects, for configuration variants, and for a small change in depression angle.
以SAR散射中心位置和震级为特征,对MSTAR公开数据获取的真实SAR图像显示了关节(即ZSU 23/4炮和T72坦克的炮塔旋转)、配置变量(如燃料桶、探照灯等)和俯角变化的不变性。这种位置和幅度的准不变性为创新的SAR识别引擎奠定了基础,该引擎可以成功识别基于非铰接式标准识别模型的真实铰接式和非标准配置车辆。识别性能结果以混淆矩阵和ROC曲线的形式给出,用于铰接对象、配置变量和俯角的小变化。
{"title":"Quasi-invariants for recognition of articulated and non-standard objects in SAR images","authors":"G. Jones, B. Bhanu","doi":"10.1109/CVBVS.1999.781098","DOIUrl":"https://doi.org/10.1109/CVBVS.1999.781098","url":null,"abstract":"Using SAR scattering center locations and magnitudes as features, invariances with articulation (i.e., turret rotation for the ZSU 23/4 gun and T72 tank), with configuration variants (e.g. fuel barrels, searchlights, etc.) and with a depression angle change are shown for real SAR images obtained from the MSTAR public data. This location and magnitude quasi-invariance forms a basis for an innovative SAR recognition engine that successfully identifies real articulated and non-standard configuration vehicles based on non-articulated, standard recognition models. Identification performance results are given as confusion matrices and ROC curves for articulated objects, for configuration variants, and for a small change in depression angle.","PeriodicalId":394469,"journal":{"name":"Proceedings IEEE Workshop on Computer Vision Beyond the Visible Spectrum: Methods and Applications (CVBVS'99)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124140336","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Moving target detection in infrared imagery using a regularized CDWT optical flow 基于正则化CDWT光流的红外图像运动目标检测
G. Castellano, J. Boyce, M. Sandler
A modified version of the CDWT optical flow algorithm developed by Magarey et al. is applied to the problem of moving target detection in noisy infrared image sequences. The optical flow algorithm is a hierarchical, phase-based approach. The modified version includes an explicit regularization of the motion field, which is of fundamental importance for the application in question. The data used consists of infrared imagery where pixel-size targets move in strongly cluttered backgrounds. To detect the targets different frames from the sequence are compared by subtraction of one from another. However, the motion of the sensor generates an apparent motion of the background across frames, and, as a consequence, the differences between background regions dominate the residue images. To avoid this effect, the estimated motion field between the frames is used to register the background spatially, so that only objects corresponding to potential targets appear in the residue images. Results of applying the method on 3 infrared image sequences are presented, which show that the target SNR is higher when the estimated motion field for the whole scene is explicitly regularized.
将Magarey等人开发的CDWT光流算法的改进版本应用于有噪声红外图像序列中的运动目标检测问题。光流算法是一种分层的、基于相位的方法。修改后的版本包括运动场的显式正则化,这对所讨论的应用至关重要。使用的数据包括红外图像,其中像素大小的目标在强烈杂乱的背景中移动。为了检测目标,将序列中的不同帧通过相减来进行比较。然而,传感器的运动产生了背景跨帧的明显运动,因此,背景区域之间的差异支配了残留图像。为了避免这种影响,利用估计的帧间运动场对背景进行空间配准,使残差图像中只出现潜在目标对应的物体。将该方法应用于3幅红外图像序列,结果表明,对整个场景的运动场估计进行显式正则化后,目标信噪比更高。
{"title":"Moving target detection in infrared imagery using a regularized CDWT optical flow","authors":"G. Castellano, J. Boyce, M. Sandler","doi":"10.1109/CVBVS.1999.781090","DOIUrl":"https://doi.org/10.1109/CVBVS.1999.781090","url":null,"abstract":"A modified version of the CDWT optical flow algorithm developed by Magarey et al. is applied to the problem of moving target detection in noisy infrared image sequences. The optical flow algorithm is a hierarchical, phase-based approach. The modified version includes an explicit regularization of the motion field, which is of fundamental importance for the application in question. The data used consists of infrared imagery where pixel-size targets move in strongly cluttered backgrounds. To detect the targets different frames from the sequence are compared by subtraction of one from another. However, the motion of the sensor generates an apparent motion of the background across frames, and, as a consequence, the differences between background regions dominate the residue images. To avoid this effect, the estimated motion field between the frames is used to register the background spatially, so that only objects corresponding to potential targets appear in the residue images. Results of applying the method on 3 infrared image sequences are presented, which show that the target SNR is higher when the estimated motion field for the whole scene is explicitly regularized.","PeriodicalId":394469,"journal":{"name":"Proceedings IEEE Workshop on Computer Vision Beyond the Visible Spectrum: Methods and Applications (CVBVS'99)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114582429","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
A ground truth tool for Synthetic Aperture Radar (SAR) imagery 合成孔径雷达(SAR)图像的地面真值工具
I. Pavlidis, Douglas P. Perrin, N. Papanikolopoulos, W. Au, S. Sawtelle
The performance of computer vision algorithms has made great strides and it is good enough to be useful in a number of civilian and military applications. Algorithm advancement in Automatic Target Recognition (ATR) in particular; has reached a critical point. State-of-the-art ATRs are capable of delivering robust performance for certain operational scenarios. As Computer Vision technology matures and algorithms enter the civilian and military marketplace as products, the lack of a formal testing theory and tools become obvious. In this paper we present the design and implementation of a Ground Truth Tool (GTT) for Synthetic Aperture Radar (SAR) imagery. The tool serves as part of an evaluation system for SAR ATRs. It features a semi-automatic method for delineating image objects that draws upon the theory of deformable models. In comparison with other deformable model implementations, our version is stable and is supported by an extensive Graphical User Interface (GUI). Preliminary experimental tests show that the system can substantially increase the productivity and accuracy of the image analyst (IA).
计算机视觉算法的性能已经取得了很大的进步,它足以在许多民用和军事应用中发挥作用。特别是自动目标识别(ATR)中的算法进展已经到了一个临界点。最先进的atr能够为某些操作场景提供强大的性能。随着计算机视觉技术的成熟和算法作为产品进入民用和军用市场,缺乏正式的测试理论和工具变得明显。本文介绍了合成孔径雷达(SAR)图像的地面真值工具(GTT)的设计和实现。该工具是SAR atr评估系统的一部分。它的特点是利用可变形模型的理论来描绘图像对象的半自动方法。与其他可变形模型实现相比,我们的版本是稳定的,并且得到了广泛的图形用户界面(GUI)的支持。初步的实验测试表明,该系统可以大大提高图像分析(IA)的生产率和准确性。
{"title":"A ground truth tool for Synthetic Aperture Radar (SAR) imagery","authors":"I. Pavlidis, Douglas P. Perrin, N. Papanikolopoulos, W. Au, S. Sawtelle","doi":"10.1109/CVBVS.1999.781097","DOIUrl":"https://doi.org/10.1109/CVBVS.1999.781097","url":null,"abstract":"The performance of computer vision algorithms has made great strides and it is good enough to be useful in a number of civilian and military applications. Algorithm advancement in Automatic Target Recognition (ATR) in particular; has reached a critical point. State-of-the-art ATRs are capable of delivering robust performance for certain operational scenarios. As Computer Vision technology matures and algorithms enter the civilian and military marketplace as products, the lack of a formal testing theory and tools become obvious. In this paper we present the design and implementation of a Ground Truth Tool (GTT) for Synthetic Aperture Radar (SAR) imagery. The tool serves as part of an evaluation system for SAR ATRs. It features a semi-automatic method for delineating image objects that draws upon the theory of deformable models. In comparison with other deformable model implementations, our version is stable and is supported by an extensive Graphical User Interface (GUI). Preliminary experimental tests show that the system can substantially increase the productivity and accuracy of the image analyst (IA).","PeriodicalId":394469,"journal":{"name":"Proceedings IEEE Workshop on Computer Vision Beyond the Visible Spectrum: Methods and Applications (CVBVS'99)","volume":"58 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125713812","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hyper-spectral image processing applications on the SIMD Pixel Processor for the digital battlefield 基于SIMD像素处理器的超光谱图像处理应用于数字战场
S. Chai, Antonio Gentile, W. Lugo-Beauchamp, J. Cruz-Rivera, D. S. Wills
Future military scenarios will rely on advanced imaging sensor technology beyond the visible spectrum to gain total battlefield awareness. Real-time processing of these data streams requires tremendous computational workloads and I/O throughputs. This paper presents three applications for hyper-spectral data streams, vector quantization, region autofocus, and K-means clustering, on the SIMD Pixel Processor (SIMPil). In SIMPil, an image sensor array (focal plane) is integrated on top of a SIMD computing layer to provide direct coupling between sensors and processors, alleviating I/O bandwidth bottlenecks while maintaining low power consumption and portability. Simulation results with sustained operation throughputs of 500-1500 Gops/sec support real-time performance and promote focal plane processing on SIMPil.
未来的军事场景将依赖于超越可见光谱的先进成像传感器技术来获得全面的战场感知。这些数据流的实时处理需要巨大的计算工作负载和I/O吞吐量。本文介绍了在SIMD像素处理器(SIMPil)上对高光谱数据流进行矢量量化、区域自动对焦和k均值聚类的三种应用。在SIMPil中,图像传感器阵列(焦平面)集成在SIMD计算层之上,提供传感器和处理器之间的直接耦合,缓解I/O带宽瓶颈,同时保持低功耗和可移植性。仿真结果表明,SIMPil的持续运行吞吐量为500-1500 Gops/sec,支持实时性能,并促进焦平面处理。
{"title":"Hyper-spectral image processing applications on the SIMD Pixel Processor for the digital battlefield","authors":"S. Chai, Antonio Gentile, W. Lugo-Beauchamp, J. Cruz-Rivera, D. S. Wills","doi":"10.1109/CVBVS.1999.781102","DOIUrl":"https://doi.org/10.1109/CVBVS.1999.781102","url":null,"abstract":"Future military scenarios will rely on advanced imaging sensor technology beyond the visible spectrum to gain total battlefield awareness. Real-time processing of these data streams requires tremendous computational workloads and I/O throughputs. This paper presents three applications for hyper-spectral data streams, vector quantization, region autofocus, and K-means clustering, on the SIMD Pixel Processor (SIMPil). In SIMPil, an image sensor array (focal plane) is integrated on top of a SIMD computing layer to provide direct coupling between sensors and processors, alleviating I/O bandwidth bottlenecks while maintaining low power consumption and portability. Simulation results with sustained operation throughputs of 500-1500 Gops/sec support real-time performance and promote focal plane processing on SIMPil.","PeriodicalId":394469,"journal":{"name":"Proceedings IEEE Workshop on Computer Vision Beyond the Visible Spectrum: Methods and Applications (CVBVS'99)","volume":"107 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128431424","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
A near-infrared fusion scheme for automatic detection of vehicle passengers 一种用于车辆乘客自动检测的近红外融合方案
I. Pavlidis, P. Symosek, B. Fritz, N. Papanikolopoulos
We undertook a study to determine if the automatic detection and counting of vehicle passengers is feasible. An automated passenger counting system would greatly facilitate the operation of freeway lanes reserved for car-pools (HOV lanes). In the present paper we report our findings regarding the appropriate sensor phenomenology and arrangement for the task. We propose a novel system based on fusion of near-infrared imaging signals and we demonstrate its adequacy with theoretical and experimental arguments.
我们进行了一项研究,以确定自动检测和计数车辆乘客是否可行。自动乘客计数系统将极大地促进为拼车保留的高速公路车道(HOV车道)的运营。在本文中,我们报告了我们关于适当的传感器现象学和任务安排的发现。我们提出了一种基于近红外成像信号融合的新系统,并通过理论和实验论证了它的充分性。
{"title":"A near-infrared fusion scheme for automatic detection of vehicle passengers","authors":"I. Pavlidis, P. Symosek, B. Fritz, N. Papanikolopoulos","doi":"10.1109/CVBVS.1999.781093","DOIUrl":"https://doi.org/10.1109/CVBVS.1999.781093","url":null,"abstract":"We undertook a study to determine if the automatic detection and counting of vehicle passengers is feasible. An automated passenger counting system would greatly facilitate the operation of freeway lanes reserved for car-pools (HOV lanes). In the present paper we report our findings regarding the appropriate sensor phenomenology and arrangement for the task. We propose a novel system based on fusion of near-infrared imaging signals and we demonstrate its adequacy with theoretical and experimental arguments.","PeriodicalId":394469,"journal":{"name":"Proceedings IEEE Workshop on Computer Vision Beyond the Visible Spectrum: Methods and Applications (CVBVS'99)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132852150","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
LADAR scene description using fuzzy morphology and rules 利用模糊形态学和规则对雷达场景进行描述
J. Keller, P. Gader, Xiaomei Wang
This paper presents a method for automatically generating descriptions of scenes represented by digital images acquired using laser radar (LADAR). A method for matching the scenes to linguistic descriptions is also presented. Both methods rely on fuzzy spatial relations. Primitive spatial relations between objects are computed using fuzzy mathematical morphology and compared to a previous method based on training a neural network to learn human preferences. For each pair of objects in a scene, the primitive spatial relations are combined into complex spatial relations using a fuzzy rule base. A scene description is generated using the highest confidence rule outputs. Scene matching is performed using the outputs of the rules that correspond to the linguistic description. The results show that seemingly significant differences in spatial relationship definitions have little impact on system performance and that reasonable match scores and descriptions can be generated from the fuzzy system.
提出了一种用激光雷达(LADAR)采集的数字图像自动生成场景描述的方法。提出了一种将场景与语言描述相匹配的方法。这两种方法都依赖于模糊空间关系。使用模糊数学形态学计算对象之间的原始空间关系,并与先前基于训练神经网络来学习人类偏好的方法进行比较。对于场景中的每对对象,使用模糊规则库将原始空间关系组合成复杂空间关系。使用最高置信度规则输出生成场景描述。使用与语言描述相对应的规则的输出执行场景匹配。结果表明,空间关系定义上看似显著的差异对系统性能影响不大,模糊系统可以生成合理的匹配分数和描述。
{"title":"LADAR scene description using fuzzy morphology and rules","authors":"J. Keller, P. Gader, Xiaomei Wang","doi":"10.1109/CVBVS.1999.781101","DOIUrl":"https://doi.org/10.1109/CVBVS.1999.781101","url":null,"abstract":"This paper presents a method for automatically generating descriptions of scenes represented by digital images acquired using laser radar (LADAR). A method for matching the scenes to linguistic descriptions is also presented. Both methods rely on fuzzy spatial relations. Primitive spatial relations between objects are computed using fuzzy mathematical morphology and compared to a previous method based on training a neural network to learn human preferences. For each pair of objects in a scene, the primitive spatial relations are combined into complex spatial relations using a fuzzy rule base. A scene description is generated using the highest confidence rule outputs. Scene matching is performed using the outputs of the rules that correspond to the linguistic description. The results show that seemingly significant differences in spatial relationship definitions have little impact on system performance and that reasonable match scores and descriptions can be generated from the fuzzy system.","PeriodicalId":394469,"journal":{"name":"Proceedings IEEE Workshop on Computer Vision Beyond the Visible Spectrum: Methods and Applications (CVBVS'99)","volume":"16 9","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120857628","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
A machine vision system using a laser radar applied to robotic fruit harvesting 一种使用激光雷达的机器视觉系统应用于机器人水果收获
A. Jiménez, R. Ceres, J. Pons
This paper describes a new laser-based computer vision system used for automatic fruit recognition. Most relevant vision studies for fruit harvesting are reviewed. Our system is based on an infrared laser-range finder sensor generating range and reflectance images and it is designed to detect spherical objects in non-structured environments. A special image restoration technique is defined and applied to improve image quality. Image analysis algorithms integrate both range and reflectance information to generate four characteristic primitives evidencing the presence of spherical objects. This machine vision system, which generates the 3-D location, radio and surface reflectivity of each spherical object has been applied to the AGRIBOT orange fruit harvester robot. Test results indicate good correct detection rates, unlikely false alarms and a robust behavior.
本文介绍了一种用于水果自动识别的激光计算机视觉系统。综述了水果收获的相关视觉研究进展。我们的系统基于红外激光测距传感器,产生距离和反射率图像,设计用于检测非结构化环境中的球形物体。定义了一种特殊的图像恢复技术,并应用于提高图像质量。图像分析算法结合距离和反射率信息来生成四个特征基元,以证明球形物体的存在。该机器视觉系统生成了每个球形物体的三维位置、无线电和表面反射率,并应用于AGRIBOT橙果收获机器人。测试结果表明,该系统具有良好的正确检测率、不太可能出现误报和鲁棒性。
{"title":"A machine vision system using a laser radar applied to robotic fruit harvesting","authors":"A. Jiménez, R. Ceres, J. Pons","doi":"10.1109/CVBVS.1999.781100","DOIUrl":"https://doi.org/10.1109/CVBVS.1999.781100","url":null,"abstract":"This paper describes a new laser-based computer vision system used for automatic fruit recognition. Most relevant vision studies for fruit harvesting are reviewed. Our system is based on an infrared laser-range finder sensor generating range and reflectance images and it is designed to detect spherical objects in non-structured environments. A special image restoration technique is defined and applied to improve image quality. Image analysis algorithms integrate both range and reflectance information to generate four characteristic primitives evidencing the presence of spherical objects. This machine vision system, which generates the 3-D location, radio and surface reflectivity of each spherical object has been applied to the AGRIBOT orange fruit harvester robot. Test results indicate good correct detection rates, unlikely false alarms and a robust behavior.","PeriodicalId":394469,"journal":{"name":"Proceedings IEEE Workshop on Computer Vision Beyond the Visible Spectrum: Methods and Applications (CVBVS'99)","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122180644","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Detecting moving objects in airborne forward looking infra-red sequences 在机载前视红外序列中检测运动物体
A. Strehl, J. Aggarwal
In this paper we propose a system that detects independently moving objects (IMOs) in forward looking infra-red (FLIR) image sequences taken from an airborne, moving platform. Ego-motion effects are removed through a robust multi-scale affine image registration process. Consequently, areas with residual motion indicate object activity. These areas are detected, refined and selected using a Bayes' classifier. The remaining regions are clustered into pairs. Each pair represents an object's front and rear end. Using motion and scene knowledge we estimate object pose and establish a region-of-interest (ROI) for each pair. Edge elements within each ROI are used to segment the convex cover containing the IMO. We show detailed results on real, complex, cluttered and noisy sequences. Moreover, we outline the integration of our robust system into a comprehensive automatic target recognition (ATR) and action classification system.
在本文中,我们提出了一个系统来检测独立运动目标(imo)的前视红外(FLIR)图像序列从机载,移动平台。通过鲁棒的多尺度仿射图像配准过程去除自我运动效应。因此,有残余运动的区域表示物体的活动。使用贝叶斯分类器检测、精炼和选择这些区域。剩下的区域成对聚集。每一对代表一个物体的前端和后端。利用运动和场景知识估计物体姿态,并为每对物体建立感兴趣区域(ROI)。每个ROI内的边缘元素用于分割包含IMO的凸盖。我们展示了真实的、复杂的、杂乱的和有噪声的序列的详细结果。此外,我们概述了我们的鲁棒系统集成到一个全面的自动目标识别(ATR)和动作分类系统。
{"title":"Detecting moving objects in airborne forward looking infra-red sequences","authors":"A. Strehl, J. Aggarwal","doi":"10.1109/CVBVS.1999.781089","DOIUrl":"https://doi.org/10.1109/CVBVS.1999.781089","url":null,"abstract":"In this paper we propose a system that detects independently moving objects (IMOs) in forward looking infra-red (FLIR) image sequences taken from an airborne, moving platform. Ego-motion effects are removed through a robust multi-scale affine image registration process. Consequently, areas with residual motion indicate object activity. These areas are detected, refined and selected using a Bayes' classifier. The remaining regions are clustered into pairs. Each pair represents an object's front and rear end. Using motion and scene knowledge we estimate object pose and establish a region-of-interest (ROI) for each pair. Edge elements within each ROI are used to segment the convex cover containing the IMO. We show detailed results on real, complex, cluttered and noisy sequences. Moreover, we outline the integration of our robust system into a comprehensive automatic target recognition (ATR) and action classification system.","PeriodicalId":394469,"journal":{"name":"Proceedings IEEE Workshop on Computer Vision Beyond the Visible Spectrum: Methods and Applications (CVBVS'99)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123769261","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 59
Morphological shared-weight neural networks: a tool for automatic target recognition beyond the visible spectrum 形态共享权神经网络:一种超越可见光谱的自动目标识别工具
M. A. Khabou, P. Gader, J. Keller
Morphological shared-weight neural networks (MSNN) combine the feature extraction capability of mathematical morphology with the function mapping capability of neural networks. This provides a trainable mechanism for translation invariant object detection using a variety of imaging sensors, including TV, forward-looking infrared (FLIR) and synthetic aperture radar (SAR). We provide an overview of previous results and new results with laser radar (LADAR). We present three sets of experiments. In the first set of experiments we use the MSNN to detect different types of targets simultaneously. In the second set we use the MSNN to detect only a particular type of target. In the third set we test a novel scenario: we train the MSNN to recognize a particular type of target using very few examples. A detection rate of 86% with a reasonable number of false alarms was achieved in the first set of experiments and a detection rate of close to 100% with very few false alarms was achieved in the second and third sets of experiments.
形态学共享权神经网络(MSNN)将数学形态学的特征提取能力与神经网络的函数映射能力相结合。这为使用各种成像传感器(包括电视、前视红外(FLIR)和合成孔径雷达(SAR))进行平移不变目标检测提供了一种可训练的机制。本文综述了激光雷达(LADAR)的研究成果。我们提出了三组实验。在第一组实验中,我们使用MSNN同时检测不同类型的目标。在第二组中,我们使用MSNN只检测特定类型的目标。在第三组中,我们测试了一个新的场景:我们训练MSNN使用很少的示例来识别特定类型的目标。第一组实验的检测率为86%,虚警数量合理;第二组和第三组实验的检测率接近100%,虚警数量很少。
{"title":"Morphological shared-weight neural networks: a tool for automatic target recognition beyond the visible spectrum","authors":"M. A. Khabou, P. Gader, J. Keller","doi":"10.1109/CVBVS.1999.781099","DOIUrl":"https://doi.org/10.1109/CVBVS.1999.781099","url":null,"abstract":"Morphological shared-weight neural networks (MSNN) combine the feature extraction capability of mathematical morphology with the function mapping capability of neural networks. This provides a trainable mechanism for translation invariant object detection using a variety of imaging sensors, including TV, forward-looking infrared (FLIR) and synthetic aperture radar (SAR). We provide an overview of previous results and new results with laser radar (LADAR). We present three sets of experiments. In the first set of experiments we use the MSNN to detect different types of targets simultaneously. In the second set we use the MSNN to detect only a particular type of target. In the third set we test a novel scenario: we train the MSNN to recognize a particular type of target using very few examples. A detection rate of 86% with a reasonable number of false alarms was achieved in the first set of experiments and a detection rate of close to 100% with very few false alarms was achieved in the second and third sets of experiments.","PeriodicalId":394469,"journal":{"name":"Proceedings IEEE Workshop on Computer Vision Beyond the Visible Spectrum: Methods and Applications (CVBVS'99)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128504251","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Passive night vision sensor comparison for unmanned ground vehicle stereo vision navigation 被动夜视传感器在无人地面车辆立体视觉导航中的比较
K. Owens, L. Matthies
One goal of the "Demo III" unmanned ground vehicle program is to enable autonomous nighttime navigation at speeds of up to 10 m.p.h. To perform obstacle detection at night with stereo vision will require night vision cameras that produce adequate image quality for the driving speeds, vehicle dynamics, obstacle sizes, and scene conditions that will be encountered. This paper analyzes the suitability of four classes of night vision cameras (3-5 /spl mu/m cooled FLIR, 8-12 /spl mu/m cooled FLIR, 8-12 /spl mu/m uncooled FLIR, and image intensifiers) for night stereo vision, using criteria based on stereo matching quality, image signal to noise ratio, motion blur and synchronization capability. We find that only cooled FLIRs will enable stereo vision performance that meets the goals of the Demo III program for nighttime autonomous mobility.
“Demo III”无人驾驶地面车辆项目的目标之一是实现夜间自主导航,速度可达每小时10英里。要在夜间使用立体视觉进行障碍物检测,需要夜视摄像头能够根据驾驶速度、车辆动态、障碍物大小和将遇到的场景条件产生足够的图像质量。以立体匹配质量、图像信噪比、运动模糊和同步能力为标准,分析了4类夜视摄像机(3-5 /spl μ m冷却型前视红外、8-12 /spl μ m冷却型前视红外、8-12 /spl非冷却型前视红外和图像增强器)对夜间立体视觉的适应性。我们发现,只有冷却后的前视红外才能实现立体视觉性能,从而满足演示III项目的夜间自主移动目标。
{"title":"Passive night vision sensor comparison for unmanned ground vehicle stereo vision navigation","authors":"K. Owens, L. Matthies","doi":"10.1109/CVBVS.1999.781095","DOIUrl":"https://doi.org/10.1109/CVBVS.1999.781095","url":null,"abstract":"One goal of the \"Demo III\" unmanned ground vehicle program is to enable autonomous nighttime navigation at speeds of up to 10 m.p.h. To perform obstacle detection at night with stereo vision will require night vision cameras that produce adequate image quality for the driving speeds, vehicle dynamics, obstacle sizes, and scene conditions that will be encountered. This paper analyzes the suitability of four classes of night vision cameras (3-5 /spl mu/m cooled FLIR, 8-12 /spl mu/m cooled FLIR, 8-12 /spl mu/m uncooled FLIR, and image intensifiers) for night stereo vision, using criteria based on stereo matching quality, image signal to noise ratio, motion blur and synchronization capability. We find that only cooled FLIRs will enable stereo vision performance that meets the goals of the Demo III program for nighttime autonomous mobility.","PeriodicalId":394469,"journal":{"name":"Proceedings IEEE Workshop on Computer Vision Beyond the Visible Spectrum: Methods and Applications (CVBVS'99)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-06-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117182197","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
Proceedings IEEE Workshop on Computer Vision Beyond the Visible Spectrum: Methods and Applications (CVBVS'99)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1