首页 > 最新文献

2014 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)最新文献

英文 中文
Accurate coverage summarization of UAV videos 无人机视频准确覆盖汇总
Pub Date : 2014-10-01 DOI: 10.1109/AIPR.2014.7041923
Chung-Ching Lin, Sharath Pankanti, John R. Smith
A predominant fraction of UAV videos are never watched or analyzed and there is growing interest in having a summary view of the UAV videos for obtaining a better overall perspective of the visual content. Real time summarization of the UAV video events is also important from tactical perspective. Our research focuses on developing resilient algorithms for summarizing videos that can be efficiently processed either onboard or offline. Our previous work [2] on the video summarization has focused on the event summarization. More recently, we have investigated the challenges in providing the coverage summarization of the video content from UAV videos. Different from the traditional coverage summarization taking SfM approach (e.g., [7]) on SIFT-based [14] feature points, there are several additional challenges including jitter, low resolution, contrast, lack of salient features in UAV videos. We propose a novel correspondence algorithm that exploits the 3D context that can potentially alleviate the correspondence ambiguity. Our results on VIRAT dataset shows that our algorithm can find many correct correspondences in low resolution imagery while avoiding many false positives from the traditional algorithms.
UAV视频的主要部分从未被观看或分析,并且对UAV视频的摘要视图越来越感兴趣,以便获得更好的视觉内容的整体视角。从战术角度看,无人机视频事件的实时总结也很重要。我们的研究重点是开发弹性算法,用于总结可以在船上或离线有效处理的视频。我们之前关于视频摘要的工作[2]主要集中在事件摘要上。最近,我们研究了在提供无人机视频内容的覆盖摘要方面的挑战。与基于sift[14]特征点的采用SfM方法的传统覆盖摘要(例如[7])不同,无人机视频还存在抖动、低分辨率、对比度、缺乏显著特征等挑战。我们提出了一种新的对应算法,利用三维上下文可以潜在地减轻对应模糊。在VIRAT数据集上的结果表明,该算法可以在低分辨率图像中找到许多正确的对应,同时避免了传统算法的许多误报。
{"title":"Accurate coverage summarization of UAV videos","authors":"Chung-Ching Lin, Sharath Pankanti, John R. Smith","doi":"10.1109/AIPR.2014.7041923","DOIUrl":"https://doi.org/10.1109/AIPR.2014.7041923","url":null,"abstract":"A predominant fraction of UAV videos are never watched or analyzed and there is growing interest in having a summary view of the UAV videos for obtaining a better overall perspective of the visual content. Real time summarization of the UAV video events is also important from tactical perspective. Our research focuses on developing resilient algorithms for summarizing videos that can be efficiently processed either onboard or offline. Our previous work [2] on the video summarization has focused on the event summarization. More recently, we have investigated the challenges in providing the coverage summarization of the video content from UAV videos. Different from the traditional coverage summarization taking SfM approach (e.g., [7]) on SIFT-based [14] feature points, there are several additional challenges including jitter, low resolution, contrast, lack of salient features in UAV videos. We propose a novel correspondence algorithm that exploits the 3D context that can potentially alleviate the correspondence ambiguity. Our results on VIRAT dataset shows that our algorithm can find many correct correspondences in low resolution imagery while avoiding many false positives from the traditional algorithms.","PeriodicalId":210982,"journal":{"name":"2014 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127526508","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
High dynamic range (HDR) video processing for the exploitation of high bit-depth sensors in human-monitored surveillance 高动态范围(HDR)视频处理技术是高位深传感器在人类监控中的应用
Pub Date : 2014-10-01 DOI: 10.1109/AIPR.2014.7041912
D. Natale, Matthew S. Baran, R. Tutwiler
High bit-depth video data is becoming more common in imaging and remote sensing because higher bit-depth cameras are becoming more affordable. Displays often represent images in lower bit-depths, and human vision is not able to completely exploit this additional information in its native form. These problems are addressed with High Dynamic Range (HDR) tone mapping, which nonlinearly maps lightness levels from a high bit-depth image into a lower bit-depth representation in a way that attempts to retain and accentuate the maximum amount of useful information therein. We have adapted the well-known Contrast Limited Adaptive Histogram Equalization (CLAHE) algorithm into the application of HDR video tone mapping by using time-adaptive local histogram transformations. In addition to lightness contrast, we use the transformations in the L*a*b* color space to amplify color contrast in the video stream. The transformed HDR video data maintains important details in local contrast while maintaining relative lightness levels locally through time. Our results show that time-adapted HDR tone mapping methods can be used in real-time video processing to store and display HDR data in low bit-depth formats with less loss of useful information compared to simple truncation.
高位深视频数据在成像和遥感中变得越来越普遍,因为更高位深的相机变得越来越便宜。显示器通常以较低的位深度表示图像,人类视觉无法完全利用其原始形式的附加信息。这些问题可以通过高动态范围(HDR)色调映射来解决,该映射将亮度级别从高位深度图像非线性地映射到较低位深度表示,以一种试图保留和强调其中最大量有用信息的方式。我们通过使用时间自适应局部直方图变换,将著名的对比度有限自适应直方图均衡化(CLAHE)算法应用于HDR视频色调映射。除了亮度对比外,我们还使用L*a*b*色彩空间中的变换来放大视频流中的色彩对比。转换后的HDR视频数据在保持局部对比度的同时保持局部相对亮度水平。我们的研究结果表明,与简单的截断相比,时间适应HDR色调映射方法可以用于实时视频处理,以低位深格式存储和显示HDR数据,减少有用信息的损失。
{"title":"High dynamic range (HDR) video processing for the exploitation of high bit-depth sensors in human-monitored surveillance","authors":"D. Natale, Matthew S. Baran, R. Tutwiler","doi":"10.1109/AIPR.2014.7041912","DOIUrl":"https://doi.org/10.1109/AIPR.2014.7041912","url":null,"abstract":"High bit-depth video data is becoming more common in imaging and remote sensing because higher bit-depth cameras are becoming more affordable. Displays often represent images in lower bit-depths, and human vision is not able to completely exploit this additional information in its native form. These problems are addressed with High Dynamic Range (HDR) tone mapping, which nonlinearly maps lightness levels from a high bit-depth image into a lower bit-depth representation in a way that attempts to retain and accentuate the maximum amount of useful information therein. We have adapted the well-known Contrast Limited Adaptive Histogram Equalization (CLAHE) algorithm into the application of HDR video tone mapping by using time-adaptive local histogram transformations. In addition to lightness contrast, we use the transformations in the L*a*b* color space to amplify color contrast in the video stream. The transformed HDR video data maintains important details in local contrast while maintaining relative lightness levels locally through time. Our results show that time-adapted HDR tone mapping methods can be used in real-time video processing to store and display HDR data in low bit-depth formats with less loss of useful information compared to simple truncation.","PeriodicalId":210982,"journal":{"name":"2014 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114271401","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Evaluating the Lidar/HSI direct method for physics-based scene modeling 评估基于物理场景建模的Lidar/HSI直接方法
Pub Date : 2014-10-01 DOI: 10.1109/AIPR.2014.7041906
Ryan N. Givens, K. Walli, M. Eismann
Recent work has been able to automate the process of generating three-dimensional, spectrally attributed scenes for use in physics-based modeling software using the Lidar/Hyperspectral Direct (LHD) method. The LHD method autonomously generates three-dimensional Digital Imaging and Remote Sensing Image Generation (DIRSIG) scenes from input high-resolution imagery, lidar data, and hyperspectral imagery and has been shown to do this successfully using both modeled and real datasets. While the output scenes look realistic and appear to match the input scenes under qualitative comparisons, a more quantitative approach is needed to evaluate the full utility of these autonomously generated scenes. This paper seeks to improve the evaluation of the spatial and spectral accuracy of autonomously generated three-dimensional scenes using the DIRSIG model. Two scenes are presented for this evaluation. The first is generated from a modeled dataset and the second is generated using data collected over a real-world site. DIRSIG-generated synthetic imagery over the recreated scenes are then compared to the original input imagery to evaluate how well the recreated scenes match the original scenes in spatial and spectral accuracy and to determine the ability of the recreated scenes to produce useful outputs for algorithm development.
最近的工作已经能够使用激光雷达/高光谱直接(LHD)方法自动生成三维光谱属性场景的过程,用于基于物理的建模软件。LHD方法从输入的高分辨率图像、激光雷达数据和高光谱图像中自动生成三维数字成像和遥感图像生成(DIRSIG)场景,并已被证明可以成功地使用建模和实际数据集。虽然输出场景看起来很逼真,并且在定性比较下似乎与输入场景相匹配,但需要更定量的方法来评估这些自主生成场景的全部效用。本文旨在利用DIRSIG模型改进对自主生成三维场景的空间和光谱精度的评估。本文给出了两个场景来进行评估。第一个是从建模的数据集生成的,第二个是使用从真实站点收集的数据生成的。然后将dirsig生成的重建场景合成图像与原始输入图像进行比较,以评估重建场景在空间和光谱精度方面与原始场景的匹配程度,并确定重建场景为算法开发提供有用输出的能力。
{"title":"Evaluating the Lidar/HSI direct method for physics-based scene modeling","authors":"Ryan N. Givens, K. Walli, M. Eismann","doi":"10.1109/AIPR.2014.7041906","DOIUrl":"https://doi.org/10.1109/AIPR.2014.7041906","url":null,"abstract":"Recent work has been able to automate the process of generating three-dimensional, spectrally attributed scenes for use in physics-based modeling software using the Lidar/Hyperspectral Direct (LHD) method. The LHD method autonomously generates three-dimensional Digital Imaging and Remote Sensing Image Generation (DIRSIG) scenes from input high-resolution imagery, lidar data, and hyperspectral imagery and has been shown to do this successfully using both modeled and real datasets. While the output scenes look realistic and appear to match the input scenes under qualitative comparisons, a more quantitative approach is needed to evaluate the full utility of these autonomously generated scenes. This paper seeks to improve the evaluation of the spatial and spectral accuracy of autonomously generated three-dimensional scenes using the DIRSIG model. Two scenes are presented for this evaluation. The first is generated from a modeled dataset and the second is generated using data collected over a real-world site. DIRSIG-generated synthetic imagery over the recreated scenes are then compared to the original input imagery to evaluate how well the recreated scenes match the original scenes in spatial and spectral accuracy and to determine the ability of the recreated scenes to produce useful outputs for algorithm development.","PeriodicalId":210982,"journal":{"name":"2014 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116467454","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adaptive automatic object recognition in single and multi-modal sensor data 单模态和多模态传感器数据的自适应自动目标识别
Pub Date : 2014-10-01 DOI: 10.1109/AIPR.2014.7041915
T. Khuon, R. Rand
For single-modal data, object recognition and classification in a 3D point cloud is a non-trivial process due to the nature of the data collected from a sensor system where the signal can be corrupted by noise from the environment, electronic system, A/D converter, etc. Therefore, an adaptive system with a specific desired tolerance is required to perform classification and recognition optimally. The feature-based pattern recognition algorithm described below, is generalized for solving a particular global problem with minimal change. Since for the given class set, a feature set must be extracted accordingly. For instance, man-made urban object classification, rural and natural objects, and human organ classification would require different and distinct feature sets. This study is to compare the adaptive automatic object recognition in single sensor and the distributed adaptive pattern recognition in multi-sensor fusion. The similarity in automatic object recognition between single-sensor and multi-sensor fusion is the ability to learn from experiences and decide on a given pattern. Their main difference is that the sensor fusion makes a decision from the decisions of all sensors whereas the single sensor requires a feature extraction for a decision.
对于单模态数据,三维点云中的物体识别和分类是一个非常重要的过程,因为从传感器系统收集的数据的性质,其中信号可能被来自环境、电子系统、a /D转换器等的噪声所破坏。因此,需要一个具有特定期望容差的自适应系统来最佳地执行分类和识别。下面描述的基于特征的模式识别算法,用于用最小的变化来解决特定的全局问题。因为对于给定的类集,必须相应地提取特征集。例如,人造的城市对象分类、农村和自然对象分类以及人体器官分类将需要不同的、不同的特征集。本研究比较了单传感器下的自适应自动目标识别与多传感器融合下的分布式自适应模式识别。单传感器自动目标识别与多传感器自动目标识别的相似之处在于从经验中学习并确定给定模式的能力。它们的主要区别在于,传感器融合从所有传感器的决策中做出决策,而单个传感器需要对决策进行特征提取。
{"title":"Adaptive automatic object recognition in single and multi-modal sensor data","authors":"T. Khuon, R. Rand","doi":"10.1109/AIPR.2014.7041915","DOIUrl":"https://doi.org/10.1109/AIPR.2014.7041915","url":null,"abstract":"For single-modal data, object recognition and classification in a 3D point cloud is a non-trivial process due to the nature of the data collected from a sensor system where the signal can be corrupted by noise from the environment, electronic system, A/D converter, etc. Therefore, an adaptive system with a specific desired tolerance is required to perform classification and recognition optimally. The feature-based pattern recognition algorithm described below, is generalized for solving a particular global problem with minimal change. Since for the given class set, a feature set must be extracted accordingly. For instance, man-made urban object classification, rural and natural objects, and human organ classification would require different and distinct feature sets. This study is to compare the adaptive automatic object recognition in single sensor and the distributed adaptive pattern recognition in multi-sensor fusion. The similarity in automatic object recognition between single-sensor and multi-sensor fusion is the ability to learn from experiences and decide on a given pattern. Their main difference is that the sensor fusion makes a decision from the decisions of all sensors whereas the single sensor requires a feature extraction for a decision.","PeriodicalId":210982,"journal":{"name":"2014 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122238144","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Change detection and classification of land cover in multispectral satellite imagery using clustering of sparse approximations (CoSA) over learned feature dictionaries 基于稀疏逼近聚类的多光谱卫星影像土地覆盖变化检测与分类
Pub Date : 2014-10-01 DOI: 10.1109/AIPR.2014.7041921
D. Moody, S. Brumby, J. Rowland, G. Altmann, Amy E. Larson
Neuromimetic machine vision and pattern recognition algorithms are of great interest for landscape characterization and change detection in satellite imagery in support of global climate change science and modeling. We present results from an ongoing effort to extend machine vision methods to the environmental sciences, using adaptive sparse signal processing combined with machine learning. A Hebbian learning rule is used to build multispectral, multiresolution dictionaries from regional satellite normalized band difference index data. Land cover labels are automatically generated via our CoSA algorithm: Clustering of Sparse Approximations, using a clustering distance metric that combines spectral and spatial textural characteristics to help separate geologic, vegetative, and hydrologie features. We demonstrate our method on example Worldview-2 satellite images of an Arctic region, and use CoSA labels to detect seasonal surface changes. Our results suggest that neuroscience-based models are a promising approach to practical pattern recognition and change detection problems in remote sensing.
模拟神经机器视觉和模式识别算法在卫星图像的景观表征和变化检测中具有重要意义,支持全球气候变化科学和建模。我们目前正在努力将机器视觉方法扩展到环境科学,使用自适应稀疏信号处理与机器学习相结合。利用Hebbian学习规则从区域卫星归一化带差指数数据中构建多光谱、多分辨率字典。土地覆盖标签通过我们的CoSA算法自动生成:稀疏逼近聚类,使用结合光谱和空间纹理特征的聚类距离度量来帮助分离地质、植被和水文特征。我们在北极地区的Worldview-2卫星图像示例上演示了我们的方法,并使用CoSA标签来检测季节性地表变化。我们的研究结果表明,基于神经科学的模型是一种很有前途的方法来解决遥感中的实际模式识别和变化检测问题。
{"title":"Change detection and classification of land cover in multispectral satellite imagery using clustering of sparse approximations (CoSA) over learned feature dictionaries","authors":"D. Moody, S. Brumby, J. Rowland, G. Altmann, Amy E. Larson","doi":"10.1109/AIPR.2014.7041921","DOIUrl":"https://doi.org/10.1109/AIPR.2014.7041921","url":null,"abstract":"Neuromimetic machine vision and pattern recognition algorithms are of great interest for landscape characterization and change detection in satellite imagery in support of global climate change science and modeling. We present results from an ongoing effort to extend machine vision methods to the environmental sciences, using adaptive sparse signal processing combined with machine learning. A Hebbian learning rule is used to build multispectral, multiresolution dictionaries from regional satellite normalized band difference index data. Land cover labels are automatically generated via our CoSA algorithm: Clustering of Sparse Approximations, using a clustering distance metric that combines spectral and spatial textural characteristics to help separate geologic, vegetative, and hydrologie features. We demonstrate our method on example Worldview-2 satellite images of an Arctic region, and use CoSA labels to detect seasonal surface changes. Our results suggest that neuroscience-based models are a promising approach to practical pattern recognition and change detection problems in remote sensing.","PeriodicalId":210982,"journal":{"name":"2014 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128407586","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Against conventional wisdom: Longitudinal inference for pattern recognition in remote sensing 反对传统智慧:遥感模式识别的纵向推理
Pub Date : 2014-10-01 DOI: 10.1109/AIPR.2014.7041932
D. Rosario, Christoph Borel-Donohue, J. Romano
In response to Democratization of Imagery, a recent leading theme in the scientific community, we discuss a persistent imaging experiment dataset, which is being considered for public release in a foreseeable future, and present our observations analyzing a subset of the dataset. The experiment is a long-term collaborative effort among the Army Research Laboratory, Army Armament RDEC, and Air Force Institute of Technology that focuses on the collection and exploitation of longwave infrared (LWIR) hyperspectral and polarimetric imagery. In this paper, we emphasize the inherent challenges associated with using remotely sensed LWIR hyperspectral imagery for material recognition, and argue that the idealized data assumptions often made by the state of the art methods are too restrictive for real operational scenarios. We treat LWIR hyperspectral imagery for the first time as Longitudinal Data and aim at proposing a more realistic framework for material recognition as a function of spectral evolution over time. The defining characteristic of a longitudinal study is that objects are measured repeatedly through time and, as a result, data are dependent. This is in contrast to cross-sectional studies in which the outcomes of a specific event are observed by randomly sampling from a large population of relevant objects, where data are assumed independent. The scientific community generally assumes the problem of object recognition to be cross-sectional. We argue that, as data evolve over a full diurnal cycle, pattern recognition problems are longitudinal in nature and that by applying this knowledge it may lead to better algorithms.
为了回应最近科学界的一个主要主题——图像民主化,我们讨论了一个持久的成像实验数据集,该数据集正在考虑在可预见的未来公开发布,并展示了我们对数据集子集的观察分析。该实验是陆军研究实验室、陆军装备RDEC和空军技术学院之间的一项长期合作,重点是长波红外(LWIR)高光谱和偏振图像的收集和开发。在本文中,我们强调了与使用遥感LWIR高光谱图像进行材料识别相关的固有挑战,并认为由最先进的方法通常做出的理想化数据假设对于实际操作场景来说过于严格。我们首次将LWIR高光谱图像视为纵向数据,旨在提出一个更现实的框架,将材料识别作为光谱随时间演变的函数。纵向研究的定义特征是,随着时间的推移,对象被反复测量,因此,数据是依赖的。这与横断面研究相反,在横断面研究中,通过从大量相关对象中随机抽样观察特定事件的结果,在横断面研究中,假设数据是独立的。科学界普遍认为物体识别问题是一个横向问题。我们认为,随着数据在一个完整的昼夜周期中演变,模式识别问题本质上是纵向的,通过应用这些知识,它可能会导致更好的算法。
{"title":"Against conventional wisdom: Longitudinal inference for pattern recognition in remote sensing","authors":"D. Rosario, Christoph Borel-Donohue, J. Romano","doi":"10.1109/AIPR.2014.7041932","DOIUrl":"https://doi.org/10.1109/AIPR.2014.7041932","url":null,"abstract":"In response to Democratization of Imagery, a recent leading theme in the scientific community, we discuss a persistent imaging experiment dataset, which is being considered for public release in a foreseeable future, and present our observations analyzing a subset of the dataset. The experiment is a long-term collaborative effort among the Army Research Laboratory, Army Armament RDEC, and Air Force Institute of Technology that focuses on the collection and exploitation of longwave infrared (LWIR) hyperspectral and polarimetric imagery. In this paper, we emphasize the inherent challenges associated with using remotely sensed LWIR hyperspectral imagery for material recognition, and argue that the idealized data assumptions often made by the state of the art methods are too restrictive for real operational scenarios. We treat LWIR hyperspectral imagery for the first time as Longitudinal Data and aim at proposing a more realistic framework for material recognition as a function of spectral evolution over time. The defining characteristic of a longitudinal study is that objects are measured repeatedly through time and, as a result, data are dependent. This is in contrast to cross-sectional studies in which the outcomes of a specific event are observed by randomly sampling from a large population of relevant objects, where data are assumed independent. The scientific community generally assumes the problem of object recognition to be cross-sectional. We argue that, as data evolve over a full diurnal cycle, pattern recognition problems are longitudinal in nature and that by applying this knowledge it may lead to better algorithms.","PeriodicalId":210982,"journal":{"name":"2014 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127878526","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Human activity detection using sparse representation 基于稀疏表示的人类活动检测
Pub Date : 2014-10-01 DOI: 10.1109/AIPR.2014.7041933
D. Killedar, S. Sasi
Human activity detection from videos is very challenging, and has got numerous applications in sports evalution, video surveillance, elder/child care, etc. In this research, a model using sparse representation is presented for the human activity detection from the video data. This is done using a linear combination of atoms from a dictionary and a sparse coefficient matrix. The dictionary is created using a Spatio Temporal Interest Points (STIP) algorithm. The Spatio temporal features are extracted for the training video data as well as the testing video data. The K-Singular Value Decomposition (KSVD) algorithm is used for learning dictionaries for the training video dataset. Finally, human action is classified using a minimum threshold residual value of the corresponding action class in the testing video dataset. Experiments are conducted on the KTH dataset which contains a number of actions. The current approach performed well in classifying activities with a success rate of 90%.
从视频中进行人体活动检测是一项非常具有挑战性的工作,在体育评估、视频监控、老人/儿童护理等领域得到了广泛的应用。本文提出了一种基于稀疏表示的视频人体活动检测模型。这是使用字典中的原子和稀疏系数矩阵的线性组合来完成的。字典是使用时空兴趣点(STIP)算法创建的。提取训练视频数据和测试视频数据的时空特征。使用k -奇异值分解(KSVD)算法学习训练视频数据集的字典。最后,使用测试视频数据集中相应动作类的最小阈值残差对人类动作进行分类。在包含多个动作的KTH数据集上进行了实验。目前的方法在分类活动方面表现良好,成功率为90%。
{"title":"Human activity detection using sparse representation","authors":"D. Killedar, S. Sasi","doi":"10.1109/AIPR.2014.7041933","DOIUrl":"https://doi.org/10.1109/AIPR.2014.7041933","url":null,"abstract":"Human activity detection from videos is very challenging, and has got numerous applications in sports evalution, video surveillance, elder/child care, etc. In this research, a model using sparse representation is presented for the human activity detection from the video data. This is done using a linear combination of atoms from a dictionary and a sparse coefficient matrix. The dictionary is created using a Spatio Temporal Interest Points (STIP) algorithm. The Spatio temporal features are extracted for the training video data as well as the testing video data. The K-Singular Value Decomposition (KSVD) algorithm is used for learning dictionaries for the training video dataset. Finally, human action is classified using a minimum threshold residual value of the corresponding action class in the testing video dataset. Experiments are conducted on the KTH dataset which contains a number of actions. The current approach performed well in classifying activities with a success rate of 90%.","PeriodicalId":210982,"journal":{"name":"2014 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115232264","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Performance benefits of sub-diffraction sized pixels in imaging sensors 成像传感器中亚衍射尺寸像素的性能优势
Pub Date : 2014-05-28 DOI: 10.1117/12.2053443
J. Caulfield, J. Curzan, N. Dhar
Infrared Focal Plane Arrays have been developed with reductions in pixel size below the Nyquist limit imposed by the optical systems Point Spread Function (PSF). These smaller sub diffraction limited pixels allows spatial oversampling of the image. We show that oversampling the PSF allows improved fidelity in imaging, resulting in sensitivity improvements due to pixel correlation, reduced false alarm rates, improved detection ranges, and an improved ability to track closely spaced objects.
红外焦平面阵列已经发展到像素尺寸低于奈奎斯特限制的光学系统点扩展函数(PSF)。这些较小的亚衍射限制像素允许图像的空间过采样。我们表明,对PSF进行过采样可以提高成像的保真度,从而提高灵敏度,因为像素相关,降低了误报率,提高了检测范围,并提高了跟踪近距离物体的能力。
{"title":"Performance benefits of sub-diffraction sized pixels in imaging sensors","authors":"J. Caulfield, J. Curzan, N. Dhar","doi":"10.1117/12.2053443","DOIUrl":"https://doi.org/10.1117/12.2053443","url":null,"abstract":"Infrared Focal Plane Arrays have been developed with reductions in pixel size below the Nyquist limit imposed by the optical systems Point Spread Function (PSF). These smaller sub diffraction limited pixels allows spatial oversampling of the image. We show that oversampling the PSF allows improved fidelity in imaging, resulting in sensitivity improvements due to pixel correlation, reduced false alarm rates, improved detection ranges, and an improved ability to track closely spaced objects.","PeriodicalId":210982,"journal":{"name":"2014 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2014-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129734283","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Physical modeling of nuclear detonations in DIRSIG DIRSIG核爆物理模拟
Pub Date : 2013-10-01 DOI: 10.1109/AIPR.2014.7041907
Ashley E. Green, T. Peery, Robert C. Slaughter, J. McClory
Digitized historic film data were used to model the fireball of a nuclear detonation and simulate the sensor response within the Digital Imaging and Remote Sensing Image Generation (DIRSIG) model. Historic films were used to determine the temperature and dimensions of the nuclear fireball and create an input to DIRSIG. DIRSIG was used to analyze how environmental interactions change the optical signal received by a realistic sensor.
数字化的历史胶片数据被用于模拟核爆炸的火球,并在数字成像和遥感图像生成(DIRSIG)模型中模拟传感器的响应。历史上的电影被用来确定核火球的温度和尺寸,并为DIRSIG创建一个输入。DIRSIG用于分析环境相互作用如何改变现实传感器接收到的光信号。
{"title":"Physical modeling of nuclear detonations in DIRSIG","authors":"Ashley E. Green, T. Peery, Robert C. Slaughter, J. McClory","doi":"10.1109/AIPR.2014.7041907","DOIUrl":"https://doi.org/10.1109/AIPR.2014.7041907","url":null,"abstract":"Digitized historic film data were used to model the fireball of a nuclear detonation and simulate the sensor response within the Digital Imaging and Remote Sensing Image Generation (DIRSIG) model. Historic films were used to determine the temperature and dimensions of the nuclear fireball and create an input to DIRSIG. DIRSIG was used to analyze how environmental interactions change the optical signal received by a realistic sensor.","PeriodicalId":210982,"journal":{"name":"2014 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2013-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123059106","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2014 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1