首页 > 最新文献

2018 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)最新文献

英文 中文
Malware Classification using Deep Convolutional Neural Networks 基于深度卷积神经网络的恶意软件分类
Pub Date : 2018-10-01 DOI: 10.1109/AIPR.2018.8707429
David Kornish, Justin Geary, Victor Sansing, Soundararajan Ezekiel, Larry Pearlstein, L. Njilla
In recent years, deep convolution neural networks (DCNNs) have won many contests in machine learning, object detection, and pattern recognition. Furthermore, deep learning techniques achieved exceptional performance in image classification, reaching accuracy levels beyond human capability. Malware variants from similar categories often contain similarities due to code reuse. Converting malware samples into images can cause these patterns to manifest as image features, which can be exploited for DCNN classification. Techniques for converting malware binaries into images for visualization and classification have been reported in the literature, and while these methods do reach a high level of classification accuracy on training datasets, they tend to be vulnerable to overfitting and perform poorly on previously unseen samples. In this paper, we explore and document a variety of techniques for representing malware binaries as images with the goal of discovering a format best suited for deep learning. We implement a database for malware binaries from several families, stored in hexadecimal format. These malware samples are converted into images using various approaches and are used to train a neural network to recognize visual patterns in the input and classify malware based on the feature vectors. Each image type is assessed using a variety of learning models, such as transfer learning with existing DCNN architectures and feature extraction for support vector machine classifier training. Each technique is evaluated in terms of classification accuracy, result consistency, and time per trial. Our preliminary results indicate that improved image representation has the potential to enable more effective classification of new malware.
近年来,深度卷积神经网络(DCNNs)在机器学习、目标检测和模式识别等领域赢得了许多竞赛。此外,深度学习技术在图像分类方面取得了卓越的表现,达到了超越人类能力的精度水平。由于代码重用,来自相似类别的恶意软件变体通常包含相似之处。将恶意软件样本转换为图像可能会导致这些模式表现为图像特征,这可以用于DCNN分类。文献中已经报道了将恶意软件二进制文件转换为图像进行可视化和分类的技术,虽然这些方法确实在训练数据集上达到了很高的分类精度,但它们往往容易受到过拟合的影响,并且在以前未见过的样本上表现不佳。在本文中,我们探索并记录了将恶意软件二进制文件表示为图像的各种技术,目的是发现最适合深度学习的格式。我们实现了一个来自几个家族的恶意软件二进制文件的数据库,以十六进制格式存储。这些恶意软件样本使用各种方法转换成图像,并用于训练神经网络来识别输入中的视觉模式并基于特征向量对恶意软件进行分类。使用各种学习模型评估每种图像类型,例如使用现有DCNN架构的迁移学习和用于支持向量机分类器训练的特征提取。根据分类准确性、结果一致性和每次试验时间对每种技术进行评估。我们的初步结果表明,改进的图像表示有可能使新的恶意软件更有效的分类。
{"title":"Malware Classification using Deep Convolutional Neural Networks","authors":"David Kornish, Justin Geary, Victor Sansing, Soundararajan Ezekiel, Larry Pearlstein, L. Njilla","doi":"10.1109/AIPR.2018.8707429","DOIUrl":"https://doi.org/10.1109/AIPR.2018.8707429","url":null,"abstract":"In recent years, deep convolution neural networks (DCNNs) have won many contests in machine learning, object detection, and pattern recognition. Furthermore, deep learning techniques achieved exceptional performance in image classification, reaching accuracy levels beyond human capability. Malware variants from similar categories often contain similarities due to code reuse. Converting malware samples into images can cause these patterns to manifest as image features, which can be exploited for DCNN classification. Techniques for converting malware binaries into images for visualization and classification have been reported in the literature, and while these methods do reach a high level of classification accuracy on training datasets, they tend to be vulnerable to overfitting and perform poorly on previously unseen samples. In this paper, we explore and document a variety of techniques for representing malware binaries as images with the goal of discovering a format best suited for deep learning. We implement a database for malware binaries from several families, stored in hexadecimal format. These malware samples are converted into images using various approaches and are used to train a neural network to recognize visual patterns in the input and classify malware based on the feature vectors. Each image type is assessed using a variety of learning models, such as transfer learning with existing DCNN architectures and feature extraction for support vector machine classifier training. Each technique is evaluated in terms of classification accuracy, result consistency, and time per trial. Our preliminary results indicate that improved image representation has the potential to enable more effective classification of new malware.","PeriodicalId":230582,"journal":{"name":"2018 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115011057","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
DCNN Augmentation via Synthetic Data from Variational Autoencoders and Generative Adversarial Networks 基于变分自编码器和生成对抗网络合成数据的DCNN增强
Pub Date : 2018-10-01 DOI: 10.1109/AIPR.2018.8707390
David Kornish, Soundararajan Ezekiel, Maria Scalzo-Cornacchia
Deep convolutional neural networks have recently demonstrated incredible capabilities in areas such as image classification and object detection, but they require large datasets of quality pre-labeled data to achieve high levels of performance. Almost all data is not properly labeled when it is captured, and the process of manually labeling large enough datasets for effective learning is impractical in many real-world applications. New studies have shown that synthetic data, generated from a simulated environment, can be effective training data for DCNNs. However, synthetic data is only as effective as the simulation from which it is gathered, and there is often a significant trade-off between designing a simulation that properly models real-world conditions and simply gathering better real-world data. Using generative network architectures, such as Generative Adversarial Networks (GANs) and Variational Autoencoders (VAEs), it is possible to produce new synthetic samples based on the features of real-world data. This data can be used to augment small datasets to increase DCNN performance, similar to traditional augmentation methods such as scaling, translation, rotation, and adding noise. In this paper, we compare the advantages of synthetic data from GANs and VAEs to traditional data augmentation techniques. Initial results are promising, indicating that using synthetic data for augmentation can improve the accuracy of DCNN classifiers.
深度卷积神经网络最近在图像分类和目标检测等领域展示了令人难以置信的能力,但它们需要大量高质量的预标记数据集来实现高水平的性能。几乎所有的数据在被捕获时都没有被正确地标记,并且在许多实际应用中,手动标记足够大的数据集以进行有效学习的过程是不切实际的。新的研究表明,从模拟环境中生成的合成数据可以作为DCNNs的有效训练数据。然而,合成数据的有效性取决于收集到的模拟,在设计一个正确模拟真实世界条件的模拟和简单地收集更好的真实世界数据之间,通常存在一个重要的权衡。使用生成网络架构,如生成对抗网络(gan)和变分自动编码器(VAEs),可以根据现实世界数据的特征生成新的合成样本。该数据可用于增强小数据集以提高DCNN性能,类似于传统的增强方法,如缩放、平移、旋转和添加噪声。在本文中,我们比较了gan和vae合成数据与传统数据增强技术的优势。初步结果是有希望的,表明使用合成数据进行增强可以提高DCNN分类器的准确率。
{"title":"DCNN Augmentation via Synthetic Data from Variational Autoencoders and Generative Adversarial Networks","authors":"David Kornish, Soundararajan Ezekiel, Maria Scalzo-Cornacchia","doi":"10.1109/AIPR.2018.8707390","DOIUrl":"https://doi.org/10.1109/AIPR.2018.8707390","url":null,"abstract":"Deep convolutional neural networks have recently demonstrated incredible capabilities in areas such as image classification and object detection, but they require large datasets of quality pre-labeled data to achieve high levels of performance. Almost all data is not properly labeled when it is captured, and the process of manually labeling large enough datasets for effective learning is impractical in many real-world applications. New studies have shown that synthetic data, generated from a simulated environment, can be effective training data for DCNNs. However, synthetic data is only as effective as the simulation from which it is gathered, and there is often a significant trade-off between designing a simulation that properly models real-world conditions and simply gathering better real-world data. Using generative network architectures, such as Generative Adversarial Networks (GANs) and Variational Autoencoders (VAEs), it is possible to produce new synthetic samples based on the features of real-world data. This data can be used to augment small datasets to increase DCNN performance, similar to traditional augmentation methods such as scaling, translation, rotation, and adding noise. In this paper, we compare the advantages of synthetic data from GANs and VAEs to traditional data augmentation techniques. Initial results are promising, indicating that using synthetic data for augmentation can improve the accuracy of DCNN classifiers.","PeriodicalId":230582,"journal":{"name":"2018 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117122942","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Principles of Dual Fusion Detection 双融合检测原理
Pub Date : 2018-10-01 DOI: 10.1109/AIPR.2018.8707417
A. Schaum
We describe a new approach to solving binary composite hypothesis testing problems and prove its equivalence to a constrained form of clairvoyant fusion. The constraint resolves an abiding theoretical conundrum: the non-commutativity of the fusion order. We then illustrate use of the new constraint by addressing a common limitation in image-based spectral detection, false alarms caused by outliers.
我们描述了一种解决二元复合假设检验问题的新方法,并证明了它等价于一种约束形式的透视融合。约束解决了一个持久的理论难题:融合顺序的非交换性。然后,我们通过解决基于图像的光谱检测中的常见限制,由异常值引起的假警报来说明新约束的使用。
{"title":"Principles of Dual Fusion Detection","authors":"A. Schaum","doi":"10.1109/AIPR.2018.8707417","DOIUrl":"https://doi.org/10.1109/AIPR.2018.8707417","url":null,"abstract":"We describe a new approach to solving binary composite hypothesis testing problems and prove its equivalence to a constrained form of clairvoyant fusion. The constraint resolves an abiding theoretical conundrum: the non-commutativity of the fusion order. We then illustrate use of the new constraint by addressing a common limitation in image-based spectral detection, false alarms caused by outliers.","PeriodicalId":230582,"journal":{"name":"2018 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126439384","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Learning for Recognizing Mobile Targets in Satellite Imagery 基于深度学习的卫星图像移动目标识别
Pub Date : 2018-10-01 DOI: 10.1109/AIPR.2018.8707415
M. D. Pritt
There is an increasing demand for software that automatically detects and classifies mobile targets such as airplanes, cars, and ships in satellite imagery. Applications of such automated target recognition (ATR) software include economic forecasting, traffic planning, maritime law enforcement, and disaster response. This paper describes the extension of a convolutional neural network (CNN) for classification to a sliding window algorithm for detection. It is evaluated on mobile targets of the xView dataset, on which it achieves detection and classification accuracies higher than 95%.
在卫星图像中自动检测和分类飞机、汽车、船舶等移动目标的软件的需求正在增加。这种自动目标识别(ATR)软件的应用包括经济预测、交通规划、海事执法和灾难响应。本文描述了将卷积神经网络(CNN)用于分类扩展到用于检测的滑动窗口算法。在xView数据集的移动目标上进行了评估,检测和分类准确率均在95%以上。
{"title":"Deep Learning for Recognizing Mobile Targets in Satellite Imagery","authors":"M. D. Pritt","doi":"10.1109/AIPR.2018.8707415","DOIUrl":"https://doi.org/10.1109/AIPR.2018.8707415","url":null,"abstract":"There is an increasing demand for software that automatically detects and classifies mobile targets such as airplanes, cars, and ships in satellite imagery. Applications of such automated target recognition (ATR) software include economic forecasting, traffic planning, maritime law enforcement, and disaster response. This paper describes the extension of a convolutional neural network (CNN) for classification to a sliding window algorithm for detection. It is evaluated on mobile targets of the xView dataset, on which it achieves detection and classification accuracies higher than 95%.","PeriodicalId":230582,"journal":{"name":"2018 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127883715","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Exploring the Effects of Class-Specific Augmentation and Class Coalescence on Deep Neural Network Performance Using a Novel Road Feature Dataset 使用新的道路特征数据集探索类别特定增强和类别合并对深度神经网络性能的影响
Pub Date : 2018-10-01 DOI: 10.1109/AIPR.2018.8707406
Tyler W. Nivin, G. Scott, J. A. Hurt, Raymond L. Chastain, C. Davis
The identification of nodal road network features in remote sensing imagery is an important object detection task due to its versatility of application. A successful capability enables urban sprawl tracking, automatic or semi-automated map accuracy validation and updating, and macro-scale infrastructure damage evaluation and tracking just to name a few. We have curated a custom, novel dataset that includes nodal road network features such as bridges, cul-de-sacs, freeway exchanges and exits, freeway overpasses, intersections, and traffic circles. From this curated data we have evaluated the use of deep machine learning for object recognition across two variations in this image dataset. These variations are expanded versus semantically coalesced classes. We have evaluated the performance of two deep convolutional neural networks, ResNet50 and Xception, to detect these features across these variations of the image datasets. We have also explored the use of class-specific data augmentation to improve the performance of the models trained for nodal road network feature detection. Cross-validation performance of the models evaluated on four variations of this nodal road network feature dataset range from 0.81 to 0.96 (F1 scores). Coalescing highly specific, semantically challenging classes into more semantically generalized classes has a significant impact on the accuracy of the models. Our analysis provides insight into if and how these techniques can improve the performance of machine learning models, facilitating application to broad area imagery analysis in numerous application domains.
遥感影像中节点路网特征的识别是一项重要的目标检测任务,具有广泛的应用前景。一个成功的功能可以实现城市扩张跟踪,自动或半自动地图准确性验证和更新,宏观规模的基础设施损坏评估和跟踪,仅举几例。我们策划了一个定制的、新颖的数据集,其中包括节点道路网络特征,如桥梁、死胡同、高速公路交换和出口、高速公路立交桥、交叉路口和交通圈。从这些精心整理的数据中,我们评估了深度机器学习在该图像数据集中的两个变量中用于对象识别的使用。这些变体是扩展的,而不是语义合并的类。我们已经评估了两个深度卷积神经网络ResNet50和Xception的性能,以在这些图像数据集的变化中检测这些特征。我们还探索了使用特定类别的数据增强来提高用于节点道路网络特征检测的训练模型的性能。在该节点路网特征数据集的四种变量上评估的模型交叉验证性能范围为0.81至0.96 (F1分数)。将高度特定的、语义上具有挑战性的类合并为语义上更一般化的类对模型的准确性有重大影响。我们的分析提供了这些技术是否以及如何提高机器学习模型的性能的见解,促进了在众多应用领域的广域图像分析的应用。
{"title":"Exploring the Effects of Class-Specific Augmentation and Class Coalescence on Deep Neural Network Performance Using a Novel Road Feature Dataset","authors":"Tyler W. Nivin, G. Scott, J. A. Hurt, Raymond L. Chastain, C. Davis","doi":"10.1109/AIPR.2018.8707406","DOIUrl":"https://doi.org/10.1109/AIPR.2018.8707406","url":null,"abstract":"The identification of nodal road network features in remote sensing imagery is an important object detection task due to its versatility of application. A successful capability enables urban sprawl tracking, automatic or semi-automated map accuracy validation and updating, and macro-scale infrastructure damage evaluation and tracking just to name a few. We have curated a custom, novel dataset that includes nodal road network features such as bridges, cul-de-sacs, freeway exchanges and exits, freeway overpasses, intersections, and traffic circles. From this curated data we have evaluated the use of deep machine learning for object recognition across two variations in this image dataset. These variations are expanded versus semantically coalesced classes. We have evaluated the performance of two deep convolutional neural networks, ResNet50 and Xception, to detect these features across these variations of the image datasets. We have also explored the use of class-specific data augmentation to improve the performance of the models trained for nodal road network feature detection. Cross-validation performance of the models evaluated on four variations of this nodal road network feature dataset range from 0.81 to 0.96 (F1 scores). Coalescing highly specific, semantically challenging classes into more semantically generalized classes has a significant impact on the accuracy of the models. Our analysis provides insight into if and how these techniques can improve the performance of machine learning models, facilitating application to broad area imagery analysis in numerous application domains.","PeriodicalId":230582,"journal":{"name":"2018 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134254228","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Cybersecurity Considerations for Image Pattern Recognition Applications 图像模式识别应用的网络安全考虑
Pub Date : 2018-10-01 DOI: 10.1109/AIPR.2018.8707427
J. Straub
Pattern recognition and image analysis find application in numerous areas. They support law enforcement, transportation security and warfighting. They also have numerous commercial purposes, ranging from agriculture to manufacturing to facility security. This paper discusses the security needs of image recognition systems and the particular concerns that the algorithms and methods used for these activities pose. It also discusses how to secure these analysis systems and the future work required in these areas.
模式识别和图像分析在许多领域都有应用。他们支持执法、运输安全和作战。它们也有许多商业用途,从农业到制造业再到设施安全。本文讨论了图像识别系统的安全需求以及用于这些活动的算法和方法所带来的特殊关注。本文还讨论了如何保护这些分析系统,以及未来在这些领域需要做的工作。
{"title":"Cybersecurity Considerations for Image Pattern Recognition Applications","authors":"J. Straub","doi":"10.1109/AIPR.2018.8707427","DOIUrl":"https://doi.org/10.1109/AIPR.2018.8707427","url":null,"abstract":"Pattern recognition and image analysis find application in numerous areas. They support law enforcement, transportation security and warfighting. They also have numerous commercial purposes, ranging from agriculture to manufacturing to facility security. This paper discusses the security needs of image recognition systems and the particular concerns that the algorithms and methods used for these activities pose. It also discusses how to secure these analysis systems and the future work required in these areas.","PeriodicalId":230582,"journal":{"name":"2018 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"37 10","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121002923","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Characterizing the Visual Social Media Environment of Eating Disorders 表征饮食失调的视觉社交媒体环境
Pub Date : 2018-10-01 DOI: 10.1109/AIPR.2018.8707400
Samsara N. Counts, J. Manning, Robert Pless
Eating disorders are often exacerbated by exposure to triggering images on social media. Standard approaches to filtering of social media by detecting hashtags or keywords are difficult to keep accurate because those migrate or change over time. In this work we present proof-of-concept demonstrations to show that Deep Learning classification algorithms are effective at classifying images related to eating disorders. We discuss some of the challenges in this domain and show that careful curation of the training data improves performance substantially.
社交媒体上的刺激性图片往往会加剧饮食失调。通过检测主题标签或关键字来过滤社交媒体的标准方法很难保持准确,因为它们会随着时间的推移而迁移或变化。在这项工作中,我们提出了概念验证演示,表明深度学习分类算法在分类与饮食失调相关的图像方面是有效的。我们讨论了该领域的一些挑战,并表明仔细管理训练数据可以大大提高性能。
{"title":"Characterizing the Visual Social Media Environment of Eating Disorders","authors":"Samsara N. Counts, J. Manning, Robert Pless","doi":"10.1109/AIPR.2018.8707400","DOIUrl":"https://doi.org/10.1109/AIPR.2018.8707400","url":null,"abstract":"Eating disorders are often exacerbated by exposure to triggering images on social media. Standard approaches to filtering of social media by detecting hashtags or keywords are difficult to keep accurate because those migrate or change over time. In this work we present proof-of-concept demonstrations to show that Deep Learning classification algorithms are effective at classifying images related to eating disorders. We discuss some of the challenges in this domain and show that careful curation of the training data improves performance substantially.","PeriodicalId":230582,"journal":{"name":"2018 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"99 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117235738","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Predicting Interpretability Loss in Thermal IR Imagery due to Compression 热红外图像压缩可解释性损失预测
Pub Date : 2018-10-01 DOI: 10.1109/AIPR.2018.8707416
Hua-mei Chen, J. Irvine, Zhonghai Wang, Genshe Chen, Erik Blasch, James Nagy
Analysis of thermal Infrared (IR) imagery is critical to many law enforcement and military missions, particularly for operations at night or in low-light conditions. Transmitting the imagery data from the sensor to the operator often relies on limited bandwidth channels, leading to information loss. This paper develops a method, known as the Compression Degradation Image Function Index (CoDIFI) framework, that predicts the degradation in interpretability associated with the specific image compression method and level of compression. Quantification of the image interpretability relies on the National Imagery Interpretability Ratings Scale (NIIRS). Building on previously reported development and validation of CoDIFI operating on electro-optical (EO) imagery collected in the visible region, this paper extends CoDIFI to imagery collected in the mid-wave infrared (MWIR) region, approximately 3 to 5 microns. For the infrared imagery application, the IR NIIRS is the standard for quantifying image interpretability and the prediction model rests on the general image quality equation (GIQE). A prediction model using the CoDIFI for IR imagery is established with empirical validation. By leveraging the CoDIFI in operational settings, mission success ensures that the compression selection is achievable in terms of the NIIRS level of imagery data delivered to users, while optimizing the use of scarce data transmission capacity.
热红外(IR)图像的分析对许多执法和军事任务至关重要,特别是在夜间或低光条件下的行动。将图像数据从传感器传输到操作员往往依赖于有限的带宽通道,导致信息丢失。本文开发了一种方法,称为压缩退化图像函数索引(CoDIFI)框架,该框架预测与特定图像压缩方法和压缩级别相关的可解释性退化。图像可解释性的量化依赖于国家图像可解释性评级量表(NIIRS)。基于先前报道的CoDIFI在可见光区域收集的光电(EO)图像上的发展和验证,本文将CoDIFI扩展到中波红外(MWIR)区域收集的图像,大约3到5微米。在红外成像应用中,红外近红外光谱是量化图像可解释性的标准,预测模型基于一般图像质量方程(GIQE)。建立了基于CoDIFI的红外图像预测模型,并进行了实证验证。通过在操作设置中利用CoDIFI,任务成功确保在交付给用户的NIIRS图像数据级别方面可以实现压缩选择,同时优化稀缺数据传输容量的使用。
{"title":"Predicting Interpretability Loss in Thermal IR Imagery due to Compression","authors":"Hua-mei Chen, J. Irvine, Zhonghai Wang, Genshe Chen, Erik Blasch, James Nagy","doi":"10.1109/AIPR.2018.8707416","DOIUrl":"https://doi.org/10.1109/AIPR.2018.8707416","url":null,"abstract":"Analysis of thermal Infrared (IR) imagery is critical to many law enforcement and military missions, particularly for operations at night or in low-light conditions. Transmitting the imagery data from the sensor to the operator often relies on limited bandwidth channels, leading to information loss. This paper develops a method, known as the Compression Degradation Image Function Index (CoDIFI) framework, that predicts the degradation in interpretability associated with the specific image compression method and level of compression. Quantification of the image interpretability relies on the National Imagery Interpretability Ratings Scale (NIIRS). Building on previously reported development and validation of CoDIFI operating on electro-optical (EO) imagery collected in the visible region, this paper extends CoDIFI to imagery collected in the mid-wave infrared (MWIR) region, approximately 3 to 5 microns. For the infrared imagery application, the IR NIIRS is the standard for quantifying image interpretability and the prediction model rests on the general image quality equation (GIQE). A prediction model using the CoDIFI for IR imagery is established with empirical validation. By leveraging the CoDIFI in operational settings, mission success ensures that the compression selection is achievable in terms of the NIIRS level of imagery data delivered to users, while optimizing the use of scarce data transmission capacity.","PeriodicalId":230582,"journal":{"name":"2018 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"283 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121059695","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Normalcy Modeling Using a Dictionary of Activities Learned from Motion Imagery Tracking Data 使用从运动图像跟踪数据中学习的活动字典进行常态化建模
Pub Date : 2018-10-01 DOI: 10.1109/AIPR.2018.8707422
J. Irvine, L. Mariano, Teal Guidici
Target tracking derived from motion imagery provides a capability to detect, recognize, and analyze activities in a manner not possible with still images. Target tracking enables automated activity analysis. In this paper, we develop methods for automatically exploiting the tracking data derived from motion imagery, or other tracking data, to detect and recognize activities, develop models of normal behavior, and detect departure from normalcy. The critical steps in our approach are to construct a syntactic representation of the track behaviors and map this representation to a small set of learned activities. We have developed methods for representing activities through syntactic analysis of the track data, by "tokenizing" the track, i.e. converting the kinematic information into strings of symbols amenable to further analysis. The syntactic analysis of target tracks is the foundation for constructing an expandable dictionary of activities. Through unsupervised learning on the tokenized track data we discovery the common activities. The probability distribution of these learned activities is the "dictionary". Newly acquired track data is compared to the dictionary to flag atypical behaviors as departures from normalcy. We demonstrate the methods with two relevant data sets: the Porto taxi data and a set of video data acquired at Draper. These data sets illustrate the flexibility and power of these methods for activity analysis.
来源于运动图像的目标跟踪提供了一种检测、识别和分析活动的能力,这是静止图像无法做到的。目标跟踪支持自动活动分析。在本文中,我们开发了自动利用来自运动图像或其他跟踪数据的跟踪数据的方法,以检测和识别活动,开发正常行为模型,并检测偏离正常。我们方法的关键步骤是构建轨迹行为的语法表示,并将这种表示映射到一组学习到的活动。我们已经开发了通过轨迹数据的语法分析来表示活动的方法,通过将轨迹“标记化”,即将运动信息转换为便于进一步分析的符号字符串。目标轨迹的句法分析是构建可扩展活动字典的基础。通过对标记化轨道数据的无监督学习,发现共同的活动。这些学习到的活动的概率分布就是“字典”。将新获得的航迹数据与字典进行比较,将非典型行为标记为偏离正常。我们用两个相关的数据集来演示这些方法:波尔图出租车数据和在德雷珀获得的一组视频数据。这些数据集说明了这些活动分析方法的灵活性和强大性。
{"title":"Normalcy Modeling Using a Dictionary of Activities Learned from Motion Imagery Tracking Data","authors":"J. Irvine, L. Mariano, Teal Guidici","doi":"10.1109/AIPR.2018.8707422","DOIUrl":"https://doi.org/10.1109/AIPR.2018.8707422","url":null,"abstract":"Target tracking derived from motion imagery provides a capability to detect, recognize, and analyze activities in a manner not possible with still images. Target tracking enables automated activity analysis. In this paper, we develop methods for automatically exploiting the tracking data derived from motion imagery, or other tracking data, to detect and recognize activities, develop models of normal behavior, and detect departure from normalcy. The critical steps in our approach are to construct a syntactic representation of the track behaviors and map this representation to a small set of learned activities. We have developed methods for representing activities through syntactic analysis of the track data, by \"tokenizing\" the track, i.e. converting the kinematic information into strings of symbols amenable to further analysis. The syntactic analysis of target tracks is the foundation for constructing an expandable dictionary of activities. Through unsupervised learning on the tokenized track data we discovery the common activities. The probability distribution of these learned activities is the \"dictionary\". Newly acquired track data is compared to the dictionary to flag atypical behaviors as departures from normalcy. We demonstrate the methods with two relevant data sets: the Porto taxi data and a set of video data acquired at Draper. These data sets illustrate the flexibility and power of these methods for activity analysis.","PeriodicalId":230582,"journal":{"name":"2018 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116796231","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Benchmark Meta-Dataset of High-Resolution Remote Sensing Imagery for Training Robust Deep Learning Models in Machine-Assisted Visual Analytics 用于训练机器辅助视觉分析中鲁棒深度学习模型的高分辨率遥感图像基准元数据集
Pub Date : 2018-10-01 DOI: 10.1109/AIPR.2018.8707433
J. A. Hurt, G. Scott, Derek T. Anderson, C. Davis
Recent years have seen the publication of various high-resolution remote sensing imagery benchmark datasets. These datasets, while diverse in design, have many co-occurring object classes that are of interest for various application domains of Earth observation. In this research, we present our evaluation of a new meta-benchmark dataset combining object classes from the UC Merced, WHU-RS19, PatternNet, and RESISC-45 benchmark datasets. We provide open-source resources to acquire the individual benchmark datasets and then agglomerate them into a new meta-dataset (MDS). Prior research has shown that contemporary deep convolutional neural networks are able to achieve cross-validation accuracies in the range of 95-100% for the 33 identified object classes. Our analysis shows that the overall accuracy for all object classes from these benchmarks is approximately 98.6%. In this work, we investigate the utility of agglomerating the benchmarks into an MDS to train more generalizable, and therefore translatable from lab to real-world, deep machine learning (DML) models. We evaluate numerous state-of-the-art architectures, as well as our data-driven DML model fusion techniques. Finally, we compare MDS performance with that of the benchmark datasets to evaluate the performance versus cost trade-off of using multiple DML in an ensemble system.
近年来出版了各种高分辨率遥感图像基准数据集。这些数据集虽然设计多样,但有许多共同发生的对象类,这些对象类对地球观测的各种应用领域都很感兴趣。在这项研究中,我们展示了我们对一个新的元基准数据集的评估,该数据集结合了来自UC Merced、WHU-RS19、PatternNet和RESISC-45基准数据集的对象类。我们提供开源资源来获取单个基准数据集,然后将它们聚合成一个新的元数据集(MDS)。先前的研究表明,当代深度卷积神经网络能够在33个已识别的对象类别中实现95-100%的交叉验证精度。我们的分析表明,这些基准测试中所有对象类的总体准确率约为98.6%。在这项工作中,我们研究了将基准聚合到MDS中的效用,以训练更一般化的,因此可以从实验室转换到现实世界的深度机器学习(DML)模型。我们评估了许多最先进的架构,以及我们的数据驱动的DML模型融合技术。最后,我们将MDS性能与基准数据集的性能进行比较,以评估在集成系统中使用多个DML的性能与成本权衡。
{"title":"Benchmark Meta-Dataset of High-Resolution Remote Sensing Imagery for Training Robust Deep Learning Models in Machine-Assisted Visual Analytics","authors":"J. A. Hurt, G. Scott, Derek T. Anderson, C. Davis","doi":"10.1109/AIPR.2018.8707433","DOIUrl":"https://doi.org/10.1109/AIPR.2018.8707433","url":null,"abstract":"Recent years have seen the publication of various high-resolution remote sensing imagery benchmark datasets. These datasets, while diverse in design, have many co-occurring object classes that are of interest for various application domains of Earth observation. In this research, we present our evaluation of a new meta-benchmark dataset combining object classes from the UC Merced, WHU-RS19, PatternNet, and RESISC-45 benchmark datasets. We provide open-source resources to acquire the individual benchmark datasets and then agglomerate them into a new meta-dataset (MDS). Prior research has shown that contemporary deep convolutional neural networks are able to achieve cross-validation accuracies in the range of 95-100% for the 33 identified object classes. Our analysis shows that the overall accuracy for all object classes from these benchmarks is approximately 98.6%. In this work, we investigate the utility of agglomerating the benchmarks into an MDS to train more generalizable, and therefore translatable from lab to real-world, deep machine learning (DML) models. We evaluate numerous state-of-the-art architectures, as well as our data-driven DML model fusion techniques. Finally, we compare MDS performance with that of the benchmark datasets to evaluate the performance versus cost trade-off of using multiple DML in an ensemble system.","PeriodicalId":230582,"journal":{"name":"2018 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"70 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131282838","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
期刊
2018 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1