首页 > 最新文献

2019 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)最新文献

英文 中文
Semantic Segmentation of Clouds in Satellite Imagery Using Deep Pre-trained U-Nets 基于深度预训练U-Nets的卫星图像云语义分割
Pub Date : 2019-10-01 DOI: 10.1109/AIPR47015.2019.9174594
Cindy Gonzales, W. Sakla
Earth observation and remote sensing technologies are widely used in various application areas. Because the abundance of collected data requires automated analytics, many communities are utilizing deep convolutional neural networks for such tasks. Automating cloud detection in remote sensing and earth observation imagery is a useful prerequisite for providing quality imagery for further analysis. In this paper, we train a model that uses a deep convolutional U-Net architecture, utilizing transfer learning to perform semantic segmentation of clouds in satellite imagery. Our proposed model outperforms state-of-the-art networks on a benchmark dataset based on several relevant segmentation metrics, including Jaccard Index (+7.69%), precision (+6.21%), and specificity (+0.37%). Moreover, we demonstrate that transfer learning utilizing a 4-channel input into a U-Net architecture is possible and highly performant by using a deep ResNet-style architecture pre-trained on ImageNet for the initialization of weights in three channels (red, green, and blue bands) and random initialization of weights in the fourth channel (near infrared band) of the first convolutional layer of the network.
对地观测和遥感技术广泛应用于各个应用领域。由于大量收集的数据需要自动分析,许多社区正在利用深度卷积神经网络来完成这些任务。在遥感和地球观测图像中实现云检测自动化是为进一步分析提供高质量图像的有用先决条件。在本文中,我们训练了一个使用深度卷积U-Net架构的模型,利用迁移学习对卫星图像中的云进行语义分割。我们提出的模型在基于几个相关分割指标的基准数据集上优于最先进的网络,包括Jaccard指数(+7.69%)、精度(+6.21%)和特异性(+0.37%)。此外,我们证明了利用4通道输入到U-Net架构的迁移学习是可能的,并且通过使用在ImageNet上预训练的深度resnet风格的架构来初始化网络第一卷积层的三个通道(红、绿、蓝波段)的权重,以及在第四个通道(近红外波段)的随机初始化权重。
{"title":"Semantic Segmentation of Clouds in Satellite Imagery Using Deep Pre-trained U-Nets","authors":"Cindy Gonzales, W. Sakla","doi":"10.1109/AIPR47015.2019.9174594","DOIUrl":"https://doi.org/10.1109/AIPR47015.2019.9174594","url":null,"abstract":"Earth observation and remote sensing technologies are widely used in various application areas. Because the abundance of collected data requires automated analytics, many communities are utilizing deep convolutional neural networks for such tasks. Automating cloud detection in remote sensing and earth observation imagery is a useful prerequisite for providing quality imagery for further analysis. In this paper, we train a model that uses a deep convolutional U-Net architecture, utilizing transfer learning to perform semantic segmentation of clouds in satellite imagery. Our proposed model outperforms state-of-the-art networks on a benchmark dataset based on several relevant segmentation metrics, including Jaccard Index (+7.69%), precision (+6.21%), and specificity (+0.37%). Moreover, we demonstrate that transfer learning utilizing a 4-channel input into a U-Net architecture is possible and highly performant by using a deep ResNet-style architecture pre-trained on ImageNet for the initialization of weights in three channels (red, green, and blue bands) and random initialization of weights in the fourth channel (near infrared band) of the first convolutional layer of the network.","PeriodicalId":167075,"journal":{"name":"2019 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134129548","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Segmentation of Infrared Breast Images Using MultiResUnet Neural Networks 基于MultiResUnet神经网络的红外乳房图像分割
Pub Date : 2019-10-01 DOI: 10.1109/AIPR47015.2019.9316541
Ange Lou, Shuyue Guan, Nada Kamona, M. Loew
Breast cancer is the second leading cause of death for women in the U.S. Early detection of breast cancer is key to higher survival rates to breast cancer patients. We are investigating infrared (IR) thermography as a noninvasive adjunct to mammography for breast cancer screening. IR imaging is radiation-free, pain-free, and non-contact. Automatic segmentation of the breast area from the acquired full-size breast IR images will help limit the area for tumor search, as well as reduce the time and effort costs of manual hand segmentation. Autoencoder-like convolutional and deconvolutional neural networks (C-DCNN) had been applied to automatically segment the breast area in IR images in previous studies. In this study, we applied a state-of-the-art deep-learning segmentation model, MultiResUnet, which consists of an encoder part to capture features and a decoder part for precise localization. It was used to segment the breast area by using a set of breast IR images, collected in our clinical trials by imaging breast cancer patients and normal volunteers with our infrared camera (N2 Imager). The database we used has 450 images, acquired from 14 patients and 16 volunteers. We used a thresholding method to remove interference in the raw images and remapped them from the original 16-bit to 8-bit, and then cropped and segmented the 8-bit images manually. Experiments using leave-one-out cross-validation (LOOCV) and comparison with the ground-truth images by using Tanimoto similarity show that the average accuracy of MultiResUnet is 91.47%, which is about 2% higher than that of the autoencoder. MultiResUnet offers a better approach to segment breast IR images than our previous model.
乳腺癌是美国女性死亡的第二大原因,早期发现乳腺癌是提高乳腺癌患者存活率的关键。我们正在研究红外(IR)热成像作为乳房x光检查的无创辅助乳腺癌筛查。红外成像无辐射、无痛、非接触。从获得的全尺寸乳房红外图像中自动分割乳房区域有助于限制肿瘤搜索的区域,并减少人工分割的时间和精力成本。类似自编码器的卷积和反卷积神经网络(C-DCNN)已被应用于红外图像中乳房区域的自动分割。在本研究中,我们应用了最先进的深度学习分割模型MultiResUnet,该模型由一个编码器部分和一个解码器部分组成,用于捕获特征和精确定位。它是通过使用一组乳房红外图像来分割乳房区域,这些图像是在我们的临床试验中收集的,这些图像是用我们的红外相机(N2成像仪)对乳腺癌患者和正常志愿者进行成像的。我们使用的数据库有450张图像,来自14名患者和16名志愿者。我们使用阈值法去除原始图像中的干扰,并将原始图像从16位重新映射到8位,然后对8位图像进行手动裁剪和分割。利用留一交叉验证(LOOCV)和谷本相似度与真实图像的对比实验表明,MultiResUnet的平均准确率为91.47%,比自编码器提高约2%。与我们之前的模型相比,MultiResUnet提供了更好的方法来分割乳房红外图像。
{"title":"Segmentation of Infrared Breast Images Using MultiResUnet Neural Networks","authors":"Ange Lou, Shuyue Guan, Nada Kamona, M. Loew","doi":"10.1109/AIPR47015.2019.9316541","DOIUrl":"https://doi.org/10.1109/AIPR47015.2019.9316541","url":null,"abstract":"Breast cancer is the second leading cause of death for women in the U.S. Early detection of breast cancer is key to higher survival rates to breast cancer patients. We are investigating infrared (IR) thermography as a noninvasive adjunct to mammography for breast cancer screening. IR imaging is radiation-free, pain-free, and non-contact. Automatic segmentation of the breast area from the acquired full-size breast IR images will help limit the area for tumor search, as well as reduce the time and effort costs of manual hand segmentation. Autoencoder-like convolutional and deconvolutional neural networks (C-DCNN) had been applied to automatically segment the breast area in IR images in previous studies. In this study, we applied a state-of-the-art deep-learning segmentation model, MultiResUnet, which consists of an encoder part to capture features and a decoder part for precise localization. It was used to segment the breast area by using a set of breast IR images, collected in our clinical trials by imaging breast cancer patients and normal volunteers with our infrared camera (N2 Imager). The database we used has 450 images, acquired from 14 patients and 16 volunteers. We used a thresholding method to remove interference in the raw images and remapped them from the original 16-bit to 8-bit, and then cropped and segmented the 8-bit images manually. Experiments using leave-one-out cross-validation (LOOCV) and comparison with the ground-truth images by using Tanimoto similarity show that the average accuracy of MultiResUnet is 91.47%, which is about 2% higher than that of the autoencoder. MultiResUnet offers a better approach to segment breast IR images than our previous model.","PeriodicalId":167075,"journal":{"name":"2019 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"1206 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131617723","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
A Visual Feature based Obstacle Avoidance Method for Autonomous Navigation 基于视觉特征的自主导航避障方法
Pub Date : 2019-10-01 DOI: 10.1109/AIPR47015.2019.9174584
Zheng Chen, Malintha Fernando, Lantao Liu
We propose a simple but effective obstacle- avoiding approach for autonomous robot navigation. The method computes local but safe navigation path and relies only on visual feature information extracted from the environment. To achieve this, we first build a discrete set of candidate navigation points in camera’s field of view; then the obstacle avoiding navigation points are selected by evaluating rewards of all candidate points, where the reward metric consists of point-wise transiting probability, safety consideration, mutual information of features, and feature density. Next, we construct a navigable passage in the free space by generating a series of convex hulls that are adjacent to each other. With the navigable passage constructed, a local path that lies within the passage is planned for the robot to safely navigate through. We evaluate the method in both a real world indoor environment as well as a simulated outdoor environment.
提出了一种简单有效的机器人自主导航避障方法。该方法计算局部但安全的导航路径,并且仅依赖于从环境中提取的视觉特征信息。为了实现这一点,我们首先在相机的视场中建立一组离散的候选导航点;然后通过评估所有候选点的奖励来选择避障导航点,其中奖励度量由逐点通过概率、安全考虑、特征互信息和特征密度组成。接下来,我们通过生成一系列彼此相邻的凸包,在自由空间中构建一个可通航的通道。通过构建可通航通道,在通道内规划一条局部路径,供机器人安全通过。我们在真实的室内环境和模拟的室外环境中对该方法进行了评估。
{"title":"A Visual Feature based Obstacle Avoidance Method for Autonomous Navigation","authors":"Zheng Chen, Malintha Fernando, Lantao Liu","doi":"10.1109/AIPR47015.2019.9174584","DOIUrl":"https://doi.org/10.1109/AIPR47015.2019.9174584","url":null,"abstract":"We propose a simple but effective obstacle- avoiding approach for autonomous robot navigation. The method computes local but safe navigation path and relies only on visual feature information extracted from the environment. To achieve this, we first build a discrete set of candidate navigation points in camera’s field of view; then the obstacle avoiding navigation points are selected by evaluating rewards of all candidate points, where the reward metric consists of point-wise transiting probability, safety consideration, mutual information of features, and feature density. Next, we construct a navigable passage in the free space by generating a series of convex hulls that are adjacent to each other. With the navigable passage constructed, a local path that lies within the passage is planned for the robot to safely navigate through. We evaluate the method in both a real world indoor environment as well as a simulated outdoor environment.","PeriodicalId":167075,"journal":{"name":"2019 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"51 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127287848","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Exploring Efficient and Tunable Convolutional Blind Image Denoising Networks 探索有效和可调的卷积盲图像去噪网络
Pub Date : 2019-10-01 DOI: 10.1109/AIPR47015.2019.9174574
Martin Jaszewski, S. Parameswaran
We address the problem of building a blind image denoising network that better adapts to user-defined efficiency and performance requirements. CNN-based architectures such as FFDNet as well as classical methods like BM3D provide fast denoising capability but require the user to specify an approximate noise level. Blind denoising networks like DnCNN and CBDNet are appealing due to their ease of use by non-experts but can be slow. Additionally, these networks are not designed to allow for selecting a reliable operating point based on constraints like available compute, affordable latency, and expected quality. To this end, we propose to develop denoising networks that are tunable to achieve a desired balance between image quality and model size. We seek inspiration from architectures that are tuned for classification, detection, and semantic segmentation on mobile phone CPUs. Incorporating recent advances in architectural building blocks and network architecture search and building upon the success of the DnCNN architectures, we present an efficient convolutional blind image denoising network.
我们解决了建立一个更好地适应用户定义的效率和性能要求的盲图像去噪网络的问题。基于cnn的架构,如FFDNet,以及经典的方法,如BM3D,提供快速去噪能力,但需要用户指定一个近似的噪声水平。像DnCNN和CBDNet这样的盲目去噪网络很有吸引力,因为它们很容易被非专家使用,但速度很慢。此外,这些网络的设计不允许基于可用计算、可负担的延迟和预期质量等约束选择可靠的操作点。为此,我们建议开发可调的去噪网络,以实现图像质量和模型大小之间的理想平衡。我们从针对手机cpu的分类、检测和语义分割的架构中寻求灵感。结合架构构建块和网络架构搜索的最新进展,在DnCNN架构成功的基础上,我们提出了一个高效的卷积盲图像去噪网络。
{"title":"Exploring Efficient and Tunable Convolutional Blind Image Denoising Networks","authors":"Martin Jaszewski, S. Parameswaran","doi":"10.1109/AIPR47015.2019.9174574","DOIUrl":"https://doi.org/10.1109/AIPR47015.2019.9174574","url":null,"abstract":"We address the problem of building a blind image denoising network that better adapts to user-defined efficiency and performance requirements. CNN-based architectures such as FFDNet as well as classical methods like BM3D provide fast denoising capability but require the user to specify an approximate noise level. Blind denoising networks like DnCNN and CBDNet are appealing due to their ease of use by non-experts but can be slow. Additionally, these networks are not designed to allow for selecting a reliable operating point based on constraints like available compute, affordable latency, and expected quality. To this end, we propose to develop denoising networks that are tunable to achieve a desired balance between image quality and model size. We seek inspiration from architectures that are tuned for classification, detection, and semantic segmentation on mobile phone CPUs. Incorporating recent advances in architectural building blocks and network architecture search and building upon the success of the DnCNN architectures, we present an efficient convolutional blind image denoising network.","PeriodicalId":167075,"journal":{"name":"2019 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"166 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121626148","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Detection and Removal of Salt and Pepper Noise by Gaussian Membership Function and Guided Filter 基于高斯隶属函数和引导滤波的椒盐噪声检测与去除
Pub Date : 2019-10-01 DOI: 10.1109/AIPR47015.2019.9174579
Ashish Kumar, Teena Sharma, N. Verma, P. Sircar, S. Vasikarla
The performance of vision-based algorithms depends on the quality of digital images. The images corrupt with salt and pepper noise during image acquisition and transmission deteriorate the performance of these algorithms. This generates the necessity of enhancement algorithms for noise removal. This paper presents an approach for salt and pepper noise removal from digital images. Firstly, the corrupt pixels are detected using a Gaussian membership function and then denoising of these corrupt pixels is performed by a combination of Gaussian and Guided filter. A digital image contains visual information in the form of pixel intensities and regions with abrupt intensity changes. The image regions with similar pixel intensities are called homogeneous and the regions with abrupt intensity changes are called edges or textures. These regions are responsible for carrying important image details. The objective of image denoising is to retrieve the actual pixel intensities in such regions. The proposed approach aims to identify and denoise the corrupt pixels by salt and pepper noise such that the details present in homogeneous regions and edges remain unchanged. For detection of corrupt pixels, two thresholds are estimated using Gaussian membership function. Then, a combination of Gaussian and Guided filter is used for denoising these detected corrupt pixels. The Gaussian filter helps to assign proper weights to the neighborhood pixel set for averaging. However, the Guided filter helps to maintain the structure of an image at very high noise level. The experiments are performed on standard images used in literature with different noise levels up to 99%. It shows that the proposed approach performs efficiently in terms of peak signal to noise ratio.
基于视觉的算法的性能取决于数字图像的质量。在图像采集和传输过程中,椒盐噪声对图像的破坏会降低算法的性能。这就产生了增强算法去噪的必要性。提出了一种去除数字图像中椒盐噪声的方法。首先,利用高斯隶属函数检测腐败像素,然后利用高斯滤波和制导滤波相结合的方法对腐败像素进行去噪。数字图像以像素强度和强度突变区域的形式包含视觉信息。具有相似像素强度的图像区域称为均匀区域,强度变化突然的区域称为边缘或纹理。这些区域负责携带重要的图像细节。图像去噪的目的是获取这些区域的实际像素强度。该方法旨在通过椒盐噪声对腐败像素进行识别和去噪,使均匀区域和边缘的细节保持不变。对于损坏像素的检测,使用高斯隶属函数估计两个阈值。然后,结合高斯滤波和制导滤波对检测到的腐败像素进行去噪。高斯滤波器有助于为邻域像素集分配适当的权值进行平均。然而,引导滤波器有助于在非常高的噪声水平下保持图像的结构。实验是在文献中使用的标准图像上进行的,不同的噪声水平高达99%。结果表明,该方法在峰值信噪比方面是有效的。
{"title":"Detection and Removal of Salt and Pepper Noise by Gaussian Membership Function and Guided Filter","authors":"Ashish Kumar, Teena Sharma, N. Verma, P. Sircar, S. Vasikarla","doi":"10.1109/AIPR47015.2019.9174579","DOIUrl":"https://doi.org/10.1109/AIPR47015.2019.9174579","url":null,"abstract":"The performance of vision-based algorithms depends on the quality of digital images. The images corrupt with salt and pepper noise during image acquisition and transmission deteriorate the performance of these algorithms. This generates the necessity of enhancement algorithms for noise removal. This paper presents an approach for salt and pepper noise removal from digital images. Firstly, the corrupt pixels are detected using a Gaussian membership function and then denoising of these corrupt pixels is performed by a combination of Gaussian and Guided filter. A digital image contains visual information in the form of pixel intensities and regions with abrupt intensity changes. The image regions with similar pixel intensities are called homogeneous and the regions with abrupt intensity changes are called edges or textures. These regions are responsible for carrying important image details. The objective of image denoising is to retrieve the actual pixel intensities in such regions. The proposed approach aims to identify and denoise the corrupt pixels by salt and pepper noise such that the details present in homogeneous regions and edges remain unchanged. For detection of corrupt pixels, two thresholds are estimated using Gaussian membership function. Then, a combination of Gaussian and Guided filter is used for denoising these detected corrupt pixels. The Gaussian filter helps to assign proper weights to the neighborhood pixel set for averaging. However, the Guided filter helps to maintain the structure of an image at very high noise level. The experiments are performed on standard images used in literature with different noise levels up to 99%. It shows that the proposed approach performs efficiently in terms of peak signal to noise ratio.","PeriodicalId":167075,"journal":{"name":"2019 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121645883","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A Method for Online Interpolation of Packet-Loss Blocks in Streaming Video 流媒体视频中丢包块的在线插值方法
Pub Date : 2019-10-01 DOI: 10.1109/AIPR47015.2019.9174590
Rumana Aktar, K. Palaniappan, J. Uhlmann
In this paper we examine and apply a linear-time matrix transformation for online interpolation of missing data blocks in frames of a streaming video sequence. We show that the resulting algorithm produces interpolated pixels that are sufficiently consistent within the context of a single frame that the missing block/tile is typically unnoticed by a viewer of the video sequence. Given the strenuous time constraints imposed by streaming video, this is essentially the only standard of performance that can be applied.
本文研究并应用线性时间矩阵变换对流视频序列帧中缺失数据块进行在线插值。我们表明,所得到的算法产生的插值像素在单个帧的上下文中是足够一致的,因此缺失的块/块通常不会被视频序列的观看者注意到。考虑到流媒体视频所施加的严格时间限制,这基本上是唯一可以应用的性能标准。
{"title":"A Method for Online Interpolation of Packet-Loss Blocks in Streaming Video","authors":"Rumana Aktar, K. Palaniappan, J. Uhlmann","doi":"10.1109/AIPR47015.2019.9174590","DOIUrl":"https://doi.org/10.1109/AIPR47015.2019.9174590","url":null,"abstract":"In this paper we examine and apply a linear-time matrix transformation for online interpolation of missing data blocks in frames of a streaming video sequence. We show that the resulting algorithm produces interpolated pixels that are sufficiently consistent within the context of a single frame that the missing block/tile is typically unnoticed by a viewer of the video sequence. Given the strenuous time constraints imposed by streaming video, this is essentially the only standard of performance that can be applied.","PeriodicalId":167075,"journal":{"name":"2019 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123902610","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Deep Learning Based Integrated Classification and Image Retrieval System for Early Skin Cancer Detection 基于深度学习的早期皮肤癌综合分类与图像检索系统
Pub Date : 2019-10-01 DOI: 10.1109/AIPR47015.2019.9174586
O. Layode, Tasmeer Alam, M. Rahman
Skin cancer is one of the most frequent cancers among human beings. Diagnosing an unknown skin lesion is the first step to determine appropriate treatment. This paper proposes an integrated classification and retrieval based Decision Support System (DSS) for skin cancer detection with an `easy to use’ user interface by applying fusion and ensemble techniques in deep feature spaces. The descriptiveness and discriminative power of features extracted from dermoscopic images are critical to achieve good classification and retrieval performances. In this work, several deep features are extracted based on using transfer learning in several pre-trained Convolutional Neural Networks (CNNs) and Logistic Regression and Support Vector Machine (SVM) models are built as ensembles of classifiers on top of these feature vectors. Furthermore, the content-based image retrieval (CBIR) technique uses the same deep features by fusing those in different feature combinations using a canonical correlation analysis. Based on image-based visual queries submitted by dermatologists, this system would respond by displaying relevant images of pigmented skin lesions of past cases as well as classifying the image category as different types of skin cancer. The system has been trained on a dermoscopic image dataset consists of 1300 images of ten different classes. The best classification (85%) and retrieval accuracies are achieved in a test data set when feature fusion and ensemble techniques are used in all available deep feature spaces. This integrated system would reduce the visual observation error of human operators and enhance clinical decision support for early screening of kin cancers.
皮肤癌是人类中最常见的癌症之一。诊断未知的皮肤病变是确定适当治疗的第一步。本文通过在深度特征空间中应用融合和集成技术,提出了一种基于分类和检索的综合皮肤癌检测决策支持系统(DSS),该系统具有易于使用的用户界面。从皮肤镜图像中提取的特征的描述性和判别能力是获得良好分类和检索性能的关键。在这项工作中,基于在几个预训练的卷积神经网络(cnn)中使用迁移学习提取几个深度特征,并在这些特征向量之上构建逻辑回归和支持向量机(SVM)模型作为分类器的集成。此外,基于内容的图像检索(CBIR)技术通过典型相关分析将不同特征组合中的图像融合,从而获得相同的深度特征。根据皮肤科医生提交的基于图像的视觉查询,该系统将显示过去病例中色素皮肤病变的相关图像,并将图像类别分类为不同类型的皮肤癌。该系统在一个皮肤镜图像数据集上进行了训练,该数据集由十个不同类别的1300张图像组成。当在所有可用的深度特征空间中使用特征融合和集成技术时,在测试数据集中实现了最佳分类(85%)和检索精度。该集成系统将减少操作人员的视觉观察误差,增强临床决策支持,为近亲癌症的早期筛查提供支持。
{"title":"Deep Learning Based Integrated Classification and Image Retrieval System for Early Skin Cancer Detection","authors":"O. Layode, Tasmeer Alam, M. Rahman","doi":"10.1109/AIPR47015.2019.9174586","DOIUrl":"https://doi.org/10.1109/AIPR47015.2019.9174586","url":null,"abstract":"Skin cancer is one of the most frequent cancers among human beings. Diagnosing an unknown skin lesion is the first step to determine appropriate treatment. This paper proposes an integrated classification and retrieval based Decision Support System (DSS) for skin cancer detection with an `easy to use’ user interface by applying fusion and ensemble techniques in deep feature spaces. The descriptiveness and discriminative power of features extracted from dermoscopic images are critical to achieve good classification and retrieval performances. In this work, several deep features are extracted based on using transfer learning in several pre-trained Convolutional Neural Networks (CNNs) and Logistic Regression and Support Vector Machine (SVM) models are built as ensembles of classifiers on top of these feature vectors. Furthermore, the content-based image retrieval (CBIR) technique uses the same deep features by fusing those in different feature combinations using a canonical correlation analysis. Based on image-based visual queries submitted by dermatologists, this system would respond by displaying relevant images of pigmented skin lesions of past cases as well as classifying the image category as different types of skin cancer. The system has been trained on a dermoscopic image dataset consists of 1300 images of ten different classes. The best classification (85%) and retrieval accuracies are achieved in a test data set when feature fusion and ensemble techniques are used in all available deep feature spaces. This integrated system would reduce the visual observation error of human operators and enhance clinical decision support for early screening of kin cancers.","PeriodicalId":167075,"journal":{"name":"2019 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129501461","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
A Comparison of Deep Learning Object Detection Models for Satellite Imagery 卫星图像中深度学习目标检测模型的比较
Pub Date : 2019-10-01 DOI: 10.1109/AIPR47015.2019.9174593
A. Groener, Gary Chern, M. D. Pritt
In this work, we compare the detection accuracy and speed of several state-of-the-art models for the task of detecting oil and gas fracking wells and small cars in commercial electrooptical satellite imagery. Several models are studied from the single-stage, two-stage, and multi-stage object detection families of techniques. For the detection of fracking well pads (50m- 250m), we find single-stage detectors provide superior prediction speed while also matching detection performance of their two and multi-stage counterparts. However, for detecting small cars, two-stage and multi-stage models provide substantially higher accuracies at the cost of some speed. We also measure timing results of the sliding window object detection algorithm to provide a baseline for comparison. Some of these models have been incorporated into the Lockheed Martin Globally-Scalable Automated Target Recognition (GATR) framework.
在这项工作中,我们比较了几种最先进的模型在商业光电卫星图像中检测油气压裂井和小型汽车的精度和速度。从单阶段、两阶段和多阶段目标检测技术中研究了几种模型。对于压裂井台(50m- 250m)的检测,我们发现单级检测器提供了更高的预测速度,同时也可以匹配两级和多级检测器的检测性能。然而,对于小型汽车的检测,两阶段和多级模型提供了更高的精度,但代价是一定的速度。我们还测量了滑动窗口目标检测算法的时序结果,为比较提供了基线。其中一些模型已被纳入洛克希德·马丁公司全球可扩展自动目标识别(GATR)框架。
{"title":"A Comparison of Deep Learning Object Detection Models for Satellite Imagery","authors":"A. Groener, Gary Chern, M. D. Pritt","doi":"10.1109/AIPR47015.2019.9174593","DOIUrl":"https://doi.org/10.1109/AIPR47015.2019.9174593","url":null,"abstract":"In this work, we compare the detection accuracy and speed of several state-of-the-art models for the task of detecting oil and gas fracking wells and small cars in commercial electrooptical satellite imagery. Several models are studied from the single-stage, two-stage, and multi-stage object detection families of techniques. For the detection of fracking well pads (50m- 250m), we find single-stage detectors provide superior prediction speed while also matching detection performance of their two and multi-stage counterparts. However, for detecting small cars, two-stage and multi-stage models provide substantially higher accuracies at the cost of some speed. We also measure timing results of the sliding window object detection algorithm to provide a baseline for comparison. Some of these models have been incorporated into the Lockheed Martin Globally-Scalable Automated Target Recognition (GATR) framework.","PeriodicalId":167075,"journal":{"name":"2019 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"327 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132705828","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Single-Period Single-Frequency (SPSF) Visualization of an EEG’s Striatal Beat Frequency 脑电图纹状体搏动频率的单周期单频可视化
Pub Date : 2019-10-01 DOI: 10.1109/AIPR47015.2019.9174571
J. LaRue
Motivated by the neuroscience concept of striatal beat frequency a new method is presented that takes an electroencephalogram of brain-wave activity and searches for spectral components using a sliding bank of windows consisting of individual singleperiod single-frequency sinusoids. This new approach is now practical due to the presence of high performance computing architectures. And note, this is a departure from legacy time-frequency spectrogram approaches which use one sliding window of constant size to calculate all frequency components simultaneously and it differs from the wavelet method because a suite of wavelets are formulated around one base unit and that unit usually includes a convolutional ramp up and ramp down structure. The proposed single-period singlefrequency search method is more tangible in identifying the existence, in time, of a frequency component due to its self-imposed constraint of using a bank of windows consisting of single period sinusoids, the length of each being a function of frequency and sampling rate. The result of this approach is a rendering of the on-off nature of the conceptualized striatal beat frequency components through a visualization of a matrix of correlation coefficients, where the rows are individual frequencies, (presented up to 1000 Hz in this paper), and where the columns indicate the position of each detected striatal beat frequency window of time.
基于纹状体跳动频率的神经科学概念,提出了一种新的方法,利用脑电波活动的脑电图,利用由单个单周期单频正弦波组成的滑动窗口来搜索频谱成分。由于高性能计算体系结构的出现,这种新方法现在是实用的。注意,这与传统的时频谱图方法不同,它使用一个恒定大小的滑动窗口来同时计算所有频率分量,它与小波方法不同,因为一组小波是围绕一个基本单位制定的,这个基本单位通常包括一个卷积斜坡上升和斜坡下降结构。所提出的单周期单频搜索方法在识别频率分量的存在方面更具体,因为它使用了一组由单周期正弦波组成的窗口,每个窗口的长度是频率和采样率的函数。这种方法的结果是通过相关系数矩阵的可视化来呈现概念化纹状体拍频成分的开关性质,其中行是单个频率,(在本文中呈现为1000 Hz),其中列表示每个检测到的纹状体拍频窗口的位置。
{"title":"Single-Period Single-Frequency (SPSF) Visualization of an EEG’s Striatal Beat Frequency","authors":"J. LaRue","doi":"10.1109/AIPR47015.2019.9174571","DOIUrl":"https://doi.org/10.1109/AIPR47015.2019.9174571","url":null,"abstract":"Motivated by the neuroscience concept of striatal beat frequency a new method is presented that takes an electroencephalogram of brain-wave activity and searches for spectral components using a sliding bank of windows consisting of individual singleperiod single-frequency sinusoids. This new approach is now practical due to the presence of high performance computing architectures. And note, this is a departure from legacy time-frequency spectrogram approaches which use one sliding window of constant size to calculate all frequency components simultaneously and it differs from the wavelet method because a suite of wavelets are formulated around one base unit and that unit usually includes a convolutional ramp up and ramp down structure. The proposed single-period singlefrequency search method is more tangible in identifying the existence, in time, of a frequency component due to its self-imposed constraint of using a bank of windows consisting of single period sinusoids, the length of each being a function of frequency and sampling rate. The result of this approach is a rendering of the on-off nature of the conceptualized striatal beat frequency components through a visualization of a matrix of correlation coefficients, where the rows are individual frequencies, (presented up to 1000 Hz in this paper), and where the columns indicate the position of each detected striatal beat frequency window of time.","PeriodicalId":167075,"journal":{"name":"2019 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"88 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134499579","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Synthetic Data Generation to Mitigate the Low/No-Shot Problem in Machine Learning 合成数据生成以缓解机器学习中的低/无镜头问题
Pub Date : 2019-10-01 DOI: 10.1109/AIPR47015.2019.9174596
Emily E. Berkson, Jared D. VanCor, Steven Esposito, Gary Chern, M. D. Pritt
The low/no-shot problem refers to a lack of available data for training deep learning algorithms. In remote sensing, complete image data sets are rare and do not always include the targets of interest. We propose a method to rapidly generate highfidelity synthetic satellite imagery featuring targets of interest over a range of solar illuminations and platform geometries. Specifically, we used the Digital Imaging and Remote Sensing Image Generation model and a custom image simulator to produce synthetic imagery of C130 aircraft in place of real Worldview-3 imagery. Our synthetic imagery was supplemented with real Worldview-3 images to test the efficacy of training deep learning algorithms with synthetic data. We deliberately chose a challenging test case of distinguishing C130s from other aircraft, or neither. Results show a negligible improvement in automatic target classification when synthetic data is supplemented with a small amount of real imagery. However, training with synthetic data alone only achieves F1-scores in line with a random classifier, suggesting that there is still significant domain mismatch between the real and synthetic datasets.
low/no-shot问题指的是缺乏训练深度学习算法的可用数据。在遥感中,完整的图像数据集很少,而且并不总是包括感兴趣的目标。我们提出了一种快速生成高保真合成卫星图像的方法,该图像具有一系列太阳光照和平台几何形状的感兴趣目标。具体来说,我们使用数字成像和遥感图像生成模型和定制图像模拟器来生成C130飞机的合成图像,以取代真实的Worldview-3图像。我们的合成图像补充了真实的Worldview-3图像,以测试使用合成数据训练深度学习算法的有效性。我们特意选择了一个具有挑战性的测试案例,将c130与其他飞机区分开来,或者两者都不区分。结果表明,当合成数据与少量真实图像相补充时,自动目标分类的改进可以忽略不计。然而,单独使用合成数据进行训练只能达到符合随机分类器的f1分数,这表明真实数据集与合成数据集之间仍然存在明显的领域不匹配。
{"title":"Synthetic Data Generation to Mitigate the Low/No-Shot Problem in Machine Learning","authors":"Emily E. Berkson, Jared D. VanCor, Steven Esposito, Gary Chern, M. D. Pritt","doi":"10.1109/AIPR47015.2019.9174596","DOIUrl":"https://doi.org/10.1109/AIPR47015.2019.9174596","url":null,"abstract":"The low/no-shot problem refers to a lack of available data for training deep learning algorithms. In remote sensing, complete image data sets are rare and do not always include the targets of interest. We propose a method to rapidly generate highfidelity synthetic satellite imagery featuring targets of interest over a range of solar illuminations and platform geometries. Specifically, we used the Digital Imaging and Remote Sensing Image Generation model and a custom image simulator to produce synthetic imagery of C130 aircraft in place of real Worldview-3 imagery. Our synthetic imagery was supplemented with real Worldview-3 images to test the efficacy of training deep learning algorithms with synthetic data. We deliberately chose a challenging test case of distinguishing C130s from other aircraft, or neither. Results show a negligible improvement in automatic target classification when synthetic data is supplemented with a small amount of real imagery. However, training with synthetic data alone only achieves F1-scores in line with a random classifier, suggesting that there is still significant domain mismatch between the real and synthetic datasets.","PeriodicalId":167075,"journal":{"name":"2019 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129624714","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
期刊
2019 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1