Augmenting Data Using Gaussian Mixture Embedding For Improving Land Cover Segmentation

Dario Augusto Borges Oliveira
{"title":"Augmenting Data Using Gaussian Mixture Embedding For Improving Land Cover Segmentation","authors":"Dario Augusto Borges Oliveira","doi":"10.1109/LAGIRS48042.2020.9165670","DOIUrl":null,"url":null,"abstract":"The use of convolutional neural networks improved greatly data synthesis in the last years and have been widely used for data augmentation in scenarios where very imbalanced data is observed, such as land cover segmentation. Balancing the proportion of classes for training segmentation models can be very challenging considering that samples where all classes are reasonably represented might constitute a small portion of a training set, and techniques for augmenting this small amount of data such as rotation, scaling and translation might be not sufficient for efficient training. In this context, this paper proposes a methodology to perform data augmentation from few samples to improve the performance of CNN-based land cover semantic segmentation. First, we estimate the latent data representation of selected training samples by means of a mixture of Gaussians, using an encoder-decoder CNN. Then, we change the latent embedding used to generate the mixture parameters, at random and in training time, to generate new mixture models slightly different from the original. Finally, we compute the displacement maps between the original and the modified mixture models, and use them to elastically deform the original images, creating new realistic samples out of the original ones. Our disentangled approach allows the spatial modification of displacement maps to preserve objects where deformation is undesired, like buildings and cars, where geometry is highly discriminant. With this simple pipeline, we managed to augment samples in training time, and improve the overall performance of two basal semantic segmentation CNN architectures for land cover semantic segmentation.","PeriodicalId":111863,"journal":{"name":"2020 IEEE Latin American GRSS & ISPRS Remote Sensing Conference (LAGIRS)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 IEEE Latin American GRSS & ISPRS Remote Sensing Conference (LAGIRS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/LAGIRS48042.2020.9165670","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

The use of convolutional neural networks improved greatly data synthesis in the last years and have been widely used for data augmentation in scenarios where very imbalanced data is observed, such as land cover segmentation. Balancing the proportion of classes for training segmentation models can be very challenging considering that samples where all classes are reasonably represented might constitute a small portion of a training set, and techniques for augmenting this small amount of data such as rotation, scaling and translation might be not sufficient for efficient training. In this context, this paper proposes a methodology to perform data augmentation from few samples to improve the performance of CNN-based land cover semantic segmentation. First, we estimate the latent data representation of selected training samples by means of a mixture of Gaussians, using an encoder-decoder CNN. Then, we change the latent embedding used to generate the mixture parameters, at random and in training time, to generate new mixture models slightly different from the original. Finally, we compute the displacement maps between the original and the modified mixture models, and use them to elastically deform the original images, creating new realistic samples out of the original ones. Our disentangled approach allows the spatial modification of displacement maps to preserve objects where deformation is undesired, like buildings and cars, where geometry is highly discriminant. With this simple pipeline, we managed to augment samples in training time, and improve the overall performance of two basal semantic segmentation CNN architectures for land cover semantic segmentation.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于高斯混合嵌入的增强数据改进土地覆盖分割
卷积神经网络的使用在过去几年中极大地改善了数据合成,并被广泛用于观察到非常不平衡数据的场景中的数据增强,例如土地覆盖分割。平衡训练分割模型的类的比例是非常具有挑战性的,因为所有类被合理表示的样本可能只占训练集的一小部分,而增加这一小部分数据的技术,如旋转、缩放和平移,可能不足以实现有效的训练。在此背景下,本文提出了一种对少量样本进行数据增强的方法,以提高基于cnn的土地覆盖语义分割的性能。首先,我们使用编码器-解码器CNN,通过混合高斯估计选定训练样本的潜在数据表示。然后,我们在随机和训练时间内改变用于生成混合参数的潜在嵌入,以生成与原始混合模型略有不同的新混合模型。最后,我们计算了原始模型和修改后的混合模型之间的位移映射,并利用它们对原始图像进行弹性变形,从原始图像中创建新的真实样本。我们的解纠缠方法允许对位移图进行空间修改,以保留不需要变形的物体,如建筑物和汽车,这些物体的几何形状是高度区分的。通过这个简单的管道,我们成功地在训练时间内增加了样本,并提高了用于土地覆盖语义分割的两种基本语义分割CNN架构的整体性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Deforestation Polygon Assessment Tool: Providing Comprehensive Information On Deforestation In The Brazilian Cerrado Biome Assessment of rainfall influence on sentinel-1 time series on amazonian tropical forests aiming deforestation detection improvement Spatial Association To Characterize The Climate Teleconnection Patterns In Ecuador Based On Satellite Precipitation Estimates Subsidence in Maceio, Brazil, Characterized by Dinsar and Inverse Modeling Preliminary Analysis For Automatic Tidal Inlets Mapping Using Google Earth Engine
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1