From simple labels to semantic image segmentation: leveraging citizen science plant photographs for tree species mapping in drone imagery

Salim Soltani, Olga Ferlian, Nico Eisenhauer, H. Feilhauer, T. Kattenborn
{"title":"From simple labels to semantic image segmentation: leveraging citizen science plant photographs for tree species mapping in drone imagery","authors":"Salim Soltani, Olga Ferlian, Nico Eisenhauer, H. Feilhauer, T. Kattenborn","doi":"10.5194/bg-21-2909-2024","DOIUrl":null,"url":null,"abstract":"Abstract. Knowledge of plant species distributions is essential for various application fields, such as nature conservation, agriculture, and forestry. Remote sensing data, especially high-resolution orthoimages from unoccupied aerial vehicles (UAVs), paired with novel pattern-recognition methods, such as convolutional neural networks (CNNs), enable accurate mapping (segmentation) of plant species. Training transferable pattern-recognition models for species segmentation across diverse landscapes and data characteristics typically requires extensive training data. Training data are usually derived from labor-intensive field surveys or visual interpretation of remote sensing images. Alternatively, pattern-recognition models could be trained more efficiently with plant photos and labels from citizen science platforms, which include millions of crowd-sourced smartphone photos and the corresponding species labels. However, these pairs of citizen-science-based photographs and simple species labels (one label for the entire image) cannot be used directly for training state-of-the-art segmentation models used for UAV image analysis, which require per-pixel labels for training (also called masks). Here, we overcome the limitation of simple labels of citizen science plant observations with a two-step approach. In the first step, we train CNN-based image classification models using the simple labels and apply them in a moving-window approach over UAV orthoimagery to create segmentation masks. In the second phase, these segmentation masks are used to train state-of-the-art CNN-based image segmentation models with an encoder–decoder structure. We tested the approach on UAV orthoimages acquired in summer and autumn at a test site comprising 10 temperate deciduous tree species in varying mixtures. Several tree species could be mapped with surprising accuracy (mean F1 score =0.47). In homogenous species assemblages, the accuracy increased considerably (mean F1 score =0.55). The results indicate that several tree species can be mapped without generating new training data and by only using preexisting knowledge from citizen science. Moreover, our analysis revealed that the variability in citizen science photographs, with respect to acquisition data and context, facilitates the generation of models that are transferable through the vegetation season. Thus, citizen science data may greatly advance our capacity to monitor hundreds of plant species and, thus, Earth's biodiversity across space and time.\n","PeriodicalId":502171,"journal":{"name":"Biogeosciences","volume":"107 3","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-06-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Biogeosciences","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.5194/bg-21-2909-2024","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Abstract. Knowledge of plant species distributions is essential for various application fields, such as nature conservation, agriculture, and forestry. Remote sensing data, especially high-resolution orthoimages from unoccupied aerial vehicles (UAVs), paired with novel pattern-recognition methods, such as convolutional neural networks (CNNs), enable accurate mapping (segmentation) of plant species. Training transferable pattern-recognition models for species segmentation across diverse landscapes and data characteristics typically requires extensive training data. Training data are usually derived from labor-intensive field surveys or visual interpretation of remote sensing images. Alternatively, pattern-recognition models could be trained more efficiently with plant photos and labels from citizen science platforms, which include millions of crowd-sourced smartphone photos and the corresponding species labels. However, these pairs of citizen-science-based photographs and simple species labels (one label for the entire image) cannot be used directly for training state-of-the-art segmentation models used for UAV image analysis, which require per-pixel labels for training (also called masks). Here, we overcome the limitation of simple labels of citizen science plant observations with a two-step approach. In the first step, we train CNN-based image classification models using the simple labels and apply them in a moving-window approach over UAV orthoimagery to create segmentation masks. In the second phase, these segmentation masks are used to train state-of-the-art CNN-based image segmentation models with an encoder–decoder structure. We tested the approach on UAV orthoimages acquired in summer and autumn at a test site comprising 10 temperate deciduous tree species in varying mixtures. Several tree species could be mapped with surprising accuracy (mean F1 score =0.47). In homogenous species assemblages, the accuracy increased considerably (mean F1 score =0.55). The results indicate that several tree species can be mapped without generating new training data and by only using preexisting knowledge from citizen science. Moreover, our analysis revealed that the variability in citizen science photographs, with respect to acquisition data and context, facilitates the generation of models that are transferable through the vegetation season. Thus, citizen science data may greatly advance our capacity to monitor hundreds of plant species and, thus, Earth's biodiversity across space and time.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
从简单标签到语义图像分割:利用公民科学植物照片绘制无人机图像中的树种图
摘要植物物种分布知识对于自然保护、农业和林业等多个应用领域至关重要。遥感数据,特别是无人飞行器(UAV)拍摄的高分辨率正射影像,与卷积神经网络(CNN)等新型模式识别方法相配合,可实现植物物种的精确绘图(分割)。要训练可在不同地貌和数据特征下转移的模式识别模型来进行物种划分,通常需要大量的训练数据。训练数据通常来自劳动密集型实地调查或遥感图像的目视判读。另外,使用来自公民科学平台的植物照片和标签也能更有效地训练模式识别模型,其中包括数以百万计的众包智能手机照片和相应的物种标签。然而,这些基于公民科学的照片和简单的物种标签(整个图像一个标签)不能直接用于训练无人机图像分析中使用的最先进的分割模型,这些模型需要每个像素的标签(也称为掩码)来进行训练。在此,我们采用两步法克服了公民科学植物观测中简单标签的限制。第一步,我们使用简单标签训练基于 CNN 的图像分类模型,并将其应用于无人机正射影像的移动窗口方法,以创建分割掩码。在第二阶段,这些分割掩码被用于训练基于 CNN 的最先进的图像分割模型,该模型具有编码器-解码器结构。我们在夏季和秋季在一个测试地点采集的无人机正射影像上测试了该方法,该地点有 10 种温带落叶树,树种混杂不一。一些树种的映射精确度令人惊讶(平均 F1 分数 =0.47)。在同质树种组合中,精确度大幅提高(平均 F1 分数 =0.55)。结果表明,无需生成新的训练数据,只需利用公民科学中已有的知识,就能绘制出多个树种的分布图。此外,我们的分析表明,公民科学照片在采集数据和背景方面的可变性有助于生成可在植被季节中转移的模型。因此,公民科学数据可能会大大提高我们监测数百种植物物种的能力,从而提高地球跨时空生物多样性的能力。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
The impacts of modelling prescribed vs. dynamic land cover in a high-CO2 future scenario – greening of the Arctic and Amazonian dieback Linking geomorphological processes and wildlife microhabitat selection: nesting birds select refuges generated by permafrost degradation in the Arctic The biological and preformed carbon pumps in perpetually slower and warmer oceans Climate-based prediction of carbon fluxes from deadwood in Australia Composite calcite and opal test in Foraminifera (Rhizaria)
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1