首页 > 最新文献

ISPRS Journal of Photogrammetry and Remote Sensing最新文献

英文 中文
Quality control tests for automated above-water hyperspectral measurements: Radiative Transfer assessment 自动水上高光谱测量的质量控制测试:辐射传输评估
IF 10.6 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2024-07-17 DOI: 10.1016/j.isprsjprs.2024.07.011

Automated above-water hyperspectral observations are often subject to inaccuracies caused by instrument malfunction and environmental conditions. This study evaluates the influence of atmospheric and water surface conditions on above-water hyperspectral measurements through statistical methods and Radiative Transfer (RT) modelling. Initially, we developed a general quality control method based on statistical assessment to detect the suspicious spectra. Subsequently, Radiative Transfer (RT) models were used to assess low light conditions, distortions in the spectral shape of above-water solar downwelling irradiance (ES(λ), mW m−2 nm−1) particularly those caused by intense atmospheric scattering and/or reddish hue of dusk or dawn radiation, the effect of atmospheric humidity and precipitation on the intensity and shape of spectra, and the influence of sun glint and surface perturbations on sky (LS(λ), mW m−2 nm−1 sr−1) and water surface (LT(λ), mW m−2 nm−1 sr−1) radiances. The proposed methods were applied to the entire archive of automated above-water hyperspectral measurements collected every ten minutes from 2020 to 2022 at the Royal Netherland Institute for Sea Research (NIOZ) at Jetty Station (NJS) located in the Marsdiep tidal inlet of the Duch Wadden Sea, the Netherlands. The findings demonstrate that low light conditions are characterized by ES(λ)max ≤ 25 mW m−2 nm−1. Red-shifted or distorted spectra are indicated by a ratio of ES(4 8 0)/ES(6 8 0) ≤ 1.0 and ESmax)/ ES(8 6 5) ≤ 1.25. High humidity/precipitation conditions are identified by the ratio of ES(9 4 0)/ES(8 6 5), which varies with the Solar Zenith Angle (SZA). Furthermore, significant sun glint and surface perturbations, such as whitecaps and foam, are indicated when the minimum ratio of LT(800 nm-950 nm)/ES(800 nm-950 nm) > 0.025 sr−1, and the ratio of LT(850 nm)/ES(850 nm) ≥ 0.025 sr−1.

自动水上高光谱观测往往会受到仪器故障和环境条件的影响而出现误差。本研究通过统计方法和辐射传递(RT)建模,评估了大气和水面条件对水上高光谱测量的影响。首先,我们开发了一种基于统计评估的通用质量控制方法来检测可疑光谱。随后,我们使用辐射传递(RT)模型来评估弱光条件、水面上太阳下沉辐照度(ES(λ), mW m-2 nm-1)光谱形状的扭曲,特别是那些由强烈的大气散射和/或黄昏或黎明辐射的红色色调引起的扭曲、大气湿度和降水对光谱强度和形状的影响,以及太阳微光和表面扰动对天空(LS(λ), mW m-2 nm-1 sr-1)和水面(LT(λ), mW m-2 nm-1 sr-1)辐射率的影响。所提出的方法被应用于荷兰皇家荷兰海洋研究所(NIOZ)位于荷兰杜赫瓦登海马尔斯迪普潮汐入口处的码头站(NJS)从 2020 年到 2022 年每十分钟收集一次的全部自动水上高光谱测量档案。研究结果表明,弱光条件下的 ES(λ)max ≤ 25 mW m-2 nm-1。ES(4 8 0)/ES(6 8 0) ≤ 1.0 和 ES(λmax)/ ES(8 6 5) ≤ 1.25 的比率表示光谱红移或扭曲。ES(9 4 0)/ES(8 6 5)的比值随太阳天顶角(SZA)的变化而变化。此外,当 LT(800 nm-950 nm)/ES(800 nm-950 nm) > 0.025 sr-1 的最小比值和 LT(850 nm)/ES(850 nm) 的比值≥ 0.025 sr-1 时,会出现明显的太阳闪光和表面扰动,如白帽和泡沫。
{"title":"Quality control tests for automated above-water hyperspectral measurements: Radiative Transfer assessment","authors":"","doi":"10.1016/j.isprsjprs.2024.07.011","DOIUrl":"10.1016/j.isprsjprs.2024.07.011","url":null,"abstract":"<div><p>Automated above-water hyperspectral observations are often subject to inaccuracies caused by instrument malfunction and environmental conditions. This study evaluates the influence of atmospheric and water surface conditions on above-water hyperspectral measurements through statistical methods and Radiative Transfer (RT) modelling. Initially, we developed a general quality control method based on statistical assessment to detect the suspicious spectra. Subsequently, Radiative Transfer (RT) models were used to assess low light conditions, distortions in the spectral shape of above-water solar downwelling irradiance (<em>E<sub>S</sub></em>(λ), mW m<sup>−2</sup> nm<sup>−1</sup>) particularly those caused by intense atmospheric scattering and/or reddish hue of dusk or dawn radiation, the effect of atmospheric humidity and precipitation on the intensity and shape of spectra, and the influence of sun glint and surface perturbations on sky (<em>L<sub>S</sub></em>(λ), mW m<sup>−2</sup> nm<sup>−1</sup> sr<sup>−1</sup>) and water surface (<em>L<sub>T</sub></em>(λ), mW m<sup>−2</sup> nm<sup>−1</sup> sr<sup>−1</sup>) radiances. The proposed methods were applied to the entire archive of automated above-water hyperspectral measurements collected every ten minutes from 2020 to 2022 at the Royal Netherland Institute for Sea Research (NIOZ) at Jetty Station (NJS) located in the Marsdiep tidal inlet of the Duch Wadden Sea, the Netherlands. The findings demonstrate that low light conditions are characterized by <em>E<sub>S</sub></em>(λ)<sub>max</sub> ≤ 25 mW m<sup>−2</sup> nm<sup>−1</sup>. Red-shifted or distorted spectra are indicated by a ratio of <em>E<sub>S</sub></em>(4<!--> <!-->8<!--> <!-->0)/<em>E<sub>S</sub></em>(6<!--> <!-->8<!--> <!-->0) ≤ 1.0 and <em>E<sub>S</sub></em>(λ<sub>max</sub>)/ <em>E<sub>S</sub></em>(8<!--> <!-->6<!--> <!-->5) ≤ 1.25. High humidity/precipitation conditions are identified by the ratio of <em>E<sub>S</sub></em>(9<!--> <!-->4<!--> <!-->0)/<em>E<sub>S</sub></em>(8<!--> <!-->6<!--> <!-->5), which varies with the Solar Zenith Angle (SZA). Furthermore, significant sun glint and surface perturbations, such as whitecaps and foam, are indicated when the minimum ratio of <em>L<sub>T</sub></em>(800 nm-950 nm)/<em>E<sub>S</sub></em>(800 nm-950 nm) &gt; 0.025 sr<sup>−1</sup>, and the ratio of <em>L<sub>T</sub></em>(850 nm)/<em>E<sub>S</sub></em>(850 nm) ≥ 0.025 sr<sup>−1</sup>.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":null,"pages":null},"PeriodicalIF":10.6,"publicationDate":"2024-07-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141638753","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Global Streetscapes — A comprehensive dataset of 10 million street-level images across 688 cities for urban science and analytics 全球街景 - 包含 688 个城市 1 000 万张街道图像的综合数据集,用于城市科学和分析
IF 10.6 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2024-07-16 DOI: 10.1016/j.isprsjprs.2024.06.023

Street view imagery (SVI) is instrumental for sensing urban environments, benefitting numerous domains such as urban morphology, health, greenery, and accessibility. Billions of images worldwide have been made available by commercial services such as Google Street View and crowdsourcing services such as Mapillary and KartaView where anyone from anywhere can upload imagery while moving. However, while the data tend to be plentiful, have high coverage and quality, and are used to derive rich insights, they remain simple and limited in metadata as characteristics such as weather, quality, and lighting conditions remain unknown, making it difficult to evaluate the suitability of the images for specific analyses. We introduce Global Streetscapes — a dataset of 10 million crowdsourced and free-to-use SVIs sampled from 688 cities across 210 countries and territories, enriched with more than 300 camera, geographical, temporal, contextual, semantic, and perceptual attributes. The cities included are well balanced and diverse, and are home to about 10% of the world’s population. Deep learning models are trained on a subset of manually labelled images for eight visual-contextual attributes pertaining to the usability of SVI — panoramic status, lighting condition, view direction, weather, platform, quality, presence of glare and reflections, achieving accuracy ranging from 68.3% to 99.9%, and used to automatically label the entire dataset. Thanks to its scale and pre-computed standard semantic information, the data can be readily used to benefit existing use cases and to unlock new applications, including multi-city comparative studies and longitudinal analyses, as affirmed by a couple of use cases in the paper. Moreover, the automated processes and open-source code facilitate the expansion and updates of the dataset and encourage users to create their own datasets. With the rich manual annotations, some of which are provided for the first time, and diverse conditions present in the images, the dataset also facilitates assessing the heterogeneous properties of crowdsourced SVIs and provides a benchmark for evaluating future computer vision models. We make the Global Streetscapes dataset and the code to reproduce and use it publicly available in https://github.com/ualsg/global-streetscapes.

街景图像(SVI)在感知城市环境方面非常重要,对城市形态、健康、绿化和无障碍环境等众多领域都大有裨益。谷歌街景等商业服务以及 Mapillary 和 KartaView 等众包服务提供了全球数十亿幅图像,任何地方的任何人都可以在移动中上传图像。然而,尽管这些数据往往数量多、覆盖面广、质量高,并可用于获得丰富的洞察力,但由于天气、质量和光照条件等特征仍然未知,这些数据仍然很简单,元数据也很有限,因此很难评估图像是否适合进行具体分析。我们介绍了全球街景--一个由来自 210 个国家和地区的 688 个城市的 1000 万张众包和免费使用的 SVIs 组成的数据集,其中包含 300 多个相机、地理、时间、上下文、语义和感知属性。所包含的城市均衡多样,人口约占世界总人口的 10%。针对与 SVI 可用性相关的八个视觉上下文属性(全景状态、照明条件、视角方向、天气、平台、质量、是否存在眩光和反光),在人工标注的图像子集上训练了深度学习模型,准确率达到 68.3% 到 99.9%,并用于自动标注整个数据集。得益于其规模和预先计算的标准语义信息,该数据可随时用于现有的使用案例,并开启新的应用,包括多城市比较研究和纵向分析,论文中的几个使用案例证实了这一点。此外,自动化流程和开放源代码为数据集的扩展和更新提供了便利,并鼓励用户创建自己的数据集。该数据集具有丰富的手动注释(其中一些是首次提供)和图像中存在的各种条件,因此还有助于评估众包 SVI 的各种属性,并为评估未来的计算机视觉模型提供基准。我们在 https://github.com/ualsg/global-streetscapes 中公开了全球街景数据集以及复制和使用该数据集的代码。
{"title":"Global Streetscapes — A comprehensive dataset of 10 million street-level images across 688 cities for urban science and analytics","authors":"","doi":"10.1016/j.isprsjprs.2024.06.023","DOIUrl":"10.1016/j.isprsjprs.2024.06.023","url":null,"abstract":"<div><p>Street view imagery (SVI) is instrumental for sensing urban environments, benefitting numerous domains such as urban morphology, health, greenery, and accessibility. Billions of images worldwide have been made available by commercial services such as Google Street View and crowdsourcing services such as Mapillary and KartaView where anyone from anywhere can upload imagery while moving. However, while the data tend to be plentiful, have high coverage and quality, and are used to derive rich insights, they remain simple and limited in metadata as characteristics such as weather, quality, and lighting conditions remain unknown, making it difficult to evaluate the suitability of the images for specific analyses. We introduce Global Streetscapes — a dataset of 10 million crowdsourced and free-to-use SVIs sampled from 688 cities across 210 countries and territories, enriched with more than 300 camera, geographical, temporal, contextual, semantic, and perceptual attributes. The cities included are well balanced and diverse, and are home to about 10% of the world’s population. Deep learning models are trained on a subset of manually labelled images for eight visual-contextual attributes pertaining to the usability of SVI — panoramic status, lighting condition, view direction, weather, platform, quality, presence of glare and reflections, achieving accuracy ranging from 68.3% to 99.9%, and used to automatically label the entire dataset. Thanks to its scale and pre-computed standard semantic information, the data can be readily used to benefit existing use cases and to unlock new applications, including multi-city comparative studies and longitudinal analyses, as affirmed by a couple of use cases in the paper. Moreover, the automated processes and open-source code facilitate the expansion and updates of the dataset and encourage users to create their own datasets. With the rich manual annotations, some of which are provided for the first time, and diverse conditions present in the images, the dataset also facilitates assessing the heterogeneous properties of crowdsourced SVIs and provides a benchmark for evaluating future computer vision models. We make the Global Streetscapes dataset and the code to reproduce and use it publicly available in <span><span>https://github.com/ualsg/global-streetscapes</span><svg><path></path></svg></span>.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":null,"pages":null},"PeriodicalIF":10.6,"publicationDate":"2024-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141630085","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Explaining the decisions and the functioning of a convolutional spatiotemporal land cover classifier with channel attention and redescription mining 利用通道关注和再描述挖掘解释卷积时空土地覆被分类器的决策和功能
IF 10.6 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2024-07-16 DOI: 10.1016/j.isprsjprs.2024.06.021

Convolutional neural networks trained with satellite image time series have demonstrated their potential in land cover classification in recent years. Nevertheless, the rationale leading to their decisions remains obscure by nature. Methods for providing relevant and simplified explanations of their decisions as well as methods for understanding their inner functioning have thus emerged. However, both kinds of methods generally work separately and no explicit connection between their findings is made available. This paper presents an innovative method for refining the explanations provided by channel-based attention mechanisms. It consists in identifying correspondence rules between neuronal activation levels and the presence of spatiotemporal patterns in the input data for each channel and target class. These rules provide both class-level and instance-level explanations, as well as an explicit understanding of the network operations. They are extracted using a state-of-the-art redescription mining algorithm. Experiments on the Reunion Island Sentinel-2 dataset show that both correct and incorrect decisions can be explained using convenient spatiotemporal visualizations.

近年来,利用卫星图像时间序列训练的卷积神经网络在土地覆被分类方面展现出了巨大潜力。然而,其决策原理本质上仍然模糊不清。因此,出现了为其决策提供相关简化解释的方法,以及了解其内部运作的方法。然而,这两种方法一般都是各自为战,没有将它们的研究结果明确地联系起来。本文提出了一种创新方法,用于完善基于通道的注意力机制所提供的解释。它包括识别神经元激活水平与每个通道和目标类别的输入数据中存在的时空模式之间的对应规则。这些规则提供了类级和实例级解释,以及对网络运行的明确理解。这些规则是使用最先进的重新描述挖掘算法提取的。在留尼汪岛哨兵-2 数据集上进行的实验表明,正确和错误的决定都可以通过便捷的时空可视化来解释。
{"title":"Explaining the decisions and the functioning of a convolutional spatiotemporal land cover classifier with channel attention and redescription mining","authors":"","doi":"10.1016/j.isprsjprs.2024.06.021","DOIUrl":"10.1016/j.isprsjprs.2024.06.021","url":null,"abstract":"<div><p>Convolutional neural networks trained with satellite image time series have demonstrated their potential in land cover classification in recent years. Nevertheless, the rationale leading to their decisions remains obscure by nature. Methods for providing relevant and simplified explanations of their decisions as well as methods for understanding their inner functioning have thus emerged. However, both kinds of methods generally work separately and no explicit connection between their findings is made available. This paper presents an innovative method for refining the explanations provided by channel-based attention mechanisms. It consists in identifying correspondence rules between neuronal activation levels and the presence of spatiotemporal patterns in the input data for each channel and target class. These rules provide both class-level and instance-level explanations, as well as an explicit understanding of the network operations. They are extracted using a state-of-the-art redescription mining algorithm. Experiments on the Reunion Island Sentinel-2 dataset show that both correct and incorrect decisions can be explained using convenient spatiotemporal visualizations.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":null,"pages":null},"PeriodicalIF":10.6,"publicationDate":"2024-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0924271624002600/pdfft?md5=d29a61ba9ff8a461eed59ce59e443f04&pid=1-s2.0-S0924271624002600-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141630086","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
EMET: An emergence-based thermal phenological framework for near real-time crop type mapping EMET:用于近实时作物类型绘图的基于出现的热物候框架
IF 10.6 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2024-07-16 DOI: 10.1016/j.isprsjprs.2024.07.007

Near real-time (NRT) crop type mapping plays a crucial role in modeling crop development, managing food supply chains, and supporting sustainable agriculture. The low-latency updates on crop type distribution also help assess the impacts of weather extremes and climate change on agricultural production in a timely fashion, aiding in identification of early risks in food insecurity as well as rapid assessments of the damage. Yet NRT crop type mapping is challenging due to the obstacle in acquiring timely crop type reference labels during the current season for crop mapping model building. Meanwhile, the crop mapping models constructed with historical crop type labels and corresponding satellite imagery may not be applicable to the current season in NRT due to spatiotemporal variability of crop phenology. The difficulty in characterizing crop phenology in NRT remains a significant hurdle in NRT crop type mapping. To tackle these issues, a novel emergence-based thermal phenological framework (EMET) is proposed in this study for field-level NRT crop type mapping. The EMET framework comprises three key components: hybrid deep learning spatiotemporal image fusion, NRT thermal-based crop phenology normalization, and NRT crop type characterization. The hybrid fusion model integrates super-resolution convolutional neural network (SRCNN) and long short-term memory (LSTM) to generate daily satellite observations with a high spatial resolution in NRT. The NRT thermal-based crop phenology normalization innovatively synthesizes within-season crop emergence (WISE) model and thermal time accumulation throughout the growing season, to timely normalize crop phenological progress derived from temporally dense fusion imagery. The NRT normalized fusion time series are then fed into an advanced deep learning classifier, the self-attention based LSTM (SAtLSTM) model, to identify crop types. Results in Illinois and Minnesota of the U.S. Corn Belt suggest that the EMET framework significantly enhances the model scalability with crop phenology normalized in NRT for timely crop mapping. A consistently higher overall accuracy is yielded by the EMET framework throughout the growing season compared to the calendar-based and WISE-based benchmark scenarios. When transferred to different study sites and testing years, EMET maintains an advantage of over 5% in overall accuracy during early- to mid-season. Moreover, EMET reaches an overall accuracy of 85% a month earlier than the benchmarks, and it can accurately characterize crop types with an overall accuracy of 90% as early as in late July. F1 scores for both corn and soybeans also achieve 90% around late July. The EMET framework paves the way for large-scale satellite-based NRT crop type mapping at the field level, which can largely help reduce food market volatility to enhance food security, as well as benefit a variety of agricultural applications to optimize crop management towards more sustainable agricultural prod

近实时(NRT)作物类型绘图在模拟作物发展、管理粮食供应链和支持可持续农业方面发挥着至关重要的作用。作物类型分布的低时延更新还有助于及时评估极端天气和气候变化对农业生产的影响,帮助识别粮食不安全的早期风险并快速评估损失。然而,由于在当季及时获取作物类型参考标签以建立作物绘图模型存在障碍,因此 NRT 作物类型绘图具有挑战性。同时,由于作物物候的时空变异性,利用历史作物类型标签和相应卫星图像构建的作物测绘模型可能并不适用于北热带地区的当前季节。难以确定北热带地区作物物候特征仍然是北热带地区作物类型绘图的一个重大障碍。为解决这些问题,本研究提出了一种新颖的基于出现的热物候框架(EMET),用于田间水平的 NRT 作物类型测绘。EMET 框架由三个关键部分组成:混合深度学习时空图像融合、基于 NRT 热的作物物候归一化和 NRT 作物类型表征。混合融合模型集成了超分辨率卷积神经网络(SRCNN)和长短期记忆(LSTM),以生成具有高空间分辨率的 NRT 每日卫星观测数据。基于 NRT 热的作物物候归一化创新性地综合了整个生长季节的季内作物出苗(WISE)模型和热时间累积,及时归一化从时间密集的融合图像中得出的作物物候进展。然后将 NRT 归一化融合时间序列输入先进的深度学习分类器--基于自我注意的 LSTM(SAtLSTM)模型,以识别作物类型。在美国玉米带伊利诺伊州和明尼苏达州的研究结果表明,EMET 框架显著增强了模型的可扩展性,通过在 NRT 中对作物物候进行归一化,可及时绘制作物图。与基于日历的基准方案和基于 WISE 的基准方案相比,EMET 框架在整个生长季节的总体精度始终较高。在不同的研究地点和测试年份中,EMET 在生长季初至中期的总体精度上保持了 5% 以上的优势。此外,EMET 的总体准确率比基准方案早一个月达到 85%,早在七月下旬就能准确描述作物类型,总体准确率达到 90%。玉米和大豆的 F1 分数也在 7 月下旬左右达到 90%。EMET 框架为基于卫星的大规模田间 NRT 作物类型测绘铺平了道路,这在很大程度上有助于减少粮食市场的波动,从而提高粮食安全,并有利于各种农业应用,优化作物管理,实现更可持续的农业生产。
{"title":"EMET: An emergence-based thermal phenological framework for near real-time crop type mapping","authors":"","doi":"10.1016/j.isprsjprs.2024.07.007","DOIUrl":"10.1016/j.isprsjprs.2024.07.007","url":null,"abstract":"<div><p>Near real-time (NRT) crop type mapping plays a crucial role in modeling crop development, managing food supply chains, and supporting sustainable agriculture. The low-latency updates on crop type distribution also help assess the impacts of weather extremes and climate change on agricultural production in a timely fashion, aiding in identification of early risks in food insecurity as well as rapid assessments of the damage. Yet NRT crop type mapping is challenging due to the obstacle in acquiring timely crop type reference labels during the current season for crop mapping model building. Meanwhile, the crop mapping models constructed with historical crop type labels and corresponding satellite imagery may not be applicable to the current season in NRT due to spatiotemporal variability of crop phenology. The difficulty in characterizing crop phenology in NRT remains a significant hurdle in NRT crop type mapping. To tackle these issues, a novel emergence-based thermal phenological framework (EMET) is proposed in this study for field-level NRT crop type mapping. The EMET framework comprises three key components: hybrid deep learning spatiotemporal image fusion, NRT thermal-based crop phenology normalization, and NRT crop type characterization. The hybrid fusion model integrates super-resolution convolutional neural network (SRCNN) and long short-term memory (LSTM) to generate daily satellite observations with a high spatial resolution in NRT. The NRT thermal-based crop phenology normalization innovatively synthesizes within-season crop emergence (WISE) model and thermal time accumulation throughout the growing season, to timely normalize crop phenological progress derived from temporally dense fusion imagery. The NRT normalized fusion time series are then fed into an advanced deep learning classifier, the self-attention based LSTM (SAtLSTM) model, to identify crop types. Results in Illinois and Minnesota of the U.S. Corn Belt suggest that the EMET framework significantly enhances the model scalability with crop phenology normalized in NRT for timely crop mapping. A consistently higher overall accuracy is yielded by the EMET framework throughout the growing season compared to the calendar-based and WISE-based benchmark scenarios. When transferred to different study sites and testing years, EMET maintains an advantage of over 5% in overall accuracy during early- to mid-season. Moreover, EMET reaches an overall accuracy of 85% a month earlier than the benchmarks, and it can accurately characterize crop types with an overall accuracy of 90% as early as in late July. F1 scores for both corn and soybeans also achieve 90% around late July. The EMET framework paves the way for large-scale satellite-based NRT crop type mapping at the field level, which can largely help reduce food market volatility to enhance food security, as well as benefit a variety of agricultural applications to optimize crop management towards more sustainable agricultural prod","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":null,"pages":null},"PeriodicalIF":10.6,"publicationDate":"2024-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0924271624002740/pdfft?md5=c768127877f879d4fb5bfe0382468a9f&pid=1-s2.0-S0924271624002740-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141630088","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Unifying remote sensing change detection via deep probabilistic change models: From principles, models to applications 通过深度概率变化模型统一遥感变化探测:从原理、模型到应用
IF 10.6 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2024-07-16 DOI: 10.1016/j.isprsjprs.2024.07.001

Change detection in high-resolution Earth observation is a fundamental Earth vision task to understand the subtle temporal dynamics of Earth’s surface, significantly promoted by generic vision technologies in recent years. Vision Transformer is a powerful component to learning spatiotemporal representation but with enormous computation complexity, especially for high-resolution images. Besides, there is still lacking principles in designing macro architectures integrating these advanced vision components for various change detection tasks. In this paper, we present a deep probabilistic change model (DPCM) to provide a unified, interpretable, modular probabilistic change process modeling to address multiple change detection tasks, including binary change detection, one-to-many semantic change detection, and many-to-many semantic change detection. DPCM describes any complex change process as a probabilistic graphical model to provide theoretical evidence for macro architecture design and generic change detection task modeling. We refer to this probabilistic graphical model as the probabilistic change model (PCM), where DPCM is the PCM parameterized by deep neural networks. For parameterization, the PCM is factorized into many easy-to-solve distributions based on task-specific assumptions, and then we can use deep neural modules to parameterize these distributions to solve the change detection problem uniformly. In this way, DPCM has both theoretical macro architecture from PCM and strong representation capability of deep networks. We also present the sparse change Transformer for better parameterization. Inspired by domain knowledge, i.e., the sparsity of change and the local correlation of change, the sparse change Transformer computes self-attention within change regions to model spatiotemporal correlations, which has a quadratic computational complexity of the change region size but independent of image size, significantly reducing computation overhead for high-resolution image change detection. We refer to this instance of DPCM with sparse change Transformer as ChangeSparse to demonstrate their effectiveness. The experiments confirm ChangeSparse’s superiority in speed and accuracy for multiple real-world application scenarios, such as disaster response and urban development monitoring. The code is available at https://github.com/Z-Zheng/pytorch-change-models. More resources can be found in http://rsidea.whu.edu.cn/resource_sharing.htm.

高分辨率地球观测中的变化检测是了解地球表面微妙时间动态的一项基本地球视觉任务,近年来通用视觉技术的发展极大地促进了这项任务的完成。视觉变换器是学习时空表征的强大组件,但其计算复杂度极高,尤其是对于高分辨率图像。此外,在为各种变化检测任务设计集成了这些高级视觉组件的宏架构时,仍缺乏相应的原则。在本文中,我们提出了一种深度概率变化模型(DPCM),以提供一种统一、可解释、模块化的概率变化过程建模,从而解决多种变化检测任务,包括二进制变化检测、一对多语义变化检测和多对多语义变化检测。DPCM 将任何复杂的变化过程描述为概率图形模型,为宏观架构设计和通用变化检测任务建模提供理论依据。我们将这种概率图形模型称为概率变化模型(PCM),其中 DPCM 是由深度神经网络参数化的 PCM。在参数化过程中,我们会根据特定任务的假设将 PCM 因子化为许多易于求解的分布,然后利用深度神经模块对这些分布进行参数化,从而均匀地求解变化检测问题。这样,DPCM 既有 PCM 的理论宏观架构,又有深度网络的强大表示能力。为了更好地进行参数化,我们还提出了稀疏变化变换器。稀疏变化变换器受领域知识(即变化的稀疏性和变化的局部相关性)的启发,在变化区域内计算自注意,以模拟时空相关性,其计算复杂度为变化区域大小的二次方,但与图像大小无关,从而显著降低了高分辨率图像变化检测的计算开销。我们将这种带有稀疏变化变换器的 DPCM 实例称为 ChangeSparse,以证明其有效性。实验证实,ChangeSparse 在多个实际应用场景(如灾难响应和城市发展监测)中都具有速度和准确性方面的优势。代码可在 https://github.com/Z-Zheng/pytorch-change-models 上获取。更多资源请访问 http://rsidea.whu.edu.cn/resource_sharing.htm。
{"title":"Unifying remote sensing change detection via deep probabilistic change models: From principles, models to applications","authors":"","doi":"10.1016/j.isprsjprs.2024.07.001","DOIUrl":"10.1016/j.isprsjprs.2024.07.001","url":null,"abstract":"<div><p>Change detection in high-resolution Earth observation is a fundamental Earth vision task to understand the subtle temporal dynamics of Earth’s surface, significantly promoted by generic vision technologies in recent years. Vision Transformer is a powerful component to learning spatiotemporal representation but with enormous computation complexity, especially for high-resolution images. Besides, there is still lacking principles in designing macro architectures integrating these advanced vision components for various change detection tasks. In this paper, we present a deep probabilistic change model (DPCM) to provide a unified, interpretable, modular probabilistic change process modeling to address multiple change detection tasks, including binary change detection, one-to-many semantic change detection, and many-to-many semantic change detection. DPCM describes any complex change process as a probabilistic graphical model to provide theoretical evidence for macro architecture design and generic change detection task modeling. We refer to this probabilistic graphical model as the probabilistic change model (PCM), where DPCM is the PCM parameterized by deep neural networks. For parameterization, the PCM is factorized into many easy-to-solve distributions based on task-specific assumptions, and then we can use deep neural modules to parameterize these distributions to solve the change detection problem uniformly. In this way, DPCM has both theoretical macro architecture from PCM and strong representation capability of deep networks. We also present the sparse change Transformer for better parameterization. Inspired by domain knowledge, i.e., the sparsity of change and the local correlation of change, the sparse change Transformer computes self-attention within change regions to model spatiotemporal correlations, which has a quadratic computational complexity of the change region size but independent of image size, significantly reducing computation overhead for high-resolution image change detection. We refer to this instance of DPCM with sparse change Transformer as ChangeSparse to demonstrate their effectiveness. The experiments confirm ChangeSparse’s superiority in speed and accuracy for multiple real-world application scenarios, such as disaster response and urban development monitoring. The code is available at <span><span>https://github.com/Z-Zheng/pytorch-change-models</span><svg><path></path></svg></span>. More resources can be found in <span><span>http://rsidea.whu.edu.cn/resource_sharing.htm</span><svg><path></path></svg></span>.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":null,"pages":null},"PeriodicalIF":10.6,"publicationDate":"2024-07-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141630089","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Navigating the publishing landscape in times of revolutionary changes 在翻天覆地的变化中驾驭出版格局
IF 10.6 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2024-07-14 DOI: 10.1016/j.isprsjprs.2024.07.008
Qihao Weng, Clement Mallet
{"title":"Navigating the publishing landscape in times of revolutionary changes","authors":"Qihao Weng,&nbsp;Clement Mallet","doi":"10.1016/j.isprsjprs.2024.07.008","DOIUrl":"https://doi.org/10.1016/j.isprsjprs.2024.07.008","url":null,"abstract":"","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":null,"pages":null},"PeriodicalIF":10.6,"publicationDate":"2024-07-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141605950","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep learning with simulated laser scanning data for 3D point cloud classification 利用模拟激光扫描数据进行三维点云分类的深度学习
IF 10.6 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2024-07-13 DOI: 10.1016/j.isprsjprs.2024.06.018
Alberto M. Esmorís , Hannah Weiser , Lukas Winiwarter , Jose C. Cabaleiro , Bernhard Höfle

Laser scanning is an active remote sensing technique applied in many disciplines to acquire state-of-the-art spatial measurements. Semantic labeling is often necessary to extract information from the raw point cloud. Deep learning methods constitute a data-hungry solution for the semantic segmentation of point clouds. In this work, we investigate the use of simulated laser scanning for training deep learning models, which are applied to real data subsequently. We show that training a deep learning model purely on virtual laser scanning data can produce results comparable to models trained on real data when evaluated on real data. For leaf-wood segmentation of trees, using the KPConv model trained with virtual data achieves 93.7% overall accuracy, while the model trained on real data reaches 94.7% overall accuracy. In urban contexts, a KPConv model trained on virtual data achieves 74.1% overall accuracy on real validation data, while the model trained on real data achieves 82.4%. Our models outperform the state-of-the-art model FSCT in terms of generalization to unseen real data as well as a baseline model trained on points randomly sampled from the tree mesh surface. From our results, we conclude that the combination of laser scanning simulation and deep learning is a cost-effective alternative to real data acquisition and manual labeling in the domain of geospatial point cloud analysis. The strengths of this approach are that (a) a large amount of diverse laser scanning training data can be generated quickly and without the need for expensive equipment, (b) the simulation configurations can be adapted so that the virtual training data have similar characteristics to the targeted real data, and (c) the whole workflow can be automated through procedural scene generation.

激光扫描是一种主动遥感技术,应用于许多学科,以获取最先进的空间测量数据。要从原始点云中提取信息,通常需要进行语义标注。深度学习方法是一种对数据要求极高的点云语义分割解决方案。在这项工作中,我们研究了如何利用模拟激光扫描来训练深度学习模型,并随后将其应用于真实数据。我们的研究表明,在真实数据上进行评估时,纯粹在虚拟激光扫描数据上训练深度学习模型所产生的结果可与在真实数据上训练的模型相媲美。在树木的叶木分割方面,使用虚拟数据训练的 KPConv 模型达到了 93.7% 的总体准确率,而使用真实数据训练的模型则达到了 94.7% 的总体准确率。在城市环境中,使用虚拟数据训练的 KPConv 模型在真实验证数据上的总体准确率为 74.1%,而使用真实数据训练的模型则达到了 82.4%。在对未见真实数据的泛化方面,我们的模型优于最先进的模型 FSCT,也优于根据从树网格表面随机取样的点训练的基线模型。根据我们的研究结果,我们得出结论:在地理空间点云分析领域,激光扫描模拟与深度学习的结合是一种经济有效的方法,可以替代真实数据采集和人工标注。这种方法的优势在于:(a) 可以快速生成大量不同的激光扫描训练数据,无需昂贵的设备;(b) 可以调整模拟配置,使虚拟训练数据具有与目标真实数据相似的特征;(c) 可以通过程序化场景生成实现整个工作流程的自动化。
{"title":"Deep learning with simulated laser scanning data for 3D point cloud classification","authors":"Alberto M. Esmorís ,&nbsp;Hannah Weiser ,&nbsp;Lukas Winiwarter ,&nbsp;Jose C. Cabaleiro ,&nbsp;Bernhard Höfle","doi":"10.1016/j.isprsjprs.2024.06.018","DOIUrl":"https://doi.org/10.1016/j.isprsjprs.2024.06.018","url":null,"abstract":"<div><p>Laser scanning is an active remote sensing technique applied in many disciplines to acquire state-of-the-art spatial measurements. Semantic labeling is often necessary to extract information from the raw point cloud. Deep learning methods constitute a data-hungry solution for the semantic segmentation of point clouds. In this work, we investigate the use of simulated laser scanning for training deep learning models, which are applied to real data subsequently. We show that training a deep learning model purely on virtual laser scanning data can produce results comparable to models trained on real data when evaluated on real data. For leaf-wood segmentation of trees, using the KPConv model trained with virtual data achieves 93.7% overall accuracy, while the model trained on real data reaches 94.7% overall accuracy. In urban contexts, a KPConv model trained on virtual data achieves 74.1% overall accuracy on real validation data, while the model trained on real data achieves 82.4%. Our models outperform the state-of-the-art model FSCT in terms of generalization to unseen real data as well as a baseline model trained on points randomly sampled from the tree mesh surface. From our results, we conclude that the combination of laser scanning simulation and deep learning is a cost-effective alternative to real data acquisition and manual labeling in the domain of geospatial point cloud analysis. The strengths of this approach are that (a) a large amount of diverse laser scanning training data can be generated quickly and without the need for expensive equipment, (b) the simulation configurations can be adapted so that the virtual training data have similar characteristics to the targeted real data, and (c) the whole workflow can be automated through procedural scene generation.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":null,"pages":null},"PeriodicalIF":10.6,"publicationDate":"2024-07-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0924271624002569/pdfft?md5=85e93537aadaaaa88cbcdae2d9791d9d&pid=1-s2.0-S0924271624002569-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141605951","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Pano2Geo: An efficient and robust building height estimation model using street-view panoramas Pano2Geo:使用街景全景图的高效稳健建筑高度估算模型
IF 10.6 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2024-07-11 DOI: 10.1016/j.isprsjprs.2024.07.005
Kaixuan Fan , Anqi Lin , Hao Wu , Zhenci Xu

Building height serves as a crucial parameter in characterizing urban vertical structure, which has a profound impact on urban sustainable development. The emergence of street-view data offers the opportunity to observe urban 3D scenarios from the human perspective, benefiting the estimation of building height. In this paper, we propose an efficient and robust building height estimation model, which we call the Pano2Geo model, by precisely projecting street-view panorama (SVP) coordinates to geospatial coordinates. Firstly, an SVP refinement stratagem is designed, incorporating NENO rules for observation quality assessment from four aspects: number of buildings, extent of the buildings, number of nodes, and orthogonal observations, followed by the application of the art gallery theorem to further refine the SVPs. Secondly, the Pano2Geo model is constructed, which provides a pixel-level projection transformation from SVP coordinates to 3D geospatial coordinates for locating the height features of buildings in the SVP. Finally, the valid building height feature points in the SVP are extracted based on a slope mutation test, and the 3D geospatial coordinates of the building height feature points are projected using the Pano2Geo model, so as to obtain the building height. The proposed model was evaluated in the city of Wuhan in China, and the results indicate that the Pano2Geo model can accurately estimate building height, with an average error of 1.85 m. Furthermore, compared with three state-of-the-art methods, the Pano2Geo model shows superior performance, with only 10.2 % of buildings have absolute errors exceeding 2 m, compared to the Map-image-based (27.2 %), Corner-based (16.8 %), and Single-view-based (13.9 %) height estimation methods. The SVP refinement method achieves optimal observation quality with less than 50 % of existing SVPs, leading to highly efficient building height estimation, particularly in areas of a high building density. Moreover, the Pano2Geo model exhibits robustness in building height estimation, maintaining errors within 2 m even as building shape complexity and occlusion degree increase within the SVP. Our source dataset and code are available at https://github.com/Giser317/Pano2Geo.git.

建筑高度是表征城市垂直结构的关键参数,对城市可持续发展有着深远的影响。街景数据的出现为从人类视角观察城市三维场景提供了机会,有利于建筑高度的估算。本文通过将街景全景(SVP)坐标精确投影到地理空间坐标,提出了一种高效、稳健的建筑高度估算模型,我们称之为 Pano2Geo 模型。首先,我们设计了一种 SVP 精化策略,从建筑物数量、建筑物范围、节点数量和正交观测四个方面结合 NENO 观察质量评估规则,然后应用艺术画廊定理进一步精化 SVP。其次,构建 Pano2Geo 模型,提供从 SVP 坐标到三维地理空间坐标的像素级投影转换,用于定位 SVP 中的建筑物高度特征。最后,根据斜率突变测试提取 SVP 中有效的建筑物高度特征点,并利用 Pano2Geo 模型对建筑物高度特征点的三维地理空间坐标进行投影,从而获得建筑物高度。在中国武汉市对所提出的模型进行了评估,结果表明 Pano2Geo 模型能够准确估计建筑高度,平均误差为 1.85 米。此外,与三种最先进的方法相比,Pano2Geo 模型表现出更优越的性能,与基于地图影像(27.2%)、基于拐角(16.8%)和基于单视角(13.9%)的高度估计方法相比,只有 10.2% 的建筑绝对误差超过 2 米。SVP 精化方法以不到 50% 的现有 SVP 实现了最佳观测质量,从而实现了高效的建筑高度估算,尤其是在建筑密度较高的地区。此外,Pano2Geo 模型在建筑高度估算方面表现出很强的鲁棒性,即使在 SVP 中建筑形状复杂度和遮挡程度增加的情况下,误差也能保持在 2 米以内。我们的源数据集和代码可在以下网址获取。
{"title":"Pano2Geo: An efficient and robust building height estimation model using street-view panoramas","authors":"Kaixuan Fan ,&nbsp;Anqi Lin ,&nbsp;Hao Wu ,&nbsp;Zhenci Xu","doi":"10.1016/j.isprsjprs.2024.07.005","DOIUrl":"10.1016/j.isprsjprs.2024.07.005","url":null,"abstract":"<div><p>Building height serves as a crucial parameter in characterizing urban vertical structure, which has a profound impact on urban sustainable development. The emergence of street-view data offers the opportunity to observe urban 3D scenarios from the human perspective, benefiting the estimation of building height. In this paper, we propose an efficient and robust building height estimation model, which we call the Pano2Geo model, by precisely projecting street-view panorama (SVP) coordinates to geospatial coordinates. Firstly, an SVP refinement stratagem is designed, incorporating NENO rules for observation quality assessment from four aspects: number of buildings, extent of the buildings, number of nodes, and orthogonal observations, followed by the application of the art gallery theorem to further refine the SVPs. Secondly, the Pano2Geo model is constructed, which provides a pixel-level projection transformation from SVP coordinates to 3D geospatial coordinates for locating the height features of buildings in the SVP. Finally, the valid building height feature points in the SVP are extracted based on a slope mutation test, and the 3D geospatial coordinates of the building height feature points are projected using the Pano2Geo model, so as to obtain the building height. The proposed model was evaluated in the city of Wuhan in China, and the results indicate that the Pano2Geo model can accurately estimate building height, with an average error of 1.85 m. Furthermore, compared with three state-of-the-art methods, the Pano2Geo model shows superior performance, with only 10.2 % of buildings have absolute errors exceeding 2 m, compared to the Map-image-based (27.2 %), Corner-based (16.8 %), and Single-view-based (13.9 %) height estimation methods. The SVP refinement method achieves optimal observation quality with less than 50 % of existing SVPs, leading to highly efficient building height estimation, particularly in areas of a high building density. Moreover, the Pano2Geo model exhibits robustness in building height estimation, maintaining errors within 2 m even as building shape complexity and occlusion degree increase within the SVP. Our source dataset and code are available at <span>https://github.com/Giser317/Pano2Geo.git</span><svg><path></path></svg>.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":null,"pages":null},"PeriodicalIF":10.6,"publicationDate":"2024-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0924271624002727/pdfft?md5=6ada921f9199fbef476b652be642f004&pid=1-s2.0-S0924271624002727-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141588660","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep neural network based on dynamic attention and layer attention for meteorological data downscaling 用于气象数据降尺度的基于动态关注和层关注的深度神经网络
IF 10.6 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2024-07-10 DOI: 10.1016/j.isprsjprs.2024.06.020
Junkai Wang, Lianlei Lin, Zongwei Zhang, Sheng Gao, Hangyi Yu

The scale of meteorological data products does not match the requirements of application scenarios, which limits their application. It is suggested that large-scale reanalysis data must be downscaled before use. Attention mechanism is the key to high-performance downscaling models. However, in different application scenarios and different locations on the network, the attention mechanism is not always beneficial. In this paper, we propose a dynamic attention module that can adaptively generate weights for each branch based on input features, thereby dynamically suppressing unnecessary attention adjustments. At the same time, we propose a layer attention module, which can independently and adaptively aggregate the feature representation of different network layers. In addition, we design a unique loss function based on homoscedasticity uncertainty, which can directly guide the model to learn the numerical mapping relationship from low resolution to high resolution at the pixel level, and implicitly motivate the model to better reconstruct the data distribution of each meteorological field by guiding the model to learn the distribution difference between different meteorological fields. Experiments show that our model is more robust in time dimension, with an MAE average reduction of about 40% compared to VDSR and other methods in downscaling composite meteorological data. It can more accurately reconstruct multivariate high-resolution meteorological fields. Codes available at https://github.com/HitKarry/SDDN.

气象数据产品的尺度与应用场景的要求不符,限制了其应用。建议大尺度再分析数据在使用前必须进行降尺度处理。注意机制是高性能降尺度模式的关键。然而,在不同的应用场景和网络的不同位置,关注机制并不总是有利的。在本文中,我们提出了一种动态注意力模块,它可以根据输入特征自适应地为每个分支生成权重,从而动态地抑制不必要的注意力调整。同时,我们还提出了层关注模块,它可以独立、自适应地聚合不同网络层的特征表示。此外,我们还设计了基于同方差不确定性的独特损失函数,可直接引导模型学习像素级从低分辨率到高分辨率的数值映射关系,并通过引导模型学习不同气象场之间的分布差异,隐含地激励模型更好地重建各气象场的数据分布。实验表明,我们的模型在时间维度上更具鲁棒性,在降尺度处理复合气象数据时,与 VDSR 和其他方法相比,平均 MAE 降低了约 40%。它能更准确地重建多变量高分辨率气象场。可在以下网址获取代码
{"title":"Deep neural network based on dynamic attention and layer attention for meteorological data downscaling","authors":"Junkai Wang,&nbsp;Lianlei Lin,&nbsp;Zongwei Zhang,&nbsp;Sheng Gao,&nbsp;Hangyi Yu","doi":"10.1016/j.isprsjprs.2024.06.020","DOIUrl":"10.1016/j.isprsjprs.2024.06.020","url":null,"abstract":"<div><p>The scale of meteorological data products does not match the requirements of application scenarios, which limits their application. It is suggested that large-scale reanalysis data must be downscaled before use. Attention mechanism is the key to high-performance downscaling models. However, in different application scenarios and different locations on the network, the attention mechanism is not always beneficial. In this paper, we propose a dynamic attention module that can adaptively generate weights for each branch based on input features, thereby dynamically suppressing unnecessary attention adjustments. At the same time, we propose a layer attention module, which can independently and adaptively aggregate the feature representation of different network layers. In addition, we design a unique loss function based on homoscedasticity uncertainty, which can directly guide the model to learn the numerical mapping relationship from low resolution to high resolution at the pixel level, and implicitly motivate the model to better reconstruct the data distribution of each meteorological field by guiding the model to learn the distribution difference between different meteorological fields. Experiments show that our model is more robust in time dimension, with an MAE average reduction of about 40% compared to VDSR and other methods in downscaling composite meteorological data. It can more accurately reconstruct multivariate high-resolution meteorological fields. Codes available at <span>https://github.com/HitKarry/SDDN</span><svg><path></path></svg>.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":null,"pages":null},"PeriodicalIF":10.6,"publicationDate":"2024-07-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141588725","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Remote sensing vegetation Indices-Driven models for sugarcane evapotranspiration estimation in the semiarid Ethiopian Rift Valley 用于埃塞俄比亚裂谷半干旱地区甘蔗蒸散量估算的遥感植被指数驱动模型
IF 10.6 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2024-07-08 DOI: 10.1016/j.isprsjprs.2024.07.004
Gezahegn W. Woldemariam , Berhan Gessesse Awoke , Raian Vargas Maretto

Evapotranspiration (ET), which represents water loss due to soil evaporation and crop transpiration, is a critical hydrological parameter for managing available water resources in irrigation systems. Traditional methods for monitoring actual evapotranspiration (ETa) involve field measurements. While accurate, they lack scalability, are labor-intensive, and incur high costs. Remote sensing satellites can help address these practical challenges by providing high-resolution imagery for spatially explicit mapping and near-real-time monitoring of ETa. This study aimed to develop simple yet robust models for estimating ETa using Sentinel-2 (S2A and S2B) satellite vegetation indices (VIs)—the normalized difference vegetation index (NDVI) and the enhanced vegetation index (EVI)—and the Google Earth Engine (GEE) cloud platform for irrigated sugarcane plantations of the Metehara Sugarcane Estate in the semiarid landscape of the Ethiopian Rift Valley. Six empirical ET-VI models that combined NDVI-based proxies (NDVIKc, NDVI*, and NDVI*scaled) and EVI-based proxies (EVIKc, EVI*, and EVI*scaled) for the crop coefficient (Kc) with the reference ET (ETo) were developed and evaluated for growing seasons between 2020 and 2022. Model validation using independently estimated sugarcane ET (ETsugarcane) and open-access remote sensing ET, Actual EvapoTranspiration and Interception (ETIa) showed that all ET-VI models captured spatiotemporal dynamics in the consumptive fraction of sugarcane water use, with a higher coefficient of determination (R2) of ≥ 0.91. However, comparative analyses of ETa retrieval models indicated improved accuracy of the ET-EVI models (root mean square error (RMSE) of ± 8 mm for ETsugarcane and ± 4 mm for ETIa) compared with the ET-NDVI models. Among the EVI models, ET-EVIKc achieved the highest R2 of 0.98, RMSE of ≤ 30 mm, and percentage bias (PBIAS) of ≤ 15 %. The results also revealed a strong correlation between the scaled VI-derived models and the reference ETIa (R2 = 0.94–0.97), which best explained the field-by-field variability, with the ET-EVI*scaled model achieving a lower RMSE of 18 mm than the ET-NDVI*scaled model (RMSE= 32 mm), while both the models showed similar levels of bias (∼17 %). Moreover, compared to the referenced ETsugarcane, the bias was minimal at − 9 % for ET-NDVI*scaled and − 1 % for ET-EVI*scaled. At the field scale, the NDVI and EVI models estimated the mean monthly ETa ranging from 99 to 129 mm m−1 and 89 to 148 mm m−1, respectively, with total annual averages of 1188–1537 mm yr−1 and 1296–1566 mm yr−1. In this context, the modeled ETa provided improved insights into consumptive water use in irrigated sugarcane plantations with limited field measurements. The statistical model evaluation metrics indic

蒸散量(ET)代表土壤蒸发和作物蒸腾造成的水分损失,是灌溉系统可用水资源管理的关键水文参数。监测实际蒸散量(ETa)的传统方法包括实地测量。这些方法虽然精确,但缺乏可扩展性、劳动密集型且成本高昂。遥感卫星可提供高分辨率图像,用于绘制空间清晰地图和近实时监测蒸散量,有助于解决这些实际挑战。本研究旨在利用哨兵-2(S2A 和 S2B)卫星植被指数(VI)--归一化差异植被指数(NDVI)和增强植被指数(EVI)--以及谷歌地球引擎(GEE)云平台,为埃塞俄比亚大裂谷半干旱地区梅特哈拉甘蔗园的灌溉甘蔗种植园开发简单而稳健的蒸散发估算模型。针对 2020 年至 2022 年的生长季节,开发并评估了六个经验性蒸散发-VI 模型,这些模型结合了基于 NDVI 的代用指标(NDVI、NDVI* 和 NDVI*)和基于 EVI 的代用指标(EVI、EVI* 和 EVI*),用于计算作物系数(Kc)和参考蒸散发(ETo)。利用独立估算的甘蔗蒸散发(ET)和开放获取的遥感蒸散发、实际蒸散和截流(ETIa)对模型进行了验证,结果表明,所有 ET-VI 模型都捕捉到了甘蔗用水消耗部分的时空动态,其判定系数(R)≥ 0.91。然而,对 ETa 检索模型的比较分析表明,与 ET-NDVI 模型相比,ET-EVI 模型的精度更高(ET 的均方根误差为 ± 8 毫米,ETIa 的均方根误差为 ± 4 毫米)。在 EVI 模型中,ET-EVI 的 R 值最高,为 0.98,RMSE ≤ 30 mm,偏差百分比 (PBIAS) ≤ 15 %。结果还显示,缩放 VI 导出模型与参考 ETIa(R = 0.94-0.97)之间具有很强的相关性,能最好地解释各田块的变异性,ET-EVI* 模型的均方根误差(RMSE)为 18 毫米,低于 ET-NDVI* 模型(RMSE= 32 毫米),而两个模型的偏差水平相似(∼17 %)。此外,与参考 ET 相比,ET-NDVI* 的偏差最小,为 - 9 %,ET-EVI* 为 - 1 %。在实地尺度上,NDVI 和 EVI 模型估计的月平均蒸散发分别为 99 至 129 毫米 m 和 89 至 148 毫米 m,年平均总量分别为 1188 至 1537 毫米 / 年和 1296 至 1566 毫米 / 年。在这种情况下,模拟的蒸散发为了解灌溉甘蔗种植园的消耗性用水提供了更深入的见解,但实地测量数据有限。统计模型评估指标表明,ET-EVI 是表征蒸散发的最佳模型,优于排名第二的 ET-NDVI 和 ET-EVI* 模型,后者优于 6%,而排名第三的 ET-NDVI* 优于 20%。我们的研究结果表明了多光谱 VI 驱动模型作为快速估算和绘制 ETa 的经济实用工具的潜力,从而支持了可持续节水实践的发展。本研究提出的经验建模框架的一个主要优点是,利用遥感 VI 和当地气象站数据,可直接对空间一致的 Kc 分布进行参数化。不过,在半干旱地区其他大型灌溉系统的耕地中进一步改进和实际应用基于标准化 VI 的蒸散发模型时,应考虑大气影响、场景特征变化和裸露地面/土壤暴露等因素。
{"title":"Remote sensing vegetation Indices-Driven models for sugarcane evapotranspiration estimation in the semiarid Ethiopian Rift Valley","authors":"Gezahegn W. Woldemariam ,&nbsp;Berhan Gessesse Awoke ,&nbsp;Raian Vargas Maretto","doi":"10.1016/j.isprsjprs.2024.07.004","DOIUrl":"10.1016/j.isprsjprs.2024.07.004","url":null,"abstract":"<div><p>Evapotranspiration (ET), which represents water loss due to soil evaporation and crop transpiration, is a critical hydrological parameter for managing available water resources in irrigation systems. Traditional methods for monitoring actual evapotranspiration (ETa) involve field measurements. While accurate, they lack scalability, are labor-intensive, and incur high costs. Remote sensing satellites can help address these practical challenges by providing high-resolution imagery for spatially explicit mapping and near-real-time monitoring of ETa. This study aimed to develop simple yet robust models for estimating ETa using Sentinel-2 (S2A and S2B) satellite vegetation indices (VIs)—the normalized difference vegetation index (NDVI) and the enhanced vegetation index (EVI)—and the Google Earth Engine (GEE) cloud platform for irrigated sugarcane plantations of the Metehara Sugarcane Estate in the semiarid landscape of the Ethiopian Rift Valley. Six empirical ET-VI models that combined NDVI-based proxies (NDVI<sub>Kc</sub>, NDVI*, and NDVI*<sub>scaled</sub>) and EVI-based proxies (EVI<sub>Kc</sub>, EVI*, and EVI*<sub>scaled</sub>) for the crop coefficient (Kc) with the reference ET (ETo) were developed and evaluated for growing seasons between 2020 and 2022. Model validation using independently estimated sugarcane ET (ET<sub>sugarcane</sub>) and open-access remote sensing ET, Actual EvapoTranspiration and Interception (ETIa) showed that all ET-VI models captured spatiotemporal dynamics in the consumptive fraction of sugarcane water use, with a higher coefficient of determination (R<sup>2</sup>) of ≥ 0.91. However, comparative analyses of ETa retrieval models indicated improved accuracy of the ET-EVI models (root mean square error (RMSE) of ± 8 mm for ET<sub>sugarcane</sub> and ± 4 mm for ETIa) compared with the ET-NDVI models. Among the EVI models, ET-EVI<sub>Kc</sub> achieved the highest R<sup>2</sup> of 0.98, RMSE of ≤ 30 mm, and percentage bias (PBIAS) of ≤ 15 %. The results also revealed a strong correlation between the scaled VI-derived models and the reference ETIa (R<sup>2</sup> = 0.94–0.97), which best explained the field-by-field variability, with the ET-EVI*<sub>scaled</sub> model achieving a lower RMSE of 18 mm than the ET-NDVI*<sub>scaled</sub> model (RMSE= 32 mm), while both the models showed similar levels of bias (∼17 %). Moreover, compared to the referenced ET<sub>sugarcane</sub>, the bias was minimal at − 9 % for ET-NDVI*<sub>scaled</sub> and − 1 % for ET-EVI*<sub>scaled</sub>. At the field scale, the NDVI and EVI models estimated the mean monthly ETa ranging from 99 to 129 mm m<sup>−1</sup> and 89 to 148 mm m<sup>−1</sup>, respectively, with total annual averages of 1188–1537 mm yr<sup>−1</sup> and 1296–1566 mm yr<sup>−1</sup>. In this context, the modeled ETa provided improved insights into consumptive water use in irrigated sugarcane plantations with limited field measurements. The statistical model evaluation metrics indic","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":null,"pages":null},"PeriodicalIF":10.6,"publicationDate":"2024-07-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141588661","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
ISPRS Journal of Photogrammetry and Remote Sensing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1