首页 > 最新文献

ISPRS Journal of Photogrammetry and Remote Sensing最新文献

英文 中文
A cost-effective and robust mapping method for diverse crop types using weakly supervised semantic segmentation with sparse point samples 利用带有稀疏点样本的弱监督语义分割技术,为不同作物类型提供经济高效且稳健的绘图方法
IF 10.6 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2024-09-20 DOI: 10.1016/j.isprsjprs.2024.09.017
Zhiwen Cai , Baodong Xu , Qiangyi Yu , Xinyu Zhang , Jingya Yang , Haodong Wei , Shiqi Li , Qian Song , Hang Xiong , Hao Wu , Wenbin Wu , Zhihua Shi , Qiong Hu

Accurate and timely information on the spatial distribution and areas of crop types is critical for yield estimation, agricultural management, and sustainable development. However, traditional crop classification methods often struggle to identify various crop types effectively due to their intricate spatiotemporal patterns and high training data demands. To address this challenge, we developed a Structure-aware Label eXpansion segmentation Network (StructLabX-Net) for diverse crop type mapping using limited point-annotated samples. StructLabX-Net features a backbone U-TempoNet, which combines CNNs and LSTM to explore intricate spatiotemporal patterns. It also incorporates multi-task weak supervision heads for edge detection and pseudo-label expansion, adding crucial structure and contextual insights. We tested the StructLabX-Net across three distinct regions in China, assessing over 10 crop types and comparing its performance against five popular classifiers based on multi-temporal Sentinel-2 images. The results showed that StructLabX-Net significantly outperformed RF, SVM, DeepCropMapping, Transformer, and patch-based CNN in identifying various crop types across three regions with sparse training samples. It achieved the highest overall accuracy and mean F1-score: 91.0% and 89.1% in Jianghan Plain, 91.5% and 90.7% in Songnen Plain, as well as 91.0% and 90.8% in Sanjiang Plain. StructLabX-Net demonstrated a particular advantage for those “hard types” characterized by limited samples and complex phenological features. Furthermore, ablation experiments highlight the crucial role of the “edge” head in guiding the model to accurately differentiate between various crop types with clearer class boundaries, and the “expansion” head in refining the understanding of target crops by providing extra details in pseudo-labels. Meanwhile, combining our backbone U-TempoNet with multi-task weak supervision heads exhibited superior results of crop type mapping than those derived by other segmentation models. Overall, StructLabX-Net maximizes the utilization of limited sparse samples from field surveys, offering a simple, cost-effective, and robust solution for accurately mapping various crop types at large scales. The code will be publicly available at https://github.com/BruceKai/StructLabX-Net.

准确及时的作物类型空间分布和面积信息对于产量估算、农业管理和可持续发展至关重要。然而,传统的作物分类方法由于其错综复杂的时空模式和对训练数据的高要求,往往难以有效识别各种作物类型。为了应对这一挑战,我们开发了结构感知标签扩展分割网络(StructLabX-Net),利用有限的点标注样本绘制多种作物类型图。StructLabX-Net 以 U-TempoNet 为骨干,结合了 CNN 和 LSTM 来探索复杂的时空模式。它还结合了用于边缘检测和伪标签扩展的多任务弱监督头,增加了重要的结构和上下文洞察力。我们在中国三个不同地区测试了 StructLabX-Net,评估了 10 多种作物类型,并将其性能与基于多时相 Sentinel-2 图像的五种流行分类器进行了比较。结果表明,StructLabX-Net 在识别三个地区的各种作物类型时,明显优于 RF、SVM、DeepCropMapping、Transformer 和基于补丁的 CNN(训练样本稀少)。它取得了最高的总体准确率和平均 F1 分数:江汉平原分别为 91.0% 和 89.1%,松嫩平原分别为 91.5% 和 90.7%,三江平原分别为 91.0% 和 90.8%。StructLabX-Net 对样本有限、物候特征复杂的 "硬类型 "具有特别的优势。此外,消融实验凸显了 "边缘 "头和 "扩展 "头的重要作用。"边缘 "头引导模型以更清晰的类别边界准确区分各种作物类型,而 "扩展 "头则通过提供伪标签中的额外细节来完善对目标作物的理解。同时,将我们的骨干 U-TempoNet 与多任务弱监督头相结合,作物类型映射结果优于其他分割模型。总之,StructLabX-Net 最大限度地利用了田野调查中有限的稀疏样本,为在大尺度上精确绘制各种作物类型提供了一个简单、经济、稳健的解决方案。代码将在 https://github.com/BruceKai/StructLabX-Net 上公开发布。
{"title":"A cost-effective and robust mapping method for diverse crop types using weakly supervised semantic segmentation with sparse point samples","authors":"Zhiwen Cai ,&nbsp;Baodong Xu ,&nbsp;Qiangyi Yu ,&nbsp;Xinyu Zhang ,&nbsp;Jingya Yang ,&nbsp;Haodong Wei ,&nbsp;Shiqi Li ,&nbsp;Qian Song ,&nbsp;Hang Xiong ,&nbsp;Hao Wu ,&nbsp;Wenbin Wu ,&nbsp;Zhihua Shi ,&nbsp;Qiong Hu","doi":"10.1016/j.isprsjprs.2024.09.017","DOIUrl":"10.1016/j.isprsjprs.2024.09.017","url":null,"abstract":"<div><p>Accurate and timely information on the spatial distribution and areas of crop types is critical for yield estimation, agricultural management, and sustainable development. However, traditional crop classification methods often struggle to identify various crop types effectively due to their intricate spatiotemporal patterns and high training data demands. To address this challenge, we developed a <strong>Struct</strong>ure-aware <strong>Lab</strong>el e<strong>X</strong>pansion segmentation Network (StructLabX-Net) for diverse crop type mapping using limited point-annotated samples. StructLabX-Net features a backbone U-TempoNet, which combines CNNs and LSTM to explore intricate spatiotemporal patterns. It also incorporates multi-task weak supervision heads for edge detection and pseudo-label expansion, adding crucial structure and contextual insights. We tested the StructLabX-Net across three distinct regions in China, assessing over 10 crop types and comparing its performance against five popular classifiers based on multi-temporal Sentinel-2 images. The results showed that StructLabX-Net significantly outperformed RF, SVM, DeepCropMapping, Transformer, and patch-based CNN in identifying various crop types across three regions with sparse training samples. It achieved the highest overall accuracy and mean <em>F1-score</em>: 91.0% and 89.1% in Jianghan Plain, 91.5% and 90.7% in Songnen Plain, as well as 91.0% and 90.8% in Sanjiang Plain. StructLabX-Net demonstrated a particular advantage for those “hard types” characterized by limited samples and complex phenological features. Furthermore, ablation experiments highlight the crucial role of the “edge” head in guiding the model to accurately differentiate between various crop types with clearer class boundaries, and the “expansion” head in refining the understanding of target crops by providing extra details in pseudo-labels. Meanwhile, combining our backbone U-TempoNet with multi-task weak supervision heads exhibited superior results of crop type mapping than those derived by other segmentation models. Overall, StructLabX-Net maximizes the utilization of limited sparse samples from field surveys, offering a simple, cost-effective, and robust solution for accurately mapping various crop types at large scales. The code will be publicly available at <span><span>https://github.com/BruceKai/StructLabX-Net</span><svg><path></path></svg></span>.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"218 ","pages":"Pages 260-276"},"PeriodicalIF":10.6,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142270434","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SoftFormer: SAR-optical fusion transformer for urban land use and land cover classification SoftFormer:用于城市土地利用和土地覆被分类的合成孔径雷达-光学融合变换器
IF 10.6 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2024-09-20 DOI: 10.1016/j.isprsjprs.2024.09.012
Rui Liu , Jing Ling , Hongsheng Zhang

Classification of urban land use and land cover is vital to many applications, and naturally becomes a popular topic in remote sensing. The finite information carried by unimodal data, the compound land use types, and the poor signal-noise ratio caused by restricted weather conditions would inevitably lead to relatively poor classification performance. Recently in remote sensing society, multimodal data fusion with deep learning technology has gained a great deal of attention. Existing research exhibit integration of multimodal data at a single level, while simultaneously lacking exploration of the immense potential provided by popular transformer and CNN structures for effectively leveraging multimodal data, which may fall into the trap that makes the information fusion inadequate. We introduce SoftFormer, a novel network that synergistically merges the strengths of CNNs with transformers, as well as achieving multi-level fusion. To extract local features from images, we propose an innovative mechanism called Interior Self-Attention, which is seamlessly integrated into the backbone network. To fully exploit the global semantic information from both modalities, in the feature-level fusion, we introduce a joint key–value learning fusion approach to integrate multimodal data within a unified semantic space. The decision and feature level information are simultaneously integrated, resulting in a multi-level fusion transformer network. Results on four remote sensing datasets show that SoftFormer is able to achieve at least 1.32%, 0.7%, and 0.99% performance improvement in overall accuracy, kappa index, and mIoU, compared to other state-of-the-art methods, the ablation studies show that multimodal fusion outperforms the unimodal data on urban land cover and land use classification, the highest overall accuracy, kappa index as well as mIoU improvement can be up to 5.71%, 10.32% and 7.91%, and the proposed modules are able to boost performance to some extent, even with cloud cover. Code will be publicly available at https://github.com/rl1024/SoftFormer.

城市土地利用和土地覆被分类对许多应用都至关重要,自然也成为遥感领域的热门话题。单模态数据所承载的信息有限,土地利用类型复杂,加上天气条件限制导致信噪比较差,这些因素都不可避免地导致分类性能相对较差。近年来,在遥感领域,利用深度学习技术进行多模态数据融合受到了广泛关注。现有研究在单一层面上展示了多模态数据的融合,但同时缺乏对流行的变换器和 CNN 结构在有效利用多模态数据方面所提供的巨大潜力的挖掘,这可能会陷入信息融合不足的陷阱。我们介绍的 SoftFormer 是一种新型网络,它协同融合了 CNN 和变换器的优势,并实现了多级融合。为了从图像中提取局部特征,我们提出了一种称为 "内部自注意 "的创新机制,并将其无缝集成到主干网络中。为了充分利用两种模态的全局语义信息,在特征级融合中,我们引入了一种联合键值学习融合方法,在统一的语义空间内整合多模态数据。决策级信息和特征级信息同时融合,形成多级融合转换器网络。对四个遥感数据集的研究结果表明,与其他最先进的方法相比,SoftFormer 在总体准确率、kappa 指数和 mIoU 方面至少能实现 1.32%、0.7% 和 0.99% 的性能提升,消融研究表明,在城市土地覆被和土地利用分类方面,多模态融合优于单模态数据,总体准确率、kappa 指数以及 mIoU 的最高提升可达 5.71%、10.32% 和 7.91%,即使在有云层覆盖的情况下,所提出的模块也能在一定程度上提高性能。代码将在 https://github.com/rl1024/SoftFormer 上公开。
{"title":"SoftFormer: SAR-optical fusion transformer for urban land use and land cover classification","authors":"Rui Liu ,&nbsp;Jing Ling ,&nbsp;Hongsheng Zhang","doi":"10.1016/j.isprsjprs.2024.09.012","DOIUrl":"10.1016/j.isprsjprs.2024.09.012","url":null,"abstract":"<div><p>Classification of urban land use and land cover is vital to many applications, and naturally becomes a popular topic in remote sensing. The finite information carried by unimodal data, the compound land use types, and the poor signal-noise ratio caused by restricted weather conditions would inevitably lead to relatively poor classification performance. Recently in remote sensing society, multimodal data fusion with deep learning technology has gained a great deal of attention. Existing research exhibit integration of multimodal data at a single level, while simultaneously lacking exploration of the immense potential provided by popular transformer and CNN structures for effectively leveraging multimodal data, which may fall into the trap that makes the information fusion inadequate. We introduce SoftFormer, a novel network that synergistically merges the strengths of CNNs with transformers, as well as achieving multi-level fusion. To extract local features from images, we propose an innovative mechanism called Interior Self-Attention, which is seamlessly integrated into the backbone network. To fully exploit the global semantic information from both modalities, in the feature-level fusion, we introduce a joint key–value learning fusion approach to integrate multimodal data within a unified semantic space. The decision and feature level information are simultaneously integrated, resulting in a multi-level fusion transformer network. Results on four remote sensing datasets show that SoftFormer is able to achieve at least 1.32%, 0.7%, and 0.99% performance improvement in overall accuracy, kappa index, and mIoU, compared to other state-of-the-art methods, the ablation studies show that multimodal fusion outperforms the unimodal data on urban land cover and land use classification, the highest overall accuracy, kappa index as well as mIoU improvement can be up to 5.71%, 10.32% and 7.91%, and the proposed modules are able to boost performance to some extent, even with cloud cover. Code will be publicly available at <span><span>https://github.com/rl1024/SoftFormer</span><svg><path></path></svg></span>.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"218 ","pages":"Pages 277-293"},"PeriodicalIF":10.6,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142270435","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An automatic procedure for mapping burned areas globally using Sentinel-2 and VIIRS/MODIS active fires in Google Earth Engine 在谷歌地球引擎中利用哨兵-2 和 VIIRS/MODIS 主动火灾绘制全球烧毁地区地图的自动程序
IF 10.6 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2024-09-19 DOI: 10.1016/j.isprsjprs.2024.08.019
Aitor Bastarrika , Armando Rodriguez-Montellano , Ekhi Roteta , Stijn Hantson , Magí Franquesa , Leyre Torre , Jon Gonzalez-Ibarzabal , Karmele Artano , Pilar Martinez-Blanco , Amaia Mesanza , Jesús A. Anaya , Emilio Chuvieco

Understanding the spatial and temporal trends of burned areas (BA) on a global scale offers a comprehensive view of the underlying mechanisms driving fire incidence and its influence on ecosystems and vegetation recovery patterns over extended periods. Such insights are invaluable for modeling fire emissions and the formulation of strategies for post-fire rehabilitation planning.

Previous research has provided strong evidence that current global BA products derived from coarse spatial resolution data underestimates global burned areas. Consequently, there is a pressing need for global high-resolution BA products. Here, we present an automatic global burned area mapping algorithm (Sentinel2BAM) based on Sentinel-2 Level-2A imagery combined with Visible Infrared Imaging Radiometer Suite (VIIRS) and Moderate Resolution Imaging Spectrometer (MODIS) active fire data. The algorithm employs a Random Forest Model trained by active fires to predict BA probabilities in each 5-day Normalized Burn Ratio (NBR) index-based temporal composites. In a second step, a time-series and object-based analysis of the estimated BA probabilities allows burned areas to be detected on a quarterly basis. The algorithm was implemented in Google Earth Engine (GEE) and applied to 576 Sentinel-2 tiles corresponding to 2019, distributed globally, to assess its ability to map burned areas across different ecosystems. Two validation sources were employed: 21 EMSR Copernicus Emergency Service perimeters obtained using high spatial resolution (<10 m) data (EMSR21) located in the Mediterranean basin and 50 20x20 km global samples selected by stratified sampling with Sentinel-2 at 10 m spatial resolution (GlobalS50). Additionally, 105 Landsat-based long sample units (GlobalL105), were employed to compare the performance of the Sentinel2BAM algorithm against the FIRECCI51 and MCD64A1 global products. Overall accuracy metrics for the Sentinel2BAM algorithm, derived from validation sources highlight higher commission (CE) than omission (OE) errors (CE=10.3 % and OE=7.6 % when using EMSR21 as reference, CE=18.9 % and OE=9.5 % when using Global S50 as reference), while GlobalL105-based inferenced global comparison metrics show similar patterns (CE=22.5 % and OE=13.4 %). Results indicate differences across ecosystems: forest fires in tropical and temperate biomes exhibit higher CE, mainly due to confusion between burned areas and croplands. According to GlobalL105, Sentinel2BAM shows greater accuracy globally (CE=22.5 %, OE=13.4 %) compared to FIRECCI51 (CE=20.8 %, OE=46.5 %) and MCD64A1 (CE=17.5 %, OE=53.1 %), substantially improving the detection of small fires and thereby reducing omission errors. The strengths and weaknesses of the algorithm are thoroughly addressed, demonstrating its potential for global application.

了解全球范围内烧毁面积(BA)的空间和时间趋势,可以全面了解火灾发生的根本原因及其对生态系统和植被恢复模式的长期影响。这些见解对于模拟火灾排放和制定火灾后恢复规划战略非常宝贵。
{"title":"An automatic procedure for mapping burned areas globally using Sentinel-2 and VIIRS/MODIS active fires in Google Earth Engine","authors":"Aitor Bastarrika ,&nbsp;Armando Rodriguez-Montellano ,&nbsp;Ekhi Roteta ,&nbsp;Stijn Hantson ,&nbsp;Magí Franquesa ,&nbsp;Leyre Torre ,&nbsp;Jon Gonzalez-Ibarzabal ,&nbsp;Karmele Artano ,&nbsp;Pilar Martinez-Blanco ,&nbsp;Amaia Mesanza ,&nbsp;Jesús A. Anaya ,&nbsp;Emilio Chuvieco","doi":"10.1016/j.isprsjprs.2024.08.019","DOIUrl":"10.1016/j.isprsjprs.2024.08.019","url":null,"abstract":"<div><p>Understanding the spatial and temporal trends of burned areas (BA) on a global scale offers a comprehensive view of the underlying mechanisms driving fire incidence and its influence on ecosystems and vegetation recovery patterns over extended periods. Such insights are invaluable for modeling fire emissions and the formulation of strategies for post-fire rehabilitation planning.</p><p>Previous research has provided strong evidence that current global BA products derived from coarse spatial resolution data underestimates global burned areas. Consequently, there is a pressing need for global high-resolution BA products. Here, we present an automatic global burned area mapping algorithm (Sentinel2BAM) based on Sentinel-2 Level-2A imagery combined with Visible Infrared Imaging Radiometer Suite (VIIRS) and Moderate Resolution Imaging Spectrometer (MODIS) active fire data. The algorithm employs a Random Forest Model trained by active fires to predict BA probabilities in each 5-day Normalized Burn Ratio (NBR) index-based temporal composites. In a second step, a time-series and object-based analysis of the estimated BA probabilities allows burned areas to be detected on a quarterly basis. The algorithm was implemented in Google Earth Engine (GEE) and applied to 576 Sentinel-2 tiles corresponding to 2019, distributed globally, to assess its ability to map burned areas across different ecosystems. Two validation sources were employed: 21 EMSR Copernicus Emergency Service perimeters obtained using high spatial resolution (&lt;10 m) data (EMSR21) located in the Mediterranean basin and 50 20x20 km global samples selected by stratified sampling with Sentinel-2 at 10 m spatial resolution (GlobalS50). Additionally, 105 Landsat-based long sample units (GlobalL105), were employed to compare the performance of the Sentinel2BAM algorithm against the FIRECCI51 and MCD64A1 global products. Overall accuracy metrics for the Sentinel2BAM algorithm, derived from validation sources highlight higher commission (CE) than omission (OE) errors (CE=10.3 % and OE=7.6 % when using EMSR21 as reference, CE=18.9 % and OE=9.5 % when using Global S50 as reference), while GlobalL105-based inferenced global comparison metrics show similar patterns (CE=22.5 % and OE=13.4 %). Results indicate differences across ecosystems: forest fires in tropical and temperate biomes exhibit higher CE, mainly due to confusion between burned areas and croplands. According to GlobalL105, Sentinel2BAM shows greater accuracy globally (CE=22.5 %, OE=13.4 %) compared to FIRECCI51 (CE=20.8 %, OE=46.5 %) and MCD64A1 (CE=17.5 %, OE=53.1 %), substantially improving the detection of small fires and thereby reducing omission errors. The strengths and weaknesses of the algorithm are thoroughly addressed, demonstrating its potential for global application.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"218 ","pages":"Pages 232-245"},"PeriodicalIF":10.6,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142246936","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A thin cloud blind correction method coupling a physical model with unsupervised deep learning for remote sensing imagery 将物理模型与遥感图像无监督深度学习相结合的薄云盲校正方法
IF 10.6 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2024-09-19 DOI: 10.1016/j.isprsjprs.2024.09.008
Liying Xu , Huifang Li , Huanfeng Shen , Chi Zhang , Liangpei Zhang

Thin cloud disturbs the observation of optical sensors, thus reducing the quality of optical remote sensing images and limiting the subsequent applications. However, the reliance of the existing thin cloud correction methods on the assistance of in-situ parameters, prior assumptions, massive paired data, or special bands severely limits their generalization. Moreover, due to the inadequate consideration of cloud characteristics, these methods struggle to obtain accurate results with complex degradations. To address the above two problems, a thin cloud blind correction (TC-BC) method coupling a cloudy image imaging model and a feature separation network (FSNet) module is proposed in this paper, based on an unsupervised self-training framework. Specifically, the FSNet module takes the independence and obscure boundary characteristics of the cloud into account to improve the correction accuracy with complex degradations. The FSNet module consists of an information interaction structure for exchanging the complementary features between cloud and ground, and a spatially adaptive structure for promoting the learning of the thin cloud distribution. Thin cloud correction experiments were conducted on an unpaired blind correction dataset (UBCSet) and the proposed TC-BC method was compared with three traditional methods. The visual results suggest that the proposed method shows obvious advantages in information recovery for thin cloud cover regions, and shows a superior global consistency between cloudy regions and clear regions. The TC-BC method also achieves the highest peak signal-to-noise ratio (PSNR) and structural similarity index measure (SSIM). The FSNet module in the TC-BC method is also proven to be effective. The FSNet module can achieve a superior precision when compared with five other deep learning networks in cloud-ground separation performance. Finally, extra experimental results show that the TC-BC method can be applied to different cloud correction scenarios with varied cloud coverage, surface types, and image scales, demonstrating its generalizability. Code: https://github.com/Liying-Xu/TCBC.

薄云会干扰光学传感器的观测,从而降低光学遥感图像的质量,限制其后续应用。然而,现有的薄云校正方法依赖于现场参数、先验假设、大量配对数据或特殊波段的辅助,这严重限制了其通用性。此外,由于没有充分考虑云的特征,这些方法很难在复杂衰减的情况下获得准确的结果。为解决上述两个问题,本文基于无监督自训练框架,提出了一种将多云图像成像模型和特征分离网络(FSNet)模块相结合的薄云盲校正(TC-BC)方法。具体来说,FSNet 模块考虑了云的独立性和模糊边界特性,以提高复杂降解情况下的校正精度。FSNet 模块包括用于交换云和地面之间互补特征的信息交互结构,以及用于促进薄云分布学习的空间自适应结构。在无配对盲校正数据集(UBCSet)上进行了薄云校正实验,并将提出的 TC-BC 方法与三种传统方法进行了比较。直观结果表明,所提出的方法在薄云覆盖区域的信息恢复方面具有明显优势,并且在多云区域和晴朗区域之间表现出更优越的全局一致性。TC-BC 方法还获得了最高的峰值信噪比(PSNR)和结构相似性指数(SSIM)。TC-BC 方法中的 FSNet 模块也被证明是有效的。与其他五个深度学习网络相比,FSNet 模块在云地分离性能方面的精度更高。最后,额外的实验结果表明,TC-BC 方法可应用于不同云覆盖率、地表类型和图像尺度的不同云校正场景,证明了该方法的普适性。代码:https://github.com/Liying-Xu/TCBC。
{"title":"A thin cloud blind correction method coupling a physical model with unsupervised deep learning for remote sensing imagery","authors":"Liying Xu ,&nbsp;Huifang Li ,&nbsp;Huanfeng Shen ,&nbsp;Chi Zhang ,&nbsp;Liangpei Zhang","doi":"10.1016/j.isprsjprs.2024.09.008","DOIUrl":"10.1016/j.isprsjprs.2024.09.008","url":null,"abstract":"<div><p>Thin cloud disturbs the observation of optical sensors, thus reducing the quality of optical remote sensing images and limiting the subsequent applications. However, the reliance of the existing thin cloud correction methods on the assistance of in-situ parameters, prior assumptions, massive paired data, or special bands severely limits their generalization. Moreover, due to the inadequate consideration of cloud characteristics, these methods struggle to obtain accurate results with complex degradations. To address the above two problems, a thin cloud blind correction (TC-BC) method coupling a cloudy image imaging model and a feature separation network (FSNet) module is proposed in this paper, based on an unsupervised self-training framework. Specifically, the FSNet module takes the independence and obscure boundary characteristics of the cloud into account to improve the correction accuracy with complex degradations. The FSNet module consists of an information interaction structure for exchanging the complementary features between cloud and ground, and a spatially adaptive structure for promoting the learning of the thin cloud distribution. Thin cloud correction experiments were conducted on an unpaired blind correction dataset (UBCSet) and the proposed TC-BC method was compared with three traditional methods. The visual results suggest that the proposed method shows obvious advantages in information recovery for thin cloud cover regions, and shows a superior global consistency between cloudy regions and clear regions. The TC-BC method also achieves the highest peak signal-to-noise ratio (PSNR) and structural similarity index measure (SSIM). The FSNet module in the TC-BC method is also proven to be effective. The FSNet module can achieve a superior precision when compared with five other deep learning networks in cloud-ground separation performance. Finally, extra experimental results show that the TC-BC method can be applied to different cloud correction scenarios with varied cloud coverage, surface types, and image scales, demonstrating its generalizability. Code: <span><span>https://github.com/Liying-Xu/TCBC</span><svg><path></path></svg></span>.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"218 ","pages":"Pages 246-259"},"PeriodicalIF":10.6,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142246972","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Nonlinear least-squares solutions to the TLS multi-station registration adjustment problem TLS 多站注册调整问题的非线性最小二乘法解决方案
IF 10.6 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2024-09-19 DOI: 10.1016/j.isprsjprs.2024.09.014
Yu Hu, Xing Fang, Wenxian Zeng

Performing multiple scans is necessary to cover an entire scene of interest, making multi-station registration adjustment a critical task in terrestrial laser scanner data processing. Existing methods either rely on pair-wise adjustment, which leads to drift accumulation and lacks global consistency, or provide an approximate solution based on a linearized model, sacrificing statistical optimality. In this study, using a multi-station stacking model, we propose a method that provides two different nonlinear least-squares (LS) solutions to this problem. We first demonstrate how a nonlinear Baarda’s S-transformation can be used to transform solutions that share the same optimal network configuration. Then, two practically meaningful LS solutions are introduced, i.e., the trivial minimal-constraints solution and the partial nearest solution. Most importantly, we derive a truncated Gauss–Newton iterative scheme to obtain numerically exact solutions to the corresponding nonlinear rank-deficient problem. We validate our method with three real-world examples, demonstrating that (1) global consistency is maintained with no drift accumulation, and (2) our nonlinear solution outperforms the approximate linearized solution. Code and data are available at https://github.com/huyuchn/Multi-station-registration.

要覆盖整个感兴趣的场景,必须进行多次扫描,因此多站配准调整成为陆地激光扫描仪数据处理中的一项关键任务。现有的方法要么依赖于成对调整,这会导致漂移累积并缺乏全局一致性;要么基于线性化模型提供近似解决方案,但牺牲了统计最优性。在本研究中,我们利用多站堆叠模型,提出了一种方法,为这一问题提供了两种不同的非线性最小二乘(LS)解决方案。我们首先展示了如何利用非线性 Baarda's S 变换来转换具有相同最优网络配置的解决方案。然后,我们介绍了两种具有实际意义的 LS 解法,即微不足道的最小约束解法和部分最近解法。最重要的是,我们推导出一种截断高斯-牛顿迭代方案,以获得相应非线性秩缺陷问题的数值精确解。我们用三个实际例子验证了我们的方法,证明:(1) 保持了全局一致性,没有漂移累积;(2) 我们的非线性解优于近似线性化解。代码和数据可在 https://github.com/huyuchn/Multi-station-registration 上获取。
{"title":"Nonlinear least-squares solutions to the TLS multi-station registration adjustment problem","authors":"Yu Hu,&nbsp;Xing Fang,&nbsp;Wenxian Zeng","doi":"10.1016/j.isprsjprs.2024.09.014","DOIUrl":"10.1016/j.isprsjprs.2024.09.014","url":null,"abstract":"<div><p>Performing multiple scans is necessary to cover an entire scene of interest, making multi-station registration adjustment a critical task in terrestrial laser scanner data processing. Existing methods either rely on pair-wise adjustment, which leads to drift accumulation and lacks global consistency, or provide an approximate solution based on a linearized model, sacrificing statistical optimality. In this study, using a multi-station stacking model, we propose a method that provides two different nonlinear least-squares (LS) solutions to this problem. We first demonstrate how a nonlinear Baarda’s S-transformation can be used to transform solutions that share the same optimal network configuration. Then, two practically meaningful LS solutions are introduced, i.e., the trivial minimal-constraints solution and the partial nearest solution. Most importantly, we derive a truncated Gauss–Newton iterative scheme to obtain numerically exact solutions to the corresponding nonlinear rank-deficient problem. We validate our method with three real-world examples, demonstrating that (1) global consistency is maintained with no drift accumulation, and (2) our nonlinear solution outperforms the approximate linearized solution. Code and data are available at <span><span>https://github.com/huyuchn/Multi-station-registration</span><svg><path></path></svg></span>.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"218 ","pages":"Pages 220-231"},"PeriodicalIF":10.6,"publicationDate":"2024-09-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142243277","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Joint block adjustment and variational optimization for global and local radiometric normalization toward multiple remote sensing image mosaicking 针对多幅遥感图像镶嵌的全局和局部辐射度归一化的联合区块调整和变异优化
IF 10.6 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2024-09-17 DOI: 10.1016/j.isprsjprs.2024.08.016
Dekun Lin , Huanfeng Shen , Xinghua Li , Chao Zeng , Tao Jiang , Yongming Ma , Mingjie Xu

Multi-temporal optical remote sensing images acquired from cross-sensor platforms often show significant radiometric differences, posing challenges when mosaicking images. These challenges include inconsistent global radiometric tones, unsmooth local radiometric transitions, and visible seamlines. In this paper, to address these challenges, we propose a two-stage approach for global and local radiometric normalization (RN) using joint block adjustment and variational optimization. In the first stage, a block adjustment based global RN (BAGRN) model is established to simultaneously perform global RN on all the images, eliminating global radiometric differences and achieving overall radiometric tonal consistency. In the second stage, a variational optimization based local RN (VOLRN) model is introduced to address the remaining local radiometric differences after global RN. The VOLRN model applies local RN to all the image blocks within a unified energy function and imposes the l1 norm constraint on the data fidelity term, providing the model with a more flexible local RN capability to radiometrically normalize the intersection and transition areas of the images. Therefore, the local radiometric discontinuities and edge artifacts can be eliminated, resulting in natural and smooth local radiometric transitions. The experimental results obtained on five challenging datasets of cross-sensor and multi-temporal remote sensing images demonstrate that the proposed approach excels in both visual quality and quantitative metrics. The proposed approach effectively eliminates global and local radiometric differences, preserves image gradients well, and has high processing efficiency. As a result, it outperforms the state-of-the-art RN approaches.

从跨传感器平台获取的多时相光学遥感图像往往显示出明显的辐射度差异,这给图像镶嵌带来了挑战。这些挑战包括不一致的全局辐射测量色调、不平滑的局部辐射测量过渡以及可见的接缝线。在本文中,为了应对这些挑战,我们提出了一种两阶段方法,利用联合块调整和变异优化实现全局和局部辐射度归一化(RN)。在第一阶段,建立基于块调整的全局归一化(BAGRN)模型,同时对所有图像执行全局归一化,消除全局辐射度差异,实现整体辐射度色调一致性。在第二阶段,引入基于变异优化的局部 RN(VOLRN)模型,以解决全局 RN 后剩余的局部辐射度差异。VOLRN 模型将局部 RN 应用于统一能量函数内的所有图像块,并对数据保真度项施加 l1 准则约束,从而使该模型具有更灵活的局部 RN 功能,可对图像的交叉和过渡区域进行辐射度归一化处理。因此,局部辐射度不连续性和边缘伪影可以被消除,从而实现自然平滑的局部辐射度过渡。在五个具有挑战性的跨传感器和多时相遥感图像数据集上获得的实验结果表明,所提出的方法在视觉质量和定量指标方面都非常出色。所提出的方法能有效消除全局和局部辐射度差异,很好地保留图像梯度,而且处理效率高。因此,它优于最先进的 RN 方法。
{"title":"Joint block adjustment and variational optimization for global and local radiometric normalization toward multiple remote sensing image mosaicking","authors":"Dekun Lin ,&nbsp;Huanfeng Shen ,&nbsp;Xinghua Li ,&nbsp;Chao Zeng ,&nbsp;Tao Jiang ,&nbsp;Yongming Ma ,&nbsp;Mingjie Xu","doi":"10.1016/j.isprsjprs.2024.08.016","DOIUrl":"10.1016/j.isprsjprs.2024.08.016","url":null,"abstract":"<div><p>Multi-temporal optical remote sensing images acquired from cross-sensor platforms often show significant radiometric differences, posing challenges when mosaicking images. These challenges include inconsistent global radiometric tones, unsmooth local radiometric transitions, and visible seamlines. In this paper, to address these challenges, we propose a two-stage approach for global and local radiometric normalization (RN) using joint block adjustment and variational optimization. In the first stage, a block adjustment based global RN (BAGRN) model is established to simultaneously perform global RN on all the images, eliminating global radiometric differences and achieving overall radiometric tonal consistency. In the second stage, a variational optimization based local RN (VOLRN) model is introduced to address the remaining local radiometric differences after global RN. The VOLRN model applies local RN to all the image blocks within a unified energy function and imposes the <span><math><mrow><msub><mi>l</mi><mn>1</mn></msub></mrow></math></span> norm constraint on the data fidelity term, providing the model with a more flexible local RN capability to radiometrically normalize the intersection and transition areas of the images. Therefore, the local radiometric discontinuities and edge artifacts can be eliminated, resulting in natural and smooth local radiometric transitions. The experimental results obtained on five challenging datasets of cross-sensor and multi-temporal remote sensing images demonstrate that the proposed approach excels in both visual quality and quantitative metrics. The proposed approach effectively eliminates global and local radiometric differences, preserves image gradients well, and has high processing efficiency. As a result, it outperforms the state-of-the-art RN approaches.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"218 ","pages":"Pages 187-203"},"PeriodicalIF":10.6,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142243271","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Weather-aware autopilot: Domain generalization for point cloud semantic segmentation in diverse weather scenarios 天气感知自动驾驶仪:不同天气情况下点云语义分割的领域泛化
IF 10.6 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2024-09-17 DOI: 10.1016/j.isprsjprs.2024.09.006
Jing Du , John Zelek , Jonathan Li

3D point cloud semantic segmentation, a pivotal task in fields such as autonomous driving and urban planning, confronts the challenge of performance degradation under adverse weather conditions. Current methodologies primarily focus on optimal weather scenarios, leaving a significant gap in handling various environmental adversities like fog, rain, and snow. To bridge this gap, we propose a comprehensive deep learning framework featuring unique components — an Adaptive Feature Normalization Module (AFNM) for effective normalization and calibration of features, a Dual-Attention Fusion Module (DAFM) for integrating cross-domain features, and a Proxy Label Generation Module (PLGM) for generating reliable proxy labels within the domain. Utilizing the SemanticKITTI and SynLiDAR datasets as source domains and the SemanticSTF dataset as the target domain, our model has been rigorously evaluated under varying weather conditions. When trained on the SemanticKITTI dataset as the source domain with the SemanticSTF dataset as the target, our approach surpasses the current state-of-the-art models by a margin of 6.2% in terms of overall mean Intersection over Union (mIoU) scores. Similarly, with the SynLiDAR dataset as the source and SemanticSTF as the target, our performance exceeds the best existing models by 3.4% in mIoU. These results substantiate the efficacy of our model in advancing the field of 3D semantic segmentation under diverse weather conditions, showcasing its notable robustness and superiority. The code is available at https://github.com/J2DU/WADG-PointSeg.

三维点云语义分割是自动驾驶和城市规划等领域的一项关键任务,它面临着在恶劣天气条件下性能下降的挑战。目前的方法主要集中在最佳天气情况下,在处理雾、雨和雪等各种环境逆境方面存在很大差距。为了弥补这一差距,我们提出了一个全面的深度学习框架,该框架具有独特的组件--用于有效归一化和校准特征的自适应特征归一化模块(AFNM)、用于整合跨域特征的双注意融合模块(DAFM)以及用于在域内生成可靠代理标签的代理标签生成模块(PLGM)。利用SemanticKITTI和SynLiDAR数据集作为源域,SemanticSTF数据集作为目标域,我们的模型在不同的天气条件下进行了严格的评估。以SemanticKITTI数据集为源域、SemanticSTF数据集为目标域进行训练时,我们的方法在总体平均交叉联合(mIoU)得分方面以6.2%的优势超过了目前最先进的模型。同样,以 SynLiDAR 数据集为源,SemanticSTF 为目标,我们的 mIoU 性能比现有最佳模型高出 3.4%。这些结果证明了我们的模型在各种天气条件下推进三维语义分割领域的功效,展示了其显著的鲁棒性和优越性。代码见 https://github.com/J2DU/WADG-PointSeg。
{"title":"Weather-aware autopilot: Domain generalization for point cloud semantic segmentation in diverse weather scenarios","authors":"Jing Du ,&nbsp;John Zelek ,&nbsp;Jonathan Li","doi":"10.1016/j.isprsjprs.2024.09.006","DOIUrl":"10.1016/j.isprsjprs.2024.09.006","url":null,"abstract":"<div><p>3D point cloud semantic segmentation, a pivotal task in fields such as autonomous driving and urban planning, confronts the challenge of performance degradation under adverse weather conditions. Current methodologies primarily focus on optimal weather scenarios, leaving a significant gap in handling various environmental adversities like fog, rain, and snow. To bridge this gap, we propose a comprehensive deep learning framework featuring unique components — an Adaptive Feature Normalization Module (AFNM) for effective normalization and calibration of features, a Dual-Attention Fusion Module (DAFM) for integrating cross-domain features, and a Proxy Label Generation Module (PLGM) for generating reliable proxy labels within the domain. Utilizing the SemanticKITTI and SynLiDAR datasets as source domains and the SemanticSTF dataset as the target domain, our model has been rigorously evaluated under varying weather conditions. When trained on the SemanticKITTI dataset as the source domain with the SemanticSTF dataset as the target, our approach surpasses the current state-of-the-art models by a margin of 6.2% in terms of overall mean Intersection over Union (mIoU) scores. Similarly, with the SynLiDAR dataset as the source and SemanticSTF as the target, our performance exceeds the best existing models by 3.4% in mIoU. These results substantiate the efficacy of our model in advancing the field of 3D semantic segmentation under diverse weather conditions, showcasing its notable robustness and superiority. The code is available at <span><span>https://github.com/J2DU/WADG-PointSeg</span><svg><path></path></svg></span>.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"218 ","pages":"Pages 204-219"},"PeriodicalIF":10.6,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142243276","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Reconstructing NDVI time series in cloud-prone regions: A fusion-and-fit approach with deep learning residual constraint 在多云地区重建 NDVI 时间序列:具有深度学习残差约束的融合拟合方法
IF 10.6 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2024-09-16 DOI: 10.1016/j.isprsjprs.2024.09.010
Peng Qin , Huabing Huang , Peimin Chen , Hailong Tang , Jie Wang , Shuang Chen

The time series data of Normalized Difference Vegetation Index (NDVI) is crucial for monitoring changes in terrestrial vegetation. Existing reconstruction methods encounter challenges in areas prone to clouds, primarily due to inadequate utilization of spatial, temporal, periodic, and multi-sensor information, as well as a lack of physical interpretations. This frequently results in limited model performance or the omission of spatial details when predicting scenarios involving land cover changes. In this study, we propose a novel approach named Residual (Re) Constraints (Co) fusion-and-fit (ReCoff), consisting of two steps: ReCoF fusion (F) and Savitzky-Golay (SG) fit. This approach addresses the challenges of reconstructing 30 m Landsat NDVI time series data in cloudy regions. The fusion-fit process captures land cover changes and maps them from MODIS to Landsat using a deep learning model with residual constraints, while simultaneously integrating multi-dimensional, multi-sensor, and long time-series information. ReCoff offers three distinct advantages. First, the fusion results are more robust to land cover change scenarios and contain richer spatial details (RMSE of 0.091 vs. 0.101, 0.164, and 0.188 for ReCoF vs. STFGAN, FSDAF, and ESTARFM). Second, ReCoff improves the effectiveness of reconstructing dense time-series data (2016–2020, 16-day interval) in cloudy areas, whereas other methods are more susceptible to the impact of prolonged data gaps. ReCoff achieves a correlation coefficient of 0.84 with the MODIS reference series, outperforming SG (0.28), HANTS (0.32), and GF-SG (0.48). Third, with the help of the GEE platform, ReCoff can be applied over large areas (771 km × 634 km) and long-time scales (bimonthly intervals from 2000 to 2020) in cloudy regions. ReCoff demonstrates potential for accurately reconstructing time-series data in cloudy areas.

归一化植被指数(NDVI)的时间序列数据对于监测陆地植被的变化至关重要。现有的重建方法在多云地区遇到了挑战,主要原因是没有充分利用空间、时间、周期和多传感器信息,以及缺乏物理解释。这经常导致在预测涉及土地覆被变化的场景时,模型性能有限或遗漏空间细节。在本研究中,我们提出了一种名为 "残差(Re)约束(Co)融合与拟合(ReCoff)"的新方法,由两个步骤组成:ReCoF 融合(F)和萨维茨基-戈莱(SG)拟合。这种方法解决了在多云地区重建 30 米大地遥感卫星 NDVI 时间序列数据的难题。融合拟合过程利用具有残差约束的深度学习模型捕捉土地覆被变化,并将其从 MODIS 映射到 Landsat,同时整合多维度、多传感器和长时间序列信息。ReCoff 具有三个显著优势。首先,融合结果对土地覆被变化情景更加稳健,并包含更丰富的空间细节(ReCoF 与 STFGAN、FSDAF 和 ESTARFM 相比,RMSE 分别为 0.091、0.101、0.164 和 0.188)。其次,ReCoff 提高了在多云地区重建密集时间序列数据(2016-2020 年,16 天间隔)的有效性,而其他方法更容易受到长时间数据间隙的影响。ReCoff 与 MODIS 参考序列的相关系数达到 0.84,优于 SG(0.28)、HANTS(0.32)和 GF-SG(0.48)。第三,在 GEE 平台的帮助下,ReCoff 可以应用于大面积(771 km × 634 km)和长时间尺度(2000 年至 2020 年的双月时间间隔)的多云地区。ReCoff 展示了在多云地区准确重建时间序列数据的潜力。
{"title":"Reconstructing NDVI time series in cloud-prone regions: A fusion-and-fit approach with deep learning residual constraint","authors":"Peng Qin ,&nbsp;Huabing Huang ,&nbsp;Peimin Chen ,&nbsp;Hailong Tang ,&nbsp;Jie Wang ,&nbsp;Shuang Chen","doi":"10.1016/j.isprsjprs.2024.09.010","DOIUrl":"10.1016/j.isprsjprs.2024.09.010","url":null,"abstract":"<div><p>The time series data of Normalized Difference Vegetation Index (NDVI) is crucial for monitoring changes in terrestrial vegetation. Existing reconstruction methods encounter challenges in areas prone to clouds, primarily due to inadequate utilization of spatial, temporal, periodic, and multi-sensor information, as well as a lack of physical interpretations. This frequently results in limited model performance or the omission of spatial details when predicting scenarios involving land cover changes. In this study, we propose a novel approach named Residual (Re) Constraints (Co) fusion-and-fit (ReCoff), consisting of two steps: ReCoF fusion (F) and Savitzky-Golay (SG) fit. This approach addresses the challenges of reconstructing 30 m Landsat NDVI time series data in cloudy regions. The fusion-fit process captures land cover changes and maps them from MODIS to Landsat using a deep learning model with residual constraints, while simultaneously integrating multi-dimensional, multi-sensor, and long time-series information. ReCoff offers three distinct advantages. First, the fusion results are more robust to land cover change scenarios and contain richer spatial details (RMSE of 0.091 vs. 0.101, 0.164, and 0.188 for ReCoF vs. STFGAN, FSDAF, and ESTARFM). Second, ReCoff improves the effectiveness of reconstructing dense time-series data (2016–2020, 16-day interval) in cloudy areas, whereas other methods are more susceptible to the impact of prolonged data gaps. ReCoff achieves a correlation coefficient of 0.84 with the MODIS reference series, outperforming SG (0.28), HANTS (0.32), and GF-SG (0.48). Third, with the help of the GEE platform, ReCoff can be applied over large areas (771 km × 634 km) and long-time scales (bimonthly intervals from 2000 to 2020) in cloudy regions. ReCoff demonstrates potential for accurately reconstructing time-series data in cloudy areas.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"218 ","pages":"Pages 170-186"},"PeriodicalIF":10.6,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142233854","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MuSRFM: Multiple scale resolution fusion based precise and robust satellite derived bathymetry model for island nearshore shallow water regions using sentinel-2 multi-spectral imagery MuSRFM:利用哨兵-2 号多光谱图像为岛屿近岸浅水区建立基于多尺度分辨率融合的精确、稳健的卫星水深模型
IF 10.6 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2024-09-14 DOI: 10.1016/j.isprsjprs.2024.09.007
Xiaoming Qin , Ziyin Wu , Xiaowen Luo , Jihong Shang , Dineng Zhao , Jieqiong Zhou , Jiaxin Cui , Hongyang Wan , Guochang Xu

The multi-spectral imagery based Satellite Derived Bathymetry (SDB) provides an efficient and cost-effective approach for acquiring bathymetry data of nearshore shallow water regions. Compared with conventional pixelwise inversion models, Deep Learning (DL) models have the theoretical capability to encompass a broader receptive field, automatically extracting comprehensive spatial features. However, enhancing spatial features by increasing the input size escalates computational complexity and model scale, challenging the hardware. To address this issue, we propose the Multiple Scale Resolution Fusion Model (MuSRFM), a novel DL-based SDB model, to integrate information of varying scales by utilizing temporally fused Sentinel-2 L2A multi-spectral imagery. The MuSRFM uses a Multi-scale Center-aligned Hierarchical Resampler (MCHR) to composite large-scale multi-spectral imagery into hierarchical scale resolution representations since the receptive field gradually narrows its focus as the spatial resolution decreases. Through this strategy, the MuSRFM gains access to rich spatial information while maintaining efficiency by progressively aggregating features of different scales through the Cropped Aligned Fusion Module (CAFM). We select St. Croix (Virgin Islands) as the training/testing dataset source, and the Root Mean Square Error (RMSE) obtained by the MuSRFM on the testing dataset is 0.8131 m (with a bathymetric range of 0–25 m), surpassing the machine learning based models and traditional semi-empirical models used as the baselines by over 35 % and 60 %, respectively. Additionally, multiple island areas worldwide, including Vieques, Oahu, Kauai, Saipan and Tinian, which exhibit distinct characteristics, are utilized to construct a real-world dataset for assessing the generalizability and transferability of the proposed MuSRFM. While the MuSRFM experiences a degradation in accuracy when applied to the diverse real-world dataset, it outperforms other baseline models considerably. Across various study areas in the real-world dataset, its RMSE lead over the second-ranked model ranges from 6.8 % to 38.1 %, indicating its accuracy and generalizability; in the Kauai area, where the performance is not ideal, a significant improvement in accuracy is achieved through fine-tuning on limited in-situ data. The code of MuSRFM is available at https://github.com/qxm1995716/musrfm.

基于多光谱图像的卫星衍生水深测量(SDB)为获取近岸浅水区域的水深测量数据提供了一种高效、经济的方法。与传统的像素反演模型相比,深度学习(DL)模型在理论上能够涵盖更广阔的感受野,自动提取全面的空间特征。然而,通过增大输入尺寸来增强空间特征会增加计算复杂度和模型规模,对硬件提出了挑战。为了解决这个问题,我们提出了多尺度分辨率融合模型(MuSRFM),这是一种基于 DL 的新型 SDB 模型,利用时间融合的哨兵-2 L2A 多光谱图像来整合不同尺度的信息。MuSRFM 采用多尺度中心对齐分层重采样器 (MCHR),将大尺度多光谱图像合成为分层尺度分辨率表示,因为随着空间分辨率的降低,感受野会逐渐缩小焦点。通过这种策略,MuSRFM 可以获取丰富的空间信息,同时通过裁剪对齐融合模块(CAFM)逐步聚合不同尺度的特征,从而保持效率。我们选择圣克罗伊岛(维尔京群岛)作为训练/测试数据集源,MuSRFM 在测试数据集上获得的均方根误差(RMSE)为 0.8131 米(水深范围为 0-25 米),分别超过基于机器学习的模型和传统半经验模型 35% 和 60% 以上。此外,世界各地的多个岛屿地区,包括别克斯岛、瓦胡岛、可爱岛、塞班岛和提尼安岛,都表现出不同的特征,利用这些岛屿地区构建了一个真实世界数据集,用于评估拟议的 MuSRFM 的通用性和可转移性。虽然 MuSRFM 在应用于多样化的真实世界数据集时精度有所下降,但其性能大大优于其他基线模型。在真实世界数据集的各个研究区域,它的 RMSE 领先于排名第二的模型 6.8 % 到 38.1 %,这表明了它的准确性和普适性;在考艾岛地区,它的性能并不理想,但通过对有限的现场数据进行微调,它的准确性得到了显著提高。MuSRFM 的代码见 https://github.com/qxm1995716/musrfm。
{"title":"MuSRFM: Multiple scale resolution fusion based precise and robust satellite derived bathymetry model for island nearshore shallow water regions using sentinel-2 multi-spectral imagery","authors":"Xiaoming Qin ,&nbsp;Ziyin Wu ,&nbsp;Xiaowen Luo ,&nbsp;Jihong Shang ,&nbsp;Dineng Zhao ,&nbsp;Jieqiong Zhou ,&nbsp;Jiaxin Cui ,&nbsp;Hongyang Wan ,&nbsp;Guochang Xu","doi":"10.1016/j.isprsjprs.2024.09.007","DOIUrl":"10.1016/j.isprsjprs.2024.09.007","url":null,"abstract":"<div><p>The multi-spectral imagery based Satellite Derived Bathymetry (SDB) provides an efficient and cost-effective approach for acquiring bathymetry data of nearshore shallow water regions. Compared with conventional pixelwise inversion models, Deep Learning (DL) models have the theoretical capability to encompass a broader receptive field, automatically extracting comprehensive spatial features. However, enhancing spatial features by increasing the input size escalates computational complexity and model scale, challenging the hardware. To address this issue, we propose the Multiple Scale Resolution Fusion Model (MuSRFM), a novel DL-based SDB model, to integrate information of varying scales by utilizing temporally fused Sentinel-2 L2A multi-spectral imagery. The MuSRFM uses a Multi-scale Center-aligned Hierarchical Resampler (MCHR) to composite large-scale multi-spectral imagery into hierarchical scale resolution representations since the receptive field gradually narrows its focus as the spatial resolution decreases. Through this strategy, the MuSRFM gains access to rich spatial information while maintaining efficiency by progressively aggregating features of different scales through the Cropped Aligned Fusion Module (CAFM). We select St. Croix (Virgin Islands) as the training/testing dataset source, and the Root Mean Square Error (RMSE) obtained by the MuSRFM on the testing dataset is 0.8131 m (with a bathymetric range of 0–25 m), surpassing the machine learning based models and traditional semi-empirical models used as the baselines by over 35 % and 60 %, respectively. Additionally, multiple island areas worldwide, including Vieques, Oahu, Kauai, Saipan and Tinian, which exhibit distinct characteristics, are utilized to construct a real-world dataset for assessing the generalizability and transferability of the proposed MuSRFM. While the MuSRFM experiences a degradation in accuracy when applied to the diverse real-world dataset, it outperforms other baseline models considerably. Across various study areas in the real-world dataset, its RMSE lead over the second-ranked model ranges from 6.8 % to 38.1 %, indicating its accuracy and generalizability; in the Kauai area, where the performance is not ideal, a significant improvement in accuracy is achieved through fine-tuning on limited in-situ data. The code of MuSRFM is available at <span><span>https://github.com/qxm1995716/musrfm</span><svg><path></path></svg></span>.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"218 ","pages":"Pages 150-169"},"PeriodicalIF":10.6,"publicationDate":"2024-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0924271624003459/pdfft?md5=4925ae29c5fd595f63a6ca31611a8d4c&pid=1-s2.0-S0924271624003459-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142232384","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Snow depth retrieval method for PolSAR data using multi-parameters snow backscattering model 利用多参数雪后散射模型的 PolSAR 数据雪深检索方法
IF 10.6 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2024-09-13 DOI: 10.1016/j.isprsjprs.2024.09.005
Haiwei Qiao , Ping Zhang , Zhen Li , Lei Huang , Zhipeng Wu , Shuo Gao , Chang Liu , Shuang Liang , Jianmin Zhou , Wei Sun

Snow depth (SD) is a crucial property of snow, its spatial and temporal variation is important for global change, snowmelt runoff simulation, disaster prediction, and freshwater storage estimation. Polarimetric Synthetic Aperture Radar (PolSAR) can precisely describe the backscattering of the target and emerge as an effective tool for SD retrieval. The backscattering component of dry snow is mainly composed of volume scattering from the snowpack and surface scattering from the snow-ground interface. However, the existing method for retrieving SD using PolSAR data has the problems of over-reliance on in-situ data and ignoring surface scattering from the snow-ground interface. We proposed a novel SD retrieval method for PolSAR data by fully considering the primary backscattering components of snow and through multi-parameter estimation to solve the snow backscattering model. Firstly, a snow backscattering model was formed by combining the small permittivity volume scattering model and the Michigan semi-empirical surface scattering model to simulate the different scattering components of snow, and the corresponding backscattering coefficients were extracted using the Yamaguchi decomposition. Then, the snow permittivity was calculated through generalized volume parameters and the extinction coefficient was further estimated through modeling. Finally, the snow backscattering model was solved by these parameters to retrieve SD. The proposed method was validated by Ku-band UAV SAR data acquired in Altay, Xinjiang, and the accuracy was evaluated by in-situ data. The correlation coefficient, root mean square error, and mean absolute error are 0.80, 4.49 cm, and 3.95 cm, respectively. Meanwhile, the uncertainties generated by different SD, model parameters estimation, solution method, and underlying surface are analyzed to enhance the generality of the proposed method.

雪深(SD)是雪的一个重要属性,其时空变化对全球变化、融雪径流模拟、灾害预测和淡水储量估算具有重要意义。极坐标合成孔径雷达(PolSAR)可以精确描述目标的后向散射,是一种有效的标度检索工具。干雪的后向散射成分主要由雪堆的体积散射和雪地界面的表面散射组成。然而,现有的利用 PolSAR 数据检索自毁的方法存在过度依赖原地数据和忽略雪地界面表面散射的问题。我们通过充分考虑雪的主要后向散射成分,并通过多参数估计来求解雪的后向散射模型,提出了一种新型的 PolSAR 数据自毁率检索方法。首先,结合小介电常数体积散射模型和密歇根半经验表面散射模型,模拟雪的不同散射分量,形成雪的后向散射模型,并利用山口分解法提取相应的后向散射系数。然后,通过广义体积参数计算雪的介电常数,并通过建模进一步估算消光系数。最后,通过这些参数求解雪的反向散射模型,从而得到标度。通过在新疆阿勒泰地区获取的 Ku 波段无人机合成孔径雷达数据对所提出的方法进行了验证,并通过原位数据对其精度进行了评估。相关系数、均方根误差和平均绝对误差分别为 0.80、4.49 厘米和 3.95 厘米。同时,分析了不同标度、模型参数估计、求解方法和底面产生的不确定性,以增强所提方法的通用性。
{"title":"Snow depth retrieval method for PolSAR data using multi-parameters snow backscattering model","authors":"Haiwei Qiao ,&nbsp;Ping Zhang ,&nbsp;Zhen Li ,&nbsp;Lei Huang ,&nbsp;Zhipeng Wu ,&nbsp;Shuo Gao ,&nbsp;Chang Liu ,&nbsp;Shuang Liang ,&nbsp;Jianmin Zhou ,&nbsp;Wei Sun","doi":"10.1016/j.isprsjprs.2024.09.005","DOIUrl":"10.1016/j.isprsjprs.2024.09.005","url":null,"abstract":"<div><p>Snow depth (SD) is a crucial property of snow, its spatial and temporal variation is important for global change, snowmelt runoff simulation, disaster prediction, and freshwater storage estimation. Polarimetric Synthetic Aperture Radar (PolSAR) can precisely describe the backscattering of the target and emerge as an effective tool for SD retrieval. The backscattering component of dry snow is mainly composed of volume scattering from the snowpack and surface scattering from the snow-ground interface. However, the existing method for retrieving SD using PolSAR data has the problems of over-reliance on in-situ data and ignoring surface scattering from the snow-ground interface. We proposed a novel SD retrieval method for PolSAR data by fully considering the primary backscattering components of snow and through multi-parameter estimation to solve the snow backscattering model. Firstly, a snow backscattering model was formed by combining the small permittivity volume scattering model and the Michigan semi-empirical surface scattering model to simulate the different scattering components of snow, and the corresponding backscattering coefficients were extracted using the Yamaguchi decomposition. Then, the snow permittivity was calculated through generalized volume parameters and the extinction coefficient was further estimated through modeling. Finally, the snow backscattering model was solved by these parameters to retrieve SD. The proposed method was validated by Ku-band UAV SAR data acquired in Altay, Xinjiang, and the accuracy was evaluated by in-situ data. The correlation coefficient, root mean square error, and mean absolute error are 0.80, 4.49 cm, and 3.95 cm, respectively. Meanwhile, the uncertainties generated by different SD, model parameters estimation, solution method, and underlying surface are analyzed to enhance the generality of the proposed method.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"218 ","pages":"Pages 136-149"},"PeriodicalIF":10.6,"publicationDate":"2024-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142229357","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
ISPRS Journal of Photogrammetry and Remote Sensing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1