首页 > 最新文献

IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing最新文献

英文 中文
Elevation and Vegetation Cover Dominate Inter-Basin Water Use Efficiency Patterns in China 高程和植被覆盖主导中国流域间水资源利用效率格局
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-05 DOI: 10.1109/JSTARS.2025.3640403
Jun Hu;Hongjun Su;Yiping Chen;Yuanwei Qin;Zhaohui Xue;Qian Du
Water use efficiency (WUE) is a fundamental indicator of the balance between ecosystem carbon assimilation and water consumption. However, its spatial variability and dominant environmental drivers across China's river basins remain unclear, posing challenges for basin-scale management. In this study, a comprehensive WUE analysis framework was established through the integration of multisource remote sensing and auxiliary datasets. In this framework, multisource vegetation, climate, topography, and land-use data were integrated to estimate WUE from the GPP-to-ET ratio, and a novel basin-scale dataset covering 25 major river basins in China from 2002 to 2021 was generated (CBS-WUE, https://doi.org/10.5281/zenodo.17402779), which was validated against FLUXNET2015 observations. With this new dataset, inter-basin comparisons were conducted to characterize spatial heterogeneity and temporal dynamics, while multivariate statistical and machine learning analyses were employed to identify the relative contributions of climatic, biotic, and land-use drivers. Results indicated that elevation and vegetation structure were the primary factors influencing basin-scale WUE differences. The national average WUE was 1.13 g C kg−1 H2O, with basin-level values ranging from 0.11 to 1.80 g C kg−1 H2O. Among them, higher WUE was in basins of moderate elevation and dense vegetation, and lower WUE was in high-elevation or arid basins. This integrative analysis highlights the dominant role of topography and vegetation in shaping WUE patterns and provides a scientific basis for enhancing water resource efficiency and ecological sustainability under changing environmental conditions.
水分利用效率(WUE)是衡量生态系统碳同化与水分消耗平衡的基本指标。然而,中国河流流域的空间变异性和主导环境驱动因素尚不清楚,这给流域尺度管理带来了挑战。本研究通过多源遥感和辅助数据集的整合,建立了一个综合的WUE分析框架。在该框架下,综合多源植被、气候、地形和土地利用数据,从gpp - et比估算WUE,并生成了覆盖中国25个主要流域的2002 - 2021年流域尺度数据集(CBS-WUE, https://doi.org/10.5281/zenodo.17402779),并与FLUXNET2015观测数据进行了验证。利用这一新的数据集,进行了流域间的比较,以表征空间异质性和时间动态,同时采用多元统计和机器学习分析来确定气候、生物和土地利用驱动因素的相对贡献。结果表明,高程和植被结构是影响流域尺度水分利用效率差异的主要因素。全国平均水分利用效率为1.13 g C kg−1 H2O,流域平均水分利用效率为0.11 ~ 1.80 g C kg−1 H2O。其中,中等海拔和植被密集的流域WUE较高,而高海拔和干旱的流域WUE较低。这种综合分析强调了地形和植被在水利用效率模式形成中的主导作用,为在变化的环境条件下提高水资源效率和生态可持续性提供了科学依据。
{"title":"Elevation and Vegetation Cover Dominate Inter-Basin Water Use Efficiency Patterns in China","authors":"Jun Hu;Hongjun Su;Yiping Chen;Yuanwei Qin;Zhaohui Xue;Qian Du","doi":"10.1109/JSTARS.2025.3640403","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3640403","url":null,"abstract":"Water use efficiency (WUE) is a fundamental indicator of the balance between ecosystem carbon assimilation and water consumption. However, its spatial variability and dominant environmental drivers across China's river basins remain unclear, posing challenges for basin-scale management. In this study, a comprehensive WUE analysis framework was established through the integration of multisource remote sensing and auxiliary datasets. In this framework, multisource vegetation, climate, topography, and land-use data were integrated to estimate WUE from the GPP-to-ET ratio, and a novel basin-scale dataset covering 25 major river basins in China from 2002 to 2021 was generated (CBS-WUE, <uri>https://doi.org/10.5281/zenodo.17402779</uri>), which was validated against FLUXNET2015 observations. With this new dataset, inter-basin comparisons were conducted to characterize spatial heterogeneity and temporal dynamics, while multivariate statistical and machine learning analyses were employed to identify the relative contributions of climatic, biotic, and land-use drivers. Results indicated that elevation and vegetation structure were the primary factors influencing basin-scale WUE differences. The national average WUE was 1.13 g C kg<sup>−1</sup> H<sub>2</sub>O, with basin-level values ranging from 0.11 to 1.80 g C kg<sup>−1</sup> H<sub>2</sub>O. Among them, higher WUE was in basins of moderate elevation and dense vegetation, and lower WUE was in high-elevation or arid basins. This integrative analysis highlights the dominant role of topography and vegetation in shaping WUE patterns and provides a scientific basis for enhancing water resource efficiency and ecological sustainability under changing environmental conditions.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"1533-1548"},"PeriodicalIF":5.3,"publicationDate":"2025-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11278658","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145778394","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TranSTD: A Wavelet-Driven Transformer-Based SAR Target Detection Framework With Adaptive Feature Enhancement and Fusion 基于自适应特征增强和融合的小波驱动变换SAR目标检测框架
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-03 DOI: 10.1109/JSTARS.2025.3639785
Bobo Xi;Jiaqi Chen;Yan Huang;Jiaojiao Li;Yunsong Li;Zan Li;Xiang-Gen Xia
Target detection in Synthetic Aperture Radar (SAR) images is of great importance in civilian monitoring and military reconnaissance. However, the unique speckle noise inherent in SAR images leads to semantic information loss, while traditional convolutional neural network downsampling methods exacerbate this issue, impacting detection accuracy and robustness. Moreover, some dense target scenarios and weak scattering features of targets make it challenging to achieve sufficient feature discriminability, adding complexity to the detection task. In addition, the multiscale characteristic of SAR targets presents difficulties in balancing detection performance with computational efficiency in complex scenes. To tackle these difficulties, this article introduces a wavelet-driven transformer-based SAR target detection framework called TranSTD. Specifically, it incorporates the Haar wavelet dynamic downsampling and semantic preserving dynamic downsampling modules, which effectively suppress noise and preserve semantic information using techniques such as Haar wavelet denoise and input-driven dynamic pooling downsampling. Furthermore, the SAR adaptive convolution (SAC) bottleneck is proposed for enhancing the discrimination of features. To optimize performance and efficiency across varying scene complexities, a multiscale SAR attention fusion encoder is developed. Extensive experiments are carried out on three datasets, showing that our proposed algorithm outperforms the current state-of-the-art benchmarks in SAR target detection, offering a robust solution for the detection of targets in complex SAR scenes.
合成孔径雷达(SAR)图像中的目标检测在民用监控和军事侦察中具有重要意义。然而,SAR图像固有的斑点噪声导致语义信息丢失,而传统的卷积神经网络降采样方法加剧了这一问题,影响了检测的准确性和鲁棒性。此外,一些密集的目标场景和目标较弱的散射特性,使得目标特征难以达到足够的可分辨性,增加了检测任务的复杂性。此外,SAR目标的多尺度特性使得在复杂场景下难以平衡检测性能和计算效率。为了解决这些困难,本文介绍了一种基于小波驱动变压器的SAR目标检测框架TranSTD。具体来说,它结合了Haar小波动态下采样和语义保持动态下采样模块,使用Haar小波去噪和输入驱动的动态池下采样等技术有效地抑制噪声并保留语义信息。在此基础上,提出了SAR自适应卷积(SAC)瓶颈来增强特征识别。为了优化不同场景下的性能和效率,研制了一种多尺度SAR注意力融合编码器。在三个数据集上进行了大量的实验,表明我们提出的算法优于当前最先进的SAR目标检测基准,为复杂SAR场景中的目标检测提供了强大的解决方案。
{"title":"TranSTD: A Wavelet-Driven Transformer-Based SAR Target Detection Framework With Adaptive Feature Enhancement and Fusion","authors":"Bobo Xi;Jiaqi Chen;Yan Huang;Jiaojiao Li;Yunsong Li;Zan Li;Xiang-Gen Xia","doi":"10.1109/JSTARS.2025.3639785","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3639785","url":null,"abstract":"Target detection in Synthetic Aperture Radar (SAR) images is of great importance in civilian monitoring and military reconnaissance. However, the unique speckle noise inherent in SAR images leads to semantic information loss, while traditional convolutional neural network downsampling methods exacerbate this issue, impacting detection accuracy and robustness. Moreover, some dense target scenarios and weak scattering features of targets make it challenging to achieve sufficient feature discriminability, adding complexity to the detection task. In addition, the multiscale characteristic of SAR targets presents difficulties in balancing detection performance with computational efficiency in complex scenes. To tackle these difficulties, this article introduces a wavelet-driven transformer-based SAR target detection framework called TranSTD. Specifically, it incorporates the Haar wavelet dynamic downsampling and semantic preserving dynamic downsampling modules, which effectively suppress noise and preserve semantic information using techniques such as Haar wavelet denoise and input-driven dynamic pooling downsampling. Furthermore, the SAR adaptive convolution (SAC) bottleneck is proposed for enhancing the discrimination of features. To optimize performance and efficiency across varying scene complexities, a multiscale SAR attention fusion encoder is developed. Extensive experiments are carried out on three datasets, showing that our proposed algorithm outperforms the current state-of-the-art benchmarks in SAR target detection, offering a robust solution for the detection of targets in complex SAR scenes.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"1197-1211"},"PeriodicalIF":5.3,"publicationDate":"2025-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11275702","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145778392","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Dual-Branch EfficientNetV2-S-Based Method for Marine Oil Spill Detection Using Multisource Satellite Data Fusion 基于多源卫星数据融合的双分支高效netv2海洋溢油检测方法
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-03 DOI: 10.1109/JSTARS.2025.3639503
Yong Wan;Liyan Peng;Rui Zhang;Ruyue Zhang;Haowen Wang
As one of the most severe forms of pollution, oil spills pose significant threats to the marine environment. Synthetic aperture radar (SAR), an active microwave remote sensing technology, enables sea surface monitoring under all weather and lighting conditions and provides high spatial resolution. It has been widely used in the field of marine oil spill detection. However, other natural phenomena, such as low wind regions and biogenic oil films, can also produce dark spot features in SAR imagery that resemble oil spills, leading to false alarms. Global navigation satellite system-reflectometry (GNSS-R), as an emerging remote sensing technique for ocean observation, offers distinct advantages, including high temporal resolution and multisource observation capabilities. By combining SAR backscattering coefficients with GNSS-R delay doppler map, it becomes possible to characterize the impact of oil spills on sea surface roughness from both backscattering and forward-scattering perspectives. This joint approach enables more accurate oil spill detection and has the potential to reduce the false alarms. Nevertheless, limited measured data for multisource remote sensing oil spill detection hinders robust multisensor fusion model development. To address this, this study proposes a synchronized data generation method, creating a joint SAR and GNSS-R oil spill dataset, and on this basis, a dual-branch EfficientNetV2-S architecture is adopted to build a multisource satellite oil spill data fusion model, which is applied to offshore oil spill detection. According to experimental results, the suggested model detects oil spills with an accuracy of 94.97%. Compared with SAR-only detection models, the false alarm rate is reduced by 3.6%, demonstrating that the dual-payload approach effectively lowers the rate of false detections in marine oil spill monitoring.
石油泄漏作为最严重的污染形式之一,对海洋环境构成了重大威胁。合成孔径雷达(SAR)是一种主动微波遥感技术,能够在所有天气和光照条件下进行海面监测,并提供高空间分辨率。它在海洋溢油检测领域得到了广泛的应用。然而,其他自然现象,如低风区和生物油膜,也会在SAR图像中产生类似石油泄漏的黑点特征,从而导致误报。全球导航卫星系统反射测量(GNSS-R)作为一种新兴的海洋遥感观测技术,具有高时间分辨率和多源观测能力等明显优势。通过将SAR后向散射系数与GNSS-R延迟多普勒图相结合,可以从后向散射和前向散射两个角度描述石油泄漏对海面粗糙度的影响。这种联合方法可以更准确地检测溢油,并有可能减少误报。然而,多源遥感溢油探测的实测数据有限,阻碍了多传感器融合模型的发展。为此,本研究提出了同步数据生成方法,创建SAR和GNSS-R联合溢油数据集,并在此基础上采用双分支的EfficientNetV2-S架构构建多源卫星溢油数据融合模型,应用于海上溢油检测。实验结果表明,该模型的溢油检测准确率为94.97%。与单纯的sar检测模型相比,虚警率降低了3.6%,表明双载荷方法有效降低了海洋溢油监测中的虚警率。
{"title":"A Dual-Branch EfficientNetV2-S-Based Method for Marine Oil Spill Detection Using Multisource Satellite Data Fusion","authors":"Yong Wan;Liyan Peng;Rui Zhang;Ruyue Zhang;Haowen Wang","doi":"10.1109/JSTARS.2025.3639503","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3639503","url":null,"abstract":"As one of the most severe forms of pollution, oil spills pose significant threats to the marine environment. Synthetic aperture radar (SAR), an active microwave remote sensing technology, enables sea surface monitoring under all weather and lighting conditions and provides high spatial resolution. It has been widely used in the field of marine oil spill detection. However, other natural phenomena, such as low wind regions and biogenic oil films, can also produce dark spot features in SAR imagery that resemble oil spills, leading to false alarms. Global navigation satellite system-reflectometry (GNSS-R), as an emerging remote sensing technique for ocean observation, offers distinct advantages, including high temporal resolution and multisource observation capabilities. By combining SAR backscattering coefficients with GNSS-R delay doppler map, it becomes possible to characterize the impact of oil spills on sea surface roughness from both backscattering and forward-scattering perspectives. This joint approach enables more accurate oil spill detection and has the potential to reduce the false alarms. Nevertheless, limited measured data for multisource remote sensing oil spill detection hinders robust multisensor fusion model development. To address this, this study proposes a synchronized data generation method, creating a joint SAR and GNSS-R oil spill dataset, and on this basis, a dual-branch EfficientNetV2-S architecture is adopted to build a multisource satellite oil spill data fusion model, which is applied to offshore oil spill detection. According to experimental results, the suggested model detects oil spills with an accuracy of 94.97%. Compared with SAR-only detection models, the false alarm rate is reduced by 3.6%, demonstrating that the dual-payload approach effectively lowers the rate of false detections in marine oil spill monitoring.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"1549-1566"},"PeriodicalIF":5.3,"publicationDate":"2025-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11275680","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145778390","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AF2-MSA Net: Attention-Fusion Focused Multiscale Architecture Network for Remote Sensing Scene Classification AF2-MSA网:面向遥感场景分类的关注融合多尺度体系结构网络
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-03 DOI: 10.1109/JSTARS.2025.3639670
Cuiping Shi;Yimin Wang;Liguo Wang
With the rapid development of deep learning technology, significant progress has been made in the field of remote sensing (RS) scene image classification. However, the large intraclass distance and high interclass similarity still pose significant challenges for RS scene classification. In addition, there are multiscale targets in RS images, which make significant differences in target characteristics. To overcome the above limitations, this article proposes a novel attention-fusion focused multiscale architecture network (AF2-MSA Net). First, a multilevel feature extraction module (MFEM) was designed to extract semantic and detail information at different scales from RS images. Subsequently, an intricately designed global context recalibration module (GCRM) was embedded into MFEM, and the features at each level were enhanced through a global context recalibration mechanism, enabling the model to dynamically focus on key semantic regions and important contextual information. Next, an axis-aligned feature harmonization module (AAFHM) was constructed to fuse multiscale features from adjacent stages layer by layer. This module combines attention mechanisms from both channel and spatial branches to adaptively coordinate and fuse multiscale contextual information, achieving deep collaborative optimization of different scale features. Finally, the GCRM and AAFHM are integrated into a unified framework called AF2-MSA Net to achieve collaborative optimization of global semantics and multiscale discriminative features. Extensive experiments on three commonly used datasets have shown that the proposed AF2-MSA Net outperforms some state-of-the-art methods in RS image scene classification tasks.
随着深度学习技术的快速发展,遥感场景图像分类领域取得了重大进展。然而,大的类内距离和高的类间相似度仍然给遥感场景分类带来了很大的挑战。此外,RS图像中存在多尺度目标,这使得目标特征存在显著差异。为了克服上述局限性,本文提出了一种新的以注意力融合为中心的多尺度架构网络(AF2-MSA网)。首先,设计多级特征提取模块,从遥感图像中提取不同尺度的语义信息和细节信息;随后,将设计复杂的全局上下文再校准模块(GCRM)嵌入到MFEM中,并通过全局上下文再校准机制增强各层特征,使模型能够动态关注关键语义区域和重要上下文信息。然后,构建一个轴向特征协调模块(AAFHM),逐层融合相邻阶段的多尺度特征;该模块结合渠道分支和空间分支的注意机制,自适应协调融合多尺度上下文信息,实现不同尺度特征的深度协同优化。最后,将GCRM和AAFHM集成到AF2-MSA Net的统一框架中,实现全局语义和多尺度判别特征的协同优化。在三个常用数据集上的大量实验表明,所提出的AF2-MSA网在RS图像场景分类任务中优于一些最先进的方法。
{"title":"AF2-MSA Net: Attention-Fusion Focused Multiscale Architecture Network for Remote Sensing Scene Classification","authors":"Cuiping Shi;Yimin Wang;Liguo Wang","doi":"10.1109/JSTARS.2025.3639670","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3639670","url":null,"abstract":"With the rapid development of deep learning technology, significant progress has been made in the field of remote sensing (RS) scene image classification. However, the large intraclass distance and high interclass similarity still pose significant challenges for RS scene classification. In addition, there are multiscale targets in RS images, which make significant differences in target characteristics. To overcome the above limitations, this article proposes a novel attention-fusion focused multiscale architecture network (AF<sup>2</sup>-MSA Net). First, a multilevel feature extraction module (MFEM) was designed to extract semantic and detail information at different scales from RS images. Subsequently, an intricately designed global context recalibration module (GCRM) was embedded into MFEM, and the features at each level were enhanced through a global context recalibration mechanism, enabling the model to dynamically focus on key semantic regions and important contextual information. Next, an axis-aligned feature harmonization module (AAFHM) was constructed to fuse multiscale features from adjacent stages layer by layer. This module combines attention mechanisms from both channel and spatial branches to adaptively coordinate and fuse multiscale contextual information, achieving deep collaborative optimization of different scale features. Finally, the GCRM and AAFHM are integrated into a unified framework called AF<sup>2</sup>-MSA Net to achieve collaborative optimization of global semantics and multiscale discriminative features. Extensive experiments on three commonly used datasets have shown that the proposed AF<sup>2</sup>-MSA Net outperforms some state-of-the-art methods in RS image scene classification tasks.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"1150-1164"},"PeriodicalIF":5.3,"publicationDate":"2025-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11275695","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145729382","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ROFANet: Residual Offset-Driven Feature Alignment Network for Unaligned Remote Sensing Image Change Detection 基于残差驱动的遥感图像变化检测特征对齐网络
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-03 DOI: 10.1109/JSTARS.2025.3639607
Guoqing Wang;He Chen;Wenchao Liu;Tianyu Wei;Panzhe Gu;Jue Wang
At present, most remote sensing change detection methods are applicable to bitemporal image alignment scenarios, that is, assuming that the pixel pairs of bitemporal images are spatially registered. Detection accuracy is highly sensitive to the alignment accuracy of the image pairs. In practical applications, obtaining well-registered image pairs is often challenging. Currently, the approach of aligning first and then detecting is both inefficient and expensive. Explicitly integrating image alignment and change detection into a framework is an effective solution. However, offset information is difficult to be reflected in the high-level features of the image, it is hard to predict accurate image offsets, and it is also difficult to correct the spatial relationship of land covers in the image using the offset. To overcome the above problems, we propose a residual offset-driven feature alignment network (ROFANet). ROFANet combines two innovative methods: residual offset prediction (ROP) and dual-branch feature correction (DFC). ROP utilizes multilevel features to achieve offset prediction from coarse to fine granularity, effectively enhancing the model's predictive ability for image offsets. DFC has established two branches: image correction and feature correction, which respectively correct distorted images and distorted features. By optimizing the spatial relationship representation of land covers, the model's change detection ability under unaligned image conditions has been enhanced. Extensive experiments conducted on three publicly available change detection datasets demonstrate that the proposed ROFANet achieves outstanding detection performance in unaligned image scenarios.
目前,大多数遥感变化检测方法都适用于双时图像对齐场景,即假定双时图像的像素对是空间配准的。检测精度对图像对的对准精度高度敏感。在实际应用中,获得配准良好的图像对往往具有挑战性。目前,先对准再检测的方法既低效又昂贵。显式地将图像对齐和变化检测集成到框架中是一种有效的解决方案。然而,偏移量信息难以反映在图像的高层特征上,难以预测准确的图像偏移量,也难以利用偏移量对图像中土地覆盖的空间关系进行校正。为了克服上述问题,我们提出了残差驱动特征对齐网络(ROFANet)。ROFANet结合了两种创新方法:残差预测(ROP)和双支路特征校正(DFC)。ROP利用多级特征实现了从粗到细粒度的偏移预测,有效增强了模型对图像偏移的预测能力。DFC建立了图像校正和特征校正两个分支,分别对畸变图像和畸变特征进行校正。通过优化土地覆盖的空间关系表示,增强了模型在非对齐图像条件下的变化检测能力。在三个公开可用的变化检测数据集上进行的大量实验表明,所提出的ROFANet在未对齐图像场景下取得了出色的检测性能。
{"title":"ROFANet: Residual Offset-Driven Feature Alignment Network for Unaligned Remote Sensing Image Change Detection","authors":"Guoqing Wang;He Chen;Wenchao Liu;Tianyu Wei;Panzhe Gu;Jue Wang","doi":"10.1109/JSTARS.2025.3639607","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3639607","url":null,"abstract":"At present, most remote sensing change detection methods are applicable to bitemporal image alignment scenarios, that is, assuming that the pixel pairs of bitemporal images are spatially registered. Detection accuracy is highly sensitive to the alignment accuracy of the image pairs. In practical applications, obtaining well-registered image pairs is often challenging. Currently, the approach of aligning first and then detecting is both inefficient and expensive. Explicitly integrating image alignment and change detection into a framework is an effective solution. However, offset information is difficult to be reflected in the high-level features of the image, it is hard to predict accurate image offsets, and it is also difficult to correct the spatial relationship of land covers in the image using the offset. To overcome the above problems, we propose a residual offset-driven feature alignment network (ROFANet). ROFANet combines two innovative methods: residual offset prediction (ROP) and dual-branch feature correction (DFC). ROP utilizes multilevel features to achieve offset prediction from coarse to fine granularity, effectively enhancing the model's predictive ability for image offsets. DFC has established two branches: image correction and feature correction, which respectively correct distorted images and distorted features. By optimizing the spatial relationship representation of land covers, the model's change detection ability under unaligned image conditions has been enhanced. Extensive experiments conducted on three publicly available change detection datasets demonstrate that the proposed ROFANet achieves outstanding detection performance in unaligned image scenarios.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"1305-1320"},"PeriodicalIF":5.3,"publicationDate":"2025-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11275653","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145778356","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SwinCTC: Efficient Network for Superresolution Reconstruction of Remote Sensing Images Based on Nonlocal Feature Enhancement by Sliding Window Mechanism 基于滑动窗口机制的非局部特征增强的遥感图像超分辨率重构网络
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-02 DOI: 10.1109/JSTARS.2025.3639298
Zhikai Wang;Yuehao Xiao;Zhumu Fu;Mengyang Li;Na Li
In transformer-based superresolution reconstruction tasks for remote sensing images, the window attention mechanism has become a key method for reducing the secondary complexity of traditional self-attention. However, the window self-attention mechanism still requires a significant amount of computational resources, especially when processing large remote sensing images. To address this issue, we propose a convolutional block structure based on the sliding window mechanism, which replaces the traditional window/sliding self-attention and drastically reduces computational complexity. It comprises a residual channel enhanced attention (RCEA) module and group convolution, which enhances the efficiency of group convolution by dynamically refining the channel weights through RCEA. In addition, the CTC-Block further refines the window feature representation by introducing a spatial attention enhancement module that focuses on key spatial details and selectively emphasizes the information regions within each window. Finally, a convolution-based feedforward network is introduced to bolster the network’s capacity to model high-frequency information in images. The experimental results demonstrate that the proposed method outperforms other classical remote sensing image superresolution reconstruction models in terms of peak signal-to-noise ratio and structural similarity evaluation metrics on the NWPU-RESISC45 and NWPU-VHR datasets. Compared with the baseline model SwinIR, the number of parameters is reduced by 39.4%, the number of floating-point operations is reduced by 42.4%, and the average inference speed reaches 22.86 ms while maintaining performance.
在基于变压器的遥感图像超分辨率重建任务中,窗口注意机制已成为降低传统自注意二次复杂度的关键方法。但是,窗口自关注机制仍然需要大量的计算资源,特别是在处理大型遥感图像时。为了解决这个问题,我们提出了一种基于滑动窗口机制的卷积块结构,取代了传统的窗口/滑动自关注,大大降低了计算复杂度。它包括残差信道增强注意(RCEA)模块和群卷积模块,通过RCEA动态细化信道权值,提高了群卷积的效率。此外,CTC-Block通过引入空间注意力增强模块进一步细化窗口特征表示,该模块专注于关键空间细节,并有选择地强调每个窗口内的信息区域。最后,引入了基于卷积的前馈网络,以增强网络对图像中高频信息的建模能力。实验结果表明,该方法在NWPU-RESISC45和NWPU-VHR数据集上的峰值信噪比和结构相似性评价指标均优于其他经典遥感图像超分辨率重建模型。与基线模型SwinIR相比,参数数量减少了39.4%,浮点运算次数减少了42.4%,在保持性能的情况下平均推理速度达到22.86 ms。
{"title":"SwinCTC: Efficient Network for Superresolution Reconstruction of Remote Sensing Images Based on Nonlocal Feature Enhancement by Sliding Window Mechanism","authors":"Zhikai Wang;Yuehao Xiao;Zhumu Fu;Mengyang Li;Na Li","doi":"10.1109/JSTARS.2025.3639298","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3639298","url":null,"abstract":"In transformer-based superresolution reconstruction tasks for remote sensing images, the window attention mechanism has become a key method for reducing the secondary complexity of traditional self-attention. However, the window self-attention mechanism still requires a significant amount of computational resources, especially when processing large remote sensing images. To address this issue, we propose a convolutional block structure based on the sliding window mechanism, which replaces the traditional window/sliding self-attention and drastically reduces computational complexity. It comprises a residual channel enhanced attention (RCEA) module and group convolution, which enhances the efficiency of group convolution by dynamically refining the channel weights through RCEA. In addition, the CTC-Block further refines the window feature representation by introducing a spatial attention enhancement module that focuses on key spatial details and selectively emphasizes the information regions within each window. Finally, a convolution-based feedforward network is introduced to bolster the network’s capacity to model high-frequency information in images. The experimental results demonstrate that the proposed method outperforms other classical remote sensing image superresolution reconstruction models in terms of peak signal-to-noise ratio and structural similarity evaluation metrics on the NWPU-RESISC45 and NWPU-VHR datasets. Compared with the baseline model SwinIR, the number of parameters is reduced by 39.4%, the number of floating-point operations is reduced by 42.4%, and the average inference speed reaches 22.86 ms while maintaining performance.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"1240-1258"},"PeriodicalIF":5.3,"publicationDate":"2025-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11271778","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145778403","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing UAV Search Under Occlusion Using Next Best View Planning 利用次优视点规划增强无人机在遮挡下的搜索
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-01 DOI: 10.1109/JSTARS.2025.3638881
Sigrid Helene Strand;Thomas Wiedemann;Bram Burczek;Dmitriy Shutin
Search and rescue missions are often critical following sudden natural disasters or in high-risk environmental situations. The most challenging search and rescue missions involve difficult-to-access terrains, such as dense forests with high occlusion. Deploying uncrewed aerial vehicles for exploration can significantly enhance search effectiveness, facilitate access to challenging environments, and reduce search time. However, in dense forests, the effectiveness of uncrewed aerial vehicles depends on their ability to capture clear views of the ground, necessitating a robust search strategy to optimize camera positioning and perspective. This work presents an optimized planning strategy and an efficient algorithm for the next best view problem in occluded environments. Two novel optimization heuristics, a geometry heuristic, and a visibility heuristic, are proposed to enhance search performance by selecting optimal camera viewpoints. Comparative evaluations in both simulated and real-world settings reveal that the visibility heuristic achieves greater performance, identifying over 90% of hidden objects in simulated forests and offering 10% better detection rates than the geometry heuristic. In addition, real-world experiments demonstrate that the visibility heuristic provides better coverage under the canopy, highlighting its potential for improving search and rescue missions in occluded environments.
在突发自然灾害或高风险环境情况下,搜索和救援任务往往至关重要。最具挑战性的搜索和救援任务涉及难以进入的地形,例如高遮挡的茂密森林。部署无人驾驶飞行器进行勘探可以显著提高搜索效率,便于进入具有挑战性的环境,并缩短搜索时间。然而,在茂密的森林中,无人驾驶飞行器的有效性取决于它们捕捉地面清晰视图的能力,这就需要一个强大的搜索策略来优化相机的定位和视角。本文提出了一种优化的规划策略和一种有效的算法来解决闭塞环境下的次优视图问题。提出了两种新的优化启发式算法,即几何启发式算法和可视性启发式算法,通过选择最佳的摄像机视点来提高搜索性能。在模拟和现实环境中的比较评估表明,可见性启发式方法取得了更好的性能,识别了模拟森林中90%以上的隐藏物体,并且提供了比几何启发式方法高10%的检测率。此外,现实世界的实验表明,可见度启发式方法在树冠下提供了更好的覆盖范围,突出了其在闭塞环境中改善搜索和救援任务的潜力。
{"title":"Enhancing UAV Search Under Occlusion Using Next Best View Planning","authors":"Sigrid Helene Strand;Thomas Wiedemann;Bram Burczek;Dmitriy Shutin","doi":"10.1109/JSTARS.2025.3638881","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3638881","url":null,"abstract":"Search and rescue missions are often critical following sudden natural disasters or in high-risk environmental situations. The most challenging search and rescue missions involve difficult-to-access terrains, such as dense forests with high occlusion. Deploying uncrewed aerial vehicles for exploration can significantly enhance search effectiveness, facilitate access to challenging environments, and reduce search time. However, in dense forests, the effectiveness of uncrewed aerial vehicles depends on their ability to capture clear views of the ground, necessitating a robust search strategy to optimize camera positioning and perspective. This work presents an optimized planning strategy and an efficient algorithm for the next best view problem in occluded environments. Two novel optimization heuristics, a geometry heuristic, and a visibility heuristic, are proposed to enhance search performance by selecting optimal camera viewpoints. Comparative evaluations in both simulated and real-world settings reveal that the visibility heuristic achieves greater performance, identifying over 90% of hidden objects in simulated forests and offering 10% better detection rates than the geometry heuristic. In addition, real-world experiments demonstrate that the visibility heuristic provides better coverage under the canopy, highlighting its potential for improving search and rescue missions in occluded environments.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"1085-1096"},"PeriodicalIF":5.3,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11271526","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145729426","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LSDFormer: Lightweight SAR Ship Detection Enhanced With Efficient Multiattention and Structural Reparameterization LSDFormer:基于高效多注意力和结构重参数化的轻型SAR舰船检测
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-01 DOI: 10.1109/JSTARS.2025.3639164
Rui Jiang;Hang Shi;Jiahong Ni;Jiatao Li;Yi Feng;Xinqiang Chen;Yinlin Li
Ship detection in synthetic aperture radar (SAR) images faces challenges such as strong background interference, varying ship appearance, and distribution and high real-time requirements. Although attention-based deep learning methods dominate this field, the design of lightweight models with efficient attention mechanisms capable of addressing the aforementioned challenges remains underexplored. To address this issue, we propose a lightweight SAR ship detection model named LSDFormer, which is built upon the MetaFormer architecture and consists of an efficient multiattention-enhanced backbone and neck and a structural reparameterization (SR)-enhanced head. We employ two lightweight modules for the backbone and neck: a PoolFormer-based feature extraction module with efficient channel modulation attention is proposed to enhance ship features and suppress background interference, and a downsampling module using efficient channel aggregation attention and group convolutions is introduced to enrich ship features. The position-sensitive attention from YOLOv11 is also introduced to handle variations in ship appearance and distribution. These three attentions are integrated into an efficient multiattention mechanism. Furthermore, an SR-based detection branch is proposed for the head of LSDFormer, which enhances ship features while reducing model complexity. Extensive experiments on SSDD and HRSID datasets demonstrate the superiority and effectiveness of LSDFormer, achieving AP50 of $mathbf {98.5pm 0.4%}$ and $mathbf {92.8pm 0.2%}$, respectively, with only $mathbf {1.5}$ M parameters and $mathbf {4.1}$ GFLOPs. The average processing time per image is $mathbf {4.9}$ ms on SSDD and $mathbf {4.2}$ ms on HRSID, confirming its real-time performance.
合成孔径雷达(SAR)图像中的船舶检测面临着背景干扰强、船舶外观分布多变、实时性要求高等挑战。尽管基于注意力的深度学习方法在这一领域占据主导地位,但具有有效注意力机制的轻量级模型的设计仍未得到充分探索。为了解决这个问题,我们提出了一种名为LSDFormer的轻型SAR船舶检测模型,该模型建立在MetaFormer架构的基础上,由高效的多注意力增强的骨干和颈部以及结构重新参数化(SR)增强的头部组成。我们在主干和颈部采用了两个轻量级模块:提出了基于poolformer的特征提取模块,利用高效的信道调制注意力增强船舶特征,抑制背景干扰;引入了基于高效信道聚合注意力和群卷积的下采样模块,丰富船舶特征。YOLOv11的位置敏感注意力也被引入来处理船舶外观和分布的变化。这三种注意力被整合成一个有效的多注意机制。在此基础上,提出了一种基于sr的LSDFormer头部检测分支,在增强船舶特征的同时降低了模型复杂度。在SSDD和HRSID数据集上的大量实验证明了LSDFormer的优越性和有效性,分别实现了$mathbf {98.5pm 0.4%}$和$mathbf {92.8pm 0.2%}$的AP50,仅使用$mathbf {1.5}$ M参数和$mathbf {4.1}$ GFLOPs。每张图像的平均处理时间在SSDD上为$mathbf {4.9}$ ms,在HRSID上为$mathbf {4.2}$ ms,证实了其实时性。
{"title":"LSDFormer: Lightweight SAR Ship Detection Enhanced With Efficient Multiattention and Structural Reparameterization","authors":"Rui Jiang;Hang Shi;Jiahong Ni;Jiatao Li;Yi Feng;Xinqiang Chen;Yinlin Li","doi":"10.1109/JSTARS.2025.3639164","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3639164","url":null,"abstract":"Ship detection in synthetic aperture radar (SAR) images faces challenges such as strong background interference, varying ship appearance, and distribution and high real-time requirements. Although attention-based deep learning methods dominate this field, the design of lightweight models with efficient attention mechanisms capable of addressing the aforementioned challenges remains underexplored. To address this issue, we propose a lightweight SAR ship detection model named LSDFormer, which is built upon the MetaFormer architecture and consists of an efficient multiattention-enhanced backbone and neck and a structural reparameterization (SR)-enhanced head. We employ two lightweight modules for the backbone and neck: a PoolFormer-based feature extraction module with efficient channel modulation attention is proposed to enhance ship features and suppress background interference, and a downsampling module using efficient channel aggregation attention and group convolutions is introduced to enrich ship features. The position-sensitive attention from YOLOv11 is also introduced to handle variations in ship appearance and distribution. These three attentions are integrated into an efficient multiattention mechanism. Furthermore, an SR-based detection branch is proposed for the head of LSDFormer, which enhances ship features while reducing model complexity. Extensive experiments on SSDD and HRSID datasets demonstrate the superiority and effectiveness of LSDFormer, achieving AP50 of <inline-formula><tex-math>$mathbf {98.5pm 0.4%}$</tex-math></inline-formula> and <inline-formula><tex-math>$mathbf {92.8pm 0.2%}$</tex-math></inline-formula>, respectively, with only <inline-formula><tex-math>$mathbf {1.5}$</tex-math></inline-formula> M parameters and <inline-formula><tex-math>$mathbf {4.1}$</tex-math></inline-formula> GFLOPs. The average processing time per image is <inline-formula><tex-math>$mathbf {4.9}$</tex-math></inline-formula> ms on SSDD and <inline-formula><tex-math>$mathbf {4.2}$</tex-math></inline-formula> ms on HRSID, confirming its real-time performance.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"1359-1377"},"PeriodicalIF":5.3,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11271640","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145778396","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Anomaly Detection of InSAR Time-Series Displacements Based on Generative Adversarial Network 基于生成对抗网络的InSAR时间序列位移异常检测
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-01 DOI: 10.1109/JSTARS.2025.3639018
Siting Xiong;Zhichao Deng;Bochen Zhang;Jiayuan Zhang;Chisheng Wang
Interferometric synthetic aperture radar (InSAR) is a widely applied and highly efficient tool for monitoring large-scale and long-term ground displacements. In most applications, evaluating InSAR results on land displacements primarily depends on the displacement rate/velocity. However, the full exploitation of time-series information is becoming increasingly important as the increasing temporal coverage of the SAR dataset can lead to the composition of different sequences of one target. Effectively and efficiently detecting abnormal timestamps from a full-time series of InSAR displacement results is critical in InSAR postanalysis. To this end, we propose a novel approach to automatically detect anomalous timestamps in InSAR-derived time-series displacements based on improved time-series anomaly detection using generative adversarial networks (TadGAN). The improved TadGAN generates a normal time-series displacement compared with the InSAR-derived time-series displacement to obtain anomaly scores for each timestamp. Based on these anomaly scores, the ratio of anomalous timestamps and maximum anomaly scores was calculated to assess the risk levels, and the anomalous sequences were classified into abrupt and trend types using a simple convolutional neural network integrated with the attention mechanism. The proposed method was applied to the InSAR-derived ground displacements of the Hong Kong–Zhuhai–Macao Bridge. The results show that the proposed method successfully detects the start and end of anomalous sequences and produces anomaly maps that are more accurate than displacement rate maps. The trained model can also be applied directly to other regions, as validated by the InSAR results of the Kowloon Peninsula in Hong Kong.
干涉合成孔径雷达(InSAR)是一种应用广泛、高效的大规模、长期地面位移监测工具。在大多数应用中,评估InSAR对陆地位移的结果主要取决于位移速率/速度。然而,时间序列信息的充分利用变得越来越重要,因为SAR数据集的时间覆盖范围越来越大,可能导致同一目标的不同序列组成。有效、高效地从InSAR位移结果的全时序列中检测异常时间戳是InSAR后期分析的关键。为此,我们提出了一种基于改进的生成对抗网络(TadGAN)时间序列异常检测的方法来自动检测insar衍生时间序列位移中的异常时间戳。改进的TadGAN生成一个正常的时间序列位移,与insar导出的时间序列位移进行比较,以获得每个时间戳的异常分数。基于这些异常得分,计算异常时间戳与最大异常得分的比值,评估异常序列的风险等级,并利用结合注意机制的简单卷积神经网络将异常序列划分为突变型和趋势型。将该方法应用于港珠澳大桥insar反演的地面位移。结果表明,该方法能够有效地检测异常序列的起止点,生成的异常图比位移率图更准确。训练后的模式亦可直接应用于其他地区,香港九龙半岛的InSAR结果证实了这一点。
{"title":"Anomaly Detection of InSAR Time-Series Displacements Based on Generative Adversarial Network","authors":"Siting Xiong;Zhichao Deng;Bochen Zhang;Jiayuan Zhang;Chisheng Wang","doi":"10.1109/JSTARS.2025.3639018","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3639018","url":null,"abstract":"Interferometric synthetic aperture radar (InSAR) is a widely applied and highly efficient tool for monitoring large-scale and long-term ground displacements. In most applications, evaluating InSAR results on land displacements primarily depends on the displacement rate/velocity. However, the full exploitation of time-series information is becoming increasingly important as the increasing temporal coverage of the SAR dataset can lead to the composition of different sequences of one target. Effectively and efficiently detecting abnormal timestamps from a full-time series of InSAR displacement results is critical in InSAR postanalysis. To this end, we propose a novel approach to automatically detect anomalous timestamps in InSAR-derived time-series displacements based on improved time-series anomaly detection using generative adversarial networks (TadGAN). The improved TadGAN generates a normal time-series displacement compared with the InSAR-derived time-series displacement to obtain anomaly scores for each timestamp. Based on these anomaly scores, the ratio of anomalous timestamps and maximum anomaly scores was calculated to assess the risk levels, and the anomalous sequences were classified into abrupt and trend types using a simple convolutional neural network integrated with the attention mechanism. The proposed method was applied to the InSAR-derived ground displacements of the Hong Kong–Zhuhai–Macao Bridge. The results show that the proposed method successfully detects the start and end of anomalous sequences and produces anomaly maps that are more accurate than displacement rate maps. The trained model can also be applied directly to other regions, as validated by the InSAR results of the Kowloon Peninsula in Hong Kong.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"1227-1239"},"PeriodicalIF":5.3,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11271588","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145778400","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Evaluation of Soil Temperature Predictions Based on the Long Short-Term Memory Model and Remote Sensing Data 基于长短期记忆模型和遥感数据的土壤温度预测评价
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-01 DOI: 10.1109/JSTARS.2025.3638765
Zihan Yuan;Jilin Gu;Yaoqi Lu;Yuwei Li
Soil temperature is a key variable in several fields of Earth science, and accurate predictions of soil temperature at different depths are of great significance for scientific research and agricultural production. Soil temperature data observations by meteorological stations are categorized as discrete and discontinuous, and moderate resolution imaging spectroradiometer (MODIS) remotely sensed data are used to perform routine soil temperature predictions on a large scale. In this study, data from three MODIS products, namely, normalized vegetation index, atmospheric precipitable water, and surface temperature, and daily average soil temperature measurements at depths of 40, 100, and 200 cm from the ground surface in Liaoning Province from 2017 to 2021 were used to establish a soil temperature prediction model based on the long short-term memory (LSTM) model. To improve the prediction accuracy and stability, an optimized LSTM model was established to perform comparative predictions of soil temperature concentrations based on the LSTM model, and it considered the hysteresis factor of soil temperature relative to the surface temperature. The LSTM soil temperature prediction models established based on the fusion of remote sensing data (NDVI, PWV, and LST) and soil temperature data at 40, 100, and 200 cm from the surface and the optimized LSTM models that considered hysteresis obtained R2 values of 0.86, 0.81, 0.69, and 0.90, 0.91, 0.88, respectively, with RMSE values of 0.30, 3.41, 3.74 °C, and 0.30, 3.41, 3.74 °C. Moreover, the SDRMSE of the optimized model considering hysteresis decreased compared to that of the LSTM. The LSTM models before and after optimization can achieve long-term daily temperature prediction of inter-annual soil temperature, although the prediction model that considers the hysteresis factor had a better fit and stability. Thus, the model considering hysteresis is more advantageous for obtaining accurate predictions of spatially continuous multidepth soil temperatures.
土壤温度是地球科学多个领域的关键变量,准确预测不同深度土壤温度对科学研究和农业生产具有重要意义。气象站的土壤温度观测数据分为离散型和不连续型,常规的大尺度土壤温度预测主要采用中分辨率成像光谱仪(MODIS)遥感数据。利用2017 - 2021年辽宁省标准化植被指数、大气可降水量和地表温度3种MODIS产品以及距地表40、100和200 cm的日平均土壤温度数据,建立了基于LSTM模型的土壤温度预测模型。为提高预测精度和稳定性,在考虑土壤温度相对于地表温度滞后因子的基础上,建立了优化的LSTM模型对土壤温度浓度进行对比预测。基于遥感数据(NDVI、PWV和LST)与地表40、100和200 cm土壤温度数据融合建立的LSTM土壤温度预测模型以及考虑滞后的优化LSTM模型的R2分别为0.86、0.81、0.69和0.90、0.91、0.88,RMSE分别为0.30、3.41、3.74°C和0.30、3.41、3.74°C。考虑迟滞的优化模型的SDRMSE比LSTM的减小。优化前后的LSTM模型可以实现年际土壤温度的长期日温度预测,但考虑滞后因素的预测模型拟合性和稳定性更好。因此,考虑迟滞的模型更有利于获得空间连续多深度土壤温度的准确预测。
{"title":"An Evaluation of Soil Temperature Predictions Based on the Long Short-Term Memory Model and Remote Sensing Data","authors":"Zihan Yuan;Jilin Gu;Yaoqi Lu;Yuwei Li","doi":"10.1109/JSTARS.2025.3638765","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3638765","url":null,"abstract":"Soil temperature is a key variable in several fields of Earth science, and accurate predictions of soil temperature at different depths are of great significance for scientific research and agricultural production. Soil temperature data observations by meteorological stations are categorized as discrete and discontinuous, and moderate resolution imaging spectroradiometer (MODIS) remotely sensed data are used to perform routine soil temperature predictions on a large scale. In this study, data from three MODIS products, namely, normalized vegetation index, atmospheric precipitable water, and surface temperature, and daily average soil temperature measurements at depths of 40, 100, and 200 cm from the ground surface in Liaoning Province from 2017 to 2021 were used to establish a soil temperature prediction model based on the long short-term memory (LSTM) model. To improve the prediction accuracy and stability, an optimized LSTM model was established to perform comparative predictions of soil temperature concentrations based on the LSTM model, and it considered the hysteresis factor of soil temperature relative to the surface temperature. The LSTM soil temperature prediction models established based on the fusion of remote sensing data (NDVI, PWV, and LST) and soil temperature data at 40, 100, and 200 cm from the surface and the optimized LSTM models that considered hysteresis obtained <italic>R</i><sup>2</sup> values of 0.86, 0.81, 0.69, and 0.90, 0.91, 0.88, respectively, with RMSE values of 0.30, 3.41, 3.74 °C, and 0.30, 3.41, 3.74 °C. Moreover, the SD<sub>RMSE</sub> of the optimized model considering hysteresis decreased compared to that of the LSTM. The LSTM models before and after optimization can achieve long-term daily temperature prediction of inter-annual soil temperature, although the prediction model that considers the hysteresis factor had a better fit and stability. Thus, the model considering hysteresis is more advantageous for obtaining accurate predictions of spatially continuous multidepth soil temperatures.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"1212-1226"},"PeriodicalIF":5.3,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11271761","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145778340","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1