首页 > 最新文献

IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing最新文献

英文 中文
SRNet: A Semantic Reasoning Network for Small Weak Object Detection in Remote Sensing Images SRNet:遥感图像中弱小目标检测的语义推理网络
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-17 DOI: 10.1109/JSTARS.2025.3645623
Zheng Li;Hao Feng;Dongdong Xu;Tianqi Zhao;Boxiao Wang;Yongcheng Wang
Small weak object detection (SWOD) is a significant but neglected task in remote sensing image interpretation. Due to limitations in imaging resolution and inherent characteristics of the objects, detection networks struggle to effectively extract semantic features, which are crucial for object identification and recognition. In recent years, graph convolutional networks (GCNs) have been developed to handle non-Euclidean data. Through GCNs, node data are enriched via aggregation and propagation across the graph. In this article, we explore the feasibility of GCNs in semantic clue extraction to address the lack of key semantics in small weak objects. First, we propose a multihead graph reasoning learning model (MGRL) that projects initial feature representations into graph space and utilizes a two-layer multihead graph network to extract essential semantic information. Second, we introduce a foreground-background binary masking technique that roughly segments the foreground region of the image. The mask is converted into a prior prompt, which is then incorporated into the adjacency matrix, emphasizing object reasoning in MGRL. Next, we present a cross learning-based feature alignment learning module to resolve feature misalignment issues caused by spatial projection. Finally, we adopt a cross-layer semantic interaction module to facilitate cross-layer communication and aggregation of features. Extensive experiments are conducted on five remote sensing datasets: DIOR, AI-TOD, NWPU VHR-10, DOTA-v1.0, and STAR. The experimental results demonstrate the superior performance and advantages of our method.
弱小目标检测是遥感图像解译中一个重要而又被忽视的课题。由于图像分辨率的限制和物体的固有特性,检测网络很难有效地提取语义特征,而语义特征对于物体的识别至关重要。近年来,图卷积网络(GCNs)被用于处理非欧几里得数据。通过GCNs,节点数据通过在图上的聚合和传播而丰富。在本文中,我们探索了GCNs在语义线索提取中的可行性,以解决小弱对象中关键语义缺失的问题。首先,我们提出了一个多头图推理学习模型(MGRL),该模型将初始特征表示投影到图空间中,并利用双层多头图网络提取基本语义信息。其次,我们引入了前景-背景二值掩蔽技术,粗略分割图像的前景区域。将掩码转换为先验提示,然后将其合并到邻接矩阵中,强调MGRL中的对象推理。接下来,我们提出了一个基于交叉学习的特征对齐学习模块来解决空间投影导致的特征不对齐问题。最后,我们采用了跨层语义交互模块,实现了特征的跨层通信和聚合。在DIOR、AI-TOD、NWPU VHR-10、DOTA-v1.0和STAR 5个遥感数据集上进行了大量实验。实验结果证明了该方法的优越性和优越性。
{"title":"SRNet: A Semantic Reasoning Network for Small Weak Object Detection in Remote Sensing Images","authors":"Zheng Li;Hao Feng;Dongdong Xu;Tianqi Zhao;Boxiao Wang;Yongcheng Wang","doi":"10.1109/JSTARS.2025.3645623","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3645623","url":null,"abstract":"Small weak object detection (SWOD) is a significant but neglected task in remote sensing image interpretation. Due to limitations in imaging resolution and inherent characteristics of the objects, detection networks struggle to effectively extract semantic features, which are crucial for object identification and recognition. In recent years, graph convolutional networks (GCNs) have been developed to handle non-Euclidean data. Through GCNs, node data are enriched via aggregation and propagation across the graph. In this article, we explore the feasibility of GCNs in semantic clue extraction to address the lack of key semantics in small weak objects. First, we propose a multihead graph reasoning learning model (MGRL) that projects initial feature representations into graph space and utilizes a two-layer multihead graph network to extract essential semantic information. Second, we introduce a foreground-background binary masking technique that roughly segments the foreground region of the image. The mask is converted into a prior prompt, which is then incorporated into the adjacency matrix, emphasizing object reasoning in MGRL. Next, we present a cross learning-based feature alignment learning module to resolve feature misalignment issues caused by spatial projection. Finally, we adopt a cross-layer semantic interaction module to facilitate cross-layer communication and aggregation of features. Extensive experiments are conducted on five remote sensing datasets: DIOR, AI-TOD, NWPU VHR-10, DOTA-v1.0, and STAR. The experimental results demonstrate the superior performance and advantages of our method.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"2680-2695"},"PeriodicalIF":5.3,"publicationDate":"2025-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11302765","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145929568","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient Spatial-Channel Spiking Neural Network for Multimodal Remote Sensing Data Classification 多模态遥感数据分类的高效空间通道峰值神经网络
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-17 DOI: 10.1109/JSTARS.2025.3645208
Xin He;Yaqin Zhao;Yushi Chen;Limin Zou
Multimodal remote sensing data always comprise hyperspectral image, light detection and ranging data, and synthetic-aperture radar. Different modalities supply complementary information that improves accuracy for multimodal remote sensing classification. Although deep learning-based methods have been a mainstream, they also introduce high computational and energy costs. Unlike existing models, spiking neural network (SNN) is intrinsically energy-efficient, which reduces computation cost and energy expenditure with only a small subset of neurons. However, SNN extracts spatial and channel features without considering the redundancy among different remote sensing modalities. To reduce cross-modal redundancy while retaining important features, this paper proposes an efficient spatial-channel SNN for multimodal remote sensing data classification. First, in the modality fusion step, considering the quality of different modalities is various (i.e., hyperspectral image is impaired by cloud), an energy-guided multimodal remote sensing fusion strategy is proposed, which allocates a high weight for the informative single modality, suppressing less-informative ones by optimizing the generation bound. Second, we leverage information from the high-quality modality. In the spatial feature learning step, an efficient spatial SNN is proposed. It transforms the spatial features to the frequency domain, which shares parameters across different time steps in the spatial dimension to reduce spatial feature redundancy. Finally, to further reduce the redundancy, an efficient channel SNN is explored, which focuses on learning important spike representations in the channel dimension by learning learnable parameters. Experimental results on the three multimodal remote sensing datasets indicate that the proposed methods are competitive compared to the state-of-the-art models.
多模态遥感数据通常包括高光谱图像、光探测和测距数据以及合成孔径雷达。不同的模态提供了互补的信息,提高了多模态遥感分类的准确性。尽管基于深度学习的方法已经成为主流,但它们也引入了高昂的计算和能源成本。与现有模型不同,峰值神经网络(SNN)本质上是节能的,它只需要一小部分神经元就可以降低计算成本和能量消耗。然而,SNN提取空间和信道特征时没有考虑不同遥感模式之间的冗余。为了在保留重要特征的同时减少多模态冗余,本文提出了一种高效的空间信道SNN多模态遥感数据分类方法。首先,在模态融合步骤中,考虑到不同模态的质量差异(即高光谱图像受云的影响),提出了一种能量引导的多模态遥感融合策略,该策略对信息丰富的单模态分配高权重,通过优化生成界来抑制信息较少的单模态。第二,我们利用高质量模式的信息。在空间特征学习步骤中,提出了一种高效的空间信噪比网络。它将空间特征转换到频域,在空间维度上跨不同时间步长共享参数,减少空间特征冗余。最后,为了进一步降低冗余,探索了一种有效的信道SNN,该SNN通过学习可学习参数来学习信道维度中重要的尖峰表示。在三种多模态遥感数据集上的实验结果表明,与现有模型相比,本文提出的方法具有一定的竞争力。
{"title":"Efficient Spatial-Channel Spiking Neural Network for Multimodal Remote Sensing Data Classification","authors":"Xin He;Yaqin Zhao;Yushi Chen;Limin Zou","doi":"10.1109/JSTARS.2025.3645208","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3645208","url":null,"abstract":"Multimodal remote sensing data always comprise hyperspectral image, light detection and ranging data, and synthetic-aperture radar. Different modalities supply complementary information that improves accuracy for multimodal remote sensing classification. Although deep learning-based methods have been a mainstream, they also introduce high computational and energy costs. Unlike existing models, spiking neural network (SNN) is intrinsically energy-efficient, which reduces computation cost and energy expenditure with only a small subset of neurons. However, SNN extracts spatial and channel features without considering the redundancy among different remote sensing modalities. To reduce cross-modal redundancy while retaining important features, this paper proposes an efficient spatial-channel SNN for multimodal remote sensing data classification. First, in the modality fusion step, considering the quality of different modalities is various (i.e., hyperspectral image is impaired by cloud), an energy-guided multimodal remote sensing fusion strategy is proposed, which allocates a high weight for the informative single modality, suppressing less-informative ones by optimizing the generation bound. Second, we leverage information from the high-quality modality. In the spatial feature learning step, an efficient spatial SNN is proposed. It transforms the spatial features to the frequency domain, which shares parameters across different time steps in the spatial dimension to reduce spatial feature redundancy. Finally, to further reduce the redundancy, an efficient channel SNN is explored, which focuses on learning important spike representations in the channel dimension by learning learnable parameters. Experimental results on the three multimodal remote sensing datasets indicate that the proposed methods are competitive compared to the state-of-the-art models.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"2879-2890"},"PeriodicalIF":5.3,"publicationDate":"2025-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11303103","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145982144","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A New Addition to Global Soil Moisture Mapping: CFOSAT Scatterometer Algorithm Development and Validation 全球土壤湿度制图的新增加:CFOSAT散射计算法的开发和验证
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-17 DOI: 10.1109/JSTARS.2025.3645399
Pan Duan;Tianjie Zhao;Haishen Lü;Shuyan Lang;Jingyao Zheng;Yu Bai;Zhiqing Peng;Wolfgang Wagner;Peng Guo;Hongtao Shi;Congrong Sun;Li Jia;Di Zhu;Xiaolong Dong;Jiancheng Shi
The monitoring of global soil moisture is crucial for understanding the hydrological cycle and managing terrestrial water resources. The China–France Oceanography Satellite (CFOSAT), equipped with the first sector-beam rotary scanning microwave scatterometer (CSCAT), provides a novel opportunity for global soil moisture mapping. However, the capability of CFOSAT’s Ku-band for soil moisture retrieval remains underexplored and lacks systematic evaluation. In this study, an Adaptive Backscatter Change Tracking (ABCT) algorithm is designed to retrieve absolute soil moisture from CFOSAT’s CSCAT measurements. The ABCT algorithm assumes a stable roughness, where changes in backscattering are primarily attributed to soil moisture variation based on a logarithmic relationship. It incorporates a vegetation influence coefficient, which quantifies how vegetation impacts the backscatter signal. This coefficient adaptively scales with changes in the Normalized Difference Vegetation Index to adjust the backscattering appropriately to include the effect of vegetation growth or decay. This allows the algorithm to isolate changes in the backscatter signal that are due to soil moisture while minimizing the false readings from vegetation growth or wilting. The CFOSAT ABCT algorithm’s performance was evaluated against extensive in-situ soil moisture data, demonstrating a robust correlation, with the Vertical-Vertical Polarization Ascending Orbit (VV Asc) result showing the highest accuracy, indicated by Pearson’s correlation coefficient (R) of 0.68 and unbiased root mean squared error (ubRMSE) of 0.057 m3/m3. Comparative analysis with the Advanced Scatterometer (ASCAT) data revealed that, while the ABCT algorithm’s correlation was slightly lower than that of the official EUMETSAT H SAF product, it notably improved the bias and ubRMSE metrics. This study underscores that the CFOSAT ABCT soil moisture retrieval algorithm and product are a valuable addition to global soil moisture mapping, complementing existing satellite missions or sensors such as SMAP, SMOS, ASCAT, AMSR2, and FY-3/MWRI.
全球土壤湿度监测对于了解水文循环和管理陆地水资源至关重要。中法海洋卫星(CFOSAT)配备了首个扇形波束旋转扫描微波散射仪(CSCAT),为全球土壤湿度测绘提供了一个新的机会。然而,CFOSAT的ku波段土壤水分检索能力仍未得到充分开发,也缺乏系统的评价。在本研究中,设计了一种自适应后向散射变化跟踪(ABCT)算法,用于从CFOSAT的CSCAT测量数据中检索绝对土壤湿度。ABCT算法假设粗糙度稳定,其中后向散射的变化主要归因于基于对数关系的土壤湿度变化。它包含植被影响系数,该系数量化了植被对后向散射信号的影响。该系数自适应缩放归一化植被指数的变化,以适当调整后向散射,以包括植被生长或腐烂的影响。这使得算法可以隔离由于土壤湿度引起的后向散射信号的变化,同时最大限度地减少植被生长或枯萎的错误读数。CFOSAT ABCT算法的性能与大量的原位土壤湿度数据进行了评估,显示出很强的相关性,其中垂直垂直极化上升轨道(VV Asc)结果显示出最高的准确性,Pearson相关系数(R)为0.68,无偏均方根误差(ubRMSE)为0.057 m3/m3。与Advanced Scatterometer (ASCAT)数据的对比分析表明,ABCT算法的相关性略低于EUMETSAT H SAF官方产品,但显著改善了偏差和ubRMSE指标。该研究强调,CFOSAT ABCT土壤水分检索算法和产品是对全球土壤水分制图的重要补充,是对现有卫星任务或传感器(如SMAP、SMOS、ASCAT、AMSR2和FY-3/MWRI)的补充。
{"title":"A New Addition to Global Soil Moisture Mapping: CFOSAT Scatterometer Algorithm Development and Validation","authors":"Pan Duan;Tianjie Zhao;Haishen Lü;Shuyan Lang;Jingyao Zheng;Yu Bai;Zhiqing Peng;Wolfgang Wagner;Peng Guo;Hongtao Shi;Congrong Sun;Li Jia;Di Zhu;Xiaolong Dong;Jiancheng Shi","doi":"10.1109/JSTARS.2025.3645399","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3645399","url":null,"abstract":"The monitoring of global soil moisture is crucial for understanding the hydrological cycle and managing terrestrial water resources. The China–France Oceanography Satellite (CFOSAT), equipped with the first sector-beam rotary scanning microwave scatterometer (CSCAT), provides a novel opportunity for global soil moisture mapping. However, the capability of CFOSAT’s Ku-band for soil moisture retrieval remains underexplored and lacks systematic evaluation. In this study, an Adaptive Backscatter Change Tracking (ABCT) algorithm is designed to retrieve absolute soil moisture from CFOSAT’s CSCAT measurements. The ABCT algorithm assumes a stable roughness, where changes in backscattering are primarily attributed to soil moisture variation based on a logarithmic relationship. It incorporates a vegetation influence coefficient, which quantifies how vegetation impacts the backscatter signal. This coefficient adaptively scales with changes in the Normalized Difference Vegetation Index to adjust the backscattering appropriately to include the effect of vegetation growth or decay. This allows the algorithm to isolate changes in the backscatter signal that are due to soil moisture while minimizing the false readings from vegetation growth or wilting. The CFOSAT ABCT algorithm’s performance was evaluated against extensive in-situ soil moisture data, demonstrating a robust correlation, with the Vertical-Vertical Polarization Ascending Orbit (VV Asc) result showing the highest accuracy, indicated by Pearson’s correlation coefficient (R) of 0.68 and unbiased root mean squared error (ubRMSE) of 0.057 m<sup>3</sup>/m<sup>3</sup>. Comparative analysis with the Advanced Scatterometer (ASCAT) data revealed that, while the ABCT algorithm’s correlation was slightly lower than that of the official EUMETSAT H SAF product, it notably improved the bias and ubRMSE metrics. This study underscores that the CFOSAT ABCT soil moisture retrieval algorithm and product are a valuable addition to global soil moisture mapping, complementing existing satellite missions or sensors such as SMAP, SMOS, ASCAT, AMSR2, and FY-3/MWRI.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"2439-2460"},"PeriodicalIF":5.3,"publicationDate":"2025-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11303030","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145929451","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Semisupervised Prototypical Network With Dynamic Threshold Pseudolabeling for Forest Classification 基于动态阈值伪标记的森林分类半监督原型网络
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-17 DOI: 10.1109/JSTARS.2025.3645613
Yifan Xie;Long Chen;Jiahao Wang;Nuermaimaitijiang Aierken;Geng Wang;Xiaoli Zhang
Forest tree species classification using remote sensing data often faces the challenge of limited labeled samples, which hampers the performance of deep learning models. While few-shot learning techniques, such as prototypical networks (PNet), show promise, overfitting remains a significant issue. Given the relatively low cost of acquiring unlabeled data, semisupervised learning offers a potential solution. However, due to class imbalance, pseudolabels based on fixed confidence thresholds tend to favor majority classes, leading to lower classification accuracy for minority classes. To address this, we propose a novel semisupervised few-shot classification model, classwise pseudolabeling squeeze-and-excitation PNet (CWPL-SEPNet). The model incorporates a channel attention module into the PNet backbone and employs a classwise adaptive pseudolabeling mechanism based on quantile thresholds. This approach balances the pseudolabeled samples and reduces bias toward majority classes. Experiments conducted using Sentinel-2 imagery in Pu’er City, China, show that incorporating unlabeled data increases the overall classification accuracy to 95.14%, with per-class accuracies of 91.82% for Tea Farm, 90.19% for Oak, 94.70% for Eucalyptus, and 94.64% for Pinus kesiya. The CWPL strategy significantly outperforms traditional fixed-threshold methods, particularly in handling class imbalance and improving classification accuracy for minority classes. Compared to baseline methods such as TPN-semi, PNet, random forest, and support vector machine, CWPL-SEPNet excels in overall accuracy, average accuracy, and Kappa value. Furthermore, the model was validated on three publicly available remote sensing datasets. CWPL-SEPNet provides a robust and efficient classification solution under few-shot conditions, offering an effective approach for tree species classification using remote sensing data.
利用遥感数据进行森林树种分类往往面临标记样本有限的挑战,这阻碍了深度学习模型的性能。虽然像原型网络(PNet)这样的少量学习技术显示出了希望,但过拟合仍然是一个重大问题。鉴于获取未标记数据的成本相对较低,半监督学习提供了一个潜在的解决方案。然而,由于类的不平衡,基于固定置信度阈值的伪标签倾向于多数类,导致对少数类的分类准确率较低。为了解决这一问题,我们提出了一种新的半监督少射分类模型,即分类伪标记挤压-激励PNet (CWPL-SEPNet)。该模型将信道关注模块集成到PNet骨干网中,并采用基于分位数阈值的分类自适应伪标记机制。这种方法平衡了伪标记的样本,减少了对大多数类别的偏见。利用中国普洱市Sentinel-2遥感影像进行的实验表明,纳入未标记数据后,整体分类准确率提高到95.14%,其中茶园、橡树、桉树和克西亚松的分类准确率分别为91.82%、90.19%、94.70%和94.64%。CWPL策略明显优于传统的固定阈值方法,特别是在处理类不平衡和提高少数类的分类精度方面。与TPN-semi、PNet、随机森林和支持向量机等基线方法相比,CWPL-SEPNet在总体精度、平均精度和Kappa值等方面具有优势。此外,该模型在三个公开的遥感数据集上进行了验证。CWPL-SEPNet提供了一种鲁棒且高效的少拍条件下的分类方案,为利用遥感数据进行树种分类提供了有效的方法。
{"title":"A Semisupervised Prototypical Network With Dynamic Threshold Pseudolabeling for Forest Classification","authors":"Yifan Xie;Long Chen;Jiahao Wang;Nuermaimaitijiang Aierken;Geng Wang;Xiaoli Zhang","doi":"10.1109/JSTARS.2025.3645613","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3645613","url":null,"abstract":"Forest tree species classification using remote sensing data often faces the challenge of limited labeled samples, which hampers the performance of deep learning models. While few-shot learning techniques, such as prototypical networks (PNet), show promise, overfitting remains a significant issue. Given the relatively low cost of acquiring unlabeled data, semisupervised learning offers a potential solution. However, due to class imbalance, pseudolabels based on fixed confidence thresholds tend to favor majority classes, leading to lower classification accuracy for minority classes. To address this, we propose a novel semisupervised few-shot classification model, classwise pseudolabeling squeeze-and-excitation PNet (CWPL-SEPNet). The model incorporates a channel attention module into the PNet backbone and employs a classwise adaptive pseudolabeling mechanism based on quantile thresholds. This approach balances the pseudolabeled samples and reduces bias toward majority classes. Experiments conducted using Sentinel-2 imagery in Pu’er City, China, show that incorporating unlabeled data increases the overall classification accuracy to 95.14%, with per-class accuracies of 91.82% for Tea Farm, 90.19% for Oak, 94.70% for <italic>Eucalyptus</i>, and 94.64% for <italic>Pinus kesiya</i>. The CWPL strategy significantly outperforms traditional fixed-threshold methods, particularly in handling class imbalance and improving classification accuracy for minority classes. Compared to baseline methods such as TPN-semi, PNet, random forest, and support vector machine, CWPL-SEPNet excels in overall accuracy, average accuracy, and Kappa value. Furthermore, the model was validated on three publicly available remote sensing datasets. CWPL-SEPNet provides a robust and efficient classification solution under few-shot conditions, offering an effective approach for tree species classification using remote sensing data.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"2405-2422"},"PeriodicalIF":5.3,"publicationDate":"2025-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11303117","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145929562","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TSP-Former: A Phenology-Guided Transformer for Tobacco Mapping Using Satellite Image Time Series TSP-Former:利用卫星图像时间序列进行烟草制图的物候导向变压器
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-17 DOI: 10.1109/JSTARS.2025.3645265
Huaming Gao;Yongqing Bai;Qing Sun;Haoran Wang;Xiangyu Tian;Hui Ma;Yixiang Li;Xianghong Che;Zhengchao Chen
Tobacco is a phenology-sensitive and economically significant crop that requires accurate and timely spatial mapping to support agricultural planning and public health regulation. However, single-date spectral similarity among crops and regional differences in planting practices limit the generalizability of existing approaches, particularly deep learning (DL) models. To address these challenges, we propose a novel phenologyguided DL framework that leverages satellite image time series (SITS) to capture crop-specific growth dynamics. Specifically, we introduce the tobacco spectral-phenological variable (TSP), which captures change rates in Red Edge-2 during peak growth. It serves as crop-specific prior knowledge for model guidance. Based on this, we develop TSP-Former, a transformer architecture that incorporates two novel modules: a central prior attention module (CPAM), which adaptively fuses spectral information with phenological priors, and an NDVI-enhanced temporal decoder (NDTD), which reinforces temporal learning by emphasizing phenologically critical stages using NDVI-weighted sequences. Extensive experiments across four major tobacco regions using Sentinel-2 imagery demonstrate the method’s superior cross-regional robustness. TSP-Former achieves an average weighted F1-score of 87.1% and an overall accuracy of 85.9%, significantly outperforming random forest and competing DL approaches. Notably, in challenging regions characterized by substantial phenological shifts, the proposed method surpasses the emerging remote sensing foundation model, AlphaEarth with a fine-tuned lightweight multilayer perceptron, by over 15% in accuracy. These findings highlight the effectiveness of integrating phenological priors into temporal deep models, enabling robust and transferable crop mapping across heterogeneous and data-constrained regions, with clear implications for scalable agricultural monitoring and policy development.
烟草是一种物候敏感且经济意义重大的作物,需要准确和及时的空间测绘,以支持农业规划和公共卫生监管。然而,作物之间的单日期光谱相似性和种植实践的区域差异限制了现有方法的泛化性,特别是深度学习(DL)模型。为了应对这些挑战,我们提出了一种新的物候引导的深度学习框架,该框架利用卫星图像时间序列(sit)来捕捉作物特定的生长动态。具体来说,我们引入了烟草光谱物候变量(TSP),它捕获了红边2在生长高峰期间的变化率。它作为特定作物的先验知识用于模型指导。基于此,我们开发了TSP-Former,这是一种包含两个新模块的变压器架构:一个是中央先验注意模块(CPAM),它自适应地融合光谱信息和物候先验,另一个是ndvi增强时间解码器(NDTD),它通过使用ndvi加权序列强调物候关键阶段来加强时间学习。使用Sentinel-2图像在四个主要烟草区进行的广泛实验表明,该方法具有优越的跨区域鲁棒性。TSP-Former的平均加权f1得分为87.1%,总体准确率为85.9%,显著优于随机森林和竞争的深度学习方法。值得注意的是,在具有大量物候变化特征的具有挑战性的地区,所提出的方法超过了新兴的遥感基础模型AlphaEarth,具有微调的轻量级多层感知器,精度超过15%。这些发现强调了将物候先验整合到时间深度模型中的有效性,从而实现跨异质和数据受限区域的可靠和可转移的作物制图,对可扩展的农业监测和政策制定具有明确的意义。
{"title":"TSP-Former: A Phenology-Guided Transformer for Tobacco Mapping Using Satellite Image Time Series","authors":"Huaming Gao;Yongqing Bai;Qing Sun;Haoran Wang;Xiangyu Tian;Hui Ma;Yixiang Li;Xianghong Che;Zhengchao Chen","doi":"10.1109/JSTARS.2025.3645265","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3645265","url":null,"abstract":"Tobacco is a phenology-sensitive and economically significant crop that requires accurate and timely spatial mapping to support agricultural planning and public health regulation. However, single-date spectral similarity among crops and regional differences in planting practices limit the generalizability of existing approaches, particularly deep learning (DL) models. To address these challenges, we propose a novel phenologyguided DL framework that leverages satellite image time series (SITS) to capture crop-specific growth dynamics. Specifically, we introduce the tobacco spectral-phenological variable (TSP), which captures change rates in Red Edge-2 during peak growth. It serves as crop-specific prior knowledge for model guidance. Based on this, we develop TSP-Former, a transformer architecture that incorporates two novel modules: a central prior attention module (CPAM), which adaptively fuses spectral information with phenological priors, and an NDVI-enhanced temporal decoder (NDTD), which reinforces temporal learning by emphasizing phenologically critical stages using NDVI-weighted sequences. Extensive experiments across four major tobacco regions using Sentinel-2 imagery demonstrate the method’s superior cross-regional robustness. TSP-Former achieves an average weighted F1-score of 87.1% and an overall accuracy of 85.9%, significantly outperforming random forest and competing DL approaches. Notably, in challenging regions characterized by substantial phenological shifts, the proposed method surpasses the emerging remote sensing foundation model, AlphaEarth with a fine-tuned lightweight multilayer perceptron, by over 15% in accuracy. These findings highlight the effectiveness of integrating phenological priors into temporal deep models, enabling robust and transferable crop mapping across heterogeneous and data-constrained regions, with clear implications for scalable agricultural monitoring and policy development.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"2423-2438"},"PeriodicalIF":5.3,"publicationDate":"2025-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11302804","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145929537","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DESR: Super-Resolution Reconstruction of FengYun-4 Multispectral Images DESR:风云四号多光谱图像的超分辨率重建
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-17 DOI: 10.1109/JSTARS.2025.3645116
Xi Kan;Xu Liu;Yonghong Zhang;Linglong Zhu;Jing Wang;Zhou Zhou;Lei Gong;Xianwu Wang
The maximum spatial resolution varies significantly across different bands in the reflectance data of FengYun-4A (FY-4A) remote sensing images. Super-resolution (SR) reconstruction of FY-4A remote sensing imagery not only enhances the spatial accuracy of low-resolution bands and achieves cross-band scale consistency, but also improves feature recognition and monitoring capabilities. This provides clearer and more reliable data support for quantitative remote sensing and applications in meteorology, ecology, agriculture, and other fields. Therefore, a differential enhancement super-resolution network (DESR) is proposed based on scale invariance. The dual paths consist of a CNN branch and a Swin Transformer branch. The CNN branch employs rep-residual blocks (RRB) to capture spatial structures, where each RRB integrates a spatial feature attention that employs large convolutional kernels with directional strides along the height and width to model long-range dependencies and spatial correlations. The Swin Transformer branch adopts residual Swin Transformer blocks to obtain a global receptive field. In addition, A differential feature enhancement module is further introduced to fuse features, highlight branch-specific deficiencies through subtraction, and achieve complementary enhancement. Experimental results show that DESR achieves more uniform error distribution and superior reconstruction quality compared with representative methods. On 2 × and 4 × SR tasks, DESR reaches PSNR values of 54.1395 and 46.0942, SSIM values of 0.9899 and 0.9749, with improvements at least 1.87% and 0.59%, respectively, while also attaining the best spectral angle mapping.
风云- 4a (FY-4A)遥感影像反射率数据在不同波段的最大空间分辨率差异显著。对FY-4A遥感影像进行超分辨率重建,不仅提高了低分辨率波段的空间精度,实现了跨波段尺度一致性,而且提高了特征识别和监测能力。这为定量遥感以及气象、生态、农业等领域的应用提供了更清晰、更可靠的数据支撑。为此,提出一种基于尺度不变性的差分增强超分辨网络(DESR)。双路径由CNN分支和Swin Transformer分支组成。CNN分支采用再现残差块(RRB)来捕获空间结构,其中每个RRB集成了一个空间特征注意,该空间特征注意采用沿高度和宽度方向跨步的大卷积核来建模远程依赖关系和空间相关性。Swin Transformer分支采用剩余的Swin Transformer块来获得全局接受场。此外,还引入差分特征增强模块,融合特征,通过减法突出分支的不足,实现互补增强。实验结果表明,与代表性方法相比,DESR具有更均匀的误差分布和更好的重建质量。在2 ×和4 × SR任务上,DESR的PSNR分别达到54.1395和46.0942,SSIM分别达到0.9899和0.9749,分别提高了1.87%和0.59%,同时也获得了最佳的光谱角映射。
{"title":"DESR: Super-Resolution Reconstruction of FengYun-4 Multispectral Images","authors":"Xi Kan;Xu Liu;Yonghong Zhang;Linglong Zhu;Jing Wang;Zhou Zhou;Lei Gong;Xianwu Wang","doi":"10.1109/JSTARS.2025.3645116","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3645116","url":null,"abstract":"The maximum spatial resolution varies significantly across different bands in the reflectance data of FengYun-4A (FY-4A) remote sensing images. Super-resolution (SR) reconstruction of FY-4A remote sensing imagery not only enhances the spatial accuracy of low-resolution bands and achieves cross-band scale consistency, but also improves feature recognition and monitoring capabilities. This provides clearer and more reliable data support for quantitative remote sensing and applications in meteorology, ecology, agriculture, and other fields. Therefore, a differential enhancement super-resolution network (DESR) is proposed based on scale invariance. The dual paths consist of a CNN branch and a Swin Transformer branch. The CNN branch employs rep-residual blocks (RRB) to capture spatial structures, where each RRB integrates a spatial feature attention that employs large convolutional kernels with directional strides along the height and width to model long-range dependencies and spatial correlations. The Swin Transformer branch adopts residual Swin Transformer blocks to obtain a global receptive field. In addition, A differential feature enhancement module is further introduced to fuse features, highlight branch-specific deficiencies through subtraction, and achieve complementary enhancement. Experimental results show that DESR achieves more uniform error distribution and superior reconstruction quality compared with representative methods. On 2 × and 4 × SR tasks, DESR reaches PSNR values of 54.1395 and 46.0942, SSIM values of 0.9899 and 0.9749, with improvements at least 1.87% and 0.59%, respectively, while also attaining the best spectral angle mapping.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"2605-2620"},"PeriodicalIF":5.3,"publicationDate":"2025-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11303137","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145929567","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Forest Stress Detection Using Feature Engineering and Selection Approach Optimized for Satellite Imagery 基于卫星图像优化的特征工程和选择方法的森林应力检测
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-15 DOI: 10.1109/JSTARS.2025.3644488
Yevhenii Salii;Volodymyr Kuzin;Nataliia Kussul;Alla Lavreniuk
Accurate detection of forest stress from satellite data depends heavily on selecting informative spectral features. Traditional approaches rely on a limited set of predefined vegetation indices, which may not generalize across environmental conditions. In this study, we introduce enhanced maximum informativeness maximum independence (E-MIMI), an efficient and interpretable feature selection strategy that identifies optimal combinations of spectral features from generalized vegetation index classes rather than fixed indices. The method combines a genetic algorithm with a caching mechanism and informativeness-based scoring to reduce computation time while maintaining high accuracy. Applied to Sentinel-2 imagery from two ecologically distinct regions, E-MIMI consistently selected index combinations involving red-edge and shortwave infrared bands—spectral domains known to reflect canopy water content and chlorophyll degradation. E-MIMI demonstrates exceptional computational efficiency, completing feature selection up to 80 times faster and using over 1000 times less memory than other traditional methods on large feature spaces. Despite this, E-MIMI achieves comparable levels of segmentation performance with a test intersection over union (IoU) of 0.61–0.63, while other methods reach an IoU of 0.60–0.64. Obtained models show a substantial improvement over previous studies in the same region (0.515–0.549 IoU). The model also generalized well to an independent dataset from Chornobyl, confirming its robustness. By integrating computer vision techniques with biophysically grounded features, our approach supports scalable, ecologically meaningful forest stress monitoring and offers a practical foundation for broader environmental applications requiring interpretable and computationally efficient feature selection.
从卫星数据中准确检测森林应力在很大程度上取决于选择信息丰富的光谱特征。传统的方法依赖于一组有限的预定义植被指数,这些指数可能无法在各种环境条件下进行推广。在本研究中,我们引入了增强最大信息量最大独立性(enhanced maximum informativeness maximum independence, E-MIMI),这是一种高效且可解释的特征选择策略,可以从广义植被指数类别而不是固定指数中识别光谱特征的最佳组合。该方法将遗传算法与缓存机制和基于信息的评分相结合,在保持较高准确率的同时减少了计算时间。E-MIMI应用于来自两个生态不同区域的Sentinel-2图像,一致地选择了包括红边和短波红外波段的指数组合,这些光谱域已知可以反映冠层含水量和叶绿素降解。E-MIMI展示了卓越的计算效率,在大型特征空间上完成特征选择的速度比其他传统方法快80倍,使用的内存比其他传统方法少1000倍以上。尽管如此,E-MIMI达到了相当水平的分割性能,测试交集超过联合(IoU)为0.61-0.63,而其他方法的IoU为0.60-0.64。获得的模型显示,在同一地区(0.515-0.549 IoU),比以往的研究有了很大的改善。该模型也可以很好地推广到来自切尔诺贝利的独立数据集,证实了它的鲁棒性。通过将计算机视觉技术与生物物理特征相结合,我们的方法支持可扩展的、有生态意义的森林应力监测,并为需要可解释和计算效率高的特征选择的更广泛的环境应用提供了实践基础。
{"title":"Forest Stress Detection Using Feature Engineering and Selection Approach Optimized for Satellite Imagery","authors":"Yevhenii Salii;Volodymyr Kuzin;Nataliia Kussul;Alla Lavreniuk","doi":"10.1109/JSTARS.2025.3644488","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3644488","url":null,"abstract":"Accurate detection of forest stress from satellite data depends heavily on selecting informative spectral features. Traditional approaches rely on a limited set of predefined vegetation indices, which may not generalize across environmental conditions. In this study, we introduce enhanced maximum informativeness maximum independence (E-MIMI), an efficient and interpretable feature selection strategy that identifies optimal combinations of spectral features from generalized vegetation index classes rather than fixed indices. The method combines a genetic algorithm with a caching mechanism and informativeness-based scoring to reduce computation time while maintaining high accuracy. Applied to Sentinel-2 imagery from two ecologically distinct regions, E-MIMI consistently selected index combinations involving red-edge and shortwave infrared bands—spectral domains known to reflect canopy water content and chlorophyll degradation. E-MIMI demonstrates exceptional computational efficiency, completing feature selection up to 80 times faster and using over 1000 times less memory than other traditional methods on large feature spaces. Despite this, E-MIMI achieves comparable levels of segmentation performance with a test intersection over union (IoU) of 0.61–0.63, while other methods reach an IoU of 0.60–0.64. Obtained models show a substantial improvement over previous studies in the same region (0.515–0.549 IoU). The model also generalized well to an independent dataset from Chornobyl, confirming its robustness. By integrating computer vision techniques with biophysically grounded features, our approach supports scalable, ecologically meaningful forest stress monitoring and offers a practical foundation for broader environmental applications requiring interpretable and computationally efficient feature selection.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"2461-2473"},"PeriodicalIF":5.3,"publicationDate":"2025-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11300936","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145929574","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Research on a Deep-Learning Model for Multisource Spaceborne GNSS-R Fusion in Sea Surface Height Retrieval 多源星载GNSS-R融合海面高度反演深度学习模型研究
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-15 DOI: 10.1109/JSTARS.2025.3643863
Yun Zhang;Tianyue Wen;Shuhu Yang;Qingjing Shi;Qifeng Qian;Chunyi Xiang;Jiaying Li;Binbin Li;Bo Peng;Yanling Han;Zhonghua Hong
The development of GNSS-R ocean altimetry technology has significantly advanced global sea surface height (SSH) monitoring. However, single-satellite systems face inherent limitations: their spatial coverage is constrained by orbital patterns and revisit cycles, making them insufficient for high-precision, high-spatiotemporal-resolution global SSH monitoring. This article innovatively proposes a deep-learning model for multisource spaceborne GNSS-R fusion-based SSH retrieval. The model corrects low-accuracy single-source inversion results and integrates inversion outputs from four GNSS-R satellite systems: FY-3E, FY-3G, CYGNSS, and Tianmu-1. This approach reduces errors caused by single sources and improves the coverage area of SSH measurements. Experiments were conducted globally between 60°N and 60°S, where the higher precision FY-3E SSH data were used to correct the retrieval results of FY-3G, CYGNSS, and Tianmu-1. Under relatively lenient data quality control criteria to construct high-coverage global SSH gridded products, the test results spanning three distinct months demonstrated significant performance improvements across all constellations. Following the error correction model optimization, the FY-3G constellation achieved a corrected mean absolute error (MAE) ranging from 1.433 to 2.158 m, representing a reduction of 40%–70% compared with precorrection values. Similarly, the corrected MAE for the CYGNSS constellation ranged from 2.178 to 4.192 m, also reflecting a 40%–70% reduction. For the Tianmu-1 constellation, the corrected MAE was refined to 1.311–1.505 m, with an MAE reduction exceeding 80%. The fusion of the four satellite systems achieved a sea surface coverage of 75.75% within an 8-h window, with high consistency against validation datasets, such as DTU18 validation model and ATL12 mean SSH. The findings of this study significantly enhance the SSH retrieval accuracy of commercial constellations not originally designed for altimetry purposes (e.g., Tianmu-1) and provide a novel approach for multisource GNSS-R fusion-based SSH monitoring. This work holds important theoretical significance and practical value, particularly offering broad application prospects in global ocean monitoring and climate change research.
GNSS-R海洋测高技术的发展极大地推动了全球海面高度监测的发展。然而,单卫星系统面临着固有的局限性:它们的空间覆盖受到轨道模式和重访周期的限制,这使得它们不足以进行高精度、高时空分辨率的全球SSH监测。创新性地提出了一种基于多源星载GNSS-R融合的SSH检索深度学习模型。该模型对低精度的单源反演结果进行了校正,并整合了4个GNSS-R卫星系统(FY-3E、FY-3G、CYGNSS和天目一号)的反演输出。这种方法减少了由单一来源引起的错误,并提高了SSH测量的覆盖范围。实验在全球60°N ~ 60°S范围内进行,使用精度更高的FY-3E SSH数据对FY-3G、CYGNSS和天目一号的检索结果进行校正。在相对宽松的数据质量控制标准下,构建高覆盖率的全球SSH网格产品,跨越三个不同月的测试结果表明,所有星座的性能都有显著提高。经过误差修正模型优化,FY-3G星座的校正平均绝对误差(MAE)在1.433 ~ 2.158 m之间,比校正前降低了40% ~ 70%。同样,CYGNSS星座的修正MAE范围为2.178至4.192米,也反映了40%-70%的减少。对于天目一号星座,修正后的MAE细化为1.311 ~ 1.505 m, MAE减小幅度超过80%。4个卫星系统的融合在8 h窗口内实现了75.75%的海面覆盖率,与验证数据集(如DTU18验证模型和ATL12平均SSH)具有很高的一致性。该研究结果显著提高了非高度计商业星座(如天母一号)的SSH检索精度,并为基于GNSS-R融合的多源SSH监测提供了一种新方法。该工作具有重要的理论意义和实践价值,尤其在全球海洋监测和气候变化研究中具有广阔的应用前景。
{"title":"Research on a Deep-Learning Model for Multisource Spaceborne GNSS-R Fusion in Sea Surface Height Retrieval","authors":"Yun Zhang;Tianyue Wen;Shuhu Yang;Qingjing Shi;Qifeng Qian;Chunyi Xiang;Jiaying Li;Binbin Li;Bo Peng;Yanling Han;Zhonghua Hong","doi":"10.1109/JSTARS.2025.3643863","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3643863","url":null,"abstract":"The development of GNSS-R ocean altimetry technology has significantly advanced global sea surface height (SSH) monitoring. However, single-satellite systems face inherent limitations: their spatial coverage is constrained by orbital patterns and revisit cycles, making them insufficient for high-precision, high-spatiotemporal-resolution global SSH monitoring. This article innovatively proposes a deep-learning model for multisource spaceborne GNSS-R fusion-based SSH retrieval. The model corrects low-accuracy single-source inversion results and integrates inversion outputs from four GNSS-R satellite systems: FY-3E, FY-3G, CYGNSS, and Tianmu-1. This approach reduces errors caused by single sources and improves the coverage area of SSH measurements. Experiments were conducted globally between 60°N and 60°S, where the higher precision FY-3E SSH data were used to correct the retrieval results of FY-3G, CYGNSS, and Tianmu-1. Under relatively lenient data quality control criteria to construct high-coverage global SSH gridded products, the test results spanning three distinct months demonstrated significant performance improvements across all constellations. Following the error correction model optimization, the FY-3G constellation achieved a corrected mean absolute error (MAE) ranging from 1.433 to 2.158 m, representing a reduction of 40%–70% compared with precorrection values. Similarly, the corrected MAE for the CYGNSS constellation ranged from 2.178 to 4.192 m, also reflecting a 40%–70% reduction. For the Tianmu-1 constellation, the corrected MAE was refined to 1.311–1.505 m, with an MAE reduction exceeding 80%. The fusion of the four satellite systems achieved a sea surface coverage of 75.75% within an 8-h window, with high consistency against validation datasets, such as DTU18 validation model and ATL12 mean SSH. The findings of this study significantly enhance the SSH retrieval accuracy of commercial constellations not originally designed for altimetry purposes (e.g., Tianmu-1) and provide a novel approach for multisource GNSS-R fusion-based SSH monitoring. This work holds important theoretical significance and practical value, particularly offering broad application prospects in global ocean monitoring and climate change research.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"2103-2119"},"PeriodicalIF":5.3,"publicationDate":"2025-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11299433","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145886606","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Semantic-Guided Hierarchical Consistency Domain Adaptation for Open-Set Remote Sensing Scene Classification 面向开放集遥感场景分类的语义引导层次一致性域自适应
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-15 DOI: 10.1109/JSTARS.2025.3644442
Yang Zhao;Jiaqi Liang;Hancheng Ma;Pingping Huang;Yifan Dong;Jing Li
Open-set domain adaptation (OSDA) aims to generalize cross-domain remote sensing scene classification by classifying unknown categories that exist in the target and not seen in the source domain. In remote sensing, the significant distribution discrepancy between two domains hinders effective knowledge transfer, which degrades the generalization performance of OSDA. In addition, the semantic similarity among different categories impacts the classification performance of both known and unknown categories. However, existing OSDA methods often neglect transferable semantic information and this limits their generalization ability. To address these issues, this article proposed a semantic-guided hierarchical consistency domain adaptation (SGHC) method to enhance semantic separability and cross-domain generalization. Specifically, an attribute guided prompt (AGP) is introduced to mine transferable semantic attributes and semantic relationships. The semantic information effectively improves fine-grained scene understanding and promotes the distinguishing of unknown categories. Then, a hierarchical consistency (HC) is employed to complement generalization in open-set scenarios. The HC retains discriminative information of categories and effectively alleviates the domain gap between the source and target domain to avoid negative transfer. To validate the proposed method's performance, experiments are conducted on six cross-domain scenarios with aerial image dataset, Northwestern Polytechnical University dataset (NWPU), and University of California Merced Land Use dataset (UCMD). Experimental results demonstrate the effectiveness of the proposed method in open-set remote sensing scene classification. Especially, the proposed method improves the overall classification accuracies by at least 5.6% on the NWPU $rightarrow$ UCMD scenario compared with the other eleven state-of-the-art methods.
开放集域自适应(Open-set domain adaptation, OSDA)是通过对目标域中存在而源域中不存在的未知类别进行分类,从而实现遥感场景跨域分类的泛化。在遥感中,两域之间的显著分布差异阻碍了知识的有效转移,从而降低了OSDA的泛化性能。此外,不同类别之间的语义相似度对已知和未知类别的分类性能都有影响。然而,现有的OSDA方法往往忽略了可转移的语义信息,这限制了它们的泛化能力。为了解决这些问题,本文提出了一种语义引导的分层一致性域自适应(SGHC)方法来增强语义可分性和跨域泛化。具体来说,引入属性引导提示(attribute guided prompt, AGP)来挖掘可转移的语义属性和语义关系。语义信息有效地提高了对场景的细粒度理解,促进了对未知类别的区分。然后,使用层次一致性(HC)来补充开集场景下的泛化。HC保留了类别的判别信息,有效地缓解了源域和目标域之间的域差距,避免了负迁移。为了验证该方法的性能,利用航空图像数据集、西北工业大学数据集(NWPU)和加州大学默塞德土地利用数据集(UCMD)在6个跨域场景下进行了实验。实验结果证明了该方法在开放集遥感场景分类中的有效性。特别是,与其他11种最先进的方法相比,所提出的方法在NWPU $right row$ UCMD场景下的总体分类精度至少提高了5.6%。
{"title":"Semantic-Guided Hierarchical Consistency Domain Adaptation for Open-Set Remote Sensing Scene Classification","authors":"Yang Zhao;Jiaqi Liang;Hancheng Ma;Pingping Huang;Yifan Dong;Jing Li","doi":"10.1109/JSTARS.2025.3644442","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3644442","url":null,"abstract":"Open-set domain adaptation (OSDA) aims to generalize cross-domain remote sensing scene classification by classifying unknown categories that exist in the target and not seen in the source domain. In remote sensing, the significant distribution discrepancy between two domains hinders effective knowledge transfer, which degrades the generalization performance of OSDA. In addition, the semantic similarity among different categories impacts the classification performance of both known and unknown categories. However, existing OSDA methods often neglect transferable semantic information and this limits their generalization ability. To address these issues, this article proposed a semantic-guided hierarchical consistency domain adaptation (SGHC) method to enhance semantic separability and cross-domain generalization. Specifically, an attribute guided prompt (AGP) is introduced to mine transferable semantic attributes and semantic relationships. The semantic information effectively improves fine-grained scene understanding and promotes the distinguishing of unknown categories. Then, a hierarchical consistency (HC) is employed to complement generalization in open-set scenarios. The HC retains discriminative information of categories and effectively alleviates the domain gap between the source and target domain to avoid negative transfer. To validate the proposed method's performance, experiments are conducted on six cross-domain scenarios with aerial image dataset, Northwestern Polytechnical University dataset (NWPU), and University of California Merced Land Use dataset (UCMD). Experimental results demonstrate the effectiveness of the proposed method in open-set remote sensing scene classification. Especially, the proposed method improves the overall classification accuracies by at least 5.6% on the NWPU <inline-formula><tex-math>$rightarrow$</tex-math></inline-formula> UCMD scenario compared with the other eleven state-of-the-art methods.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"2088-2102"},"PeriodicalIF":5.3,"publicationDate":"2025-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11300939","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145886642","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hybrid Attention Driven CNN-Mamba Multimodal Fusion Network for Remote Sensing Image Semantic Segmentation 混合注意力驱动的CNN-Mamba多模态融合网络遥感图像语义分割
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-15 DOI: 10.1109/JSTARS.2025.3644588
Shu Tian;Minglei Li;Lin Cao;Lihong Kang;Jing Tian;Xiangwei Xing;Bo Shen;Kangning Du;Chong Fu;Ye Zhang
In recent years, the increase of multimodal image data has offered a broader prospect for multimodal semantic segmentation. However, the data heterogeneity between different modalities make it difficult to leverage complementary information and create semantic understanding deviations, which limits the fusion quality and segmentation accuracy. To overcome these challenges, we propose a hybrid attention driven CNN-Mamba multimodal fusion network (HACMNet) for semantic segmentation. It aims to fully exploit the strengths of optical images in texture and semantic representation, along with the complementary structural and elevation information from the digital surface model (DSM). This enables the effective extraction and combination of global and local complementary information to achieve higher accuracy and robustness in semantic segmentation. Specifically, we propose a progressive cross-modal feature interaction (PCMFI) mechanism in the encoder. It integrates the fine-grained textures and semantic information of optical images with the structural boundaries and spatial information of DSM, thereby facilitating more precise cross-modal feature interaction. Second, we design an adaptive dual-stream Mamba cross-modal fusion (ADMCF) module, which leverages a learnable variable mechanism to deeply represent global semantic and spatial structural information. This enhances deep semantic feature interaction and improves the ability of the model to distinguish complex land cover categories. Together, these modules progressively refine cross-modal cues and strengthen semantic interactions, enabling more coherent and discriminative multimodal fusion. Finally, we introduce a global-local feature decoder to effectively integrate the global and local information from the fused multimodal features. It preserves the structural integrity of target objects while enhancing edge detail representation, thus enhancing segmentation results. Through rigorous testing on standard datasets like ISPRS Vaihingen and Potsdam, the proposed HACMNet demonstrates advantages over prevailing methods in multimodal remote sensing analysis, particularly on challenging object classes.
近年来,多模态图像数据的增加为多模态语义分割提供了更广阔的前景。然而,不同模态之间的数据异构性使得难以利用互补信息并产生语义理解偏差,从而限制了融合质量和分割精度。为了克服这些挑战,我们提出了一种混合注意力驱动的CNN-Mamba多模态融合网络(HACMNet)用于语义分割。它旨在充分利用光学图像在纹理和语义表示方面的优势,以及来自数字表面模型(DSM)的互补结构和高程信息。这使得全局和局部互补信息的有效提取和组合能够在语义分割中达到更高的准确性和鲁棒性。具体来说,我们在编码器中提出了一种渐进式跨模态特征交互(PCMFI)机制。它将光学图像的细粒度纹理和语义信息与DSM的结构边界和空间信息相结合,从而实现更精确的跨模态特征交互。其次,设计了自适应双流曼巴跨模态融合(ADMCF)模块,利用可学习的变量机制深度表征全局语义和空间结构信息。这增强了深层语义特征交互,提高了模型区分复杂土地覆盖类别的能力。总之,这些模块逐步完善跨模态线索,加强语义交互,使多模态融合更加连贯和有区别。最后,我们引入了一个全局-局部特征解码器,以有效地整合融合后的多模态特征的全局和局部信息。在保持目标物体结构完整性的同时,增强边缘细节的表示,从而提高分割效果。通过对ISPRS Vaihingen和Potsdam等标准数据集的严格测试,拟议的HACMNet在多模态遥感分析中显示出优于主流方法的优势,特别是在具有挑战性的目标类上。
{"title":"Hybrid Attention Driven CNN-Mamba Multimodal Fusion Network for Remote Sensing Image Semantic Segmentation","authors":"Shu Tian;Minglei Li;Lin Cao;Lihong Kang;Jing Tian;Xiangwei Xing;Bo Shen;Kangning Du;Chong Fu;Ye Zhang","doi":"10.1109/JSTARS.2025.3644588","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3644588","url":null,"abstract":"In recent years, the increase of multimodal image data has offered a broader prospect for multimodal semantic segmentation. However, the data heterogeneity between different modalities make it difficult to leverage complementary information and create semantic understanding deviations, which limits the fusion quality and segmentation accuracy. To overcome these challenges, we propose a hybrid attention driven CNN-Mamba multimodal fusion network (HACMNet) for semantic segmentation. It aims to fully exploit the strengths of optical images in texture and semantic representation, along with the complementary structural and elevation information from the digital surface model (DSM). This enables the effective extraction and combination of global and local complementary information to achieve higher accuracy and robustness in semantic segmentation. Specifically, we propose a progressive cross-modal feature interaction (PCMFI) mechanism in the encoder. It integrates the fine-grained textures and semantic information of optical images with the structural boundaries and spatial information of DSM, thereby facilitating more precise cross-modal feature interaction. Second, we design an adaptive dual-stream Mamba cross-modal fusion (ADMCF) module, which leverages a learnable variable mechanism to deeply represent global semantic and spatial structural information. This enhances deep semantic feature interaction and improves the ability of the model to distinguish complex land cover categories. Together, these modules progressively refine cross-modal cues and strengthen semantic interactions, enabling more coherent and discriminative multimodal fusion. Finally, we introduce a global-local feature decoder to effectively integrate the global and local information from the fused multimodal features. It preserves the structural integrity of target objects while enhancing edge detail representation, thus enhancing segmentation results. Through rigorous testing on standard datasets like ISPRS Vaihingen and Potsdam, the proposed HACMNet demonstrates advantages over prevailing methods in multimodal remote sensing analysis, particularly on challenging object classes.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"2254-2272"},"PeriodicalIF":5.3,"publicationDate":"2025-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11300934","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145885035","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1