首页 > 最新文献

IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing最新文献

英文 中文
LESMI: Integrating Linear-Exponential Model, Shapelets, and Multirocket for Wetland Vegetation Inundation Monitoring With Time Series SAR 基于线性-指数模型、Shapelets和Multirocket的时间序列SAR湿地植被淹没监测
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-30 DOI: 10.1109/JSTARS.2025.3649200
Yuanye Cao;Xiuguo Liu;Yuannan Long;Hui Yang;Shixiong Yan;Qihao Chen
Accurate monitoring of wetland vegetation inundation is crucial for maintaining regional ecological balance and conserving biodiversity, serving as a fundamental prerequisite for wetland environmental monitoring and protection. The complex scattering characteristics of vegetation under different inundation conditions, combined with spatial and seasonal heterogeneity, pose significant challenges to precise vegetation inundation state identification. Therefore, this study proposes a novel approach named the linear-exponential model, shapelets, and multirocket integration (LESMI), for monitoring the inundation state and temporal changes of wetland vegetation using radar backscatter variation patterns. First, a new linear-exponential model is developed to characterize the backscatter-water depth relationship and represent the inundation state characteristics of wetland vegetation. Second, based on the typical inundated state of historical stages determined by the linear-exponential model, LESMI method innovatively combines the Shapelets with multirocket classification to efficiently extract multivariate key time periods features for inundation state identification and achieve large-scale, near real-time inundation state classification. Experimental results in the Dongting Lake wetland show that the proposed method achieves inundation recognition accuracies of 96.84% for reeds and 92.59% for grassland, outperforming traditional methods and LSTM deep learning by average margins of 12.95% and 1.87%, respectively. The linear-exponential model significantly enhances identification performance, improving accuracy by 5.64% and 3.83% compared to linear and normal distribution models. Monitoring from 2019 to 2021 demonstrates that LESMI effectively captures flood peak impacts on vegetation inundation and provides detailed classification of noninundated, shallow inundated, and deep inundated states, offering reliable technical support for dynamic wetland ecosystem monitoring and refined management.
湿地植被淹没的准确监测对维护区域生态平衡和保护生物多样性至关重要,是湿地环境监测与保护的基本前提。不同淹没条件下植被的复杂散射特性,加之空间和季节异质性,给植被淹没状态的精确识别带来了重大挑战。为此,本研究提出了一种基于雷达后向散射变化模式的线性指数模型、shapelets和多火箭集成(LESMI)方法来监测湿地植被的淹没状态和时间变化。首先,建立了一种新的线性-指数模型来表征湿地植被的后向散射-水深关系和淹没状态特征。其次,基于线性指数模型确定的历史阶段典型淹没状态,LESMI方法创新地将Shapelets与多火箭分类相结合,高效提取多变量关键时间段特征用于淹没状态识别,实现大规模、近实时的淹没状态分类。在洞庭湖湿地的实验结果表明,该方法对芦苇和草地的洪水识别准确率分别达到96.84%和92.59%,优于传统方法和LSTM深度学习,平均差值分别为12.95%和1.87%。线性-指数模型显著提高了识别性能,与线性和正态分布模型相比,准确率分别提高了5.64%和3.83%。2019 - 2021年监测结果表明,LESMI有效捕获了洪峰对植被淹没的影响,并提供了详细的非淹没、浅淹没和深淹没状态分类,为湿地生态系统动态监测和精细化管理提供了可靠的技术支持。
{"title":"LESMI: Integrating Linear-Exponential Model, Shapelets, and Multirocket for Wetland Vegetation Inundation Monitoring With Time Series SAR","authors":"Yuanye Cao;Xiuguo Liu;Yuannan Long;Hui Yang;Shixiong Yan;Qihao Chen","doi":"10.1109/JSTARS.2025.3649200","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3649200","url":null,"abstract":"Accurate monitoring of wetland vegetation inundation is crucial for maintaining regional ecological balance and conserving biodiversity, serving as a fundamental prerequisite for wetland environmental monitoring and protection. The complex scattering characteristics of vegetation under different inundation conditions, combined with spatial and seasonal heterogeneity, pose significant challenges to precise vegetation inundation state identification. Therefore, this study proposes a novel approach named the linear-exponential model, shapelets, and multirocket integration (LESMI), for monitoring the inundation state and temporal changes of wetland vegetation using radar backscatter variation patterns. First, a new linear-exponential model is developed to characterize the backscatter-water depth relationship and represent the inundation state characteristics of wetland vegetation. Second, based on the typical inundated state of historical stages determined by the linear-exponential model, LESMI method innovatively combines the Shapelets with multirocket classification to efficiently extract multivariate key time periods features for inundation state identification and achieve large-scale, near real-time inundation state classification. Experimental results in the Dongting Lake wetland show that the proposed method achieves inundation recognition accuracies of 96.84% for reeds and 92.59% for grassland, outperforming traditional methods and LSTM deep learning by average margins of 12.95% and 1.87%, respectively. The linear-exponential model significantly enhances identification performance, improving accuracy by 5.64% and 3.83% compared to linear and normal distribution models. Monitoring from 2019 to 2021 demonstrates that LESMI effectively captures flood peak impacts on vegetation inundation and provides detailed classification of noninundated, shallow inundated, and deep inundated states, offering reliable technical support for dynamic wetland ecosystem monitoring and refined management.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"4242-4256"},"PeriodicalIF":5.3,"publicationDate":"2025-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11319178","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146026501","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Study on Harmful Algal Blooms in the Waters Near the Yangtze River Estuary Based on Twin Satellites HY-1C/D COCTS Data 基于HY-1C/D COCTS双卫星数据的长江口海域有害藻华研究
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-30 DOI: 10.1109/JSTARS.2025.3649548
Xuan Liu;Lina Cai;Jiahua Li;Tianle Mao
This study examined a novel harmful algal bloom (HAB) inversion model (HABI) using domestic Chinese ocean color and temperature scanner multispectral data from HY-1C/D satellites. The model achieves the dual capabilities of HAB presence detection and density quantification, a key advancement over conventional binary classification models that lack the ability to delineate HAB density gradients. Key findings of this article include the following. 1) The HABI model uses spectral bands at 443, 490, and 565 nm, demonstrating superior performance in quantifying HAB density gradients compared to existing methods, with design adaptability to sensors featuring similar spectral configurations. 2) HABI achieved high inversion accuracy (R2 = 0.8682, RMSE = 0.09195, Recall = 0.9300, Precision = 0.949, F1-score = 0.939), showing strong consistency with Bulletin of China Marine Disaster and in situ HAB measurements in the Waters near the Yangtze River Estuary. 3) The distribution of HABs takes on obvious temporal and spatial change characteristics, with high density clusters localized in coastal zones, peaking in spring/summer, and changed seasonally. Their seasonal factors contributing to the change of HAB mainly include Yangtze River freshwater discharge and coastal upwelling, and modulated by physical (e.g., sea surface temperature), anthropogenic (e.g., industrial wastewater), and biogeochemical factors (e.g., dissolved inorganic nitrogen) as well as biodiversity. These findings are conceptually integrated in Fig. 14, synthesizing the model mechanics and spatio-temporal dynamics. The HABI algorithm proposed in this article can effectively applied for HAB monitoring and quantification, providing a technical support for near-shore ecological assessment and management.
利用海- 1c /D卫星海洋色温扫描仪多光谱数据,建立了一种新的有害藻华(HAB)反演模型。该模型实现了有害藻华存在检测和密度量化的双重功能,这是传统二元分类模型的一个关键进步,后者缺乏描述有害藻华密度梯度的能力。本文的主要发现包括以下内容。1) HABI模型使用443、490和565 nm的光谱波段,与现有方法相比,在量化HAB密度梯度方面表现出更好的性能,并且具有对具有相似光谱配置的传感器的设计适应性。2) HABI具有较高的反演精度(R2 = 0.8682, RMSE = 0.09195, Recall = 0.9300, Precision = 0.949, s1 -score = 0.939),与《中国海情公报》和长江口附近海域原位HAB测量结果具有较强的一致性。3)HABs分布具有明显的时空变化特征,高密度聚集集中在海岸带,春夏季达到峰值,季节性变化明显。影响赤潮变化的季节性因子主要包括长江淡水排放和沿岸上升流,并受物理因子(如海表温度)、人为因子(如工业废水)、生物地球化学因子(如溶解无机氮)和生物多样性的调节。综合模型力学和时空动力学,这些发现在概念上集成在图14中。本文提出的HABI算法可有效应用于有害藻华的监测与量化,为近岸生态评价与管理提供技术支持。
{"title":"Study on Harmful Algal Blooms in the Waters Near the Yangtze River Estuary Based on Twin Satellites HY-1C/D COCTS Data","authors":"Xuan Liu;Lina Cai;Jiahua Li;Tianle Mao","doi":"10.1109/JSTARS.2025.3649548","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3649548","url":null,"abstract":"This study examined a novel harmful algal bloom (HAB) inversion model (HABI) using domestic Chinese ocean color and temperature scanner multispectral data from HY-1C/D satellites. The model achieves the dual capabilities of HAB presence detection and density quantification, a key advancement over conventional binary classification models that lack the ability to delineate HAB density gradients. Key findings of this article include the following. 1) The HABI model uses spectral bands at 443, 490, and 565 nm, demonstrating superior performance in quantifying HAB density gradients compared to existing methods, with design adaptability to sensors featuring similar spectral configurations. 2) HABI achieved high inversion accuracy (<italic>R</i><sup>2</sup> = 0.8682, RMSE = 0.09195, Recall = 0.9300, Precision = 0.949, F1-score = 0.939), showing strong consistency with Bulletin of China Marine Disaster and in situ HAB measurements in the Waters near the Yangtze River Estuary. 3) The distribution of HABs takes on obvious temporal and spatial change characteristics, with high density clusters localized in coastal zones, peaking in spring/summer, and changed seasonally. Their seasonal factors contributing to the change of HAB mainly include Yangtze River freshwater discharge and coastal upwelling, and modulated by physical (e.g., sea surface temperature), anthropogenic (e.g., industrial wastewater), and biogeochemical factors (e.g., dissolved inorganic nitrogen) as well as biodiversity. These findings are conceptually integrated in Fig. 14, synthesizing the model mechanics and spatio-temporal dynamics. The HABI algorithm proposed in this article can effectively applied for HAB monitoring and quantification, providing a technical support for near-shore ecological assessment and management.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"4872-4886"},"PeriodicalIF":5.3,"publicationDate":"2025-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11319152","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146082029","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Novel Network for Change Detection Based on a Divide-and-Conquer Fusion Strategy 一种基于分而治之融合策略的变化检测网络
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-26 DOI: 10.1109/JSTARS.2025.3648330
Mengmeng Wang;Xu Lin;Yuanxin Ye;Wenhui Wu;Bai Zhu;Yanshuai Dai
Change detection (CD) is a fundamental task that is pivotal in understanding surface changes. Recently, CD methods have advanced rapidly and attained impressive results, driven by deep learning technology. However, existing methods generally employ fusion modules with the same design for multilevel features, overlooking the inherent distinctions between low-level spatial features and deep-level semantic features generated by deep networks. To overcome this limitation, this article proposes a novel CD network, referred to as DACNet. This method introduces a divide-and-conquer fusion strategy designed to fuse multilevel features using different fusion strategies. Specifically, the widely used MobileNetV2 is employed within a dual-branch architecture to extract multilevel features from bitemporal images. Subsequently, the proposed divide-and-conquer fusion strategy comprises two specialized modules: the change region localization module and the edge complementarity module, which are tailored to fuse deep-level semantic features and low-level spatial features, respectively. In addition, to mitigate the unnecessary noise introduced by the conventional UNet architectures, attention gates are introduced into the UNet decoder to enhance the changed information and suppress background noises. Extensive experiments are conducted on three available CD datasets: LEVIR-CD, Google-CD, and MSRS-CD. The proposed network achieved favorable results compared to the nine state-of-the-art methods across all experiments, improving the F1 score by 0.93%, 1.10%, and 0.81% on the LEVIR-CD, Google-CD, and MSRS-CD datasets, respectively.
变化检测是了解地表变化的一项基础性工作。近年来,在深度学习技术的推动下,CD方法发展迅速,取得了令人印象深刻的成果。然而,现有方法通常采用相同设计的融合模块来处理多层特征,忽略了深层网络生成的底层空间特征和深层语义特征之间的内在区别。为了克服这一限制,本文提出了一种新的CD网络,称为DACNet。该方法引入了一种分而治之的融合策略,旨在使用不同的融合策略融合多层次特征。具体而言,广泛使用的MobileNetV2在双分支架构中用于从双时间图像中提取多层特征。随后,本文提出的分而治之融合策略包括两个专门的模块:变化区域定位模块和边缘互补模块,分别针对深层语义特征和底层空间特征进行融合。此外,为了减轻传统UNet结构带来的不必要的噪声,在UNet解码器中引入了注意门,以增强变化信息并抑制背景噪声。在三个可用的CD数据集:LEVIR-CD、Google-CD和MSRS-CD上进行了广泛的实验。该网络在所有实验中均取得了较好的结果,在LEVIR-CD、Google-CD和MSRS-CD数据集上的F1得分分别提高了0.93%、1.10%和0.81%。
{"title":"A Novel Network for Change Detection Based on a Divide-and-Conquer Fusion Strategy","authors":"Mengmeng Wang;Xu Lin;Yuanxin Ye;Wenhui Wu;Bai Zhu;Yanshuai Dai","doi":"10.1109/JSTARS.2025.3648330","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3648330","url":null,"abstract":"Change detection (CD) is a fundamental task that is pivotal in understanding surface changes. Recently, CD methods have advanced rapidly and attained impressive results, driven by deep learning technology. However, existing methods generally employ fusion modules with the same design for multilevel features, overlooking the inherent distinctions between low-level spatial features and deep-level semantic features generated by deep networks. To overcome this limitation, this article proposes a novel CD network, referred to as DACNet. This method introduces a divide-and-conquer fusion strategy designed to fuse multilevel features using different fusion strategies. Specifically, the widely used MobileNetV2 is employed within a dual-branch architecture to extract multilevel features from bitemporal images. Subsequently, the proposed divide-and-conquer fusion strategy comprises two specialized modules: the change region localization module and the edge complementarity module, which are tailored to fuse deep-level semantic features and low-level spatial features, respectively. In addition, to mitigate the unnecessary noise introduced by the conventional UNet architectures, attention gates are introduced into the UNet decoder to enhance the changed information and suppress background noises. Extensive experiments are conducted on three available CD datasets: LEVIR-CD, Google-CD, and MSRS-CD. The proposed network achieved favorable results compared to the nine state-of-the-art methods across all experiments, improving the F1 score by 0.93%, 1.10%, and 0.81% on the LEVIR-CD, Google-CD, and MSRS-CD datasets, respectively.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"2891-2904"},"PeriodicalIF":5.3,"publicationDate":"2025-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11316246","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145982153","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HyM3S: Integrating Multiscale Spatial–Spectral Features With Sequence Modeling for Hyperspectral Classification HyM3S:融合多尺度空间光谱特征与序列建模的高光谱分类
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-24 DOI: 10.1109/JSTARS.2025.3647643
Yin Chen;Shaoqun Qi;Luhe Wan;Chunlong Du;Zhiwei Lin;Ling Zhu;Xiaona Yu
Transformer models have been widely adopted for hyperspectral image (HSI) classification due to their exceptional long-sequence modeling capabilities. However, the self-attention mechanism in Transformers incurs quadratic computational complexity, posing challenges in both speed and memory consumption. Recently, a novel state-space model—the Mamba model—has emerged, overcoming the quadratic complexity of self-attention by achieving linear computational complexity while retaining powerful long-sequence modeling. Yet, the original Mamba design does not account for the unique spectral–spatial characteristics of HSI data, making it difficult to capture multiscale features. This limitation can lead to the loss of critical spectral–spatial cues at fine targets and complex boundaries, resulting in increased classification noise, blurred boundary segmentation, and reduced overall accuracy. To address the loss of fine–grained spectral–spatial information in HSI, we propose HyM3S: a hyperspectral multiscale spatial–spectral sequence model that integrates multiscale spatial–spectral convolutions with Mamba’s linear sequence modeling.HyM3S first extracts multiscale spatial and spectral features in parallel along horizontal and vertical branches, and reinforces salient channels via channel–wise attention. Features are then adaptively fused across modality and directional dimensions to form a unified joint representation. Finally, this representation is fed into the Mamba module for long–range dependency modeling under linear complexity, thereby significantly improving classification accuracy and suppressing noise. Experiments on four benchmark HSI datasets—Pavia University (PaviaU), Houston2013, WHU–Hi–HanChuan, and WHU–Hi–HongHu—demonstrate the clear superiority of the proposed HyM3S model for HSI classification.
变压器模型由于其特殊的长序列建模能力而被广泛应用于高光谱图像(HSI)分类。然而,变形金刚的自注意机制带来了二次计算复杂度,在速度和内存消耗方面都提出了挑战。最近出现了一种新的状态空间模型——曼巴模型,它在保持强大的长序列建模的同时,通过实现线性计算复杂度,克服了自注意的二次复杂性。然而,最初的曼巴设计并没有考虑到HSI数据独特的光谱空间特征,这使得捕捉多尺度特征变得困难。这种限制可能导致在精细目标和复杂边界处丢失关键的光谱空间线索,从而导致分类噪声增加、边界分割模糊和整体精度降低。为了解决HSI中细粒度光谱空间信息的丢失问题,我们提出了HyM3S:一种集成了多尺度空间光谱卷积和曼巴线性序列建模的高光谱多尺度空间光谱序列模型。HyM3S首先沿水平和垂直分支平行提取多尺度空间和光谱特征,并通过通道智能关注强化突出通道。然后在模态和方向维度上自适应融合特征,形成统一的联合表示。最后,将该表示形式输入到Mamba模块中进行线性复杂度下的远程依赖建模,从而显著提高了分类精度并抑制了噪声。在帕维亚大学(PaviaU)、休斯顿2013、WHU-Hi-HanChuan和whu - hi - honghu四个基准HSI数据集上的实验表明,所提出的HyM3S模型在HSI分类方面具有明显的优势。
{"title":"HyM3S: Integrating Multiscale Spatial–Spectral Features With Sequence Modeling for Hyperspectral Classification","authors":"Yin Chen;Shaoqun Qi;Luhe Wan;Chunlong Du;Zhiwei Lin;Ling Zhu;Xiaona Yu","doi":"10.1109/JSTARS.2025.3647643","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3647643","url":null,"abstract":"Transformer models have been widely adopted for hyperspectral image (HSI) classification due to their exceptional long-sequence modeling capabilities. However, the self-attention mechanism in Transformers incurs quadratic computational complexity, posing challenges in both speed and memory consumption. Recently, a novel state-space model—the Mamba model—has emerged, overcoming the quadratic complexity of self-attention by achieving linear computational complexity while retaining powerful long-sequence modeling. Yet, the original Mamba design does not account for the unique spectral–spatial characteristics of HSI data, making it difficult to capture multiscale features. This limitation can lead to the loss of critical spectral–spatial cues at fine targets and complex boundaries, resulting in increased classification noise, blurred boundary segmentation, and reduced overall accuracy. To address the loss of fine–grained spectral–spatial information in HSI, we propose HyM3S: a hyperspectral multiscale spatial–spectral sequence model that integrates multiscale spatial–spectral convolutions with Mamba’s linear sequence modeling.HyM3S first extracts multiscale spatial and spectral features in parallel along horizontal and vertical branches, and reinforces salient channels via channel–wise attention. Features are then adaptively fused across modality and directional dimensions to form a unified joint representation. Finally, this representation is fed into the Mamba module for long–range dependency modeling under linear complexity, thereby significantly improving classification accuracy and suppressing noise. Experiments on four benchmark HSI datasets—Pavia University (PaviaU), Houston2013, WHU–Hi–HanChuan, and WHU–Hi–HongHu—demonstrate the clear superiority of the proposed HyM3S model for HSI classification.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"3190-3205"},"PeriodicalIF":5.3,"publicationDate":"2025-12-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11314656","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145982175","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automatic RFI Detection, Location, and Classification System in GNSS Bands GNSS波段自动RFI检测、定位和分类系统
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-24 DOI: 10.1109/JSTARS.2025.3647776
Adrian Perez-Portero;Jorge Querol;Adriano Camps
Global navigation satellite systems (GNSSs) are critical infrastructure components in modern positioning, navigation, and timing (PNT) services, playing a vital role in both civilian and defense applications. These systems operate in specific frequency bands that are also utilized by other Earth Observation technologies, such as GNSS-radio occultations and GNSS-reflectometry. Other passive microwave remote sensing techniques such as microwave radiometers, work with very faint signals in nearby frequency bands within the L-Band. However, the increasing prevalence of radio-frequency interferences (RFIs) poses a significant threat, potentially compromising the integrity and reliability of PNT services, and corrupting geophysical observations. Effective RFI mitigation relies on accurate detection and classification of interference sources, a task that becomes increasingly challenging due to the complexity and diversity of RFI signals. This work presents an automated classification system for RFI detection and characterization in GNSS bands. The methodology employs advanced digital signal processing techniques and statistical algorithms to improve RFI detection and classification. RFI events are then stored in a long-term database to provide insights into the local spectrum, and to aid in mitigation and law enforcement efforts. This study provides a description of the classification system, including its architecture, implementation, and performance analysis. The results highlight the potential of this system to enhance the resilience of GNSS PNT services against RFI.
全球导航卫星系统(gnss)是现代定位、导航和授时(PNT)服务的关键基础设施组件,在民用和国防应用中都发挥着至关重要的作用。这些系统在其他地球观测技术(如gnss无线电掩星和gnss反射计)也利用的特定频带中运行。其他无源微波遥感技术,如微波辐射计,在l波段内附近频段的非常微弱的信号中工作。然而,日益普遍的射频干扰(rfi)构成了重大威胁,可能损害PNT服务的完整性和可靠性,并破坏地球物理观测。有效的RFI缓解依赖于对干扰源的准确检测和分类,由于RFI信号的复杂性和多样性,这一任务变得越来越具有挑战性。这项工作提出了一个自动分类系统,用于GNSS波段的RFI检测和表征。该方法采用先进的数字信号处理技术和统计算法来改进射频信号的检测和分类。然后将RFI事件存储在一个长期数据库中,以提供对当地频谱的见解,并有助于缓解和执法工作。本研究提供了分类系统的描述,包括其体系结构、实现和性能分析。结果强调了该系统在增强GNSS PNT服务对RFI的弹性方面的潜力。
{"title":"Automatic RFI Detection, Location, and Classification System in GNSS Bands","authors":"Adrian Perez-Portero;Jorge Querol;Adriano Camps","doi":"10.1109/JSTARS.2025.3647776","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3647776","url":null,"abstract":"Global navigation satellite systems (GNSSs) are critical infrastructure components in modern positioning, navigation, and timing (PNT) services, playing a vital role in both civilian and defense applications. These systems operate in specific frequency bands that are also utilized by other Earth Observation technologies, such as GNSS-radio occultations and GNSS-reflectometry. Other passive microwave remote sensing techniques such as microwave radiometers, work with very faint signals in nearby frequency bands within the L-Band. However, the increasing prevalence of radio-frequency interferences (RFIs) poses a significant threat, potentially compromising the integrity and reliability of PNT services, and corrupting geophysical observations. Effective RFI mitigation relies on accurate detection and classification of interference sources, a task that becomes increasingly challenging due to the complexity and diversity of RFI signals. This work presents an automated classification system for RFI detection and characterization in GNSS bands. The methodology employs advanced digital signal processing techniques and statistical algorithms to improve RFI detection and classification. RFI events are then stored in a long-term database to provide insights into the local spectrum, and to aid in mitigation and law enforcement efforts. This study provides a description of the classification system, including its architecture, implementation, and performance analysis. The results highlight the potential of this system to enhance the resilience of GNSS PNT services against RFI.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"3072-3084"},"PeriodicalIF":5.3,"publicationDate":"2025-12-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11314687","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145982198","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TFST: Two-Frame Ship Tracking for SAR Using YOLOv12 and Feature-Based Matching TFST:基于YOLOv12和特征匹配的SAR双帧船舶跟踪
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-24 DOI: 10.1109/JSTARS.2025.3647680
Muhammad Yasir;Shanwei Liu;Mingming Xu;Fernando J. Aguilar;Jianhua Wan;Shiqing Wei;Saied Pirasteh;Hong Fan;Qamar Ul Islam
Tracking objects in synthetic aperture radar (SAR) imagery is critical for maritime surveillance, traffic monitoring, and security applications, but remains a major challenge due to speckle noise, sea clutter, and limited temporal continuity. Most existing tracking-by-detection methods process frames independently, often resulting in weak associations and frequent identity switches (IDs). To overcome these limitations, we propose TFST, a two-frame SAR ship tracking framework that integrates detection, feature encoding, and optimal assignment. In this way, the goal of this work is to address the current gaps in SAR ship tracking by strengthening cross-frame partnerships and reducing IDs through an integrated two-frame tracking framework. In our approach, a deep detector first processes consecutive frames to generate candidate bounding boxes. A lightweight feature extractor encodes both appearance and structural cues, while a matching module constructs a cost matrix that combines feature similarity and positional consistency. Gating is applied to remove infeasible associations, and the Hungarian algorithm is employed to achieve a globally optimal assignment. Quantitative evaluations performed on three widely known and publicly available SAR-Ship datasets (SSTD, SSDD, and SAR-Ship) further highlight the advantages of TFST. In terms of ship detection performance, TFST achieved an average mAP@50 improvement of 2.2% over the YOLOv12 baseline model on all three tested datasets. Regarding tracking results, the superiority of TFST over state-of-the-art multiobject trackers becomes even more evident. In fact, the proposed model achieved the highest multiple object tracking accuracy (MOTA) (86.9%) and the best IDF1 score (82.7%), thus outperforming strong baselines such as Siam-SORT (82.1% MOTA and 79.8% IDF1) and TrackFormer (80.7% MOTA and 78.7% IDF1). In conclusion, TFST demonstrated improved robustness, fewer ID switches, and higher tracking accuracy compared to baseline methods, underscoring its effectiveness in complex maritime environments.
合成孔径雷达(SAR)图像中的目标跟踪对于海上监视、交通监控和安全应用至关重要,但由于散斑噪声、海杂波和有限的时间连续性,仍然是一个主要挑战。大多数现有的检测跟踪方法都是独立处理帧的,往往导致弱关联和频繁的身份切换(id)。为了克服这些限制,我们提出了TFST,这是一种集成了检测、特征编码和最优分配的两帧SAR船舶跟踪框架。通过这种方式,这项工作的目标是通过加强跨框架伙伴关系和通过集成的两框架跟踪框架减少id来解决当前SAR船舶跟踪方面的差距。在我们的方法中,深度检测器首先处理连续帧以生成候选边界框。轻量级特征提取器对外观和结构线索进行编码,而匹配模块构建结合特征相似性和位置一致性的代价矩阵。采用门控法去除不可行关联,采用匈牙利算法实现全局最优分配。对三个广为人知且公开可用的SAR-Ship数据集(SSTD、SSDD和SAR-Ship)进行的定量评估进一步突出了TFST的优势。在船舶检测性能方面,TFST在所有三个测试数据集上比YOLOv12基线模型平均mAP@50提高了2.2%。在跟踪结果方面,TFST相对于最先进的多目标跟踪器的优势更加明显。事实上,该模型实现了最高的多目标跟踪精度(MOTA)(86.9%)和最佳的IDF1得分(82.7%),从而优于Siam-SORT (82.1% MOTA和79.8% IDF1)和TrackFormer (80.7% MOTA和78.7% IDF1)等强基线。总之,与基线方法相比,TFST具有更好的鲁棒性、更少的ID切换和更高的跟踪精度,强调了其在复杂海洋环境中的有效性。
{"title":"TFST: Two-Frame Ship Tracking for SAR Using YOLOv12 and Feature-Based Matching","authors":"Muhammad Yasir;Shanwei Liu;Mingming Xu;Fernando J. Aguilar;Jianhua Wan;Shiqing Wei;Saied Pirasteh;Hong Fan;Qamar Ul Islam","doi":"10.1109/JSTARS.2025.3647680","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3647680","url":null,"abstract":"Tracking objects in synthetic aperture radar (SAR) imagery is critical for maritime surveillance, traffic monitoring, and security applications, but remains a major challenge due to speckle noise, sea clutter, and limited temporal continuity. Most existing tracking-by-detection methods process frames independently, often resulting in weak associations and frequent identity switches (IDs). To overcome these limitations, we propose TFST, a two-frame SAR ship tracking framework that integrates detection, feature encoding, and optimal assignment. In this way, the goal of this work is to address the current gaps in SAR ship tracking by strengthening cross-frame partnerships and reducing IDs through an integrated two-frame tracking framework. In our approach, a deep detector first processes consecutive frames to generate candidate bounding boxes. A lightweight feature extractor encodes both appearance and structural cues, while a matching module constructs a cost matrix that combines feature similarity and positional consistency. Gating is applied to remove infeasible associations, and the Hungarian algorithm is employed to achieve a globally optimal assignment. Quantitative evaluations performed on three widely known and publicly available SAR-Ship datasets (<italic>SSTD, SSDD</i>, and <italic>SAR-Ship</i>) further highlight the advantages of TFST. In terms of ship detection performance, TFST achieved an average <italic>mAP@50</i> improvement of 2.2% over the YOLOv12 baseline model on all three tested datasets. Regarding tracking results, the superiority of TFST over state-of-the-art multiobject trackers becomes even more evident. In fact, the proposed model achieved the highest multiple object tracking accuracy (MOTA) (86.9%) and the best <italic>IDF1</i> score (82.7%), thus outperforming strong baselines such as <italic>Siam-SORT</i> (82.1% <italic>MOTA</i> and 79.8% <italic>IDF1</i>) and <italic>TrackFormer</i> (80.7% <italic>MOTA</i> and 78.7% <italic>IDF1</i>). In conclusion, TFST demonstrated improved robustness, fewer <italic>ID</i> switches, and higher tracking accuracy compared to baseline methods, underscoring its effectiveness in complex maritime environments.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"3175-3189"},"PeriodicalIF":5.3,"publicationDate":"2025-12-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11314707","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145982142","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SI-Mamba: High-Resolution Sea Ice Recognition via RB-NDI Guided State-Space Model SI-Mamba:基于RB-NDI引导状态空间模型的高分辨率海冰识别
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-24 DOI: 10.1109/JSTARS.2025.3648014
Wenjun Hong;Zhanchao Huang;Yongke Yang;Luping You;Junchao Cai;Jiajun Zhou;Weiwang Guan;Hua Su
Sea ice recognition is of great significance for reflecting climate change and ensuring ship navigation safety. In recent years, many deep learning-based methods have been proposed and applied to the segmentation and recognition of sea ice regions. However, existing deep learning models often struggle to effectively capture the subtle spectral differences between sea ice and seawater, as well as the large-scale spatial dependencies in high-resolution remote sensing images, resulting in limited segmentation accuracy in areas with ambiguous ice–water boundaries. This article proposes a red–blue normalized difference index (RB-NDI) guided state-space model (SSM) approach for sea ice segmentation, termed SI-Mamba. In the proposed SI-Mamba, the index-guided collaborative enhancement module employs an RB-NDI index-guided SSM mechanism to overcome the limitations in explicit modeling of spectral features, achieving efficient modeling of long-range spatial dependencies in sea ice distribution. Furthermore, the designed dynamic boundary focus loss function adjusts the model’s expressive capability in edge-sensitive regions through collaborative optimization between the main segmentation head and the index-assisted head. Experiments on multiple generated sea ice datasets demonstrate that the proposed SI-Mamba achieves improved performance in sea ice segmentation and identification in optical remote sensing images. It significantly enhances the accuracy of ice–water boundaries recognition and generalization capability in complex scenarios, offering a novel and effective solution for remote sensing-based sea ice recognition.
海冰识别对于反映气候变化、保障船舶航行安全具有重要意义。近年来,许多基于深度学习的方法被提出并应用于海冰区域的分割和识别。然而,现有的深度学习模型往往难以有效捕获海冰和海水之间细微的光谱差异,以及高分辨率遥感图像中的大尺度空间依赖关系,导致在冰-水边界不明确的地区分割精度有限。本文提出了一种红蓝归一化差分指数(RB-NDI)引导的海冰分割状态空间模型(SSM)方法,称为SI-Mamba。在SI-Mamba中,指数导向的协同增强模块采用RB-NDI指数导向的SSM机制,克服了光谱特征显式建模的局限性,实现了海冰分布远程空间依赖关系的高效建模。此外,设计的动态边界焦点损失函数通过主分割头和索引辅助头之间的协同优化来调整模型在边缘敏感区域的表达能力。在多个生成的海冰数据集上进行的实验表明,SI-Mamba在光学遥感图像中海冰分割和识别方面取得了较好的效果。该方法显著提高了复杂场景下海冰边界识别的精度和概化能力,为基于遥感的海冰识别提供了一种新颖有效的解决方案。
{"title":"SI-Mamba: High-Resolution Sea Ice Recognition via RB-NDI Guided State-Space Model","authors":"Wenjun Hong;Zhanchao Huang;Yongke Yang;Luping You;Junchao Cai;Jiajun Zhou;Weiwang Guan;Hua Su","doi":"10.1109/JSTARS.2025.3648014","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3648014","url":null,"abstract":"Sea ice recognition is of great significance for reflecting climate change and ensuring ship navigation safety. In recent years, many deep learning-based methods have been proposed and applied to the segmentation and recognition of sea ice regions. However, existing deep learning models often struggle to effectively capture the subtle spectral differences between sea ice and seawater, as well as the large-scale spatial dependencies in high-resolution remote sensing images, resulting in limited segmentation accuracy in areas with ambiguous ice–water boundaries. This article proposes a red–blue normalized difference index (RB-NDI) guided state-space model (SSM) approach for sea ice segmentation, termed SI-Mamba. In the proposed SI-Mamba, the index-guided collaborative enhancement module employs an RB-NDI index-guided SSM mechanism to overcome the limitations in explicit modeling of spectral features, achieving efficient modeling of long-range spatial dependencies in sea ice distribution. Furthermore, the designed dynamic boundary focus loss function adjusts the model’s expressive capability in edge-sensitive regions through collaborative optimization between the main segmentation head and the index-assisted head. Experiments on multiple generated sea ice datasets demonstrate that the proposed SI-Mamba achieves improved performance in sea ice segmentation and identification in optical remote sensing images. It significantly enhances the accuracy of ice–water boundaries recognition and generalization capability in complex scenarios, offering a novel and effective solution for remote sensing-based sea ice recognition.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"2783-2795"},"PeriodicalIF":5.3,"publicationDate":"2025-12-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11314692","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145982154","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Lighter: A Lightweight Full-Information and Dual-Guide Network for Remote Sensing Image Change Detection 打火机:用于遥感图像变化检测的轻量级全信息双导网络
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-24 DOI: 10.1109/JSTARS.2025.3647926
Yuan Wang;Sixian Chan;Guoyu Yang;Jian Tao;Tianyang Dong;Xiaolong Zhou;Xiaoqin Zhang
Remote sensing (RS) change detection (CD), a technique focused on identifying surface alterations from bitemporal images, holds substantial significance for various applications, such as land management and disaster surveillance. In the last decade, deep learning-based CD methods have advanced rapidly. However, recognizing sporadic distributional changes over time in complex scenes remains a significant challenge. However, many existing solutions rely on large model capacities and high computational costs, yet still fail to incorporate sufficient semantic information for accurately recognizing complex real changes. To tackle this challenge, a lightweight full-information and dual-guide network (Lighter) for RSCD is presented. Specifically, we design a lightweight full-information mingling module that emphasizes the injection of multiperspective information during the feature interaction. This approach leverages rich semantics as cues to reason about diverse changes. Furthermore, we propose a lightweight dual-guide difference capture module, which utilizes the unique information of each guide to teach each other thereby reducing the interference of pseudovariations. Extensive experiments on four datasets demonstrate that our lightweight architecture achieves state-of-the-art performance with only 1.10 M parameters and 2.01 G FLOPs.
遥感(RS)变化检测(CD)是一种从双时相图像中识别地表变化的技术,在土地管理和灾害监测等各种应用中具有重要意义。在过去的十年中,基于深度学习的CD方法发展迅速。然而,在复杂的场景中识别零星的分布变化仍然是一个重大的挑战。然而,许多现有的解决方案依赖于大的模型容量和高的计算成本,但仍然不能包含足够的语义信息来准确识别复杂的真实变化。为了应对这一挑战,RSCD提出了一种轻量级的全信息双导网络(Lighter)。具体来说,我们设计了一个轻量级的全信息混合模块,强调在功能交互过程中注入多视角信息。这种方法利用丰富的语义作为推理各种变化的线索。此外,我们提出了一种轻量级的双导差分捕获模块,该模块利用每个导的独特信息相互教导,从而减少伪变异的干扰。在四个数据集上进行的大量实验表明,我们的轻量级架构仅以1.10 M参数和2.01 G FLOPs实现了最先进的性能。
{"title":"Lighter: A Lightweight Full-Information and Dual-Guide Network for Remote Sensing Image Change Detection","authors":"Yuan Wang;Sixian Chan;Guoyu Yang;Jian Tao;Tianyang Dong;Xiaolong Zhou;Xiaoqin Zhang","doi":"10.1109/JSTARS.2025.3647926","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3647926","url":null,"abstract":"Remote sensing (RS) change detection (CD), a technique focused on identifying surface alterations from bitemporal images, holds substantial significance for various applications, such as land management and disaster surveillance. In the last decade, deep learning-based CD methods have advanced rapidly. However, recognizing sporadic distributional changes over time in complex scenes remains a significant challenge. However, many existing solutions rely on large model capacities and high computational costs, yet still fail to incorporate sufficient semantic information for accurately recognizing complex real changes. To tackle this challenge, a lightweight full-information and dual-guide network (Lighter) for RSCD is presented. Specifically, we design a lightweight full-information mingling module that emphasizes the injection of multiperspective information during the feature interaction. This approach leverages rich semantics as cues to reason about diverse changes. Furthermore, we propose a lightweight dual-guide difference capture module, which utilizes the unique information of each guide to teach each other thereby reducing the interference of pseudovariations. Extensive experiments on four datasets demonstrate that our lightweight architecture achieves state-of-the-art performance with only 1.10 M parameters and 2.01 G FLOPs.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"2924-2937"},"PeriodicalIF":5.3,"publicationDate":"2025-12-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11314198","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145982250","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multimodal Collaborative Interactive Soft Fusion Network for RGB-Infrared Aerial Image Object Detection rgb -红外航拍图像目标检测的多模态协同交互软融合网络
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-24 DOI: 10.1109/JSTARS.2025.3648023
Bei Cheng;Beihao Xu;Wenjie Gan;Qingwang Wang
Multimodal object detection plays a crucial role in all-weather and multiscene applications of aerial imagery. Existing studies mainly focus on multimodal fusion and interlevel feature interaction during feature extraction, that is, correcting or enhancing dual-branch weights through fused features or multimodal interactions, while neglecting the supplementation of missing modality features. This limitation can lead to noise propagation across layers and a reduction in interaction capability caused by feature absence. In this article, we propose a multimodal collaborative interactive soft fusion network (MCISFNet) for RGB-infrared aerial image object detection. The proposed method introduces a saliency-guided multimodal soft fusion mechanism (SMSFM), which explicitly directs attention to and enhances critical regions, dynamically adjusts feature weights, and integrates complementary information to mitigate the problem of missing data in dual-branch representations. To address the complexity of aerial scenarios, we further develop a multiscale interactive gating module (MIGM) that explicitly incorporates multiscale contextual information, enabling fine-grained refinement of primary modality features and enhancing the discriminability of fused representations. Moreover, we design a cross-modal global context collaborative modeling (CGCCM) strategy, in which a cross-modal shared branch is constructed to jointly perform context extraction and feature fusion. This collaborative design not only improves the alignment of deep semantic features but also ensures that the learned RGB and IR features are more consistent and complementary, while reducing computational cost. Extensive experiments conducted on three multimodal aerial image detection datasets (DroneVehicle, VEDAI, and ODinMJ) demonstrate the robustness and generalization capability of the proposed MCISFNet framework.
多模态目标检测在航空影像的全天候、多场景应用中起着至关重要的作用。现有研究主要集中在特征提取过程中的多模态融合和层间特征交互,即通过融合特征或多模态交互对双分支权值进行校正或增强,而忽略了缺失模态特征的补充。这种限制会导致噪声跨层传播,并且由于特征缺失导致交互能力降低。在本文中,我们提出了一种用于rgb -红外航拍图像目标检测的多模态协同交互软融合网络(MCISFNet)。该方法引入了一种显著性引导的多模态软融合机制(SMSFM),明确引导和增强关键区域的注意力,动态调整特征权重,并集成互补信息,以缓解双分支表示中的数据缺失问题。为了解决空中场景的复杂性,我们进一步开发了一个多尺度交互式门控模块(MIGM),该模块明确地融合了多尺度上下文信息,实现了对主要模态特征的细粒度细化,并增强了融合表征的可辨别性。此外,我们设计了一种跨模态全局上下文协同建模(CGCCM)策略,该策略构建了一个跨模态共享分支,共同进行上下文提取和特征融合。这种协同设计不仅提高了深层语义特征的对齐,而且保证了学习到的RGB和IR特征更加一致和互补,同时降低了计算成本。在三个多模式航空图像检测数据集(dronevvehicle、VEDAI和ODinMJ)上进行的大量实验证明了所提出的MCISFNet框架的鲁棒性和泛化能力。
{"title":"Multimodal Collaborative Interactive Soft Fusion Network for RGB-Infrared Aerial Image Object Detection","authors":"Bei Cheng;Beihao Xu;Wenjie Gan;Qingwang Wang","doi":"10.1109/JSTARS.2025.3648023","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3648023","url":null,"abstract":"Multimodal object detection plays a crucial role in all-weather and multiscene applications of aerial imagery. Existing studies mainly focus on multimodal fusion and interlevel feature interaction during feature extraction, that is, correcting or enhancing dual-branch weights through fused features or multimodal interactions, while neglecting the supplementation of missing modality features. This limitation can lead to noise propagation across layers and a reduction in interaction capability caused by feature absence. In this article, we propose a multimodal collaborative interactive soft fusion network (MCISFNet) for RGB-infrared aerial image object detection. The proposed method introduces a saliency-guided multimodal soft fusion mechanism (SMSFM), which explicitly directs attention to and enhances critical regions, dynamically adjusts feature weights, and integrates complementary information to mitigate the problem of missing data in dual-branch representations. To address the complexity of aerial scenarios, we further develop a multiscale interactive gating module (MIGM) that explicitly incorporates multiscale contextual information, enabling fine-grained refinement of primary modality features and enhancing the discriminability of fused representations. Moreover, we design a cross-modal global context collaborative modeling (CGCCM) strategy, in which a cross-modal shared branch is constructed to jointly perform context extraction and feature fusion. This collaborative design not only improves the alignment of deep semantic features but also ensures that the learned RGB and IR features are more consistent and complementary, while reducing computational cost. Extensive experiments conducted on three multimodal aerial image detection datasets (DroneVehicle, VEDAI, and ODinMJ) demonstrate the robustness and generalization capability of the proposed MCISFNet framework.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"3206-3218"},"PeriodicalIF":5.3,"publicationDate":"2025-12-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11314552","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145982263","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Multimodal Remote Sensing Method for Mangrove Species Classification Based on Sentinel-2 Imagery 基于Sentinel-2影像的红树林物种分类多模态遥感方法
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-24 DOI: 10.1109/JSTARS.2025.3647732
Haiwei Yuan;Shikuan Wang;Jianzhou Gong
Intertidal mangrove communities present significant challenges for remote sensing classification due to tidal dynamics, intertwining canopies, and spectral confusion between species. These factors often lead to classification errors and indistinct species boundaries, primarily due to intraspectral variation and interspectral similarity, and also limit the capture of vertical structural features. To address these issues, we propose MSMNet, a specialized multimodal deep learning method for fine-grained mangrove species classification based on Sentinel-2 multispectral imagery. The model uses ResNet50 as its backbone architecture and integrates the Mamba state-space module to model long-range spatial correlations. MSMNet incorporates wavelet transform technology to enhance its ability to represent textures and uses three pathways to extract three key mangrove remote sensing modalities: morphological texture, spectral features, and vegetation physiological characteristics. The design of the multimodal dynamic fusion and enhanced multiscale integration modules supports cross-modal adaptive weight allocation and efficient cross-scale feature aggregation, leveraging complementary information across dimensions. It is demonstrated by experimental results that all baseline models are significantly outperformed by MSMNet with 73.80% mIoU, 84.47% $F1$-score, and 99.37% overall accuracy. Compared to the second-best approach, its metrics improved by 2.25%, 1.59%, and 0.06%, respectively. MSMNet notably demonstrated exceptional performance in classifying key mangrove species, such as Rhizophora stylosa, achieving a 4.06% accuracy improvement and significantly reducing misclassification rates at species level. These findings confirm that multimodal feature fusion and multiscale information integration are crucial for improving the accuracy of mangrove species classification. MSMNet provides an efficient solution for precise intertidal mangrove mapping.
潮间带红树林群落由于潮汐动态、缠绕的冠层和物种之间的光谱混淆,对遥感分类提出了重大挑战。这些因素往往导致分类错误和物种边界模糊,这主要是由于光谱内的变化和光谱间的相似性,也限制了垂直结构特征的捕捉。为了解决这些问题,我们提出了MSMNet,一种基于Sentinel-2多光谱图像的细粒度红树林物种分类的专用多模态深度学习方法。该模型使用ResNet50作为其主干架构,并集成了Mamba状态空间模块来建模远程空间相关性。MSMNet结合小波变换技术增强纹理表征能力,采用三种路径提取红树林遥感形态纹理、光谱特征和植被生理特征。多模态动态融合和增强型多尺度集成模块的设计支持跨模态自适应权重分配和高效的跨尺度特征聚合,利用了跨维度的互补信息。实验结果表明,MSMNet以73.80%的mIoU、84.47%的$F1$-score和99.37%的总体准确率显著优于所有基线模型。与次优方法相比,其指标分别提高了2.25%、1.59%和0.06%。MSMNet在主要红树林物种(如茎尖根hora stylosa)的分类中表现出优异的表现,准确率提高了4.06%,在物种水平上显著降低了误分类率。这些结果证实了多模态特征融合和多尺度信息集成对于提高红树林物种分类的准确性至关重要。MSMNet为潮间带红树林的精确测绘提供了有效的解决方案。
{"title":"A Multimodal Remote Sensing Method for Mangrove Species Classification Based on Sentinel-2 Imagery","authors":"Haiwei Yuan;Shikuan Wang;Jianzhou Gong","doi":"10.1109/JSTARS.2025.3647732","DOIUrl":"https://doi.org/10.1109/JSTARS.2025.3647732","url":null,"abstract":"Intertidal mangrove communities present significant challenges for remote sensing classification due to tidal dynamics, intertwining canopies, and spectral confusion between species. These factors often lead to classification errors and indistinct species boundaries, primarily due to intraspectral variation and interspectral similarity, and also limit the capture of vertical structural features. To address these issues, we propose MSMNet, a specialized multimodal deep learning method for fine-grained mangrove species classification based on Sentinel-2 multispectral imagery. The model uses ResNet50 as its backbone architecture and integrates the Mamba state-space module to model long-range spatial correlations. MSMNet incorporates wavelet transform technology to enhance its ability to represent textures and uses three pathways to extract three key mangrove remote sensing modalities: morphological texture, spectral features, and vegetation physiological characteristics. The design of the multimodal dynamic fusion and enhanced multiscale integration modules supports cross-modal adaptive weight allocation and efficient cross-scale feature aggregation, leveraging complementary information across dimensions. It is demonstrated by experimental results that all baseline models are significantly outperformed by MSMNet with 73.80% mIoU, 84.47% <inline-formula><tex-math>$F1$</tex-math></inline-formula>-score, and 99.37% overall accuracy. Compared to the second-best approach, its metrics improved by 2.25%, 1.59%, and 0.06%, respectively. MSMNet notably demonstrated exceptional performance in classifying key mangrove species, such as Rhizophora stylosa, achieving a 4.06% accuracy improvement and significantly reducing misclassification rates at species level. These findings confirm that multimodal feature fusion and multiscale information integration are crucial for improving the accuracy of mangrove species classification. MSMNet provides an efficient solution for precise intertidal mangrove mapping.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"4766-4778"},"PeriodicalIF":5.3,"publicationDate":"2025-12-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11314702","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146082032","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1