首页 > 最新文献

IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society最新文献

英文 中文
K-Means Clustering for Improved Data-Driven Satellite Aerosol Retrieval 改进数据驱动卫星气溶胶反演的k均值聚类
Shangshang Zhang;Yulong Fan;Lin Sun
Accurate retrieval of the spatiotemporal distribution of atmospheric aerosols is essential for studying aerosolradiationcloud interactions, air-quality forecasting, and climate-change assessment. Although data-driven methods have significantly advanced aerosol retrieval, the existing models often neglect the influence of aerosol type on retrieval accuracy. To address this gap, this study presents an improved data-driven aerosol retrieval framework that explicitly incorporates aerosol type information into model training. Aerosol classification is performed using the $K$ -means unsupervised clustering algorithm to optimize training samples, thereby enhancing model adaptability and retrieval accuracy. The refined samples are then used to train an extremely randomized trees (ERTs) model, achieving an optimal balance between accuracy and computational efficiency. Validation results demonstrate strong performance, with a correlation coefficient of 0.93, a root mean square error (RMSE) of 0.072, and over 89% of results falling within the expected error range [(EE: ± (0.05+20% $times $ in situ observations)], better than that of the traditional model. The findings demonstrate that integrating aerosol-type information into data-driven retrievals substantially improves accuracy and applicability for aerosol remote sensing. Future research should focus on refining aerosol classification techniques and integrating multisource remote sensing data to enhance model robustness and global applicability further.
准确获取大气气溶胶的时空分布对于研究气溶胶-云相互作用、空气质量预报和气候变化评估至关重要。虽然数据驱动的方法显著提高了气溶胶的检索精度,但现有的模型往往忽略了气溶胶类型对检索精度的影响。为了解决这一差距,本研究提出了一个改进的数据驱动的气溶胶检索框架,该框架明确地将气溶胶类型信息纳入模型训练中。使用$K$ means无监督聚类算法对训练样本进行优化,从而提高模型的适应性和检索精度。然后使用精炼的样本来训练极度随机树(ERTs)模型,在准确性和计算效率之间实现最佳平衡。验证结果表现出较强的性能,相关系数为0.93,均方根误差(RMSE)为0.072,超过89%的结果落在预期误差范围内[(EE:±(0.05+20% $times $原位观测值)],优于传统模型。研究结果表明,将气溶胶类型信息整合到数据驱动的检索中,大大提高了气溶胶遥感的准确性和适用性。未来的研究重点应放在完善气溶胶分类技术和整合多源遥感数据上,以进一步提高模型的鲁棒性和全球适用性。
{"title":"K-Means Clustering for Improved Data-Driven Satellite Aerosol Retrieval","authors":"Shangshang Zhang;Yulong Fan;Lin Sun","doi":"10.1109/LGRS.2025.3631867","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3631867","url":null,"abstract":"Accurate retrieval of the spatiotemporal distribution of atmospheric aerosols is essential for studying aerosolradiationcloud interactions, air-quality forecasting, and climate-change assessment. Although data-driven methods have significantly advanced aerosol retrieval, the existing models often neglect the influence of aerosol type on retrieval accuracy. To address this gap, this study presents an improved data-driven aerosol retrieval framework that explicitly incorporates aerosol type information into model training. Aerosol classification is performed using the <inline-formula> <tex-math>$K$ </tex-math></inline-formula>-means unsupervised clustering algorithm to optimize training samples, thereby enhancing model adaptability and retrieval accuracy. The refined samples are then used to train an extremely randomized trees (ERTs) model, achieving an optimal balance between accuracy and computational efficiency. Validation results demonstrate strong performance, with a correlation coefficient of 0.93, a root mean square error (RMSE) of 0.072, and over 89% of results falling within the expected error range [(EE: ± (0.05+20% <inline-formula> <tex-math>$times $ </tex-math></inline-formula> in situ observations)], better than that of the traditional model. The findings demonstrate that integrating aerosol-type information into data-driven retrievals substantially improves accuracy and applicability for aerosol remote sensing. Future research should focus on refining aerosol classification techniques and integrating multisource remote sensing data to enhance model robustness and global applicability further.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145560637","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Method for Reconstructing Surface Spectral Reflectance With Missing RadCalNet Data 一种利用RadCalNet数据缺失重建地表光谱反射率的方法
Shutian Zhu;Qiyue Liu;Chuanzhao Tian;Hanlie Xu;Jie Han;Wenhao Zhang;Na Xu
Data gaps exist in the measured spectral reflectance and atmospheric data from the radiometric calibration network (RadCalNet) due to instrument malfunctions or weather-related interferences, which severely impedes the application of the data. Therefore, developing a method to fill these missing RadCalNet data is a pressing issue. This study focuses on four RadCalNet sites with distinct surface types and proposes a high-precision bottom-of-atmosphere (BOA) spectral reflectance model. With on-site atmospheric data from RadCalNet, the predicted results achieve a root mean square error (RMSE) of no more than 1.26%. In scenarios where in situ atmospheric conditions are completely missing, the ERA5 dataset is used as a substitute and validated with Landsat 8 surface reflectance products; the absolute errors for all sites did not exceed 4.58%, validating the proposed method’s effectiveness. Additionally, the importance of input parameters and the impact of their uncertainties on prediction accuracy are discussed.
RadCalNet辐射定标网(radiometric calibration network, RadCalNet)的实测光谱反射率和大气数据由于仪器故障或天气干扰存在数据空白,严重阻碍了数据的应用。因此,开发一种方法来填补这些缺失的RadCalNet数据是一个紧迫的问题。本研究以四个不同地表类型的RadCalNet站点为研究对象,提出了一个高精度的大气底部(BOA)光谱反射率模型。利用RadCalNet的现场大气数据,预测结果的均方根误差(RMSE)不超过1.26%。在完全没有现场大气条件的情况下,使用ERA5数据集作为替代,并使用Landsat 8表面反射率产品进行验证;所有位点的绝对误差均不超过4.58%,验证了方法的有效性。此外,还讨论了输入参数的重要性及其不确定性对预测精度的影响。
{"title":"A Method for Reconstructing Surface Spectral Reflectance With Missing RadCalNet Data","authors":"Shutian Zhu;Qiyue Liu;Chuanzhao Tian;Hanlie Xu;Jie Han;Wenhao Zhang;Na Xu","doi":"10.1109/LGRS.2025.3631876","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3631876","url":null,"abstract":"Data gaps exist in the measured spectral reflectance and atmospheric data from the radiometric calibration network (RadCalNet) due to instrument malfunctions or weather-related interferences, which severely impedes the application of the data. Therefore, developing a method to fill these missing RadCalNet data is a pressing issue. This study focuses on four RadCalNet sites with distinct surface types and proposes a high-precision bottom-of-atmosphere (BOA) spectral reflectance model. With on-site atmospheric data from RadCalNet, the predicted results achieve a root mean square error (RMSE) of no more than 1.26%. In scenarios where in situ atmospheric conditions are completely missing, the ERA5 dataset is used as a substitute and validated with Landsat 8 surface reflectance products; the absolute errors for all sites did not exceed 4.58%, validating the proposed method’s effectiveness. Additionally, the importance of input parameters and the impact of their uncertainties on prediction accuracy are discussed.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145560640","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LD-YOLO: A Lightweight Dynamic Convolution-Based YOLOv8n Framework for Robust Ship Detection in SAR Imagery LD-YOLO:一种轻量级的基于动态卷积的YOLOv8n框架,用于SAR图像的鲁棒舰船检测
Jiqiang Niu;Mengyang Li;Hao Lin;Yichen Liu;Zijian Liu;Hongrui Li;Shaomian Niu
Deep learning has emerged as the predominant approach for ship detection in synthetic aperture radar (SAR) imagery. Nevertheless, persistent challenges such as densely clustered vessels, intricate background complexity, and multiscale target variations often lead to incomplete feature extraction, resulting in false alarms and missed detections. To address these limitations, this study presents LD-YOLO, an enhanced model based on YOLOv8n, which incorporates three critical innovations. Dynamic convolution layers are strategically embedded within key backbone stages to adaptively adjust kernel parameters, enhancing multiscale feature discriminability while maintaining computational efficiency. The proposed C2f-LSK module combines decomposed large-kernel convolution with attention mechanisms, enabling dynamic optimization of receptive field contributions across different detection stages and effective modeling of global contextual information. Considering the characteristics of small vessels in SAR imagery and the impact of downsampling rates on image quality, a dedicated $160times 160$ detection head is further integrated to preserve fine-grained details of small targets, complemented by bidirectional feature fusion to strengthen semantic context propagation. Extensive experiments validate the model’s superiority, achieving 98.2% of AP50 and 73.1% of AP50-95 on the SSDD benchmark, with consistent performance improvements demonstrated on HRSID (94.6% AP50) datasets. These advancements position LD-YOLO as a robust solution for maritime surveillance applications requiring high-precision SAR image analysis under complex operational conditions.
深度学习已成为合成孔径雷达(SAR)图像中船舶检测的主要方法。然而,持续存在的挑战,如密集聚集的血管、复杂的背景复杂性和多尺度目标变化,往往导致不完整的特征提取,从而导致误报和漏检。为了解决这些限制,本研究提出了基于YOLOv8n的增强模型LD-YOLO,其中包含三个关键创新。动态卷积层战略性地嵌入关键骨干阶段,自适应调整核参数,在保持计算效率的同时增强了多尺度特征的可分辨性。提出的C2f-LSK模块将分解的大核卷积与注意机制相结合,实现了不同检测阶段感受野贡献的动态优化,并有效地对全局上下文信息进行建模。考虑到SAR图像中小船只的特点以及下采样率对图像质量的影响,进一步集成了专用的160 × 160检测头,以保留小目标的细粒度细节,并辅以双向特征融合,以加强语义上下文传播。大量的实验验证了该模型的优越性,在SSDD基准上实现了98.2%的AP50和73.1%的AP50-95,在HRSID数据集上表现出了一致的性能改进(94.6% AP50)。这些进步使LD-YOLO成为需要在复杂操作条件下进行高精度SAR图像分析的海上监视应用的强大解决方案。
{"title":"LD-YOLO: A Lightweight Dynamic Convolution-Based YOLOv8n Framework for Robust Ship Detection in SAR Imagery","authors":"Jiqiang Niu;Mengyang Li;Hao Lin;Yichen Liu;Zijian Liu;Hongrui Li;Shaomian Niu","doi":"10.1109/LGRS.2025.3630098","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3630098","url":null,"abstract":"Deep learning has emerged as the predominant approach for ship detection in synthetic aperture radar (SAR) imagery. Nevertheless, persistent challenges such as densely clustered vessels, intricate background complexity, and multiscale target variations often lead to incomplete feature extraction, resulting in false alarms and missed detections. To address these limitations, this study presents LD-YOLO, an enhanced model based on YOLOv8n, which incorporates three critical innovations. Dynamic convolution layers are strategically embedded within key backbone stages to adaptively adjust kernel parameters, enhancing multiscale feature discriminability while maintaining computational efficiency. The proposed C2f-LSK module combines decomposed large-kernel convolution with attention mechanisms, enabling dynamic optimization of receptive field contributions across different detection stages and effective modeling of global contextual information. Considering the characteristics of small vessels in SAR imagery and the impact of downsampling rates on image quality, a dedicated <inline-formula> <tex-math>$160times 160$ </tex-math></inline-formula> detection head is further integrated to preserve fine-grained details of small targets, complemented by bidirectional feature fusion to strengthen semantic context propagation. Extensive experiments validate the model’s superiority, achieving 98.2% of AP50 and 73.1% of AP50-95 on the SSDD benchmark, with consistent performance improvements demonstrated on HRSID (94.6% AP50) datasets. These advancements position LD-YOLO as a robust solution for maritime surveillance applications requiring high-precision SAR image analysis under complex operational conditions.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145612149","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Spatial–Temporal and Wavenumber--Frequency Inversion Algorithms for Ocean Surface Current Using Coherent S-Band Radar 基于相干s波段雷达的海流时空和波数频率反演算法
Xinyu Fu;Chen Zhao;Zezong Chen;Sitao Wu;Fan Ding;Rui Liu;Guoxing Zheng
Coherent S-band radar has recently emerged as a promising technique for ocean surface wave and current detection. It can measure ocean surface current by estimating Doppler frequency shifts from sea surface signals. However, the conventional time averaging (TA) method neglects spatial dimension information and is unavailable under low wind speed conditions. Two algorithms for ocean current inversion are proposed in this letter: the spatial–temporal averaging (STA) method and the wavenumber--frequency (WF) method. In the STA method, the TA method is extended to the spatial–temporal domain. This approach fully exploits the spatial continuity of radar signals. In the WF method, a 2-D Fast Fourier Transform (2-D FFT) is applied to transform the spatial–temporal radial velocities into the WF domain. After employing dual filtering to eliminate nonlinear components, the radial current velocity is estimated through a modified dispersion relation fitting. The two methods are based on different physical mechanisms: the STA method measurements include wind drift components, while the WF method remains unaffected by wind drift. Therefore, wind drift can be effectively estimated by calculating the difference between the two methods’ measurements. Validation using observational data collected at Beishuang Island during Typhoon Catfish shows that the estimated wind drifts achieve a correlation coefficient (COR) of 0.90 with the “empirical model predictions.” This confirms the effectiveness of the proposed algorithms.
相干s波段雷达是近年来发展起来的一种很有前途的海面波流探测技术。它可以通过估计海面信号的多普勒频移来测量海面电流。然而,传统的时间平均(TA)方法忽略了空间维度信息,在低风速条件下无法使用。本文提出了两种海流反演算法:时空平均法(STA)和波数频率法(WF)。在STA方法中,将TA方法扩展到时空域。这种方法充分利用了雷达信号的空间连续性。在WF方法中,采用二维快速傅里叶变换(2-D FFT)将时空径向速度变换到WF域中。采用双重滤波去除非线性分量后,通过修正色散关系拟合估计径向电流速度。两种方法基于不同的物理机制:STA方法测量包含风漂移分量,而WF方法不受风漂移的影响。因此,通过计算两种方法测量值的差值,可以有效地估计风漂。台风鲇鱼期间北双岛观测资料的验证表明,风量估算值与“经验模型预测值”的相关系数(COR)为0.90。这证实了所提算法的有效性。
{"title":"Spatial–Temporal and Wavenumber--Frequency Inversion Algorithms for Ocean Surface Current Using Coherent S-Band Radar","authors":"Xinyu Fu;Chen Zhao;Zezong Chen;Sitao Wu;Fan Ding;Rui Liu;Guoxing Zheng","doi":"10.1109/LGRS.2025.3629684","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3629684","url":null,"abstract":"Coherent S-band radar has recently emerged as a promising technique for ocean surface wave and current detection. It can measure ocean surface current by estimating Doppler frequency shifts from sea surface signals. However, the conventional time averaging (TA) method neglects spatial dimension information and is unavailable under low wind speed conditions. Two algorithms for ocean current inversion are proposed in this letter: the spatial–temporal averaging (STA) method and the wavenumber--frequency (WF) method. In the STA method, the TA method is extended to the spatial–temporal domain. This approach fully exploits the spatial continuity of radar signals. In the WF method, a 2-D Fast Fourier Transform (2-D FFT) is applied to transform the spatial–temporal radial velocities into the WF domain. After employing dual filtering to eliminate nonlinear components, the radial current velocity is estimated through a modified dispersion relation fitting. The two methods are based on different physical mechanisms: the STA method measurements include wind drift components, while the WF method remains unaffected by wind drift. Therefore, wind drift can be effectively estimated by calculating the difference between the two methods’ measurements. Validation using observational data collected at Beishuang Island during Typhoon Catfish shows that the estimated wind drifts achieve a correlation coefficient (COR) of 0.90 with the “empirical model predictions.” This confirms the effectiveness of the proposed algorithms.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145560638","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SDWPNet: A Downsampling-Driven Network for SAR Ship Detection With Refined Features and Optimized Loss SDWPNet:一种特征细化、损失优化的下采样驱动SAR舰船检测网络
Xingyu Hu;Hongyu Chen;Yugang Chang;Xue Yang;Weiming Zeng
Ship detection in remote sensing images plays an important role in various maritime activities. However, the existing deep learning methods face challenges, such as changes in ship target size, complex backgrounds, and noise interference in remote sensing images, which can lead to low detection accuracy and incomplete target detection. To address these issues, we proposed a synthetic aperture radar (SAR) image target detection framework called SDWPNet, aimed at improving target detection performance in complex scenes. First, we proposed SDWavetpool (SDW), which optimizes feature downsampling through multiscale wavelet features, effectively reducing the dimensionality of the feature map while preserving the detailed information of small targets. It can more accurately identify medium and large targets in complex backgrounds, fully utilizing multilevel features. Then, the network structure was optimized using a feature extraction module that combines the PPA mechanism, making it more focused on the details of small targets. In addition, we further improved the detection accuracy by improving the loss function (ICMPIoU). The experiments on the SAR ship detection dataset (SSDD) and high-resolution SAR image dataset (HRSID) show that this framework performs well in both accuracy and response speed of target detection, achieving 74.5% and 67.6% in $mathbf {mAP_{.50:.95}}$ , using only parameter 2.97 M.
遥感图像中的船舶检测在各种海事活动中发挥着重要作用。然而,现有的深度学习方法面临着船舶目标尺寸变化、背景复杂、遥感图像中存在噪声干扰等问题,导致检测精度低、目标检测不完整。为了解决这些问题,我们提出了一种合成孔径雷达(SAR)图像目标检测框架SDWPNet,旨在提高复杂场景下的目标检测性能。首先,我们提出SDWavetpool (SDW),通过多尺度小波特征优化特征降采样,在保留小目标详细信息的同时有效降低特征映射的维数;它可以更准确地识别复杂背景下的大中型目标,充分利用多层次特征。然后,利用结合PPA机制的特征提取模块对网络结构进行优化,使其更加关注小目标的细节。此外,我们通过改进损失函数(ICMPIoU)进一步提高了检测精度。在SAR船舶检测数据集(SSDD)和高分辨率SAR图像数据集(HRSID)上的实验表明,该框架在目标检测精度和响应速度上都有良好的表现,在$mathbf {mAP_{.50:上分别达到了74.5%和67.6%。$,只使用参数2.97 M。
{"title":"SDWPNet: A Downsampling-Driven Network for SAR Ship Detection With Refined Features and Optimized Loss","authors":"Xingyu Hu;Hongyu Chen;Yugang Chang;Xue Yang;Weiming Zeng","doi":"10.1109/LGRS.2025.3629377","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3629377","url":null,"abstract":"Ship detection in remote sensing images plays an important role in various maritime activities. However, the existing deep learning methods face challenges, such as changes in ship target size, complex backgrounds, and noise interference in remote sensing images, which can lead to low detection accuracy and incomplete target detection. To address these issues, we proposed a synthetic aperture radar (SAR) image target detection framework called SDWPNet, aimed at improving target detection performance in complex scenes. First, we proposed SDWavetpool (SDW), which optimizes feature downsampling through multiscale wavelet features, effectively reducing the dimensionality of the feature map while preserving the detailed information of small targets. It can more accurately identify medium and large targets in complex backgrounds, fully utilizing multilevel features. Then, the network structure was optimized using a feature extraction module that combines the PPA mechanism, making it more focused on the details of small targets. In addition, we further improved the detection accuracy by improving the loss function (ICMPIoU). The experiments on the SAR ship detection dataset (SSDD) and high-resolution SAR image dataset (HRSID) show that this framework performs well in both accuracy and response speed of target detection, achieving 74.5% and 67.6% in <inline-formula> <tex-math>$mathbf {mAP_{.50:.95}}$ </tex-math></inline-formula>, using only parameter 2.97 M.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145612117","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient Remote Sensing Change Detection With Change State Space Models 基于变化状态空间模型的高效遥感变化检测
Elman Ghazaei;Erchan Aptoula
ConvNets and Vision Transformers (ViTs) have been widely used for change detection (CD), though they exhibit limitations: long-range dependencies are not effectively captured by the former, while the latter are associated with high computational demands. Vision Mamba, based on State Space Models, has been proposed as an alternative, yet has been primarily utilized as a feature extraction backbone. In this work, the change state space model (CSSM) is introduced as a task-specific approach for CD, designed to focus exclusively on relevant changes between bitemporal images while filtering out irrelevant information. Through this design, the number of parameters is reduced, computational efficiency is improved, and robustness is enhanced. CSSM is evaluated on three benchmark datasets, where superior performance is achieved compared to ConvNets, ViTs, and Mamba-based models, at a significantly lower computational cost. The code will be made publicly available at https://github.com/Elman295/CSSM upon acceptance
卷积神经网络和视觉变压器(ViTs)已经广泛用于变化检测(CD),尽管它们显示出局限性:前者不能有效地捕获远程依赖关系,而后者与高计算需求相关。基于状态空间模型的视觉曼巴(Vision Mamba)已被提出作为一种替代方法,但目前主要用作特征提取主干。在这项工作中,变化状态空间模型(CSSM)作为一种特定于CD的任务方法被引入,旨在专门关注双时间图像之间的相关变化,同时过滤掉无关信息。通过这种设计,减少了参数的数量,提高了计算效率,增强了鲁棒性。CSSM在三个基准数据集上进行了评估,与基于ConvNets、ViTs和mamba的模型相比,CSSM在计算成本显著降低的情况下取得了卓越的性能。一经接受,代码将在https://github.com/Elman295/CSSM上公开发布
{"title":"Efficient Remote Sensing Change Detection With Change State Space Models","authors":"Elman Ghazaei;Erchan Aptoula","doi":"10.1109/LGRS.2025.3629303","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3629303","url":null,"abstract":"ConvNets and Vision Transformers (ViTs) have been widely used for change detection (CD), though they exhibit limitations: long-range dependencies are not effectively captured by the former, while the latter are associated with high computational demands. Vision Mamba, based on State Space Models, has been proposed as an alternative, yet has been primarily utilized as a feature extraction backbone. In this work, the change state space model (CSSM) is introduced as a task-specific approach for CD, designed to focus exclusively on relevant changes between bitemporal images while filtering out irrelevant information. Through this design, the number of parameters is reduced, computational efficiency is improved, and robustness is enhanced. CSSM is evaluated on three benchmark datasets, where superior performance is achieved compared to ConvNets, ViTs, and Mamba-based models, at a significantly lower computational cost. The code will be made publicly available at <uri>https://github.com/Elman295/CSSM</uri> upon acceptance","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145560636","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
XSNet: Lightweight Object Detection Model Using X-Shaped Architecture in Remote Sensing Images XSNet:基于x形结构的遥感图像轻量化目标检测模型
Dat Minh-Tien Nguyen;Thien Huynh-The
Remote sensing object detection faces challenges such as small object sizes, complex backgrounds, and computational constraints. To overcome these challenges, we propose XSNet, an efficient deep learning (DL) model proficiently designed to enhance feature representation and multiscale detection. Concretely, XSNet introduces three key innovations: swin-involution transformer (SIner) to improve local self-attention and spatial adaptability, positional weight bi-level routing attention (PosWeightRA) to refine spatial awareness and preserve positional encoding, and an X-shaped multiscale feature fusion strategy to optimize feature aggregation while reducing computational cost. These components collectively improve detection accuracy, particularly for small and overlapping objects. Through extensive experiments, XSNet achieves impressive mAP0.5 and mAP0.95 scores of 47.1% and 28.2% on VisDrone2019, and 92.9% and 66.0% on RSOD. It outperforms state-of-the-art models while maintaining a compact size of 7.11 million parameters and fast inference time of 35.5 ms, making it well-suited for real-time remote sensing in resource-constrained environments.
遥感目标检测面临着小目标尺寸、复杂背景和计算限制等挑战。为了克服这些挑战,我们提出了XSNet,一种高效的深度学习(DL)模型,旨在增强特征表示和多尺度检测。具体而言,XSNet引入了三个关键创新:旋转-对合变压器(SIner)提高局部自关注和空间适应性,位置权重双级路由关注(PosWeightRA)改进空间感知和保留位置编码,x形多尺度特征融合策略优化特征聚合,同时降低计算成本。这些组件共同提高了检测精度,特别是对于小的和重叠的物体。通过广泛的实验,XSNet在VisDrone2019上取得了令人印象深刻的mAP0.5和mAP0.95分数,分别为47.1%和28.2%,在RSOD上取得了92.9%和66.0%的分数。它优于最先进的模型,同时保持711万个参数的紧凑尺寸和35.5 ms的快速推断时间,使其非常适合资源受限环境下的实时遥感。
{"title":"XSNet: Lightweight Object Detection Model Using X-Shaped Architecture in Remote Sensing Images","authors":"Dat Minh-Tien Nguyen;Thien Huynh-The","doi":"10.1109/LGRS.2025.3626855","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3626855","url":null,"abstract":"Remote sensing object detection faces challenges such as small object sizes, complex backgrounds, and computational constraints. To overcome these challenges, we propose XSNet, an efficient deep learning (DL) model proficiently designed to enhance feature representation and multiscale detection. Concretely, XSNet introduces three key innovations: swin-involution transformer (SIner) to improve local self-attention and spatial adaptability, positional weight bi-level routing attention (PosWeightRA) to refine spatial awareness and preserve positional encoding, and an X-shaped multiscale feature fusion strategy to optimize feature aggregation while reducing computational cost. These components collectively improve detection accuracy, particularly for small and overlapping objects. Through extensive experiments, XSNet achieves impressive mAP0.5 and mAP0.95 scores of 47.1% and 28.2% on VisDrone2019, and 92.9% and 66.0% on RSOD. It outperforms state-of-the-art models while maintaining a compact size of 7.11 million parameters and fast inference time of 35.5 ms, making it well-suited for real-time remote sensing in resource-constrained environments.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"22 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145510076","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TopoSegNet: Enhancing Geometric Fidelity of Coastline Extraction via a Joint Segmentation and Topological Reasoning Framework TopoSegNet:通过联合分割和拓扑推理框架提高海岸线提取的几何保真度
Binge Cui;Shengyun Liu;Jing Zhang;Yan Lu
Coastline extraction from remote sensing imagery is persistently challenged by intra-class heterogeneity (e.g., diverse coastline types) and boundary ambiguity. Existing methods often exhibit suboptimal performance in complex scenes mixing artificial and natural landforms, as they tend to ignore coastline morphological priors and struggle to recover details in low-contrast regions. To address these issues, this letter introduces TopoSegNet, a novel collaborative framework centered on a dual-decoder architecture. A segmentation decoder utilizes a morphology-aware attention (MAA) module to adaptively decouple and model diverse coastline morphologies and a structure-detail synergistic enhancement (SDSE) module to reconstruct weak boundaries with high fidelity. Meanwhile, a learnable topology decoder frames topology construction as a graph reasoning task, which ensures the geometric and topological integrity of the final vector output. TopoSegNet was evaluated on the public Landsat-8 and a custom Lianyungang Gaofen-1 (GF-1) dataset. The experimental results show that the proposed method reached 98.64%, 66.80%, and 0.795% on the mIoU, BIoU, and average path length similarity (APLS) metrics, respectively, verifying its validity and superiority. Compared to the state-of-the-art methods, the TopoSegNet model demonstrates significantly higher accuracy and topological fidelity.
从遥感影像中提取海岸线一直受到类内异质性(如海岸线类型的多样性)和边界模糊性的挑战。现有的方法往往在人工和自然地形混合的复杂场景中表现不佳,因为它们往往忽略海岸线形态先验,并且难以在低对比度区域恢复细节。为了解决这些问题,本文介绍了TopoSegNet,这是一种以双解码器架构为中心的新型协作框架。分割解码器利用形态感知注意(MAA)模块自适应解耦和建模不同的海岸线形态,利用结构-细节协同增强(SDSE)模块重建高保真的弱边界。同时,一个可学习的拓扑解码器将拓扑构造作为一个图推理任务,保证了最终矢量输出的几何和拓扑完整性。TopoSegNet在公共Landsat-8和自定义连云港高分一号(GF-1)数据集上进行了评估。实验结果表明,该方法在mIoU、BIoU和平均路径长度相似度(apl)指标上分别达到98.64%、66.80%和0.795%,验证了该方法的有效性和优越性。与最先进的方法相比,TopoSegNet模型显示出更高的精度和拓扑保真度。
{"title":"TopoSegNet: Enhancing Geometric Fidelity of Coastline Extraction via a Joint Segmentation and Topological Reasoning Framework","authors":"Binge Cui;Shengyun Liu;Jing Zhang;Yan Lu","doi":"10.1109/LGRS.2025.3626786","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3626786","url":null,"abstract":"Coastline extraction from remote sensing imagery is persistently challenged by intra-class heterogeneity (e.g., diverse coastline types) and boundary ambiguity. Existing methods often exhibit suboptimal performance in complex scenes mixing artificial and natural landforms, as they tend to ignore coastline morphological priors and struggle to recover details in low-contrast regions. To address these issues, this letter introduces TopoSegNet, a novel collaborative framework centered on a dual-decoder architecture. A segmentation decoder utilizes a morphology-aware attention (MAA) module to adaptively decouple and model diverse coastline morphologies and a structure-detail synergistic enhancement (SDSE) module to reconstruct weak boundaries with high fidelity. Meanwhile, a learnable topology decoder frames topology construction as a graph reasoning task, which ensures the geometric and topological integrity of the final vector output. TopoSegNet was evaluated on the public Landsat-8 and a custom Lianyungang Gaofen-1 (GF-1) dataset. The experimental results show that the proposed method reached 98.64%, 66.80%, and 0.795% on the mIoU, BIoU, and average path length similarity (APLS) metrics, respectively, verifying its validity and superiority. Compared to the state-of-the-art methods, the TopoSegNet model demonstrates significantly higher accuracy and topological fidelity.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145612124","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Robust Joint Optimization Network for Feature Detection and Description in Optical and SAR Image Matching 光学与SAR图像匹配中特征检测与描述的鲁棒联合优化网络
Xinshan Zhang;Zhitao Fu;Menghua Li;Shaochen Zhang;Han Nie;Bo-Hui Tang
Deep learning approaches that jointly learn feature extraction have achieved remarkable progress in image matching. However, current methods often treat central and neighboring pixels uniformly and use static feature selection strategies that fail to account for environmental variations. This results in limited robustness of descriptors and keypoints, thereby affecting matching accuracy. To address these limitations, we propose a robust joint optimization network for feature detection and description in optical and SAR image matching. A center-weighted module (CWM) is designed to enhance local feature representation by emphasizing the hierarchical relationship between central and surrounding features. Furthermore, a multiscale gated aggregation (MSGA) module is introduced to suppress redundant responses and improve keypoint discriminability through a gating mechanism. To address the inconsistency of score maps across heterogeneous modalities, we design a position-constrained repeatability loss to guide the network in learning stable and consistent keypoint correspondences. Experimental results across various scenarios demonstrate that the proposed method outperforms state-of-the-art techniques in terms of both matching accuracy and the number of correct matches, highlighting its robustness and effectiveness.
联合学习特征提取的深度学习方法在图像匹配方面取得了显著进展。然而,目前的方法通常是均匀地处理中心和邻近像素,并使用静态特征选择策略,无法考虑环境变化。这导致描述符和关键点的鲁棒性有限,从而影响匹配精度。为了解决这些限制,我们提出了一个鲁棒的联合优化网络,用于光学和SAR图像匹配中的特征检测和描述。设计了中心加权模块(CWM),通过强调中心特征和周围特征之间的层次关系来增强局部特征的表示。在此基础上,引入了多尺度门控聚合(MSGA)模块,通过门控机制抑制冗余响应,提高关键点的可分辨性。为了解决跨异构模式的分数图不一致的问题,我们设计了一个位置约束的可重复性损失来指导网络学习稳定和一致的关键点对应。各种场景下的实验结果表明,该方法在匹配精度和正确匹配数量方面都优于目前最先进的技术,突出了其鲁棒性和有效性。
{"title":"A Robust Joint Optimization Network for Feature Detection and Description in Optical and SAR Image Matching","authors":"Xinshan Zhang;Zhitao Fu;Menghua Li;Shaochen Zhang;Han Nie;Bo-Hui Tang","doi":"10.1109/LGRS.2025.3626750","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3626750","url":null,"abstract":"Deep learning approaches that jointly learn feature extraction have achieved remarkable progress in image matching. However, current methods often treat central and neighboring pixels uniformly and use static feature selection strategies that fail to account for environmental variations. This results in limited robustness of descriptors and keypoints, thereby affecting matching accuracy. To address these limitations, we propose a robust joint optimization network for feature detection and description in optical and SAR image matching. A center-weighted module (CWM) is designed to enhance local feature representation by emphasizing the hierarchical relationship between central and surrounding features. Furthermore, a multiscale gated aggregation (MSGA) module is introduced to suppress redundant responses and improve keypoint discriminability through a gating mechanism. To address the inconsistency of score maps across heterogeneous modalities, we design a position-constrained repeatability loss to guide the network in learning stable and consistent keypoint correspondences. Experimental results across various scenarios demonstrate that the proposed method outperforms state-of-the-art techniques in terms of both matching accuracy and the number of correct matches, highlighting its robustness and effectiveness.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145537627","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MambaCast: An Efficient Precipitation Nowcasting Model With Dual-Branch Mamba MambaCast:一种有效的双分支Mamba降水临近预报模式
Haowen Jin;Yuankang Ye;Chang Liu;Feng Gao
Precipitation nowcasting using radar echo data is critical for issuing timely extreme weather warnings, yet the existing models struggle to balance computational efficiency with prediction accuracy when modeling complex, nonlinear echo sequences. To address these challenges, we propose MambaCast, a novel dual-branch precipitation nowcasting model built upon the Mamba framework. Specifically, MambaCast incorporates three key components: a state-space model (SSM) branch, a convolutional neural network (CNN) branch and a CastFusion module. The SSM branch captures global low-frequency evolution features in the radar echo field through a selective scanning mechanism, while the CNN branch extracts local high-frequency transient features using gated spatiotemporal attention (gSTA). The CastFusion module dynamically integrates features across different frequency scales, enabling adaptive fusion of spatiotemporal distribution. Experiments on two public radar datasets show that MambaCast consistently outperforms baseline models.
利用雷达回波数据进行降水临近预报对于及时发布极端天气预警至关重要,然而现有模型在模拟复杂的非线性回波序列时难以平衡计算效率和预测精度。为了解决这些挑战,我们提出了MambaCast,一种基于Mamba框架的新型双分支降水临近预报模型。具体来说,MambaCast包含三个关键组件:一个状态空间模型(SSM)分支,一个卷积神经网络(CNN)分支和一个CastFusion模块。SSM分支通过选择性扫描机制捕获雷达回波场的全局低频演化特征,CNN分支利用门控时空注意(gSTA)提取局部高频瞬态特征。CastFusion模块动态集成不同频率尺度的特征,实现时空分布的自适应融合。在两个公共雷达数据集上的实验表明,MambaCast始终优于基线模型。
{"title":"MambaCast: An Efficient Precipitation Nowcasting Model With Dual-Branch Mamba","authors":"Haowen Jin;Yuankang Ye;Chang Liu;Feng Gao","doi":"10.1109/LGRS.2025.3626369","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3626369","url":null,"abstract":"Precipitation nowcasting using radar echo data is critical for issuing timely extreme weather warnings, yet the existing models struggle to balance computational efficiency with prediction accuracy when modeling complex, nonlinear echo sequences. To address these challenges, we propose MambaCast, a novel dual-branch precipitation nowcasting model built upon the Mamba framework. Specifically, MambaCast incorporates three key components: a state-space model (SSM) branch, a convolutional neural network (CNN) branch and a CastFusion module. The SSM branch captures global low-frequency evolution features in the radar echo field through a selective scanning mechanism, while the CNN branch extracts local high-frequency transient features using gated spatiotemporal attention (gSTA). The CastFusion module dynamically integrates features across different frequency scales, enabling adaptive fusion of spatiotemporal distribution. Experiments on two public radar datasets show that MambaCast consistently outperforms baseline models.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145537628","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1