首页 > 最新文献

IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society最新文献

英文 中文
SA-RTDETR: A High-Precision Real-Time Detection Transformer Based on Complex Scenarios for SAR Object Detection SA-RTDETR:基于复杂场景SAR目标检测的高精度实时检测变压器
Zhaoyu Liu;Wei Chen;Lixia Yang
To address core challenges in synthetic aperture radar (SAR) image target detection, including complex background interference, weak small-target features, and multiscale target coexistence, this study proposes the synthetic aperture-optimized real-time detection transformer (SA-RTDETR) model. The framework incorporates three core modules to enhance detection efficacy. First, the bidirectional receptive field boosting module synergistically integrates local details with global contextual information and substantially improves discriminative feature extraction while preserving spatial resolution. Second, the deformable attention-based intrascale feature interaction module employs adaptive sampling of critical scattering regions to address localization difficulties of small targets in SAR imagery. Third, the attention upsampling module mitigates detail loss and aliasing artifacts inherent in traditional interpolation methods through feature compensation strategies. Experimental results on the SARDet-100K dataset demonstrate that SA-RTDETR achieves 90.1% mAP@50, 56.0% mAP@50-95, and 84.7% recall rate representing improvements of 2.7%, 2.6%, and 2.2% over the baseline model, respectively. The end-to-end architecture enables high-precision SAR image analysis and offers considerable potential for military reconnaissance and maritime surveillance applications. The SA-RTDETR model establishes a novel technical paradigm for reliable all-weather remote sensing target detection by harmonizing feature robustness, scale adaptability, and operational efficiency.
针对合成孔径雷达(SAR)图像目标检测中存在的复杂背景干扰、弱小目标特征和多尺度目标共存等核心问题,提出了合成孔径优化实时检测变压器(SA-RTDETR)模型。该框架包含三个核心模块,以提高检测效率。首先,双向感受野增强模块将局部细节与全局上下文信息协同集成,在保持空间分辨率的同时显著提高了判别特征提取。其次,基于形变注意力的尺度内特征交互模块采用关键散射区域的自适应采样,解决了SAR图像中小目标的定位难题。第三,注意力上采样模块通过特征补偿策略减轻了传统插值方法固有的细节损失和混叠现象。在SARDet-100K数据集上的实验结果表明,SA-RTDETR的召回率达到了90.1% mAP@50、56.0% mAP@50-95和84.7%,分别比基线模型提高了2.7%、2.6%和2.2%。端到端架构实现了高精度SAR图像分析,并为军事侦察和海上监视应用提供了相当大的潜力。SA-RTDETR模型通过协调特征鲁棒性、规模适应性和操作效率,为全天候遥感目标的可靠探测建立了一种新的技术范式。
{"title":"SA-RTDETR: A High-Precision Real-Time Detection Transformer Based on Complex Scenarios for SAR Object Detection","authors":"Zhaoyu Liu;Wei Chen;Lixia Yang","doi":"10.1109/LGRS.2025.3632153","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3632153","url":null,"abstract":"To address core challenges in synthetic aperture radar (SAR) image target detection, including complex background interference, weak small-target features, and multiscale target coexistence, this study proposes the synthetic aperture-optimized real-time detection transformer (SA-RTDETR) model. The framework incorporates three core modules to enhance detection efficacy. First, the bidirectional receptive field boosting module synergistically integrates local details with global contextual information and substantially improves discriminative feature extraction while preserving spatial resolution. Second, the deformable attention-based intrascale feature interaction module employs adaptive sampling of critical scattering regions to address localization difficulties of small targets in SAR imagery. Third, the attention upsampling module mitigates detail loss and aliasing artifacts inherent in traditional interpolation methods through feature compensation strategies. Experimental results on the SARDet-100K dataset demonstrate that SA-RTDETR achieves 90.1% mAP@50, 56.0% mAP@50-95, and 84.7% recall rate representing improvements of 2.7%, 2.6%, and 2.2% over the baseline model, respectively. The end-to-end architecture enables high-precision SAR image analysis and offers considerable potential for military reconnaissance and maritime surveillance applications. The SA-RTDETR model establishes a novel technical paradigm for reliable all-weather remote sensing target detection by harmonizing feature robustness, scale adaptability, and operational efficiency.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145612118","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An End-to-End Sea Clutter Suppression Method Using Wavelet Convolution-Enhanced Attentional Complex-Valued Neural Network 基于小波卷积增强注意复值神经网络的端到端海杂波抑制方法
Haoxuan Xu;Meiguo Gao
Marine radar is widely employed in ocean monitoring systems. However, sea clutter significantly impairs radar data interpretability and degrades maritime target detection performance. Effective clutter suppression methods are thus essential to enhance target characteristics for improved detection. However, environmental sea clutter often exhibits complex statistical characteristics, causing traditional model-based methods to suffer from performance degradation. To address this challenge, this letter proposes a sea clutter suppression method based on a complex-valued neural network (CVNN). First, the network incorporates a wavelet convolution (WTConv) block to expand the receptive field. Second, complex-valued convolutional blocks integrated with an attention mechanism are designed to enhance latent feature extraction. Finally, the model’s performance is rigorously validated using real-measured data. Experimental results demonstrate that the proposed model achieves superior clutter suppression performance.
船用雷达在海洋监测系统中有着广泛的应用。然而,海杂波极大地削弱了雷达数据的可解释性,降低了海上目标探测性能。因此,有效的杂波抑制方法对于提高目标特性以改进检测至关重要。然而,环境海杂波往往表现出复杂的统计特征,导致传统的基于模型的方法性能下降。为了解决这一挑战,本文提出了一种基于复值神经网络(CVNN)的海杂波抑制方法。首先,该网络采用小波卷积(WTConv)块来扩展接受域。其次,设计了结合注意机制的复值卷积块来增强潜在特征的提取。最后,利用实测数据对模型的性能进行了严格验证。实验结果表明,该模型具有较好的杂波抑制性能。
{"title":"An End-to-End Sea Clutter Suppression Method Using Wavelet Convolution-Enhanced Attentional Complex-Valued Neural Network","authors":"Haoxuan Xu;Meiguo Gao","doi":"10.1109/LGRS.2025.3631806","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3631806","url":null,"abstract":"Marine radar is widely employed in ocean monitoring systems. However, sea clutter significantly impairs radar data interpretability and degrades maritime target detection performance. Effective clutter suppression methods are thus essential to enhance target characteristics for improved detection. However, environmental sea clutter often exhibits complex statistical characteristics, causing traditional model-based methods to suffer from performance degradation. To address this challenge, this letter proposes a sea clutter suppression method based on a complex-valued neural network (CVNN). First, the network incorporates a wavelet convolution (WTConv) block to expand the receptive field. Second, complex-valued convolutional blocks integrated with an attention mechanism are designed to enhance latent feature extraction. Finally, the model’s performance is rigorously validated using real-measured data. Experimental results demonstrate that the proposed model achieves superior clutter suppression performance.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145612148","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
RSNet-Lite: A Lightweight Perception Subnetwork for Remote Sensing Object Detection RSNet-Lite:一种用于遥感目标检测的轻量级感知子网
Haotian Li;Jiaqi Ma;Wenna Guo;Xiaoxia Li;Xiaohui Qin;Zhenhua Ma
With the rapid development of applications such as unmanned aerial vehicle (UAV)-based remote sensing, smart cities, and intelligent transportation, small-object detection has become increasingly important in the field of object recognition. However, existing methods often struggle to balance detection accuracy and inference efficiency under large-scale variations, dense small-object distributions, and complex background interference. To address these challenges, this letter proposes a lightweight perception subnetwork, RSNet-Lite. The network integrates a multiscale attention mechanism to enhance small-object perception, dynamic convolution, and long-range spatial modeling units to improve feature representation, and lightweight convolution with efficient sampling strategies to significantly reduce computational complexity. As a result, RSNet-Lite achieves real-time inference while maintaining high detection accuracy, striking a balance between speed and performance. Finally, the proposed method is validated on the Aerial Image–Tiny Object Detection (AI-TOD) and Vision Meets Drone (VisDrone) datasets, demonstrating its effectiveness and strong potential for small-object detection tasks.
随着无人机遥感、智慧城市、智能交通等应用的快速发展,小目标检测在目标识别领域的地位日益重要。然而,在大规模变化、密集小目标分布和复杂背景干扰下,现有方法往往难以平衡检测精度和推理效率。为了解决这些挑战,这封信提出了一个轻量级感知子网,RSNet-Lite。该网络集成了多尺度注意机制来增强小目标感知,动态卷积和远程空间建模单元来改善特征表示,轻量级卷积和高效采样策略来显著降低计算复杂度。因此,RSNet-Lite在保持高检测精度的同时实现了实时推理,在速度和性能之间取得了平衡。最后,在航空图像微小目标检测(AI-TOD)和视觉与无人机(VisDrone)数据集上对该方法进行了验证,证明了该方法在小目标检测任务中的有效性和强大潜力。
{"title":"RSNet-Lite: A Lightweight Perception Subnetwork for Remote Sensing Object Detection","authors":"Haotian Li;Jiaqi Ma;Wenna Guo;Xiaoxia Li;Xiaohui Qin;Zhenhua Ma","doi":"10.1109/LGRS.2025.3631871","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3631871","url":null,"abstract":"With the rapid development of applications such as unmanned aerial vehicle (UAV)-based remote sensing, smart cities, and intelligent transportation, small-object detection has become increasingly important in the field of object recognition. However, existing methods often struggle to balance detection accuracy and inference efficiency under large-scale variations, dense small-object distributions, and complex background interference. To address these challenges, this letter proposes a lightweight perception subnetwork, RSNet-Lite. The network integrates a multiscale attention mechanism to enhance small-object perception, dynamic convolution, and long-range spatial modeling units to improve feature representation, and lightweight convolution with efficient sampling strategies to significantly reduce computational complexity. As a result, RSNet-Lite achieves real-time inference while maintaining high detection accuracy, striking a balance between speed and performance. Finally, the proposed method is validated on the Aerial Image–Tiny Object Detection (AI-TOD) and Vision Meets Drone (VisDrone) datasets, demonstrating its effectiveness and strong potential for small-object detection tasks.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145612151","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
K-Means Clustering for Improved Data-Driven Satellite Aerosol Retrieval 改进数据驱动卫星气溶胶反演的k均值聚类
Shangshang Zhang;Yulong Fan;Lin Sun
Accurate retrieval of the spatiotemporal distribution of atmospheric aerosols is essential for studying aerosolradiationcloud interactions, air-quality forecasting, and climate-change assessment. Although data-driven methods have significantly advanced aerosol retrieval, the existing models often neglect the influence of aerosol type on retrieval accuracy. To address this gap, this study presents an improved data-driven aerosol retrieval framework that explicitly incorporates aerosol type information into model training. Aerosol classification is performed using the $K$ -means unsupervised clustering algorithm to optimize training samples, thereby enhancing model adaptability and retrieval accuracy. The refined samples are then used to train an extremely randomized trees (ERTs) model, achieving an optimal balance between accuracy and computational efficiency. Validation results demonstrate strong performance, with a correlation coefficient of 0.93, a root mean square error (RMSE) of 0.072, and over 89% of results falling within the expected error range [(EE: ± (0.05+20% $times $ in situ observations)], better than that of the traditional model. The findings demonstrate that integrating aerosol-type information into data-driven retrievals substantially improves accuracy and applicability for aerosol remote sensing. Future research should focus on refining aerosol classification techniques and integrating multisource remote sensing data to enhance model robustness and global applicability further.
准确获取大气气溶胶的时空分布对于研究气溶胶-云相互作用、空气质量预报和气候变化评估至关重要。虽然数据驱动的方法显著提高了气溶胶的检索精度,但现有的模型往往忽略了气溶胶类型对检索精度的影响。为了解决这一差距,本研究提出了一个改进的数据驱动的气溶胶检索框架,该框架明确地将气溶胶类型信息纳入模型训练中。使用$K$ means无监督聚类算法对训练样本进行优化,从而提高模型的适应性和检索精度。然后使用精炼的样本来训练极度随机树(ERTs)模型,在准确性和计算效率之间实现最佳平衡。验证结果表现出较强的性能,相关系数为0.93,均方根误差(RMSE)为0.072,超过89%的结果落在预期误差范围内[(EE:±(0.05+20% $times $原位观测值)],优于传统模型。研究结果表明,将气溶胶类型信息整合到数据驱动的检索中,大大提高了气溶胶遥感的准确性和适用性。未来的研究重点应放在完善气溶胶分类技术和整合多源遥感数据上,以进一步提高模型的鲁棒性和全球适用性。
{"title":"K-Means Clustering for Improved Data-Driven Satellite Aerosol Retrieval","authors":"Shangshang Zhang;Yulong Fan;Lin Sun","doi":"10.1109/LGRS.2025.3631867","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3631867","url":null,"abstract":"Accurate retrieval of the spatiotemporal distribution of atmospheric aerosols is essential for studying aerosolradiationcloud interactions, air-quality forecasting, and climate-change assessment. Although data-driven methods have significantly advanced aerosol retrieval, the existing models often neglect the influence of aerosol type on retrieval accuracy. To address this gap, this study presents an improved data-driven aerosol retrieval framework that explicitly incorporates aerosol type information into model training. Aerosol classification is performed using the <inline-formula> <tex-math>$K$ </tex-math></inline-formula>-means unsupervised clustering algorithm to optimize training samples, thereby enhancing model adaptability and retrieval accuracy. The refined samples are then used to train an extremely randomized trees (ERTs) model, achieving an optimal balance between accuracy and computational efficiency. Validation results demonstrate strong performance, with a correlation coefficient of 0.93, a root mean square error (RMSE) of 0.072, and over 89% of results falling within the expected error range [(EE: ± (0.05+20% <inline-formula> <tex-math>$times $ </tex-math></inline-formula> in situ observations)], better than that of the traditional model. The findings demonstrate that integrating aerosol-type information into data-driven retrievals substantially improves accuracy and applicability for aerosol remote sensing. Future research should focus on refining aerosol classification techniques and integrating multisource remote sensing data to enhance model robustness and global applicability further.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145560637","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Method for Reconstructing Surface Spectral Reflectance With Missing RadCalNet Data 一种利用RadCalNet数据缺失重建地表光谱反射率的方法
Shutian Zhu;Qiyue Liu;Chuanzhao Tian;Hanlie Xu;Jie Han;Wenhao Zhang;Na Xu
Data gaps exist in the measured spectral reflectance and atmospheric data from the radiometric calibration network (RadCalNet) due to instrument malfunctions or weather-related interferences, which severely impedes the application of the data. Therefore, developing a method to fill these missing RadCalNet data is a pressing issue. This study focuses on four RadCalNet sites with distinct surface types and proposes a high-precision bottom-of-atmosphere (BOA) spectral reflectance model. With on-site atmospheric data from RadCalNet, the predicted results achieve a root mean square error (RMSE) of no more than 1.26%. In scenarios where in situ atmospheric conditions are completely missing, the ERA5 dataset is used as a substitute and validated with Landsat 8 surface reflectance products; the absolute errors for all sites did not exceed 4.58%, validating the proposed method’s effectiveness. Additionally, the importance of input parameters and the impact of their uncertainties on prediction accuracy are discussed.
RadCalNet辐射定标网(radiometric calibration network, RadCalNet)的实测光谱反射率和大气数据由于仪器故障或天气干扰存在数据空白,严重阻碍了数据的应用。因此,开发一种方法来填补这些缺失的RadCalNet数据是一个紧迫的问题。本研究以四个不同地表类型的RadCalNet站点为研究对象,提出了一个高精度的大气底部(BOA)光谱反射率模型。利用RadCalNet的现场大气数据,预测结果的均方根误差(RMSE)不超过1.26%。在完全没有现场大气条件的情况下,使用ERA5数据集作为替代,并使用Landsat 8表面反射率产品进行验证;所有位点的绝对误差均不超过4.58%,验证了方法的有效性。此外,还讨论了输入参数的重要性及其不确定性对预测精度的影响。
{"title":"A Method for Reconstructing Surface Spectral Reflectance With Missing RadCalNet Data","authors":"Shutian Zhu;Qiyue Liu;Chuanzhao Tian;Hanlie Xu;Jie Han;Wenhao Zhang;Na Xu","doi":"10.1109/LGRS.2025.3631876","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3631876","url":null,"abstract":"Data gaps exist in the measured spectral reflectance and atmospheric data from the radiometric calibration network (RadCalNet) due to instrument malfunctions or weather-related interferences, which severely impedes the application of the data. Therefore, developing a method to fill these missing RadCalNet data is a pressing issue. This study focuses on four RadCalNet sites with distinct surface types and proposes a high-precision bottom-of-atmosphere (BOA) spectral reflectance model. With on-site atmospheric data from RadCalNet, the predicted results achieve a root mean square error (RMSE) of no more than 1.26%. In scenarios where in situ atmospheric conditions are completely missing, the ERA5 dataset is used as a substitute and validated with Landsat 8 surface reflectance products; the absolute errors for all sites did not exceed 4.58%, validating the proposed method’s effectiveness. Additionally, the importance of input parameters and the impact of their uncertainties on prediction accuracy are discussed.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145560640","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LD-YOLO: A Lightweight Dynamic Convolution-Based YOLOv8n Framework for Robust Ship Detection in SAR Imagery LD-YOLO:一种轻量级的基于动态卷积的YOLOv8n框架,用于SAR图像的鲁棒舰船检测
Jiqiang Niu;Mengyang Li;Hao Lin;Yichen Liu;Zijian Liu;Hongrui Li;Shaomian Niu
Deep learning has emerged as the predominant approach for ship detection in synthetic aperture radar (SAR) imagery. Nevertheless, persistent challenges such as densely clustered vessels, intricate background complexity, and multiscale target variations often lead to incomplete feature extraction, resulting in false alarms and missed detections. To address these limitations, this study presents LD-YOLO, an enhanced model based on YOLOv8n, which incorporates three critical innovations. Dynamic convolution layers are strategically embedded within key backbone stages to adaptively adjust kernel parameters, enhancing multiscale feature discriminability while maintaining computational efficiency. The proposed C2f-LSK module combines decomposed large-kernel convolution with attention mechanisms, enabling dynamic optimization of receptive field contributions across different detection stages and effective modeling of global contextual information. Considering the characteristics of small vessels in SAR imagery and the impact of downsampling rates on image quality, a dedicated $160times 160$ detection head is further integrated to preserve fine-grained details of small targets, complemented by bidirectional feature fusion to strengthen semantic context propagation. Extensive experiments validate the model’s superiority, achieving 98.2% of AP50 and 73.1% of AP50-95 on the SSDD benchmark, with consistent performance improvements demonstrated on HRSID (94.6% AP50) datasets. These advancements position LD-YOLO as a robust solution for maritime surveillance applications requiring high-precision SAR image analysis under complex operational conditions.
深度学习已成为合成孔径雷达(SAR)图像中船舶检测的主要方法。然而,持续存在的挑战,如密集聚集的血管、复杂的背景复杂性和多尺度目标变化,往往导致不完整的特征提取,从而导致误报和漏检。为了解决这些限制,本研究提出了基于YOLOv8n的增强模型LD-YOLO,其中包含三个关键创新。动态卷积层战略性地嵌入关键骨干阶段,自适应调整核参数,在保持计算效率的同时增强了多尺度特征的可分辨性。提出的C2f-LSK模块将分解的大核卷积与注意机制相结合,实现了不同检测阶段感受野贡献的动态优化,并有效地对全局上下文信息进行建模。考虑到SAR图像中小船只的特点以及下采样率对图像质量的影响,进一步集成了专用的160 × 160检测头,以保留小目标的细粒度细节,并辅以双向特征融合,以加强语义上下文传播。大量的实验验证了该模型的优越性,在SSDD基准上实现了98.2%的AP50和73.1%的AP50-95,在HRSID数据集上表现出了一致的性能改进(94.6% AP50)。这些进步使LD-YOLO成为需要在复杂操作条件下进行高精度SAR图像分析的海上监视应用的强大解决方案。
{"title":"LD-YOLO: A Lightweight Dynamic Convolution-Based YOLOv8n Framework for Robust Ship Detection in SAR Imagery","authors":"Jiqiang Niu;Mengyang Li;Hao Lin;Yichen Liu;Zijian Liu;Hongrui Li;Shaomian Niu","doi":"10.1109/LGRS.2025.3630098","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3630098","url":null,"abstract":"Deep learning has emerged as the predominant approach for ship detection in synthetic aperture radar (SAR) imagery. Nevertheless, persistent challenges such as densely clustered vessels, intricate background complexity, and multiscale target variations often lead to incomplete feature extraction, resulting in false alarms and missed detections. To address these limitations, this study presents LD-YOLO, an enhanced model based on YOLOv8n, which incorporates three critical innovations. Dynamic convolution layers are strategically embedded within key backbone stages to adaptively adjust kernel parameters, enhancing multiscale feature discriminability while maintaining computational efficiency. The proposed C2f-LSK module combines decomposed large-kernel convolution with attention mechanisms, enabling dynamic optimization of receptive field contributions across different detection stages and effective modeling of global contextual information. Considering the characteristics of small vessels in SAR imagery and the impact of downsampling rates on image quality, a dedicated <inline-formula> <tex-math>$160times 160$ </tex-math></inline-formula> detection head is further integrated to preserve fine-grained details of small targets, complemented by bidirectional feature fusion to strengthen semantic context propagation. Extensive experiments validate the model’s superiority, achieving 98.2% of AP50 and 73.1% of AP50-95 on the SSDD benchmark, with consistent performance improvements demonstrated on HRSID (94.6% AP50) datasets. These advancements position LD-YOLO as a robust solution for maritime surveillance applications requiring high-precision SAR image analysis under complex operational conditions.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-11-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145612149","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Spatial–Temporal and Wavenumber--Frequency Inversion Algorithms for Ocean Surface Current Using Coherent S-Band Radar 基于相干s波段雷达的海流时空和波数频率反演算法
Xinyu Fu;Chen Zhao;Zezong Chen;Sitao Wu;Fan Ding;Rui Liu;Guoxing Zheng
Coherent S-band radar has recently emerged as a promising technique for ocean surface wave and current detection. It can measure ocean surface current by estimating Doppler frequency shifts from sea surface signals. However, the conventional time averaging (TA) method neglects spatial dimension information and is unavailable under low wind speed conditions. Two algorithms for ocean current inversion are proposed in this letter: the spatial–temporal averaging (STA) method and the wavenumber--frequency (WF) method. In the STA method, the TA method is extended to the spatial–temporal domain. This approach fully exploits the spatial continuity of radar signals. In the WF method, a 2-D Fast Fourier Transform (2-D FFT) is applied to transform the spatial–temporal radial velocities into the WF domain. After employing dual filtering to eliminate nonlinear components, the radial current velocity is estimated through a modified dispersion relation fitting. The two methods are based on different physical mechanisms: the STA method measurements include wind drift components, while the WF method remains unaffected by wind drift. Therefore, wind drift can be effectively estimated by calculating the difference between the two methods’ measurements. Validation using observational data collected at Beishuang Island during Typhoon Catfish shows that the estimated wind drifts achieve a correlation coefficient (COR) of 0.90 with the “empirical model predictions.” This confirms the effectiveness of the proposed algorithms.
相干s波段雷达是近年来发展起来的一种很有前途的海面波流探测技术。它可以通过估计海面信号的多普勒频移来测量海面电流。然而,传统的时间平均(TA)方法忽略了空间维度信息,在低风速条件下无法使用。本文提出了两种海流反演算法:时空平均法(STA)和波数频率法(WF)。在STA方法中,将TA方法扩展到时空域。这种方法充分利用了雷达信号的空间连续性。在WF方法中,采用二维快速傅里叶变换(2-D FFT)将时空径向速度变换到WF域中。采用双重滤波去除非线性分量后,通过修正色散关系拟合估计径向电流速度。两种方法基于不同的物理机制:STA方法测量包含风漂移分量,而WF方法不受风漂移的影响。因此,通过计算两种方法测量值的差值,可以有效地估计风漂。台风鲇鱼期间北双岛观测资料的验证表明,风量估算值与“经验模型预测值”的相关系数(COR)为0.90。这证实了所提算法的有效性。
{"title":"Spatial–Temporal and Wavenumber--Frequency Inversion Algorithms for Ocean Surface Current Using Coherent S-Band Radar","authors":"Xinyu Fu;Chen Zhao;Zezong Chen;Sitao Wu;Fan Ding;Rui Liu;Guoxing Zheng","doi":"10.1109/LGRS.2025.3629684","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3629684","url":null,"abstract":"Coherent S-band radar has recently emerged as a promising technique for ocean surface wave and current detection. It can measure ocean surface current by estimating Doppler frequency shifts from sea surface signals. However, the conventional time averaging (TA) method neglects spatial dimension information and is unavailable under low wind speed conditions. Two algorithms for ocean current inversion are proposed in this letter: the spatial–temporal averaging (STA) method and the wavenumber--frequency (WF) method. In the STA method, the TA method is extended to the spatial–temporal domain. This approach fully exploits the spatial continuity of radar signals. In the WF method, a 2-D Fast Fourier Transform (2-D FFT) is applied to transform the spatial–temporal radial velocities into the WF domain. After employing dual filtering to eliminate nonlinear components, the radial current velocity is estimated through a modified dispersion relation fitting. The two methods are based on different physical mechanisms: the STA method measurements include wind drift components, while the WF method remains unaffected by wind drift. Therefore, wind drift can be effectively estimated by calculating the difference between the two methods’ measurements. Validation using observational data collected at Beishuang Island during Typhoon Catfish shows that the estimated wind drifts achieve a correlation coefficient (COR) of 0.90 with the “empirical model predictions.” This confirms the effectiveness of the proposed algorithms.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145560638","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SDWPNet: A Downsampling-Driven Network for SAR Ship Detection With Refined Features and Optimized Loss SDWPNet:一种特征细化、损失优化的下采样驱动SAR舰船检测网络
Xingyu Hu;Hongyu Chen;Yugang Chang;Xue Yang;Weiming Zeng
Ship detection in remote sensing images plays an important role in various maritime activities. However, the existing deep learning methods face challenges, such as changes in ship target size, complex backgrounds, and noise interference in remote sensing images, which can lead to low detection accuracy and incomplete target detection. To address these issues, we proposed a synthetic aperture radar (SAR) image target detection framework called SDWPNet, aimed at improving target detection performance in complex scenes. First, we proposed SDWavetpool (SDW), which optimizes feature downsampling through multiscale wavelet features, effectively reducing the dimensionality of the feature map while preserving the detailed information of small targets. It can more accurately identify medium and large targets in complex backgrounds, fully utilizing multilevel features. Then, the network structure was optimized using a feature extraction module that combines the PPA mechanism, making it more focused on the details of small targets. In addition, we further improved the detection accuracy by improving the loss function (ICMPIoU). The experiments on the SAR ship detection dataset (SSDD) and high-resolution SAR image dataset (HRSID) show that this framework performs well in both accuracy and response speed of target detection, achieving 74.5% and 67.6% in $mathbf {mAP_{.50:.95}}$ , using only parameter 2.97 M.
遥感图像中的船舶检测在各种海事活动中发挥着重要作用。然而,现有的深度学习方法面临着船舶目标尺寸变化、背景复杂、遥感图像中存在噪声干扰等问题,导致检测精度低、目标检测不完整。为了解决这些问题,我们提出了一种合成孔径雷达(SAR)图像目标检测框架SDWPNet,旨在提高复杂场景下的目标检测性能。首先,我们提出SDWavetpool (SDW),通过多尺度小波特征优化特征降采样,在保留小目标详细信息的同时有效降低特征映射的维数;它可以更准确地识别复杂背景下的大中型目标,充分利用多层次特征。然后,利用结合PPA机制的特征提取模块对网络结构进行优化,使其更加关注小目标的细节。此外,我们通过改进损失函数(ICMPIoU)进一步提高了检测精度。在SAR船舶检测数据集(SSDD)和高分辨率SAR图像数据集(HRSID)上的实验表明,该框架在目标检测精度和响应速度上都有良好的表现,在$mathbf {mAP_{.50:上分别达到了74.5%和67.6%。$,只使用参数2.97 M。
{"title":"SDWPNet: A Downsampling-Driven Network for SAR Ship Detection With Refined Features and Optimized Loss","authors":"Xingyu Hu;Hongyu Chen;Yugang Chang;Xue Yang;Weiming Zeng","doi":"10.1109/LGRS.2025.3629377","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3629377","url":null,"abstract":"Ship detection in remote sensing images plays an important role in various maritime activities. However, the existing deep learning methods face challenges, such as changes in ship target size, complex backgrounds, and noise interference in remote sensing images, which can lead to low detection accuracy and incomplete target detection. To address these issues, we proposed a synthetic aperture radar (SAR) image target detection framework called SDWPNet, aimed at improving target detection performance in complex scenes. First, we proposed SDWavetpool (SDW), which optimizes feature downsampling through multiscale wavelet features, effectively reducing the dimensionality of the feature map while preserving the detailed information of small targets. It can more accurately identify medium and large targets in complex backgrounds, fully utilizing multilevel features. Then, the network structure was optimized using a feature extraction module that combines the PPA mechanism, making it more focused on the details of small targets. In addition, we further improved the detection accuracy by improving the loss function (ICMPIoU). The experiments on the SAR ship detection dataset (SSDD) and high-resolution SAR image dataset (HRSID) show that this framework performs well in both accuracy and response speed of target detection, achieving 74.5% and 67.6% in <inline-formula> <tex-math>$mathbf {mAP_{.50:.95}}$ </tex-math></inline-formula>, using only parameter 2.97 M.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145612117","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient Remote Sensing Change Detection With Change State Space Models 基于变化状态空间模型的高效遥感变化检测
Elman Ghazaei;Erchan Aptoula
ConvNets and Vision Transformers (ViTs) have been widely used for change detection (CD), though they exhibit limitations: long-range dependencies are not effectively captured by the former, while the latter are associated with high computational demands. Vision Mamba, based on State Space Models, has been proposed as an alternative, yet has been primarily utilized as a feature extraction backbone. In this work, the change state space model (CSSM) is introduced as a task-specific approach for CD, designed to focus exclusively on relevant changes between bitemporal images while filtering out irrelevant information. Through this design, the number of parameters is reduced, computational efficiency is improved, and robustness is enhanced. CSSM is evaluated on three benchmark datasets, where superior performance is achieved compared to ConvNets, ViTs, and Mamba-based models, at a significantly lower computational cost. The code will be made publicly available at https://github.com/Elman295/CSSM upon acceptance
卷积神经网络和视觉变压器(ViTs)已经广泛用于变化检测(CD),尽管它们显示出局限性:前者不能有效地捕获远程依赖关系,而后者与高计算需求相关。基于状态空间模型的视觉曼巴(Vision Mamba)已被提出作为一种替代方法,但目前主要用作特征提取主干。在这项工作中,变化状态空间模型(CSSM)作为一种特定于CD的任务方法被引入,旨在专门关注双时间图像之间的相关变化,同时过滤掉无关信息。通过这种设计,减少了参数的数量,提高了计算效率,增强了鲁棒性。CSSM在三个基准数据集上进行了评估,与基于ConvNets、ViTs和mamba的模型相比,CSSM在计算成本显著降低的情况下取得了卓越的性能。一经接受,代码将在https://github.com/Elman295/CSSM上公开发布
{"title":"Efficient Remote Sensing Change Detection With Change State Space Models","authors":"Elman Ghazaei;Erchan Aptoula","doi":"10.1109/LGRS.2025.3629303","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3629303","url":null,"abstract":"ConvNets and Vision Transformers (ViTs) have been widely used for change detection (CD), though they exhibit limitations: long-range dependencies are not effectively captured by the former, while the latter are associated with high computational demands. Vision Mamba, based on State Space Models, has been proposed as an alternative, yet has been primarily utilized as a feature extraction backbone. In this work, the change state space model (CSSM) is introduced as a task-specific approach for CD, designed to focus exclusively on relevant changes between bitemporal images while filtering out irrelevant information. Through this design, the number of parameters is reduced, computational efficiency is improved, and robustness is enhanced. CSSM is evaluated on three benchmark datasets, where superior performance is achieved compared to ConvNets, ViTs, and Mamba-based models, at a significantly lower computational cost. The code will be made publicly available at <uri>https://github.com/Elman295/CSSM</uri> upon acceptance","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145560636","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
XSNet: Lightweight Object Detection Model Using X-Shaped Architecture in Remote Sensing Images XSNet:基于x形结构的遥感图像轻量化目标检测模型
Dat Minh-Tien Nguyen;Thien Huynh-The
Remote sensing object detection faces challenges such as small object sizes, complex backgrounds, and computational constraints. To overcome these challenges, we propose XSNet, an efficient deep learning (DL) model proficiently designed to enhance feature representation and multiscale detection. Concretely, XSNet introduces three key innovations: swin-involution transformer (SIner) to improve local self-attention and spatial adaptability, positional weight bi-level routing attention (PosWeightRA) to refine spatial awareness and preserve positional encoding, and an X-shaped multiscale feature fusion strategy to optimize feature aggregation while reducing computational cost. These components collectively improve detection accuracy, particularly for small and overlapping objects. Through extensive experiments, XSNet achieves impressive mAP0.5 and mAP0.95 scores of 47.1% and 28.2% on VisDrone2019, and 92.9% and 66.0% on RSOD. It outperforms state-of-the-art models while maintaining a compact size of 7.11 million parameters and fast inference time of 35.5 ms, making it well-suited for real-time remote sensing in resource-constrained environments.
遥感目标检测面临着小目标尺寸、复杂背景和计算限制等挑战。为了克服这些挑战,我们提出了XSNet,一种高效的深度学习(DL)模型,旨在增强特征表示和多尺度检测。具体而言,XSNet引入了三个关键创新:旋转-对合变压器(SIner)提高局部自关注和空间适应性,位置权重双级路由关注(PosWeightRA)改进空间感知和保留位置编码,x形多尺度特征融合策略优化特征聚合,同时降低计算成本。这些组件共同提高了检测精度,特别是对于小的和重叠的物体。通过广泛的实验,XSNet在VisDrone2019上取得了令人印象深刻的mAP0.5和mAP0.95分数,分别为47.1%和28.2%,在RSOD上取得了92.9%和66.0%的分数。它优于最先进的模型,同时保持711万个参数的紧凑尺寸和35.5 ms的快速推断时间,使其非常适合资源受限环境下的实时遥感。
{"title":"XSNet: Lightweight Object Detection Model Using X-Shaped Architecture in Remote Sensing Images","authors":"Dat Minh-Tien Nguyen;Thien Huynh-The","doi":"10.1109/LGRS.2025.3626855","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3626855","url":null,"abstract":"Remote sensing object detection faces challenges such as small object sizes, complex backgrounds, and computational constraints. To overcome these challenges, we propose XSNet, an efficient deep learning (DL) model proficiently designed to enhance feature representation and multiscale detection. Concretely, XSNet introduces three key innovations: swin-involution transformer (SIner) to improve local self-attention and spatial adaptability, positional weight bi-level routing attention (PosWeightRA) to refine spatial awareness and preserve positional encoding, and an X-shaped multiscale feature fusion strategy to optimize feature aggregation while reducing computational cost. These components collectively improve detection accuracy, particularly for small and overlapping objects. Through extensive experiments, XSNet achieves impressive mAP0.5 and mAP0.95 scores of 47.1% and 28.2% on VisDrone2019, and 92.9% and 66.0% on RSOD. It outperforms state-of-the-art models while maintaining a compact size of 7.11 million parameters and fast inference time of 35.5 ms, making it well-suited for real-time remote sensing in resource-constrained environments.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"22 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145510076","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1