首页 > 最新文献

IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing最新文献

英文 中文
MEETNet: Morphology-Edge Enhanced Triple-Cascaded Network for Infrared Small Target Detection 用于红外小目标检测的形态学边缘增强三级联网络
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-01-12 DOI: 10.1109/JSTARS.2026.3651900
Enyu Zhao;Yu Shi;Nianxin Qu;Yulei Wang;Hang Zhao
Infrared small target detection is focused on accurately identifying tiny targets with low signal-to-noise ratio against complex backgrounds, representing a critical challenge in the field of infrared image processing. Existing approaches frequently fail to retain small target information during global semantic extraction and struggle with preserving detailed features and achieving effective feature fusion. To address these limitations, this article proposes a morphology-edge enhanced triple-cascaded network (MEETNet) for infrared small target detection. The network employs a triple-cascaded architecture that maintains high resolution and enhances information interaction between different stages, facilitating effective multilevel feature fusion while safeguarding deep small-target characteristics. MEETNet integrates an edge-detail enhanced module (EDEM) and a detail-aware multi-scale fusion module (DMSFM). These modules introduce edge-detail enhanced features that amalgamate contrast and edge information, thereby amplifying target saliency and improving edge representation. Specifically, EDEM augments target contrast and edge structures by integrating edge-detail-enhanced features with shallow details. This integration improves the discriminability capacity of shallow features for detecting small targets. Moreover, DMSFM implements a multireceptive field mechanism to merge target details with deep semantic insights, enabling the capture of more distinctive global contextual features. Experimental evaluations conducted using two public datasets—NUAA-SIRST and NUDT-SIRST—demonstrate that the proposed MEETNet surpasses existing state-of-the-art methods for infrared small target detection in terms of detection accuracy.
红外小目标检测的重点是在复杂背景下准确识别低信噪比的微小目标,是红外图像处理领域的一个关键挑战。现有的方法在全局语义提取过程中往往不能保留小目标信息,难以保留细节特征并实现有效的特征融合。为了解决这些限制,本文提出了一种用于红外小目标检测的形态学边缘增强三级联网络(MEETNet)。该网络采用三级联架构,既保持了高分辨率,又增强了各阶段之间的信息交互,在保证深度小目标特征的同时,实现了有效的多级特征融合。MEETNet集成了边缘细节增强模块(EDEM)和细节感知多尺度融合模块(DMSFM)。这些模块引入边缘细节增强特征,合并对比度和边缘信息,从而放大目标显著性并改善边缘表示。具体来说,EDEM通过将边缘细节增强特征与浅层细节相结合来增强目标对比度和边缘结构。这种融合提高了浅层特征对小目标的识别能力。此外,DMSFM实现了一种多接受场机制,将目标细节与深度语义洞察合并在一起,从而能够捕获更多独特的全局上下文特征。使用两个公共数据集(nuaa - sirst和nudt - sirst)进行的实验评估表明,所提出的MEETNet在检测精度方面优于现有的最先进的红外小目标检测方法。
{"title":"MEETNet: Morphology-Edge Enhanced Triple-Cascaded Network for Infrared Small Target Detection","authors":"Enyu Zhao;Yu Shi;Nianxin Qu;Yulei Wang;Hang Zhao","doi":"10.1109/JSTARS.2026.3651900","DOIUrl":"https://doi.org/10.1109/JSTARS.2026.3651900","url":null,"abstract":"Infrared small target detection is focused on accurately identifying tiny targets with low signal-to-noise ratio against complex backgrounds, representing a critical challenge in the field of infrared image processing. Existing approaches frequently fail to retain small target information during global semantic extraction and struggle with preserving detailed features and achieving effective feature fusion. To address these limitations, this article proposes a morphology-edge enhanced triple-cascaded network (MEETNet) for infrared small target detection. The network employs a triple-cascaded architecture that maintains high resolution and enhances information interaction between different stages, facilitating effective multilevel feature fusion while safeguarding deep small-target characteristics. MEETNet integrates an edge-detail enhanced module (EDEM) and a detail-aware multi-scale fusion module (DMSFM). These modules introduce edge-detail enhanced features that amalgamate contrast and edge information, thereby amplifying target saliency and improving edge representation. Specifically, EDEM augments target contrast and edge structures by integrating edge-detail-enhanced features with shallow details. This integration improves the discriminability capacity of shallow features for detecting small targets. Moreover, DMSFM implements a multireceptive field mechanism to merge target details with deep semantic insights, enabling the capture of more distinctive global contextual features. Experimental evaluations conducted using two public datasets—NUAA-SIRST and NUDT-SIRST—demonstrate that the proposed MEETNet surpasses existing state-of-the-art methods for infrared small target detection in terms of detection accuracy.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"4748-4765"},"PeriodicalIF":5.3,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11340625","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146082064","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Feature-Screened and Structure-Constrained Deep Forest for Unsupervised SAR Image Change Detection 基于特征筛选和结构约束的无监督SAR图像变化检测深度森林
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-01-12 DOI: 10.1109/JSTARS.2026.3651534
Wanying Song;Ruijing Zhu;Jie Wang;Yinyin Jiang;Yan Wu
Deep forest-based models for synthetic aperture radar (SAR) image change detection are generally challenged by noise sensitivity and high feature redundancy, which significantly degrade the prediction performance. To address these issues, this article proposes a structure-constrained and feature-screened deep forest, abbreviated as SC-FS-DF, for SAR image change detection. In preclassification, a fuzzy multineighborhood information C-means clustering is proposed to generate high-quality pseudo-labels. It introduces the edge information, the nonlocal and intrasuperpixel neighborhoods into the objective function of fuzzy local information C-means, thus suppressing the speckle noise and constraining structures of targets. In the sample learning and label prediction module, a feature-screened deep forest (FS-DF) framework is constructed by combining feature importance and redundancy analysis with a dropout strategy, thus screening out the noninformative features and meanwhile retaining the informative ones for learning at each cascade layer. Finally, a novel energy function fusing the nonlocal and superpixel information is derived for refining the detection map generated by FS-DF, further preserving fine details and edge locations. Extensive comparison and ablation experiments on five real SAR datasets verify the effectiveness and robustness of the proposed SC-FS-DF, and demonstrate that the SC-FS-DF can well screen the high-dimensional features in change detection and constrain the structures of targets.
基于深度森林的合成孔径雷达(SAR)图像变化检测模型存在噪声敏感性和特征冗余度高的问题,严重影响了预测效果。为了解决这些问题,本文提出了一种结构约束和特征筛选的深森林,简称SC-FS-DF,用于SAR图像变化检测。在预分类中,提出了一种模糊多邻域信息c均值聚类方法来生成高质量的伪标签。在模糊局部信息C-means的目标函数中引入边缘信息、非局部和超像素内邻域,从而抑制散斑噪声和约束目标结构。在样本学习和标签预测模块中,将特征重要性和冗余分析与dropout策略相结合,构建了特征筛选深度森林(FS-DF)框架,从而筛选出非信息特征,同时保留每个级联层学习的信息特征。最后,导出了一种融合非局部和超像素信息的能量函数,用于细化FS-DF生成的检测图,进一步保留了精细细节和边缘位置。在5个真实SAR数据集上进行了大量对比和烧蚀实验,验证了该算法的有效性和鲁棒性,并证明了该算法在变化检测中能够很好地筛选高维特征并约束目标结构。
{"title":"Feature-Screened and Structure-Constrained Deep Forest for Unsupervised SAR Image Change Detection","authors":"Wanying Song;Ruijing Zhu;Jie Wang;Yinyin Jiang;Yan Wu","doi":"10.1109/JSTARS.2026.3651534","DOIUrl":"https://doi.org/10.1109/JSTARS.2026.3651534","url":null,"abstract":"Deep forest-based models for synthetic aperture radar (SAR) image change detection are generally challenged by noise sensitivity and high feature redundancy, which significantly degrade the prediction performance. To address these issues, this article proposes a structure-constrained and feature-screened deep forest, abbreviated as SC-FS-DF, for SAR image change detection. In preclassification, a fuzzy multineighborhood information C-means clustering is proposed to generate high-quality pseudo-labels. It introduces the edge information, the nonlocal and intrasuperpixel neighborhoods into the objective function of fuzzy local information C-means, thus suppressing the speckle noise and constraining structures of targets. In the sample learning and label prediction module, a feature-screened deep forest (FS-DF) framework is constructed by combining feature importance and redundancy analysis with a dropout strategy, thus screening out the noninformative features and meanwhile retaining the informative ones for learning at each cascade layer. Finally, a novel energy function fusing the nonlocal and superpixel information is derived for refining the detection map generated by FS-DF, further preserving fine details and edge locations. Extensive comparison and ablation experiments on five real SAR datasets verify the effectiveness and robustness of the proposed SC-FS-DF, and demonstrate that the SC-FS-DF can well screen the high-dimensional features in change detection and constrain the structures of targets.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"4056-4068"},"PeriodicalIF":5.3,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11339914","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146026482","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DTWSTSR: Dual-Tree Complex Wavelet and Swin Transformer Based Remote Sensing Images Super-Resolution Network 基于双树复小波和Swin变压器的遥感图像超分辨网络
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-01-12 DOI: 10.1109/JSTARS.2026.3651075
Yu Yao;Hengbin Wang;Xiang Gao;Ziyao Xing;Xiaodong Zhang;Yuanyuan Zhao;Shaoming Li;Zhe Liu
High-resolution remote sensing images provide crucial data support for applications such as precision agriculture and water resource management. However, super-resolution reconstructions often suffer from over-smoothed textures and structural distortions, failing to accurately recover the intricate details of ground objects. To address this issue, this article proposes a remote sensing image super-resolution network (DTWSTSR) that combines the Dual-Tree Complex Wavelet Transform and Swin Transformer, which enhances the ability of texture detail reconstruction by fusing frequency-domain and spatial-domain features. This model includes a Dual-Tree Complex Wavelet Texture Feature Sensing Module (DWTFSM) for integrating frequency and spatial features, and a Multiscale Efficient Channel Attention mechanism to enhance attention to multiscale and global details. In addition, we design a Kolmogorov–Arnold Network based on a branch attention mechanism, which improves the model’s ability to represent complex nonlinear features. During the training process, we investigate the impact of hyperparameters and propose the two-stage SSIM&SL1 loss function to reduce structural differences between images. Experimental results show that DTWSTSR outperforms existing mainstream methods under different magnification factors (×2, ×3, ×4), ranking among the top two in multiple metrics. For example, at ×2 magnification, its PSNR value is 0.64–2.68 dB higher than that of other models. Visual comparisons demonstrate that the proposed model achieves clearer and more accurate detail reconstruction of target ground objects. Furthermore, the model exhibits excellent generalization ability in cross-sensor image (OLI2MSI dataset) reconstruction.
高分辨率遥感图像为精准农业和水资源管理等应用提供了重要的数据支持。然而,超分辨率重建往往存在纹理过度平滑和结构扭曲的问题,无法准确恢复地物的复杂细节。针对这一问题,本文提出了一种结合双树复小波变换和Swin变压器的遥感图像超分辨率网络(DTWSTSR),通过融合频域和空域特征,增强了纹理细节的重建能力。该模型采用双树复小波纹理特征感知模块(DWTFSM)对频率和空间特征进行融合,采用多尺度高效通道关注机制对多尺度和全局细节进行关注。此外,我们设计了一个基于分支注意机制的Kolmogorov-Arnold网络,提高了模型表征复杂非线性特征的能力。在训练过程中,我们研究了超参数的影响,并提出了两阶段SSIM&SL1损失函数来减少图像之间的结构差异。实验结果表明,DTWSTSR在不同放大倍数(×2, ×3, ×4)下优于现有主流方法,多项指标均排名前两位。例如,在×2放大倍数下,其PSNR值比其他模型高0.64-2.68 dB。目视对比表明,该模型对目标地物的细节重建更加清晰、准确。此外,该模型在跨传感器图像(OLI2MSI数据集)重建中表现出良好的泛化能力。
{"title":"DTWSTSR: Dual-Tree Complex Wavelet and Swin Transformer Based Remote Sensing Images Super-Resolution Network","authors":"Yu Yao;Hengbin Wang;Xiang Gao;Ziyao Xing;Xiaodong Zhang;Yuanyuan Zhao;Shaoming Li;Zhe Liu","doi":"10.1109/JSTARS.2026.3651075","DOIUrl":"https://doi.org/10.1109/JSTARS.2026.3651075","url":null,"abstract":"High-resolution remote sensing images provide crucial data support for applications such as precision agriculture and water resource management. However, super-resolution reconstructions often suffer from over-smoothed textures and structural distortions, failing to accurately recover the intricate details of ground objects. To address this issue, this article proposes a remote sensing image super-resolution network (DTWSTSR) that combines the Dual-Tree Complex Wavelet Transform and Swin Transformer, which enhances the ability of texture detail reconstruction by fusing frequency-domain and spatial-domain features. This model includes a Dual-Tree Complex Wavelet Texture Feature Sensing Module (DWTFSM) for integrating frequency and spatial features, and a Multiscale Efficient Channel Attention mechanism to enhance attention to multiscale and global details. In addition, we design a Kolmogorov–Arnold Network based on a branch attention mechanism, which improves the model’s ability to represent complex nonlinear features. During the training process, we investigate the impact of hyperparameters and propose the two-stage SSIM&SL1 loss function to reduce structural differences between images. Experimental results show that DTWSTSR outperforms existing mainstream methods under different magnification factors (×2, ×3, ×4), ranking among the top two in multiple metrics. For example, at ×2 magnification, its PSNR value is 0.64–2.68 dB higher than that of other models. Visual comparisons demonstrate that the proposed model achieves clearer and more accurate detail reconstruction of target ground objects. Furthermore, the model exhibits excellent generalization ability in cross-sensor image (OLI2MSI dataset) reconstruction.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"4730-4747"},"PeriodicalIF":5.3,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11329193","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146082033","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Estimation of Ships’ Complex High-Resolution Range Profiles Based on Sparse Optimization Method in Non-Gaussian Sea Clutter 非高斯海杂波下基于稀疏优化方法的舰船复杂高分辨率距离轮廓估计
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-01-12 DOI: 10.1109/JSTARS.2026.3651639
Yang Liu;Kun Zhang;Chun-Yi Song;Zhi-Wei Xu
In high-resolution maritime radar working in scanning mode, the classification and identification of ships require the recovery of the ship’s high-resolution range profiles (HRRPs) from radar returns. The return signal from the ship is a complex sparse signal interfered by non-Gaussian sea clutter. In this article, three sparse optimization methods matching the non-Gaussian characteristics of sea clutter, i.e., the sparse optimization matching K-distribution method, the sparse optimization matching generalized Pareto distribution method, the sparse optimization matching CGIG distribution method, are proposed to estimate complex HRRPs of ships. The compound Gaussian model is used to describe the non-Gaussianity of sea clutter, and the sparsity of ships’ complex HRRPs is constrained by the random distribution with one parameter. In the three methods, the Anderson–Darling test is used to search the parameters of the sparse constraint model. Besides, the non-Gaussian characteristics of sea clutter depend on the marine environment parameters and radar operating parameters. For different scenarios, the minimal criterion of the Kolmogorov–Smirnov distance is used to select the best model from the three compound Gaussian models, and then select the corresponding proposed methods. Simulated and measured radar data are used to evaluate the performance of the proposed methods and the results show that the proposed methods obtain better estimates of ship HRRPs compared to the recent SRIM method and the classical SLIM method.
在扫描模式下工作的高分辨率海事雷达中,船舶的分类和识别需要从雷达回波中恢复船舶的高分辨率距离像(hrrp)。船舶回波信号是受非高斯海杂波干扰的复稀疏信号。本文针对海杂波的非高斯特性,提出了稀疏优化匹配k -分布法、稀疏优化匹配广义Pareto分布法、稀疏优化匹配CGIG分布法三种稀疏优化方法来估计舰船的复杂hrrp。采用复合高斯模型描述海杂波的非高斯性,舰船复杂hrrp的稀疏性受单参数随机分布的约束。在这三种方法中,使用Anderson-Darling检验来搜索稀疏约束模型的参数。此外,海杂波的非高斯特性取决于海洋环境参数和雷达工作参数。针对不同的场景,采用Kolmogorov-Smirnov距离最小准则从三种复合高斯模型中选择最佳模型,然后选择相应的建议方法。利用模拟和实测雷达数据对所提方法的性能进行了评价,结果表明,与现有的SRIM方法和经典的SLIM方法相比,所提方法获得了更好的舰船hrrp估计。
{"title":"Estimation of Ships’ Complex High-Resolution Range Profiles Based on Sparse Optimization Method in Non-Gaussian Sea Clutter","authors":"Yang Liu;Kun Zhang;Chun-Yi Song;Zhi-Wei Xu","doi":"10.1109/JSTARS.2026.3651639","DOIUrl":"https://doi.org/10.1109/JSTARS.2026.3651639","url":null,"abstract":"In high-resolution maritime radar working in scanning mode, the classification and identification of ships require the recovery of the ship’s high-resolution range profiles (HRRPs) from radar returns. The return signal from the ship is a complex sparse signal interfered by non-Gaussian sea clutter. In this article, three sparse optimization methods matching the non-Gaussian characteristics of sea clutter, i.e., the sparse optimization matching K-distribution method, the sparse optimization matching generalized Pareto distribution method, the sparse optimization matching CGIG distribution method, are proposed to estimate complex HRRPs of ships. The compound Gaussian model is used to describe the non-Gaussianity of sea clutter, and the sparsity of ships’ complex HRRPs is constrained by the random distribution with one parameter. In the three methods, the Anderson–Darling test is used to search the parameters of the sparse constraint model. Besides, the non-Gaussian characteristics of sea clutter depend on the marine environment parameters and radar operating parameters. For different scenarios, the minimal criterion of the Kolmogorov–Smirnov distance is used to select the best model from the three compound Gaussian models, and then select the corresponding proposed methods. Simulated and measured radar data are used to evaluate the performance of the proposed methods and the results show that the proposed methods obtain better estimates of ship HRRPs compared to the recent SRIM method and the classical SLIM method.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"3998-4013"},"PeriodicalIF":5.3,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11339885","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146026530","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IR-DETR: An Infrared Small Object Detector Combining Edge-Aware Mechanism and Multiscale Feature Fusion 结合边缘感知机制和多尺度特征融合的红外小目标探测器IR-DETR
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-01-12 DOI: 10.1109/JSTARS.2026.3652368
Jinlong Hu;Ming Zhao;Biying Liu;Xing Chen
Infrared target detection plays a critical role in military reconnaissance, forest-fire prevention, and search-and-rescue operations, owing to its all-weather capability and strong penetration performance. Nevertheless, detecting ultra-small infrared targets in remote sensing and monitoring scenarios remains exceedingly challenging due to the targets’ diminutive size and the long standoff distances involved. In this work, we address this problem by introducing IR-DETR, a lightweight, high-precision infrared small-target detection model built upon the real-time detection Transformer (RT-DETR) framework. First, we propose the multiscale edge-aware convolution module (LDConv), which integrates parallel multiscale Laplacian-of-Gaussian filtering and dilated convolutions within the backbone’s shallow layers, augmented by lightweight channel attention, to markedly enhance the extraction of weak-texture features and boost target–background contrast. Second, we devise the MSCShiftCSP multiscale fusion module: by orchestrating parallel multiscale convolutional branches and parameter-free channel shifting within a CSP structure (replacing the standard RepC3 unit) to strengthen spatial–channel interactions and global context fusion while preserving model efficiency. Third, we replace the original P5 large-target detection head with a high-resolution P2 branch head, halving the channel count from 256 to 128 to capture edges and fine details of targets as small as 6–7 pixels while substantially reducing model complexity. Extensive ablation studies on four public datasets (IRSTD-1 K, SIRST-UAVB, SIRST-v1, and NUDT-SIRST) show that IR-DETR achieves up to a 22% increase in mAP$_{50text{-}95}$ over the RT-DETR baseline with only 9.5 million parameters; notably, on SIRST-UAVB (average target size 6–7pixels), mAP$_{75}$ improves by 47%. Compared against 23 mainstream detectors (including Faster R-CNN, SSD, YOLOv5/8/10/11, and various DETR variants), IR-DETR consistently attains the highest detection accuracy across benchmarks. In summary, IR-DETR delivers a powerful yet lightweight solution for real-time detection of diverse small infrared targets in complex environments, achieving superior accuracy without significant parameter overhead and advancing the state of the art in infrared small-target detection.
红外目标探测以其全天候的能力和较强的突防性能,在军事侦察、森林防火和搜救行动中发挥着重要作用。然而,在遥感和监测场景中探测超小型红外目标仍然非常具有挑战性,因为目标体积小,距离远。在这项工作中,我们通过引入IR-DETR来解决这个问题,IR-DETR是一种轻量级、高精度的红外小目标检测模型,建立在实时检测变压器(RT-DETR)框架之上。首先,我们提出了多尺度边缘感知卷积模块(LDConv),该模块集成了并行多尺度拉普拉斯高斯滤波和骨干浅层内的扩展卷积,并通过轻量级通道关注增强,显著增强了弱纹理特征的提取并提高了目标背景对比度。其次,我们设计了MSCShiftCSP多尺度融合模块:通过在CSP结构内协调并行多尺度卷积分支和无参数信道移动(取代标准的RepC3单元)来加强空间信道相互作用和全局上下文融合,同时保持模型效率。第三,我们将原来的P5大目标检测头替换为高分辨率的P2分支头,将通道数从256个减少到128个,可以捕获小到6-7像素的目标边缘和精细细节,同时大大降低了模型复杂度。在四个公共数据集(irstd - 1k、SIRST-UAVB、SIRST-v1和NUDT-SIRST)上进行的广泛烧蚀研究表明,IR-DETR在仅使用950万个参数的情况下,mAP$ {50text{-}95}$比RT-DETR基线增加了22%;值得注意的是,在SIRST-UAVB(平均目标大小为6 - 7像素)上,mAP$_{75}$提高了47%。与23种主流检测器(包括Faster R-CNN, SSD, YOLOv5/8/10/11和各种DETR变体)相比,IR-DETR在基准测试中始终达到最高的检测精度。总之,IR-DETR为复杂环境中各种小型红外目标的实时检测提供了强大而轻便的解决方案,在没有显著参数开销的情况下实现了卓越的精度,并推进了红外小目标检测的最新状态。
{"title":"IR-DETR: An Infrared Small Object Detector Combining Edge-Aware Mechanism and Multiscale Feature Fusion","authors":"Jinlong Hu;Ming Zhao;Biying Liu;Xing Chen","doi":"10.1109/JSTARS.2026.3652368","DOIUrl":"https://doi.org/10.1109/JSTARS.2026.3652368","url":null,"abstract":"Infrared target detection plays a critical role in military reconnaissance, forest-fire prevention, and search-and-rescue operations, owing to its all-weather capability and strong penetration performance. Nevertheless, detecting ultra-small infrared targets in remote sensing and monitoring scenarios remains exceedingly challenging due to the targets’ diminutive size and the long standoff distances involved. In this work, we address this problem by introducing <italic>IR-DETR</i>, a lightweight, high-precision infrared small-target detection model built upon the real-time detection Transformer (RT-DETR) framework. First, we propose the multiscale edge-aware convolution module (LDConv), which integrates parallel multiscale Laplacian-of-Gaussian filtering and dilated convolutions within the backbone’s shallow layers, augmented by lightweight channel attention, to markedly enhance the extraction of weak-texture features and boost target–background contrast. Second, we devise the MSCShiftCSP multiscale fusion module: by orchestrating parallel multiscale convolutional branches and parameter-free channel shifting within a CSP structure (replacing the standard RepC3 unit) to strengthen spatial–channel interactions and global context fusion while preserving model efficiency. Third, we replace the original P5 large-target detection head with a high-resolution P2 branch head, halving the channel count from 256 to 128 to capture edges and fine details of targets as small as 6–7 pixels while substantially reducing model complexity. Extensive ablation studies on four public datasets (IRSTD-1 K, SIRST-UAVB, SIRST-v1, and NUDT-SIRST) show that IR-DETR achieves up to a 22% increase in mAP<inline-formula><tex-math>$_{50text{-}95}$</tex-math></inline-formula> over the RT-DETR baseline with only 9.5 million parameters; notably, on SIRST-UAVB (average target size 6–7pixels), mAP<inline-formula><tex-math>$_{75}$</tex-math></inline-formula> improves by 47%. Compared against 23 mainstream detectors (including Faster R-CNN, SSD, YOLOv5/8/10/11, and various DETR variants), IR-DETR consistently attains the highest detection accuracy across benchmarks. In summary, IR-DETR delivers a powerful yet lightweight solution for real-time detection of diverse small infrared targets in complex environments, achieving superior accuracy without significant parameter overhead and advancing the state of the art in infrared small-target detection.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"7264-7279"},"PeriodicalIF":5.3,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11339883","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147299583","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SAR-W-MixMAE: Polarization-Aware Self-Supervised Pretraining for Masked Autoencoders on SAR Data 基于SAR数据的掩膜编码器偏振感知自监督预训练
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-01-12 DOI: 10.1109/JSTARS.2026.3652404
Ali Caglayan;Nevrez Imamoglu;Toru Kouyama
Self-supervised pretraining has emerged as a powerful approach for learning transferable representations from large-scale unlabeled data, significantly reducing reliance on task-specific labeled datasets. Although masked autoencoders (MAEs) have shown considerable success in optical remote sensing, such as RGB and multispectral imagery, their application to synthetic aperture radar (SAR) data remains underexplored due to their unique imaging characteristics, including speckle content and intensity variability. In this work, we investigate the effectiveness of MAEs for SAR pretraining, specifically applying MixMAE [Liu, et al.,(2023)] to Sentinel-1 SAR imagery. We introduce SAR-W-MixMAE, a domain-aware self-supervised learning approach that incorporates an SAR-specific pixelwise weighting strategy into the reconstruction loss, mitigating the effects of speckle content and high-intensity backscatter variations. Experimental results demonstrate that SAR-W-MixMAE consistently improves baseline models in multilabel SAR image classification and flood detection tasks, extending the state-of-the-art performance on the popular BigEarthNet dataset. Extensive ablation studies reveal that pretraining duration and fine-tuning dataset size significantly impact downstream performance. In particular, early stopping during pretraining can yield optimal downstream task accuracy, challenging the assumption that prolonged pretraining enhances results. These insights contribute to the development of foundation models tailored for SAR imagery and provide practical guidelines for optimizing pretraining strategies in remote sensing applications.
自监督预训练已经成为一种从大规模未标记数据中学习可转移表征的强大方法,显著减少了对特定任务标记数据集的依赖。尽管掩膜自动编码器(MAEs)在光学遥感(如RGB和多光谱成像)中取得了相当大的成功,但由于其独特的成像特性(包括散斑含量和强度可变性),它们在合成孔径雷达(SAR)数据中的应用仍未得到充分探索。在这项工作中,我们研究了MAEs在SAR预训练中的有效性,特别是将MixMAE [Liu, et .,(2023)]应用于Sentinel-1 SAR图像。我们引入了SAR-W-MixMAE,这是一种领域感知的自监督学习方法,它将sar特定的像素加权策略纳入重建损失,减轻了散斑内容和高强度后向散射变化的影响。实验结果表明,SAR- w - mixmae在多标签SAR图像分类和洪水检测任务中不断改进基线模型,扩展了流行的BigEarthNet数据集的最先进性能。广泛的消融研究表明,预训练时间和微调数据集大小显著影响下游性能。特别是,在预训练期间提前停止可以产生最佳的下游任务准确性,挑战了延长预训练可以提高结果的假设。这些见解有助于开发适合SAR图像的基础模型,并为优化遥感应用中的预训练策略提供实用指南。
{"title":"SAR-W-MixMAE: Polarization-Aware Self-Supervised Pretraining for Masked Autoencoders on SAR Data","authors":"Ali Caglayan;Nevrez Imamoglu;Toru Kouyama","doi":"10.1109/JSTARS.2026.3652404","DOIUrl":"https://doi.org/10.1109/JSTARS.2026.3652404","url":null,"abstract":"Self-supervised pretraining has emerged as a powerful approach for learning transferable representations from large-scale unlabeled data, significantly reducing reliance on task-specific labeled datasets. Although masked autoencoders (MAEs) have shown considerable success in optical remote sensing, such as RGB and multispectral imagery, their application to synthetic aperture radar (SAR) data remains underexplored due to their unique imaging characteristics, including speckle content and intensity variability. In this work, we investigate the effectiveness of MAEs for SAR pretraining, specifically applying MixMAE [Liu, et al.,(2023)] to Sentinel-1 SAR imagery. We introduce SAR-W-MixMAE, a domain-aware self-supervised learning approach that incorporates an SAR-specific pixelwise weighting strategy into the reconstruction loss, mitigating the effects of speckle content and high-intensity backscatter variations. Experimental results demonstrate that SAR-W-MixMAE consistently improves baseline models in multilabel SAR image classification and flood detection tasks, extending the state-of-the-art performance on the popular BigEarthNet dataset. Extensive ablation studies reveal that pretraining duration and fine-tuning dataset size significantly impact downstream performance. In particular, early stopping during pretraining can yield optimal downstream task accuracy, challenging the assumption that prolonged pretraining enhances results. These insights contribute to the development of foundation models tailored for SAR imagery and provide practical guidelines for optimizing pretraining strategies in remote sensing applications.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"5590-5601"},"PeriodicalIF":5.3,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11344788","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146175697","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SFCFNet: A Spatial–Frequency Cross-Attention Fusion Network for Hyperspectral Image Classification SFCFNet:一种用于高光谱图像分类的空频交叉关注融合网络
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-01-12 DOI: 10.1109/JSTARS.2026.3651577
Wei Huang;JiaLu Li;Qiqiang Chen;Junru Yin;Jiqiang Niu;Le Sun
In recent years, the integration of convolutional neural networks and Transformers has significantly advanced hyperspectral image (HSI) classification by jointly capturing local and global features. However, most existing methods primarily focus on the fusion of spectral–spatial features while neglecting the complementary information contained in frequency-domain features. To address this issue, we propose a spatial–frequency cross-attention fusion network (SFCFNet) that jointly models spectral, spatial, and frequency-domain features for HSI classification. The framework consists of three core modules: first, the multiscale spectral–spatial feature learning module extracts joint spectral spatial features using multiscale 3-D and 2-D convolutions. Next, the triple-branch representation module employs three branches to capture global spatial features of large-scale structures, local spatial features of fine-grained textures, and multiscale frequency features based on Haar wavelet decomposition, providing complementary multidomain representations for subsequent deep fusion. Finally, the dual-domain feature cross-attention fusion module achieves effective fusion of spatial structures and frequency-domain textures, enhancing the model’s ability to separate complex backgrounds from fine-grained targets and thereby improving classification performance. Compared with other methods, SFCFNet achieves higher overall accuracy on the Salinas, Houston2013, WHU-Hi-LongKou, and Xuzhou datasets, reaching 99.05%, 98.07%, 98.76%, and 98.18%, respectively.
近年来,卷积神经网络与transformer的融合通过联合捕获局部和全局特征,极大地推进了高光谱图像(HSI)分类。然而,现有的方法大多侧重于频谱空间特征的融合,而忽略了频域特征中包含的互补信息。为了解决这个问题,我们提出了一个空间-频率交叉关注融合网络(SFCFNet),该网络联合建模频谱、空间和频域特征,用于HSI分类。该框架由三个核心模块组成:第一,多尺度光谱-空间特征学习模块使用多尺度三维和二维卷积提取联合光谱空间特征;接下来,三分支表示模块利用三分支捕获大尺度结构的全局空间特征、细粒度纹理的局部空间特征以及基于Haar小波分解的多尺度频率特征,为后续深度融合提供互补的多域表示。最后,双域特征交叉关注融合模块实现了空间结构和频域纹理的有效融合,增强了模型从细粒度目标中分离复杂背景的能力,从而提高了分类性能。与其他方法相比,SFCFNet在Salinas、Houston2013、WHU-Hi-LongKou和徐州数据集上的总体精度分别达到99.05%、98.07%、98.76%和98.18%。
{"title":"SFCFNet: A Spatial–Frequency Cross-Attention Fusion Network for Hyperspectral Image Classification","authors":"Wei Huang;JiaLu Li;Qiqiang Chen;Junru Yin;Jiqiang Niu;Le Sun","doi":"10.1109/JSTARS.2026.3651577","DOIUrl":"https://doi.org/10.1109/JSTARS.2026.3651577","url":null,"abstract":"In recent years, the integration of convolutional neural networks and Transformers has significantly advanced hyperspectral image (HSI) classification by jointly capturing local and global features. However, most existing methods primarily focus on the fusion of spectral–spatial features while neglecting the complementary information contained in frequency-domain features. To address this issue, we propose a spatial–frequency cross-attention fusion network (SFCFNet) that jointly models spectral, spatial, and frequency-domain features for HSI classification. The framework consists of three core modules: first, the multiscale spectral–spatial feature learning module extracts joint spectral spatial features using multiscale 3-D and 2-D convolutions. Next, the triple-branch representation module employs three branches to capture global spatial features of large-scale structures, local spatial features of fine-grained textures, and multiscale frequency features based on Haar wavelet decomposition, providing complementary multidomain representations for subsequent deep fusion. Finally, the dual-domain feature cross-attention fusion module achieves effective fusion of spatial structures and frequency-domain textures, enhancing the model’s ability to separate complex backgrounds from fine-grained targets and thereby improving classification performance. Compared with other methods, SFCFNet achieves higher overall accuracy on the Salinas, Houston2013, WHU-Hi-LongKou, and Xuzhou datasets, reaching 99.05%, 98.07%, 98.76%, and 98.18%, respectively.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"4994-5008"},"PeriodicalIF":5.3,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11340627","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146082022","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Spatiotemporal Heterogeneity in Greenland Firn From the Synthesis of Satellite Radar Altimetry and Passive Microwave Measurements 基于卫星雷达测高和被动微波测量综合的格陵兰岛植被时空异质性
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-01-12 DOI: 10.1109/JSTARS.2026.3651847
Kirk M. Scanlan;Anja Rutishauser;Sebastian B. Simonsen
The spatiotemporal properties of the Greenland Ice Sheet firn layer are an important factor when assessing overall ice sheet mass balance and internal meltwater storage capacity. Increasingly a target for the satellite remote sensing community, this study investigates the recovery of vertical firn density heterogeneity over a ten-year period from the synthesis of passive microwave and active radar altimetry measurements. The mismatch between ESA SMOS observations and a passive microwave forward model, initialized with surface densities estimated from the backscatter strength of ISRO/CNES SARAL and ESA CryoSat-2, serves as a proxy for vertical density variability. Validated with in situ measurements, the results demonstrate clear long-term patterns in Greenland firn heterogeneity characterized by spatially expansive sharp increases in firn heterogeneity following extreme melt seasons that require multiple quiescent years to rehabilitate. The results demonstrate that by the start of the 2023 melt season (i.e., the end of the timeframe considered), the Greenland firn layer had reached its most heterogeneous state of the preceding decade. Continued investigation into the synthesis of different remote sensing datasets represents a pathway toward generating novel insights into the spatiotemporal evolution of Greenland Ice Sheet surface conditions.
格陵兰冰盖冰层的时空特征是评估冰盖整体质量平衡和内部融水储存能力的重要因素。作为卫星遥感界日益关注的一个目标,本研究从被动微波和主动雷达测高的综合测量中恢复了十年来垂直植被密度的非均质性。ESA SMOS观测与被动微波正演模型之间的不匹配可以作为垂直密度变化的代理,该模型初始化了ISRO/CNES SARAL和ESA CryoSat-2的后向散射强度估算的表面密度。通过原位测量验证,结果表明格陵兰冰质异质性具有明确的长期模式,其特征是在极端融化季节之后,冰质异质性在空间上急剧增加,需要多个静息年才能恢复。结果表明,到2023年融化季节开始时(即所考虑的时间框架结束时),格陵兰冰层已达到前十年最不均匀的状态。对不同遥感数据集的持续综合研究代表了对格陵兰冰盖表面条件的时空演变产生新见解的途径。
{"title":"Spatiotemporal Heterogeneity in Greenland Firn From the Synthesis of Satellite Radar Altimetry and Passive Microwave Measurements","authors":"Kirk M. Scanlan;Anja Rutishauser;Sebastian B. Simonsen","doi":"10.1109/JSTARS.2026.3651847","DOIUrl":"https://doi.org/10.1109/JSTARS.2026.3651847","url":null,"abstract":"The spatiotemporal properties of the Greenland Ice Sheet firn layer are an important factor when assessing overall ice sheet mass balance and internal meltwater storage capacity. Increasingly a target for the satellite remote sensing community, this study investigates the recovery of vertical firn density heterogeneity over a ten-year period from the synthesis of passive microwave and active radar altimetry measurements. The mismatch between ESA SMOS observations and a passive microwave forward model, initialized with surface densities estimated from the backscatter strength of ISRO/CNES SARAL and ESA CryoSat-2, serves as a proxy for vertical density variability. Validated with in situ measurements, the results demonstrate clear long-term patterns in Greenland firn heterogeneity characterized by spatially expansive sharp increases in firn heterogeneity following extreme melt seasons that require multiple quiescent years to rehabilitate. The results demonstrate that by the start of the 2023 melt season (i.e., the end of the timeframe considered), the Greenland firn layer had reached its most heterogeneous state of the preceding decade. Continued investigation into the synthesis of different remote sensing datasets represents a pathway toward generating novel insights into the spatiotemporal evolution of Greenland Ice Sheet surface conditions.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"4085-4098"},"PeriodicalIF":5.3,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11339888","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146026391","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning Boundary-Aware Semantic Context Network for Remote Sensing Change Detection 基于学习边界感知的语义上下文网络遥感变化检测
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-01-12 DOI: 10.1109/JSTARS.2026.3651696
Weiran Zhou;Guanting Guo;Huihui Song;Xu Zhang;Kaihua Zhang
Remote sensing change detection aims to identify changes on the Earth's surface from remote sensing images acquired at different times. However, the identification of changed areas is often hindered by pseudochanges in similar objects, leading to inaccurate identification of change boundaries. To address this issue, we propose a novel network named boundary-guided semantic context network (BSCNet), which decouples features to improve the feature representation ability for changing objects. Specifically, we design a selective context fusion module that selectively fuses semantically rich features by computing the similarity between features from adjacent stages of the backbone network, thereby preventing detailed features from being overwhelmed by contextual information. In addition, to enhance the ability to perceive changes, we design a context fast aggregation module that leverages a pyramid structure to help the model simultaneously extract and fuse detailed and semantic information at different scales, enabling more accurate change detection. Finally, we design a boundary-guided feature fusion module to aggregate edge-level, texture-level, and semantic-level information, which enables the network to represent change regions more comprehensively and precisely. Experimental results on the WHU-CD, LEVIR-CD, and SYSU-CD datasets show that BSCNet achieves F1 scores of 94.92%, 92.19%, and 82.55%, respectively.
遥感变化检测的目的是从不同时间获取的遥感图像中识别地球表面的变化。然而,变化区域的识别往往受到类似对象的伪变化的阻碍,导致变化边界的不准确识别。为了解决这一问题,我们提出了一种新的网络,即边界引导语义上下文网络(BSCNet),该网络将特征解耦,以提高对变化对象的特征表示能力。具体而言,我们设计了一个选择性上下文融合模块,该模块通过计算骨干网相邻阶段特征之间的相似性来选择性地融合语义丰富的特征,从而防止详细特征被上下文信息淹没。此外,为了增强感知变化的能力,我们设计了一个上下文快速聚合模块,该模块利用金字塔结构帮助模型同时提取和融合不同尺度的细节和语义信息,从而实现更准确的变化检测。最后,设计了边界引导的特征融合模块,对边缘级、纹理级和语义级信息进行聚合,使网络能够更全面、更准确地表示变化区域。在WHU-CD、levird - cd和SYSU-CD数据集上的实验结果表明,BSCNet的F1得分分别为94.92%、92.19%和82.55%。
{"title":"Learning Boundary-Aware Semantic Context Network for Remote Sensing Change Detection","authors":"Weiran Zhou;Guanting Guo;Huihui Song;Xu Zhang;Kaihua Zhang","doi":"10.1109/JSTARS.2026.3651696","DOIUrl":"https://doi.org/10.1109/JSTARS.2026.3651696","url":null,"abstract":"Remote sensing change detection aims to identify changes on the Earth's surface from remote sensing images acquired at different times. However, the identification of changed areas is often hindered by pseudochanges in similar objects, leading to inaccurate identification of change boundaries. To address this issue, we propose a novel network named boundary-guided semantic context network (BSCNet), which decouples features to improve the feature representation ability for changing objects. Specifically, we design a selective context fusion module that selectively fuses semantically rich features by computing the similarity between features from adjacent stages of the backbone network, thereby preventing detailed features from being overwhelmed by contextual information. In addition, to enhance the ability to perceive changes, we design a context fast aggregation module that leverages a pyramid structure to help the model simultaneously extract and fuse detailed and semantic information at different scales, enabling more accurate change detection. Finally, we design a boundary-guided feature fusion module to aggregate edge-level, texture-level, and semantic-level information, which enables the network to represent change regions more comprehensively and precisely. Experimental results on the WHU-CD, LEVIR-CD, and SYSU-CD datasets show that BSCNet achieves F1 scores of 94.92%, 92.19%, and 82.55%, respectively.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"4177-4187"},"PeriodicalIF":5.3,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11339892","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146026466","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PyramidMamba: An Effective Hyperspectral Remote Sensing Image Target Detection Network 金字塔曼巴:一种有效的高光谱遥感图像目标检测网络
IF 5.3 2区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-01-12 DOI: 10.1109/JSTARS.2026.3650961
Shixin Liu;Pingyu Liu;Xiaofei Wang
The lack of prior knowledge is a challenging issue in target detection tasks for hyperspectral remote sensing images. In this article, we propose an effective network for object detection in hyperspectral remote sensing images. First, through spectral data augmentation methods, all surrounding pixels within a data block are encoded as the transformed spectral signature of the central pixel, thereby constructing a sufficient number of training sample pairs. Subsequently, a backbone network (PyramidMamba) was designed to establish long-term dependencies across the frequency domain and multiscale dimensions using the Mamba residual module and pyramid wavelet transform module. A residual self-attention module is further developed, integrating self-attention with convolutional operations to enhance feature extraction while improving the network's depth and stability. A backbone network was employed to extract representative vectors from augmented sample pairs, which were then optimized through a spectral contrast head to enhance the distinction between target and background features. Experimental results demonstrate that compared to mainstream algorithms, the proposed algorithm achieves higher detection accuracy and computational efficiency. It successfully learns deep nonlinear feature representations with stronger discriminative power, enabling effective separation of targets from background and delivering state-of-the-art performance.
在高光谱遥感图像的目标检测任务中,缺乏先验知识是一个具有挑战性的问题。在本文中,我们提出了一种有效的高光谱遥感图像目标检测网络。首先,通过光谱数据增强方法,将数据块内所有周围像素编码为中心像素变换后的光谱签名,从而构造足够数量的训练样本对。随后,利用Mamba残差模块和金字塔小波变换模块,设计了一个骨干网络(PyramidMamba),以建立跨频域和多尺度维度的长期依赖关系。进一步开发残差自注意模块,将自注意与卷积运算相结合,增强特征提取,同时提高网络的深度和稳定性。利用骨干网络从增强的样本对中提取代表性向量,然后通过光谱对比头对其进行优化,以增强目标和背景特征的区别。实验结果表明,与主流算法相比,该算法具有更高的检测精度和计算效率。它成功地学习了深度非线性特征表示,具有更强的判别能力,能够有效地将目标与背景分离,并提供最先进的性能。
{"title":"PyramidMamba: An Effective Hyperspectral Remote Sensing Image Target Detection Network","authors":"Shixin Liu;Pingyu Liu;Xiaofei Wang","doi":"10.1109/JSTARS.2026.3650961","DOIUrl":"https://doi.org/10.1109/JSTARS.2026.3650961","url":null,"abstract":"The lack of prior knowledge is a challenging issue in target detection tasks for hyperspectral remote sensing images. In this article, we propose an effective network for object detection in hyperspectral remote sensing images. First, through spectral data augmentation methods, all surrounding pixels within a data block are encoded as the transformed spectral signature of the central pixel, thereby constructing a sufficient number of training sample pairs. Subsequently, a backbone network (PyramidMamba) was designed to establish long-term dependencies across the frequency domain and multiscale dimensions using the Mamba residual module and pyramid wavelet transform module. A residual self-attention module is further developed, integrating self-attention with convolutional operations to enhance feature extraction while improving the network's depth and stability. A backbone network was employed to extract representative vectors from augmented sample pairs, which were then optimized through a spectral contrast head to enhance the distinction between target and background features. Experimental results demonstrate that compared to mainstream algorithms, the proposed algorithm achieves higher detection accuracy and computational efficiency. It successfully learns deep nonlinear feature representations with stronger discriminative power, enabling effective separation of targets from background and delivering state-of-the-art performance.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"4163-4176"},"PeriodicalIF":5.3,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11329180","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146026481","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1