首页 > 最新文献

IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society最新文献

英文 中文
Multiscale Occlusion-Robust Scene Classification in Remote Sensing Images via Supervised Contrastive Learning
Zhao Yang;Jia Chen;Jun Li;Xiangan Zheng
Scene classification of remote sensing images plays a vital role in Earth observation applications. Among various challenges, occlusion is a prevalent and critical issue in practical applications, particularly when dealing with large-area occlusions caused by clouds, shadows, and man-made structures. Current methods, whether based on occlusion recovery or occlusion-robust feature extraction, generally show limited performance when processing extensive occluded regions due to ignoring the inconsistency in feature representation caused by multiscale occlusions. To address the occlusion challenge, this letter proposes a novel contrastive learning-based multiscale occlusion framework with three key components: 1) a pretext task module that distinguishes between small and large occlusions to enable occlusion-invariant feature learning; 2) a multibranch feature extraction network based on ResNet-50’s shared-weight convolutional layers for consistent feature extraction across occlusion levels; and 3) a joint loss function that adaptively balances contrastive feature learning and classification. Extensive experiments were evaluated on the DIOR-Occ and LEVIR-Occ benchmark datasets, demonstrating significant improvements in classification accuracy across different occlusion scenarios. Compared with existing approaches, the proposed framework achieves superior robustness and generalization capabilities, with notable advantages in the analysis of highly occluded data. Future research will explore the adaptation of this framework to detection and segmentation tasks.
{"title":"Multiscale Occlusion-Robust Scene Classification in Remote Sensing Images via Supervised Contrastive Learning","authors":"Zhao Yang;Jia Chen;Jun Li;Xiangan Zheng","doi":"10.1109/LGRS.2025.3537104","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3537104","url":null,"abstract":"Scene classification of remote sensing images plays a vital role in Earth observation applications. Among various challenges, occlusion is a prevalent and critical issue in practical applications, particularly when dealing with large-area occlusions caused by clouds, shadows, and man-made structures. Current methods, whether based on occlusion recovery or occlusion-robust feature extraction, generally show limited performance when processing extensive occluded regions due to ignoring the inconsistency in feature representation caused by multiscale occlusions. To address the occlusion challenge, this letter proposes a novel contrastive learning-based multiscale occlusion framework with three key components: 1) a pretext task module that distinguishes between small and large occlusions to enable occlusion-invariant feature learning; 2) a multibranch feature extraction network based on ResNet-50’s shared-weight convolutional layers for consistent feature extraction across occlusion levels; and 3) a joint loss function that adaptively balances contrastive feature learning and classification. Extensive experiments were evaluated on the DIOR-Occ and LEVIR-Occ benchmark datasets, demonstrating significant improvements in classification accuracy across different occlusion scenarios. Compared with existing approaches, the proposed framework achieves superior robustness and generalization capabilities, with notable advantages in the analysis of highly occluded data. Future research will explore the adaptation of this framework to detection and segmentation tasks.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"22 ","pages":"1-5"},"PeriodicalIF":0.0,"publicationDate":"2025-01-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143496498","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A High-Resolution Imaging Method of Ionosonde Based on Spatial-Frequency 2-D Spectrum Estimation Technology
Tongxin Liu;Guobin Yang;Chunhua Jiang;Hao Li;Chongzhe Lao
The ionosonde is one of the most widely used ionospheric detection devices. It has significant implications for the study of space physics and wireless communication technology. To solve the problem of low resolution with traditional iososonde imaging method, this letter proposes a high-resolution 2-D spatial-frequency spectral estimation imaging method. By reconstructing the echo signal matrix and constructing the guidance vector that simultaneously reflects the range and motion characteristics of the target, high-precision estimation of the range-Doppler spectrum is achieved in 2-D Capon estimation. The analysis of experimental sporadic-E (Es) sounding data of August 13, 2021 in Wuhan ( $114^{circ } 22^{prime } $ E, $30^{circ } 30^{prime } $ N), China, demonstrates its ability of the precise measurement for single-layer echoes and the identification ability for multilayer fine structures. This method further enhances the potential of traditional ionosonde for precise observation and analysis of the ionosphere.
{"title":"A High-Resolution Imaging Method of Ionosonde Based on Spatial-Frequency 2-D Spectrum Estimation Technology","authors":"Tongxin Liu;Guobin Yang;Chunhua Jiang;Hao Li;Chongzhe Lao","doi":"10.1109/LGRS.2025.3537288","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3537288","url":null,"abstract":"The ionosonde is one of the most widely used ionospheric detection devices. It has significant implications for the study of space physics and wireless communication technology. To solve the problem of low resolution with traditional iososonde imaging method, this letter proposes a high-resolution 2-D spatial-frequency spectral estimation imaging method. By reconstructing the echo signal matrix and constructing the guidance vector that simultaneously reflects the range and motion characteristics of the target, high-precision estimation of the range-Doppler spectrum is achieved in 2-D Capon estimation. The analysis of experimental sporadic-E (Es) sounding data of August 13, 2021 in Wuhan (<inline-formula> <tex-math>$114^{circ } 22^{prime } $ </tex-math></inline-formula>E, <inline-formula> <tex-math>$30^{circ } 30^{prime } $ </tex-math></inline-formula>N), China, demonstrates its ability of the precise measurement for single-layer echoes and the identification ability for multilayer fine structures. This method further enhances the potential of traditional ionosonde for precise observation and analysis of the ionosphere.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"22 ","pages":"1-5"},"PeriodicalIF":0.0,"publicationDate":"2025-01-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143403908","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Simplified Predictor With Adjustable Compression Ratio Based on CCSDS 123.0-B-2
Lei Zhang;Peng Zhang;Chengpeng Song;Longsheng Zhang
The Consultative Committee for Space Data Systems (CCSDS) has released the CCSDS 123.0-B-2 standard, which supports lossless and near-lossless compression for multispectral and hyperspectral images. This standard introduces a quantization feedback loop in the predictor. However, this feedback loop significantly reduces throughput efficiency in hardware implementations, hindering real-time on-board compression. To address this issue, this letter proposes a single-step prediction strategy. In the modes of band interleaved by line (BIL) data transmission mode, throughput is improved from 17.86 to 101 Msamples/s, achieving real-time processing. Furthermore, this letter introduces an adaptive maximum reconstruction error to achieve a controllable compression ratio, thus addressing the challenges posed by fixed and limited downlink bandwidth in harsh environments.
空间数据系统协商委员会(CCSDS)发布了 CCSDS 123.0-B-2 标准,该标准支持多光谱和高光谱图像的无损和近无损压缩。该标准在预测器中引入了量化反馈回路。然而,这一反馈回路大大降低了硬件实现的吞吐效率,阻碍了实时板载压缩。为解决这一问题,本文提出了一种单步预测策略。在带交错线(BIL)数据传输模式下,吞吐量从 17.86 Msamples/s 提高到 101 Msamples/s,实现了实时处理。此外,本文还引入了自适应最大重构误差,以实现可控的压缩比,从而应对恶劣环境下固定和有限的下行链路带宽所带来的挑战。
{"title":"A Simplified Predictor With Adjustable Compression Ratio Based on CCSDS 123.0-B-2","authors":"Lei Zhang;Peng Zhang;Chengpeng Song;Longsheng Zhang","doi":"10.1109/LGRS.2025.3536038","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3536038","url":null,"abstract":"The Consultative Committee for Space Data Systems (CCSDS) has released the CCSDS 123.0-B-2 standard, which supports lossless and near-lossless compression for multispectral and hyperspectral images. This standard introduces a quantization feedback loop in the predictor. However, this feedback loop significantly reduces throughput efficiency in hardware implementations, hindering real-time on-board compression. To address this issue, this letter proposes a single-step prediction strategy. In the modes of band interleaved by line (BIL) data transmission mode, throughput is improved from 17.86 to 101 Msamples/s, achieving real-time processing. Furthermore, this letter introduces an adaptive maximum reconstruction error to achieve a controllable compression ratio, thus addressing the challenges posed by fixed and limited downlink bandwidth in harsh environments.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"22 ","pages":"1-5"},"PeriodicalIF":0.0,"publicationDate":"2025-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143388607","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Feature Enhancement and Feedback Network for Change Detection in Remote Sensing Images
Zhenghao Jiang;Biao Wang;Xiao Xu;YaoBo Zhang;Peng Zhang;Yanlan Wu;Hui Yang
Remote sensing change detection (CD) has garnered extensive research and application due to its ability to identify changes in land features within the same area across different periods. CD tasks require features with strong intraclass distinctions and precise spatial boundary details. Existing methods enhance the extraction of difference features but significantly increase computational complexity in high-resolution remote sensing imagery. Moreover, these methods focus on pixel-level difference extraction while neglecting feedback from the overall change object. As a result, they lack global information perception, leading to blurred edges and fragmented interiors in the change areas. To address these challenges, we propose a feature enhancement and feedback network (FEFNet) for CD. First, we designed a multilevel dual-feature fusion enhancement module (DFFM) to improve the representation of latent features between the bitemporal images. Second, we developed a feature coupling feedback module (FCFM) that efficiently decodes multiscale change features to generate extraction results. The experimental results show that FEFNet outperforms recent models in both computational efficiency and detection performance. With only 10.56G FLOPs and 2.26M parameters, FEFNet achieves an ${F}_{1}$ score of 92.32% on the LEVIR-CD dataset and 93.77% on the WHU-CD dataset. The code will be available at https://github.com/XiaoJ058/RS-CD.
{"title":"Feature Enhancement and Feedback Network for Change Detection in Remote Sensing Images","authors":"Zhenghao Jiang;Biao Wang;Xiao Xu;YaoBo Zhang;Peng Zhang;Yanlan Wu;Hui Yang","doi":"10.1109/LGRS.2025.3536164","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3536164","url":null,"abstract":"Remote sensing change detection (CD) has garnered extensive research and application due to its ability to identify changes in land features within the same area across different periods. CD tasks require features with strong intraclass distinctions and precise spatial boundary details. Existing methods enhance the extraction of difference features but significantly increase computational complexity in high-resolution remote sensing imagery. Moreover, these methods focus on pixel-level difference extraction while neglecting feedback from the overall change object. As a result, they lack global information perception, leading to blurred edges and fragmented interiors in the change areas. To address these challenges, we propose a feature enhancement and feedback network (FEFNet) for CD. First, we designed a multilevel dual-feature fusion enhancement module (DFFM) to improve the representation of latent features between the bitemporal images. Second, we developed a feature coupling feedback module (FCFM) that efficiently decodes multiscale change features to generate extraction results. The experimental results show that FEFNet outperforms recent models in both computational efficiency and detection performance. With only 10.56G FLOPs and 2.26M parameters, FEFNet achieves an <inline-formula> <tex-math>${F}_{1}$ </tex-math></inline-formula> score of 92.32% on the LEVIR-CD dataset and 93.77% on the WHU-CD dataset. The code will be available at <uri>https://github.com/XiaoJ058/RS-CD</uri>.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"22 ","pages":"1-5"},"PeriodicalIF":0.0,"publicationDate":"2025-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143360899","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SHAP-Assisted Resilience Enhancement Against Adversarial Perturbations in Optical and SAR Image Classification
Amir Hosein Oveis;Alessandro Cantelli-Forti;Elisa Giusti;Meysam Soltanpour;Neda Rojhani;Marco Martorella
The increasing reliance on convolutional neural networks (CNNs) for automatic target recognition (ATR) in critical applications necessitates robust defenses against adversarial attacks, which can undermine their reliability. To address this challenge, this letter proposes a novel classification framework that enhances CNN robustness for ATR under adversarial perturbations. Although CNNs are renowned for their high recognition accuracy, their performance can be compromised by subtle adversarial perturbations designed to deceive the classifier. Our methodology is based on extracting specific features from Shapley additive explanations (SHAP) analysis within and outside the detected target area. These features are then used to train a multinomial logistic regression model using the training labels, and the trained regressor performs the classification. The key strength of our framework relies on robustness enhancement against adversarial attacks, particularly designed by the fast gradient sign method (FGSM). We validate our findings through extensive evaluations using two publicly available datasets: the multitype aircraft remote sensing images (MTARSI) dataset, which contains optical images of various aircraft types, and the moving and stationary target acquisition and recognition (MSTAR) dataset, which contains radar images.
{"title":"SHAP-Assisted Resilience Enhancement Against Adversarial Perturbations in Optical and SAR Image Classification","authors":"Amir Hosein Oveis;Alessandro Cantelli-Forti;Elisa Giusti;Meysam Soltanpour;Neda Rojhani;Marco Martorella","doi":"10.1109/LGRS.2025.3536005","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3536005","url":null,"abstract":"The increasing reliance on convolutional neural networks (CNNs) for automatic target recognition (ATR) in critical applications necessitates robust defenses against adversarial attacks, which can undermine their reliability. To address this challenge, this letter proposes a novel classification framework that enhances CNN robustness for ATR under adversarial perturbations. Although CNNs are renowned for their high recognition accuracy, their performance can be compromised by subtle adversarial perturbations designed to deceive the classifier. Our methodology is based on extracting specific features from Shapley additive explanations (SHAP) analysis within and outside the detected target area. These features are then used to train a multinomial logistic regression model using the training labels, and the trained regressor performs the classification. The key strength of our framework relies on robustness enhancement against adversarial attacks, particularly designed by the fast gradient sign method (FGSM). We validate our findings through extensive evaluations using two publicly available datasets: the multitype aircraft remote sensing images (MTARSI) dataset, which contains optical images of various aircraft types, and the moving and stationary target acquisition and recognition (MSTAR) dataset, which contains radar images.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"22 ","pages":"1-5"},"PeriodicalIF":0.0,"publicationDate":"2025-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143360968","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SOLSTM: Multisource Information Fusion Semantic Segmentation Network Based on SAR-OPT Matching Attention and Long Short-Term Memory Network
Hao Chang;Xiongjun Fu;Kunyi Guo;Jian Dong;Jialin Guan;Chuyi Liu
With the significant advancements in deep learning technology and the substantial improvement in remote sensing image resolution, remote sensing semantic segmentation has garnered widespread attention. Synthetic aperture radar (SAR) and optical images are the primary sources of remote sensing data, offering complementary information. SAR images can capture surface information even under cloud cover and at night, whereas optical images provide higher resolution in clear weather conditions. Deep learning-based feature fusion methods can effectively integrate multisource information to obtain more comprehensive surface data. However, there are significant spatiotemporal differences in multisource information, making it challenging to select and extract the most discriminative features for segmentation tasks. To address this, we propose a lightweight and efficient fusion semantic segmentation network, SOLSTM, which mixes SAR and optical images as inputs and performs cyclic cross-fusion to establish a new network paradigm. To tackle multisource data heterogeneity, we introduce SAR-OPT matching attention, which aggregates multisource image features by adaptively adjusting fusion weights, thereby achieving comprehensive perception of feature channels and contextual information. Additionally, to mitigate the high computational complexity of processing multidimensional data, we introduce the mLSTM block, which employs linear operations to mine global contextual information in fused images, thus reducing computational complexity and enhancing image segmentation performance. Experiments on the WHU-OPT-SAR dataset show that SOLSTM has excellent performance, achieving up to 52.9 mIoU and outperforming single source image segmentation, verifying the effective fusion of OPT-SAR.
{"title":"SOLSTM: Multisource Information Fusion Semantic Segmentation Network Based on SAR-OPT Matching Attention and Long Short-Term Memory Network","authors":"Hao Chang;Xiongjun Fu;Kunyi Guo;Jian Dong;Jialin Guan;Chuyi Liu","doi":"10.1109/LGRS.2025.3535524","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3535524","url":null,"abstract":"With the significant advancements in deep learning technology and the substantial improvement in remote sensing image resolution, remote sensing semantic segmentation has garnered widespread attention. Synthetic aperture radar (SAR) and optical images are the primary sources of remote sensing data, offering complementary information. SAR images can capture surface information even under cloud cover and at night, whereas optical images provide higher resolution in clear weather conditions. Deep learning-based feature fusion methods can effectively integrate multisource information to obtain more comprehensive surface data. However, there are significant spatiotemporal differences in multisource information, making it challenging to select and extract the most discriminative features for segmentation tasks. To address this, we propose a lightweight and efficient fusion semantic segmentation network, SOLSTM, which mixes SAR and optical images as inputs and performs cyclic cross-fusion to establish a new network paradigm. To tackle multisource data heterogeneity, we introduce SAR-OPT matching attention, which aggregates multisource image features by adaptively adjusting fusion weights, thereby achieving comprehensive perception of feature channels and contextual information. Additionally, to mitigate the high computational complexity of processing multidimensional data, we introduce the mLSTM block, which employs linear operations to mine global contextual information in fused images, thus reducing computational complexity and enhancing image segmentation performance. Experiments on the WHU-OPT-SAR dataset show that SOLSTM has excellent performance, achieving up to 52.9 mIoU and outperforming single source image segmentation, verifying the effective fusion of OPT-SAR.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"22 ","pages":"1-5"},"PeriodicalIF":0.0,"publicationDate":"2025-01-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143396319","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Gamma Log Inversion of Seismic Data Based on Transformer With Stratigraphic Position Encoding
Yongjian Zhou;Haochen Qi;Wang Zhang;Xiaocai Shan
As an indispensable part of geophysical exploration, seismic inversion can obtain the properties of subsurface media based on seismic data and available well-log information. With the nonlinear mapping ability, deep neural networks can map seismic data to well-log of interest. Interpreting gamma is crucial as it is essential for determining lithology and indicating sediment characteristics. Stratigraphic frameworks can approximate low-frequency trends in subsurface properties and are often used to guide well-log interpolation effectively. However, the existing deep neural network models cannot effectively explicitly fuse critical stratigraphic information, which will restrict the physical explainability and correctness of the seismic inversion. Thus, we propose a stratigraphic-encoded transformer algorithm, named SeisWellTrans, to build a gamma log inversion model using horizon position encoding and seismic trace as inputs. Specifically, the incorporation of stratigraphic information from several horizons is crucial for improving the resolution of the output; and SeisWellTrans can efficiently model context in seismic sequences by capturing the interactions between horizon position encodings. We take the Volve field data as an example and use several gamma curves as training labels, and numerical experiments demonstrate the geologically reasonable performance and high validation accuracy of this network and the crucial role that stratigraphic information plays. On the four validation wells, stratigraphic-encoded SeisWellTrans obtained an average correlation coefficient of 86%, exceeding 79% of stratigraphic-encoded convolutional neural network (CNN).
{"title":"Gamma Log Inversion of Seismic Data Based on Transformer With Stratigraphic Position Encoding","authors":"Yongjian Zhou;Haochen Qi;Wang Zhang;Xiaocai Shan","doi":"10.1109/LGRS.2025.3535723","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3535723","url":null,"abstract":"As an indispensable part of geophysical exploration, seismic inversion can obtain the properties of subsurface media based on seismic data and available well-log information. With the nonlinear mapping ability, deep neural networks can map seismic data to well-log of interest. Interpreting gamma is crucial as it is essential for determining lithology and indicating sediment characteristics. Stratigraphic frameworks can approximate low-frequency trends in subsurface properties and are often used to guide well-log interpolation effectively. However, the existing deep neural network models cannot effectively explicitly fuse critical stratigraphic information, which will restrict the physical explainability and correctness of the seismic inversion. Thus, we propose a stratigraphic-encoded transformer algorithm, named SeisWellTrans, to build a gamma log inversion model using horizon position encoding and seismic trace as inputs. Specifically, the incorporation of stratigraphic information from several horizons is crucial for improving the resolution of the output; and SeisWellTrans can efficiently model context in seismic sequences by capturing the interactions between horizon position encodings. We take the Volve field data as an example and use several gamma curves as training labels, and numerical experiments demonstrate the geologically reasonable performance and high validation accuracy of this network and the crucial role that stratigraphic information plays. On the four validation wells, stratigraphic-encoded SeisWellTrans obtained an average correlation coefficient of 86%, exceeding 79% of stratigraphic-encoded convolutional neural network (CNN).","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"22 ","pages":"1-5"},"PeriodicalIF":0.0,"publicationDate":"2025-01-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10856238","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143465881","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Efficient Hybrid Domain Algorithm for Accurate SAR Raw Data Generation With Trajectory Deviations
Bo Wan;Lei Zhang;Jianxin Wu;Guanyong Wang;Zirui Xi
An efficient raw data generation (RDG) algorithm in the hybrid domain is proposed, which can be applied to the accurate echo generation of spotlight mode synthetic aperture radar (SAR) with trajectory deviation, even in cases of terrain undulation. Generally, the accuracy required for the signal’s envelope is on the order of the range resolution cell. However, the requirement for the phase is significantly higher, on the order of the wavelength. Therefore, the proposed algorithm calculates the phase of the SAR raw data in the time domain through a point-by-point approach to ensure accuracy and computes the envelope of the raw data through subblock processing in the frequency domain to improve efficiency. To balance the computational efficiency and accuracy, the optimal selection of subblock size is discussed in detail. Simulation experiments verify the accuracy and efficiency of the method.
{"title":"An Efficient Hybrid Domain Algorithm for Accurate SAR Raw Data Generation With Trajectory Deviations","authors":"Bo Wan;Lei Zhang;Jianxin Wu;Guanyong Wang;Zirui Xi","doi":"10.1109/LGRS.2025.3528935","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3528935","url":null,"abstract":"An efficient raw data generation (RDG) algorithm in the hybrid domain is proposed, which can be applied to the accurate echo generation of spotlight mode synthetic aperture radar (SAR) with trajectory deviation, even in cases of terrain undulation. Generally, the accuracy required for the signal’s envelope is on the order of the range resolution cell. However, the requirement for the phase is significantly higher, on the order of the wavelength. Therefore, the proposed algorithm calculates the phase of the SAR raw data in the time domain through a point-by-point approach to ensure accuracy and computes the envelope of the raw data through subblock processing in the frequency domain to improve efficiency. To balance the computational efficiency and accuracy, the optimal selection of subblock size is discussed in detail. Simulation experiments verify the accuracy and efficiency of the method.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"22 ","pages":"1-5"},"PeriodicalIF":0.0,"publicationDate":"2025-01-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143422804","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SAG-Net: Spectrum Adaptive Gate Network for Learning Feature Representation From Multispectral Imagery
Yong Li;Bohan Li;Zhongqun Chen;Yixuan Li;Guohan Zhang
Feature representation plays a key role in matching keypoints, especially for the multispectral images of large spectral difference. On such image pairs, existing methods typically use the two images only, but it is challenging to directly learn spectrum-invariant feature representation due to the complex nonlinear distortion between them. To address this issue, this letter proposes using intermediate-band images to facilitate learning spectrum-invariant feature representation. For this purpose, this work designs a spectrum adaptive gate network (SAG-Net) that consists of a SPectral gate (SPeG) module and a deep feature extractor. The SPeG module selectively activates the spectrum-invariant features according to input image content on-the-fly. It hence allows for training on the images of over two bands simultaneously with a single network without the need of an individual branch per band. To investigate the SPeG module, we also constructed a Landsat 9 Multi-Spectral Images (L9-MSI) dataset including 3167 scenes of aligned images across five spectral bands (visible, B5, B6, B7, and B10) from the Landsat 9 imagery. The experimental results demonstrate the SPeG module can learn common feature representation for varying-band images, and the intermediate B5, B6, and B7 images are useful for the SAG-Net to learn the common feature between visible and B10. On the L9-MSI dataset, the SAG-Net significantly improved the number of correct matches and the matching score (MS). Our dataset will be released at https://github.com/bohanlee/L9MSI-Dataset.git.
{"title":"SAG-Net: Spectrum Adaptive Gate Network for Learning Feature Representation From Multispectral Imagery","authors":"Yong Li;Bohan Li;Zhongqun Chen;Yixuan Li;Guohan Zhang","doi":"10.1109/LGRS.2025.3535635","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3535635","url":null,"abstract":"Feature representation plays a key role in matching keypoints, especially for the multispectral images of large spectral difference. On such image pairs, existing methods typically use the two images only, but it is challenging to directly learn spectrum-invariant feature representation due to the complex nonlinear distortion between them. To address this issue, this letter proposes using intermediate-band images to facilitate learning spectrum-invariant feature representation. For this purpose, this work designs a spectrum adaptive gate network (SAG-Net) that consists of a SPectral gate (SPeG) module and a deep feature extractor. The SPeG module selectively activates the spectrum-invariant features according to input image content on-the-fly. It hence allows for training on the images of over two bands simultaneously with a single network without the need of an individual branch per band. To investigate the SPeG module, we also constructed a Landsat 9 Multi-Spectral Images (L9-MSI) dataset including 3167 scenes of aligned images across five spectral bands (visible, B5, B6, B7, and B10) from the Landsat 9 imagery. The experimental results demonstrate the SPeG module can learn common feature representation for varying-band images, and the intermediate B5, B6, and B7 images are useful for the SAG-Net to learn the common feature between visible and B10. On the L9-MSI dataset, the SAG-Net significantly improved the number of correct matches and the matching score (MS). Our dataset will be released at <uri>https://github.com/bohanlee/L9MSI-Dataset.git</uri>.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"22 ","pages":"1-5"},"PeriodicalIF":0.0,"publicationDate":"2025-01-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143455229","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IcGAN4ColSAR: A Novel Multispectral Conditional Generative Adversarial Network Approach for SAR Image Colorization
Kangqing Shen;Gemine Vivone;Simone Lolli;Michael Schmitt;Xiaoyuan Yang;Jocelyn Chanussot
SAR colorization aims to enrich gray-scale SAR images with color while ensuring the preservation of original radiometric and spatial details. However, researchers often limit themselves to using only the red, green, and blue bands of a multispectral image as the source of color information, coupled with a single-polarization channel from the SAR image. This approach neglects the intrinsic characteristics of remote sensing data and thus fails to fully leverage available information. To overcome this limitation, this research attempts to explore inclusion of all available bands from multispectral images along with dual-polarization channels from SAR imagery in the colorization process. Furthermore, we present a new colorization method called improved conditional generative adversarial network for SAR colorization (IcGAN4ColSAR). This method tries to include the spectral angle mapper index within its loss function. Sufficient experiments show that our explorations in the number of data channels and the loss function are helpful in improving the colorization performance of the SAR image.
{"title":"IcGAN4ColSAR: A Novel Multispectral Conditional Generative Adversarial Network Approach for SAR Image Colorization","authors":"Kangqing Shen;Gemine Vivone;Simone Lolli;Michael Schmitt;Xiaoyuan Yang;Jocelyn Chanussot","doi":"10.1109/LGRS.2025.3534795","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3534795","url":null,"abstract":"SAR colorization aims to enrich gray-scale SAR images with color while ensuring the preservation of original radiometric and spatial details. However, researchers often limit themselves to using only the red, green, and blue bands of a multispectral image as the source of color information, coupled with a single-polarization channel from the SAR image. This approach neglects the intrinsic characteristics of remote sensing data and thus fails to fully leverage available information. To overcome this limitation, this research attempts to explore inclusion of all available bands from multispectral images along with dual-polarization channels from SAR imagery in the colorization process. Furthermore, we present a new colorization method called improved conditional generative adversarial network for SAR colorization (IcGAN4ColSAR). This method tries to include the spectral angle mapper index within its loss function. Sufficient experiments show that our explorations in the number of data channels and the loss function are helpful in improving the colorization performance of the SAR image.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"22 ","pages":"1-5"},"PeriodicalIF":0.0,"publicationDate":"2025-01-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143396318","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1