Pub Date : 2026-01-16DOI: 10.1109/JSTARS.2026.3654602
Ming Tong;Shenghua Fan;Jiu Jiang;Hezhi Sun;Jisan Yang;Chu He
Recently, detectors based on deep learning have boosted the state-of-the-art of application on ship detection in synthetic aperture radar (SAR) images. However, constructing discriminative feature from scattering of background and distinguishing contour of ship precisely still present challenging subject to the inherent scattering mechanism of SAR. In this article, a dual-branch detection framework with perception of scattering characteristic and geometric contour is introduced to deal with the problem. First, a scattering characteristic perception branch is proposed to fit the scattering distribution of SAR ship through conditional diffusion model, which introduces learnable scattering feature. Second, a convex contour perception branch is designed as two-stage coarse-to-fine pipeline to delimit the irregular boundary of ship by learning scattering key points. Finally, a cross-token integration module following Bayesian framework is introduced to couple features of scattering and texture adaptively to learn construction of discriminative feature. Furthermore, comprehensive experiments on three authoritative SAR datasets for oriented ship detection demonstrate the effectiveness of proposed method.
{"title":"Dual-Perception Detector for Ship Detection in SAR Images","authors":"Ming Tong;Shenghua Fan;Jiu Jiang;Hezhi Sun;Jisan Yang;Chu He","doi":"10.1109/JSTARS.2026.3654602","DOIUrl":"https://doi.org/10.1109/JSTARS.2026.3654602","url":null,"abstract":"Recently, detectors based on deep learning have boosted the state-of-the-art of application on ship detection in synthetic aperture radar (SAR) images. However, constructing discriminative feature from scattering of background and distinguishing contour of ship precisely still present challenging subject to the inherent scattering mechanism of SAR. In this article, a dual-branch detection framework with perception of scattering characteristic and geometric contour is introduced to deal with the problem. First, a scattering characteristic perception branch is proposed to fit the scattering distribution of SAR ship through conditional diffusion model, which introduces learnable scattering feature. Second, a convex contour perception branch is designed as two-stage coarse-to-fine pipeline to delimit the irregular boundary of ship by learning scattering key points. Finally, a cross-token integration module following Bayesian framework is introduced to couple features of scattering and texture adaptively to learn construction of discriminative feature. Furthermore, comprehensive experiments on three authoritative SAR datasets for oriented ship detection demonstrate the effectiveness of proposed method.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"4790-4808"},"PeriodicalIF":5.3,"publicationDate":"2026-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11355870","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146082030","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-14DOI: 10.1109/JSTARS.2026.3654241
Yuan Li;Tianzhu Zhang;Ziyi Xiong;Junying Lv;Yinning Pang
Detecting three-dimensional (3-D) windows is vital for creating semantic building models with high level of detail, furnishing smart city and digital twin programs. Existing studies on window extraction using street imagery or laser scanning data often rely on limited types of features, resulting in compromised accuracy and completeness due to shadows and geometric decorations caused by curtains, balconies, plants, and other objects. To enhance the effectiveness and robustness of building window extraction in 3-D, this article proposes an automatic method that leverages synergistic information from multiview-stereo (MVS) point clouds, through an adaptive divide-and-combine pipeline. Color information inherited from the imagery serves as a main clue to acquire the point clouds of individual building façades that may be coplanar and connected. The geometric information associated with normal vectors is then combined with color, to adaptively divide individual building façade into an irregular grid that conforms to the window edges. Subsequently, HSV color and depth distances within each grid cell are computed, and the grid cells are encoded to quantify the global arrangement features of windows. Finally, the multitype features are fused in an integer programming model, by solving which the optimal combination of grid cells corresponding to windows is obtained. Benefitting from the informative MVS point clouds and the fusion of multitype features, our method is able to directly produce 3-D models with high regularity for buildings with different appearances. Experimental results demonstrate that the proposed method is effective in 3-D window extraction while overcoming variations in façade appearances caused by foreign objects and missing data, with a high point-wise precision of 92.7%, recall of 77.09%, IoU of 71.95%, and F1-score of 83.42%. The results also exhibit a high level of integrity, with the accuracy of correctly extracted windows reaching 89.81%. In the future, we will focus on the development of a more universal façade dividing method to deal with even more complicated windows.
{"title":"Automated Extraction of 3-D Windows From MVS Point Clouds by Comprehensive Fusion of Multitype Features","authors":"Yuan Li;Tianzhu Zhang;Ziyi Xiong;Junying Lv;Yinning Pang","doi":"10.1109/JSTARS.2026.3654241","DOIUrl":"https://doi.org/10.1109/JSTARS.2026.3654241","url":null,"abstract":"Detecting three-dimensional (3-D) windows is vital for creating semantic building models with high level of detail, furnishing smart city and digital twin programs. Existing studies on window extraction using street imagery or laser scanning data often rely on limited types of features, resulting in compromised accuracy and completeness due to shadows and geometric decorations caused by curtains, balconies, plants, and other objects. To enhance the effectiveness and robustness of building window extraction in 3-D, this article proposes an automatic method that leverages synergistic information from multiview-stereo (MVS) point clouds, through an adaptive divide-and-combine pipeline. Color information inherited from the imagery serves as a main clue to acquire the point clouds of individual building façades that may be coplanar and connected. The geometric information associated with normal vectors is then combined with color, to adaptively divide individual building façade into an irregular grid that conforms to the window edges. Subsequently, HSV color and depth distances within each grid cell are computed, and the grid cells are encoded to quantify the global arrangement features of windows. Finally, the multitype features are fused in an integer programming model, by solving which the optimal combination of grid cells corresponding to windows is obtained. Benefitting from the informative MVS point clouds and the fusion of multitype features, our method is able to directly produce 3-D models with high regularity for buildings with different appearances. Experimental results demonstrate that the proposed method is effective in 3-D window extraction while overcoming variations in façade appearances caused by foreign objects and missing data, with a high point-wise precision of 92.7%, recall of 77.09%, IoU of 71.95%, and F1-score of 83.42%. The results also exhibit a high level of integrity, with the accuracy of correctly extracted windows reaching 89.81%. In the future, we will focus on the development of a more universal façade dividing method to deal with even more complicated windows.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"4918-4934"},"PeriodicalIF":5.3,"publicationDate":"2026-01-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11353237","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146082003","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-14DOI: 10.1109/JSTARS.2026.3654195
Daniel Carcereri;Luca Dell’Amore;Stefano Tebaldini;Paola Rizzoli
The increasing use of artificial intelligence (AI) models in Earth Observation (EO) applications, such as forest height estimation, has led to a growing need for explainable AI (XAI) methods. Despite their high accuracy, AI models are often criticized for their “black-box” nature, making it difficult to understand the inner decision-making process. In this study, we propose a multifaceted approach to XAI for a convolutional neural network (CNN)-based model that estimates forest height from TanDEM-X single-pass InSAR data. By combining domain knowledge, saliency maps, and feature importance analysis through exhaustive model permutations, we provide a comprehensive investigation of the network working principles. Our results suggests that the proposed model is implicitly capable of recognizing and compensating for the SAR acquisition geometry-related distortions. We find that the mean phase center height and its local variability represents the most informative predictor. We also find evidence that the interferometric coherence and the backscatter maps capture complementary but equally relevant views of the vegetation. This work contributes to advance the understanding of the model’s inner workings, and targets the development of more transparent and trustworthy AI for EO applications, ultimately leading to improved accuracy and reliability in the estimation of forest parameters.
{"title":"Insights on the Working Principles of a CNN for Forest Height Regression From Single-Pass InSAR Data","authors":"Daniel Carcereri;Luca Dell’Amore;Stefano Tebaldini;Paola Rizzoli","doi":"10.1109/JSTARS.2026.3654195","DOIUrl":"https://doi.org/10.1109/JSTARS.2026.3654195","url":null,"abstract":"The increasing use of artificial intelligence (AI) models in Earth Observation (EO) applications, such as forest height estimation, has led to a growing need for explainable AI (XAI) methods. Despite their high accuracy, AI models are often criticized for their “black-box” nature, making it difficult to understand the inner decision-making process. In this study, we propose a multifaceted approach to XAI for a convolutional neural network (CNN)-based model that estimates forest height from TanDEM-X single-pass InSAR data. By combining domain knowledge, saliency maps, and feature importance analysis through exhaustive model permutations, we provide a comprehensive investigation of the network working principles. Our results suggests that the proposed model is implicitly capable of recognizing and compensating for the SAR acquisition geometry-related distortions. We find that the mean phase center height and its local variability represents the most informative predictor. We also find evidence that the interferometric coherence and the backscatter maps capture complementary but equally relevant views of the vegetation. This work contributes to advance the understanding of the model’s inner workings, and targets the development of more transparent and trustworthy AI for EO applications, ultimately leading to improved accuracy and reliability in the estimation of forest parameters.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"4809-4824"},"PeriodicalIF":5.3,"publicationDate":"2026-01-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11352840","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146082019","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-14DOI: 10.1109/JSTARS.2026.3654017
Ali Al Bataineh;Bandi Vamsi;Scott Alan Smith
Accurate prediction of the water quality index is essential for protecting public health and managing freshwater resources. Existing models often rely on arbitrary weight initialization and make limited use of ensemble learning, which results in unstable performance and reduced interpretability. This study introduces a hybrid machine learning framework that combines feature-informed neural network initialization with gradient boosting (XGBoost) to address these limitations. Neural network weights are initialized using feature significance scores derived from SHapley Additive exPlanations (SHAP) and predictions are iteratively refined using XGBoost. The model was trained and evaluated using the public quality of freshwater dataset and compared against several baselines, including random forest, support vector regression, a conventional artificial neural network with Xavier initialization, and an XGBoost-only model. Our framework achieved an accuracy of 86.9%, an F1-score of 0.849, and a receiver operating characteristic–area under the curve of 0.894, outperforming all comparative methods. Ablation experiments showed that both the SHAP-based initialization and the boosting component each improved performance over simpler baselines.
{"title":"A Hybrid Machine Learning Framework for Water Quality Index Prediction Using Feature-Based Neural Network Initialization","authors":"Ali Al Bataineh;Bandi Vamsi;Scott Alan Smith","doi":"10.1109/JSTARS.2026.3654017","DOIUrl":"https://doi.org/10.1109/JSTARS.2026.3654017","url":null,"abstract":"Accurate prediction of the water quality index is essential for protecting public health and managing freshwater resources. Existing models often rely on arbitrary weight initialization and make limited use of ensemble learning, which results in unstable performance and reduced interpretability. This study introduces a hybrid machine learning framework that combines feature-informed neural network initialization with gradient boosting (XGBoost) to address these limitations. Neural network weights are initialized using feature significance scores derived from SHapley Additive exPlanations (SHAP) and predictions are iteratively refined using XGBoost. The model was trained and evaluated using the public quality of freshwater dataset and compared against several baselines, including random forest, support vector regression, a conventional artificial neural network with Xavier initialization, and an XGBoost-only model. Our framework achieved an accuracy of 86.9%, an <italic>F</i>1-score of 0.849, and a receiver operating characteristic–area under the curve of 0.894, outperforming all comparative methods. Ablation experiments showed that both the SHAP-based initialization and the boosting component each improved performance over simpler baselines.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"4887-4905"},"PeriodicalIF":5.3,"publicationDate":"2026-01-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11353250","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146082021","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-13DOI: 10.1109/JSTARS.2026.3653626
Xiaole Lin;Guangping Li;Jiahua Xie;Zhuokun Zhi
While convolutional neural network (CNN)-based methods for small object detection in remote sensing imagery have advanced considerably, substantial challenges remain unresolved, primarily stemming from complex backgrounds and insufficient feature representation. To address these issues, we propose a novel architecture specifically designed to accommodate the unique demands of small objects, termed AMFC-DEIM. This framework introduces three key innovations: first, the adaptive one-to-one (O2O) matching mechanism, which enhances dense O2O matching by adaptively adjusting the matching grid configuration to the object distribution, thereby preserving the resolution of small objects throughout training; second, the focal convolution module, engineered to explicitly align with the spatial characteristics of small objects for extracting fine-grained features; and third, the enhanced normalized Wasserstein distance, which stabilizes the training process and bolsters performance on small targets. Comprehensive experiments conducted on three benchmark remote sensing small object detection datasets: RSOD, LEVIR-SHIP and NWPU VHR-10, demonstrate that AMFC-DEIM achieves remarkable performance, attaining AP$_{50}$ scores of 96.2%, 86.2%, and 95.1%, respectively, while maintaining only 5.27 M parameters. These results substantially outperform several established benchmark models and state-of-the-art methods.
{"title":"AMFC-DEIM: Improved DEIM With Adaptive Matching and Focal Convolution for Remote Sensing Small Object Detection","authors":"Xiaole Lin;Guangping Li;Jiahua Xie;Zhuokun Zhi","doi":"10.1109/JSTARS.2026.3653626","DOIUrl":"https://doi.org/10.1109/JSTARS.2026.3653626","url":null,"abstract":"While convolutional neural network (CNN)-based methods for small object detection in remote sensing imagery have advanced considerably, substantial challenges remain unresolved, primarily stemming from complex backgrounds and insufficient feature representation. To address these issues, we propose a novel architecture specifically designed to accommodate the unique demands of small objects, termed AMFC-DEIM. This framework introduces three key innovations: first, the adaptive one-to-one (O2O) matching mechanism, which enhances dense O2O matching by adaptively adjusting the matching grid configuration to the object distribution, thereby preserving the resolution of small objects throughout training; second, the focal convolution module, engineered to explicitly align with the spatial characteristics of small objects for extracting fine-grained features; and third, the enhanced normalized Wasserstein distance, which stabilizes the training process and bolsters performance on small targets. Comprehensive experiments conducted on three benchmark remote sensing small object detection datasets: RSOD, LEVIR-SHIP and NWPU VHR-10, demonstrate that AMFC-DEIM achieves remarkable performance, attaining AP<inline-formula><tex-math>$_{50}$</tex-math></inline-formula> scores of 96.2%, 86.2%, and 95.1%, respectively, while maintaining only 5.27 M parameters. These results substantially outperform several established benchmark models and state-of-the-art methods.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"5021-5034"},"PeriodicalIF":5.3,"publicationDate":"2026-01-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11347584","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146082020","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-13DOI: 10.1109/JSTARS.2026.3653676
Jiapeng Huang;Yue Zhang;Xiaozhu Yang;Fan Mo
Forest canopy height is a critical structural parameter for accurately assessing forest carbon storage. This study integrates Global Ecosystem Dynamics Investigation (GEDI) LiDAR data with multisource remote sensing features to construct a multidimensional feature space comprising 13 parameters. By employing high-dimensional feature vectors of “spatial coordinates + environmental features,” the proposed deep learning-based neural network-guided interpolation (NNGI) model effectively harnesses the capacity of deep learning to model complex nonlinear relationships and adaptively extract local features. This method adopts a dual-network collaborative architecture to dynamically learn interpolation weights based on environmental similarity in the feature space, rather than relying on fixed parameters or merely considering spatial distance, thereby effectively fusing the complex nonlinear relationship modeling capability of deep learning with the concept of spatial interpolation. Experiments conducted across five representative regions in the United States demonstrate that the overall accuracy of the NNGI model significantly outperforms traditional machine learning methods, Pearson correlation coefffcient (r) = 0.79, root-mean-square error (RMSE) = 5.38 m, mean absolute error = 4.04 m, bias = –0.15 m. In areas with low (0% –20% ) and high (61% –80% ) vegetation cover fractions, the RMSE decreased by 37.52% and 5.37%, respectively, while the r-value increased by 15.87% and 35.90%, respectively. Regarding different slope aspects, the RMSE for southeastern and western slopes decreased by 30.38% and 18.70%, respectively. This study provides a more reliable solution for the accurate estimation of forest structural parameters in complex environments.
{"title":"A Deep Learning-Based Model for Forest Canopy Height Mapping Using Multisource Remote Sensing Data","authors":"Jiapeng Huang;Yue Zhang;Xiaozhu Yang;Fan Mo","doi":"10.1109/JSTARS.2026.3653676","DOIUrl":"https://doi.org/10.1109/JSTARS.2026.3653676","url":null,"abstract":"Forest canopy height is a critical structural parameter for accurately assessing forest carbon storage. This study integrates Global Ecosystem Dynamics Investigation (GEDI) LiDAR data with multisource remote sensing features to construct a multidimensional feature space comprising 13 parameters. By employing high-dimensional feature vectors of “spatial coordinates + environmental features,” the proposed deep learning-based neural network-guided interpolation (NNGI) model effectively harnesses the capacity of deep learning to model complex nonlinear relationships and adaptively extract local features. This method adopts a dual-network collaborative architecture to dynamically learn interpolation weights based on environmental similarity in the feature space, rather than relying on fixed parameters or merely considering spatial distance, thereby effectively fusing the complex nonlinear relationship modeling capability of deep learning with the concept of spatial interpolation. Experiments conducted across five representative regions in the United States demonstrate that the overall accuracy of the NNGI model significantly outperforms traditional machine learning methods, Pearson correlation coefffcient (<italic>r</i>) = 0.79, root-mean-square error (RMSE) = 5.38 m, mean absolute error = 4.04 m, bias = –0.15 m. In areas with low (0% –20% ) and high (61% –80% ) vegetation cover fractions, the RMSE decreased by 37.52% and 5.37%, respectively, while the <italic>r</i>-value increased by 15.87% and 35.90%, respectively. Regarding different slope aspects, the RMSE for southeastern and western slopes decreased by 30.38% and 18.70%, respectively. This study provides a more reliable solution for the accurate estimation of forest structural parameters in complex environments.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"4842-4857"},"PeriodicalIF":5.3,"publicationDate":"2026-01-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11348094","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146082031","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-12DOI: 10.1109/JSTARS.2026.3651900
Enyu Zhao;Yu Shi;Nianxin Qu;Yulei Wang;Hang Zhao
Infrared small target detection is focused on accurately identifying tiny targets with low signal-to-noise ratio against complex backgrounds, representing a critical challenge in the field of infrared image processing. Existing approaches frequently fail to retain small target information during global semantic extraction and struggle with preserving detailed features and achieving effective feature fusion. To address these limitations, this article proposes a morphology-edge enhanced triple-cascaded network (MEETNet) for infrared small target detection. The network employs a triple-cascaded architecture that maintains high resolution and enhances information interaction between different stages, facilitating effective multilevel feature fusion while safeguarding deep small-target characteristics. MEETNet integrates an edge-detail enhanced module (EDEM) and a detail-aware multi-scale fusion module (DMSFM). These modules introduce edge-detail enhanced features that amalgamate contrast and edge information, thereby amplifying target saliency and improving edge representation. Specifically, EDEM augments target contrast and edge structures by integrating edge-detail-enhanced features with shallow details. This integration improves the discriminability capacity of shallow features for detecting small targets. Moreover, DMSFM implements a multireceptive field mechanism to merge target details with deep semantic insights, enabling the capture of more distinctive global contextual features. Experimental evaluations conducted using two public datasets—NUAA-SIRST and NUDT-SIRST—demonstrate that the proposed MEETNet surpasses existing state-of-the-art methods for infrared small target detection in terms of detection accuracy.
{"title":"MEETNet: Morphology-Edge Enhanced Triple-Cascaded Network for Infrared Small Target Detection","authors":"Enyu Zhao;Yu Shi;Nianxin Qu;Yulei Wang;Hang Zhao","doi":"10.1109/JSTARS.2026.3651900","DOIUrl":"https://doi.org/10.1109/JSTARS.2026.3651900","url":null,"abstract":"Infrared small target detection is focused on accurately identifying tiny targets with low signal-to-noise ratio against complex backgrounds, representing a critical challenge in the field of infrared image processing. Existing approaches frequently fail to retain small target information during global semantic extraction and struggle with preserving detailed features and achieving effective feature fusion. To address these limitations, this article proposes a morphology-edge enhanced triple-cascaded network (MEETNet) for infrared small target detection. The network employs a triple-cascaded architecture that maintains high resolution and enhances information interaction between different stages, facilitating effective multilevel feature fusion while safeguarding deep small-target characteristics. MEETNet integrates an edge-detail enhanced module (EDEM) and a detail-aware multi-scale fusion module (DMSFM). These modules introduce edge-detail enhanced features that amalgamate contrast and edge information, thereby amplifying target saliency and improving edge representation. Specifically, EDEM augments target contrast and edge structures by integrating edge-detail-enhanced features with shallow details. This integration improves the discriminability capacity of shallow features for detecting small targets. Moreover, DMSFM implements a multireceptive field mechanism to merge target details with deep semantic insights, enabling the capture of more distinctive global contextual features. Experimental evaluations conducted using two public datasets—NUAA-SIRST and NUDT-SIRST—demonstrate that the proposed MEETNet surpasses existing state-of-the-art methods for infrared small target detection in terms of detection accuracy.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"4748-4765"},"PeriodicalIF":5.3,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11340625","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146082064","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Deep forest-based models for synthetic aperture radar (SAR) image change detection are generally challenged by noise sensitivity and high feature redundancy, which significantly degrade the prediction performance. To address these issues, this article proposes a structure-constrained and feature-screened deep forest, abbreviated as SC-FS-DF, for SAR image change detection. In preclassification, a fuzzy multineighborhood information C-means clustering is proposed to generate high-quality pseudo-labels. It introduces the edge information, the nonlocal and intrasuperpixel neighborhoods into the objective function of fuzzy local information C-means, thus suppressing the speckle noise and constraining structures of targets. In the sample learning and label prediction module, a feature-screened deep forest (FS-DF) framework is constructed by combining feature importance and redundancy analysis with a dropout strategy, thus screening out the noninformative features and meanwhile retaining the informative ones for learning at each cascade layer. Finally, a novel energy function fusing the nonlocal and superpixel information is derived for refining the detection map generated by FS-DF, further preserving fine details and edge locations. Extensive comparison and ablation experiments on five real SAR datasets verify the effectiveness and robustness of the proposed SC-FS-DF, and demonstrate that the SC-FS-DF can well screen the high-dimensional features in change detection and constrain the structures of targets.
{"title":"Feature-Screened and Structure-Constrained Deep Forest for Unsupervised SAR Image Change Detection","authors":"Wanying Song;Ruijing Zhu;Jie Wang;Yinyin Jiang;Yan Wu","doi":"10.1109/JSTARS.2026.3651534","DOIUrl":"https://doi.org/10.1109/JSTARS.2026.3651534","url":null,"abstract":"Deep forest-based models for synthetic aperture radar (SAR) image change detection are generally challenged by noise sensitivity and high feature redundancy, which significantly degrade the prediction performance. To address these issues, this article proposes a structure-constrained and feature-screened deep forest, abbreviated as SC-FS-DF, for SAR image change detection. In preclassification, a fuzzy multineighborhood information C-means clustering is proposed to generate high-quality pseudo-labels. It introduces the edge information, the nonlocal and intrasuperpixel neighborhoods into the objective function of fuzzy local information C-means, thus suppressing the speckle noise and constraining structures of targets. In the sample learning and label prediction module, a feature-screened deep forest (FS-DF) framework is constructed by combining feature importance and redundancy analysis with a dropout strategy, thus screening out the noninformative features and meanwhile retaining the informative ones for learning at each cascade layer. Finally, a novel energy function fusing the nonlocal and superpixel information is derived for refining the detection map generated by FS-DF, further preserving fine details and edge locations. Extensive comparison and ablation experiments on five real SAR datasets verify the effectiveness and robustness of the proposed SC-FS-DF, and demonstrate that the SC-FS-DF can well screen the high-dimensional features in change detection and constrain the structures of targets.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"4056-4068"},"PeriodicalIF":5.3,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11339914","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146026482","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-12DOI: 10.1109/JSTARS.2026.3651075
Yu Yao;Hengbin Wang;Xiang Gao;Ziyao Xing;Xiaodong Zhang;Yuanyuan Zhao;Shaoming Li;Zhe Liu
High-resolution remote sensing images provide crucial data support for applications such as precision agriculture and water resource management. However, super-resolution reconstructions often suffer from over-smoothed textures and structural distortions, failing to accurately recover the intricate details of ground objects. To address this issue, this article proposes a remote sensing image super-resolution network (DTWSTSR) that combines the Dual-Tree Complex Wavelet Transform and Swin Transformer, which enhances the ability of texture detail reconstruction by fusing frequency-domain and spatial-domain features. This model includes a Dual-Tree Complex Wavelet Texture Feature Sensing Module (DWTFSM) for integrating frequency and spatial features, and a Multiscale Efficient Channel Attention mechanism to enhance attention to multiscale and global details. In addition, we design a Kolmogorov–Arnold Network based on a branch attention mechanism, which improves the model’s ability to represent complex nonlinear features. During the training process, we investigate the impact of hyperparameters and propose the two-stage SSIM&SL1 loss function to reduce structural differences between images. Experimental results show that DTWSTSR outperforms existing mainstream methods under different magnification factors (×2, ×3, ×4), ranking among the top two in multiple metrics. For example, at ×2 magnification, its PSNR value is 0.64–2.68 dB higher than that of other models. Visual comparisons demonstrate that the proposed model achieves clearer and more accurate detail reconstruction of target ground objects. Furthermore, the model exhibits excellent generalization ability in cross-sensor image (OLI2MSI dataset) reconstruction.
{"title":"DTWSTSR: Dual-Tree Complex Wavelet and Swin Transformer Based Remote Sensing Images Super-Resolution Network","authors":"Yu Yao;Hengbin Wang;Xiang Gao;Ziyao Xing;Xiaodong Zhang;Yuanyuan Zhao;Shaoming Li;Zhe Liu","doi":"10.1109/JSTARS.2026.3651075","DOIUrl":"https://doi.org/10.1109/JSTARS.2026.3651075","url":null,"abstract":"High-resolution remote sensing images provide crucial data support for applications such as precision agriculture and water resource management. However, super-resolution reconstructions often suffer from over-smoothed textures and structural distortions, failing to accurately recover the intricate details of ground objects. To address this issue, this article proposes a remote sensing image super-resolution network (DTWSTSR) that combines the Dual-Tree Complex Wavelet Transform and Swin Transformer, which enhances the ability of texture detail reconstruction by fusing frequency-domain and spatial-domain features. This model includes a Dual-Tree Complex Wavelet Texture Feature Sensing Module (DWTFSM) for integrating frequency and spatial features, and a Multiscale Efficient Channel Attention mechanism to enhance attention to multiscale and global details. In addition, we design a Kolmogorov–Arnold Network based on a branch attention mechanism, which improves the model’s ability to represent complex nonlinear features. During the training process, we investigate the impact of hyperparameters and propose the two-stage SSIM&SL1 loss function to reduce structural differences between images. Experimental results show that DTWSTSR outperforms existing mainstream methods under different magnification factors (×2, ×3, ×4), ranking among the top two in multiple metrics. For example, at ×2 magnification, its PSNR value is 0.64–2.68 dB higher than that of other models. Visual comparisons demonstrate that the proposed model achieves clearer and more accurate detail reconstruction of target ground objects. Furthermore, the model exhibits excellent generalization ability in cross-sensor image (OLI2MSI dataset) reconstruction.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"4730-4747"},"PeriodicalIF":5.3,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11329193","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146082033","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-12DOI: 10.1109/JSTARS.2026.3651639
Yang Liu;Kun Zhang;Chun-Yi Song;Zhi-Wei Xu
In high-resolution maritime radar working in scanning mode, the classification and identification of ships require the recovery of the ship’s high-resolution range profiles (HRRPs) from radar returns. The return signal from the ship is a complex sparse signal interfered by non-Gaussian sea clutter. In this article, three sparse optimization methods matching the non-Gaussian characteristics of sea clutter, i.e., the sparse optimization matching K-distribution method, the sparse optimization matching generalized Pareto distribution method, the sparse optimization matching CGIG distribution method, are proposed to estimate complex HRRPs of ships. The compound Gaussian model is used to describe the non-Gaussianity of sea clutter, and the sparsity of ships’ complex HRRPs is constrained by the random distribution with one parameter. In the three methods, the Anderson–Darling test is used to search the parameters of the sparse constraint model. Besides, the non-Gaussian characteristics of sea clutter depend on the marine environment parameters and radar operating parameters. For different scenarios, the minimal criterion of the Kolmogorov–Smirnov distance is used to select the best model from the three compound Gaussian models, and then select the corresponding proposed methods. Simulated and measured radar data are used to evaluate the performance of the proposed methods and the results show that the proposed methods obtain better estimates of ship HRRPs compared to the recent SRIM method and the classical SLIM method.
{"title":"Estimation of Ships’ Complex High-Resolution Range Profiles Based on Sparse Optimization Method in Non-Gaussian Sea Clutter","authors":"Yang Liu;Kun Zhang;Chun-Yi Song;Zhi-Wei Xu","doi":"10.1109/JSTARS.2026.3651639","DOIUrl":"https://doi.org/10.1109/JSTARS.2026.3651639","url":null,"abstract":"In high-resolution maritime radar working in scanning mode, the classification and identification of ships require the recovery of the ship’s high-resolution range profiles (HRRPs) from radar returns. The return signal from the ship is a complex sparse signal interfered by non-Gaussian sea clutter. In this article, three sparse optimization methods matching the non-Gaussian characteristics of sea clutter, i.e., the sparse optimization matching K-distribution method, the sparse optimization matching generalized Pareto distribution method, the sparse optimization matching CGIG distribution method, are proposed to estimate complex HRRPs of ships. The compound Gaussian model is used to describe the non-Gaussianity of sea clutter, and the sparsity of ships’ complex HRRPs is constrained by the random distribution with one parameter. In the three methods, the Anderson–Darling test is used to search the parameters of the sparse constraint model. Besides, the non-Gaussian characteristics of sea clutter depend on the marine environment parameters and radar operating parameters. For different scenarios, the minimal criterion of the Kolmogorov–Smirnov distance is used to select the best model from the three compound Gaussian models, and then select the corresponding proposed methods. Simulated and measured radar data are used to evaluate the performance of the proposed methods and the results show that the proposed methods obtain better estimates of ship HRRPs compared to the recent SRIM method and the classical SLIM method.","PeriodicalId":13116,"journal":{"name":"IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing","volume":"19 ","pages":"3998-4013"},"PeriodicalIF":5.3,"publicationDate":"2026-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11339885","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146026530","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}