首页 > 最新文献

ISPRS Journal of Photogrammetry and Remote Sensing最新文献

英文 中文
Explainable urban renewal prediction at building-scale using hierarchical graph neural networks 基于层次图神经网络的建筑尺度可解释的城市更新预测
IF 12.2 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2026-03-01 Epub Date: 2026-02-11 DOI: 10.1016/j.isprsjprs.2026.02.013
Xiaoqin Yan , Zhou Huang , Shuliang Ren , Qia Zhu , Ganmin Yin , Junnan Qi , Yi Bao
Urban renewal prediction is critical for sustainable development, yet existing methods often overlook complex spatial dependencies and lack explainability. This study proposes an explainable hierarchical graph network for urban renewal (URHGN) prediction. It employs a hierarchical graph to model building and community level spatial interactions with GNNExplainer to predict renewal potential and quantify driving factors. Applied to Beijing, URHGN achieves an F1-score of 0.855 ± 0.012, outperforming traditional machine learning (0.700–0.775) and single-layer graph methods (0.789–0.810). Explainability analysis reveals that building-level features like floor (importance: 0.485 ± 0.030) are primary drivers while community-level context like house price (0.343 ± 0.003) provides essential supplements. Spatial relationships prove more influential than node features, with contribution scores of 0.271 ± 0.019 and 0.124 ± 0.017 respectively. The model identifies 15,990 buildings with very high renewal potential (scores > 0.8), advancing explainable GeoAI (XGeoAI) methodologies for evidence-based urban planning and establishing a foundation for future dynamic models incorporating temporal changes. The code is available at: https://github.com/kkxiaoqin/URHGN.
城市更新预测对可持续发展至关重要,但现有方法往往忽略了复杂的空间依赖关系,缺乏可解释性。本文提出了一种可解释的层次图网络用于城市更新预测。它采用层次图来模拟建筑和社区层面的空间相互作用,以预测更新潜力和量化驱动因素。应用于北京,URHGN的f1得分为0.855±0.012,优于传统的机器学习方法(0.700-0.775)和单层图方法(0.789-0.810)。可解释性分析表明,楼层等建筑层面特征(重要性:0.485±0.030)是主要驱动因素,而房价等社区层面背景(重要性:0.343±0.003)是必不可少的补充因素。空间关系比节点特征的影响更大,贡献值分别为0.271±0.019和0.124±0.017。该模型确定了15,990座具有很高更新潜力的建筑(得分>; 0.8),为基于证据的城市规划推进了可解释的GeoAI (XGeoAI)方法,并为纳入时间变化的未来动态模型奠定了基础。代码可从https://github.com/kkxiaoqin/URHGN获得。
{"title":"Explainable urban renewal prediction at building-scale using hierarchical graph neural networks","authors":"Xiaoqin Yan ,&nbsp;Zhou Huang ,&nbsp;Shuliang Ren ,&nbsp;Qia Zhu ,&nbsp;Ganmin Yin ,&nbsp;Junnan Qi ,&nbsp;Yi Bao","doi":"10.1016/j.isprsjprs.2026.02.013","DOIUrl":"10.1016/j.isprsjprs.2026.02.013","url":null,"abstract":"<div><div>Urban renewal prediction is critical for sustainable development, yet existing methods often overlook complex spatial dependencies and lack explainability. This study proposes an explainable hierarchical graph network for urban renewal (URHGN) prediction. It employs a hierarchical graph to model building and community level spatial interactions with GNNExplainer to predict renewal potential and quantify driving factors. Applied to Beijing, URHGN achieves an F1-score of 0.855 ± 0.012, outperforming traditional machine learning (0.700–0.775) and single-layer graph methods (0.789–0.810). Explainability analysis reveals that building-level features like floor (importance: 0.485 ± 0.030) are primary drivers while community-level context like house price (0.343 ± 0.003) provides essential supplements. Spatial relationships prove more influential than node features, with contribution scores of 0.271 ± 0.019 and 0.124 ± 0.017 respectively. The model identifies 15,990 buildings with very high renewal potential (scores &gt; 0.8), advancing explainable GeoAI (XGeoAI) methodologies for evidence-based urban planning and establishing a foundation for future dynamic models incorporating temporal changes. The code is available at: https://github.com/kkxiaoqin/URHGN.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"233 ","pages":"Pages 609-622"},"PeriodicalIF":12.2,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146153116","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An advanced decoupled polarimetric calibration method for the LuTan-1 hybrid- and quadrature-polarimetric modes LuTan-1混合偏振模式和正交偏振模式的一种高级解耦偏振定标方法
IF 12.2 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2026-03-01 Epub Date: 2026-01-30 DOI: 10.1016/j.isprsjprs.2026.01.035
Lizhi Liu, Lijie Huang, Yiding Wang, Pingping Lu, Bo Li, Liang Li, Robert Wang, Yirong Wu
During solar maximum, low-frequency spaceborne Polarimetric Synthetic Aperture Radar (PolSAR) systems suffer ionosphere-induced distortions that couple with system-induced polarimetric distortions. High-precision decoupled polarimetric calibration is therefore essential for obtaining high-fidelity PolSAR data. Existing point-target calibration methods lack a general approach for unbiased estimation of polarimetric distortion across multiple polarimetric modes and calibrator combinations, particularly under spatiotemporally varying ionospheric conditions. To address this, we derive the necessary conditions for unbiased estimation and propose a General Polarimetric Calibration Method (GPCM) applicable to various configurations. In addition, Enhanced Multi-Look Autofocus (EMLA), a modified STEC inversion method, is introduced for precise inversion of Slant Total Electron Content (STEC), enabling estimation of the spatiotemporally varying Faraday rotation angle for system distortion decoupling and PolSAR data compensation. GPCM applied to LuTan-1 HP and QP data results in HH/VV amplitude and phase imbalances of 0.0433  dB (STD: 0.017) and − 0.60° (STD: 1.02°), respectively, measured on trihedral corner reflectors. Calibration results also indicate that QP mode isolation exceeds 39 dB, while estimated axial ratios for HP mode are lower than 0.115 dB. Under comparable conditions, the results of GPCM are consistent with the Freeman analytical method. Furthermore, EMLA outperforms existing STEC inversion methods (COA, MLA, and GIM-based mapping), achieving a mean absolute difference of 1.95 TECU compared with in-situ measurements while demonstrating applicability to general scenes. Overall, the effectiveness of GPCM and EMLA in the LuTan-1 calibration mission is confirmed, indicating their potential for future PolSAR calibration tasks. The primary calibrated experimental dataset is publicly available at https://radars.ac.cn/web/data/getData?dataType=HPSAREADEN&pageType=en.
在太阳活动极大期,低频星载极化合成孔径雷达(PolSAR)系统遭受电离层诱导的畸变,这种畸变与系统诱导的极化畸变耦合。因此,高精度解耦极化校准对于获得高保真的PolSAR数据至关重要。现有的点目标校准方法缺乏一种通用的方法来无偏估计跨多个极化模式和校准器组合的极化失真,特别是在时空变化的电离层条件下。为了解决这个问题,我们推导了无偏估计的必要条件,并提出了一种适用于各种配置的通用偏振校准方法(GPCM)。此外,提出了一种改进的STEC反演方法——Enhanced Multi-Look Autofocus (EMLA),用于精确反演倾斜总电子含量(STEC),从而估算出法拉第旋转角的时空变化,从而实现系统畸变解耦和PolSAR数据补偿。采用GPCM对鲁坦1号的HP和QP数据进行处理,在三面角反射镜上测得的HH/VV振幅和相位不平衡分别为0.0433 dB (STD: 0.017)和- 0.60°(STD: 1.02°)。校准结果还表明,QP模式隔离度超过39 dB,而HP模式的估计轴向比低于0.115 dB。在可比条件下,GPCM的计算结果与Freeman分析方法一致。此外,EMLA优于现有的STEC反演方法(基于COA、MLA和基于gimm的制图),与原位测量相比,平均绝对差为1.95 TECU,同时证明了对一般场景的适用性。总体而言,GPCM和EMLA在LuTan-1校准任务中的有效性得到了证实,表明它们在未来PolSAR校准任务中的潜力。主要校准的实验数据集可在https://radars.ac.cn/web/data/getData?dataType=HPSAREADEN&pageType=en上公开获取。
{"title":"An advanced decoupled polarimetric calibration method for the LuTan-1 hybrid- and quadrature-polarimetric modes","authors":"Lizhi Liu,&nbsp;Lijie Huang,&nbsp;Yiding Wang,&nbsp;Pingping Lu,&nbsp;Bo Li,&nbsp;Liang Li,&nbsp;Robert Wang,&nbsp;Yirong Wu","doi":"10.1016/j.isprsjprs.2026.01.035","DOIUrl":"10.1016/j.isprsjprs.2026.01.035","url":null,"abstract":"<div><div>During solar maximum, low-frequency spaceborne Polarimetric Synthetic Aperture Radar (PolSAR) systems suffer ionosphere-induced distortions that couple with system-induced polarimetric distortions. High-precision decoupled polarimetric calibration is therefore essential for obtaining high-fidelity PolSAR data. Existing point-target calibration methods lack a general approach for unbiased estimation of polarimetric distortion across multiple polarimetric modes and calibrator combinations, particularly under spatiotemporally varying ionospheric conditions. To address this, we derive the necessary conditions for unbiased estimation and propose a General Polarimetric Calibration Method (GPCM) applicable to various configurations. In addition, Enhanced Multi-Look Autofocus (EMLA), a modified STEC inversion method, is introduced for precise inversion of Slant Total Electron Content (STEC), enabling estimation of the spatiotemporally varying Faraday rotation angle for system distortion decoupling and PolSAR data compensation. GPCM applied to LuTan-1 HP and QP data results in HH/VV amplitude and phase imbalances of 0.0433  dB (STD: 0.017) and − 0.60° (STD: 1.02°), respectively, measured on trihedral corner reflectors. Calibration results also indicate that QP mode isolation exceeds 39 dB, while estimated axial ratios for HP mode are lower than 0.115 dB. Under comparable conditions, the results of GPCM are consistent with the Freeman analytical method. Furthermore, EMLA outperforms existing STEC inversion methods (COA, MLA, and GIM-based mapping), achieving a mean absolute difference of 1.95 TECU compared with in-situ measurements while demonstrating applicability to general scenes. Overall, the effectiveness of GPCM and EMLA in the LuTan-1 calibration mission is confirmed, indicating their potential for future PolSAR calibration tasks. The primary calibrated experimental dataset is publicly available at <span><span>https://radars.ac.cn/web/data/getData?dataType=HPSAREADEN&amp;pageType=en</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"233 ","pages":"Pages 310-327"},"PeriodicalIF":12.2,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146079953","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PANet: A multi-scale temporal decoupling network and its high-resolution benchmark dataset for detecting pseudo changes in cropland non-agriculturalization PANet:基于多尺度时间解耦网络及其高分辨率基准数据集的耕地非农化伪变化检测
IF 12.2 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2026-03-01 Epub Date: 2026-01-22 DOI: 10.1016/j.isprsjprs.2026.01.029
Songman Sui , Jixian Zhang , Haiyan Gu , Yue Chang
Cropland non-agriculturalization (CNA) refers to the conversion of cropland into non-agricultural land such as construction land or ponds, posing threats to food security and ecological balance. Remote sensing technology enables precise monitoring of this process, but bi-temporal methods are susceptible to errors caused by seasonal spectral fluctuations, weather interference, and imaging discrepancies, often leading to false detections. Existing methods, which lack support from temporal datasets, struggle to disentangle the spectral confusion of gradual non-agriculturalization and short-term disturbances, thereby limiting the accuracy of dynamic cropland resource monitoring. To address this issue, a novel phenology-aware temporal change detection network (PANet) is proposed to solve the misclassification challenges in CNA detection caused by “same object with different spectra” and “different objects with similar spectra” issues. A phenology-aware module (PATM) is designed, leveraging a dual-driven decoupling model to dynamically weight phenology-sensitive periods and adaptively represent non-uniform temporal intervals. Through a time-aligned feature enhancement strategy and dual-driven (intra-annual/inter-annual) temporal decay functions, PANet simultaneously focuses on short-term anomalies and robustly models long-term trends. Additionally, a sample balance adjustment module (DFBL) is developed to mitigate the impact of sample imbalance by incorporating prior knowledge of changes and dynamic adjustment factors, enhancing the model’s sensitivity to non-agriculturalization changes. Furthermore, the first high-resolution CNA dataset based on actual production data is constructed, containing 1295 pairs of 512 × 512 masked images. Compared to existing datasets, this dataset offers extensive temporal coverage, capturing comprehensive seasonal periodic characteristics of cropland. Comparative experiments with several classical time-series methods and bi-temporal methods validate the effectiveness of PANet. Experimental results on the LHCD dataset demonstrate that PANet achieves the highest F1 score, specifically, 61.01% and 61.70%. PANet accurately captures CNA information, making it vital for the scientific management and sustainable utilization of limited cropland resources. The LHCD can be downloaded from https://github.com/mss-s/LHCD.
耕地非农化是指将耕地转化为建设用地或池塘等非农业用地,对粮食安全和生态平衡构成威胁。遥感技术能够对这一过程进行精确监测,但双时间方法容易受到季节光谱波动、天气干扰和成像差异造成的误差的影响,往往导致错误的检测。现有的方法缺乏时间数据集的支持,难以理清逐渐非农业化和短期扰动的光谱混淆,从而限制了动态耕地资源监测的准确性。针对这一问题,提出了一种新的物候感知时间变化检测网络(PANet),以解决CNA检测中由于“同一物体具有不同光谱”和“不同物体具有相似光谱”造成的误分类问题。设计了物候感知模块(PATM),利用双驱动解耦模型动态加权物候敏感期,并自适应表示非均匀时间间隔。通过时间同步特征增强策略和双驱动(年内/年际)时间衰减函数,PANet同时关注短期异常并稳健地模拟长期趋势。此外,通过引入变化的先验知识和动态调整因子,构建了样本平衡调整模块(DFBL)来缓解样本失衡的影响,增强了模型对非农业变化的敏感性。构建了第一个基于实际生产数据的高分辨率CNA数据集,包含1295对512 × 512的掩膜图像。与现有数据集相比,该数据集提供了广泛的时间覆盖,捕获了农田的全面季节性周期性特征。与几种经典时间序列方法和双时间方法的对比实验验证了PANet的有效性。在LHCD数据集上的实验结果表明,PANet的F1得分最高,分别为61.01%和61.70%。PANet准确捕获CNA信息,对有限耕地资源的科学管理和可持续利用至关重要。LHCD可从https://github.com/mss-s/LHCD下载。
{"title":"PANet: A multi-scale temporal decoupling network and its high-resolution benchmark dataset for detecting pseudo changes in cropland non-agriculturalization","authors":"Songman Sui ,&nbsp;Jixian Zhang ,&nbsp;Haiyan Gu ,&nbsp;Yue Chang","doi":"10.1016/j.isprsjprs.2026.01.029","DOIUrl":"10.1016/j.isprsjprs.2026.01.029","url":null,"abstract":"<div><div>Cropland non-agriculturalization (CNA) refers to the conversion of cropland into non-agricultural land such as construction land or ponds, posing threats to food security and ecological balance. Remote sensing technology enables precise monitoring of this process, but bi-temporal methods are susceptible to errors caused by seasonal spectral fluctuations, weather interference, and imaging discrepancies, often leading to false detections. Existing methods, which lack support from temporal datasets, struggle to disentangle the spectral confusion of gradual non-agriculturalization and short-term disturbances, thereby limiting the accuracy of dynamic cropland resource monitoring. To address this issue, a novel phenology-aware temporal change detection network (PANet) is proposed to solve the misclassification challenges in CNA detection caused by “same object with different spectra” and “different objects with similar spectra” issues. A phenology-aware module (PATM) is designed, leveraging a dual-driven decoupling model to dynamically weight phenology-sensitive periods and adaptively represent non-uniform temporal intervals. Through a time-aligned feature enhancement strategy and dual-driven (intra-annual/inter-annual) temporal decay functions, PANet simultaneously focuses on short-term anomalies and robustly models long-term trends. Additionally, a sample balance adjustment module (DFBL) is developed to mitigate the impact of sample imbalance by incorporating prior knowledge of changes and dynamic adjustment factors, enhancing the model’s sensitivity to non-agriculturalization changes. Furthermore, the first high-resolution CNA dataset based on actual production data is constructed, containing 1295 pairs of 512 × 512 masked images. Compared to existing datasets, this dataset offers extensive temporal coverage, capturing comprehensive seasonal periodic characteristics of cropland. Comparative experiments with several classical time-series methods and bi-temporal methods validate the effectiveness of PANet. Experimental results on the LHCD dataset demonstrate that PANet achieves the highest F1 score, specifically, 61.01% and 61.70%. PANet accurately captures CNA information, making it vital for the scientific management and sustainable utilization of limited cropland resources. The LHCD can be downloaded from <span><span>https://github.com/mss-s/LHCD</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"233 ","pages":"Pages 126-143"},"PeriodicalIF":12.2,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146033659","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Weak supervision makes strong details: fine-grained object recognition in remote sensing images via regional diffusion with VLM 弱监督生成强细节:利用VLM进行区域扩散的遥感图像细粒度目标识别
IF 12.2 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2026-03-01 Epub Date: 2026-01-23 DOI: 10.1016/j.isprsjprs.2026.01.024
Liuqian Wang, Jing Zhang, Guangming Mi, Li Zhuo
Fine-grained object recognition (FGOR) is gaining increasing attention in automated remote sensing analysis and interpretation (RSAI). However, the full potential of FGOR in remote sensing images (RSIs) is still constrained by several key issues: the reliance on high-quality labeled data, the difficulty of reconstructing fine details in low-resolution images, and the limited robustness of FGOR model for distinguishing similar object categories. In response, we propose an automatic fine-grained object recognition network (AutoFGOR) that follows a hierarchical dual-pipeline architecture for object analysis at global and regional levels. Specifically, Pipeline I: region detection network, which leverages geometric invariance module for weakly-supervised learning to improve the detection accuracy of sparsely labeled RSIs and extract category-free regions; and on top of that, Pipeline II: regional diffusion with vision language model (RD-VLM), which pioneers the combination of stable diffusion XL (SDXL) and large language and vision assistant (LLaVA) through a specially designed adaptive resolution adaptor (ARA) for object region super-resolution reconstruction, fundamentally solving the difficulties of feature extraction from low-quality regions and fine-grained feature mining. In addition, we introduce a winner-takes-all (WTA) strategy that utilizes a voting mechanism to enhance the reliability of fine-grained classification in complex scenes. Experimental results on FAIR1M-v2.0, VEDAI, and HRSC2016 datasets demonstrate our AutoFGOR achieving 31.72%, 80.25%, and 88.05% mAP, respectively, with highly competitive performance. In addition, the × 4 reconstruction results achieve scores of 0.5275 and 0.8173 on the MANIQA and CLIP-IQA indicators, respectively. The code will be available on GitHub: https://github.com/BJUT-AIVBD/AutoFGOR.
细粒度目标识别(FGOR)在自动遥感分析与解释(RSAI)中越来越受到关注。然而,FGOR在遥感图像(rsi)中的全部潜力仍然受到几个关键问题的限制:对高质量标记数据的依赖,在低分辨率图像中重建精细细节的困难,以及FGOR模型在区分相似目标类别方面的有限鲁棒性。作为回应,我们提出了一种自动细粒度目标识别网络(AutoFGOR),该网络遵循分层双管道架构,用于全球和区域层面的目标分析。其中,管道1:区域检测网络,利用几何不变性模块进行弱监督学习,提高稀疏标记rsi的检测精度,提取无类别区域;在此基础上,Pipeline II:区域扩散与视觉语言模型(RD-VLM),通过专门设计的自适应分辨率适配器(ARA),率先将稳定扩散XL (SDXL)与大型语言视觉助手(LLaVA)相结合,进行目标区域超分辨率重建,从根本上解决了低质量区域特征提取和细粒度特征挖掘的难题。此外,我们引入了赢家通吃(WTA)策略,该策略利用投票机制来增强复杂场景中细粒度分类的可靠性。在FAIR1M-v2.0、VEDAI和HRSC2016数据集上的实验结果表明,我们的AutoFGOR分别实现了31.72%、80.25%和88.05%的mAP,具有很强的竞争力。此外,× 4重建结果在MANIQA和CLIP-IQA指标上分别达到0.5275和0.8173分。代码将在GitHub上提供:https://github.com/BJUT-AIVBD/AutoFGOR。
{"title":"Weak supervision makes strong details: fine-grained object recognition in remote sensing images via regional diffusion with VLM","authors":"Liuqian Wang,&nbsp;Jing Zhang,&nbsp;Guangming Mi,&nbsp;Li Zhuo","doi":"10.1016/j.isprsjprs.2026.01.024","DOIUrl":"10.1016/j.isprsjprs.2026.01.024","url":null,"abstract":"<div><div>Fine-grained object recognition (FGOR) is gaining increasing attention in automated remote sensing analysis and interpretation (RSAI). However, the full potential of FGOR in remote sensing images (RSIs) is still constrained by several key issues: the reliance on high-quality labeled data, the difficulty of reconstructing fine details in low-resolution images, and the limited robustness of FGOR model for distinguishing similar object categories. In response, we propose an automatic fine-grained object recognition network (AutoFGOR) that follows a hierarchical dual-pipeline architecture for object analysis at global and regional levels. Specifically, Pipeline I: region detection network, which leverages geometric invariance module for weakly-supervised learning to improve the detection accuracy of sparsely labeled RSIs and extract category-free regions; and on top of that, Pipeline II: regional diffusion with vision language model (RD-VLM), which pioneers the combination of stable diffusion XL (SDXL) and large language and vision assistant (LLaVA) through a specially designed adaptive resolution adaptor (ARA) for object region super-resolution reconstruction, fundamentally solving the difficulties of feature extraction from low-quality regions and fine-grained feature mining. In addition, we introduce a winner-takes-all (WTA) strategy that utilizes a voting mechanism to enhance the reliability of fine-grained classification in complex scenes. Experimental results on FAIR1M-v2.0, VEDAI, and HRSC2016 datasets demonstrate our AutoFGOR achieving 31.72%, 80.25%, and 88.05% mAP, respectively, with highly competitive performance. In addition, the × 4 reconstruction results achieve scores of 0.5275 and 0.8173 on the MANIQA and CLIP-IQA indicators, respectively. <u>The code will be available on GitHub:</u> <span><span><u>https://github.com/BJUT-AIVBD/AutoFGOR</u></span><svg><path></path></svg></span><u>.</u></div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"233 ","pages":"Pages 231-246"},"PeriodicalIF":12.2,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146033657","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Detecting global ocean subsurface density change with high-resolution via dual-task densely-former 利用双任务密度仪高分辨率探测全球海洋地下密度变化
IF 12.2 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2026-03-01 Epub Date: 2026-01-22 DOI: 10.1016/j.isprsjprs.2026.01.026
Hua Su , Weiqi Xie , Luping You , Sihui Li , Dian Lin , An Wang
High-resolution ocean subsurface density is crucial for studying dynamic processes and stratification within the ocean under recent global ocean warming. This study proposes a novel deep learning-based model, named DDFNet (Dual-task Densely-Former Network), for reconstructing ocean subsurface density, to address the challenges in reconstructing high-resolution and high-reliability global ocean subsurface density. DDFNet employs multi-scale feature extraction, attention mechanisms, and a dual-label design, combining an encoder-decoder backbone network with a global spatial attention module to capture the complex spatiotemporal relationships in ocean data effectively. The model utilizes multisource surface remote sensing data as input and incorporates Argo profile data and ORAS5 reanalysis data as labels. An adaptive weighted loss function dynamically balances the contributions of the two label types, improving reconstruction accuracy and achieving a spatial resolution of 0.25°×0.25°. By constructing dual tasks with in situ observations and reanalysis data for joint learning, the true state of the ocean and the consistency of physical processes are enhanced, improving the model’s reconstruction accuracy and physical consistency. Experimental results demonstrate that DDFNet outperforms well-used LightGBM and CNN models, with the reconstructed DDFNet-SD dataset achieving an R2 of 0.9863 and an RMSE of 0.2804 kg/m3. The dataset further reveals a declining trend in global ocean subsurface density at a rate of −4.47 × 10-4 kg/m3/decade, particularly pronounced in the upper 0–700 m, which is likely associated with global ocean warming and salinity changes. The high-resolution dataset facilitates studies on mesoscale ocean dynamics, stratification variability, and climate change impacts.
高分辨率的海洋地下密度对于研究近年来全球海洋变暖背景下海洋内部的动态过程和分层至关重要。本文提出了一种基于深度学习的海洋地下密度重建模型DDFNet (Dual-task dense - former Network),以解决重建高分辨率、高可靠性全球海洋地下密度的挑战。DDFNet采用多尺度特征提取、关注机制和双标签设计,将编码器-解码器骨干网与全局空间关注模块相结合,有效捕获海洋数据中复杂的时空关系。该模型采用多源地表遥感数据作为输入,并结合Argo剖面数据和ORAS5再分析数据作为标签。自适应加权损失函数动态平衡了两种标签类型的贡献,提高了重建精度,实现了0.25°×0.25°的空间分辨率。通过构建现场观测和再分析数据联合学习的双重任务,增强了海洋真实状态和物理过程的一致性,提高了模型的重建精度和物理一致性。实验结果表明,DDFNet优于常用的LightGBM和CNN模型,重建的DDFNet- sd数据集的R2为0.9863,RMSE为0.2804 kg/m3。数据集进一步揭示了全球海洋地下密度以- 4.47 × 10-4 kg/m3/ 10年的速率下降的趋势,特别是在0-700 m以上,这可能与全球海洋变暖和盐度变化有关。高分辨率数据集有助于中尺度海洋动力学、分层变率和气候变化影响的研究。
{"title":"Detecting global ocean subsurface density change with high-resolution via dual-task densely-former","authors":"Hua Su ,&nbsp;Weiqi Xie ,&nbsp;Luping You ,&nbsp;Sihui Li ,&nbsp;Dian Lin ,&nbsp;An Wang","doi":"10.1016/j.isprsjprs.2026.01.026","DOIUrl":"10.1016/j.isprsjprs.2026.01.026","url":null,"abstract":"<div><div>High-resolution ocean subsurface density is crucial for studying dynamic processes and stratification within the ocean under recent global ocean warming. This study proposes a novel deep learning-based model, named DDFNet (Dual-task Densely-Former Network), for reconstructing ocean subsurface density, to address the challenges in reconstructing high-resolution and high-reliability global ocean subsurface density. DDFNet employs multi-scale feature extraction, attention mechanisms, and a dual-label design, combining an encoder-decoder backbone network with a global spatial attention module to capture the complex spatiotemporal relationships in ocean data effectively. The model utilizes multisource surface remote sensing data as input and incorporates Argo profile data and ORAS5 reanalysis data as labels. An adaptive weighted loss function dynamically balances the contributions of the two label types, improving reconstruction accuracy and achieving a spatial resolution of 0.25°×0.25°. By constructing dual tasks with <em>in situ</em> observations and reanalysis data for joint learning, the true state of the ocean and the consistency of physical processes are enhanced, improving the model’s reconstruction accuracy and physical consistency. Experimental results demonstrate that DDFNet outperforms well-used LightGBM and CNN models, with the reconstructed DDFNet-SD dataset achieving an <em>R<sup>2</sup></em> of 0.9863 and an RMSE of 0.2804 kg/m<sup>3</sup>. The dataset further reveals a declining trend in global ocean subsurface density at a rate of −4.47 × 10<sup>-4</sup> kg/m<sup>3</sup>/decade, particularly pronounced in the upper 0–700 m, which is likely associated with global ocean warming and salinity changes. The high-resolution dataset facilitates studies on mesoscale ocean dynamics, stratification variability, and climate change impacts.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"233 ","pages":"Pages 158-179"},"PeriodicalIF":12.2,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146033934","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
WEGLA-NormGAN: wavelet-enhanced Cycle-GAN with global-local attention for radiometric normalization of remote sensing images WEGLA-NormGAN:基于全局-局部关注的遥感图像辐射归一化小波增强循环gan
IF 12.2 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2026-03-01 Epub Date: 2026-01-19 DOI: 10.1016/j.isprsjprs.2026.01.020
Wenxia Gan , Yu Feng , Jianhao Miao , Xinghua Li , Huanfeng Shen
The diversity of satellite remote sensing images has significantly enhanced the capability to observe surface information on Earth. However, multi-temporal optical remote sensing images acquired from different sensor platforms often exhibit substantial radiometric discrepancies, and it is difficult to obtain overlapping reference images, which poses critical challenges for seamless large-scale mosaicking, including global radiometric inconsistency, unsmooth local transitions, and visible seamlines. Existing traditional and deep learning methods can achieve reasonable performance on paired datasets, but often face challenges in balancing spatial structural integrity with enhanced radiometric consistency and generalizing to unseen images. To address these issues, a wavelet-enhanced radiometric normalization network called WEGLA-NormGAN is proposed to generate radiometrically normalized imagery with sound radiometric consistency and spatial fidelity. This framework integrates frequency-domain and spatial-domain information to achieve consistent multi-scale radiometric feature modeling while ensuring spatial structural fidelity. Firstly, wavelet transform is introduced to effectively decouple radiometric information and structural features from images, explicitly enhancing radiometric feature representation and edge-texture preservation. Secondly, a U-Net architecture with multi-scale modeling advantages is fused with an adaptive attention mechanism incorporating residual structures. This hybrid design employs a statistical alignment strategy to efficiently extract global shallow features and local statistical information, adaptively adjust the dynamic attention of unseen data, and alleviate local distortions, improving radiometric consistency and achieving high-fidelity spatial structure preservation. The proposed framework generates radiometrically normalized imagery that harmonizes radiometric consistency with spatial fidelity, while achieving outstanding radiometric normalization even in unseen scenarios. Extensive experiments were conducted on two public datasets and a self-constructed dataset. The results demonstrate that WEGLA-NormGAN outperforms seven state-of-the-art methods in cross-temporal scenarios and five in cross-spatiotemporal scenarios in terms of radiometric consistency, structural fidelity, and robustness. The code is available at https://github.com/WITRS/WeGLA-Norm.git.
卫星遥感影像的多样性大大提高了观测地球表面信息的能力。然而,从不同传感器平台获取的多时相光学遥感图像往往存在较大的辐射差异,并且难以获得重叠的参考图像,这对无缝大规模拼接提出了严峻的挑战,包括全局辐射不一致、局部过渡不平滑和接缝可见。现有的传统和深度学习方法可以在配对数据集上获得合理的性能,但在平衡空间结构完整性与增强辐射一致性以及推广到未见图像方面往往面临挑战。为了解决这些问题,提出了一种名为WEGLA-NormGAN的小波增强辐射归一化网络,以生成具有声音辐射一致性和空间保真度的辐射归一化图像。该框架集成了频域和空域信息,在保证空间结构保真度的同时实现一致的多尺度辐射特征建模。首先,引入小波变换,有效解耦图像中的辐射特征信息和结构特征,明显增强辐射特征表示和边缘纹理保持;其次,将具有多尺度建模优势的U-Net体系结构与包含残余结构的自适应注意机制相融合;该混合设计采用统计对齐策略,有效提取全局浅特征和局部统计信息,自适应调整未见数据的动态注意力,减轻局部失真,提高辐射一致性,实现高保真的空间结构保存。所提出的框架生成辐射标准化图像,使辐射一致性与空间保真度相协调,同时即使在看不见的场景中也能实现出色的辐射标准化。在两个公共数据集和一个自建数据集上进行了大量的实验。结果表明,在辐射一致性、结构保真度和鲁棒性方面,WEGLA-NormGAN在跨时间情景下优于7种最先进的方法,在跨时空情景下优于5种最先进的方法。代码可在https://github.com/WITRS/WeGLA-Norm.git上获得。
{"title":"WEGLA-NormGAN: wavelet-enhanced Cycle-GAN with global-local attention for radiometric normalization of remote sensing images","authors":"Wenxia Gan ,&nbsp;Yu Feng ,&nbsp;Jianhao Miao ,&nbsp;Xinghua Li ,&nbsp;Huanfeng Shen","doi":"10.1016/j.isprsjprs.2026.01.020","DOIUrl":"10.1016/j.isprsjprs.2026.01.020","url":null,"abstract":"<div><div>The diversity of satellite remote sensing images has significantly enhanced the capability to observe surface information on Earth. However, multi-temporal optical remote sensing images acquired from different sensor platforms often exhibit substantial radiometric discrepancies, and it is difficult to obtain overlapping reference images, which poses critical challenges for seamless large-scale mosaicking, including global radiometric inconsistency, unsmooth local transitions, and visible seamlines. Existing traditional and deep learning methods can achieve reasonable performance on paired datasets, but often face challenges in balancing spatial structural integrity with enhanced radiometric consistency and generalizing to unseen images. To address these issues, a wavelet-enhanced radiometric normalization network called WEGLA-NormGAN is proposed to generate radiometrically normalized imagery with sound radiometric consistency and spatial fidelity. This framework integrates frequency-domain and spatial-domain information to achieve consistent multi-scale radiometric feature modeling while ensuring spatial structural fidelity. Firstly, wavelet transform is introduced to effectively decouple radiometric information and structural features from images, explicitly enhancing radiometric feature representation and edge-texture preservation. Secondly, a U-Net architecture with multi-scale modeling advantages is fused with an adaptive attention mechanism incorporating residual structures. This hybrid design employs a statistical alignment strategy to efficiently extract global shallow features and local statistical information, adaptively adjust the dynamic attention of unseen data, and alleviate local distortions, improving radiometric consistency and achieving high-fidelity spatial structure preservation. The proposed framework generates radiometrically normalized imagery that harmonizes radiometric consistency with spatial fidelity, while achieving outstanding radiometric normalization even in unseen scenarios. Extensive experiments were conducted on two public datasets and a self-constructed dataset. The results demonstrate that WEGLA-NormGAN outperforms seven state-of-the-art methods in cross-temporal scenarios and five in cross-spatiotemporal scenarios in terms of radiometric consistency, structural fidelity, and robustness. The code is available at <span><span>https://github.com/WITRS/WeGLA-Norm.git</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"233 ","pages":"Pages 39-54"},"PeriodicalIF":12.2,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146001037","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
RTPSeg: A multi-modality dataset for LiDAR point cloud semantic segmentation assisted with RGB-thermal images in autonomous driving RTPSeg:自动驾驶中辅助rgb热图像的激光雷达点云语义分割多模态数据集
IF 12.2 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2026-03-01 Epub Date: 2026-01-16 DOI: 10.1016/j.isprsjprs.2026.01.008
Yifan Sun , Chenguang Dai , Wenke Li , Xinpu Liu , Yongqi Sun , Ye Zhang , Weijun Guan , Yongsheng Zhang , Yulan Guo , Hanyun Wang
LiDAR point cloud semantic segmentation is crucial for scene understanding in autonomous driving, yet the sparse and textureless characteristics of point clouds cause huge challenges for this task. To address this, numerous studies have explored to leverage the dense color and fine-grained texture from RGB images for multi-modality 3D semantic segmentation. Nevertheless, these methods still encounter certain limitations when facing complex scenarios, as RGB images degrade under poor lighting conditions. In contrast, thermal infrared (TIR) images can provide thermal radiation information of road objects and are robust to illumination change, offering complementary advantages to RGB images. Therefore, in this work we introduce RTPSeg, the first and only multi-modality dataset to simultaneously provide RGB and TIR images for point cloud semantic segmentation. RTPSeg includes over 3000 synchronized frames collected by RGB camera, infrared camera, and LiDAR, providing over 248M pointwise annotations for 18 semantic categories in autonomous driving, involving urban and village scenes during both daytime and nighttime. Based on RTPSeg, we also propose RTPSegNet, a baseline model for point cloud semantic segmentation jointly assisted with RGB and TIR images. Extensive experiments demonstrate that the RTPSeg dataset presents considerable challenges and opportunities to existing point cloud semantic segmentation approaches, and our RTPSegNet exhibits promising effectiveness in jointly leveraging the complementary information between point clouds, RGB images, and TIR images. More importantly, the experimental results also confirm that 3D semantic segmentation can be effectively enhanced by introducing additional TIR image modality, revealing the promising potential of this innovative research and application. We anticipate that the RTPSeg will catalyze in-depth research in this field. Both RTPSeg and RTPSegNet will be released at https://github.com/sssssyf/RTPSeg
激光雷达点云语义分割对于自动驾驶场景理解至关重要,但点云的稀疏性和无纹理性给这一任务带来了巨大的挑战。为了解决这个问题,许多研究探索了利用RGB图像的密集颜色和细粒度纹理进行多模态3D语义分割。然而,这些方法在面对复杂场景时仍然会遇到一定的局限性,因为RGB图像在较差的光照条件下会退化。相比之下,热红外(TIR)图像可以提供道路物体的热辐射信息,并且对光照变化具有鲁棒性,是RGB图像的互补优势。因此,在这项工作中,我们引入了RTPSeg,这是第一个也是唯一一个同时提供RGB和TIR图像用于点云语义分割的多模态数据集。RTPSeg包括由RGB相机、红外相机和激光雷达采集的3000多帧同步帧,提供超过248M的自动驾驶18个语义类别的点向注释,包括白天和夜间的城市和乡村场景。在RTPSeg的基础上,提出了RTPSegNet——RGB和TIR图像联合辅助的点云语义分割基线模型。大量的实验表明,RTPSeg数据集对现有的点云语义分割方法提出了相当大的挑战和机遇,我们的RTPSegNet在联合利用点云、RGB图像和TIR图像之间的互补信息方面表现出了良好的效果。更重要的是,实验结果也证实了通过引入额外的TIR图像模态可以有效地增强三维语义分割,揭示了这一创新研究和应用的广阔潜力。我们期待RTPSeg将促进这一领域的深入研究。RTPSeg和rtpsenet都将在https://github.com/sssssyf/RTPSeg上发布
{"title":"RTPSeg: A multi-modality dataset for LiDAR point cloud semantic segmentation assisted with RGB-thermal images in autonomous driving","authors":"Yifan Sun ,&nbsp;Chenguang Dai ,&nbsp;Wenke Li ,&nbsp;Xinpu Liu ,&nbsp;Yongqi Sun ,&nbsp;Ye Zhang ,&nbsp;Weijun Guan ,&nbsp;Yongsheng Zhang ,&nbsp;Yulan Guo ,&nbsp;Hanyun Wang","doi":"10.1016/j.isprsjprs.2026.01.008","DOIUrl":"10.1016/j.isprsjprs.2026.01.008","url":null,"abstract":"<div><div>LiDAR point cloud semantic segmentation is crucial for scene understanding in autonomous driving, yet the sparse and textureless characteristics of point clouds cause huge challenges for this task. To address this, numerous studies have explored to leverage the dense color and fine-grained texture from RGB images for multi-modality 3D semantic segmentation. Nevertheless, these methods still encounter certain limitations when facing complex scenarios, as RGB images degrade under poor lighting conditions. In contrast, thermal infrared (TIR) images can provide thermal radiation information of road objects and are robust to illumination change, offering complementary advantages to RGB images. Therefore, in this work we introduce RTPSeg, the first and only multi-modality dataset to simultaneously provide RGB and TIR images for point cloud semantic segmentation. RTPSeg includes over 3000 synchronized frames collected by RGB camera, infrared camera, and LiDAR, providing over 248M pointwise annotations for 18 semantic categories in autonomous driving, involving urban and village scenes during both daytime and nighttime. Based on RTPSeg, we also propose RTPSegNet, a baseline model for point cloud semantic segmentation jointly assisted with RGB and TIR images. Extensive experiments demonstrate that the RTPSeg dataset presents considerable challenges and opportunities to existing point cloud semantic segmentation approaches, and our RTPSegNet exhibits promising effectiveness in jointly leveraging the complementary information between point clouds, RGB images, and TIR images. More importantly, the experimental results also confirm that 3D semantic segmentation can be effectively enhanced by introducing additional TIR image modality, revealing the promising potential of this innovative research and application. We anticipate that the RTPSeg will catalyze in-depth research in this field. Both RTPSeg and RTPSegNet will be released at <span><span>https://github.com/sssssyf/RTPSeg</span><svg><path></path></svg></span></div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"233 ","pages":"Pages 25-38"},"PeriodicalIF":12.2,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145982030","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
RECREATE: Supervised contrastive learning and inpainting based hyperspectral image denoising 再现:基于高光谱图像去噪的监督对比学习和图像修复
IF 12.2 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2026-03-01 Epub Date: 2026-01-16 DOI: 10.1016/j.isprsjprs.2026.01.022
Aditya Dixit , Anup Kumar Gupta , Puneet Gupta , Ankur Garg
Hyperspectral image (HSI) contains information at various spectra, making it valuable in several real-world applications such as environmental monitoring, agriculture, and remote sensing. However, the acquisition process often introduces noise, necessitating effective HSI denoising methods to maintain its applicability. Deep Learning (DL) is considered as the de-facto for HSI denoising, but it requires a significant number of training samples to optimize network parameters for effective denoising outcomes. However, obtaining extensive datasets is challenging in HSI, leading to epistemic uncertainties and thereby deteriorating the denoising performance. This paper introduces a novel supervised contrastive learning (SCL) method, RECREATE, to enhance feature learning and mitigate the issue of epistemic uncertainty for HSI denoising. Furthermore, we introduce the exploration of image inpainting as an auxiliary task to enhance the HSI denoising performance. By adding HSI inpainting to CL, our method essentially enhances HSI denoising by increasing training datasets and enforcing improved feature learning. Experimental outcomes on various HSI datasets validate the efficacy of RECREATE, showcasing its potential for integration with existing HSI denoising techniques to enhance their performance, both qualitatively and quantitatively. This innovative method holds promise for addressing the limitations posed by limited training data and thereby advancing the field toward proposing better HSI denoising methods.
高光谱图像(HSI)包含各种光谱的信息,使其在环境监测,农业和遥感等几个现实世界的应用中具有价值。然而,采集过程往往会引入噪声,需要有效的HSI去噪方法来保持其适用性。深度学习(DL)被认为是HSI去噪的事实,但它需要大量的训练样本来优化网络参数以获得有效的去噪结果。然而,在HSI中获得广泛的数据集是具有挑战性的,导致认知不确定性,从而降低了去噪性能。本文介绍了一种新的监督对比学习(SCL)方法——recrere,以增强特征学习并减轻HSI去噪的认知不确定性问题。此外,我们还介绍了对图像着色的探索,作为增强HSI去噪性能的辅助任务。通过将HSI图像添加到CL中,我们的方法通过增加训练数据集和加强改进的特征学习,本质上增强了HSI去噪。在各种HSI数据集上的实验结果验证了re的有效性,展示了其与现有HSI去噪技术集成的潜力,以提高其定性和定量性能。这种创新的方法有望解决有限的训练数据所带来的限制,从而推动该领域提出更好的HSI去噪方法。
{"title":"RECREATE: Supervised contrastive learning and inpainting based hyperspectral image denoising","authors":"Aditya Dixit ,&nbsp;Anup Kumar Gupta ,&nbsp;Puneet Gupta ,&nbsp;Ankur Garg","doi":"10.1016/j.isprsjprs.2026.01.022","DOIUrl":"10.1016/j.isprsjprs.2026.01.022","url":null,"abstract":"<div><div>Hyperspectral image (HSI) contains information at various spectra, making it valuable in several real-world applications such as environmental monitoring, agriculture, and remote sensing. However, the acquisition process often introduces noise, necessitating effective HSI denoising methods to maintain its applicability. Deep Learning (DL) is considered as the de-facto for HSI denoising, but it requires a significant number of training samples to optimize network parameters for effective denoising outcomes. However, obtaining extensive datasets is challenging in HSI, leading to epistemic uncertainties and thereby deteriorating the denoising performance. This paper introduces a novel supervised contrastive learning (SCL) method, <em>RECREATE</em>, to enhance feature learning and mitigate the issue of epistemic uncertainty for HSI denoising. Furthermore, we introduce the exploration of image inpainting as an auxiliary task to enhance the HSI denoising performance. By adding HSI inpainting to CL, our method essentially enhances HSI denoising by increasing training datasets and enforcing improved feature learning. Experimental outcomes on various HSI datasets validate the efficacy of <em>RECREATE</em>, showcasing its potential for integration with existing HSI denoising techniques to enhance their performance, both qualitatively and quantitatively. This innovative method holds promise for addressing the limitations posed by limited training data and thereby advancing the field toward proposing better HSI denoising methods.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"233 ","pages":"Pages 14-24"},"PeriodicalIF":12.2,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145982031","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Progressive uncertainty-guided network for binary segmentation in high-resolution remote sensing imagery 高分辨率遥感图像二值分割的渐进式不确定性引导网络
IF 12.2 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2026-02-01 Epub Date: 2026-01-07 DOI: 10.1016/j.isprsjprs.2026.01.010
Jiepan Li , Wei He , Ting Hu , Minghao Tang , Liangpei Zhang
Binary semantic segmentation in remote sensing (RS) imagery faces persistent challenges due to complex object appearances, ambiguous boundaries, and high similarity between foreground and background, all of which introduce significant uncertainty into the prediction process. Existing approaches often treat uncertainty as either a global attribute or a pixel-level estimate, overlooking the critical role of spatial and contextual interactions. To address these limitations, we propose the Progressive Uncertainty-Guided Segmentation Network (PUGNet), a unified framework that explicitly models uncertainty in a context-aware manner. PUGNet decomposes uncertainty into three distinct components: foreground uncertainty, background uncertainty, and contextual uncertainty. This tripartite modeling enables more precise handling of local ambiguities and global inconsistencies. We adopt a coarse-to-fine decoding strategy that progressively refines features through two specialized modules. The Dynamic Uncertainty-Aware Module enhances regions of high foreground and background uncertainty using Gaussian-based modeling and contrastive learning. The Entropy-Driven Refinement Module quantifies contextual uncertainty via entropy and facilitates adaptive refinement through multi-scale context aggregation. Extensive experiments on ten public benchmark datasets, covering both single-temporal (e.g., building and cropland extraction) and bi-temporal (e.g., building change detection) binary segmentation tasks, demonstrate that PUGNet consistently achieves superior segmentation accuracy and uncertainty reduction, establishing a new state of the art in RS binary segmentation. The full implementation of the proposed framework and all experimental results can be accessed at https://github.com/Henryjiepanli/PU_RS.
由于物体外观复杂、边界模糊、前景与背景高度相似等特点,给遥感图像的二值语义分割带来了很大的不确定性。现有的方法通常将不确定性视为全局属性或像素级估计,忽略了空间和上下文相互作用的关键作用。为了解决这些限制,我们提出了渐进式不确定性引导分割网络(PUGNet),这是一个统一的框架,以上下文感知的方式明确地建模不确定性。PUGNet将不确定性分解为三个不同的组成部分:前景不确定性、背景不确定性和上下文不确定性。这种三方建模可以更精确地处理局部模糊和全局不一致。我们采用一种从粗到精的解码策略,通过两个专门的模块逐步细化特征。动态不确定性感知模块使用基于高斯的建模和对比学习来增强前景和背景不确定性高的区域。熵驱动的细化模块通过熵量化上下文不确定性,并通过多尺度上下文聚合促进自适应细化。在10个公共基准数据集上进行了广泛的实验,涵盖了单时间(如建筑物和农田提取)和双时间(如建筑物变化检测)二值分割任务,表明PUGNet始终如一地实现了卓越的分割精度和不确定性降低,建立了RS二值分割的新状态。提出的框架的完整实施和所有实验结果可以在https://github.com/Henryjiepanli/PU_RS上访问。
{"title":"Progressive uncertainty-guided network for binary segmentation in high-resolution remote sensing imagery","authors":"Jiepan Li ,&nbsp;Wei He ,&nbsp;Ting Hu ,&nbsp;Minghao Tang ,&nbsp;Liangpei Zhang","doi":"10.1016/j.isprsjprs.2026.01.010","DOIUrl":"10.1016/j.isprsjprs.2026.01.010","url":null,"abstract":"<div><div>Binary semantic segmentation in remote sensing (RS) imagery faces persistent challenges due to complex object appearances, ambiguous boundaries, and high similarity between foreground and background, all of which introduce significant uncertainty into the prediction process. Existing approaches often treat uncertainty as either a global attribute or a pixel-level estimate, overlooking the critical role of spatial and contextual interactions. To address these limitations, we propose the <strong>Progressive Uncertainty-Guided Segmentation Network (PUGNet)</strong>, a unified framework that explicitly models uncertainty in a context-aware manner. PUGNet decomposes uncertainty into three distinct components: <strong>foreground uncertainty</strong>, <strong>background uncertainty</strong>, and <strong>contextual uncertainty</strong>. This tripartite modeling enables more precise handling of local ambiguities and global inconsistencies. We adopt a coarse-to-fine decoding strategy that progressively refines features through two specialized modules. The <strong>Dynamic Uncertainty-Aware Module</strong> enhances regions of high foreground and background uncertainty using Gaussian-based modeling and contrastive learning. The <strong>Entropy-Driven Refinement Module</strong> quantifies contextual uncertainty via entropy and facilitates adaptive refinement through multi-scale context aggregation. Extensive experiments on ten public benchmark datasets, covering both single-temporal (<em>e.g.</em>, building and cropland extraction) and bi-temporal (<em>e.g.</em>, building change detection) binary segmentation tasks, demonstrate that PUGNet consistently achieves superior segmentation accuracy and uncertainty reduction, establishing a new state of the art in RS binary segmentation. The full implementation of the proposed framework and all experimental results can be accessed at <span><span>https://github.com/Henryjiepanli/PU_RS</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"232 ","pages":"Pages 561-577"},"PeriodicalIF":12.2,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145925511","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Change tensor: Estimating complex topographic changes from point clouds using Riemann manifold surfaces 变化张量:估计复杂的地形变化从点云使用黎曼流形表面
IF 12.2 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2026-02-01 Epub Date: 2026-01-12 DOI: 10.1016/j.isprsjprs.2026.01.009
Shoujun Jia , Lotte de Vugt , Andreas Mayr , Katharina Anders , Chun Liu , Martin Rutzinger
Estimating complex 3D topographic surface changes including rigid spatial movement and non-rigid morphological deformation is an essential task to investigate Earth surface dynamics. However, for current 3D point comparison approaches, it is challenging to separate rigid and non-rigid topographic surface changes from multi-temporal 3D point clouds. Additionally, these methods are affected by challenges including topographic surface roughness and point cloud heterogeneities (i.e., discrete and irregular point distributions). To address these challenges, in this paper, we consider the dynamic evolution of topographic surfaces as the geometric changes of Riemann manifold surfaces. By building Euclidean (straight) and non-Euclidean (curved) coordinate systems on Riemann manifold surfaces that are represented from point clouds, the rigid transformation and non-rigid deformation of the Riemann manifold surfaces are solved to conceptualize rigid and non-rigid change tensors, respectively. On this basis, we design rigid (i.e., translation and rotation) and non-rigid (i.e., stretch and distortion) change features to describe various topographic surface changes and quantify the associated uncertainties to capture significant changes. The proposed method is tested on pairwise point clouds with simulated and real topographic surface changes in mountain regions. Simulation experiments demonstrate that the proposed method performed better than the baseline (i.e., M3C2) and state-of-the-art methods (i.e., LOG), with a higher translation accuracy (more than 50% improvement), a lower translation uncertainty (more than 61% reduction), and strong robustness to varying point densities. These results also show that the proposed method accurately quantifies three additional types of change features (i.e., the mean accuracies of rotation, stretch, and distortion are 1.5°, 0.5%, and 3.5°, respectively). Moreover, the real-scene experimental results demonstrate the effectiveness and superiority of the proposed method in estimating various topographic changes in real environments, the applicability in analyzing geomorphological processes, and the potential contribution for understanding spatiotemporal patterns of Earth surface dynamics.
估计复杂的三维地形表面变化,包括刚性空间运动和非刚性形态变形,是研究地球表面动力学的重要任务。然而,对于目前的三维点比较方法来说,从多时相三维点云中分离出刚性和非刚性地形表面变化是一个挑战。此外,这些方法还受到地形表面粗糙度和点云异质性(即离散和不规则点分布)等挑战的影响。为了应对这些挑战,本文将地形表面的动态演变视为黎曼流形表面的几何变化。通过在点云表示的黎曼流形表面上建立欧几里得(直)和非欧几里得(弯)坐标系,求解黎曼流形表面的刚性变换和非刚性变形,分别将刚性和非刚性变化张量概念化。在此基础上,我们设计了刚性(即平移和旋转)和非刚性(即拉伸和扭曲)变化特征来描述各种地形表面变化,并量化相关的不确定性以捕捉重大变化。该方法在模拟和真实地形变化的山地成对点云上进行了测试。仿真实验表明,该方法优于基准方法(即M3C2)和最先进的方法(即LOG),具有更高的翻译精度(提高50%以上),更低的翻译不确定性(降低61%以上),并且对变点密度具有较强的鲁棒性。这些结果还表明,该方法准确量化了另外三种类型的变化特征(即旋转、拉伸和畸变的平均精度分别为1.5°、0.5%和3.5°)。此外,实际实验结果表明,该方法在估算真实环境中各种地形变化方面具有有效性和优越性,在分析地貌过程方面具有适用性,对理解地表动力学时空格局具有潜在贡献。
{"title":"Change tensor: Estimating complex topographic changes from point clouds using Riemann manifold surfaces","authors":"Shoujun Jia ,&nbsp;Lotte de Vugt ,&nbsp;Andreas Mayr ,&nbsp;Katharina Anders ,&nbsp;Chun Liu ,&nbsp;Martin Rutzinger","doi":"10.1016/j.isprsjprs.2026.01.009","DOIUrl":"10.1016/j.isprsjprs.2026.01.009","url":null,"abstract":"<div><div>Estimating complex 3D topographic surface changes including rigid spatial movement and non-rigid morphological deformation is an essential task to investigate Earth surface dynamics. However, for current 3D point comparison approaches, it is challenging to separate rigid and non-rigid topographic surface changes from multi-temporal 3D point clouds. Additionally, these methods are affected by challenges including topographic surface roughness and point cloud heterogeneities (i.e., discrete and irregular point distributions). To address these challenges, in this paper, we consider the dynamic evolution of topographic surfaces as the geometric changes of Riemann manifold surfaces. By building Euclidean (straight) and non-Euclidean (curved) coordinate systems on Riemann manifold surfaces that are represented from point clouds, the rigid transformation and non-rigid deformation of the Riemann manifold surfaces are solved to conceptualize rigid and non-rigid change tensors, respectively. On this basis, we design rigid (i.e., translation and rotation) and non-rigid (i.e., stretch and distortion) change features to describe various topographic surface changes and quantify the associated uncertainties to capture significant changes. The proposed method is tested on pairwise point clouds with simulated and real topographic surface changes in mountain regions. Simulation experiments demonstrate that the proposed method performed better than the baseline (i.e., M3C2) and state-of-the-art methods (i.e., LOG), with a higher translation accuracy (more than 50% improvement), a lower translation uncertainty (more than 61% reduction), and strong robustness to varying point densities. These results also show that the proposed method accurately quantifies three additional types of change features (i.e., the mean accuracies of rotation, stretch, and distortion are 1.5°, 0.5%, and 3.5°, respectively). Moreover, the real-scene experimental results demonstrate the effectiveness and superiority of the proposed method in estimating various topographic changes in real environments, the applicability in analyzing geomorphological processes, and the potential contribution for understanding spatiotemporal patterns of Earth surface dynamics.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"232 ","pages":"Pages 766-786"},"PeriodicalIF":12.2,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145957294","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
ISPRS Journal of Photogrammetry and Remote Sensing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1