首页 > 最新文献

ISPRS Journal of Photogrammetry and Remote Sensing最新文献

英文 中文
PGVS: A probabilistic graph-theoretic framework for view-graph selection in structure-from-motion PGVS:一种基于运动的结构视图选择的概率图论框架
IF 12.2 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2025-11-19 DOI: 10.1016/j.isprsjprs.2025.11.004
Pengwei Zhou , Hongche Yin , Guozheng Xu , Xiaosong Wei , Annan Zhou , Jian Yao , Li Li , Huang Jing
<div><div>The underlying view-graph is generally constructed through the matching of unordered image paris, which establish a crucial step in ensuring the accuracy and efficiency of the structure-from-motion (SfM) process. However, the initial graph often contain redundant and erroneous edges, which arise from incorrect image retrieval and ambiguous structures(e.g., symmetric buildings with identical or opposing facets), leading to the emergence of ghosting effects and superimposed reconstruction artifacts. Most contemporary approaches employ bespoke solutions to attain specific reconstruction goals, such as efficiency, precision, or disambiguation. In contrast to these task-specific methods, we propose a probabilistic graph-theoretic framework, termed PGVS, which formulates the view-graph selection problem as a weighted maximum clique optimization problem, achieving both sparsification and disambiguation simultaneously. Furthermore, we develop a sophisticated binary penalty continuous relaxation technique to derive a solution that is guaranteed to correspond to the optimal outcome of the original problem. In contrast to techniques for verifying pose consistency, we introduce a context-aware graph similarity assessment mechanism that is based on view triplets with a multi-view patch tracking strategy. This approach helps alleviate the effects of vanishing keypoints and environmental occlusions and reduces the impact of erroneous image correspondences that often undermine the reliability of pose estimation. Moreover, we develop a Bayesian inference framework to evaluate edge-level consistency analysis over the context graph, enabling us to estimate the likelihood that each edge reflects a globally coherent match. This probabilistic characterization is then leveraged to construct the adjacency matrix for a weighted maximum clique formulation. To solve this combinatorial problem, we employ a continuous binary-penalty relaxation technique, which enables us to obtain an optimal solution reflecting global consistency with the highest matching affinity and confidence. The resulting selected view-graph constitutes a novel and efficient algorithmic component that can be seamlessly integrated as a preprocessing module into any SfM pipeline, thereby enhancing its adaptability and general applicability. We validate the efficacy of our method on both generic and ambiguous datasets, which cover a wide spectrum of small, medium, and large-scale datasets, each exhibiting distinct statistical characteristics. In generic datasets, our approach significantly reduces reconstruction time by removing redundant edges to sparsify the view-graph while preserving accuracy and mitigating ghosting artifacts. For ambiguous datasets, our method excels in identifying erroneous matches, even under highly challenging conditions, leading to accurate, disambiguated, and unsuperimposed 3D reconstructions. The source code of our approach is publicly available at <span><span>https://
底层视图通常是通过对无序图像的匹配来构建的,这是保证结构-运动(SfM)过程精度和效率的关键步骤。然而,初始图通常包含冗余和错误的边缘,这是由于不正确的图像检索和模糊的结构(例如:例如,具有相同或相反侧面的对称建筑),导致重影效果和叠加重建伪影的出现。大多数现代方法采用定制的解决方案来实现特定的重建目标,例如效率、精度或消歧义。与这些特定任务的方法相比,我们提出了一个概率图论框架,称为PGVS,它将视图图选择问题表述为加权最大团优化问题,同时实现了稀疏化和消歧。此外,我们开发了一种复杂的二元惩罚连续松弛技术,以获得保证对应于原始问题最优结果的解。与验证姿态一致性的技术相比,我们引入了一种基于视图三元组和多视图补丁跟踪策略的上下文感知图相似性评估机制。这种方法有助于减轻关键点消失和环境遮挡的影响,并减少错误图像对应的影响,错误图像对应通常会破坏姿态估计的可靠性。此外,我们开发了一个贝叶斯推理框架来评估上下文图上的边缘级一致性分析,使我们能够估计每个边缘反映全局连贯匹配的可能性。然后利用这种概率特征来构造邻接矩阵,用于加权最大团公式。为了解决这一组合问题,我们采用了连续二元惩罚松弛技术,使我们能够获得具有最高匹配亲和度和置信度的反映全局一致性的最优解。所选择的视图图构成了一种新颖而高效的算法组件,可以作为预处理模块无缝集成到任何SfM管道中,从而增强了其适应性和通用适用性。我们验证了我们的方法在通用和模糊数据集上的有效性,这些数据集涵盖了广泛的小型,中型和大型数据集,每个数据集都表现出不同的统计特征。在通用数据集中,我们的方法通过去除冗余边缘来稀疏视图图,同时保持准确性并减轻重影工件,从而显着减少了重建时间。对于模棱两可的数据集,我们的方法在识别错误匹配方面表现出色,即使在极具挑战性的条件下,也能实现准确、消除歧义和无叠加的3D重建。我们的方法的源代码可以在https://github.com/zhoupengwei/PGVS上公开获得。
{"title":"PGVS: A probabilistic graph-theoretic framework for view-graph selection in structure-from-motion","authors":"Pengwei Zhou ,&nbsp;Hongche Yin ,&nbsp;Guozheng Xu ,&nbsp;Xiaosong Wei ,&nbsp;Annan Zhou ,&nbsp;Jian Yao ,&nbsp;Li Li ,&nbsp;Huang Jing","doi":"10.1016/j.isprsjprs.2025.11.004","DOIUrl":"10.1016/j.isprsjprs.2025.11.004","url":null,"abstract":"&lt;div&gt;&lt;div&gt;The underlying view-graph is generally constructed through the matching of unordered image paris, which establish a crucial step in ensuring the accuracy and efficiency of the structure-from-motion (SfM) process. However, the initial graph often contain redundant and erroneous edges, which arise from incorrect image retrieval and ambiguous structures(e.g., symmetric buildings with identical or opposing facets), leading to the emergence of ghosting effects and superimposed reconstruction artifacts. Most contemporary approaches employ bespoke solutions to attain specific reconstruction goals, such as efficiency, precision, or disambiguation. In contrast to these task-specific methods, we propose a probabilistic graph-theoretic framework, termed PGVS, which formulates the view-graph selection problem as a weighted maximum clique optimization problem, achieving both sparsification and disambiguation simultaneously. Furthermore, we develop a sophisticated binary penalty continuous relaxation technique to derive a solution that is guaranteed to correspond to the optimal outcome of the original problem. In contrast to techniques for verifying pose consistency, we introduce a context-aware graph similarity assessment mechanism that is based on view triplets with a multi-view patch tracking strategy. This approach helps alleviate the effects of vanishing keypoints and environmental occlusions and reduces the impact of erroneous image correspondences that often undermine the reliability of pose estimation. Moreover, we develop a Bayesian inference framework to evaluate edge-level consistency analysis over the context graph, enabling us to estimate the likelihood that each edge reflects a globally coherent match. This probabilistic characterization is then leveraged to construct the adjacency matrix for a weighted maximum clique formulation. To solve this combinatorial problem, we employ a continuous binary-penalty relaxation technique, which enables us to obtain an optimal solution reflecting global consistency with the highest matching affinity and confidence. The resulting selected view-graph constitutes a novel and efficient algorithmic component that can be seamlessly integrated as a preprocessing module into any SfM pipeline, thereby enhancing its adaptability and general applicability. We validate the efficacy of our method on both generic and ambiguous datasets, which cover a wide spectrum of small, medium, and large-scale datasets, each exhibiting distinct statistical characteristics. In generic datasets, our approach significantly reduces reconstruction time by removing redundant edges to sparsify the view-graph while preserving accuracy and mitigating ghosting artifacts. For ambiguous datasets, our method excels in identifying erroneous matches, even under highly challenging conditions, leading to accurate, disambiguated, and unsuperimposed 3D reconstructions. The source code of our approach is publicly available at &lt;span&gt;&lt;span&gt;https://","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"231 ","pages":"Pages 641-663"},"PeriodicalIF":12.2,"publicationDate":"2025-11-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145560227","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Bundle adjustment-based co-registration with high geolocation accuracy for UAV photogrammetry 基于束平差的无人机摄影测量高精度协同配准
IF 12.2 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2025-11-17 DOI: 10.1016/j.isprsjprs.2025.11.014
Nan Jiang , Hai-bo Li , Cong-jiang Li , Yu-xiang Hu , Jia-wen Zhou
Co-registration of UAV photogrammetry data has emerged as a vital technique for aligning multi-temporal aerial datasets due to its operational simplicity and low cost. However, its persistently low geolocation accuracy, resulting from the unavailability of real ground control points (GCPs), remains challenging. To resolve this limitation, we identify and quantify the “Error-Correction Effect (ECE)” in bundle adjustment (BA) and propose a novel method leveraging this phenomenon to estimate horizontal and vertical geographic coordinates of assumed control points (ACPs), significantly enhancing geolocation accuracy in multi-temporal UAV co-registration. Through phenomenological description and formula derivation, we first conduct theoretical analysis of ECE to elucidate its formation mechanism. Field experiments and sensitivity analysis then clarify ECE’s triggering conditions and quantify its characteristics. Subsequently, we develop ECE-based solutions for deriving horizontal and vertical coordinates of ACPs, with parametric sensitivity and computational accuracy validated through field-measured data. Results demonstrate that accuracy primarily correlates with distance from ACPs to the controlled areas. At operational distances of 650–900 m, mean horizontal coordinate errors range from 8-13 mm, while vertical errors range from 10-60 mm. This represents a substantial improvement over conventional co-registration, which exhibits mean errors of 214 mm (horizontal) and 295 mm (vertical) under identical conditions.
Our code is available at https://github.com/meowxu/ECE-fitting.
无人机摄影测量数据的协同配准由于其操作简单和成本低而成为多时相航空数据集对准的重要技术。然而,由于无法获得真正的地面控制点(gcp),其持续的低地理定位精度仍然具有挑战性。为了解决这一限制,我们识别并量化了束平差(BA)中的“误差校正效应(ECE)”,并提出了一种利用这一现象估计假设控制点(acp)的水平和垂直地理坐标的新方法,显著提高了多时相无人机协同配准的地理定位精度。通过现象学描述和公式推导,首先对ECE进行理论分析,阐明其形成机制。现场实验和灵敏度分析明确了ECE的触发条件,量化了其特性。随后,我们开发了基于ece的解决方案,用于推导acp的水平和垂直坐标,并通过现场测量数据验证了参数灵敏度和计算精度。结果表明,准确度主要与acp到控制区的距离有关。在650 ~ 900 m的操作距离内,平均水平坐标误差为8 ~ 13 mm,垂直坐标误差为10 ~ 60 mm。这比传统的共配准有了很大的改进,传统的共配准在相同条件下的平均误差为214毫米(水平)和295毫米(垂直)。
{"title":"Bundle adjustment-based co-registration with high geolocation accuracy for UAV photogrammetry","authors":"Nan Jiang ,&nbsp;Hai-bo Li ,&nbsp;Cong-jiang Li ,&nbsp;Yu-xiang Hu ,&nbsp;Jia-wen Zhou","doi":"10.1016/j.isprsjprs.2025.11.014","DOIUrl":"10.1016/j.isprsjprs.2025.11.014","url":null,"abstract":"<div><div>Co-registration of UAV photogrammetry data has emerged as a vital technique for aligning multi-temporal aerial datasets due to its operational simplicity and low cost. However, its persistently low geolocation accuracy, resulting from the unavailability of real ground control points (GCPs), remains challenging. To resolve this limitation, we identify and quantify the “Error-Correction Effect (ECE)” in bundle adjustment (BA) and propose a novel method leveraging this phenomenon to estimate horizontal and vertical geographic coordinates of assumed control points (ACPs), significantly enhancing geolocation accuracy in multi-temporal UAV co-registration. Through phenomenological description and formula derivation, we first conduct theoretical analysis of ECE to elucidate its formation mechanism. Field experiments and sensitivity analysis then clarify ECE’s triggering conditions and quantify its characteristics. Subsequently, we develop ECE-based solutions for deriving horizontal and vertical coordinates of ACPs, with parametric sensitivity and computational accuracy validated through field-measured data. Results demonstrate that accuracy primarily correlates with distance from ACPs to the controlled areas. At operational distances of 650–900 m, mean horizontal coordinate errors range from 8-13 mm, while vertical errors range from 10-60 mm. This represents a substantial improvement over conventional co-registration, which exhibits mean errors of 214 mm (horizontal) and 295 mm (vertical) under identical conditions.</div><div>Our code is available at <span><span>https://github.com/meowxu/ECE-fitting</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"231 ","pages":"Pages 622-640"},"PeriodicalIF":12.2,"publicationDate":"2025-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145535926","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
UAV-based monocular 3D panoptic mapping for fruit shape completion in orchard 基于无人机的果园水果形状补全单眼三维全景映射
IF 12.2 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2025-11-14 DOI: 10.1016/j.isprsjprs.2025.11.013
Kaiwen Wang , Yue Pan , Federico Magistri , Lammert Kooistra , Cyrill Stachniss , Wensheng Wang , João Valente
Accurate fruit shape reconstruction under real-world field conditions is essential for high-throughput phenotyping, sensor-based yield estimation, and orchard management. Existing approaches based on 2D imaging or explicit 3D reconstruction often suffer from occlusions, sparse views, and complex scene dynamics as a result of the plant geometries. This paper presents a novel UAV-based monocular 3D panoptic mapping framework for robust and scalable fruit shape completion in orchards. The proposed method integrates (1) Grounded-SAM2 for multi-object tracking and segmentation (MOTS), (2) photogrammetric structure-from-motion for 3D scene reconstruction, and (3) DeepSDF, an implicit neural representation, for completing occluded fruit geometries with a neural network. We furthermore propose a new MOTS evaluation protocol to assess tracking performance without requiring ground truth annotations. Experiments conducted in both controlled laboratory conditions and an operational apple orchard demonstrate the accuracy of our 3D fruit reconstruction at the centimeter level. The Chamfer distance error of the proposed shape completion method using the DeepSDF shape prior reduces this to the millimeter level, and outperforms the traditional method, while Grounded-SAM2 enables robust fruit tracking across challenging viewpoints. The approach is highly scalable and applicable to real-world agricultural scenarios, offering a promising solution to reconstruct complete fruits with visibility higher than 10% for precise 3D fruit phenotyping at a large scale under occluded conditions.
在真实的田间条件下,精确的果实形状重建对于高通量表型分析、基于传感器的产量估计和果园管理至关重要。由于植物的几何形状,现有的基于2D成像或明确的3D重建的方法经常受到遮挡、稀疏视图和复杂场景动态的影响。本文提出了一种新的基于无人机的单眼三维全景映射框架,用于果园水果形状的鲁棒和可扩展完成。该方法集成了(1)基于ground - sam2的多目标跟踪与分割(MOTS),(2)基于运动的摄影测量结构(photogrammetric structure- frommotion),以及(3)基于隐式神经表示的DeepSDF,利用神经网络完成被遮挡水果的几何形状。我们进一步提出了一种新的MOTS评估协议来评估跟踪性能,而不需要地面真值注释。在受控的实验室条件和可操作的苹果园中进行的实验表明,我们的三维水果重建在厘米水平上的准确性。使用DeepSDF形状先验的形状补全方法的倒角距离误差将其降低到毫米级别,并且优于传统方法,而ground - sam2可以在具有挑战性的视点上进行稳健的水果跟踪。该方法具有高度可扩展性,适用于现实农业场景,为在闭塞条件下大规模重建完整水果的精确3D表型提供了一个有希望的解决方案,其可见度高于10%。
{"title":"UAV-based monocular 3D panoptic mapping for fruit shape completion in orchard","authors":"Kaiwen Wang ,&nbsp;Yue Pan ,&nbsp;Federico Magistri ,&nbsp;Lammert Kooistra ,&nbsp;Cyrill Stachniss ,&nbsp;Wensheng Wang ,&nbsp;João Valente","doi":"10.1016/j.isprsjprs.2025.11.013","DOIUrl":"10.1016/j.isprsjprs.2025.11.013","url":null,"abstract":"<div><div>Accurate fruit shape reconstruction under real-world field conditions is essential for high-throughput phenotyping, sensor-based yield estimation, and orchard management. Existing approaches based on 2D imaging or explicit 3D reconstruction often suffer from occlusions, sparse views, and complex scene dynamics as a result of the plant geometries. This paper presents a novel UAV-based monocular 3D panoptic mapping framework for robust and scalable fruit shape completion in orchards. The proposed method integrates (1) Grounded-SAM2 for multi-object tracking and segmentation (MOTS), (2) photogrammetric structure-from-motion for 3D scene reconstruction, and (3) DeepSDF, an implicit neural representation, for completing occluded fruit geometries with a neural network. We furthermore propose a new MOTS evaluation protocol to assess tracking performance without requiring ground truth annotations. Experiments conducted in both controlled laboratory conditions and an operational apple orchard demonstrate the accuracy of our 3D fruit reconstruction at the centimeter level. The Chamfer distance error of the proposed shape completion method using the DeepSDF shape prior reduces this to the millimeter level, and outperforms the traditional method, while Grounded-SAM2 enables robust fruit tracking across challenging viewpoints. The approach is highly scalable and applicable to real-world agricultural scenarios, offering a promising solution to reconstruct complete fruits with visibility higher than 10% for precise 3D fruit phenotyping at a large scale under occluded conditions.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"231 ","pages":"Pages 608-621"},"PeriodicalIF":12.2,"publicationDate":"2025-11-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145520460","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Mapping aboveground tree biomass and uncertainty using an upscaling approach: A case study of the larch forests in northeastern China using UAV laser scanning data 利用升级尺度方法绘制地表树木生物量和不确定性——以中国东北落叶松森林为例
IF 12.2 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2025-11-13 DOI: 10.1016/j.isprsjprs.2025.11.008
Yuanshuo Hao , Timo Pukkala , Xin Liu , Ying Quan , Lihu Dong , Fengri Li
Forest biomass mapping, monitoring and verification are vital to understand carbon cycling and mitigate climate change. Nevertheless, a remote sensing–based framework that upscales biomass estimation from the individual-tree level remains underdeveloped, especially for rigorously quantifying the propagation of associated uncertainties. This study proposed an upscaling framework for aboveground biomass (AGB) mapping and prediction uncertainty estimation using unmanned aerial vehicle laser scanning (UAVLS) data. Tree-level metrics were first generated from delineated point clouds. The per-tree AGB was predicted with a developed allometric equation and then upscaled for coarser-scale prediction. A case study was conducted on larch (Larix olgensis) forests in Northeast China. An AGB allometric equation for remote sensing application purposes was established, adopting the crown width (CW) and tree height (H) as predictors, based on 147 destructive tree samples. Tree- and plot-level (30 m × 30 m) AGB values were estimated with root-mean-squared differences (RMSDs) of 33.84 % and 10.74 %, respectively, relative to field-based AGB estimates, and the prediction accuracy improved as the estimation was aggregated from the tree to the plot scale. Furthermore, this study introduced an analytical framework to characterize the AGB prediction uncertainty considering error propagation throughout the whole upscaling workflow. At the tree level, the total uncertainty in AGB prediction was approximately 38.11 %. The errors associated with the UAVLS-measured CW and H contributed the most to the total uncertainty, at approximately 61.60 %, followed by allometry residual errors, which contributed approximately 37.71 % to the total uncertainty, while model parameters only contributed approximately 0.69 % to the total uncertainty. The per-plot error accounted for approximately 11.67 % of the estimated AGB, of which omission, commission, and aggregated tree-level errors accounted for approximately 78.97 %, 7.27 %, and 13.76 %, respectively, of the total variance and generally decreased for the plots with higher AGB. A simulation experiment revealed that the aggregated tree-level errors decreased the most with the spatial resolution over the other errors. This study not only contributes to the upscaling of AGB estimation using UAVLS data but also provides an intuitive framework for quantifying the associated uncertainties.
森林生物量制图、监测和验证对于了解碳循环和减缓气候变化至关重要。然而,从单个树的水平提高生物量估算的基于遥感的框架仍然不发达,特别是对相关不确定性传播的严格量化。本文提出了利用无人机激光扫描(UAVLS)数据进行地上生物量(AGB)制图和预测不确定性估算的升级框架。树级度量首先由划定的点云生成。利用发展的异速生长方程预测每棵树的AGB,然后进行更大尺度的预测。以东北地区落叶松(Larix olgensis)为例进行了研究。以147个破坏性树木样本为基础,以冠宽(CW)和树高(H)为预测因子,建立了遥感应用的AGB异速生长方程。相对于基于田间的AGB估计值,树级和样地级(30 m × 30 m) AGB估计值的均方根差(rmsd)分别为33.84%和10.74%,并且随着估计值从树级到样地尺度的汇总,预测精度有所提高。此外,本研究引入了一个分析框架来表征AGB预测的不确定性,考虑误差在整个升级工作流程中的传播。在树水平上,AGB预测的总不确定性约为38.11%。与uavls测量的CW和H相关的误差对总不确定度的贡献最大,约为61.60%,其次是异速残差,占总不确定度的约37.71%,而模型参数仅占总不确定度的约0.69%。每样地误差约占估计AGB的11.67%,其中遗漏、委托和汇总树级误差分别约占总方差的78.97%、7.27%和13.76%,并且在AGB较高的样地总体上减小。仿真实验表明,随着空间分辨率的提高,树级综合误差的减小幅度最大。该研究不仅有助于提高利用UAVLS数据估计AGB的尺度,而且为量化相关不确定性提供了一个直观的框架。
{"title":"Mapping aboveground tree biomass and uncertainty using an upscaling approach: A case study of the larch forests in northeastern China using UAV laser scanning data","authors":"Yuanshuo Hao ,&nbsp;Timo Pukkala ,&nbsp;Xin Liu ,&nbsp;Ying Quan ,&nbsp;Lihu Dong ,&nbsp;Fengri Li","doi":"10.1016/j.isprsjprs.2025.11.008","DOIUrl":"10.1016/j.isprsjprs.2025.11.008","url":null,"abstract":"<div><div>Forest biomass mapping, monitoring and verification are vital to understand carbon cycling and mitigate climate change. Nevertheless, a remote sensing–based framework that upscales biomass estimation from the individual-tree level remains underdeveloped, especially for rigorously quantifying the propagation of associated uncertainties. This study proposed an upscaling framework for aboveground biomass (AGB) mapping and prediction uncertainty estimation using unmanned aerial vehicle laser scanning (UAVLS) data. Tree-level metrics were first generated from delineated point clouds. The per-tree AGB was predicted with a developed allometric equation and then upscaled for coarser-scale prediction. A case study was conducted on larch (<em>Larix olgensis</em>) forests in Northeast China. An AGB allometric equation for remote sensing application purposes was established, adopting the crown width (CW) and tree height (H) as predictors, based on 147 destructive tree samples. Tree- and plot-level (30 m × 30 m) AGB values were estimated with root-mean-squared differences (RMSDs) of 33.84 % and 10.74 %, respectively, relative to field-based AGB estimates, and the prediction accuracy improved as the estimation was aggregated from the tree to the plot scale. Furthermore, this study introduced an analytical framework to characterize the AGB prediction uncertainty considering error propagation throughout the whole upscaling workflow. At the tree level, the total uncertainty in AGB prediction was approximately 38.11 %. The errors associated with the UAVLS-measured CW and H contributed the most to the total uncertainty, at approximately 61.60 %, followed by allometry residual errors, which contributed approximately 37.71 % to the total uncertainty, while model parameters only contributed approximately 0.69 % to the total uncertainty. The per-plot error accounted for approximately 11.67 % of the estimated AGB, of which omission, commission, and aggregated tree-level errors accounted for approximately 78.97 %, 7.27 %, and 13.76 %, respectively, of the total variance and generally decreased for the plots with higher AGB. A simulation experiment revealed that the aggregated tree-level errors decreased the most with the spatial resolution over the other errors. This study not only contributes to the upscaling of AGB estimation using UAVLS data but also provides an intuitive framework for quantifying the associated uncertainties.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"231 ","pages":"Pages 595-607"},"PeriodicalIF":12.2,"publicationDate":"2025-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145520459","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing snow depth estimation in forested regions of the northern hemisphere: A physically-constrained machine learning approach with spatiotemporal dynamics 增强北半球森林地区雪深估计:一种具有时空动态的物理约束机器学习方法
IF 12.2 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2025-11-13 DOI: 10.1016/j.isprsjprs.2025.11.006
Yanlin Wei , Xiaofeng Li , Lingjia Gu , Zhaojun Zheng , Yingjie Shi , YiJing Li , Xingming Zheng , Tao Jiang
Snow cover is a crucial factor in climate change and water resource management, particularly in forested regions, which affects vegetation carbon stocks by regulating local hydrological and temperature conditions. However, accurate snow depth (SD) estimation in forested regions remains significantly challenging because of the dual effects of forest canopy attenuation and radiation. To address this issue, a novel machine learning (ML) framework model by coupling microwave radiative transfer model (RTM) was proposed for SD retrieval in the forested regions of the Northern Hemisphere (NH), named the RTM-ML SD model. The physical mechanisms of forest–snow radiative transfer were incorporated into the ML model by introducing physically constrained brightness temperatures (TBs) that capture the radiative contributions of snow cover and forest canopy within the satellite field of view. Additionally, the spatiotemporal dynamic modeling strategy was implemented to ensure RTM-ML SD model flexibility and stability across different regions and seasons. Independent validations revealed that the retrieved SDs suitably agreed with in situ ground observations, with high correlation (R = 0.88) and low uncertainties (RMSE = 14.98 cm, MAE = 8.44 cm, and bias =  − 1.46 cm), which are 8.0 % and 6.1 % lower, respectively, than those without introducing the RTM and spatiotemporal dynamic modeling strategy. Compared with that of well-known SD products and inversion algorithms (Chang, AMSR2, ERA5-Land, and Globsnow), the SD inversion accuracy increased by more than 50 %, effectively mitigating SD underestimation in forested regions. Overall, we conclude that 1) the impact of forest canopies on SD retrieval is complex, and the physical knowledge of forest–snow radiative transfer should be considered in SD estimation, and 2) the RTM-ML SD model exhibits notable temporal and spatial generalizability in the forested regions of the NH, thereby enhancing its capacity to support large-scale environmental monitoring, enable more reliable seasonal water resource forecasting, and inform climate adaptation strategies in snow-dependent regions.
积雪是气候变化和水资源管理的一个关键因素,特别是在森林地区,它通过调节当地水文和温度条件影响植被碳储量。然而,由于森林冠层衰减和辐射的双重影响,森林地区准确的雪深估算仍然具有很大的挑战性。为了解决这一问题,提出了一种基于耦合微波辐射传输模型(RTM)的机器学习(ML)框架模型,并将其命名为RTM-ML SD模型。通过引入捕获卫星视场内积雪和森林冠层辐射贡献的物理约束亮度温度(TBs),将森林-雪辐射转移的物理机制纳入ML模型。此外,采用时空动态建模策略,确保RTM-ML SD模型在不同地区和季节的灵活性和稳定性。独立验证结果表明,与未引入RTM和时空动态建模策略的结果相比,反演的SDs与地面观测结果吻合较好,具有较高的相关性(R = 0.88)和较低的不确定性(RMSE = 14.98 cm, MAE = 8.44 cm,偏差= - 1.46 cm),分别降低了8.0%和6.1%。与常用的SD产品和反演算法(Chang、AMSR2、ERA5-Land、Globsnow)相比,反演精度提高50%以上,有效缓解了森林地区的SD低估。结果表明:1)森林冠层对SD反演的影响较为复杂,在进行SD估算时应考虑森林-雪辐射转移的物理知识;2)RTM-ML SD模型在NH林区具有显著的时空泛化性,从而增强了其支持大尺度环境监测的能力,使季节性水资源预测更加可靠;并为依赖雪的地区的气候适应战略提供信息。
{"title":"Enhancing snow depth estimation in forested regions of the northern hemisphere: A physically-constrained machine learning approach with spatiotemporal dynamics","authors":"Yanlin Wei ,&nbsp;Xiaofeng Li ,&nbsp;Lingjia Gu ,&nbsp;Zhaojun Zheng ,&nbsp;Yingjie Shi ,&nbsp;YiJing Li ,&nbsp;Xingming Zheng ,&nbsp;Tao Jiang","doi":"10.1016/j.isprsjprs.2025.11.006","DOIUrl":"10.1016/j.isprsjprs.2025.11.006","url":null,"abstract":"<div><div>Snow cover is a crucial factor in climate change and water resource management, particularly in forested regions, which affects vegetation carbon stocks by regulating local hydrological and temperature conditions. However, accurate snow depth (SD) estimation in forested regions remains significantly challenging because of the dual effects of forest canopy attenuation and radiation. To address this issue, a novel machine learning (ML) framework model by coupling microwave radiative transfer model (RTM) was proposed for SD retrieval in the forested regions of the Northern Hemisphere (NH), named the RTM-ML SD model. The physical mechanisms of forest–snow radiative transfer were incorporated into the ML model by introducing physically constrained brightness temperatures (TBs) that capture the radiative contributions of snow cover and forest canopy within the satellite field of view. Additionally, the spatiotemporal dynamic modeling strategy was implemented to ensure RTM-ML SD model flexibility and stability across different regions and seasons. Independent validations revealed that the retrieved SDs suitably agreed with in situ ground observations, with high correlation (<em>R =</em> 0.88) and low uncertainties (<em>RMSE =</em> 14.98 cm<em>, MAE =</em> 8.44 cm<em>,</em> and bias <em>=  −</em> 1.46 cm), which are 8.0 % and 6.1 % lower, respectively, than those without introducing the RTM and spatiotemporal dynamic modeling strategy. Compared with that of well-known SD products and inversion algorithms (Chang, AMSR2, ERA5-Land, and Globsnow), the SD inversion accuracy increased by more than 50 %, effectively mitigating SD underestimation in forested regions. Overall, we conclude that 1) the impact of forest canopies on SD retrieval is complex, and the physical knowledge of forest–snow radiative transfer should be considered in SD estimation, and 2) the RTM-ML SD model exhibits notable temporal and spatial generalizability in the forested regions of the NH, thereby enhancing its capacity to support large-scale environmental monitoring, enable more reliable seasonal water resource forecasting, and inform climate adaptation strategies in snow-dependent regions.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"231 ","pages":"Pages 576-594"},"PeriodicalIF":12.2,"publicationDate":"2025-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145520457","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Compact-pol to quad-pol SAR reconstruction via a joint mathematical-physical-constrained multimodal correlation-preserving latent diffusion framework 基于数学-物理约束多模态保持相关的潜在扩散框架的紧凑极点到四极点SAR重建
IF 12.2 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2025-11-13 DOI: 10.1016/j.isprsjprs.2025.11.005
Zihuan Guo , Hong Zhang , Ji Ge , Xiao-Ming Li , Peifeng Ma , Haoxuan Duan , Lu Xu , Chao Wang
Compact-polarimetric (CP) synthetic aperture radar (SAR) offers wide swath coverage and relatively rich polarimetric information, making it a promising mode of earth observation. However, compared to quad-polarimetric (QP) data, CP data are still limited in fully characterizing polarimetric scattering mechanisms. Existing methods for reconstructing QP data from CP data face challenges such as the lack of physical properties in the QP covariance matrix and insufficient utilization of polarimetric, terrain, and imaging information. To address these limitations, this paper proposes a multimodal correlation-preserving latent diffusion framework under mathematical-physical constraints for CP to QP SAR data reconstruction (SAR-C2QM). First, we propose a mathematical-physical hard constraint through the Cholesky decomposition method and a two-dimensional phase vector encoding approach, ensuring both the positive semi-definiteness and phase continuity of the reconstructed QP data. Next, SAR image modality and corresponding terrain-imaging information modality are constructed based on multi-source data, providing rich polarimetric and effective terrain-imaging multimodal information for model training. Furthermore, a multimodal correlation-preserving latent diffusion model with a hierarchical bidirectional cross-attention module is designed, enabling the fusion of multimodal and multi-scale features, with mathematical-physical constraints guiding the reconstruction process to achieve high-fidelity QP reconstruction across large-area coverage and complex terrain conditions. Experiments on simulated CP data derived from Gaofen-3 and Radarsat-2 QP data across diverse wave codes demonstrate that the proposed method outperforms existing methods in both accuracy and stability. Specifically, the Wishart distance error between reconstructed and original QP data is reduced by over 12.1%. Moreover, the overall accuracy of land cover classification using the reconstructed QP data improves by an average of 15.1% compared to CP data, closely approaching the classification accuracy of the original QP data. The code is available in https://github.com/zihuan21/SAR-C2QM.
紧凑偏振合成孔径雷达(SAR)具有较宽的波段覆盖范围和相对丰富的偏振信息,是一种很有前途的对地观测方式。然而,与四极化(QP)数据相比,CP数据在充分表征极化散射机制方面仍然受到限制。现有的从CP数据重建QP数据的方法面临着诸如缺乏QP协方差矩阵的物理性质以及极化、地形和成像信息的充分利用等挑战。为了解决这些问题,本文提出了一种基于数学-物理约束的多模态保持相关的潜在扩散框架,用于CP到QP SAR数据重构(SAR- c2qm)。首先,我们通过Cholesky分解方法和二维相位矢量编码方法提出了数学物理硬约束,保证了重构的QP数据的正半确定性和相位连续性。其次,基于多源数据构建SAR图像模态和相应的地形成像信息模态,为模型训练提供丰富的极化有效的地形成像多模态信息。此外,设计了具有分层双向交叉关注模块的多模态保持相关的潜在扩散模型,实现了多模态和多尺度特征的融合,以数学物理约束指导重建过程,实现了大范围覆盖和复杂地形条件下的高保真QP重建。对高分三号和Radarsat-2 QP数据在不同波码下的模拟CP数据进行的实验表明,该方法在精度和稳定性方面都优于现有方法。具体而言,重构数据与原始QP数据之间的Wishart距离误差降低了12.1%以上。与CP数据相比,重构QP数据的土地覆盖分类总体精度平均提高了15.1%,接近原始QP数据的分类精度。该代码可在https://github.com/zihuan21/SAR-C2QM中获得。
{"title":"Compact-pol to quad-pol SAR reconstruction via a joint mathematical-physical-constrained multimodal correlation-preserving latent diffusion framework","authors":"Zihuan Guo ,&nbsp;Hong Zhang ,&nbsp;Ji Ge ,&nbsp;Xiao-Ming Li ,&nbsp;Peifeng Ma ,&nbsp;Haoxuan Duan ,&nbsp;Lu Xu ,&nbsp;Chao Wang","doi":"10.1016/j.isprsjprs.2025.11.005","DOIUrl":"10.1016/j.isprsjprs.2025.11.005","url":null,"abstract":"<div><div>Compact-polarimetric (CP) synthetic aperture radar (SAR) offers wide swath coverage and relatively rich polarimetric information, making it a promising mode of earth observation. However, compared to quad-polarimetric (QP) data, CP data are still limited in fully characterizing polarimetric scattering mechanisms. Existing methods for reconstructing QP data from CP data face challenges such as the lack of physical properties in the QP covariance matrix and insufficient utilization of polarimetric, terrain, and imaging information. To address these limitations, this paper proposes a multimodal correlation-preserving latent diffusion framework under mathematical-physical constraints for CP to QP SAR data reconstruction (SAR-C2QM). First, we propose a mathematical-physical hard constraint through the Cholesky decomposition method and a two-dimensional phase vector encoding approach, ensuring both the positive semi-definiteness and phase continuity of the reconstructed QP data. Next, SAR image modality and corresponding terrain-imaging information modality are constructed based on multi-source data, providing rich polarimetric and effective terrain-imaging multimodal information for model training. Furthermore, a multimodal correlation-preserving latent diffusion model with a hierarchical bidirectional cross-attention module is designed, enabling the fusion of multimodal and multi-scale features, with mathematical-physical constraints guiding the reconstruction process to achieve high-fidelity QP reconstruction across large-area coverage and complex terrain conditions. Experiments on simulated CP data derived from Gaofen-3 and Radarsat-2 QP data across diverse wave codes demonstrate that the proposed method outperforms existing methods in both accuracy and stability. Specifically, the Wishart distance error between reconstructed and original QP data is reduced by over 12.1%. Moreover, the overall accuracy of land cover classification using the reconstructed QP data improves by an average of 15.1% compared to CP data, closely approaching the classification accuracy of the original QP data. The code is available in <span><span>https://github.com/zihuan21/SAR-C2QM</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"231 ","pages":"Pages 552-575"},"PeriodicalIF":12.2,"publicationDate":"2025-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145520458","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
High-precision flood change detection with lightweight SAR transformer network and context-aware attention for enriched-diverse and complex flooding scenarios 基于轻型SAR变压器网络的高精度洪水变化检测和上下文感知关注,用于丰富多样和复杂的洪水场景
IF 12.2 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2025-11-12 DOI: 10.1016/j.isprsjprs.2025.11.011
Menghao Du, Zhenfeng Shao, Xiongwu Xiao, Jindou Zhang, Duowang Zhu, Jinyang Wang, Timo Balz, Deren Li
<div><div>Floods are highly destructive natural disasters that threaten both society and the environment. Given the all-weather, all-time imaging capability of synthetic aperture radar (SAR), analyzing flood events using SAR imagery across diverse scenarios is essential for developing high-precision and robust detection models. However, existing transformer-based change detection methods achieve high precision, but their high computational cost and large parameter sizes necessitate lightweight design while maintaining detection accuracy. Moreover, existing studies focus on a few specific scenarios without thorough validation and in-depth analysis of model strengths across diverse flood conditions with imbalanced inundation ratios. To address these challenges, this paper proposes an adaptive window and context-aware attention network (AWCA-Net) for high-precision SAR-based flood change detection under diverse flooding scenarios, achieving a lightweight model while maintaining the highest detection accuracy. AWCA-Net has three key advantages: Firstly, the neighborhood feature enhancement module with contextual information (NECM) strengthens the discrimination of subtle and heterogeneous flood changes. Secondly, the large kernel grouping attention gate module based on high-low layer feature difference (LGDM) leverages difference-weighted attention to effectively guide the selection of flood-relevant features. Thirdly, the multi-scale convolutional attention module based on adaptive window selection (MSAWM) dynamically adjusts kernel sizes to capture diverse flood change patterns. To better train and evaluate AWCA-Net, we constructed VarFloods, the first large-scale and enriched-diverse benchmark dataset for flood change detection that spans five continents and includes diverse regions, scenarios, land cover types, causes, years, and inundation ratios, which includes both GRD and preprocessed versions. We evaluated AWCA-Net’s performance on three datasets. We found that: (1) AWCA-Net achieves the highest-precision while maintaining significantly lower computational cost, outperforming other state-of-the-art (SOTA) methods. On the two representative public datasets and the enriched-diverse benchmark dataset (VarFloods-G and VarFloods-P), AWCA-Net improves the IoU by 11.59 % to 42.57 % over the basic model, 2.16 % to 11.48 % over an advanced transformer-based model, and 1.31 % to 1.68 % over the best comparative model, while maintaining a computational cost of only 17.53G, which is just 8.7 % to 60 % of existing SOTA models. (2) Difference-guided attention enhances detection in complex background regions, neighborhood-based fusion improves performance in irregular terrains, and adaptive convolution contributes to stable results across diverse flood scenarios. And the proposed AWCA-Net demonstrates strong generalization and stability under diverse flood scenarios with imbalanced inundation ratios. The dataset and code of AWCA-Net will be released at: <s
洪水是极具破坏性的自然灾害,对社会和环境都构成威胁。考虑到合成孔径雷达(SAR)的全天候、全天候成像能力,利用不同场景的SAR图像分析洪水事件对于开发高精度、鲁棒的探测模型至关重要。然而,现有的基于变压器的变化检测方法虽然精度较高,但计算成本高,参数尺寸大,需要在保持检测精度的同时进行轻量化设计。此外,现有的研究主要集中在几个特定的场景上,没有对不同淹没比不平衡洪水条件下的模型优势进行彻底的验证和深入的分析。为了解决这些问题,本文提出了一种自适应窗口和上下文感知关注网络(AWCA-Net),用于在不同洪水场景下基于sar的高精度洪水变化检测,在保持最高检测精度的同时实现了轻量级模型。AWCA-Net具有三个关键优势:首先,基于上下文信息的邻域特征增强模块(NECM)增强了对洪水微妙和非均质变化的识别;其次,基于高低层特征差(LGDM)的大核分组注意门模块利用差分加权注意有效指导洪水相关特征的选择。第三,基于自适应窗口选择(MSAWM)的多尺度卷积关注模块动态调整核大小,捕捉不同的洪水变化模式。为了更好地训练和评估AWCA-Net,我们构建了varflood,这是第一个大规模和丰富多样的洪水变化检测基准数据集,跨越五大洲,包括不同的区域、场景、土地覆盖类型、原因、年份和淹没比率,包括GRD和预处理版本。我们在三个数据集上评估了AWCA-Net的性能。我们发现:(1)AWCA-Net在保持较低的计算成本的同时实现了最高的精度,优于其他最先进的(SOTA)方法。在两个具有代表性的公共数据集和丰富的基准数据集(VarFloods-G和VarFloods-P)上,AWCA-Net将IoU比基本模型提高了11.59%至42.57%,比先进的基于变压器的模型提高了2.16%至11.48%,比最佳比较模型提高了1.31%至1.68%,同时保持了17.53 3g的计算成本,仅为现有SOTA模型的8.7%至60%。(2)差分引导的注意力增强了复杂背景区域的检测能力,基于邻域的融合提高了不规则地形的检测性能,自适应卷积有助于在不同洪水场景下获得稳定的结果。在不同淹没比不平衡的洪水情景下,AWCA-Net具有较强的通用性和稳定性。AWCA-Net的数据集和代码将在https://github.com/Dumh1998/AWCA-Net上发布。
{"title":"High-precision flood change detection with lightweight SAR transformer network and context-aware attention for enriched-diverse and complex flooding scenarios","authors":"Menghao Du,&nbsp;Zhenfeng Shao,&nbsp;Xiongwu Xiao,&nbsp;Jindou Zhang,&nbsp;Duowang Zhu,&nbsp;Jinyang Wang,&nbsp;Timo Balz,&nbsp;Deren Li","doi":"10.1016/j.isprsjprs.2025.11.011","DOIUrl":"10.1016/j.isprsjprs.2025.11.011","url":null,"abstract":"&lt;div&gt;&lt;div&gt;Floods are highly destructive natural disasters that threaten both society and the environment. Given the all-weather, all-time imaging capability of synthetic aperture radar (SAR), analyzing flood events using SAR imagery across diverse scenarios is essential for developing high-precision and robust detection models. However, existing transformer-based change detection methods achieve high precision, but their high computational cost and large parameter sizes necessitate lightweight design while maintaining detection accuracy. Moreover, existing studies focus on a few specific scenarios without thorough validation and in-depth analysis of model strengths across diverse flood conditions with imbalanced inundation ratios. To address these challenges, this paper proposes an adaptive window and context-aware attention network (AWCA-Net) for high-precision SAR-based flood change detection under diverse flooding scenarios, achieving a lightweight model while maintaining the highest detection accuracy. AWCA-Net has three key advantages: Firstly, the neighborhood feature enhancement module with contextual information (NECM) strengthens the discrimination of subtle and heterogeneous flood changes. Secondly, the large kernel grouping attention gate module based on high-low layer feature difference (LGDM) leverages difference-weighted attention to effectively guide the selection of flood-relevant features. Thirdly, the multi-scale convolutional attention module based on adaptive window selection (MSAWM) dynamically adjusts kernel sizes to capture diverse flood change patterns. To better train and evaluate AWCA-Net, we constructed VarFloods, the first large-scale and enriched-diverse benchmark dataset for flood change detection that spans five continents and includes diverse regions, scenarios, land cover types, causes, years, and inundation ratios, which includes both GRD and preprocessed versions. We evaluated AWCA-Net’s performance on three datasets. We found that: (1) AWCA-Net achieves the highest-precision while maintaining significantly lower computational cost, outperforming other state-of-the-art (SOTA) methods. On the two representative public datasets and the enriched-diverse benchmark dataset (VarFloods-G and VarFloods-P), AWCA-Net improves the IoU by 11.59 % to 42.57 % over the basic model, 2.16 % to 11.48 % over an advanced transformer-based model, and 1.31 % to 1.68 % over the best comparative model, while maintaining a computational cost of only 17.53G, which is just 8.7 % to 60 % of existing SOTA models. (2) Difference-guided attention enhances detection in complex background regions, neighborhood-based fusion improves performance in irregular terrains, and adaptive convolution contributes to stable results across diverse flood scenarios. And the proposed AWCA-Net demonstrates strong generalization and stability under diverse flood scenarios with imbalanced inundation ratios. The dataset and code of AWCA-Net will be released at: &lt;s","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"231 ","pages":"Pages 507-531"},"PeriodicalIF":12.2,"publicationDate":"2025-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145509781","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Toward noise-resilient retrieval of land surface temperature and emissivity using airborne thermal infrared hyperspectral imagery 基于机载热红外高光谱图像的地表温度和发射率的抗噪反演研究
IF 12.2 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2025-11-12 DOI: 10.1016/j.isprsjprs.2025.10.039
Du Wang , Li-Qin Cao , Yu-Hao Du , Hai-Yang Xiong , Fa-Wang Ye , Yan-Fei Zhong
As thermal infrared remote sensing advances toward higher spectral and spatial resolutions, the need for precise surface parameters is becoming increasingly important for earth observation applications. However, effective retrieval remains fundamentally challenging due to degraded spectral quality caused by narrow bandwidths of thermal infrared hyperspectral imagers, atmospheric line absorption interference, and limitations in sensor manufacturing. To address this, this study introduces a Noise-Resilient Atmospheric Compensation with Temperature and Emissivity Separation (NRAC-TES) method, where the noise-resistant capability is mainly achieved through the NRAC module during the atmospheric compensation (AC) stage. The robust retrieval strategy uses data-model correlations constrained by atmospheric parameter features, enabling adaptive adjustment of model parameters to match real-world noisy data and reducing biases caused by spectral noise in at-sensor radiance. The temperature and emissivity separation of ASTER-TES incorporates an atmospheric downwelling radiance lookup table, allowing direct retrieval of land surface temperature and emissivity without needing prior surface-atmospheric information. Ground validation experiments on Hypercam-LW demonstrate that the proposed NRAC-TES achieves an AC accuracy of 1.2 K for ground-leaving radiance, with corresponding errors of 1.2 K and 0.025 for LST and LSE, respectively. Additionally, comparison with traditional methods like MODTRAN-TES, ICCAAC-TES, and FSW-TES, based on airborne datasets from HyTES and TASI, highlights the importance and effectiveness of the noise-resilient design incorporated in our approach.
随着热红外遥感技术向更高的光谱和空间分辨率发展,对精确地表参数的需求在地球观测应用中变得越来越重要。然而,由于热红外高光谱成像仪的窄带带宽、大气线吸收干扰以及传感器制造的限制导致光谱质量下降,有效的检索仍然具有根本性的挑战性。为了解决这一问题,本研究引入了一种具有温度和发射率分离的抗噪声大气补偿(NRAC- tes)方法,该方法主要通过NRAC模块在大气补偿(AC)阶段实现抗噪声能力。鲁棒检索策略利用受大气参数特征约束的数据模型相关性,实现模型参数的自适应调整,以匹配真实世界的噪声数据,并减少由红外传感器辐射光谱噪声引起的偏差。ASTER-TES的温度和发射率分离包含一个大气下井辐射查找表,允许直接检索地表温度和发射率,而不需要事先的地表大气信息。在Hypercam-LW上进行的地面验证实验表明,NRAC-TES的离地辐射精度为1.2 K, LST和LSE的误差分别为1.2 K和0.025。此外,基于HyTES和TASI的机载数据集,与modtrans - tes、ICCAAC-TES和FSW-TES等传统方法进行比较,突出了我们方法中抗噪设计的重要性和有效性。
{"title":"Toward noise-resilient retrieval of land surface temperature and emissivity using airborne thermal infrared hyperspectral imagery","authors":"Du Wang ,&nbsp;Li-Qin Cao ,&nbsp;Yu-Hao Du ,&nbsp;Hai-Yang Xiong ,&nbsp;Fa-Wang Ye ,&nbsp;Yan-Fei Zhong","doi":"10.1016/j.isprsjprs.2025.10.039","DOIUrl":"10.1016/j.isprsjprs.2025.10.039","url":null,"abstract":"<div><div>As thermal infrared remote sensing advances toward higher spectral and spatial resolutions, the need for precise surface parameters is becoming increasingly important for earth observation applications. However, effective retrieval remains fundamentally challenging due to degraded spectral quality caused by narrow bandwidths of thermal infrared hyperspectral imagers, atmospheric line absorption interference, and limitations in sensor manufacturing. To address this, this study introduces a Noise-Resilient Atmospheric Compensation with Temperature and Emissivity Separation (NRAC-TES) method, where the noise-resistant capability is mainly achieved through the NRAC module during the atmospheric compensation (AC) stage. The robust retrieval strategy uses data-model correlations constrained by atmospheric parameter features, enabling adaptive adjustment of model parameters to match real-world noisy data and reducing biases caused by spectral noise in at-sensor radiance. The temperature and emissivity separation of ASTER-TES incorporates an atmospheric downwelling radiance lookup table, allowing direct retrieval of land surface temperature and emissivity without needing prior surface-atmospheric information. Ground validation experiments on Hypercam-LW demonstrate that the proposed NRAC-TES achieves an AC accuracy of 1.2 K for ground-leaving radiance, with corresponding errors of 1.2 K and 0.025 for LST and LSE, respectively. Additionally, comparison with traditional methods like MODTRAN-TES, ICCAAC-TES, and FSW-TES, based on airborne datasets from HyTES and TASI, highlights the importance and effectiveness of the noise-resilient design incorporated in our approach.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"231 ","pages":"Pages 532-551"},"PeriodicalIF":12.2,"publicationDate":"2025-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145520462","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Identifying rural roads in remote sensing imagery: From benchmark dataset to coarse-to-fine extraction network—A case study in China 遥感影像中的农村道路识别:从基准数据集到粗到精提取网络——以中国为例
IF 12.2 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2025-11-11 DOI: 10.1016/j.isprsjprs.2025.11.010
Ningjing Wang , Xinyu Wang , Yang Pan , Lei Lei , Wanqiang Yao , Yanfei Zhong
Rural roads are vital infrastructure for promoting agricultural modernization and rural revitalization, and their accurate and efficient extraction is of great significance. However, rural road extraction remains highly challenging, as these roads are typically extremely narrow and often span only a few pixels in high-resolution imagery, which easily leads to attenuation or loss of critical pixels during deep feature encoding. In addition, low contrast, weak textures, and spectral similarity to the surrounding environment further reduce their distinguishability. The combined effect makes extracting narrow and spectrally similar rural rsoads particularly difficult, compromising both the topological connectivity and structural integrity of road networks. To address these challenges, we propose the Narrow-Road-Aware Network (NANet), a coarse-to-fine two-stage framework. In the positioning stage, NANet effectively strengthens weak road responses at multiple scales while preserving global consistency. In the refinement stage, a foreground-background collaborative mechanism enhances road saliency and suppresses background interference, thereby significantly improving the continuity and completeness of rural road extraction. In addition, we construct the WHU-CR dataset, a large-scale benchmark specifically designed for rural road extraction in China, which covers 14 provinces across China’s seven major grain-producing regions and contains 130,412 pairs of 512 × 512 images with annotated masks, providing a solid foundation for effectively developing and evaluating deep learning models for rural road extraction. Experimental results on the WHU-CR and DeepGlobe datasets demonstrate that NANet outperforms state-of-the-art methods. In large-scale rural road mapping, NANet exhibits strong practical applicability and generalization. The dataset and code will be available at: https://rsidea.whu.edu.cn/resource_sharing.htm.
乡村公路是推进农业现代化和乡村振兴的重要基础设施,其准确高效的提取具有重要意义。然而,农村道路提取仍然具有很高的挑战性,因为这些道路通常非常狭窄,并且在高分辨率图像中通常只跨越几个像素,这很容易导致深度特征编码过程中关键像素的衰减或丢失。此外,低对比度、弱纹理以及与周围环境的光谱相似性进一步降低了它们的可识别性。综合效应使得提取狭窄和频谱相似的农村道路特别困难,损害了道路网络的拓扑连通性和结构完整性。为了应对这些挑战,我们提出了窄路感知网络(NANet),这是一个从粗到精的两阶段框架。在定位阶段,NANet在保持全局一致性的同时,有效地增强了多尺度的弱道路响应。在细化阶段,前景-背景协同机制增强了道路显著性,抑制了背景干扰,显著提高了农村道路提取的连续性和完整性。此外,我们构建了中国农村道路提取的大规模基准数据集WHU-CR,该数据集涵盖中国七大粮食主产区的14个省份,包含130,412对512 × 512带注释掩码的图像,为有效开发和评估农村道路提取的深度学习模型提供了坚实的基础。在WHU-CR和DeepGlobe数据集上的实验结果表明,NANet优于最先进的方法。在大尺度农村道路制图中,NANet具有很强的实用性和通用性。数据集和代码可在https://rsidea.whu.edu.cn/resource_sharing.htm上获得。
{"title":"Identifying rural roads in remote sensing imagery: From benchmark dataset to coarse-to-fine extraction network—A case study in China","authors":"Ningjing Wang ,&nbsp;Xinyu Wang ,&nbsp;Yang Pan ,&nbsp;Lei Lei ,&nbsp;Wanqiang Yao ,&nbsp;Yanfei Zhong","doi":"10.1016/j.isprsjprs.2025.11.010","DOIUrl":"10.1016/j.isprsjprs.2025.11.010","url":null,"abstract":"<div><div>Rural roads are vital infrastructure for promoting agricultural modernization and rural revitalization, and their accurate and efficient extraction is of great significance. However, rural road extraction remains highly challenging, as these roads are typically extremely narrow and often span only a few pixels in high-resolution imagery, which easily leads to attenuation or loss of critical pixels during deep feature encoding. In addition, low contrast, weak textures, and spectral similarity to the surrounding environment further reduce their distinguishability. The combined effect makes extracting narrow and spectrally similar rural rsoads particularly difficult, compromising both the topological connectivity and structural integrity of road networks. To address these challenges, we propose the Narrow-Road-Aware Network (NANet), a coarse-to-fine two-stage framework. In the positioning stage, NANet effectively strengthens weak road responses at multiple scales while preserving global consistency. In the refinement stage, a foreground-background collaborative mechanism enhances road saliency and suppresses background interference, thereby significantly improving the continuity and completeness of rural road extraction. In addition, we construct the WHU-CR dataset, a large-scale benchmark specifically designed for rural road extraction in China, which covers 14 provinces across China’s seven major grain-producing regions and contains 130,412 pairs of 512 × 512 images with annotated masks, providing a solid foundation for effectively developing and evaluating deep learning models for rural road extraction. Experimental results on the WHU-CR and DeepGlobe datasets demonstrate that NANet outperforms state-of-the-art methods. In large-scale rural road mapping, NANet exhibits strong practical applicability and generalization. The dataset and code will be available at: <span><span>https://rsidea.whu.edu.cn/resource_sharing.htm</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"231 ","pages":"Pages 487-506"},"PeriodicalIF":12.2,"publicationDate":"2025-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145509819","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Bias-aware learning for unbiased scene graph generation in remote sensing imagery 遥感图像中无偏场景图生成的偏差感知学习
IF 12.2 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL Pub Date : 2025-11-11 DOI: 10.1016/j.isprsjprs.2025.11.009
Tingzhu Wang, Linlin Wang, Junwei Luo, Kang Wu, Yansheng Li
Scene graph generation (SGG) in remote sensing imagery (RSI) plays a critical role in advancing the understanding of geospatial scenarios. The incorporation of scene graphs, which capture the relationships between objects within imagery, has significantly enhanced the performance of various RS tasks, such as RS image retrieval and RS image captioning. However, the long-tailed distribution of relationships has adversely affected the SGG quality, leading to biased relationship predictions. In our previous Sample-Level Bias Prediction (SBP) method, we find that the union region of one object pair (i.e., one sample) contains rich and dedicated contextual information, enabling the prediction of the sample-specific bias for correcting the biased relationship prediction. Nevertheless, SBP employs a single model to correct all categories, which results in the suppression of certain relationship categories while enhancing the detection accuracy of others. To address this challenge, we present a Bias-Aware Learning (BAL) method that employs group collaboration to alleviate the burden of a single model correcting all categories. During BAL, we initially group the relationship categories and deploy group-specific SBP models to learn and predict the corresponding biases within their relationships. Subsequently, we design a Group-Collaborative Distillation (GCD) strategy to facilitate the collaboration among SBP models from different groups, thereby alleviating category suppression and enhancing the overall performance of SGG. The extensive experimental results on STAR and AUG datasets for RSI demonstrate that our BAL outperforms the state-of-the-art methods. Compared to the best model RPCM on STAR, RPCM equipped with BAL shows a significant average improvement of 5.08%/5.00%, 3.87%/3.64% and 0.22%/0.25% on HMR@K for PredCls, SGCls, and SGDet tasks, respectively. Moreover, we have also validated the effectiveness and generalization of our BAL on VG dataset for natural imagery. The code will be available at https://github.com/Zhuzi24/BAL.
遥感影像(RSI)中的场景图生成(SGG)对于提高对地理空间场景的理解起着至关重要的作用。场景图捕获了图像中物体之间的关系,其结合显著提高了各种遥感任务的性能,如遥感图像检索和遥感图像字幕。然而,关系的长尾分布对SGG质量产生了不利影响,导致关系预测有偏差。在我们之前的样本水平偏差预测(SBP)方法中,我们发现一个对象对(即一个样本)的联合区域包含丰富且专用的上下文信息,从而能够预测样本特定偏差以纠正偏差关系预测。然而,SBP采用单一模型来纠正所有类别,这导致某些关系类别被抑制,而其他关系类别的检测精度得到提高。为了解决这一挑战,我们提出了一种偏见感知学习(BAL)方法,该方法采用小组协作来减轻单个模型纠正所有类别的负担。在BAL过程中,我们首先对关系类别进行分组,并部署特定于组的SBP模型来学习和预测其关系中相应的偏差。随后,我们设计了一种群体协同蒸馏(GCD)策略,促进不同群体的SBP模型之间的协作,从而减轻类别抑制,提高SBP模型的整体性能。在STAR和AUG数据集上的广泛实验结果表明,我们的BAL优于最先进的方法。与STAR上的最佳模型RPCM相比,配备BAL的RPCM在HMR@K上对PredCls、SGCls和SGDet任务的平均效率分别提高了5.08%/5.00%、3.87%/3.64%和0.22%/0.25%。此外,我们还验证了我们的BAL在VG数据集上的有效性和泛化性。代码可在https://github.com/Zhuzi24/BAL上获得。
{"title":"Bias-aware learning for unbiased scene graph generation in remote sensing imagery","authors":"Tingzhu Wang,&nbsp;Linlin Wang,&nbsp;Junwei Luo,&nbsp;Kang Wu,&nbsp;Yansheng Li","doi":"10.1016/j.isprsjprs.2025.11.009","DOIUrl":"10.1016/j.isprsjprs.2025.11.009","url":null,"abstract":"<div><div>Scene graph generation (SGG) in remote sensing imagery (RSI) plays a critical role in advancing the understanding of geospatial scenarios. The incorporation of scene graphs, which capture the relationships between objects within imagery, has significantly enhanced the performance of various RS tasks, such as RS image retrieval and RS image captioning. However, the long-tailed distribution of relationships has adversely affected the SGG quality, leading to biased relationship predictions. In our previous Sample-Level Bias Prediction (SBP) method, we find that the union region of one object pair (i.e., one sample) contains rich and dedicated contextual information, enabling the prediction of the sample-specific bias for correcting the biased relationship prediction. Nevertheless, SBP employs a single model to correct all categories, which results in the suppression of certain relationship categories while enhancing the detection accuracy of others. To address this challenge, we present a Bias-Aware Learning (BAL) method that employs group collaboration to alleviate the burden of a single model correcting all categories. During BAL, we initially group the relationship categories and deploy group-specific SBP models to learn and predict the corresponding biases within their relationships. Subsequently, we design a Group-Collaborative Distillation (GCD) strategy to facilitate the collaboration among SBP models from different groups, thereby alleviating category suppression and enhancing the overall performance of SGG. The extensive experimental results on STAR and AUG datasets for RSI demonstrate that our BAL outperforms the state-of-the-art methods. Compared to the best model RPCM on STAR, RPCM equipped with BAL shows a significant average improvement of 5.08%/5.00%, 3.87%/3.64% and 0.22%/0.25% on HMR@K for PredCls, SGCls, and SGDet tasks, respectively. Moreover, we have also validated the effectiveness and generalization of our BAL on VG dataset for natural imagery. The code will be available at <span><span>https://github.com/Zhuzi24/BAL</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"231 ","pages":"Pages 473-486"},"PeriodicalIF":12.2,"publicationDate":"2025-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145499145","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
ISPRS Journal of Photogrammetry and Remote Sensing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1