首页 > 最新文献

PFG-Journal of Photogrammetry Remote Sensing and Geoinformation Science最新文献

英文 中文
Self-Supervised 3D Semantic Occupancy Prediction from Multi-View 2D Surround Images 根据多视角二维环绕图像进行自监督三维语义占用预测
IF 4.1 4区 地球科学 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2024-09-18 DOI: 10.1007/s41064-024-00308-9
S. Abualhanud, E. Erahan, M. Mehltretter

An accurate 3D representation of the geometry and semantics of an environment builds the basis for a large variety of downstream tasks and is essential for autonomous driving related tasks such as path planning and obstacle avoidance. The focus of this work is put on 3D semantic occupancy prediction, i.e., the reconstruction of a scene as a voxel grid where each voxel is assigned both an occupancy and a semantic label. We present a Convolutional Neural Network-based method that utilizes multiple color images from a surround-view setup with minimal overlap, together with the associated interior and exterior camera parameters as input, to reconstruct the observed environment as a 3D semantic occupancy map. To account for the ill-posed nature of reconstructing a 3D representation from monocular 2D images, the image information is integrated over time: Under the assumption that the camera setup is moving, images from consecutive time steps are used to form a multi-view stereo setup. In exhaustive experiments, we investigate the challenges presented by dynamic objects and the possibilities of training the proposed method with either 3D or 2D reference data. Latter being motivated by the comparably higher costs of generating and annotating 3D ground truth data. Moreover, we present and investigate a novel self-supervised training scheme that does not require any geometric reference data, but only relies on sparse semantic ground truth. An evaluation on the Occ3D dataset, including a comparison against current state-of-the-art self-supervised methods from the literature, demonstrates the potential of our self-supervised variant.

对环境的几何形状和语义进行精确的三维表示,为各种下游任务奠定了基础,对于自动驾驶相关任务(如路径规划和避障)也至关重要。这项工作的重点是三维语义占位预测,即以体素网格的形式重建场景,其中每个体素都被分配了占位和语义标签。我们提出了一种基于卷积神经网络的方法,该方法利用环视设置中重叠最少的多幅彩色图像以及相关的内部和外部相机参数作为输入,将观察到的环境重建为三维语义占位图。考虑到从单目二维图像重建三维表示的不确定性,图像信息是随时间整合的:假设摄像机是移动的,那么连续时间步骤的图像将被用于形成多视角立体设置。在详尽的实验中,我们研究了动态物体带来的挑战,以及使用三维或二维参考数据训练所提方法的可能性。后者是因为生成和注释三维地面实况数据的成本相对较高。此外,我们还提出并研究了一种新颖的自监督训练方案,它不需要任何几何参考数据,而只依赖于稀疏的语义地面实况。在 Occ3D 数据集上进行的评估,包括与目前文献中最先进的自监督方法的比较,证明了我们的自监督变体的潜力。
{"title":"Self-Supervised 3D Semantic Occupancy Prediction from Multi-View 2D Surround Images","authors":"S. Abualhanud, E. Erahan, M. Mehltretter","doi":"10.1007/s41064-024-00308-9","DOIUrl":"https://doi.org/10.1007/s41064-024-00308-9","url":null,"abstract":"<p>An accurate 3D representation of the geometry and semantics of an environment builds the basis for a large variety of downstream tasks and is essential for autonomous driving related tasks such as path planning and obstacle avoidance. The focus of this work is put on 3D semantic occupancy prediction, i.e., the reconstruction of a scene as a voxel grid where each voxel is assigned both an occupancy and a semantic label. We present a Convolutional Neural Network-based method that utilizes multiple color images from a surround-view setup with minimal overlap, together with the associated interior and exterior camera parameters as input, to reconstruct the observed environment as a 3D semantic occupancy map. To account for the ill-posed nature of reconstructing a 3D representation from monocular 2D images, the image information is integrated over time: Under the assumption that the camera setup is moving, images from consecutive time steps are used to form a multi-view stereo setup. In exhaustive experiments, we investigate the challenges presented by dynamic objects and the possibilities of training the proposed method with either 3D or 2D reference data. Latter being motivated by the comparably higher costs of generating and annotating 3D ground truth data. Moreover, we present and investigate a novel self-supervised training scheme that does not require any geometric reference data, but only relies on sparse semantic ground truth. An evaluation on the Occ3D dataset, including a comparison against current state-of-the-art self-supervised methods from the literature, demonstrates the potential of our self-supervised variant.</p>","PeriodicalId":56035,"journal":{"name":"PFG-Journal of Photogrammetry Remote Sensing and Geoinformation Science","volume":"6 1","pages":""},"PeriodicalIF":4.1,"publicationDate":"2024-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142258861","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Characterization of transient movements within the Joshimath hillslope complex: Results from multi-sensor InSAR observations 乔希马特山坡复合体瞬时移动的特征:多传感器 InSAR 观测结果
IF 4.1 4区 地球科学 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2024-09-17 DOI: 10.1007/s41064-024-00315-w
Wandi Wang, Mahdi Motagh, Zhuge Xia, Zhong Lu, Sadra Karimzadeh, Chao Zhou, Alina V. Shevchenko, Sigrid Roessner

This paper investigates the spatiotemporal characteristics and life-cycle of movements within the Joshimath landslide-prone slope over the period from 2015 to 2024, utilizing multi-sensor interferometric data from Sentinel‑1, ALOS‑2, and TerraSAR‑X satellites. Multi-temporal InSAR analysis before the 2023 slope destabilization crisis, when the region experienced significant ground deformation acceleration, revealed two distinct deformation clusters within the eastern and middle parts of the slope. These active deformation regions have been creeping up to −200 mm/yr. Slope deformation analysis indicates that the entire Joshimath landslide-prone slope can be categorized kinematically as either Extremely-Slow (ES) or Very-Slow (VS) moving slope, with the eastern cluster mainly exhibiting ES movements, while the middle cluster showing VS movements. Two episodes of significant acceleration occurred on August 21, 2019 and November 2, 2021, with the rate of slope deformation increasing by 20% (from −50 to −60 mm/yr) and around threefold (from −60 to −249 mm/yr), respectively. Following the 2023 destabilization crisis, the rate of ground deformation notably increased across all datasets for both clusters, except for the Sentinel‑1 ascending data in the eastern cluster. Pre-crisis, horizontal deformation was dominant both in the eastern and middle clusters. Horizontal deformation remained dominant and increased significantly in the eastern cluster post-crisis phase, whereas vertical deformation became predominant in the middle cluster. Wavelet analysis reveals a strong correlation between two acceleration episodes and extreme precipitation in 2019 and 2021, but no similar correlation was detected in other years. This indicates that while extreme rainfall significantly influenced the dynamics of slope movements during these episodes, less strong precipitation had a minimal impact on slope movements during other periods.

本文利用哨兵-1、ALOS-2 和 TerraSAR-X 卫星的多传感器干涉测量数据,研究了 2015 年至 2024 年期间乔希马特山滑坡易发斜坡的时空特征和运动生命周期。在 2023 年斜坡失稳危机之前,该地区经历了显著的地面变形加速,多时相 InSAR 分析揭示了斜坡东部和中部两个不同的变形群。这些活跃的变形区域的蠕变速度高达-200 毫米/年。斜坡变形分析表明,整个乔希马特易滑坡斜坡在运动学上可分为极慢(ES)移动斜坡和极慢(VS)移动斜坡,其中东部斜坡群主要表现为 ES 移动,而中部斜坡群则表现为 VS 移动。2019年8月21日和2021年11月2日发生了两次明显的加速运动,斜坡变形速度分别增加了20%(从-50毫米/年增加到-60毫米/年)和3倍左右(从-60毫米/年增加到-249毫米/年)。2023 年失稳危机发生后,除东部集群的哨兵-1 号上升数据外,两个集群的所有数据集的地面变形速率都显著增加。危机前,水平形变在东部和中部组群中均占主导地位。危机后阶段,水平形变在东部星群中仍占主导地位并显著增加,而垂直形变则在中部星群中占主导地位。小波分析显示,2019 年和 2021 年的两次加速事件与极端降水之间存在很强的相关性,但在其他年份没有发现类似的相关性。这表明,虽然极端降水在这些事件中对斜坡运动的动态产生了重大影响,但在其他时期,强度较小的降水对斜坡运动的影响微乎其微。
{"title":"Characterization of transient movements within the Joshimath hillslope complex: Results from multi-sensor InSAR observations","authors":"Wandi Wang, Mahdi Motagh, Zhuge Xia, Zhong Lu, Sadra Karimzadeh, Chao Zhou, Alina V. Shevchenko, Sigrid Roessner","doi":"10.1007/s41064-024-00315-w","DOIUrl":"https://doi.org/10.1007/s41064-024-00315-w","url":null,"abstract":"<p>This paper investigates the spatiotemporal characteristics and life-cycle of movements within the Joshimath landslide-prone slope over the period from 2015 to 2024, utilizing multi-sensor interferometric data from Sentinel‑1, ALOS‑2, and TerraSAR‑X satellites. Multi-temporal InSAR analysis before the 2023 slope destabilization crisis, when the region experienced significant ground deformation acceleration, revealed two distinct deformation clusters within the eastern and middle parts of the slope. These active deformation regions have been creeping up to −200 mm/yr. Slope deformation analysis indicates that the entire Joshimath landslide-prone slope can be categorized kinematically as either Extremely-Slow (ES) or Very-Slow (VS) moving slope, with the eastern cluster mainly exhibiting ES movements, while the middle cluster showing VS movements. Two episodes of significant acceleration occurred on August 21, 2019 and November 2, 2021, with the rate of slope deformation increasing by 20% (from −50 to −60 mm/yr) and around threefold (from −60 to −249 mm/yr), respectively. Following the 2023 destabilization crisis, the rate of ground deformation notably increased across all datasets for both clusters, except for the Sentinel‑1 ascending data in the eastern cluster. Pre-crisis, horizontal deformation was dominant both in the eastern and middle clusters. Horizontal deformation remained dominant and increased significantly in the eastern cluster post-crisis phase, whereas vertical deformation became predominant in the middle cluster. Wavelet analysis reveals a strong correlation between two acceleration episodes and extreme precipitation in 2019 and 2021, but no similar correlation was detected in other years. This indicates that while extreme rainfall significantly influenced the dynamics of slope movements during these episodes, less strong precipitation had a minimal impact on slope movements during other periods.</p>","PeriodicalId":56035,"journal":{"name":"PFG-Journal of Photogrammetry Remote Sensing and Geoinformation Science","volume":"1 1","pages":""},"PeriodicalIF":4.1,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142258853","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Monocular Pose and Shape Reconstruction of Vehicles in UAV imagery using a Multi-task CNN 利用多任务 CNN 对无人机图像中的车辆进行单目姿态和形状重构
IF 4.1 4区 地球科学 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2024-09-16 DOI: 10.1007/s41064-024-00311-0
S. El Amrani Abouelassad, M. Mehltretter, F. Rottensteiner

Estimating the pose and shape of vehicles from aerial images is an important, yet challenging task. While there are many existing approaches that use stereo images from street-level perspectives to reconstruct objects in 3D, the majority of aerial configurations used for purposes like traffic surveillance are limited to monocular images. Addressing this challenge, a Convolutional Neural Network-based method is presented in this paper, which jointly performs detection, pose, type and 3D shape estimation for vehicles observed in monocular UAV imagery. For this purpose, a robust 3D object model is used following the concept of an Active Shape Model. In addition, different variants of loss functions for learning 3D shape estimation are presented, focusing on the height component, which is particularly challenging to estimate from monocular near-nadir images. We also introduce a UAV-based dataset to evaluate our model in addition to an augmented version of the publicly available Hessigheim benchmark dataset. Our method yields promising results in pose and shape estimation: utilising images with a ground sampling distance (GSD) of 3 cm, it achieves median errors of up to 4 cm in position and 3° in orientation. Additionally, it achieves root mean square (RMS) errors of (pm 6) cm in planimetry and (pm 18) cm in height for keypoints defining the car shape.

从航空图像中估计车辆的姿态和形状是一项重要而又具有挑战性的任务。虽然现有的许多方法都使用街景立体图像来重建三维物体,但大多数用于交通监控等目的的航空配置都仅限于单目图像。为了应对这一挑战,本文提出了一种基于卷积神经网络的方法,该方法可联合执行单目无人机图像中观察到的车辆的检测、姿态、类型和三维形状估计。为此,根据主动形状模型的概念,使用了一个稳健的三维物体模型。此外,我们还介绍了用于学习三维形状估计的不同损失函数变体,重点关注高度分量,因为从单目近天底图像中估计高度分量尤其具有挑战性。除了公开的 Hessigheim 基准数据集的增强版外,我们还引入了一个基于无人机的数据集来评估我们的模型。我们的方法在姿态和形状估计方面取得了可喜的成果:利用地面采样距离(GSD)为 3 厘米的图像,我们的方法在位置和方向上的中值误差分别达到了 4 厘米和 3°。此外,对于定义汽车形状的关键点,它的平面测量均方根(RMS)误差为 (pm 6) 厘米,高度误差为 (pm 18) 厘米。
{"title":"Monocular Pose and Shape Reconstruction of Vehicles in UAV imagery using a Multi-task CNN","authors":"S. El Amrani Abouelassad, M. Mehltretter, F. Rottensteiner","doi":"10.1007/s41064-024-00311-0","DOIUrl":"https://doi.org/10.1007/s41064-024-00311-0","url":null,"abstract":"<p>Estimating the pose and shape of vehicles from aerial images is an important, yet challenging task. While there are many existing approaches that use stereo images from street-level perspectives to reconstruct objects in 3D, the majority of aerial configurations used for purposes like traffic surveillance are limited to monocular images. Addressing this challenge, a Convolutional Neural Network-based method is presented in this paper, which jointly performs detection, pose, type and 3D shape estimation for vehicles observed in monocular UAV imagery. For this purpose, a robust 3D object model is used following the concept of an Active Shape Model. In addition, different variants of loss functions for learning 3D shape estimation are presented, focusing on the height component, which is particularly challenging to estimate from monocular near-nadir images. We also introduce a UAV-based dataset to evaluate our model in addition to an augmented version of the publicly available Hessigheim benchmark dataset. Our method yields promising results in pose and shape estimation: utilising images with a ground sampling distance (GSD) of 3 cm, it achieves median errors of up to 4 cm in position and 3° in orientation. Additionally, it achieves root mean square (RMS) errors of <span>(pm 6)</span> cm in planimetry and <span>(pm 18)</span> cm in height for keypoints defining the car shape.</p>","PeriodicalId":56035,"journal":{"name":"PFG-Journal of Photogrammetry Remote Sensing and Geoinformation Science","volume":"81 1","pages":""},"PeriodicalIF":4.1,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142258854","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Assessing the Impact of Data-resolution On Ocean Frontal Characteristics 评估数据分辨率对海洋锋面特征的影响
IF 4.1 4区 地球科学 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2024-09-16 DOI: 10.1007/s41064-024-00318-7
Kai Yang, Andrew M. Fischer

Easy access to and advances in satellite remote sensing data has enabled enhanced analysis of ocean fronts, physical and ecologically important areas where water masses converge. Recent development of higher-resolution satellite imagery to detect ocean fronts provides the potential to better capture patterns and trends of ocean change and improve modelling and prediction efforts. This study examines the relationship between satellite data spatial resolution and its influence on the quantification of frontal characteristics, frontal quantity, length, strength and density. We also examine the relationship between Finite-Size Lyapunov Exponents and image resolution. We found higher spatial resolution leads to increased frontal quantity and decreased frontal length. Also, both strength and spatial density of fronts differ at various resolutions. The Finite-Size Lyapunov Exponent value does not change significantly with resolution. Knowledge of the impact of resolution on the quantification of frontal characteristics is crucial as it enables the exploration of novel experimental design to further facilitate the development of improved parameterization and uncertainties in ocean modelling/studies.

卫星遥感数据的便捷获取和进步加强了对海洋前沿的分析,海洋前沿是水团汇聚的重要物理和生态区域。最近开发的用于探测海洋锋面的高分辨率卫星图像有可能更好地捕捉海洋变化的模式和趋势,并改进建模和预测工作。本研究探讨了卫星数据空间分辨率之间的关系及其对锋面特征、锋面数量、长度、强度和密度量化的影响。我们还研究了有限大小 Lyapunov 指数与图像分辨率之间的关系。我们发现,空间分辨率越高,额部数量越多,额部长度越短。此外,在不同分辨率下,锋面的强度和空间密度也不同。有限大小 Lyapunov 指数值不会随着分辨率的提高而发生显著变化。了解分辨率对锋面特征量化的影响至关重要,因为这有助于探索新的实验设计,进一步促进海洋建模/研究中参数化和不确定性的改进。
{"title":"Assessing the Impact of Data-resolution On Ocean Frontal Characteristics","authors":"Kai Yang, Andrew M. Fischer","doi":"10.1007/s41064-024-00318-7","DOIUrl":"https://doi.org/10.1007/s41064-024-00318-7","url":null,"abstract":"<p>Easy access to and advances in satellite remote sensing data has enabled enhanced analysis of ocean fronts, physical and ecologically important areas where water masses converge. Recent development of higher-resolution satellite imagery to detect ocean fronts provides the potential to better capture patterns and trends of ocean change and improve modelling and prediction efforts. This study examines the relationship between satellite data spatial resolution and its influence on the quantification of frontal characteristics, frontal quantity, length, strength and density. We also examine the relationship between Finite-Size Lyapunov Exponents and image resolution. We found higher spatial resolution leads to increased frontal quantity and decreased frontal length. Also, both strength and spatial density of fronts differ at various resolutions. The Finite-Size Lyapunov Exponent value does not change significantly with resolution. Knowledge of the impact of resolution on the quantification of frontal characteristics is crucial as it enables the exploration of novel experimental design to further facilitate the development of improved parameterization and uncertainties in ocean modelling/studies.</p>","PeriodicalId":56035,"journal":{"name":"PFG-Journal of Photogrammetry Remote Sensing and Geoinformation Science","volume":"39 1","pages":""},"PeriodicalIF":4.1,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142258855","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Challenges and Opportunities of Sentinel-1 InSAR for Transport Infrastructure Monitoring 用于交通基础设施监测的哨兵-1 InSAR 的挑战与机遇
IF 4.1 4区 地球科学 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2024-09-16 DOI: 10.1007/s41064-024-00314-x
Andreas Piter, Mahmud Haghshenas Haghighi, Mahdi Motagh

Monitoring displacement at transport infrastructure using Sentinel‑1 Interferometric Synthetic Aperture Radar (InSAR) faces challenges due to the sensor’s medium spatial resolution, which limits the pixel coverage over the infrastructure. Therefore, carefully selecting coherent pixels is crucial to achieve a high density of reliable measurement points and to minimize noisy observations. This study evaluates the effectiveness of various pixel selection methods for displacement monitoring within transport infrastructures. We employ a two-step InSAR time series processing approach. First, high-quality first-order pixels are selected using temporal phase coherence (TPC) to estimate and correct atmospheric contributions. Then, a combination of different pixel selection methods is applied to identify coherent second-order pixels for displacement analysis. These methods include amplitude dispersion index (ADI), TPC, phase linking coherence (PLC), and top eigenvalue percentage (TEP), targeting both point-like scatterer (PS) and distributed scatterer (DS) pixels. Experiments are conducted in two case studies: one in Germany, characterized by dense vegetation, and one in Spain, with sparse vegetation. In Germany, the density of measurement points was approximately 30 points/km², with the longest segment of the infrastructure without any coherent pixels being 2.8 km. In Spain, the density of measurement points exceeded 500 points/km², with the longest section without coherent pixels being 700 meters. The results indicate that despite the challenges posed by medium-resolution data, the sensor is capable of providing adequate measurement points when suitable pixel selection methods are employed. However, careful consideration is necessary to exclude noisy pixels from the analysis. The findings highlight the importance of choosing a proper method tailored to infrastructure characteristics. Specifically, combining TPC and PLC methods offers a complementary set of pixels suitable for displacement measurements, whereas ADI and TEP are less effective in this context. This study demonstrates the potential of Sentinel‑1 InSAR for capturing both regional-scale and localized displacements at transport infrastructure.

使用哨兵-1 干涉合成孔径雷达(InSAR)监测交通基础设施的位移面临挑战,因为传感器的空间分辨率较低,限制了对基础设施的像素覆盖。因此,要获得高密度的可靠测量点并最大限度地减少噪声观测,精心选择相干像素至关重要。本研究评估了各种像素选择方法在交通基础设施位移监测中的有效性。我们采用了两步 InSAR 时间序列处理方法。首先,利用时相相干性(TPC)选择高质量的一阶像素,以估计和修正大气贡献。然后,结合不同的像素选择方法,确定相干的二阶像素,进行位移分析。这些方法包括振幅色散指数 (ADI)、时相相干性 (TPC)、相位链接相干性 (PLC) 和顶特征值百分比 (TEP),针对点状散射体 (PS) 和分布式散射体 (DS) 像素。实验在两个案例研究中进行:一个在植被茂密的德国,另一个在植被稀疏的西班牙。在德国,测量点密度约为 30 点/平方公里,基础设施中没有任何连贯像素的最长路段为 2.8 公里。在西班牙,测量点密度超过 500 点/平方公里,没有连贯像素的最长路段为 700 米。结果表明,尽管中等分辨率数据带来了挑战,但如果采用适当的像素选择方法,传感器还是能够提供足够的测量点。不过,在分析时需要仔细考虑如何排除噪声像素。研究结果强调了根据基础设施特点选择适当方法的重要性。具体来说,将 TPC 和 PLC 方法结合起来可提供一组适合位移测量的互补像素,而 ADI 和 TEP 在这方面的效果较差。这项研究证明了 Sentinel-1 InSAR 在捕捉交通基础设施的区域和局部位移方面的潜力。
{"title":"Challenges and Opportunities of Sentinel-1 InSAR for Transport Infrastructure Monitoring","authors":"Andreas Piter, Mahmud Haghshenas Haghighi, Mahdi Motagh","doi":"10.1007/s41064-024-00314-x","DOIUrl":"https://doi.org/10.1007/s41064-024-00314-x","url":null,"abstract":"<p>Monitoring displacement at transport infrastructure using Sentinel‑1 Interferometric Synthetic Aperture Radar (InSAR) faces challenges due to the sensor’s medium spatial resolution, which limits the pixel coverage over the infrastructure. Therefore, carefully selecting coherent pixels is crucial to achieve a high density of reliable measurement points and to minimize noisy observations. This study evaluates the effectiveness of various pixel selection methods for displacement monitoring within transport infrastructures. We employ a two-step InSAR time series processing approach. First, high-quality first-order pixels are selected using temporal phase coherence (TPC) to estimate and correct atmospheric contributions. Then, a combination of different pixel selection methods is applied to identify coherent second-order pixels for displacement analysis. These methods include amplitude dispersion index (ADI), TPC, phase linking coherence (PLC), and top eigenvalue percentage (TEP), targeting both point-like scatterer (PS) and distributed scatterer (DS) pixels. Experiments are conducted in two case studies: one in Germany, characterized by dense vegetation, and one in Spain, with sparse vegetation. In Germany, the density of measurement points was approximately 30 points/km², with the longest segment of the infrastructure without any coherent pixels being 2.8 km. In Spain, the density of measurement points exceeded 500 points/km², with the longest section without coherent pixels being 700 meters. The results indicate that despite the challenges posed by medium-resolution data, the sensor is capable of providing adequate measurement points when suitable pixel selection methods are employed. However, careful consideration is necessary to exclude noisy pixels from the analysis. The findings highlight the importance of choosing a proper method tailored to infrastructure characteristics. Specifically, combining TPC and PLC methods offers a complementary set of pixels suitable for displacement measurements, whereas ADI and TEP are less effective in this context. This study demonstrates the potential of Sentinel‑1 InSAR for capturing both regional-scale and localized displacements at transport infrastructure.</p>","PeriodicalId":56035,"journal":{"name":"PFG-Journal of Photogrammetry Remote Sensing and Geoinformation Science","volume":"11 1","pages":""},"PeriodicalIF":4.1,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142258856","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Weighted Multiple Point Cloud Fusion 加权多点云融合
IF 4.1 4区 地球科学 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2024-09-12 DOI: 10.1007/s41064-024-00310-1
Kwasi Nyarko Poku-Agyemang, Alexander Reiterer

Multiple viewpoint 3D reconstruction has been used in recent years to create accurate complete scenes and objects used for various applications. This is to overcome limitations of single viewpoint 3D digital imaging such as occlusion within the scene during the reconstruction process. In this paper, we propose a weighted point cloud fusion process using both local and global spatial information of the point clouds to fuse them together. The process aims to minimize duplication and remove noise while maintaining a consistent level of details using spatial information from point clouds to compute a weight to fuse them. The algorithm improves the overall accuracy of the fused point cloud while maintaining a similar degree of coverage comparable with state-of-the-art point cloud fusion algorithms.

近年来,多视点三维重建技术已被用于创建精确完整的场景和对象,并被广泛应用于各种领域。这是为了克服单视点三维数字成像的局限性,如重建过程中场景内的遮挡。在本文中,我们提出了一种加权点云融合流程,利用点云的局部和全局空间信息将它们融合在一起。该流程旨在利用点云的空间信息来计算融合点云的权重,从而最大限度地减少重复和去除噪音,同时保持细节的一致性。该算法提高了融合点云的整体精度,同时保持了与最先进的点云融合算法类似的覆盖程度。
{"title":"Weighted Multiple Point Cloud Fusion","authors":"Kwasi Nyarko Poku-Agyemang, Alexander Reiterer","doi":"10.1007/s41064-024-00310-1","DOIUrl":"https://doi.org/10.1007/s41064-024-00310-1","url":null,"abstract":"<p>Multiple viewpoint 3D reconstruction has been used in recent years to create accurate complete scenes and objects used for various applications. This is to overcome limitations of single viewpoint 3D digital imaging such as occlusion within the scene during the reconstruction process. In this paper, we propose a weighted point cloud fusion process using both local and global spatial information of the point clouds to fuse them together. The process aims to minimize duplication and remove noise while maintaining a consistent level of details using spatial information from point clouds to compute a weight to fuse them. The algorithm improves the overall accuracy of the fused point cloud while maintaining a similar degree of coverage comparable with state-of-the-art point cloud fusion algorithms.</p>","PeriodicalId":56035,"journal":{"name":"PFG-Journal of Photogrammetry Remote Sensing and Geoinformation Science","volume":"3 1","pages":""},"PeriodicalIF":4.1,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142200717","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Stripe Error Correction for Landsat-7 Using Deep Learning 利用深度学习对 Landsat-7 进行条纹误差校正
IF 4.1 4区 地球科学 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2024-08-29 DOI: 10.1007/s41064-024-00306-x
Hilal Adıyaman, Yunus Emre Varul, Tolga Bakırman, Bülent Bayram

Long-term time series satellite imagery became highly essential for analyzing earth cycles such as global warming, climate change, and urbanization. Landsat‑7 satellite imagery plays a key role in this domain since it provides open-access data with expansive coverage and consistent temporal resolution for more than two decades. This paper addresses the challenge of stripe errors induced by Scan Line Corrector sensor malfunction in Landsat‑7 ETM+ satellite imagery, resulting in data loss and degradation. To overcome this problem, we propose a Generative Adversarial Networks approach to fill the gaps in the Landsat‑7 ETM+ panchromatic images. First, we introduce the YTU_STRIPE dataset, comprising Landsat‑8 OLI panchromatic images with synthetically induced stripe errors, for model training and testing. Our results indicate sufficient performance of the Pix2Pix GAN for this purpose. We demonstrate the efficiency of our approach through systematic experimentation and evaluation using various accuracy metrics, including Peak Signal-to-Noise Ratio, Structural Similarity Index Measurement, Universal Image Quality Index, Correlation Coefficient, and Root Mean Square Error which were calculated as 38.5570, 0.9206, 0.7670, 0.7753 and 3.8212, respectively. Our findings suggest promising prospects for utilizing synthetic imagery from Landsat‑8 OLI to mitigate stripe errors in Landsat‑7 ETM+ SLC-off imagery, thereby enhancing image reconstruction efforts. The datasets and model weights generated in this study are publicly available for further research and development: https://github.com/ynsemrevrl/eliminating-stripe-errors.

长期的时间序列卫星图像对于分析全球变暖、气候变化和城市化等地球周期非常重要。Landsat-7 卫星图像在这一领域发挥着关键作用,因为它提供了二十多年来覆盖范围广、时间分辨率一致的开放式数据。本文探讨了 Landsat-7 ETM+ 卫星图像中因扫描线校正器传感器故障而引起的条纹错误,从而导致数据丢失和质量下降的难题。为了克服这一问题,我们提出了一种生成对抗网络方法来填补 Landsat-7 ETM+ 全色图像中的空白。首先,我们引入了 YTU_STRIPE 数据集,该数据集由具有合成条纹误差的 Landsat-8 OLI 全色图像组成,用于模型训练和测试。我们的结果表明,Pix2Pix GAN 在这方面具有足够的性能。我们通过系统实验和使用各种精度指标(包括峰值信噪比、结构相似性指数测量、通用图像质量指数、相关系数和均方根误差)进行评估,证明了我们方法的效率,计算结果分别为 38.5570、0.9206、0.7670、0.7753 和 3.8212。我们的研究结果表明,利用来自 Landsat-8 OLI 的合成图像来减少 Landsat-7 ETM+ SLC-off 图像中的条纹误差,从而提高图像重建工作的效率,前景十分广阔。本研究生成的数据集和模型权重可公开用于进一步的研究和开发:https://github.com/ynsemrevrl/eliminating-stripe-errors。
{"title":"Stripe Error Correction for Landsat-7 Using Deep Learning","authors":"Hilal Adıyaman, Yunus Emre Varul, Tolga Bakırman, Bülent Bayram","doi":"10.1007/s41064-024-00306-x","DOIUrl":"https://doi.org/10.1007/s41064-024-00306-x","url":null,"abstract":"<p>Long-term time series satellite imagery became highly essential for analyzing earth cycles such as global warming, climate change, and urbanization. Landsat‑7 satellite imagery plays a key role in this domain since it provides open-access data with expansive coverage and consistent temporal resolution for more than two decades. This paper addresses the challenge of stripe errors induced by Scan Line Corrector sensor malfunction in Landsat‑7 ETM+ satellite imagery, resulting in data loss and degradation. To overcome this problem, we propose a Generative Adversarial Networks approach to fill the gaps in the Landsat‑7 ETM+ panchromatic images. First, we introduce the YTU_STRIPE dataset, comprising Landsat‑8 OLI panchromatic images with synthetically induced stripe errors, for model training and testing. Our results indicate sufficient performance of the Pix2Pix GAN for this purpose. We demonstrate the efficiency of our approach through systematic experimentation and evaluation using various accuracy metrics, including Peak Signal-to-Noise Ratio, Structural Similarity Index Measurement, Universal Image Quality Index, Correlation Coefficient, and Root Mean Square Error which were calculated as 38.5570, 0.9206, 0.7670, 0.7753 and 3.8212, respectively. Our findings suggest promising prospects for utilizing synthetic imagery from Landsat‑8 OLI to mitigate stripe errors in Landsat‑7 ETM+ SLC-off imagery, thereby enhancing image reconstruction efforts. The datasets and model weights generated in this study are publicly available for further research and development: https://github.com/ynsemrevrl/eliminating-stripe-errors.</p>","PeriodicalId":56035,"journal":{"name":"PFG-Journal of Photogrammetry Remote Sensing and Geoinformation Science","volume":"14 1","pages":""},"PeriodicalIF":4.1,"publicationDate":"2024-08-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142200722","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
EnhancedNet, an End-to-End Network for Dense Disparity Estimation and its Application to Aerial Images 增强型网络(EnhancedNet)--用于密集差异估计的端到端网络及其在航空图像中的应用
IF 4.1 4区 地球科学 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2024-08-28 DOI: 10.1007/s41064-024-00307-w
Junhua Kang, Lin Chen, Christian Heipke

Recent developments in deep learning technology have boosted the performance of dense stereo reconstruction. However, the state-of-the-art deep learning-based stereo matching methods are mainly trained using close-range synthetic images. Consequently, the application of these methods in aerial photogrammetry and remote sensing is currently far from straightforward. In this paper, we propose a new disparity estimation network for stereo matching and investigate its generalization abilities in regard to aerial images. First, we propose an end-to-end deep learning network for stereo matching, regularized by disparity gradients, which includes a residual cost volume and a reconstruction error volume in a refinement module, and multiple losses. In order to investigate the influence of the multiple losses, a comprehensive analysis is presented. Second, based on this network trained with synthetic close-range data, we propose a new pipeline for matching high-resolution aerial imagery. The experimental results show that the proposed network improves the disparity accuracy by up to 40% in terms of errors larger than 1 px compared to results when not including the refinement network, especially in areas containing detailed small objects. In addition, in qualitative and quantitative experiments, we are able to show that our model, pre-trained on a synthetic stereo dataset, achieves very competitive sub-pixel geometric accuracy on aerial images. These results confirm that the domain gap between synthetic close-range and real aerial images can be satisfactorily bridged using the proposed new deep learning method for dense image matching.

深度学习技术的最新发展提升了密集立体重建的性能。然而,基于深度学习的最新立体匹配方法主要是利用近距离合成图像进行训练的。因此,这些方法目前在航空摄影测量和遥感中的应用还不够直接。在本文中,我们提出了一种新的用于立体匹配的差异估计网络,并研究了其在航空图像方面的泛化能力。首先,我们提出了一种用于立体匹配的端到端深度学习网络,该网络由差距梯度正则化,包括细化模块中的残差成本卷和重建误差卷,以及多重损失。为了研究多重损失的影响,本文进行了综合分析。其次,基于这个用合成近距离数据训练的网络,我们提出了一种新的高分辨率航空图像匹配管道。实验结果表明,与不包含细化网络的结果相比,在误差大于 1 px 的情况下,所提出的网络可将差异精度提高 40%,尤其是在包含细节小物体的区域。此外,在定性和定量实验中,我们还证明了我们在合成立体数据集上预先训练的模型在航空图像上实现了极具竞争力的亚像素几何精度。这些结果证实,利用所提出的新深度学习方法进行密集图像匹配,可以令人满意地缩小合成近距离图像与真实航空图像之间的领域差距。
{"title":"EnhancedNet, an End-to-End Network for Dense Disparity Estimation and its Application to Aerial Images","authors":"Junhua Kang, Lin Chen, Christian Heipke","doi":"10.1007/s41064-024-00307-w","DOIUrl":"https://doi.org/10.1007/s41064-024-00307-w","url":null,"abstract":"<p>Recent developments in deep learning technology have boosted the performance of dense stereo reconstruction. However, the state-of-the-art deep learning-based stereo matching methods are mainly trained using close-range synthetic images. Consequently, the application of these methods in aerial photogrammetry and remote sensing is currently far from straightforward. In this paper, we propose a new disparity estimation network for stereo matching and investigate its generalization abilities in regard to aerial images. First, we propose an end-to-end deep learning network for stereo matching, regularized by disparity gradients, which includes a residual cost volume and a reconstruction error volume in a refinement module, and multiple losses. In order to investigate the influence of the multiple losses, a comprehensive analysis is presented. Second, based on this network trained with synthetic close-range data, we propose a new pipeline for matching high-resolution aerial imagery. The experimental results show that the proposed network improves the disparity accuracy by up to 40% in terms of errors larger than 1 px compared to results when not including the refinement network, especially in areas containing detailed small objects. In addition, in qualitative and quantitative experiments, we are able to show that our model, pre-trained on a synthetic stereo dataset, achieves very competitive sub-pixel geometric accuracy on aerial images. These results confirm that the domain gap between synthetic close-range and real aerial images can be satisfactorily bridged using the proposed new deep learning method for dense image matching.</p>","PeriodicalId":56035,"journal":{"name":"PFG-Journal of Photogrammetry Remote Sensing and Geoinformation Science","volume":"25 1","pages":""},"PeriodicalIF":4.1,"publicationDate":"2024-08-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142200718","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fresh Concrete Properties from Stereoscopic Image Sequences 从立体图像序列中获取新鲜混凝土特性
IF 4.1 4区 地球科学 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2024-08-26 DOI: 10.1007/s41064-024-00303-0
Max Meyer, Amadeus Langer, Max Mehltretter, Dries Beyer, Max Coenen, Tobias Schack, Michael Haist, Christian Heipke

Increasing the degree of digitization and automation in concrete production can make a decisive contribution to reducing the associated (text{CO}_{2}) emissions. This paper presents a method which predicts the properties of fresh concrete during the mixing process on the basis of stereoscopic image sequences of the moving concrete and mix design information or a variation of these. A Convolutional Neural Network (CNN) is used for the prediction, which receives the images supported by information about the mix design as input. In addition, the network receives temporal information in the form of the time difference between image acquisition and the point in time for which the concrete properties are to be predicted. During training, the times at which the reference values were captured are used for the latter. With this temporal information, the network implicitly learns the time-dependent behavior of the concrete properties. The network predicts the slump flow diameter, the yield stress and the plastic viscosity. The time-dependent prediction opens up the possibility of forecasting the temporal development of the fresh concrete properties during mixing. This is a significant advantage for the concrete industry, as countermeasures can then be taken in a timely manner, if the properties deviate from the desired ones. In various experiments it is shown that both the stereoscopic observations and the mix design information contain valuable information for the time-dependent prediction of the fresh concrete properties.

提高混凝土生产的数字化和自动化程度可以为减少相关排放做出决定性贡献。本文介绍了一种方法,该方法可根据移动混凝土的立体图像序列和混合设计信息或这些信息的变体,预测搅拌过程中新拌混凝土的特性。预测使用了卷积神经网络 (CNN),该网络接收由混合设计信息支持的图像作为输入。此外,该网络还能接收时间信息,即图像采集与预测混凝土特性的时间点之间的时间差。在训练过程中,参考值的采集时间被用于后者。有了这些时间信息,网络就能隐式地学习随时间变化的混凝土特性行为。该网络可预测坍落度流动直径、屈服应力和塑性粘度。随时间变化的预测为预报搅拌过程中新拌混凝土性能的随时间变化提供了可能。这对混凝土行业来说是一个重大优势,因为如果性能偏离预期,就可以及时采取应对措施。各种实验表明,立体观测和混合设计信息都包含了对新拌混凝土性能随时间变化进行预测的宝贵信息。
{"title":"Fresh Concrete Properties from Stereoscopic Image Sequences","authors":"Max Meyer, Amadeus Langer, Max Mehltretter, Dries Beyer, Max Coenen, Tobias Schack, Michael Haist, Christian Heipke","doi":"10.1007/s41064-024-00303-0","DOIUrl":"https://doi.org/10.1007/s41064-024-00303-0","url":null,"abstract":"<p>Increasing the degree of digitization and automation in concrete production can make a decisive contribution to reducing the associated <span>(text{CO}_{2})</span> emissions. This paper presents a method which predicts the properties of fresh concrete during the mixing process on the basis of stereoscopic image sequences of the moving concrete and mix design information or a variation of these. A Convolutional Neural Network (CNN) is used for the prediction, which receives the images supported by information about the mix design as input. In addition, the network receives temporal information in the form of the time difference between image acquisition and the point in time for which the concrete properties are to be predicted. During training, the times at which the reference values were captured are used for the latter. With this temporal information, the network implicitly learns the time-dependent behavior of the concrete properties. The network predicts the slump flow diameter, the yield stress and the plastic viscosity. The time-dependent prediction opens up the possibility of forecasting the temporal development of the fresh concrete properties during mixing. This is a significant advantage for the concrete industry, as countermeasures can then be taken in a timely manner, if the properties deviate from the desired ones. In various experiments it is shown that both the stereoscopic observations and the mix design information contain valuable information for the time-dependent prediction of the fresh concrete properties.</p>","PeriodicalId":56035,"journal":{"name":"PFG-Journal of Photogrammetry Remote Sensing and Geoinformation Science","volume":"3 1","pages":""},"PeriodicalIF":4.1,"publicationDate":"2024-08-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142200719","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Assessing Patterns and Trends in Urbanization and Land Use Efficiency Across the Philippines: A Comprehensive Analysis Using Global Earth Observation Data and SDG 11.3.1 Indicators 评估菲律宾各地城市化和土地使用效率的模式和趋势:利用全球地球观测数据和可持续发展目标 11.3.1 指标进行综合分析
IF 4.1 4区 地球科学 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2024-08-13 DOI: 10.1007/s41064-024-00305-y
Jojene R. Santillan, Christian Heipke

Urbanization, a global phenomenon with profound implications for sustainable development, is a focal point of Sustainable Development Goal 11 (SDG 11). Aimed at fostering inclusive, resilient, and sustainable urbanization by 2030, SDG 11 emphasizes the importance of monitoring land use efficiency (LUE) through indicator 11.3.1. In the Philippines, urbanization has surged over recent decades. Despite its importance, research on urbanization and LUE has predominantly focused on the country’s national capital region (Metro Manila), while little to no attention is given to comprehensive investigations across different regions, provinces, cities, and municipalities of the country. Additionally, challenges in acquiring consistent spatial data, especially due to the Philippines’ archipelagic nature, have hindered comprehensive analysis. To address these gaps, this study conducts a thorough examination of urbanization patterns and LUE dynamics in the Philippines from 1975 to 2020, leveraging Global Human Settlement Layers (GHSL) data and secondary indicators associated with SDG 11.3.1. Our study examines spatial patterns and temporal trends in built-up area expansion, population growth, and LUE characteristics at both city and municipal levels. Among the major findings are the substantial growth in built-up areas and population across the country. We also found a shift in urban growth dynamics, with Metro Manila showing limited expansion in recent years while new urban growth emerges in other regions of the country. Our analysis of the spatiotemporal patterns of Land Consumption Rate (LCR) revealed three distinct evolutional phases: a growth phase between 1975–1990, followed by a decline phase between 1990–2005, and a resurgence phase from 2005–2020. Generally declining trends in LCR and Population Growth Rate (PGR) were evident, demonstrating the country’s direction towards efficient built-up land utilization. However, this efficiency coincides with overcrowding issues as revealed by additional indicators such as the Abstract Achieved Population Density in Expansion Areas (AAPDEA) and Marginal Land Consumption per New Inhabitant (MLCNI). We also analyzed the spatial patterns and temporal trends of LUE across the country and found distinct clusters of transitioning urban centers, densely inhabited metropolises, expanding metropolitan regions, and rapidly growing urban hubs. The study’s findings suggest the need for policy interventions that promote compact and sustainable urban development, equitable regional development, and measures to address overcrowding in urban areas. By aligning policies with the observed spatial and temporal trends, decision-makers can work towards achieving SDG 11, fostering inclusive, resilient, and sustainable urbanization in the Philippines.

城市化是一个对可持续发展具有深远影响的全球现象,是可持续发展目标 11(SDG 11)的一个焦点。可持续发展目标 11 旨在到 2030 年促进包容、有韧性和可持续的城市化,通过指标 11.3.1 强调了监测土地使用效率(LUE)的重要性。近几十年来,菲律宾的城市化进程迅猛发展。尽管城市化和土地使用效率非常重要,但有关城市化和土地使用效率的研究却主要集中在国家首都地区(大马尼拉),而对全国不同地区、省、市和直辖市的全面调查却几乎没有给予关注。此外,在获取一致的空间数据方面存在的挑战,尤其是菲律宾的群岛性质,也阻碍了综合分析的进行。为了弥补这些不足,本研究利用全球人类住区图层(GHSL)数据和与可持续发展目标 11.3.1 相关的二级指标,对 1975 年至 2020 年菲律宾的城市化模式和土地使用效率动态进行了全面研究。我们的研究考察了城市和市镇层面建成区扩张、人口增长和土地利用效率特征的空间模式和时间趋势。主要发现包括全国范围内建成区和人口的大幅增长。我们还发现城市增长动态发生了变化,近年来大马尼拉市的扩张有限,而全国其他地区则出现了新的城市增长。我们对土地消耗率(LCR)时空模式的分析表明了三个不同的演变阶段:1975-1990 年为增长阶段,1990-2005 年为下降阶段,2005-2020 年为恢复阶段。土地消耗率(LCR)和人口增长率(PGR)总体呈下降趋势,这表明该国正朝着高效利用建筑用地的方向发展。然而,这种效率与过度拥挤问题同时存在,扩展区人口密度(AAPDEA)和每新增居民边际土地消耗(MLCNI)等其他指标也揭示了这一问题。我们还分析了全国土地利用效率的空间模式和时间趋势,发现了明显的城市中心转型群、人口稠密的大都市、扩张中的大都市地区和快速发展的城市中心。研究结果表明,有必要采取政策干预措施,促进紧凑和可持续的城市发展、公平的区域发展以及解决城市地区过度拥挤问题的措施。通过使政策与观察到的空间和时间趋势相一致,决策者可以努力实现可持续发展目标 11,促进菲律宾的包容性、弹性和可持续城市化。
{"title":"Assessing Patterns and Trends in Urbanization and Land Use Efficiency Across the Philippines: A Comprehensive Analysis Using Global Earth Observation Data and SDG 11.3.1 Indicators","authors":"Jojene R. Santillan, Christian Heipke","doi":"10.1007/s41064-024-00305-y","DOIUrl":"https://doi.org/10.1007/s41064-024-00305-y","url":null,"abstract":"<p>Urbanization, a global phenomenon with profound implications for sustainable development, is a focal point of Sustainable Development Goal 11 (SDG 11). Aimed at fostering inclusive, resilient, and sustainable urbanization by 2030, SDG 11 emphasizes the importance of monitoring land use efficiency (LUE) through indicator 11.3.1. In the Philippines, urbanization has surged over recent decades. Despite its importance, research on urbanization and LUE has predominantly focused on the country’s national capital region (Metro Manila), while little to no attention is given to comprehensive investigations across different regions, provinces, cities, and municipalities of the country. Additionally, challenges in acquiring consistent spatial data, especially due to the Philippines’ archipelagic nature, have hindered comprehensive analysis. To address these gaps, this study conducts a thorough examination of urbanization patterns and LUE dynamics in the Philippines from 1975 to 2020, leveraging Global Human Settlement Layers (GHSL) data and secondary indicators associated with SDG 11.3.1. Our study examines spatial patterns and temporal trends in built-up area expansion, population growth, and LUE characteristics at both city and municipal levels. Among the major findings are the substantial growth in built-up areas and population across the country. We also found a shift in urban growth dynamics, with Metro Manila showing limited expansion in recent years while new urban growth emerges in other regions of the country. Our analysis of the spatiotemporal patterns of Land Consumption Rate (LCR) revealed three distinct evolutional phases: a growth phase between 1975–1990, followed by a decline phase between 1990–2005, and a resurgence phase from 2005–2020. Generally declining trends in LCR and Population Growth Rate (PGR) were evident, demonstrating the country’s direction towards efficient built-up land utilization. However, this efficiency coincides with overcrowding issues as revealed by additional indicators such as the Abstract Achieved Population Density in Expansion Areas (AAPDEA) and Marginal Land Consumption per New Inhabitant (MLCNI). We also analyzed the spatial patterns and temporal trends of LUE across the country and found distinct clusters of transitioning urban centers, densely inhabited metropolises, expanding metropolitan regions, and rapidly growing urban hubs. The study’s findings suggest the need for policy interventions that promote compact and sustainable urban development, equitable regional development, and measures to address overcrowding in urban areas. By aligning policies with the observed spatial and temporal trends, decision-makers can work towards achieving SDG 11, fostering inclusive, resilient, and sustainable urbanization in the Philippines.</p>","PeriodicalId":56035,"journal":{"name":"PFG-Journal of Photogrammetry Remote Sensing and Geoinformation Science","volume":"7 1","pages":""},"PeriodicalIF":4.1,"publicationDate":"2024-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142200720","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
PFG-Journal of Photogrammetry Remote Sensing and Geoinformation Science
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1