首页 > 最新文献

ISPRS Journal of Photogrammetry and Remote Sensing最新文献

英文 中文
A temporal-spatial deep learning network for winter wheat mapping using time-series Sentinel-2 imagery 利用时间序列 Sentinel-2 图像绘制冬小麦地图的时空深度学习网络
IF 12.7 1区 地球科学 Q1 Computer Science Pub Date : 2024-06-14 DOI: 10.1016/j.isprsjprs.2024.06.005
Lingling Fan , Lang Xia , Jing Yang , Xiao Sun , Shangrong Wu , Bingwen Qiu , Jin Chen , Wenbin Wu , Peng Yang

Accurate mapping of winter wheat provides essential information for food security and ecosystem protection. Deep learning approaches have achieved promising crop discrimination performance based on multitemporal satellite imagery. However, due to the high dimensionality of the data, sequential relations, and complex semantic information in time-series imagery, effective methods that can automatically capture temporal-spatial features with high separability and generalizability have received less attention. In this study, we proposed a U-shaped CNN-Transformer hybrid framework based on an attention mechanism, named the U-Temporal-Spatial-Transformer network (UTS-Former), for winter wheat mapping using Sentinel-2 imagery. This model includes an “encoder-decoder” structure for multiscale information mining of time series images and a temporal-spatial transformer module (TST) for learning comprehensive temporal sequence features and spatial semantic information. The results obtained from two study areas indicated that our UTS-Former achieved the best accuracy, with a mean MCC of 0.928 and an F1-score of 0.950, and the results of different band combinations also showed better performance than other popular time-series methods. We found that the MCC (MCC/All) of the UTS-Former using only RGB bands decreased by 4.53 %, while it decreased by 13.36 % and 35.02 % for UNet2d-LSTM and CNN-BiLSTM, respectively, compared with that of all the band combinations. The comparison demonstrated that the proposed UTS-Former could capture more global temporal-spatial information from winter wheat fields and achieve greater precision in terms of local details than other methods, resulting in high-quality mapping. The analysis of attention scores for the available acquisition dates revealed significant contributions of both beginning and ending growth images in winter wheat mapping, which is valuable for making appropriate selections of image dates. These findings suggest that the proposed approach has great potential for accurate, cost-effective, and high-quality winter wheat mapping.

冬小麦的精确测绘为粮食安全和生态系统保护提供了重要信息。基于多时相卫星图像的深度学习方法已经取得了良好的作物判别性能。然而,由于时间序列图像中的数据维度高、顺序关系和语义信息复杂,能够自动捕捉具有高分离性和可泛化性的时空特征的有效方法受到的关注较少。在本研究中,我们提出了一种基于注意力机制的 U 型 CNN-Transformer 混合框架,命名为 U-Temporal-Spatial-Transformer 网络(UTS-Former),用于利用哨兵-2 图像绘制冬小麦图。该模型包括一个用于时间序列图像多尺度信息挖掘的 "编码器-解码器 "结构,以及一个用于学习综合时间序列特征和空间语义信息的时空变换器模块(TST)。两个研究领域的结果表明,我们的UTS-Former达到了最佳精度,平均MCC为0.928,F1-score为0.950,不同波段组合的结果也比其他流行的时间序列方法表现更好。我们发现,与所有波段组合相比,仅使用 RGB 波段的UTS-Former 的 MCC(MCC/全部)下降了 4.53%,而 UNet2d-LSTM 和 CNN-BiLSTM 的 MCC(MCC/全部)分别下降了 13.36% 和 35.02%。对比结果表明,与其他方法相比,所提出的UTS-Former能捕捉到更多冬小麦田的全局时空信息,并在局部细节方面达到更高的精度,从而获得高质量的绘图。对现有采集日期的关注度评分分析表明,生长初期和生长末期图像在冬小麦绘图中的贡献都很大,这对合理选择图像日期很有价值。这些研究结果表明,所提出的方法在准确、经济、高质量地绘制冬小麦地图方面具有巨大潜力。
{"title":"A temporal-spatial deep learning network for winter wheat mapping using time-series Sentinel-2 imagery","authors":"Lingling Fan ,&nbsp;Lang Xia ,&nbsp;Jing Yang ,&nbsp;Xiao Sun ,&nbsp;Shangrong Wu ,&nbsp;Bingwen Qiu ,&nbsp;Jin Chen ,&nbsp;Wenbin Wu ,&nbsp;Peng Yang","doi":"10.1016/j.isprsjprs.2024.06.005","DOIUrl":"https://doi.org/10.1016/j.isprsjprs.2024.06.005","url":null,"abstract":"<div><p>Accurate mapping of winter wheat provides essential information for food security and ecosystem protection. Deep learning approaches have achieved promising crop discrimination performance based on multitemporal satellite imagery. However, due to the high dimensionality of the data, sequential relations, and complex semantic information in time-series imagery, effective methods that can automatically capture temporal-spatial features with high separability and generalizability have received less attention. In this study, we proposed a U-shaped CNN-Transformer hybrid framework based on an attention mechanism, named the U-Temporal-Spatial-Transformer network (UTS-Former), for winter wheat mapping using Sentinel-2 imagery. This model includes an “encoder-decoder” structure for multiscale information mining of time series images and a temporal-spatial transformer module (TST) for learning comprehensive temporal sequence features and spatial semantic information. The results obtained from two study areas indicated that our UTS-Former achieved the best accuracy, with a mean MCC of 0.928 and an F1-score of 0.950, and the results of different band combinations also showed better performance than other popular time-series methods. We found that the MCC (MCC/All) of the UTS-Former using only RGB bands decreased by 4.53 %, while it decreased by 13.36 % and 35.02 % for UNet2d-LSTM and CNN-BiLSTM, respectively, compared with that of all the band combinations. The comparison demonstrated that the proposed UTS-Former could capture more global temporal-spatial information from winter wheat fields and achieve greater precision in terms of local details than other methods, resulting in high-quality mapping. The analysis of attention scores for the available acquisition dates revealed significant contributions of both beginning and ending growth images in winter wheat mapping, which is valuable for making appropriate selections of image dates. These findings suggest that the proposed approach has great potential for accurate, cost-effective, and high-quality winter wheat mapping.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":null,"pages":null},"PeriodicalIF":12.7,"publicationDate":"2024-06-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141325196","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Legged robot-aided 3D tunnel mapping via residual compensation and anomaly detection 通过残差补偿和异常检测实现支腿机器人辅助三维隧道测绘
IF 12.7 1区 地球科学 Q1 Computer Science Pub Date : 2024-06-14 DOI: 10.1016/j.isprsjprs.2024.05.025
Xing Zhang , Zhanpeng Huang , Qingquan Li , Ruisheng Wang , Baoding Zhou

Three-dimensional (3D) mapping is important to achieve early warning for construction safety and support the long-term safety maintenance of tunnels. However, generating 3D point cloud maps of excavation tunnels that tend to be deficient in features, have rough lining structures, and suffer from dynamic construction interference, can be a challenging task. In this paper, we propose a novel legged robot-aided 3D tunnel mapping method to address the influence of point clouds in the mapping phase. First, a method of kinematic model construction that integrates information from both the robot’s motors and the inertial measurement unit (IMU) is proposed to correct the motion distortion of point clouds. Then, a residual compensation model for unreliable regions (abbreviated as the URC model) is proposed to eliminate the inherent alignment errors in the 3D structures. The structural regions of a tunnel are divided into different reliabilities using the K-means method, and an inherent alignment metric is compensated based on region residual estimation. The compensated alignment metric is then incorporated into a rotation-guided anomaly consistency detection (RAD) model. An isolation forest-based anomaly consistency indicator is designed to remove anomalous light detection and ranging (LiDAR) points and reduce sensor noise caused by ultralong distances. To verify the proposed method, we conduct numerous experiments in three tunnels, namely, a drilling and blasting tunnel, a TBM tunnel, and an underground pedestrian tunnel. According to the experimental results, the proposed method achieves 0.84 ‰, 0.40 ‰, and 0.31 ‰ closure errors (CEs) for the three tunnels, respectively, and the absolute map error (AME) and relative map error (RME) are approximately 1.45 cm and 0.57 %, respectively. The trajectory estimation and mapping errors of our method are smaller than those of existing methods, such as FAST-LIO2, Faster-LIO and LiLi-OM. In addition, ablation tests are conducted to further reveal the roles of the different models used in our method for legged robot-aided 3D mapping in tunnels.

三维(3D)测绘对于实现施工安全预警和支持隧道的长期安全维护非常重要。然而,开挖隧道往往特征不全、衬砌结构粗糙,并且受到动态施工干扰,因此生成开挖隧道的三维点云图是一项具有挑战性的任务。本文提出了一种新型的腿式机器人辅助三维隧道测绘方法,以解决测绘阶段点云的影响问题。首先,我们提出了一种集成机器人电机和惯性测量单元(IMU)信息的运动学模型构建方法,以纠正点云的运动失真。然后,提出了不可靠区域的残余补偿模型(简称为 URC 模型),以消除三维结构中固有的对齐误差。利用 K-means 方法将隧道的结构区域划分为不同的可靠度,并根据区域残差估算补偿固有的对齐度量。然后将补偿后的对齐度量纳入旋转引导的异常一致性检测(RAD)模型。我们设计了一种基于隔离林的异常一致性指标,以去除异常光探测和测距(LiDAR)点,并减少超长距离造成的传感器噪声。为了验证所提出的方法,我们在三条隧道(钻爆隧道、TBM 隧道和地下人行隧道)中进行了大量实验。实验结果表明,所提方法在三条隧道中的闭合误差(CE)分别为 0.84 ‰、0.40 ‰ 和 0.31 ‰,绝对地图误差(AME)和相对地图误差(RME)分别约为 1.45 cm 和 0.57 %。与 FAST-LIO2、Faster-LIO 和 LiLi-OM 等现有方法相比,我们的方法的轨迹估计和绘图误差更小。此外,我们还进行了烧蚀测试,以进一步揭示我们的方法中使用的不同模型在隧道中的腿部机器人辅助三维绘图中的作用。
{"title":"Legged robot-aided 3D tunnel mapping via residual compensation and anomaly detection","authors":"Xing Zhang ,&nbsp;Zhanpeng Huang ,&nbsp;Qingquan Li ,&nbsp;Ruisheng Wang ,&nbsp;Baoding Zhou","doi":"10.1016/j.isprsjprs.2024.05.025","DOIUrl":"https://doi.org/10.1016/j.isprsjprs.2024.05.025","url":null,"abstract":"<div><p>Three-dimensional (3D) mapping is important to achieve early warning for construction safety and support the long-term safety maintenance of tunnels. However, generating 3D point cloud maps of excavation tunnels that tend to be deficient in features, have rough lining structures, and suffer from dynamic construction interference, can be a challenging task. In this paper, we propose a novel legged robot-aided 3D tunnel mapping method to address the influence of point clouds in the mapping phase. First, a method of kinematic model construction that integrates information from both the robot’s motors and the inertial measurement unit (IMU) is proposed to correct the motion distortion of point clouds. Then, a residual compensation model for unreliable regions (abbreviated as the URC model) is proposed to eliminate the inherent alignment errors in the 3D structures. The structural regions of a tunnel are divided into different reliabilities using the K-means method, and an inherent alignment metric is compensated based on region residual estimation. The compensated alignment metric is then incorporated into a rotation-guided anomaly consistency detection (RAD) model. An isolation forest-based anomaly consistency indicator is designed to remove anomalous light detection and ranging (LiDAR) points and reduce sensor noise caused by ultralong distances. To verify the proposed method, we conduct numerous experiments in three tunnels, namely, a drilling and blasting tunnel, a TBM tunnel, and an underground pedestrian tunnel. According to the experimental results, the proposed method achieves 0.84 ‰, 0.40 ‰, and 0.31 ‰ closure errors (CEs) for the three tunnels, respectively, and the absolute map error (AME) and relative map error (RME) are approximately 1.45 cm and 0.57 %, respectively. The trajectory estimation and mapping errors of our method are smaller than those of existing methods, such as FAST-LIO2, Faster-LIO and LiLi-OM. In addition, ablation tests are conducted to further reveal the roles of the different models used in our method for legged robot-aided 3D mapping in tunnels.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":null,"pages":null},"PeriodicalIF":12.7,"publicationDate":"2024-06-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141325195","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HRVQA: A Visual Question Answering benchmark for high-resolution aerial images HRVQA:高分辨率航空图像的视觉问题解答基准
IF 12.7 1区 地球科学 Q1 Computer Science Pub Date : 2024-06-14 DOI: 10.1016/j.isprsjprs.2024.06.002
Kun Li , George Vosselman , Michael Ying Yang

Visual question answering (VQA) is an important and challenging multimodal task in computer vision and photogrammetry. Recently, efforts have been made to bring the VQA task to aerial images, due to its potential real-world applications in disaster monitoring, urban planning, and digital earth product generation. However, the development of VQA in this domain is restricted by the huge variation in the appearance, scale, and orientation of the concepts in aerial images, along with the scarcity of well-annotated datasets. In this paper, we introduce a new dataset, HRVQA, which provides a collection of 53,512 aerial images of 1024 × 1024 pixels and semi-automatically generated 1,070,240 QA pairs. To benchmark the understanding capability of VQA models for aerial images, we evaluate the recent methods on the HRVQA dataset. Moreover, we propose a novel model, GFTransformer, with gated attention modules and a mutual fusion module. The experiments show that the proposed dataset is quite challenging, especially the specific attribute-related questions. Our method achieves superior performance in comparison to the previous state-of-the-art approaches. The dataset and the source code are released at https://hrvqa.nl/.

视觉问题解答(VQA)是计算机视觉和摄影测量学中一项重要而具有挑战性的多模态任务。最近,由于航空图像在灾害监测、城市规划和数字地球产品生成方面的潜在实际应用,人们开始努力将 VQA 任务引入航空图像。然而,由于航空图像中的概念在外观、比例和方向上存在巨大差异,加上缺乏注释完善的数据集,VQA 在这一领域的发展受到了限制。在本文中,我们引入了一个新的数据集 HRVQA,该数据集收集了 53,512 幅 1024 × 1024 像素的航空图像,并半自动生成了 1,070,240 个 QA 对。为了衡量 VQA 模型对航空图像的理解能力,我们在 HRVQA 数据集上评估了最近的方法。此外,我们还提出了一种新型模型--GFTransformer,它具有门控注意模块和相互融合模块。实验表明,所提出的数据集相当具有挑战性,尤其是与特定属性相关的问题。与之前最先进的方法相比,我们的方法取得了卓越的性能。数据集和源代码发布于 https://hrvqa.nl/。
{"title":"HRVQA: A Visual Question Answering benchmark for high-resolution aerial images","authors":"Kun Li ,&nbsp;George Vosselman ,&nbsp;Michael Ying Yang","doi":"10.1016/j.isprsjprs.2024.06.002","DOIUrl":"https://doi.org/10.1016/j.isprsjprs.2024.06.002","url":null,"abstract":"<div><p>Visual question answering (VQA) is an important and challenging multimodal task in computer vision and photogrammetry. Recently, efforts have been made to bring the VQA task to aerial images, due to its potential real-world applications in disaster monitoring, urban planning, and digital earth product generation. However, the development of VQA in this domain is restricted by the huge variation in the appearance, scale, and orientation of the concepts in aerial images, along with the scarcity of well-annotated datasets. In this paper, we introduce a new dataset, HRVQA, which provides a collection of 53,512 aerial images of 1024 × 1024 pixels and semi-automatically generated 1,070,240 QA pairs. To benchmark the understanding capability of VQA models for aerial images, we evaluate the recent methods on the HRVQA dataset. Moreover, we propose a novel model, GFTransformer, with gated attention modules and a mutual fusion module. The experiments show that the proposed dataset is quite challenging, especially the specific attribute-related questions. Our method achieves superior performance in comparison to the previous state-of-the-art approaches. The dataset and the source code are released at <span>https://hrvqa.nl/</span><svg><path></path></svg>.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":null,"pages":null},"PeriodicalIF":12.7,"publicationDate":"2024-06-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0924271624002326/pdfft?md5=f0ff0854512e9027f5be14632f4bef39&pid=1-s2.0-S0924271624002326-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141325197","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CloudSeg: A multi-modal learning framework for robust land cover mapping under cloudy conditions CloudSeg:用于在多云条件下绘制稳健土地覆被图的多模式学习框架
IF 12.7 1区 地球科学 Q1 Computer Science Pub Date : 2024-06-10 DOI: 10.1016/j.isprsjprs.2024.06.001
Fang Xu , Yilei Shi , Wen Yang , Gui-Song Xia , Xiao Xiang Zhu

Cloud coverage poses a significant challenge to optical image interpretation, degrading ground information on Earth’s surface. Synthetic aperture radar (SAR), with its ability to penetrate clouds, provides supplementary information to optical data. However, existing optical-SAR fusion methods predominantly focus on cloud-free scenarios, neglecting the practical challenge of semantic segmentation under cloudy conditions. To tackle this issue, we propose CloudSeg, a novel framework tailored for land cover mapping in the presence of clouds. It addresses the challenges posed by cloud cover from two aspects: reducing semantic ambiguity in areas of the cloudy image that are obscured by clouds and enhancing effective information in the unobstructed portions. Specifically, CloudSeg employs a multi-task learning strategy to simultaneously handle low-level visual task and high-level semantic understanding task, mitigating the semantic ambiguity caused by cloud cover by acquiring discriminative features through an auxiliary cloud removal task. Additionally, CloudSeg incorporates a knowledge distillation strategy, which utilizes the knowledge learned by the teacher network under cloud-free conditions to guide the student network to overcome the interference of cloud-covered areas, enhancing the valuable information from the unobstructed parts of cloud-covered images. Extensive experiments conducted on two datasets, M3M-CR and WHU-OPT-SAR, demonstrate the effectiveness and superiority of the proposed CloudSeg method for land cover mapping under cloudy conditions. Specifically, CloudSeg outperforms the state-of-the-art competitors by 3.16% in terms of mIoU on M3M-CR and by 5.56% on WHU-OPT-SAR, highlighting its substantial advantages for analyzing regions frequently obscured by clouds. Codes are available at https://github.com/xufangchn/CloudSeg.

云层覆盖对光学图像判读构成重大挑战,降低了地球表面的地面信息质量。合成孔径雷达(SAR)具有穿透云层的能力,可为光学数据提供补充信息。然而,现有的光学-合成孔径雷达融合方法主要关注无云场景,忽视了在多云条件下进行语义分割的实际挑战。为了解决这个问题,我们提出了 CloudSeg,这是一个专为有云环境下的土地覆盖制图而量身定制的新框架。它从两个方面解决了云层带来的挑战:减少云层图像中被云层遮挡区域的语义模糊性,以及增强未被遮挡部分的有效信息。具体来说,CloudSeg 采用多任务学习策略,同时处理低层次的视觉任务和高层次的语义理解任务,通过辅助的云雾去除任务获取判别特征,从而减轻云层造成的语义模糊。此外,CloudSeg 还采用了知识提炼策略,利用教师网络在无云条件下学习到的知识来指导学生网络克服云层覆盖区域的干扰,从而增强云层覆盖图像中未被遮挡部分的有价值信息。在M3M-CR和WHU-OPT-SAR两个数据集上进行的大量实验证明了所提出的CloudSeg方法在多云条件下绘制土地覆盖图的有效性和优越性。具体来说,在 M3M-CR 和 WHU-OPT-SAR 两个数据集上,CloudSeg 的 mIoU 分别比最先进的竞争对手高出 3.16% 和 5.56%,这凸显了它在分析经常被云层遮挡的区域方面的巨大优势。代码见 https://github.com/xufangchn/CloudSeg。
{"title":"CloudSeg: A multi-modal learning framework for robust land cover mapping under cloudy conditions","authors":"Fang Xu ,&nbsp;Yilei Shi ,&nbsp;Wen Yang ,&nbsp;Gui-Song Xia ,&nbsp;Xiao Xiang Zhu","doi":"10.1016/j.isprsjprs.2024.06.001","DOIUrl":"https://doi.org/10.1016/j.isprsjprs.2024.06.001","url":null,"abstract":"<div><p>Cloud coverage poses a significant challenge to optical image interpretation, degrading ground information on Earth’s surface. Synthetic aperture radar (SAR), with its ability to penetrate clouds, provides supplementary information to optical data. However, existing optical-SAR fusion methods predominantly focus on cloud-free scenarios, neglecting the practical challenge of semantic segmentation under cloudy conditions. To tackle this issue, we propose CloudSeg, a novel framework tailored for land cover mapping in the presence of clouds. It addresses the challenges posed by cloud cover from two aspects: reducing semantic ambiguity in areas of the cloudy image that are obscured by clouds and enhancing effective information in the unobstructed portions. Specifically, CloudSeg employs a multi-task learning strategy to simultaneously handle low-level visual task and high-level semantic understanding task, mitigating the semantic ambiguity caused by cloud cover by acquiring discriminative features through an auxiliary cloud removal task. Additionally, CloudSeg incorporates a knowledge distillation strategy, which utilizes the knowledge learned by the teacher network under cloud-free conditions to guide the student network to overcome the interference of cloud-covered areas, enhancing the valuable information from the unobstructed parts of cloud-covered images. Extensive experiments conducted on two datasets, <em>M3M-CR</em> and <em>WHU-OPT-SAR</em>, demonstrate the effectiveness and superiority of the proposed CloudSeg method for land cover mapping under cloudy conditions. Specifically, CloudSeg outperforms the state-of-the-art competitors by 3.16% in terms of mIoU on <em>M3M-CR</em> and by 5.56% on <em>WHU-OPT-SAR</em>, highlighting its substantial advantages for analyzing regions frequently obscured by clouds. Codes are available at <span>https://github.com/xufangchn/CloudSeg</span><svg><path></path></svg>.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":null,"pages":null},"PeriodicalIF":12.7,"publicationDate":"2024-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141303571","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Simultaneous invariant normalization of waveform features from bathymetric lidar, SINWav: A Saipan case study 测深激光雷达波形特征的同步不变归一化,SINWav:塞班岛案例研究
IF 12.7 1区 地球科学 Q1 Computer Science Pub Date : 2024-06-08 DOI: 10.1016/j.isprsjprs.2024.05.024
Jaehoon Jung , Christopher E. Parrish , Bryan Costa , Suhong Yoo

Over the past two decades, a major advance that enabled airborne bathymetric lidar to benefit a much wider range of marine science applications was the development of procedures for creating seafloor reflectance mosaics from recorded intensity data. It was recognized that intensity data, derived from the amplitudes of laser returns from the seafloor, contained information related to seafloor albedo and composition. However, the raw intensity data were also found to be related to a number of nuisance parameters, such that, when grided, they exhibited discontinuities, seamlines and other artifacts, hindering their use in benthic habitat mapping. These realizations led to the development of tools and workflows for correcting lidar intensity data to produce seamless seafloor reflectance mosaics. At present, an opportunity exists for another major advance in airborne bathymetric lidar by utilizing not only intensity data, but a large suite of waveform features that describe the shape of the return signal from the seafloor, to characterize benthic habitats and perform ecological assessments. However, similar to raw intensity data, other waveform features exhibit salient discontinuities, seamlines, and other artifacts, if uncorrected. Furthermore, in contrast to the case of intensity data, little work has been done on correction of an entire suite of waveform features to create a set of seamless seafloor mosaics. This study aims to address this need through a novel normalization method that integrates two image blending techniques: Gaussian weighted color matching and Laplacian pyramid blending. The proposed approach, Simultaneous Invariant Normalization of Waveform Features (SINWav), is designed to be invariant to the type of input waveform features, such that feature-specific tuning is unnecessary. To handle vast amounts of data efficiently, we developed a memory-efficient sparse matrix representation. The methods were applied to bathymetric lidar data from Saipan containing 16 different waveform features. Both visual assessments and quantitative analyses using quality metrics indicated that the proposed approach outperforms results derived from raw data and conventional linear transform.

在过去的二十年里,机载测深激光雷达取得了一项重大进展,使更广泛的海洋科学应用受益,这就是开发了从记录的强度数据创建海底反射率镶嵌图的程序。人们认识到,从海底激光回波振幅得出的强度数据包含与海底反照率和成分有关的信息。然而,人们还发现原始的强度数据与一些干扰参数有关,因此,在进行网格划分时,它们会显示出不连续性、缝合线和其他伪影,从而妨碍了它们在海底生境绘图中的应用。认识到这一点后,开发了校正激光雷达强度数据的工具和工作流程,以生成无缝海底反射率镶嵌图。目前,机载测深激光雷达不仅可以利用强度数据,还可以利用描述海底回波信号形状的大量波形特征来描述海底生境特征和进行生态评估,从而有机会取得另一项重大进展。然而,与原始强度数据类似,其他波形特征如果未经校正,也会表现出明显的不连续性、缝合线和其他伪影。此外,与强度数据的情况不同,目前还很少有人对整套波形特征进行校正,以创建一套无缝海底镶嵌图。本研究旨在通过整合两种图像混合技术的新型归一化方法来满足这一需求:高斯加权颜色匹配和拉普拉斯金字塔混合。所提出的方法--波形特征同步不变归一化(SINWav)--旨在对输入波形特征的类型保持不变,因此无需对特定特征进行调整。为了高效处理海量数据,我们开发了一种内存效率高的稀疏矩阵表示法。我们将这些方法应用于塞班岛的测深激光雷达数据,其中包含 16 种不同的波形特征。直观评估和使用质量指标的定量分析都表明,所提出的方法优于通过原始数据和传统线性变换得出的结果。
{"title":"Simultaneous invariant normalization of waveform features from bathymetric lidar, SINWav: A Saipan case study","authors":"Jaehoon Jung ,&nbsp;Christopher E. Parrish ,&nbsp;Bryan Costa ,&nbsp;Suhong Yoo","doi":"10.1016/j.isprsjprs.2024.05.024","DOIUrl":"https://doi.org/10.1016/j.isprsjprs.2024.05.024","url":null,"abstract":"<div><p>Over the past two decades, a major advance that enabled airborne bathymetric lidar to benefit a much wider range of marine science applications was the development of procedures for creating seafloor reflectance mosaics from recorded intensity data. It was recognized that intensity data, derived from the amplitudes of laser returns from the seafloor, contained information related to seafloor albedo and composition. However, the raw intensity data were also found to be related to a number of nuisance parameters, such that, when grided, they exhibited discontinuities, seamlines and other artifacts, hindering their use in benthic habitat mapping. These realizations led to the development of tools and workflows for correcting lidar intensity data to produce seamless seafloor reflectance mosaics. At present, an opportunity exists for another major advance in airborne bathymetric lidar by utilizing not only intensity data, but a large suite of waveform features that describe the shape of the return signal from the seafloor, to characterize benthic habitats and perform ecological assessments. However, similar to raw intensity data, other waveform features exhibit salient discontinuities, seamlines, and other artifacts, if uncorrected. Furthermore, in contrast to the case of intensity data, little work has been done on correction of an entire suite of waveform features to create a set of seamless seafloor mosaics. This study aims to address this need through a novel normalization method that integrates two image blending techniques: Gaussian weighted color matching and Laplacian pyramid blending. The proposed approach, Simultaneous Invariant Normalization of Waveform Features (SINWav), is designed to be invariant to the type of input waveform features, such that feature-specific tuning is unnecessary. To handle vast amounts of data efficiently, we developed a memory-efficient sparse matrix representation. The methods were applied to bathymetric lidar data from Saipan containing 16 different waveform features. Both visual assessments and quantitative analyses using quality metrics indicated that the proposed approach outperforms results derived from raw data and conventional linear transform.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":null,"pages":null},"PeriodicalIF":12.7,"publicationDate":"2024-06-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0924271624002284/pdfft?md5=25922c37d31f077da03bf2525d3fd035&pid=1-s2.0-S0924271624002284-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141294177","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TSI-Siamnet: A Siamese network for cloud and shadow detection based on time-series cloudy images TSI-Siamnet:基于时间序列多云图像的云影检测连体网络
IF 12.7 1区 地球科学 Q1 Computer Science Pub Date : 2024-06-05 DOI: 10.1016/j.isprsjprs.2024.05.022
Qunming Wang , Jiayi Li , Xiaohua Tong , Peter M. Atkinson

Accurate cloud and shadow detection is a crucial prerequisite for optical remote sensing image analysis and application. Multi-temporal-based cloud and shadow detection methods are a preferable choice to detect clouds in complex scenes (e.g., thin clouds, broken clouds and clouds with interference from artificial surfaces with high reflectivity). However, such methods commonly require cloud-free reference images, and this may be difficult to achieve in time-series data since clouds are often prevalent and of varying spatial distribution in optical remote sensing images. Furthermore, current multi-temporal-based methods have limited feature extraction capability and rely heavily on prior assumptions. To address these issues, this paper proposes a Siamese network (Siamnet) for cloud and shadow detection based on Time-Series cloudy Images, namely TSI-Siamnet, which consists of two steps: 1) low-rank and sparse component decomposition of time-series cloudy images is conducted to construct a composite reference image to cope with dynamic changes in the cloud distribution in time-series images; 2) an extended Siamnet with optimal difference calculation module (DM) and multi-scale difference features fusion module (MDFM) is constructed to extract reliable disparity features and alleviate semantic information feature dilution during the decoder part. TSI-Siamnet was tested extensively on seven land cover types in the well-known Landsat 8 Biome dataset. Compared to six state-of-the-art methods (including four deep learning-based methods and two classical non-deep learning-based methods), TSI-Siamnet produced the best performance with an overall accuracy of 95.05% and MIoU of 84.37%. In three more challenging experiments, TSI-Siamnet showed enhanced detection of thin and broken clouds and greater anti-interference to highly reflective surfaces. TSI-Siamnet provides a novel strategy to explore comprehensively the valid information in time-series cloudy images and integrate the extracted spectral-spatial–temporal features for reliable cloud and shadow detection.

准确的云影探测是光学遥感图像分析和应用的重要前提。基于多时相的云影检测方法是在复杂场景(如薄云、破碎云和受到高反射率人工表面干扰的云)中检测云的首选方法。然而,这类方法通常需要无云的参考图像,而这在时间序列数据中可能很难实现,因为云在光学遥感图像中通常很普遍,而且空间分布各不相同。此外,目前基于多时相的方法的特征提取能力有限,而且严重依赖于先验假设。为了解决这些问题,本文提出了一种基于时间序列多云图像的云影检测连体网络(Siamnet),即 TSI-Siamnet,它包括两个步骤:1) 对时间序列多云图像进行低秩和稀疏分量分解,构建复合参考图像,以应对时间序列图像中云层分布的动态变化;2) 构建带有最优差分计算模块(DM)和多尺度差分特征融合模块(MDFM)的扩展 Siamnet,以提取可靠的差分特征,减轻解码器部分对语义信息特征的稀释。TSI-Siamnet 在著名的 Landsat 8 生物群落数据集中的七种土地覆被类型上进行了广泛测试。与六种最先进的方法(包括四种基于深度学习的方法和两种基于非深度学习的经典方法)相比,TSI-Siamnet 的性能最佳,总体准确率为 95.05%,MIoU 为 84.37%。在三个更具挑战性的实验中,TSI-Siamnet 增强了对薄云和碎云的检测,并提高了对高反射表面的抗干扰能力。TSI-Siamnet 提供了一种新颖的策略,可全面探索时间序列多云图像中的有效信息,并整合提取的光谱-空间-时间特征,从而实现可靠的云和阴影检测。
{"title":"TSI-Siamnet: A Siamese network for cloud and shadow detection based on time-series cloudy images","authors":"Qunming Wang ,&nbsp;Jiayi Li ,&nbsp;Xiaohua Tong ,&nbsp;Peter M. Atkinson","doi":"10.1016/j.isprsjprs.2024.05.022","DOIUrl":"https://doi.org/10.1016/j.isprsjprs.2024.05.022","url":null,"abstract":"<div><p>Accurate cloud and shadow detection is a crucial prerequisite for optical remote sensing image analysis and application. Multi-temporal-based cloud and shadow detection methods are a preferable choice to detect clouds in complex scenes (e.g., thin clouds, broken clouds and clouds with interference from artificial surfaces with high reflectivity). However, such methods commonly require cloud-free reference images, and this may be difficult to achieve in time-series data since clouds are often prevalent and of varying spatial distribution in optical remote sensing images. Furthermore, current multi-temporal-based methods have limited feature extraction capability and rely heavily on prior assumptions. To address these issues, this paper proposes a Siamese network (Siamnet) for cloud and shadow detection based on Time-Series cloudy Images, namely TSI-Siamnet, which consists of two steps: 1) low-rank and sparse component decomposition of time-series cloudy images is conducted to construct a composite reference image to cope with dynamic changes in the cloud distribution in time-series images; 2) an extended Siamnet with optimal difference calculation module (DM) and multi-scale difference features fusion module (MDFM) is constructed to extract reliable disparity features and alleviate semantic information feature dilution during the decoder part. TSI-Siamnet was tested extensively on seven land cover types in the well-known Landsat 8 Biome dataset. Compared to six state-of-the-art methods (including four deep learning-based methods and two classical non-deep learning-based methods), TSI-Siamnet produced the best performance with an overall accuracy of 95.05% and MIoU of 84.37%. In three more challenging experiments, TSI-Siamnet showed enhanced detection of thin and broken clouds and greater anti-interference to highly reflective surfaces. TSI-Siamnet provides a novel strategy to explore comprehensively the valid information in time-series cloudy images and integrate the extracted spectral-spatial–temporal features for reliable cloud and shadow detection.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":null,"pages":null},"PeriodicalIF":12.7,"publicationDate":"2024-06-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141250807","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Evaluating the spatial–temporal transferability of models for agricultural land cover mapping using Landsat archive 评估利用大地遥感卫星档案绘制农业土地覆被图的模型的时空可转移性
IF 12.7 1区 地球科学 Q1 Computer Science Pub Date : 2024-05-31 DOI: 10.1016/j.isprsjprs.2024.05.020
Jayan Wijesingha, Ilze Dzene, Michael Wachendorf

Changes in policy and new plans can significantly influence land use and trigger land use change in the long term. The data for pre- and post-policy implementation is necessary to assess the specific policy’s impact on land use. In the early nineties, Germany started promoting renewable energy production, including bioenergy, which changed the agricultural landscape. Remote sensing (RS) image-based machine learning models can be beneficial for mapping agricultural land use in the present and the past. However, machine learning classification models trained on RS data from specific training sites and time may not be able to predict data for unknown sites and unknown temporal points due to changes in crop phenology, field features, or ecological site circumstances because most of the models are limited in their performances according to variations of the training data set. Therefore, this study aims to assess the spatial–temporal transferability of Landsat-based agricultural land use type classification. The study was developed to map agricultural land cover (5 classes: maize, grasslands, summer crops, winter crops, and mixed crops) in two regions in Germany (North Hesse and Weser-Ems) between the years 2010 and 2018 using Landsat archive data (i.e., Landsat 5, 7, and 8). Two machine learning models (random forest − RF and 2D convolution neural network – 2DCNN) were trained and evaluated according to no transferability (reference) scenario and three spatial–temporal scenarios using mF1 and class level F1 values. Three model transferability scenarios were evaluated: a) temporal – S1, b) spatial – S2, and c) spatiotemporal – S3. The reference scenario, without transferability, achieved an overall accuracy of 89.1% and a macro F1 score of 0.74 for RF and 89.9% and 0.75 for CNN, respectively. Under three transferability scenarios (S1, S2, and S3), the macro F1 scores decreased to 0.67, 0.66, and 0.62 for RF, and 0.68, 0.62, and 0.58 for CNN, respectively. The dissimilarity between the data employed to train the model and data from the new domain indicated a clear link that could explain the reduction in model predictability. Moreover, the performance degradation could be attributed to the disparity in environmental, climatic, and crop calendar conditions between the two domains. Understanding the extent of model performance degradation during transferability is crucial for developing effective strategies to mitigate these issues and enhance the generalisability of machine learning models for agriculture land cover mapping.

政策和新规划的变化会对土地利用产生重大影响,并引发土地利用的长期变化。要评估具体政策对土地利用的影响,就必须获得政策实施前后的数据。九十年代初,德国开始推广可再生能源生产,包括生物能源,这改变了农业景观。基于遥感(RS)图像的机器学习模型可用于绘制当前和过去的农业土地利用图。然而,根据特定训练地点和时间的 RS 数据训练的机器学习分类模型可能无法预测未知地点和未知时间点的数据,原因是作物物候、田间特征或生态地点环境的变化,因为大多数模型的性能会因训练数据集的变化而受到限制。因此,本研究旨在评估基于大地遥感卫星的农用地类型分类的时空可转移性。该研究利用大地遥感卫星档案数据(即大地遥感卫星 5 号、7 号和 8 号)绘制了 2010 年至 2018 年期间德国两个地区(北黑森州和威悉河-埃姆斯州)的农业用地覆盖图(5 个类别:玉米、草地、夏季作物、冬季作物和混合作物)。使用 mF1 和类级 F1 值对两种机器学习模型(随机森林 - RF 和二维卷积神经网络 - 2DCNN)进行了训练,并根据无可移植性(参考)情景和三种时空情景进行了评估。评估了三种模型可转移性情景:a) 时间情景--S1;b) 空间情景--S2;c) 时空情景--S3。在没有可移植性的参考方案中,RF 的总体准确率为 89.1%,宏观 F1 得分为 0.74;CNN 的准确率为 89.9%,宏观 F1 得分为 0.75。在三种可转移方案(S1、S2 和 S3)下,RF 的宏观 F1 分数分别降至 0.67、0.66 和 0.62,CNN 的宏观 F1 分数分别降至 0.68、0.62 和 0.58。用于训练模型的数据与来自新领域的数据之间的差异表明,两者之间存在明显的联系,可以解释模型预测能力下降的原因。此外,性能下降还可能是由于两个领域的环境、气候和作物日历条件不同造成的。了解可转移性过程中模型性能下降的程度,对于制定有效策略以缓解这些问题并提高农业土地覆被制图机器学习模型的通用性至关重要。
{"title":"Evaluating the spatial–temporal transferability of models for agricultural land cover mapping using Landsat archive","authors":"Jayan Wijesingha,&nbsp;Ilze Dzene,&nbsp;Michael Wachendorf","doi":"10.1016/j.isprsjprs.2024.05.020","DOIUrl":"https://doi.org/10.1016/j.isprsjprs.2024.05.020","url":null,"abstract":"<div><p>Changes in policy and new plans can significantly influence land use and trigger land use change in the long term. The data for pre- and post-policy implementation is necessary to assess the specific policy’s impact on land use. In the early nineties, Germany started promoting renewable energy production, including bioenergy, which changed the agricultural landscape. Remote sensing (RS) image-based machine learning models can be beneficial for mapping agricultural land use in the present and the past. However, machine learning classification models trained on RS data from specific training sites and time may not be able to predict data for unknown sites and unknown temporal points due to changes in crop phenology, field features, or ecological site circumstances because most of the models are limited in their performances according to variations of the training data set. Therefore, this study aims to assess the spatial–temporal transferability of Landsat-based agricultural land use type classification. The study was developed to map agricultural land cover (5 classes: maize, grasslands, summer crops, winter crops, and mixed crops) in two regions in Germany (North Hesse and Weser-Ems) between the years 2010 and 2018 using Landsat archive data (i.e., Landsat 5, 7, and 8). Two machine learning models (random forest − RF and 2D convolution neural network – 2DCNN) were trained and evaluated according to no transferability (reference) scenario and three spatial–temporal scenarios using mF1 and class level F1 values. Three model transferability scenarios were evaluated: a) temporal – S1, b) spatial – S2, and c) spatiotemporal – S3. The reference scenario, without transferability, achieved an overall accuracy of 89.1% and a macro F1 score of 0.74 for RF and 89.9% and 0.75 for CNN, respectively. Under three transferability scenarios (S1, S2, and S3), the macro F1 scores decreased to 0.67, 0.66, and 0.62 for RF, and 0.68, 0.62, and 0.58 for CNN, respectively. The dissimilarity between the data employed to train the model and data from the new domain indicated a clear link that could explain the reduction in model predictability. Moreover, the performance degradation could be attributed to the disparity in environmental, climatic, and crop calendar conditions between the two domains. Understanding the extent of model performance degradation during transferability is crucial for developing effective strategies to mitigate these issues and enhance the generalisability of machine learning models for agriculture land cover mapping.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":null,"pages":null},"PeriodicalIF":12.7,"publicationDate":"2024-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0924271624002181/pdfft?md5=bac8266465f22f34cfa9e494111b48a8&pid=1-s2.0-S0924271624002181-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141243789","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Incremental registration towards large-scale heterogeneous point clouds by hierarchical graph matching 通过分层图匹配实现大规模异构点云的增量注册
IF 12.7 1区 地球科学 Q1 Computer Science Pub Date : 2024-05-31 DOI: 10.1016/j.isprsjprs.2024.05.017
Shoujun Jia, Chun Liu, Hangbin Wu, Weihua Huan, Shufan Wang

The increasing availability of point cloud acquisition techniques makes it possible to significantly increase 3D observation capacity by the registration of multi-sensor, multi-platform, and multi-temporal point clouds. However, there are geometric heterogeneities (point density variations and point distribution differences), small overlaps (30 % ∼ 50 %), and large data amounts (a few millions) among these large-scale heterogeneous point clouds, which pose great challenges for effective and efficient registration. In this paper, considering the structural representation capacity of graph model, we propose an incremental registration method for large-scale heterogeneous point clouds by hierarchical graph matching. More specifically, we first construct a novel graph model to discriminatively and robustly represent heterogeneous point clouds. In addition to conventional nodes and edges, our graph model particularly designs discriminative and robust feature descriptors for local node description and captures spatial relationships from both locations and orientations for global edge description. We further devise a matching strategy to accurately estimate node matches for our graph models with partial even small overlaps. This effectiveness benefits from the comprehensiveness of node and edge dissimilarities and the constraint of geometric consistency in the optimization objective. On this basis, we design a coarse-to-fine registration framework for effective and efficient point cloud registration. In this incremental framework, graph matching is hierarchically utilized to achieve sparse-to-dense point matching by global extraction and local propagation, which provides dense correspondences for robust coarse registration and predicts overlap ratio for accurate fine registration, and also avoids huge computation costs for large-scale point clouds. Extensive experiments on one benchmark and three changing self-built datasets with large scales, outliers, changing densities, and small overlaps show the excellent transformation and correspondence accuracies of our registration method for large-scale heterogeneous point clouds. Compared to the state-of-the-art methods (i.e., TrimICP, CoBigICP, GROR, VPFBR, DPCR, and PRR), our registration method performs approximate even higher efficiency while achieves an improvement of 33 % − 88 % regarding registration accuracy (OE).

随着点云采集技术的日益普及,通过对多传感器、多平台和多时相点云进行注册,可显著提高三维观测能力。然而,这些大规模异构点云之间存在几何异质性(点密度变化和点分布差异)、小重叠(30% ∼ 50%)和大数据量(数百万),这给有效和高效的配准带来了巨大挑战。本文考虑到图模型的结构表示能力,提出了一种分层图匹配的大规模异构点云增量配准方法。更具体地说,我们首先构建了一个新颖的图模型,以区分并稳健地表示异构点云。除了传统的节点和边缘外,我们的图模型还特别设计了用于局部节点描述的辨别性和鲁棒性特征描述符,并从位置和方向两方面捕捉空间关系,用于全局边缘描述。我们还进一步设计了一种匹配策略,以准确估计具有部分甚至微小重叠的图模型的节点匹配情况。这种有效性得益于节点和边缘差异的全面性以及优化目标中的几何一致性约束。在此基础上,我们设计了一个从粗到细的注册框架,以实现高效的点云注册。在这个增量框架中,图匹配被分层利用,通过全局提取和局部传播实现从稀疏到密集的点匹配,从而为稳健的粗注册提供密集的对应关系,为精确的细注册预测重叠率,同时也避免了大规模点云的巨大计算成本。在一个基准数据集和三个不断变化的自建数据集上进行的大量实验表明,我们的配准方法在大规模异构点云上具有出色的变换和对应精度。与最先进的方法(即 TrimICP、CoBigICP、GROR、VPFBR、DPCR 和 PRR)相比,我们的配准方法效率更高,配准精度(OE)提高了 33% - 88%。
{"title":"Incremental registration towards large-scale heterogeneous point clouds by hierarchical graph matching","authors":"Shoujun Jia,&nbsp;Chun Liu,&nbsp;Hangbin Wu,&nbsp;Weihua Huan,&nbsp;Shufan Wang","doi":"10.1016/j.isprsjprs.2024.05.017","DOIUrl":"https://doi.org/10.1016/j.isprsjprs.2024.05.017","url":null,"abstract":"<div><p>The increasing availability of point cloud acquisition techniques makes it possible to significantly increase 3D observation capacity by the registration of multi-sensor, multi-platform, and multi-temporal point clouds. However, there are geometric heterogeneities (point density variations and point distribution differences), small overlaps (30 % ∼ 50 %), and large data amounts (a few millions) among these large-scale heterogeneous point clouds, which pose great challenges for effective and efficient registration. In this paper, considering the structural representation capacity of graph model, we propose an incremental registration method for large-scale heterogeneous point clouds by hierarchical graph matching. More specifically, we first construct a novel graph model to discriminatively and robustly represent heterogeneous point clouds. In addition to conventional nodes and edges, our graph model particularly designs discriminative and robust feature descriptors for local node description and captures spatial relationships from both locations and orientations for global edge description. We further devise a matching strategy to accurately estimate node matches for our graph models with partial even small overlaps. This effectiveness benefits from the comprehensiveness of node and edge dissimilarities and the constraint of geometric consistency in the optimization objective. On this basis, we design a coarse-to-fine registration framework for effective and efficient point cloud registration. In this incremental framework, graph matching is hierarchically utilized to achieve sparse-to-dense point matching by global extraction and local propagation, which provides dense correspondences for robust coarse registration and predicts overlap ratio for accurate fine registration, and also avoids huge computation costs for large-scale point clouds. Extensive experiments on one benchmark and three changing self-built datasets with large scales, outliers, changing densities, and small overlaps show the excellent transformation and correspondence accuracies of our registration method for large-scale heterogeneous point clouds. Compared to the state-of-the-art methods (<em>i.e.,</em> TrimICP, CoBigICP, GROR, VPFBR, DPCR, and PRR), our registration method performs approximate even higher efficiency while achieves an improvement of 33 % − 88 % regarding registration accuracy (OE).</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":null,"pages":null},"PeriodicalIF":12.7,"publicationDate":"2024-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141240270","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Vectorized building extraction from high-resolution remote sensing images using spatial cognitive graph convolution model 利用空间认知图卷积模型从高分辨率遥感图像中提取矢量化建筑物
IF 12.7 1区 地球科学 Q1 Computer Science Pub Date : 2024-05-31 DOI: 10.1016/j.isprsjprs.2024.05.015
Zhuotong Du , Haigang Sui , Qiming Zhou , Mingting Zhou , Weiyue Shi , Jianxun Wang , Junyi Liu

Traditional approach from source image to application vectors in building extraction needs additional complex regularization of converted intermediate raster results. While in conversion, the lost detailed artifacts, unnecessary nodes, and messy paths would be labor-intensive to repair errors and topological issues, even aside the inherent problems of blob-like objects and blurry, jagged edges in first-stage extraction. This research explores new graph convolution-driven solution, the spatial-cognitive shaping model (SCShaping), to directly access vectorization form of individual buildings through spatial cognitive approximation to coordinates that form building boundaries. To strengthen graph nodes expressivity, this method enriches topological feature embedding travelling along the model architecture along with features contributed from convolutional neural network (CNN) extractor. To stimulate the neighboring aggregation in graphs, Graph-Encoder-Decoder mechanism is introduced to augment feature reuse integrating complementary graph convolution layers. The strong embedding guarantees effective feature tapping and the robust structure guarantees the feature mining. Comparative studies have been conducted between the proposed approach with five other methods on three challenging datasets. The results demonstrate the proposed approach yields unanimous and significant improvements in mask-wise metrics, which evaluate object integrity and accuracy, as well as edge-wise metrics, which assess contour regularity and precision. The outperformance also indicates better multi-scale object adaptability of SCShaping. The obtain-and-play SCShaping commands a pleasurable implementation way to advance ideal manmachine collaboration.

在建筑物提取中,从源图像到应用矢量的传统方法需要对转换后的中间光栅结果进行额外的复杂正则化处理。而在转换过程中,丢失的细节伪影、不必要的节点和杂乱的路径将耗费大量人力物力来修复错误和拓扑问题,甚至还要撇开第一阶段提取中固有的球状物体和模糊锯齿状边缘等问题。本研究探索了新的图卷积驱动解决方案--空间认知塑造模型(SCShaping),通过对建筑物边界坐标的空间认知近似,直接获取单个建筑物的矢量化形式。为了加强图节点的表现力,该方法丰富了沿模型架构行进的拓扑特征嵌入,以及卷积神经网络(CNN)提取器提供的特征。为促进图中的邻近聚合,引入了图-编码器-解码器机制,通过整合互补的图卷积层来增强特征重用性。强嵌入保证了有效的特征挖掘,稳健的结构保证了特征挖掘。在三个具有挑战性的数据集上,对所提出的方法与其他五种方法进行了比较研究。结果表明,在评估对象完整性和准确性的掩膜指标以及评估轮廓规则性和精确性的边缘指标方面,所提出的方法都取得了一致且显著的改进。优异的表现还表明 SCShaping 具有更好的多尺度对象适应性。SCShaping 的 "即取即用 "命令是一种推进理想人机协作的愉悦实施方式。
{"title":"Vectorized building extraction from high-resolution remote sensing images using spatial cognitive graph convolution model","authors":"Zhuotong Du ,&nbsp;Haigang Sui ,&nbsp;Qiming Zhou ,&nbsp;Mingting Zhou ,&nbsp;Weiyue Shi ,&nbsp;Jianxun Wang ,&nbsp;Junyi Liu","doi":"10.1016/j.isprsjprs.2024.05.015","DOIUrl":"https://doi.org/10.1016/j.isprsjprs.2024.05.015","url":null,"abstract":"<div><p>Traditional approach from source image to application vectors in building extraction needs additional complex regularization of converted intermediate raster results. While in conversion, the lost detailed artifacts, unnecessary nodes, and messy paths would be labor-intensive to repair errors and topological issues, even aside the inherent problems of blob-like objects and blurry, jagged edges in first-stage extraction. This research explores new graph convolution-driven solution, the spatial-cognitive shaping model (SCShaping), to directly access vectorization form of individual buildings through spatial cognitive approximation to coordinates that form building boundaries. To strengthen graph nodes expressivity, this method enriches topological feature embedding travelling along the model architecture along with features contributed from convolutional neural network (CNN) extractor. To stimulate the neighboring aggregation in graphs, Graph-Encoder-Decoder mechanism is introduced to augment feature reuse integrating complementary graph convolution layers. The strong embedding guarantees effective feature tapping and the robust structure guarantees the feature mining. Comparative studies have been conducted between the proposed approach with five other methods on three challenging datasets. The results demonstrate the proposed approach yields unanimous and significant improvements in mask-wise metrics, which evaluate object integrity and accuracy, as well as edge-wise metrics, which assess contour regularity and precision. The outperformance also indicates better multi-scale object adaptability of SCShaping. The obtain-and-play SCShaping commands a pleasurable implementation way to advance ideal <em>man</em>–<em>machine</em> collaboration.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":null,"pages":null},"PeriodicalIF":12.7,"publicationDate":"2024-05-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141243788","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Detection and attribution of cereal yield losses using Sentinel-2 and weather data: A case study in South Australia 利用 Sentinel-2 和气象数据检测谷物产量损失并确定损失原因:南澳大利亚案例研究
IF 12.7 1区 地球科学 Q1 Computer Science Pub Date : 2024-05-30 DOI: 10.1016/j.isprsjprs.2024.05.021
Keke Duan , Anton Vrieling , Michael Schlund , Uday Bhaskar Nidumolu , Christina Ratcliff , Simon Collings , Andrew Nelson

Weather extremes affect crop production. Remote sensing can help to detect crop damage and estimate lost yield due to weather extremes over large spatial extents. We propose a novel scalable method to predict in-season yield losses at the sub-field level and attribute these to weather extremes. To assess our method’s potential, we conducted a proof-of-concept case study on winter cereal paddocks in South Australia using data from 2017 to 2022. To detect crop growth anomalies throughout the growing season, we aligned a two-band Enhanced Vegetation Index (EVI2) time series from Sentinel-2 with thermal time. The deviation between the expected and observed EVI2 time series was defined as the Crop Damage Index (CDI). We assessed the performance of the CDI within specific phenological windows to predict yield loss. Finally, by comparing instances of substantial increase in CDI with different extreme weather indicators, we explored which (combinations of) extreme weather events were likely responsible for the experienced yield reduction. We found that the use of thermal time diminished the temporal deviation of EVI2 time series between years, resulting in the effective construction of typical stress-free crop growth curves. Thermal-time-based EVI2 time series resulted in better prediction of yield reduction than those based on calendar dates. Yield reduction could be predicted before grain-filling (approximately two months before harvest) with an R2 of 0.83 for wheat and 0.91 for barley. Finally, the combined analysis of CDI curves and extreme weather indices allowed for timely detection of weather-related causes of crop damage, which also captured the spatial variations of crop damage attribution at sub-field level. The proposed framework provides a basis for early warning of crop damage and attributing the damage to weather extremes in near real-time, which should help to adopt appropriate crop protection strategies.

极端天气会影响作物产量。遥感技术可帮助检测作物受损情况,并估算大范围极端天气造成的产量损失。我们提出了一种可扩展的新方法,用于预测亚田块层面的当季产量损失,并将其归因于极端天气。为了评估我们的方法的潜力,我们利用 2017 年至 2022 年的数据对南澳大利亚的冬季谷物围场进行了概念验证案例研究。为了检测整个生长季节的作物生长异常,我们将来自哨兵-2 的双波段增强植被指数 (EVI2) 时间序列与热时间进行了比对。预期和观测到的 EVI2 时间序列之间的偏差被定义为作物损害指数(CDI)。我们评估了 CDI 在特定物候窗口内预测产量损失的性能。最后,通过比较 CDI 大幅上升与不同极端天气指标的关系,我们探讨了哪些(组合)极端天气事件可能是造成减产的原因。我们发现,热时间的使用减小了 EVI2 时间序列在不同年份之间的时间偏差,从而有效地构建了典型的无压力作物生长曲线。与基于日历日期的时间序列相比,基于热时间的 EVI2 时间序列能更好地预测减产。小麦和大麦的 R2 分别为 0.83 和 0.91。最后,CDI 曲线和极端天气指数的综合分析有助于及时发现与天气相关的作物损害原因,同时还能捕捉子田块层面作物损害归因的空间变化。所提出的框架为作物损害的早期预警以及近实时地将损害归因于极端天气提供了基础,这将有助于采取适当的作物保护策略。
{"title":"Detection and attribution of cereal yield losses using Sentinel-2 and weather data: A case study in South Australia","authors":"Keke Duan ,&nbsp;Anton Vrieling ,&nbsp;Michael Schlund ,&nbsp;Uday Bhaskar Nidumolu ,&nbsp;Christina Ratcliff ,&nbsp;Simon Collings ,&nbsp;Andrew Nelson","doi":"10.1016/j.isprsjprs.2024.05.021","DOIUrl":"https://doi.org/10.1016/j.isprsjprs.2024.05.021","url":null,"abstract":"<div><p>Weather extremes affect crop production. Remote sensing can help to detect crop damage and estimate lost yield due to weather extremes over large spatial extents. We propose a novel scalable method to predict in-season yield losses at the sub-field level and attribute these to weather extremes. To assess our method’s potential, we conducted a proof-of-concept case study on winter cereal paddocks in South Australia using data from 2017 to 2022. To detect crop growth anomalies throughout the growing season, we aligned a two-band Enhanced Vegetation Index (EVI2) time series from Sentinel-2 with thermal time. The deviation between the expected and observed EVI2 time series was defined as the Crop Damage Index (CDI). We assessed the performance of the CDI within specific phenological windows to predict yield loss. Finally, by comparing instances of substantial increase in CDI with different extreme weather indicators, we explored which (combinations of) extreme weather events were likely responsible for the experienced yield reduction. We found that the use of thermal time diminished the temporal deviation of EVI2 time series between years, resulting in the effective construction of typical stress-free crop growth curves. Thermal-time-based EVI2 time series resulted in better prediction of yield reduction than those based on calendar dates. Yield reduction could be predicted before grain-filling (approximately two months before harvest) with an R<sup>2</sup> of 0.83 for wheat and 0.91 for barley. Finally, the combined analysis of CDI curves and extreme weather indices allowed for timely detection of weather-related causes of crop damage, which also captured the spatial variations of crop damage attribution at sub-field level. The proposed framework provides a basis for early warning of crop damage and attributing the damage to weather extremes in near real-time, which should help to adopt appropriate crop protection strategies.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":null,"pages":null},"PeriodicalIF":12.7,"publicationDate":"2024-05-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0924271624002193/pdfft?md5=a7c5d07fec49f19fc05bd71d62c4e830&pid=1-s2.0-S0924271624002193-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141243765","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
ISPRS Journal of Photogrammetry and Remote Sensing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1