首页 > 最新文献

The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences最新文献

英文 中文
A Robust Registration Method for Multi-view SAR Images based on Best Buddy Similarity 基于最佳好友相似性的多视角合成孔径雷达图像稳健配准方法
Pub Date : 2024-05-11 DOI: 10.5194/isprs-archives-xlviii-1-2024-881-2024
Yifan Zhang, Zhiwei Li, Wen Wang, Minzheng Mu, Bangwei Zuo
Abstract. Due to the influence of imaging angle and terrain undulation, multi-view synthetic aperture radar (SAR) images are difficult to be directly registered by traditional methods. Although feature matching solves the issue of image rotation and maintains scale invariance, these methods often lead to non-uniformity of interest points and may not achieve subpixel accuracy. The traditional template matching method makes it difficult to generate correct matches for multi-view SAR oblique images. In this paper, a multi-view SAR image template matching method based on Best Buddy Similarity (BBS) is proposed to solve the traditional methods' problem. Firstly, the initial correspondences between images are established according to the Range-Doppler model of SAR images. Secondly, a sliding window search is performed on the established correspondence, the BBS is calculated, and the subpixel locations of the peaks on the similarity map are estimated to achieve a fine match. In the calculation process of BBS, the SAR-ROEWA operator is used to suppress the speckle noise of SAR images. The experiment demonstrated that SAR-BBS can accurately match SAR images with large rotation angle. The peak value on the search window is significant. The registration accuracy of SAR-BBS outperforms the other state-of-the-art methods.
摘要由于成像角度和地形起伏的影响,多视角合成孔径雷达(SAR)图像难以用传统方法直接注册。虽然特征匹配可以解决图像旋转问题并保持比例不变性,但这些方法往往会导致兴趣点的不均匀性,而且可能无法达到亚像素精度。传统的模板匹配方法很难为多视角 SAR 倾斜图像生成正确的匹配结果。本文提出了一种基于最佳好友相似度(BBS)的多视角 SAR 图像模板匹配方法来解决传统方法的问题。首先,根据 SAR 图像的测距-多普勒模型建立图像之间的初始对应关系。其次,在建立的对应关系上执行滑动窗口搜索,计算 BBS,并估计相似性图上峰值的子像素位置,以实现精细匹配。在计算 BBS 的过程中,使用了 SAR-ROEWA 算子来抑制 SAR 图像的斑点噪声。实验证明,SAR-BBS 可以精确匹配大旋转角度的 SAR 图像。搜索窗口上的峰值非常显著。SAR-BBS 的配准精度优于其他先进方法。
{"title":"A Robust Registration Method for Multi-view SAR Images based on Best Buddy Similarity","authors":"Yifan Zhang, Zhiwei Li, Wen Wang, Minzheng Mu, Bangwei Zuo","doi":"10.5194/isprs-archives-xlviii-1-2024-881-2024","DOIUrl":"https://doi.org/10.5194/isprs-archives-xlviii-1-2024-881-2024","url":null,"abstract":"Abstract. Due to the influence of imaging angle and terrain undulation, multi-view synthetic aperture radar (SAR) images are difficult to be directly registered by traditional methods. Although feature matching solves the issue of image rotation and maintains scale invariance, these methods often lead to non-uniformity of interest points and may not achieve subpixel accuracy. The traditional template matching method makes it difficult to generate correct matches for multi-view SAR oblique images. In this paper, a multi-view SAR image template matching method based on Best Buddy Similarity (BBS) is proposed to solve the traditional methods' problem. Firstly, the initial correspondences between images are established according to the Range-Doppler model of SAR images. Secondly, a sliding window search is performed on the established correspondence, the BBS is calculated, and the subpixel locations of the peaks on the similarity map are estimated to achieve a fine match. In the calculation process of BBS, the SAR-ROEWA operator is used to suppress the speckle noise of SAR images. The experiment demonstrated that SAR-BBS can accurately match SAR images with large rotation angle. The peak value on the search window is significant. The registration accuracy of SAR-BBS outperforms the other state-of-the-art methods.\u0000","PeriodicalId":505918,"journal":{"name":"The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences","volume":" 4","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140988885","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Transformer and Visual Foundation Model-Based Method for Cross-View Remote Sensing Image Retrieval 基于变换器和视觉基础模型的跨视角遥感图像检索方法
Pub Date : 2024-05-11 DOI: 10.5194/isprs-archives-xlviii-1-2024-821-2024
Changjiang Yin, Qin Ye, Junqi Luo
Abstract. Retrieving UAV images that lack POS information with georeferenced satellite orthoimagery is challenging due to the differences in angles of views. Most existing methods rely on deep neural networks with a large number of parameters, leading to substantial time and financial investments in network training. Consequently, these methods may not be well-suited for downstream tasks that have high timeliness requirements. In this work, we propose a cross-view remote sensing image retrieval method based on transformer and visual foundation model. We investigated the potential of visual foundation model for extracting common features from cross-view images. Training is only conducted on a small, self-designed retrieval head, alleviating the burden of network training. Specifically, we designed a CVV module to optimize the features extracted from the visual foundation model, making these features more adept for cross-view image retrieval tasks. And we designed an MLP head to achieve similarity discrimination. The method is verified on a publicly available dataset containing multiple scenes. Our method shows excellent results in terms of both efficiency and accuracy on 15 sub-datasets (10 or 50 scene categories) derived from the public dataset, which holds practical value in engineering applications with streamlined scene categories and constrained computational resources. Furthermore, we initiated a comprehensive discussion and conducted ablation experiments on the network design to validate its efficacy. Additionally, we analyzed the presence of overfitting within the network and deliberated on the limitations of our study, proposing potential avenues for future enhancements.
摘要由于视角的不同,利用地理参照卫星正射影像检索缺乏 POS 信息的无人机图像具有挑战性。现有的大多数方法都依赖于具有大量参数的深度神经网络,从而导致在网络训练方面投入大量的时间和资金。因此,这些方法可能不太适合对时效性要求较高的下游任务。在这项工作中,我们提出了一种基于变换器和视觉基础模型的跨视角遥感图像检索方法。我们研究了视觉基础模型从跨视角图像中提取共同特征的潜力。训练只在一个自行设计的小型检索头上进行,减轻了网络训练的负担。具体来说,我们设计了一个 CVV 模块来优化从视觉基础模型中提取的特征,使这些特征更适合跨视图图像检索任务。我们还设计了一个 MLP 头来实现相似性判别。我们在一个包含多个场景的公开数据集上对该方法进行了验证。我们的方法在公共数据集衍生的 15 个子数据集(10 或 50 个场景类别)上显示出了卓越的效率和准确性,这在场景类别精简、计算资源有限的工程应用中具有实用价值。此外,我们还对网络设计进行了全面讨论和消融实验,以验证其有效性。此外,我们还分析了网络中是否存在过拟合现象,并讨论了我们研究的局限性,提出了未来改进的潜在途径。
{"title":"A Transformer and Visual Foundation Model-Based Method for Cross-View Remote Sensing Image Retrieval","authors":"Changjiang Yin, Qin Ye, Junqi Luo","doi":"10.5194/isprs-archives-xlviii-1-2024-821-2024","DOIUrl":"https://doi.org/10.5194/isprs-archives-xlviii-1-2024-821-2024","url":null,"abstract":"Abstract. Retrieving UAV images that lack POS information with georeferenced satellite orthoimagery is challenging due to the differences in angles of views. Most existing methods rely on deep neural networks with a large number of parameters, leading to substantial time and financial investments in network training. Consequently, these methods may not be well-suited for downstream tasks that have high timeliness requirements. In this work, we propose a cross-view remote sensing image retrieval method based on transformer and visual foundation model. We investigated the potential of visual foundation model for extracting common features from cross-view images. Training is only conducted on a small, self-designed retrieval head, alleviating the burden of network training. Specifically, we designed a CVV module to optimize the features extracted from the visual foundation model, making these features more adept for cross-view image retrieval tasks. And we designed an MLP head to achieve similarity discrimination. The method is verified on a publicly available dataset containing multiple scenes. Our method shows excellent results in terms of both efficiency and accuracy on 15 sub-datasets (10 or 50 scene categories) derived from the public dataset, which holds practical value in engineering applications with streamlined scene categories and constrained computational resources. Furthermore, we initiated a comprehensive discussion and conducted ablation experiments on the network design to validate its efficacy. Additionally, we analyzed the presence of overfitting within the network and deliberated on the limitations of our study, proposing potential avenues for future enhancements.\u0000","PeriodicalId":505918,"journal":{"name":"The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences","volume":" 19","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140988542","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Polar-vision1: A Novel Collinearity Equation of Perspective Projection in Polar Coordinate System Polar-vision1: 极坐标系中透视投影的新型共线性方程
Pub Date : 2024-05-11 DOI: 10.5194/isprs-archives-xlviii-1-2024-917-2024
Zhengkang Zuo, Bin Zhang
Abstract. Progress has been made in the community of photogrammetry and 3d computer vision in addressing the mathematical challenge posed by the collinearity equation. We introduce a new method for establishing the coordinate reference for 2d pixels and 3d landmarks using 'angular coordinates'. The mathematical relationships required for converting 3d landmarks, expressed in angular coordinates, to the camera framework are presented. The landmarks are then projected using perspective projection to obtain 2d pixels represented in angular coordinates. This framework is formally nominated as the 'Polar-vision1', which has been developed and integrated into the commercial software G3D-Cluster. Its application to pinhole camera image processing has demonstrated superior efficiency and admission rates of tie points, as well as reconstruction detail capabilities, compared to OpenMVG, achieving approximately a 1.4× improvement. The project 'Key Technologies and Tool System for Realistic 3D Modeling through Integration of Multi-Source Information in the Space-Air-Ground Domain' was awarded First Prize at the 2023 Surveying Science and Technology Awards, with Polar-vision1 as one of the innovative points.
摘要摄影测量学和三维计算机视觉界在解决共线性方程带来的数学挑战方面取得了进展。我们介绍了一种使用 "角度坐标 "建立二维像素和三维地标的坐标参考的新方法。我们介绍了将以角度坐标表示的 3D 地标转换为摄像机框架所需的数学关系。然后使用透视投影法对地标进行投影,从而获得以角度坐标表示的 2d 像素。这一框架被正式命名为 "Polar-vision1",已被开发并集成到商业软件 G3D-Cluster 中。与 OpenMVG 相比,它在针孔摄像机图像处理中的应用显示出了更高的效率和领带点的接纳率,以及重建细节的能力,大约提高了 1.4 倍。天-空-地领域多源信息集成真实三维建模关键技术与工具系统 "项目荣获 2023 年度测绘科学技术奖一等奖,Polar-vision1 是创新点之一。
{"title":"Polar-vision1: A Novel Collinearity Equation of Perspective Projection in Polar Coordinate System","authors":"Zhengkang Zuo, Bin Zhang","doi":"10.5194/isprs-archives-xlviii-1-2024-917-2024","DOIUrl":"https://doi.org/10.5194/isprs-archives-xlviii-1-2024-917-2024","url":null,"abstract":"Abstract. Progress has been made in the community of photogrammetry and 3d computer vision in addressing the mathematical challenge posed by the collinearity equation. We introduce a new method for establishing the coordinate reference for 2d pixels and 3d landmarks using 'angular coordinates'. The mathematical relationships required for converting 3d landmarks, expressed in angular coordinates, to the camera framework are presented. The landmarks are then projected using perspective projection to obtain 2d pixels represented in angular coordinates. This framework is formally nominated as the 'Polar-vision1', which has been developed and integrated into the commercial software G3D-Cluster. Its application to pinhole camera image processing has demonstrated superior efficiency and admission rates of tie points, as well as reconstruction detail capabilities, compared to OpenMVG, achieving approximately a 1.4× improvement. The project 'Key Technologies and Tool System for Realistic 3D Modeling through Integration of Multi-Source Information in the Space-Air-Ground Domain' was awarded First Prize at the 2023 Surveying Science and Technology Awards, with Polar-vision1 as one of the innovative points.\u0000","PeriodicalId":505918,"journal":{"name":"The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences","volume":"2 20","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140988181","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Technical Framework and Preliminary Practices of Global Geographic Information Resource Construction 全球地理信息资源建设的技术框架和初步实践
Pub Date : 2024-05-11 DOI: 10.5194/isprs-archives-xlviii-1-2024-837-2024
Hongwei Zhang, Jiage Chen, Chenchen Wu, Lijun Chen
Abstract. High-precision and real-time global geographic information data are fundamental and strategic resources in various fields such as safeguarding global strategic interests, studying global environmental changes, and planning for sustainable development. However, due to challenges related to ground control and obtaining reference information, the development of global geographic information resources faces significant hurdles in terms of geometric positioning, information extraction, and data mining. This paper starts with the characteristics of domestically produced remote sensing images and proposes a comprehensive technical framework centered around "uncontrolled geometric positioning, intelligent interpretation of typical elements, mining of multi-source data from abroad, and intelligent hybrid collection and compilation of Digital Elevation Models (DEMs)." The paper elaborates on the key technical challenges that need to be overcome and their corresponding solutions. It also outlines the development of relevant data products and production technical specifications. Multiple production-oriented software tools were developed, leading to the creation of a variety of data products in multiple types and scales, including global 30-meter land cover data, DEM data, core vector data, and more.
摘要高精度、实时性的全球地理信息数据是维护全球战略利益、研究全球环境变化、规划可持续发展等各个领域的基础性、战略性资源。然而,由于地面控制和参考信息获取等方面的挑战,全球地理信息资源的开发在几何定位、信息提取和数据挖掘等方面面临着巨大障碍。本文从国产遥感影像的特点入手,提出了以 "几何定位失控、典型要素智能解译、国外多源数据挖掘、数字高程模型(DEM)智能混合采集与编制 "为核心的综合技术框架。论文详细阐述了需要克服的关键技术挑战及其相应的解决方案。文件还概述了相关数据产品和生产技术规范的开发情况。开发了多种面向生产的软件工具,从而创建了多种类型和尺度的数据产品,包括全球 30 米土地覆被数据、DEM 数据、核心矢量数据等。
{"title":"Technical Framework and Preliminary Practices of Global Geographic Information Resource Construction","authors":"Hongwei Zhang, Jiage Chen, Chenchen Wu, Lijun Chen","doi":"10.5194/isprs-archives-xlviii-1-2024-837-2024","DOIUrl":"https://doi.org/10.5194/isprs-archives-xlviii-1-2024-837-2024","url":null,"abstract":"Abstract. High-precision and real-time global geographic information data are fundamental and strategic resources in various fields such as safeguarding global strategic interests, studying global environmental changes, and planning for sustainable development. However, due to challenges related to ground control and obtaining reference information, the development of global geographic information resources faces significant hurdles in terms of geometric positioning, information extraction, and data mining. This paper starts with the characteristics of domestically produced remote sensing images and proposes a comprehensive technical framework centered around \"uncontrolled geometric positioning, intelligent interpretation of typical elements, mining of multi-source data from abroad, and intelligent hybrid collection and compilation of Digital Elevation Models (DEMs).\" The paper elaborates on the key technical challenges that need to be overcome and their corresponding solutions. It also outlines the development of relevant data products and production technical specifications. Multiple production-oriented software tools were developed, leading to the creation of a variety of data products in multiple types and scales, including global 30-meter land cover data, DEM data, core vector data, and more.\u0000","PeriodicalId":505918,"journal":{"name":"The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences","volume":"3 18","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140988350","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards Integration of IndoorGML and GDF for Robot Navigation in Warehouses 将 IndoorGML 和 GDF 整合用于仓库机器人导航
Pub Date : 2024-05-11 DOI: 10.5194/isprs-archives-xlviii-1-2024-737-2024
Ziwei Xiang, Kunlin Yu, Zhiyong Wang
Abstract. With the development of the navigation technology, the outdoor navigation has made great progress, whereas the indoor navigation has some areas which is underdeveloped, insufficient to meet the rapidly increasing demands of people as well as the robotics. Even though, the advance in indoor navigation technology still has really brought a wide range of applications and a broad market, for instance, the flourishing intelligent warehouse system utilizes multi-robot operation which have the certain requirement for an accurate indoor navigation system. As for the indoor navigation, the OGC standard IndoorGML has been released and undergoing revision constantly. While the document really provides more advantageous support for the applications of Indoor Location-Based Services (LBS), in some aspects, especially the door-to-door navigation and the warehouse environment, it is not sufficiently adaptable, with still some room for improvement. IndoorGML is powerful for the common indoor scenarios like malls and offices, while as for carefully-arranged warehouse environment and other large-scale operation scenarios with multi-robots that is more similar to an ordered system, it is obviously insufficient. In this paper, we discuss about the potential to combination of IndoorGML and ITS standard ISO 20524 (GDF5.1), and extend the OGC standard indoorGML. We analyze the definition as well as function of related concepts, making some comparisons between these two standards. We conclude that these two standards are well-matched with vital potential to merge and unify the indoor and outdoor systems for spatial information.
摘要随着导航技术的发展,室外导航取得了长足的进步,而室内导航在某些方面还不够发达,无法满足人们以及机器人快速增长的需求。尽管如此,室内导航技术的进步确实带来了广泛的应用和广阔的市场,例如,方兴未艾的智能仓储系统利用多机器人操作,对精确的室内导航系统有一定的要求。在室内导航方面,OGC 标准 IndoorGML 已经发布,并在不断修订中。虽然该文件确实为室内位置服务(LBS)的应用提供了较为有利的支持,但在某些方面,尤其是门到门导航和仓库环境方面,其适应性还不够强,仍有一定的改进空间。IndoorGML 对于商场、办公室等常见的室内场景来说功能强大,但对于布局严谨的仓库环境和其他更类似于有序系统的多机器人大规模作业场景来说,它显然还不够。本文探讨了 IndoorGML 与 ITS 标准 ISO 20524(GDF5.1)结合的可能性,并对 OGC 标准 indoorGML 进行了扩展。我们分析了相关概念的定义和功能,并对这两个标准进行了比较。我们的结论是,这两个标准非常匹配,具有合并和统一室内外空间信息系统的巨大潜力。
{"title":"Towards Integration of IndoorGML and GDF for Robot Navigation in Warehouses","authors":"Ziwei Xiang, Kunlin Yu, Zhiyong Wang","doi":"10.5194/isprs-archives-xlviii-1-2024-737-2024","DOIUrl":"https://doi.org/10.5194/isprs-archives-xlviii-1-2024-737-2024","url":null,"abstract":"Abstract. With the development of the navigation technology, the outdoor navigation has made great progress, whereas the indoor navigation has some areas which is underdeveloped, insufficient to meet the rapidly increasing demands of people as well as the robotics. Even though, the advance in indoor navigation technology still has really brought a wide range of applications and a broad market, for instance, the flourishing intelligent warehouse system utilizes multi-robot operation which have the certain requirement for an accurate indoor navigation system. As for the indoor navigation, the OGC standard IndoorGML has been released and undergoing revision constantly. While the document really provides more advantageous support for the applications of Indoor Location-Based Services (LBS), in some aspects, especially the door-to-door navigation and the warehouse environment, it is not sufficiently adaptable, with still some room for improvement. IndoorGML is powerful for the common indoor scenarios like malls and offices, while as for carefully-arranged warehouse environment and other large-scale operation scenarios with multi-robots that is more similar to an ordered system, it is obviously insufficient. In this paper, we discuss about the potential to combination of IndoorGML and ITS standard ISO 20524 (GDF5.1), and extend the OGC standard indoorGML. We analyze the definition as well as function of related concepts, making some comparisons between these two standards. We conclude that these two standards are well-matched with vital potential to merge and unify the indoor and outdoor systems for spatial information.\u0000","PeriodicalId":505918,"journal":{"name":"The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences","volume":" 671","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140989563","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Soybean seedling detection and counting from UAV images based on an improved YOLOv8 Network 基于改进的 YOLOv8 网络从无人机图像中检测和计算大豆幼苗
Pub Date : 2024-05-11 DOI: 10.5194/isprs-archives-xlviii-1-2024-727-2024
Haotian Wu, Junhua Kang, Heli Li
Abstract. The utilization of unmanned aerial vehicle (UAV) for soybean seedling detection is an effective way to estimate soybean yield, which plays a crucial role in agricultural planning and decision-making. However, the soybean seedlings objects in the UAV image are small, in clusters, and occluded each other, which makes it very challenging to achieve accurate object detection and counting. To address these issues, we optimize the YOLOv8 model and propose a GAS-YOLOv8 network, aiming to enhance the detection accuracy for the task of soybean seedling detection based on UAV images. Firstly, a global attention mechanism (GAM) is incorporated into the neck module of YOLOv8, which reallocates weights and prioritizes global information to more effectively extract soybean seedling features. Secondly, the CIOU loss function is replaced with the SIOU loss, which includes an angle loss term to guide the regression of bounding boxes. Experimental results show that, on the soybean seedling dataset, the proposed GAS-YOLOv8 model achieves a 1.3% improvement in mAP@0.5 and a 6% enhancement in detection performance in dense seedling areas, when compared to the baseline model YOLOv8s.When compared to other object detection models (YOLOv5, Faster R-CNN, etc.), the GAS-YOLOv8 model similarly achieved the best detection performance. These results demonstrate the effectiveness of the GAS-YOLOv8 in detecting dense soybean seedlings, providing more accurate theoretical support for subsequent yield estimation.
摘要利用无人飞行器(UAV)进行大豆幼苗检测是估算大豆产量的一种有效方法,在农业规划和决策中发挥着至关重要的作用。然而,无人机图像中的大豆幼苗物体较小、成团且相互遮挡,这给实现精确的物体检测和计数带来了很大挑战。针对这些问题,我们对 YOLOv8 模型进行了优化,并提出了 GAS-YOLOv8 网络,旨在提高基于无人机图像的大豆幼苗检测任务的检测精度。首先,在 YOLOv8 的颈部模块中加入全局关注机制(GAM),重新分配权重并优先考虑全局信息,从而更有效地提取大豆幼苗特征。其次,将 CIOU 损失函数替换为 SIOU 损失,后者包含一个角度损失项,用于指导边界框的回归。实验结果表明,在大豆秧苗数据集上,与基线模型 YOLOv8s 相比,所提出的 GAS-YOLOv8 模型在 mAP@0.5 上提高了 1.3%,在秧苗密集区域的检测性能上提高了 6%。与其他物体检测模型(YOLOv5、Faster R-CNN 等)相比,GAS-YOLOv8 模型同样取得了最佳检测性能。这些结果证明了 GAS-YOLOv8 在检测密集大豆幼苗方面的有效性,为后续的产量估算提供了更准确的理论支持。
{"title":"Soybean seedling detection and counting from UAV images based on an improved YOLOv8 Network","authors":"Haotian Wu, Junhua Kang, Heli Li","doi":"10.5194/isprs-archives-xlviii-1-2024-727-2024","DOIUrl":"https://doi.org/10.5194/isprs-archives-xlviii-1-2024-727-2024","url":null,"abstract":"Abstract. The utilization of unmanned aerial vehicle (UAV) for soybean seedling detection is an effective way to estimate soybean yield, which plays a crucial role in agricultural planning and decision-making. However, the soybean seedlings objects in the UAV image are small, in clusters, and occluded each other, which makes it very challenging to achieve accurate object detection and counting. To address these issues, we optimize the YOLOv8 model and propose a GAS-YOLOv8 network, aiming to enhance the detection accuracy for the task of soybean seedling detection based on UAV images. Firstly, a global attention mechanism (GAM) is incorporated into the neck module of YOLOv8, which reallocates weights and prioritizes global information to more effectively extract soybean seedling features. Secondly, the CIOU loss function is replaced with the SIOU loss, which includes an angle loss term to guide the regression of bounding boxes. Experimental results show that, on the soybean seedling dataset, the proposed GAS-YOLOv8 model achieves a 1.3% improvement in mAP@0.5 and a 6% enhancement in detection performance in dense seedling areas, when compared to the baseline model YOLOv8s.When compared to other object detection models (YOLOv5, Faster R-CNN, etc.), the GAS-YOLOv8 model similarly achieved the best detection performance. These results demonstrate the effectiveness of the GAS-YOLOv8 in detecting dense soybean seedlings, providing more accurate theoretical support for subsequent yield estimation.\u0000","PeriodicalId":505918,"journal":{"name":"The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences","volume":" 1133","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140989097","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-temporal Monitoring for Road Slope Collapse by Means of LUTAN-1 SAR Data and High Resolution Optical Data 利用 LUTAN-1 SAR 数据和高分辨率光学数据对路基边坡坍塌进行多时监测
Pub Date : 2024-05-11 DOI: 10.5194/isprs-archives-xlviii-1-2024-867-2024
Xiang Zhang, Xinming Tang, Tao Li, Xiaoqing Zhou, Haifeng Hu, Xuefei Zhang, Jing Lu
Abstract. Collapse is one of the most destructive natural disaster, being sudden, frequent, and highly concealed, causing large-scale damage. On August 10, 2023, the slope of 108 national highway in Weinan, Shaanxi Province collapsed. The lower edge of the collapse slope body is the Luohe river, and the collapse body rushes into the river to form a barrier lake. Remote sensing technique can provide multiple dimensional information for disaster emergency and management. Lutan-1 SAR satellites are the first group L-band SAR constellation for multiple applications in China. Owing to the precise orbit control ability and high revisit characteristics for Lutan-1 SAR satellites, surface deformation monitoring with centimeter even millimeter accuracy may be achieved. Based on the multi-temporal pre-disaster and post-disaster Lutan-1 SAR data and high resolution optical data, the collapse information including the pre-disaster and post-disaster were extracted and analysed. From July 11 to 27, 2023, the pre-collapse deformation was obtained with the maximum value of 6 cm, and obvious deformation occurred before the collapse. Lutan-1 monitored results pre-collapse can provide certain information for disaster early identification. From July 27 to August 24, 2023, due to the serious incoherence caused by large deformation and ground changes, effective deformation information cannot be obtained based on the InSAR technique. In addition, the collapse information was clearly extracted by the high resolution optical data acquired pre-collapse and post collapse. After the collapse, significant deformation was extracted from August 24 to September 21 with the maximum value of 6 cm, indicating that obvious deformation still occurred over the collapse area. Through the analysis for the series results obtained by SAR and optical data, it is favourable for disaster emergency and management.
摘要崩塌是最具破坏性的自然灾害之一,具有突发性、频发性、隐蔽性强等特点,会造成大规模的破坏。2023年8月10日,陕西省渭南市108国道边坡发生崩塌。塌方坡体下缘为洛河,塌方体冲入洛河形成堰塞湖。遥感技术可以为灾害应急和管理提供多维信息。芦滩一号合成孔径雷达卫星是我国第一组多用途 L 波段合成孔径雷达星座。由于 "路坦一号 "合成孔径雷达卫星具有精确的轨道控制能力和高重访特性,可实现厘米级甚至毫米级精度的地表形变监测。基于多时相的灾前、灾后 "路坦一号 "合成孔径雷达数据和高分辨率光学数据,提取并分析了包括灾前、灾后在内的塌陷信息。从 2023 年 7 月 11 日到 27 日,获得了坍塌前的变形,最大值为 6 厘米,坍塌前发生了明显的变形。路坦一号 "坍塌前的监测结果可为灾害早期识别提供一定的信息。2023年7月27日至8月24日,由于大变形和地面变化造成的严重不一致性,基于InSAR技术无法获得有效的变形信息。此外,通过坍塌前和坍塌后获取的高分辨率光学数据,可以清晰地提取坍塌信息。塌陷后,从 8 月 24 日至 9 月 21 日提取到了明显的形变,最大值为 6 厘米,表明塌陷区仍发生了明显的形变。通过对合成孔径雷达和光学数据得到的序列结果进行分析,有利于灾害应急和管理。
{"title":"Multi-temporal Monitoring for Road Slope Collapse by Means of LUTAN-1 SAR Data and High Resolution Optical Data","authors":"Xiang Zhang, Xinming Tang, Tao Li, Xiaoqing Zhou, Haifeng Hu, Xuefei Zhang, Jing Lu","doi":"10.5194/isprs-archives-xlviii-1-2024-867-2024","DOIUrl":"https://doi.org/10.5194/isprs-archives-xlviii-1-2024-867-2024","url":null,"abstract":"Abstract. Collapse is one of the most destructive natural disaster, being sudden, frequent, and highly concealed, causing large-scale damage. On August 10, 2023, the slope of 108 national highway in Weinan, Shaanxi Province collapsed. The lower edge of the collapse slope body is the Luohe river, and the collapse body rushes into the river to form a barrier lake. Remote sensing technique can provide multiple dimensional information for disaster emergency and management. Lutan-1 SAR satellites are the first group L-band SAR constellation for multiple applications in China. Owing to the precise orbit control ability and high revisit characteristics for Lutan-1 SAR satellites, surface deformation monitoring with centimeter even millimeter accuracy may be achieved. Based on the multi-temporal pre-disaster and post-disaster Lutan-1 SAR data and high resolution optical data, the collapse information including the pre-disaster and post-disaster were extracted and analysed. From July 11 to 27, 2023, the pre-collapse deformation was obtained with the maximum value of 6 cm, and obvious deformation occurred before the collapse. Lutan-1 monitored results pre-collapse can provide certain information for disaster early identification. From July 27 to August 24, 2023, due to the serious incoherence caused by large deformation and ground changes, effective deformation information cannot be obtained based on the InSAR technique. In addition, the collapse information was clearly extracted by the high resolution optical data acquired pre-collapse and post collapse. After the collapse, significant deformation was extracted from August 24 to September 21 with the maximum value of 6 cm, indicating that obvious deformation still occurred over the collapse area. Through the analysis for the series results obtained by SAR and optical data, it is favourable for disaster emergency and management.\u0000","PeriodicalId":505918,"journal":{"name":"The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences","volume":" 387","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140990001","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Analysis of Multiple Scattering Characteristics of Cable-Stayed Bridges with Multi-band SAR 利用多波段合成孔径雷达分析斜拉桥的多重散射特性
Pub Date : 2024-05-11 DOI: 10.5194/isprs-archives-xlviii-1-2024-761-2024
Yanhao Xu, Yangmao Wen, Tao Li, Sijie Ma, Jie Liu
Abstract. Accurate localization of multi-scattering features of cable-stayed bridges in multi-band Synthetic Aperture Radar (SAR) imagery is crucial for intelligent recognition of bridge targets within images, as well as for precise water level extraction. This study focuses on the Badong Yangtze River Bridge, utilizing Unmanned Aerial Vehicle (UAV) LiDAR data of the bridge, and analyzes the multi-scattering characteristics of different bridge structural targets based on Geometric Optics (GO) methods and the Range-Doppler principle. Furthermore, the study integrates LiDAR data of the bridge's cable-stays to examine their multi-scattering phenomena, finding that the undulations of the Yangtze River's surface waves significantly contribute to the pronounced double scattering features of the bridge's cable-stays. Additionally, statistical analysis of multi-source SAR data indicates that this phenomenon is not directly correlated with radar wavelength, implying no direct connection to surface roughness. Utilizing LiDAR point cloud data from the bridge's street lamps, this paper proposes a novel method for estimating water level elevation by identifying the center position of spots formed by double scattering from lamp posts. The results show that using TerraSAR ascending and descending orbit images, this method achieves a water level elevation accuracy of approximately 0.2 meters.
摘要精确定位多波段合成孔径雷达(SAR)图像中斜拉桥的多散射特征对于智能识别图像中的桥梁目标以及精确提取水位至关重要。本研究以巴东长江大桥为研究对象,利用无人机(UAV)获取的大桥激光雷达数据,基于几何光学(GO)方法和测距-多普勒原理,分析了不同桥梁结构目标的多重散射特征。此外,研究还整合了大桥斜拉索的激光雷达数据,对其多重散射现象进行了研究,发现长江表面波的起伏极大地促进了大桥斜拉索明显的双重散射特征。此外,对多源合成孔径雷达数据的统计分析表明,这种现象与雷达波长没有直接关系,这意味着与表面粗糙度没有直接联系。本文利用大桥路灯的激光雷达点云数据,提出了一种新方法,通过识别灯柱双重散射形成的光斑中心位置来估算水位高程。结果表明,利用 TerraSAR 上升和下降轨道图像,该方法可实现约 0.2 米的水位标高精度。
{"title":"Analysis of Multiple Scattering Characteristics of Cable-Stayed Bridges with Multi-band SAR","authors":"Yanhao Xu, Yangmao Wen, Tao Li, Sijie Ma, Jie Liu","doi":"10.5194/isprs-archives-xlviii-1-2024-761-2024","DOIUrl":"https://doi.org/10.5194/isprs-archives-xlviii-1-2024-761-2024","url":null,"abstract":"Abstract. Accurate localization of multi-scattering features of cable-stayed bridges in multi-band Synthetic Aperture Radar (SAR) imagery is crucial for intelligent recognition of bridge targets within images, as well as for precise water level extraction. This study focuses on the Badong Yangtze River Bridge, utilizing Unmanned Aerial Vehicle (UAV) LiDAR data of the bridge, and analyzes the multi-scattering characteristics of different bridge structural targets based on Geometric Optics (GO) methods and the Range-Doppler principle. Furthermore, the study integrates LiDAR data of the bridge's cable-stays to examine their multi-scattering phenomena, finding that the undulations of the Yangtze River's surface waves significantly contribute to the pronounced double scattering features of the bridge's cable-stays. Additionally, statistical analysis of multi-source SAR data indicates that this phenomenon is not directly correlated with radar wavelength, implying no direct connection to surface roughness. Utilizing LiDAR point cloud data from the bridge's street lamps, this paper proposes a novel method for estimating water level elevation by identifying the center position of spots formed by double scattering from lamp posts. The results show that using TerraSAR ascending and descending orbit images, this method achieves a water level elevation accuracy of approximately 0.2 meters.\u0000","PeriodicalId":505918,"journal":{"name":"The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences","volume":" 670","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140989564","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Comparison of Sentinel-1 Biased and Unbiased Coherence for Crop Monitoring and Classification 用于作物监测和分类的哨兵-1 有偏相干性和无偏相干性比较
Pub Date : 2024-05-11 DOI: 10.5194/isprs-archives-xlviii-1-2024-903-2024
Qinxin Zhao, Qinghua Xie, Xing Peng, Yusong Bao, Tonglu Jia, Linwei Yue, Haiqiang Fu, Jianjun Zhu
Abstract. Synthetic Aperture Radar (SAR) holds significant potential for applications in crop monitoring and classification. Interferometric SAR (InSAR) coherence proves effective in monitoring crop growth. Currently, the coherence based on the maximum likelihood estimator is biased towards low coherence values. Therefore, the main aim of this work is to access the performance of Sentinel-1 time-series biased coherence and unbiased coherence in crop monitoring and classification. This study was conducted during the 2018 growing season (April-October) in Komoka, an agricultural region in southwestern Ontario, Canada, primarily cultivating three crops: soybean, corn, and winter wheat. To verify the ability of coherence to monitor crops, a linear correlation coefficient between temporal coherence and dual polarimetric radar vegetation index (DpRVI) was fitted. The results revealed a stable correlation between temporal coherence and DpRVI time-series, with the highest correlation observed for soybean (0.7 < R < 0.8), followed by wheat and corn. Notably, unbiased coherence of the VV channel exhibited the highest correlation (R > 0.75). In addition, we applied unbiased coherence to crop classification. The results show that unbiased coherence exhibits very promising classification performance, with the overall accuracy (84.83%) and kappa coefficient (0.76) of VV improved by 8.35% and 0.12, respectively, over biased coherence, and the overall accuracy (73.25%) and kappa coefficient (0.57) of VH improved by 7.56% and 0.14, respectively, over biased coherence, and all crop classification accuracies were also effectively improved. This study demonstrates the feasibility of coherence monitoring of crops and provides new insights in enhancing the higher separability of crops.
摘要合成孔径雷达(SAR)在农作物监测和分类方面具有巨大的应用潜力。干涉合成孔径雷达(InSAR)的相干性证明可有效监测作物生长。目前,基于最大似然估计的相干性偏向于低相干性值。因此,这项工作的主要目的是了解哨兵-1 时间序列有偏相干性和无偏相干性在作物监测和分类中的性能。本研究于 2018 年生长季节(4 月至 10 月)在加拿大安大略省西南部的农业区科莫卡进行,主要种植三种作物:大豆、玉米和冬小麦。为验证相干性监测作物的能力,拟合了时间相干性与双偏振雷达植被指数(DpRVI)之间的线性相关系数。结果显示,时间相干性与 DpRVI 时间序列之间存在稳定的相关性,其中大豆的相关性最高(0.7 < R < 0.8),其次是小麦和玉米。值得注意的是,VV 通道的无偏相干性表现出最高的相关性(R > 0.75)。此外,我们还将无偏相干性应用于作物分类。结果表明,无偏相干性表现出非常好的分类性能,VV 的总体准确率(84.83%)和卡帕系数(0.76)比有偏相干性分别提高了 8.35% 和 0.12,VH 的总体准确率(73.25%)和卡帕系数(0.57)比有偏相干性分别提高了 7.56% 和 0.14,所有作物的分类准确率也都得到了有效提高。这项研究证明了农作物相干性监测的可行性,并为提高农作物的可分离性提供了新的见解。
{"title":"A Comparison of Sentinel-1 Biased and Unbiased Coherence for Crop Monitoring and Classification","authors":"Qinxin Zhao, Qinghua Xie, Xing Peng, Yusong Bao, Tonglu Jia, Linwei Yue, Haiqiang Fu, Jianjun Zhu","doi":"10.5194/isprs-archives-xlviii-1-2024-903-2024","DOIUrl":"https://doi.org/10.5194/isprs-archives-xlviii-1-2024-903-2024","url":null,"abstract":"Abstract. Synthetic Aperture Radar (SAR) holds significant potential for applications in crop monitoring and classification. Interferometric SAR (InSAR) coherence proves effective in monitoring crop growth. Currently, the coherence based on the maximum likelihood estimator is biased towards low coherence values. Therefore, the main aim of this work is to access the performance of Sentinel-1 time-series biased coherence and unbiased coherence in crop monitoring and classification. This study was conducted during the 2018 growing season (April-October) in Komoka, an agricultural region in southwestern Ontario, Canada, primarily cultivating three crops: soybean, corn, and winter wheat. To verify the ability of coherence to monitor crops, a linear correlation coefficient between temporal coherence and dual polarimetric radar vegetation index (DpRVI) was fitted. The results revealed a stable correlation between temporal coherence and DpRVI time-series, with the highest correlation observed for soybean (0.7 < R < 0.8), followed by wheat and corn. Notably, unbiased coherence of the VV channel exhibited the highest correlation (R > 0.75). In addition, we applied unbiased coherence to crop classification. The results show that unbiased coherence exhibits very promising classification performance, with the overall accuracy (84.83%) and kappa coefficient (0.76) of VV improved by 8.35% and 0.12, respectively, over biased coherence, and the overall accuracy (73.25%) and kappa coefficient (0.57) of VH improved by 7.56% and 0.14, respectively, over biased coherence, and all crop classification accuracies were also effectively improved. This study demonstrates the feasibility of coherence monitoring of crops and provides new insights in enhancing the higher separability of crops.\u0000","PeriodicalId":505918,"journal":{"name":"The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences","volume":"121 42","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140987567","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Crop Phenology Estimation in Rice Fields Using Sentinel-1 GRD SAR Data and Machine Learning-Aided Particle Filtering Approach 利用哨兵-1 GRD合成孔径雷达数据和机器学习辅助粒子滤波方法估算稻田作物物候期
Pub Date : 2024-05-11 DOI: 10.5194/isprs-archives-xlviii-1-2024-799-2024
Jie Yang, Hongtao Shi, Qinghua Xie, Juan M. Lopez-Sanchez, Xing Peng, Jianghao Yu, Lei Chen
Abstract. Monitoring crop phenology is essential for managing field disasters, protecting the environment, and making decisions about agricultural productivity. Because of its high timeliness, high resolution, great penetration, and sensitivity to specific structural elements, synthetic aperture radar (SAR) is a valuable technique for crop phenology estimation. Particle filtering (PF) belongs to the family of dynamical approach and has the ability to predict crop phenology with SAR data in real time. The observation equation is a key factor affecting the accuracy of particle filtering estimation and depends on fitting. Compared to the common polynomial fitting (POLY), machine learning methods can automatically learn features and handle complex data structures, offering greater flexibility and generalization capabilities. Therefore, incorporating two ensemble learning algorithms consisting of support vector machine regression (SVR), random forest regression (RFR), respectively, we proposed two machine learning-aided particle filtering approaches (PF-SVR, PF-RFR) to estimate crop phenology. One year of time-series Sentinel-1 GRD SAR data in 2017 covering rice fields in Sevilla region in Spain was used for establishing the observation and prediction equations, and the other year of data in 2018 was used for validating the prediction accuracy of PF methods. Four polarization features (VV, VH, VH/VV and Radar Vegetation Index (RVI)) were exploited as the observations in modeling. Experimental results reveals that the machine learning-aided methods are superior than the PF-POLY method. The PF-SVR exhibited better performance than the PF-RFR and PF-POLY methods. The optimal outcome from PF-SVR yielded a root-mean-square error (RMSE) of 7.79, compared to 7.94 for PF-RFR and 9.1 for PF-POLY. Moreover, the results suggest that the RVI is generally more sensitive than other features to crop phenology and the performance of polarization features presented consistent among all methods, i.e., RVI>VV>VH>VH/VV. Our findings offer valuable references for real-time crop phenology monitoring with SAR data.
摘要监测作物物候对于管理田间灾害、保护环境和制定农业生产决策至关重要。合成孔径雷达(SAR)具有高时效性、高分辨率、高穿透性以及对特定结构元素的敏感性等特点,是估测农作物物候的重要技术。粒子滤波(PF)属于动力学方法,能够利用合成孔径雷达数据实时预测作物物候。观测方程是影响粒子滤波估计精度的关键因素,并取决于拟合。与常见的多项式拟合(POLY)相比,机器学习方法可以自动学习特征并处理复杂的数据结构,具有更大的灵活性和泛化能力。因此,我们结合由支持向量机回归(SVR)和随机森林回归(RFR)分别组成的两种集合学习算法,提出了两种机器学习辅助粒子滤波方法(PF-SVR、PF-RFR)来估计作物物候。其中,2017 年的一年时间序列 Sentinel-1 GRD SAR 数据用于建立观测和预测方程,覆盖西班牙塞维利亚地区的水稻田;2018 年的另一年数据用于验证 PF 方法的预测精度。建模时利用了四个偏振特征(VV、VH、VH/VV 和雷达植被指数(RVI))作为观测值。实验结果表明,机器学习辅助方法优于 PF-POLY 方法。PF-SVR 的性能优于 PF-RFR 和 PF-POLY 方法。PF-SVR 的最优结果均方根误差(RMSE)为 7.79,而 PF-RFR 为 7.94,PF-POLY 为 9.1。此外,研究结果表明,RVI 通常比其他特征对作物物候更为敏感,而且偏振特征在所有方法中表现一致,即 RVI>V>VH>VH/VV。我们的研究结果为利用合成孔径雷达数据实时监测作物物候提供了有价值的参考。
{"title":"Crop Phenology Estimation in Rice Fields Using Sentinel-1 GRD SAR Data and Machine Learning-Aided Particle Filtering Approach","authors":"Jie Yang, Hongtao Shi, Qinghua Xie, Juan M. Lopez-Sanchez, Xing Peng, Jianghao Yu, Lei Chen","doi":"10.5194/isprs-archives-xlviii-1-2024-799-2024","DOIUrl":"https://doi.org/10.5194/isprs-archives-xlviii-1-2024-799-2024","url":null,"abstract":"Abstract. Monitoring crop phenology is essential for managing field disasters, protecting the environment, and making decisions about agricultural productivity. Because of its high timeliness, high resolution, great penetration, and sensitivity to specific structural elements, synthetic aperture radar (SAR) is a valuable technique for crop phenology estimation. Particle filtering (PF) belongs to the family of dynamical approach and has the ability to predict crop phenology with SAR data in real time. The observation equation is a key factor affecting the accuracy of particle filtering estimation and depends on fitting. Compared to the common polynomial fitting (POLY), machine learning methods can automatically learn features and handle complex data structures, offering greater flexibility and generalization capabilities. Therefore, incorporating two ensemble learning algorithms consisting of support vector machine regression (SVR), random forest regression (RFR), respectively, we proposed two machine learning-aided particle filtering approaches (PF-SVR, PF-RFR) to estimate crop phenology. One year of time-series Sentinel-1 GRD SAR data in 2017 covering rice fields in Sevilla region in Spain was used for establishing the observation and prediction equations, and the other year of data in 2018 was used for validating the prediction accuracy of PF methods. Four polarization features (VV, VH, VH/VV and Radar Vegetation Index (RVI)) were exploited as the observations in modeling. Experimental results reveals that the machine learning-aided methods are superior than the PF-POLY method. The PF-SVR exhibited better performance than the PF-RFR and PF-POLY methods. The optimal outcome from PF-SVR yielded a root-mean-square error (RMSE) of 7.79, compared to 7.94 for PF-RFR and 9.1 for PF-POLY. Moreover, the results suggest that the RVI is generally more sensitive than other features to crop phenology and the performance of polarization features presented consistent among all methods, i.e., RVI>VV>VH>VH/VV. Our findings offer valuable references for real-time crop phenology monitoring with SAR data.\u0000","PeriodicalId":505918,"journal":{"name":"The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences","volume":" 1149","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140988735","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1