首页 > 最新文献

IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society最新文献

英文 中文
Source Independent Reflection Waveform Inversion 源独立反射波形反演
Zhanyuan Liang;Xiaoyu Zhang;Guoqiang Shen;Zhentao Wang;Xiping Wang
Reflection waveform inversion (RWI) updates the low- to mid-wavenumber components of the velocity model accurately by projecting the waveform errors between the observed and synthetic data onto the reflection wave paths. However, the synthetic data, generated with the aid of migration/demigration, exhibit unexpected waveform deviations from the observed data due to unknown source wavelets, potentially interfering with inversion outcomes. To address this issue, we propose a source-independent RWI (SI-RWI) method. Initially, the equivalent source spectrum of the migration/demigration process is derived in the frequency domain. Subsequently, the misfit function for RWI is designed to ensure that the observed and synthetic data share the same equivalent source spectrum. Finally, based on this novel misfit function, an RWI method is formulated that does not rely on the phase distortions of source wavelets. The proposed approach has been demonstrated successfully using 2-D examples.
反射波形反演(RWI)通过将观测数据和合成数据之间的波形误差投影到反射波路径上,准确地更新速度模型的中低波数分量。然而,在偏移/反偏移的帮助下生成的合成数据,由于未知源小波的存在,与观测数据出现了意想不到的波形偏差,可能会干扰反演结果。为了解决这个问题,我们提出了一种独立于源的RWI (SI-RWI)方法。首先,在频域中推导出偏移/反偏移过程的等效源谱。随后,设计RWI失配函数,以确保观测数据和合成数据共享相同的等效源谱。最后,在此基础上,提出了一种不依赖于源小波相位畸变的RWI方法。该方法已成功地通过二维实例进行了验证。
{"title":"Source Independent Reflection Waveform Inversion","authors":"Zhanyuan Liang;Xiaoyu Zhang;Guoqiang Shen;Zhentao Wang;Xiping Wang","doi":"10.1109/LGRS.2025.3640124","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3640124","url":null,"abstract":"Reflection waveform inversion (RWI) updates the low- to mid-wavenumber components of the velocity model accurately by projecting the waveform errors between the observed and synthetic data onto the reflection wave paths. However, the synthetic data, generated with the aid of migration/demigration, exhibit unexpected waveform deviations from the observed data due to unknown source wavelets, potentially interfering with inversion outcomes. To address this issue, we propose a source-independent RWI (SI-RWI) method. Initially, the equivalent source spectrum of the migration/demigration process is derived in the frequency domain. Subsequently, the misfit function for RWI is designed to ensure that the observed and synthetic data share the same equivalent source spectrum. Finally, based on this novel misfit function, an RWI method is formulated that does not rely on the phase distortions of source wavelets. The proposed approach has been demonstrated successfully using 2-D examples.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145830730","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Seismic Denoising via Multiround SCU-Net 基于多轮scu网的地震去噪
Yuli Qi;Guoxin Chen;Jinxin Chen;Jun Li;Rongsen Du;Haiyang Lu;Naijian Wang;Xingguo Huang
As oil and gas explorations progressively advance toward deeper and more complex geological formations, the imperative for precise characterization of subsurface structures has become increasingly prominent. The efficacy of noise suppression is a critical determinant for the quality of subsequent inversion and imaging processes. In recent years, deep learning methodologies have garnered significant attention and widespread application in seismic denoising, primarily due to their inherent data-driven advantages. While conventional deep learning implementations have achieved notable denoising performance, they are confronted with inherent limitations, including incomplete noise reduction and potential signal degradation. To address these challenges, this study proposes an innovative multiround SCU-Net (MR-SCU) denoising approach. The MR-SCU methodology based on SCU-Net employs noise as labeled data to generate an initial denoised outcome in the first round. Denoising results are used as input while utilizing the residuals between the labeled and predicted data as labels for subsequent denoising round. Multiple rounds are iteratively repeated to achieve more thorough denoising effect while preserving effective signals from being compromised. The incorporation of SSIM (structural similarity index measure) as the loss function further enhances the method’s precision in detail-oriented denoising tasks. Numerical experiments conducted on synthetic data and field data acquired from a specific region in western China substantiate the efficacy of the MR-SCU, demonstrating its capability to deliver superior denoising performance while optimally preserve valuable seismic information.
随着油气勘探逐渐向更深、更复杂的地质构造推进,精确表征地下构造的必要性日益突出。噪声抑制的有效性是后续反演和成像过程质量的关键决定因素。近年来,深度学习方法由于其固有的数据驱动优势,在地震去噪中得到了广泛的关注和应用。虽然传统的深度学习实现已经取得了显著的去噪性能,但它们面临着固有的局限性,包括不完全的降噪和潜在的信号退化。为了应对这些挑战,本研究提出了一种创新的多轮SCU-Net (MR-SCU)去噪方法。基于SCU-Net的MR-SCU方法使用噪声作为标记数据,在第一轮中生成初始去噪结果。去噪结果用作输入,同时利用标记数据和预测数据之间的残差作为后续去噪轮的标签。多轮迭代重复,实现更彻底的去噪效果,同时保持有效信号不被破坏。将SSIM (structural similarity index measure)作为损失函数,进一步提高了该方法在面向细节的去噪任务中的精度。对中国西部特定地区的合成数据和现场数据进行的数值实验证实了MR-SCU的有效性,证明了它能够提供优越的去噪性能,同时最佳地保留有价值的地震信息。
{"title":"Seismic Denoising via Multiround SCU-Net","authors":"Yuli Qi;Guoxin Chen;Jinxin Chen;Jun Li;Rongsen Du;Haiyang Lu;Naijian Wang;Xingguo Huang","doi":"10.1109/LGRS.2025.3640968","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3640968","url":null,"abstract":"As oil and gas explorations progressively advance toward deeper and more complex geological formations, the imperative for precise characterization of subsurface structures has become increasingly prominent. The efficacy of noise suppression is a critical determinant for the quality of subsequent inversion and imaging processes. In recent years, deep learning methodologies have garnered significant attention and widespread application in seismic denoising, primarily due to their inherent data-driven advantages. While conventional deep learning implementations have achieved notable denoising performance, they are confronted with inherent limitations, including incomplete noise reduction and potential signal degradation. To address these challenges, this study proposes an innovative multiround SCU-Net (MR-SCU) denoising approach. The MR-SCU methodology based on SCU-Net employs noise as labeled data to generate an initial denoised outcome in the first round. Denoising results are used as input while utilizing the residuals between the labeled and predicted data as labels for subsequent denoising round. Multiple rounds are iteratively repeated to achieve more thorough denoising effect while preserving effective signals from being compromised. The incorporation of SSIM (structural similarity index measure) as the loss function further enhances the method’s precision in detail-oriented denoising tasks. Numerical experiments conducted on synthetic data and field data acquired from a specific region in western China substantiate the efficacy of the MR-SCU, demonstrating its capability to deliver superior denoising performance while optimally preserve valuable seismic information.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145778222","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TEFormer: Texture-Aware and Edge-Guided Transformer for Semantic Segmentation of Urban Remote Sensing Images TEFormer:用于城市遥感图像语义分割的纹理感知和边缘引导转换器
Guoyu Zhou;Jing Zhang;Yi Yan;Hui Zhang;Li Zhuo
Accurate semantic segmentation of urban remote sensing images (URSIs) is essential for urban planning and environmental monitoring. However, it remains challenging due to the subtle texture differences and similar spatial structures among geospatial objects, which cause semantic ambiguity and misclassification. Additional complexities arise from irregular object shapes, blurred boundaries, and overlapping spatial distributions of objects, resulting in diverse and intricate edge morphologies. To address these issues, we propose TEFormer, a texture-aware and edge-guided Transformer. Our model features a texture-aware module (TaM) in the encoder to capture fine-grained texture distinctions between visually similar categories, thereby enhancing semantic discrimination. The decoder incorporates an edge-guided tri-branch decoder (Eg3Head) to preserve local edges and details while maintaining multiscale context-awareness. Finally, an edge-guided feature fusion module (EgFFM) effectively integrates contextual, detail, and edge information to achieve refined semantic segmentation. Extensive evaluation demonstrates that TEFormer yields mean intersection over union (mIoU) scores of 88.57% on Potsdam and 81.46% on Vaihingen, exceeding the next best methods by 0.73% and 0.22%. On the LoveDA dataset, it secures the second position with an overall mIoU of 53.55%, trailing the optimal performance by a narrow margin of 0.19%.
城市遥感图像的准确语义分割是城市规划和环境监测的基础。然而,由于地理空间对象之间存在细微的纹理差异和相似的空间结构,从而导致语义模糊和误分类,这一问题仍然具有挑战性。由于物体形状不规则,边界模糊,物体空间分布重叠,导致边缘形态多样而复杂,从而产生额外的复杂性。为了解决这些问题,我们提出了TEFormer,一个纹理感知和边缘引导的变压器。我们的模型在编码器中具有纹理感知模块(TaM),用于捕获视觉上相似类别之间的细粒度纹理差异,从而增强语义区分。该解码器结合了一个边缘引导的三分支解码器(Eg3Head),以保留局部边缘和细节,同时保持多尺度上下文感知。最后,利用边缘引导特征融合模块(EgFFM)有效地整合上下文、细节和边缘信息,实现精细化的语义分割。广泛评价表明,TEFormer方法在波茨坦和瓦伊辛根的平均mIoU分数分别为88.57%和81.46%,分别比次优方法高出0.73%和0.22%。在LoveDA数据集上,它以53.55%的总体mIoU排名第二,以0.19%的微弱差距落后于最佳表现。
{"title":"TEFormer: Texture-Aware and Edge-Guided Transformer for Semantic Segmentation of Urban Remote Sensing Images","authors":"Guoyu Zhou;Jing Zhang;Yi Yan;Hui Zhang;Li Zhuo","doi":"10.1109/LGRS.2025.3639147","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3639147","url":null,"abstract":"Accurate semantic segmentation of urban remote sensing images (URSIs) is essential for urban planning and environmental monitoring. However, it remains challenging due to the subtle texture differences and similar spatial structures among geospatial objects, which cause semantic ambiguity and misclassification. Additional complexities arise from irregular object shapes, blurred boundaries, and overlapping spatial distributions of objects, resulting in diverse and intricate edge morphologies. To address these issues, we propose TEFormer, a texture-aware and edge-guided Transformer. Our model features a texture-aware module (TaM) in the encoder to capture fine-grained texture distinctions between visually similar categories, thereby enhancing semantic discrimination. The decoder incorporates an edge-guided tri-branch decoder (Eg3Head) to preserve local edges and details while maintaining multiscale context-awareness. Finally, an edge-guided feature fusion module (EgFFM) effectively integrates contextual, detail, and edge information to achieve refined semantic segmentation. Extensive evaluation demonstrates that TEFormer yields mean intersection over union (mIoU) scores of 88.57% on Potsdam and 81.46% on Vaihingen, exceeding the next best methods by 0.73% and 0.22%. On the LoveDA dataset, it secures the second position with an overall mIoU of 53.55%, trailing the optimal performance by a narrow margin of 0.19%.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145778331","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Applying ViT Masked Autoencoders to Seismic Data for Feature Extraction and Few-Shot Learning 应用ViT掩码自编码器对地震数据进行特征提取和少镜头学习
Fernando G. Marques;Carlos A. Astudillo;Alan Souza;Daniel Miranda;Edson Borin
We apply the self-supervised learning (SSL) technique of vision transformers masked autoencoder (ViTs MAE) models with the goal of producing a feature extractor ViT backbone for neural networks that receive seismic data as an input. We then evaluate the quality of these backbones by coupling them to a simple linear prediction head and fine-tuning these models in a seismic semantic segmentation task. We compare domain-specific ViT MAE against cross-domain pretrained and randomly initialized ViTs and show that it yields superior performance in low-data regimes. We also demonstrate that pretraining loss correlates with downstream performance, supporting its use as a proxy for feature quality.
我们应用视觉变压器掩膜自编码器(ViTs MAE)模型的自监督学习(SSL)技术,目的是为接收地震数据作为输入的神经网络生成特征提取器ViT主干。然后,我们通过将这些骨干耦合到一个简单的线性预测头并在地震语义分割任务中微调这些模型来评估它们的质量。我们将特定领域的ViT MAE与跨领域预训练和随机初始化的ViT进行了比较,并表明它在低数据区具有优越的性能。我们还证明了预训练损失与下游性能相关,支持将其用作特征质量的代理。
{"title":"Applying ViT Masked Autoencoders to Seismic Data for Feature Extraction and Few-Shot Learning","authors":"Fernando G. Marques;Carlos A. Astudillo;Alan Souza;Daniel Miranda;Edson Borin","doi":"10.1109/LGRS.2025.3639172","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3639172","url":null,"abstract":"We apply the self-supervised learning (SSL) technique of vision transformers masked autoencoder (ViTs MAE) models with the goal of producing a feature extractor ViT backbone for neural networks that receive seismic data as an input. We then evaluate the quality of these backbones by coupling them to a simple linear prediction head and fine-tuning these models in a seismic semantic segmentation task. We compare domain-specific ViT MAE against cross-domain pretrained and randomly initialized ViTs and show that it yields superior performance in low-data regimes. We also demonstrate that pretraining loss correlates with downstream performance, supporting its use as a proxy for feature quality.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145830823","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Stochastic Frequency-Dependent Velocity and Attenuation Inversion for Hydrocarbon Detection 油气探测的随机频率相关速度和衰减反演
Fang Ouyang;Jianguo Zhao;Xinze Liu;Bin Wang;Yu Zhang;Bohong Yan
Rock-physics theories and experiments have demonstrated that seismic wave velocity dispersion and attenuation are closely related to hydrocarbon deposits. To obtain the velocity at different seismic frequencies, the frequency-dependent amplitude variation with angle (AVA) inversion method has been developed to invert the dispersive velocity from frequency-domain P-wave reflection coefficients. Such a method can overcome the disability of the conventional AVA inversion in terms of seismic dispersion. However, the limitation is that only velocity dispersion is considered while the effects of seismic attenuation are neglected. In this letter, we proposed a new frequency-dependent AVA method, in which the thickness of reservoir and the complex dispersive P-wave velocity that includes the information of both dispersion and attenuation are simultaneously inverted. To better catch the characteristics of the reflections and transmissions between layers, the reflectivity method is adopted as the forward modeling engine. Furthermore, a modified simulated annealing method that takes advantages of the parameter-by-parameter optimization idea in heat-bath algorithm as well as the acceptance criteria used in Metropolis algorithm is developed, so as to achieve efficient and better global optimization for the complex inversion problem of high-degree nonlinearity and ill-posedness. Compared with previous frequency-dependent AVA methods, our improved approach can not only predict the P-wave velocity dispersion but also the frequency-dependent inverse quality factor of the reservoir layer. Using synthetic records and field data through a drilling well, the effectiveness and applicability of the proposed method in hydrocarbon indication are verified.
岩石物理理论和实验表明,地震波速度的频散和衰减与油气沉积密切相关。为了获得不同地震频率下的速度,提出了频率相关振幅随角度变化(AVA)反演方法,利用频域纵波反射系数反演频散速度。该方法克服了常规AVA反演在地震频散方面的不足。然而,其局限性在于只考虑了速度频散,而忽略了地震衰减的影响。在本文中,我们提出了一种新的频率相关的AVA方法,该方法同时反演储层厚度和包含色散和衰减信息的复色散纵波速度。为了更好地捕捉层间反射和透射的特征,采用反射率法作为正演引擎。在此基础上,利用热浴算法的逐参数优化思想和Metropolis算法的接受准则,提出了一种改进的模拟退火方法,对具有高度非线性和病态性的复杂反演问题实现高效、更好的全局优化。与以往的频率相关AVA方法相比,改进的方法不仅可以预测纵波速度频散,而且可以预测储层的频率相关逆质量因子。通过一口井的合成记录和现场资料,验证了该方法在油气指示方面的有效性和适用性。
{"title":"Stochastic Frequency-Dependent Velocity and Attenuation Inversion for Hydrocarbon Detection","authors":"Fang Ouyang;Jianguo Zhao;Xinze Liu;Bin Wang;Yu Zhang;Bohong Yan","doi":"10.1109/LGRS.2025.3637108","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3637108","url":null,"abstract":"Rock-physics theories and experiments have demonstrated that seismic wave velocity dispersion and attenuation are closely related to hydrocarbon deposits. To obtain the velocity at different seismic frequencies, the frequency-dependent amplitude variation with angle (AVA) inversion method has been developed to invert the dispersive velocity from frequency-domain P-wave reflection coefficients. Such a method can overcome the disability of the conventional AVA inversion in terms of seismic dispersion. However, the limitation is that only velocity dispersion is considered while the effects of seismic attenuation are neglected. In this letter, we proposed a new frequency-dependent AVA method, in which the thickness of reservoir and the complex dispersive P-wave velocity that includes the information of both dispersion and attenuation are simultaneously inverted. To better catch the characteristics of the reflections and transmissions between layers, the reflectivity method is adopted as the forward modeling engine. Furthermore, a modified simulated annealing method that takes advantages of the parameter-by-parameter optimization idea in heat-bath algorithm as well as the acceptance criteria used in Metropolis algorithm is developed, so as to achieve efficient and better global optimization for the complex inversion problem of high-degree nonlinearity and ill-posedness. Compared with previous frequency-dependent AVA methods, our improved approach can not only predict the P-wave velocity dispersion but also the frequency-dependent inverse quality factor of the reservoir layer. Using synthetic records and field data through a drilling well, the effectiveness and applicability of the proposed method in hydrocarbon indication are verified.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145674823","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
EFD-YOLO: An Improved YOLOv8 Network for River Floating Debris Object Detection EFD-YOLO:一种改进的YOLOv8网络用于河流漂浮物检测
Yier Yan;Zhibin Liang;Changhong Liu;Tao Zou
With the rapid development of uncrewed aerial vehicle (UAV) technology, UAVs have provided an innovative solution for floating debris monitoring. However, object detection in UAV images remains challenging due to high miss rates for small objects, insufficient low-level feature extraction, and computational redundancy. This letter proposes an efficient floating debris detection model based on YOLOv8n, named EFD-you only look once (YOLO), to address these issues. First, the edge fusion stem (EFStem) module is proposed to enhance low-level feature extraction through an integrated gate-attention mechanism. Second, the multibranch efficient reparameterization block (MBERB) is designed to achieve efficient cross-layer feature fusion. Experimental results demonstrate that compared to YOLOv8n, our model achieves a 6.3% improvement in mean average precision (mAP) on the UAV floating debris dataset, while simultaneously reducing parameters by 26.7% and improving small object recall by 21.9%. The inference time of EFD-YOLO on the RK3588 edge device is as low as 30.5 ms, demonstrating real-time capability.
随着无人飞行器(UAV)技术的快速发展,无人机为浮物监测提供了创新的解决方案。然而,无人机图像中的目标检测仍然具有挑战性,因为小目标的高缺失率、低层次特征提取不足和计算冗余。为了解决这些问题,本信函提出了一种基于YOLOv8n的高效漂浮碎片检测模型,称为EFD-you only look once (YOLO)。首先,提出边缘融合干(EFStem)模块,通过集成门-注意机制增强底层特征提取;其次,设计多分支高效重参数化块(MBERB),实现高效的跨层特征融合;实验结果表明,与YOLOv8n相比,我们的模型在无人机漂浮碎片数据集上的平均精度(mAP)提高了6.3%,同时参数减少了26.7%,小目标召回率提高了21.9%。EFD-YOLO在RK3588边缘器件上的推理时间低至30.5 ms,显示出实时性。
{"title":"EFD-YOLO: An Improved YOLOv8 Network for River Floating Debris Object Detection","authors":"Yier Yan;Zhibin Liang;Changhong Liu;Tao Zou","doi":"10.1109/LGRS.2025.3636279","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3636279","url":null,"abstract":"With the rapid development of uncrewed aerial vehicle (UAV) technology, UAVs have provided an innovative solution for floating debris monitoring. However, object detection in UAV images remains challenging due to high miss rates for small objects, insufficient low-level feature extraction, and computational redundancy. This letter proposes an efficient floating debris detection model based on YOLOv8n, named EFD-you only look once (YOLO), to address these issues. First, the edge fusion stem (EFStem) module is proposed to enhance low-level feature extraction through an integrated gate-attention mechanism. Second, the multibranch efficient reparameterization block (MBERB) is designed to achieve efficient cross-layer feature fusion. Experimental results demonstrate that compared to YOLOv8n, our model achieves a 6.3% improvement in mean average precision (mAP) on the UAV floating debris dataset, while simultaneously reducing parameters by 26.7% and improving small object recall by 21.9%. The inference time of EFD-YOLO on the RK3588 edge device is as low as 30.5 ms, demonstrating real-time capability.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-11-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145674820","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
User-Driven Land Cover Change Prediction Map Tool for Land Conservation Planning 基于用户驱动的土地保护规划土地覆盖变化预测地图工具
Pui-Yu Ling;Laura Nunes;Jonathan Srinivasan;Nasir Popalzay;Palmer Wilson;Jameson Quisenberry;Alex Borowicz
To be effective, ecosystem and habitat conservation must not only look at past losses but also understand the effects of current and future decisions on landscapes. Here, we present a transformative, user-driven land cover change prediction tool designed to aid land planners in strategic decision-making for conservation and habitat protection. Within an integrated map-based prediction pipeline, the tool uses machine learning (ML) and deep learning (DL) models to classify satellite images and make predictions of near-term land cover changes. The tool facilitates user interaction with a cloud-hosted ML model, making it accessible to nontechnical users for generating map-based predictions using big data. The tool’s key strength lies in its dynamic variable adjustment feature, empowering users to tailor scenarios related to potential future development planning. Through the integration of cloud-hosted ML and DL models with a user-centric interface, the tool has the potential to allow stakeholders and land planners to make informed decisions, actively minimizing habitat destruction and aligning with broader conservation objectives. We tested our approach in the context of central Texas, USA to evaluate its effectiveness in diverse conservation scenarios, with an average overall accuracy of 88% for the land cover class maps over four years and over 72% for the five-year land cover change prediction. While our approach has the potential to improve land management and planning for conservation, we also acknowledge the importance of rigorous model validation and ongoing refinement and highlight the need for technological advancement to be developed with strong stakeholder engagement.
为了有效地保护生态系统和栖息地,不仅要关注过去的损失,还要了解当前和未来的决策对景观的影响。在这里,我们提出了一个变革性的、用户驱动的土地覆盖变化预测工具,旨在帮助土地规划者进行保护和栖息地保护的战略决策。在集成的基于地图的预测管道中,该工具使用机器学习(ML)和深度学习(DL)模型对卫星图像进行分类,并对近期土地覆盖变化进行预测。该工具促进了用户与云托管ML模型的交互,使非技术用户可以使用大数据生成基于地图的预测。该工具的关键优势在于其动态变量调整功能,使用户能够根据潜在的未来发展规划定制场景。通过将云托管的ML和DL模型与以用户为中心的界面集成,该工具有可能使利益相关者和土地规划者做出明智的决策,积极减少栖息地破坏,并与更广泛的保护目标保持一致。我们在美国德克萨斯州中部测试了我们的方法,以评估其在不同保护情景下的有效性,4年土地覆盖分类图的平均总体精度为88%,5年土地覆盖变化预测的平均总体精度超过72%。虽然我们的方法有可能改善土地管理和保护规划,但我们也承认严格的模型验证和不断改进的重要性,并强调需要在利益相关者的大力参与下开发技术进步。
{"title":"User-Driven Land Cover Change Prediction Map Tool for Land Conservation Planning","authors":"Pui-Yu Ling;Laura Nunes;Jonathan Srinivasan;Nasir Popalzay;Palmer Wilson;Jameson Quisenberry;Alex Borowicz","doi":"10.1109/LGRS.2025.3636286","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3636286","url":null,"abstract":"To be effective, ecosystem and habitat conservation must not only look at past losses but also understand the effects of current and future decisions on landscapes. Here, we present a transformative, user-driven land cover change prediction tool designed to aid land planners in strategic decision-making for conservation and habitat protection. Within an integrated map-based prediction pipeline, the tool uses machine learning (ML) and deep learning (DL) models to classify satellite images and make predictions of near-term land cover changes. The tool facilitates user interaction with a cloud-hosted ML model, making it accessible to nontechnical users for generating map-based predictions using big data. The tool’s key strength lies in its dynamic variable adjustment feature, empowering users to tailor scenarios related to potential future development planning. Through the integration of cloud-hosted ML and DL models with a user-centric interface, the tool has the potential to allow stakeholders and land planners to make informed decisions, actively minimizing habitat destruction and aligning with broader conservation objectives. We tested our approach in the context of central Texas, USA to evaluate its effectiveness in diverse conservation scenarios, with an average overall accuracy of 88% for the land cover class maps over four years and over 72% for the five-year land cover change prediction. While our approach has the potential to improve land management and planning for conservation, we also acknowledge the importance of rigorous model validation and ongoing refinement and highlight the need for technological advancement to be developed with strong stakeholder engagement.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-11-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11265796","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145729300","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Self-Prompt Calibration Network Based on Segment Anything Model 2 for High-Resolution Remote Sensing Image Segmentation 基于分段任意模型2的高分辨率遥感图像自提示标定网络
Yizhou Lan;Daoyuan Zheng;Xinge Zhao;Ke Shang;Feizhou Zhang
Remote sensing image segmentation is particularly difficult due to the coexistence of large-scale variations and fine-grained structures in very high-resolution imagery. Conventional CNN-based or transformer-based networks often struggle to capture global context while preserving boundary details, leading to degraded performance on small or thin objects. To address these challenges, we propose a self-prompt calibration network based on segment anything model 2 (SC-SAM). The SC-SAM achieves self-prompt by feeding mask prompts from a lightweight decoder into frozen prompt encoder. Output calibration is achieved through the proposed cross-probability guided calibration (CPGC) module, which employs cross-probability uncertainty as complementary guidance to refine final predictions via self-prompted outputs. Furthermore, to better preserve contextual and structural information across multiple scales, a scale-decoupled kernel mixture (SDKM) module is designed. Experimental results on the ISPRS Vaihingen and Potsdam dataset demonstrate that the proposed approach surpasses the state-of-the-art methods by 1.02% and 1.34% in mIoU, highlighting its effectiveness. This study provides new insights into adapting SAM for domain-specific remote sensing segmentation tasks.
由于在高分辨率影像中同时存在大尺度变化和细粒度结构,因此遥感影像分割尤其困难。传统的基于cnn或基于变压器的网络通常难以捕捉全局上下文,同时保留边界细节,导致在小或薄对象上的性能下降。为了解决这些挑战,我们提出了一个基于分段任意模型2 (SC-SAM)的自提示校准网络。SC-SAM通过从轻量级解码器向冻结提示编码器馈送掩码提示来实现自提示。输出校准通过提出的交叉概率引导校准(CPGC)模块实现,该模块采用交叉概率不确定性作为补充指导,通过自提示输出来改进最终预测。此外,为了更好地跨尺度保存上下文信息和结构信息,设计了尺度解耦核混合(SDKM)模块。在ISPRS Vaihingen和Potsdam数据集上的实验结果表明,该方法的mIoU比现有方法分别高出1.02%和1.34%,表明了该方法的有效性。该研究为将SAM应用于特定领域的遥感分割任务提供了新的见解。
{"title":"A Self-Prompt Calibration Network Based on Segment Anything Model 2 for High-Resolution Remote Sensing Image Segmentation","authors":"Yizhou Lan;Daoyuan Zheng;Xinge Zhao;Ke Shang;Feizhou Zhang","doi":"10.1109/LGRS.2025.3636177","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3636177","url":null,"abstract":"Remote sensing image segmentation is particularly difficult due to the coexistence of large-scale variations and fine-grained structures in very high-resolution imagery. Conventional CNN-based or transformer-based networks often struggle to capture global context while preserving boundary details, leading to degraded performance on small or thin objects. To address these challenges, we propose a self-prompt calibration network based on segment anything model 2 (SC-SAM). The SC-SAM achieves self-prompt by feeding mask prompts from a lightweight decoder into frozen prompt encoder. Output calibration is achieved through the proposed cross-probability guided calibration (CPGC) module, which employs cross-probability uncertainty as complementary guidance to refine final predictions via self-prompted outputs. Furthermore, to better preserve contextual and structural information across multiple scales, a scale-decoupled kernel mixture (SDKM) module is designed. Experimental results on the ISPRS Vaihingen and Potsdam dataset demonstrate that the proposed approach surpasses the state-of-the-art methods by 1.02% and 1.34% in mIoU, highlighting its effectiveness. This study provides new insights into adapting SAM for domain-specific remote sensing segmentation tasks.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-11-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145674840","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Federated Aerial Video Captioning With Effective Temporal Adaptation 有效时间自适应的联邦航空视频字幕
Nguyen Anh Tu;Nursultan Makhanov;Kenzhebek Taniyev;Ton Duc Do
Aerial video captioning (VC) facilitates the automatic interpretation of dynamic scenes in remote sensing (RS), supporting critical applications, such as disaster response, traffic monitoring, and environmental surveillance. However, challenges, such as extreme angles and continuous camera motion, require adaptive modeling of complex temporal relationships. To tackle these challenges, we leverage an image-language model as the vision encoder and introduce a temporal adaptation module that combines convolution with self-attention layers to both capture local semantics across neighboring frames and model global temporal dependencies. This design allows our model to exploit the multimodal knowledge of the vision encoder while effectively reasoning over the spatiotemporal dynamics. In addition, privacy concerns often restrict access to annotated aerial datasets, posing further challenges for model training. To address this, we develop a federated learning (FL) framework that enables collaborative model training across decentralized clients. Within this framework, we establish a unified benchmark for systematic comparison of temporal adapters, text decoders, and FL strategies, hence filling a gap in the existing literature. Extensive experiments validate the robustness of our approach and its potential for advancing aerial VC.
航空视频字幕(VC)促进了遥感(RS)中动态场景的自动解释,支持关键应用,如灾害响应、交通监控和环境监测。然而,极端角度和连续摄像机运动等挑战需要对复杂的时间关系进行自适应建模。为了应对这些挑战,我们利用图像语言模型作为视觉编码器,并引入时间适应模块,该模块将卷积与自关注层结合起来,既可以捕获相邻帧之间的局部语义,又可以对全局时间依赖性进行建模。这种设计允许我们的模型利用视觉编码器的多模态知识,同时有效地对时空动态进行推理。此外,隐私问题往往限制了对带注释的航空数据集的访问,这给模型训练带来了进一步的挑战。为了解决这个问题,我们开发了一个联邦学习(FL)框架,支持跨分散客户端的协作模型训练。在这个框架内,我们建立了一个统一的基准,用于系统比较时间适配器、文本解码器和FL策略,从而填补了现有文献中的空白。大量的实验验证了我们的方法的鲁棒性及其推进空中VC的潜力。
{"title":"Federated Aerial Video Captioning With Effective Temporal Adaptation","authors":"Nguyen Anh Tu;Nursultan Makhanov;Kenzhebek Taniyev;Ton Duc Do","doi":"10.1109/LGRS.2025.3636165","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3636165","url":null,"abstract":"Aerial video captioning (VC) facilitates the automatic interpretation of dynamic scenes in remote sensing (RS), supporting critical applications, such as disaster response, traffic monitoring, and environmental surveillance. However, challenges, such as extreme angles and continuous camera motion, require adaptive modeling of complex temporal relationships. To tackle these challenges, we leverage an image-language model as the vision encoder and introduce a temporal adaptation module that combines convolution with self-attention layers to both capture local semantics across neighboring frames and model global temporal dependencies. This design allows our model to exploit the multimodal knowledge of the vision encoder while effectively reasoning over the spatiotemporal dynamics. In addition, privacy concerns often restrict access to annotated aerial datasets, posing further challenges for model training. To address this, we develop a federated learning (FL) framework that enables collaborative model training across decentralized clients. Within this framework, we establish a unified benchmark for systematic comparison of temporal adapters, text decoders, and FL strategies, hence filling a gap in the existing literature. Extensive experiments validate the robustness of our approach and its potential for advancing aerial VC.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-11-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145674849","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Integrated Framework for Estimating the All-Sky Surface Downward Longwave Radiation From FY-3D/MERSI-II Imagery 基于FY-3D/MERSI-II影像的全天空表面向下长波辐射综合估算框架
Qi Zeng;Wanchun Zhang;Jie Cheng
This study develops an integrated framework for all-sky surface longwave downward radiation (SLDR) estimate for the medium resolution spectral imager-II (MERSI-II) onboard the Fengyun-3D (FY-3D) satellite. The framework comprises a hybrid method for the clear-sky SLDR estimate and a cloud base temperature (CBT)-based single-layer cloud model (SLCM) for the cloudy-sky SLDR estimate. In situ validation indicates that the hybrid method yields a bias/RMSE of −0.78/21.70 W/m2, whereas the SLCM achieves a bias/RMSE of 5.79/23.61 W/m2. The bias/RMSE of the all-sky SLDR is 3.37/22.93 W/m2. The estimated all-sky instantaneous SLDR was combined with ERA5 temporal information to derive daily SLDR using a bias-corrected sinusoidal integration method, yielding a bias of 0.04 W/m2 and an RMSE of 16.77 W/m2. These results demonstrate the robustness of the proposed framework and its substantial potential in generating both instantaneous and daily SLDR products at 1 km spatial resolution.
为风云三号卫星(FY-3D)上的中分辨率光谱成像仪ii (MERSI-II)开发了全天表面长波向下辐射(SLDR)估算的集成框架。该框架包括用于晴空SLDR估计的混合方法和用于云基温度(CBT)估计的单层云模型(SLCM)。原位验证表明,混合方法的偏差/RMSE为- 0.78/21.70 W/m2,而SLCM方法的偏差/RMSE为5.79/23.61 W/m2。全天SLDR的偏差/均方根误差为3.37/22.93 W/m2。将估计的全天瞬时SLDR与ERA5时间信息结合使用偏差校正正弦积分法得到日SLDR,偏差为0.04 W/m2, RMSE为16.77 W/m2。这些结果证明了所提出的框架的鲁棒性及其在生成1公里空间分辨率的瞬时和每日SLDR产品方面的巨大潜力。
{"title":"An Integrated Framework for Estimating the All-Sky Surface Downward Longwave Radiation From FY-3D/MERSI-II Imagery","authors":"Qi Zeng;Wanchun Zhang;Jie Cheng","doi":"10.1109/LGRS.2025.3636236","DOIUrl":"https://doi.org/10.1109/LGRS.2025.3636236","url":null,"abstract":"This study develops an integrated framework for all-sky surface longwave downward radiation (SLDR) estimate for the medium resolution spectral imager-II (MERSI-II) onboard the Fengyun-3D (FY-3D) satellite. The framework comprises a hybrid method for the clear-sky SLDR estimate and a cloud base temperature (CBT)-based single-layer cloud model (SLCM) for the cloudy-sky SLDR estimate. In situ validation indicates that the hybrid method yields a bias/RMSE of −0.78/21.70 W/m2, whereas the SLCM achieves a bias/RMSE of 5.79/23.61 W/m2. The bias/RMSE of the all-sky SLDR is 3.37/22.93 W/m2. The estimated all-sky instantaneous SLDR was combined with ERA5 temporal information to derive daily SLDR using a bias-corrected sinusoidal integration method, yielding a bias of 0.04 W/m2 and an RMSE of 16.77 W/m2. These results demonstrate the robustness of the proposed framework and its substantial potential in generating both instantaneous and daily SLDR products at 1 km spatial resolution.","PeriodicalId":91017,"journal":{"name":"IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society","volume":"23 ","pages":"1-5"},"PeriodicalIF":4.4,"publicationDate":"2025-11-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145830854","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE geoscience and remote sensing letters : a publication of the IEEE Geoscience and Remote Sensing Society
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1