首页 > 最新文献

Artificial Intelligence in Geosciences最新文献

英文 中文
Microseismic moment tensor inversion based on ResNet model 基于ResNet模型的微震矩张量反演
Pub Date : 2025-03-01 DOI: 10.1016/j.aiig.2025.100107
Jiaqi Yan , Li Ma , Tianqi Jiang , Jing Zheng , Dewei Li , Xingzhi Teng
This paper proposed a moment tensor regression prediction technology based on ResNet for microseismic events. Taking the great advantages of deep networks in classification and regression tasks, it can realize the great potential of fast and accurate inversion of microseismic moment tensors after the network trained. This ResNet-based moment tensor prediction technology, whose input is raw recordings, does not require the extraction of data features in advance. First, we tested the network using synthetic data and performed a quantitative assessment of the errors. The results demonstrate that the network exhibits high accuracy and efficiency during the prediction phase. Next, we tested the network using real microseismic data and compared the results with those from traditional inversion methods. The error in the results was relatively small compared to traditional methods. However, the network operates more efficiently without requiring manual intervention, making it highly valuable for near-real-time monitoring applications.
提出了一种基于ResNet的矩张量回归预测技术。利用深度网络在分类和回归任务上的巨大优势,可以实现网络训练后快速准确反演微震矩张量的巨大潜力。这种基于resnet的矩张量预测技术,其输入为原始记录,不需要提前提取数据特征。首先,我们使用合成数据测试了网络,并对误差进行了定量评估。结果表明,该网络在预测阶段具有较高的精度和效率。接下来,我们使用真实微震数据对网络进行了测试,并将结果与传统反演方法进行了比较。与传统方法相比,结果误差相对较小。然而,该网络在不需要人工干预的情况下更有效地运行,使其对近实时监控应用具有很高的价值。
{"title":"Microseismic moment tensor inversion based on ResNet model","authors":"Jiaqi Yan ,&nbsp;Li Ma ,&nbsp;Tianqi Jiang ,&nbsp;Jing Zheng ,&nbsp;Dewei Li ,&nbsp;Xingzhi Teng","doi":"10.1016/j.aiig.2025.100107","DOIUrl":"10.1016/j.aiig.2025.100107","url":null,"abstract":"<div><div>This paper proposed a moment tensor regression prediction technology based on ResNet for microseismic events. Taking the great advantages of deep networks in classification and regression tasks, it can realize the great potential of fast and accurate inversion of microseismic moment tensors after the network trained. This ResNet-based moment tensor prediction technology, whose input is raw recordings, does not require the extraction of data features in advance. First, we tested the network using synthetic data and performed a quantitative assessment of the errors. The results demonstrate that the network exhibits high accuracy and efficiency during the prediction phase. Next, we tested the network using real microseismic data and compared the results with those from traditional inversion methods. The error in the results was relatively small compared to traditional methods. However, the network operates more efficiently without requiring manual intervention, making it highly valuable for near-real-time monitoring applications.</div></div>","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"6 1","pages":"Article 100107"},"PeriodicalIF":0.0,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143551571","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Innovative cone resistance and sleeve friction prediction from geophysics based on a coupled geo-statistical and machine learning process 基于地球统计和机器学习耦合过程的创新地球物理锥体阻力和滑套摩擦预测
Pub Date : 2025-02-26 DOI: 10.1016/j.aiig.2025.100110
A. Bolève, R. Eddies, M. Staring, Y. Benboudiaf, H. Pournaki, M. Nepveaux
Geotechnical parameters derived from an intrusive cone penetration test (CPT) are used to asses mechanical properties to inform the design phase of infrastructure projects. However, local, in situ 1D measurements can fail to capture 3D subsurface variations, which could mean less than optimal design decisions for foundation engineering. By coupling the localised measurements from CPTs with more global 3D measurements derived from geophysical methods, a higher fidelity 3D overview of the subsurface can be obtained. Machine Learning (ML) may offer an effective means to capture all types of geophysical information associated with CPT data at a site scale to build a 2D or 3D ground model. In this paper, we present an ML approach to build a 3D ground model of cone resistance and sleeve friction by combining several CPT measurements with Multichannel Analysis of Surface Waves (MASW) and Electrical Resistivity Tomography (ERT) data on a land site characterisation project in the United Arab Emirates (UAE). To avoid a potential overfitting problem inherent to the use of machine learning and a lack of data at certain locations, we explore the possibility of using a prior Geo-Statistical (GS) approach that attempts to constrain the overfitting process by “artificially” increasing the amount of input data. A sensitivity study is also performed on input features used to train the ML algorithm to better define the optimal combination of input features for the prediction. Our results showed that ERT data were not useful in capturing 3D variations of geotechnical properties compared to Vs due to the geographical location of the site (200 m east from the Oman Gulf) and the possible effect of saline water intrusion. Additionally, we demonstrate that the use of a prior GS phase could be a promising and interesting means to make the prediction of ground properties more robust, especially for this specific case study described in this paper. Looking ahead, better representation of the subsurface can lead to a number of benefits for stakeholders involved in developing assets. Better ground/geotechnical models mean better site calibration of design methods and fewer design assumptions for reliability-based design, creating an opportunity for value engineering in the form of lighter construction without compromising safety, shorter construction timelines, and reduced resource requirements.
从侵入式锥体穿透测试(CPT)中获得的岩土参数用于评估机械性能,为基础设施项目的设计阶段提供信息。然而,局部的原位1D测量可能无法捕获三维地下变化,这可能意味着基础工程的最佳设计决策不足。通过将来自cpt的局部测量与来自地球物理方法的更多全局3D测量相结合,可以获得更高保真度的地下3D概况。机器学习(ML)可以提供一种有效的方法,在现场尺度上捕获与CPT数据相关的所有类型的地球物理信息,以建立2D或3D地面模型。在本文中,我们提出了一种ML方法,通过将多个CPT测量结果与多通道表面波分析(MASW)和电阻率层析成像(ERT)数据相结合,在阿拉伯联合酋长国(UAE)的一个地块特征描述项目中建立锥体阻力和套筒摩擦的三维地面模型。为了避免使用机器学习和某些位置缺乏数据所固有的潜在过拟合问题,我们探索了使用先前地理统计学(GS)方法的可能性,该方法试图通过“人为”增加输入数据量来限制过拟合过程。对用于训练ML算法的输入特征进行敏感性研究,以更好地定义用于预测的输入特征的最佳组合。我们的研究结果表明,由于场地的地理位置(距阿曼湾以东200米)和盐水入侵的可能影响,与v相比,ERT数据在捕获岩土力学特性的三维变化方面并不有用。此外,我们证明,使用先前的GS相位可能是一种有前途和有趣的方法,可以使地面性质的预测更加可靠,特别是对于本文中描述的具体案例研究。展望未来,更好地代表地下资源可以为参与开发资产的利益相关者带来许多好处。更好的地面/岩土模型意味着更好的现场校准设计方法和更少的基于可靠性设计的设计假设,以更轻的结构形式创造价值工程的机会,而不影响安全,更短的施工时间,减少资源需求。
{"title":"Innovative cone resistance and sleeve friction prediction from geophysics based on a coupled geo-statistical and machine learning process","authors":"A. Bolève,&nbsp;R. Eddies,&nbsp;M. Staring,&nbsp;Y. Benboudiaf,&nbsp;H. Pournaki,&nbsp;M. Nepveaux","doi":"10.1016/j.aiig.2025.100110","DOIUrl":"10.1016/j.aiig.2025.100110","url":null,"abstract":"<div><div>Geotechnical parameters derived from an intrusive cone penetration test (CPT) are used to asses mechanical properties to inform the design phase of infrastructure projects. However, local, in situ 1D measurements can fail to capture 3D subsurface variations, which could mean less than optimal design decisions for foundation engineering. By coupling the localised measurements from CPTs with more global 3D measurements derived from geophysical methods, a higher fidelity 3D overview of the subsurface can be obtained. Machine Learning (ML) may offer an effective means to capture all types of geophysical information associated with CPT data at a site scale to build a 2D or 3D ground model. In this paper, we present an ML approach to build a 3D ground model of cone resistance and sleeve friction by combining several CPT measurements with Multichannel Analysis of Surface Waves (MASW) and Electrical Resistivity Tomography (ERT) data on a land site characterisation project in the United Arab Emirates (UAE). To avoid a potential overfitting problem inherent to the use of machine learning and a lack of data at certain locations, we explore the possibility of using a prior Geo-Statistical (GS) approach that attempts to constrain the overfitting process by “artificially” increasing the amount of input data. A sensitivity study is also performed on input features used to train the ML algorithm to better define the optimal combination of input features for the prediction. Our results showed that ERT data were not useful in capturing 3D variations of geotechnical properties compared to V<sub>s</sub> due to the geographical location of the site (200 m east from the Oman Gulf) and the possible effect of saline water intrusion. Additionally, we demonstrate that the use of a prior GS phase could be a promising and interesting means to make the prediction of ground properties more robust, especially for this specific case study described in this paper. Looking ahead, better representation of the subsurface can lead to a number of benefits for stakeholders involved in developing assets. Better ground/geotechnical models mean better site calibration of design methods and fewer design assumptions for reliability-based design, creating an opportunity for value engineering in the form of lighter construction without compromising safety, shorter construction timelines, and reduced resource requirements.</div></div>","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"6 1","pages":"Article 100110"},"PeriodicalIF":0.0,"publicationDate":"2025-02-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143526930","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Robust low frequency seismic bandwidth extension with a U-net and synthetic training data 鲁棒低频地震带宽扩展与U-net和综合训练数据
Pub Date : 2025-02-25 DOI: 10.1016/j.aiig.2025.100109
P. Zwartjes, J. Yoo
This work focuses on enhancing low frequency seismic data using a convolutional neural network trained on synthetic data. Traditional seismic data often lack both high and low frequencies, which are essential for detailed geological interpretation and various geophysical applications. Low frequency data is particularly valuable for reducing wavelet sidelobes and improving full waveform inversion (FWI). Conventional methods for bandwidth extension include seismic deconvolution and sparse inversion, which have limitations in recovering low frequencies. The study explores the potential of the U-net, which has been successful in other geophysical applications such as noise attenuation and seismic resolution enhancement. The novelty in our approach is that we do not rely on computationally expensive finite difference modelling to create training data. Instead, our synthetic training data is created from individual randomly perturbed events with variations in bandwidth, making it more adaptable to different data sets compared to previous deep learning methods. The method was tested on both synthetic and real seismic data, demonstrating effective low frequency reconstruction and sidelobe reduction. With a synthetic full waveform inversion to recover a velocity model and a seismic amplitude inversion to estimate acoustic impedance we demonstrate the validity and benefit of the proposed method. Overall, the study presents a robust approach to seismic bandwidth extension using deep learning, emphasizing the importance of diverse and well-designed but computationally inexpensive synthetic training data.
这项工作的重点是使用在合成数据上训练的卷积神经网络来增强低频地震数据。传统的地震数据往往缺乏高频和低频,这是详细的地质解释和各种地球物理应用所必需的。低频数据对于减少小波副瓣和改善全波形反演(FWI)特别有价值。传统的带宽扩展方法包括地震反褶积和稀疏反演,但在恢复低频方面存在局限性。该研究探索了U-net的潜力,U-net已经在其他地球物理应用中取得了成功,例如噪声衰减和地震分辨率提高。我们的方法的新颖之处在于,我们不依赖于计算昂贵的有限差分建模来创建训练数据。相反,我们的合成训练数据是由带宽变化的单个随机扰动事件创建的,与以前的深度学习方法相比,它更能适应不同的数据集。在合成地震和真实地震数据上进行了测试,结果表明该方法具有有效的低频重建和旁瓣抑制效果。通过合成全波形反演恢复速度模型和地震振幅反演估计声阻抗,验证了该方法的有效性和优越性。总的来说,该研究提出了一种利用深度学习扩展地震带宽的鲁棒方法,强调了多样化和精心设计但计算成本低廉的合成训练数据的重要性。
{"title":"Robust low frequency seismic bandwidth extension with a U-net and synthetic training data","authors":"P. Zwartjes,&nbsp;J. Yoo","doi":"10.1016/j.aiig.2025.100109","DOIUrl":"10.1016/j.aiig.2025.100109","url":null,"abstract":"<div><div>This work focuses on enhancing low frequency seismic data using a convolutional neural network trained on synthetic data. Traditional seismic data often lack both high and low frequencies, which are essential for detailed geological interpretation and various geophysical applications. Low frequency data is particularly valuable for reducing wavelet sidelobes and improving full waveform inversion (FWI). Conventional methods for bandwidth extension include seismic deconvolution and sparse inversion, which have limitations in recovering low frequencies. The study explores the potential of the U-net, which has been successful in other geophysical applications such as noise attenuation and seismic resolution enhancement. The novelty in our approach is that we do not rely on computationally expensive finite difference modelling to create training data. Instead, our synthetic training data is created from individual randomly perturbed events with variations in bandwidth, making it more adaptable to different data sets compared to previous deep learning methods. The method was tested on both synthetic and real seismic data, demonstrating effective low frequency reconstruction and sidelobe reduction. With a synthetic full waveform inversion to recover a velocity model and a seismic amplitude inversion to estimate acoustic impedance we demonstrate the validity and benefit of the proposed method. Overall, the study presents a robust approach to seismic bandwidth extension using deep learning, emphasizing the importance of diverse and well-designed but computationally inexpensive synthetic training data.</div></div>","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"6 1","pages":"Article 100109"},"PeriodicalIF":0.0,"publicationDate":"2025-02-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143512149","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Applying deep learning to teleseismic phase detection and picking: PcP and PKiKP cases 将深度学习应用于远震相位检测和拾取:PcP和PKiKP案例
Pub Date : 2025-02-21 DOI: 10.1016/j.aiig.2025.100108
Congcong Yuan , Jie Zhang
The availability of a tremendous amount of seismic data demands seismological researchers to analyze seismic phases efficiently. Recently, deep learning algorithms exhibit a powerful capability of detecting and picking on P- and S-wave phases. However, it remains a challenge to effeciently process enormous teleseismic phases, which are crucial to probe Earth's interior structures and their dynamics. In this study, we propose a scheme to detect and pick teleseismic phases, such as seismic phase that reflects off the core-mantle boundary (i.e., PcP) and that reflects off the inner-core boundary (i.e., PKiKP), from a seismic dataset in Japan. The scheme consists of three steps: 1) latent phase traces are truncated from the whole seismogram with theoretical arrival times; 2) latent phases are recognized and evaluated by convolutional neural network (CNN) models; 3) arrivals of good or fair phase are picked with another CNN models. The testing detection result on 7386 seismograms shows that the scheme recognizes 92.15% and 94.13% of PcP and PKiKP phases. The testing picking result has a mean absolute error of 0.0742 s and 0.0636 s for the PcP and PKiKP phases, respectively. These seismograms were processed in just 5 min for phase detection and picking, demonstrating the efficiency of the proposed scheme in automatic teleseismic phase analysis.
大量地震资料的可用性要求地震学研究人员高效地分析地震相。最近,深度学习算法在探测和挑选P波和s波相位方面表现出强大的能力。然而,有效地处理巨大的远震相位仍然是一个挑战,而远震相位对于探测地球内部结构及其动力学至关重要。在这项研究中,我们提出了一种从日本地震数据集中检测和提取远震相位的方案,例如从核幔边界反射的地震相位(即PcP)和从内核边界反射的地震相位(即PKiKP)。该方案包括三个步骤:1)从具有理论到达时间的整个地震记录中截断潜相迹;2)利用卷积神经网络(CNN)模型对潜在相位进行识别和评估;3)用另一个CNN模型选择好的或一般的相位到达。7386张地震图的测试检测结果表明,该方案对PcP相位和PKiKP相位的识别率分别为92.15%和94.13%。PcP期和PKiKP期的平均绝对误差分别为0.0742 s和0.0636 s。这些地震记录在5分钟内进行了相位检测和拾取,证明了该方案在自动远震相位分析中的有效性。
{"title":"Applying deep learning to teleseismic phase detection and picking: PcP and PKiKP cases","authors":"Congcong Yuan ,&nbsp;Jie Zhang","doi":"10.1016/j.aiig.2025.100108","DOIUrl":"10.1016/j.aiig.2025.100108","url":null,"abstract":"<div><div>The availability of a tremendous amount of seismic data demands seismological researchers to analyze seismic phases efficiently. Recently, deep learning algorithms exhibit a powerful capability of detecting and picking on P- and S-wave phases. However, it remains a challenge to effeciently process enormous teleseismic phases, which are crucial to probe Earth's interior structures and their dynamics. In this study, we propose a scheme to detect and pick teleseismic phases, such as seismic phase that reflects off the core-mantle boundary (i.e., PcP) and that reflects off the inner-core boundary (i.e., PKiKP), from a seismic dataset in Japan. The scheme consists of three steps: 1) latent phase traces are truncated from the whole seismogram with theoretical arrival times; 2) latent phases are recognized and evaluated by convolutional neural network (CNN) models; 3) arrivals of good or fair phase are picked with another CNN models. The testing detection result on 7386 seismograms shows that the scheme recognizes 92.15% and 94.13% of PcP and PKiKP phases. The testing picking result has a mean absolute error of 0.0742 s and 0.0636 s for the PcP and PKiKP phases, respectively. These seismograms were processed in just 5 min for phase detection and picking, demonstrating the efficiency of the proposed scheme in automatic teleseismic phase analysis.</div></div>","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"6 1","pages":"Article 100108"},"PeriodicalIF":0.0,"publicationDate":"2025-02-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143488648","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimizing zero-shot text-based segmentation of remote sensing imagery using SAM and Grounding DINO 基于SAM和ground DINO的遥感影像零射击文本分割优化
Pub Date : 2025-02-13 DOI: 10.1016/j.aiig.2025.100105
Mohanad Diab , Polychronis Kolokoussis , Maria Antonia Brovelli
The use of AI technologies in remote sensing (RS) tasks has been the focus of many individuals in both the professional and academic domains. Having more accessible interfaces and tools that allow people of little or no experience to intuitively interact with RS data of multiple formats is a potential provided by this integration. However, the use of AI and AI agents to help automate RS-related tasks is still in its infancy stage, with some frameworks and interfaces built on top of well-known vision language models (VLM) such as GPT-4, segment anything model (SAM), and grounding DINO. These tools do promise and draw guidelines on the potentials and limitations of existing solutions concerning the use of said models. In this work, the state of the art AI foundation models (FM) are reviewed and used in a multi-modal manner to ingest RS imagery input and perform zero-shot object detection using natural language. The natural language input is then used to define the classes or labels the model should look for, then, both inputs are fed to the pipeline. The pipeline presented in this work makes up for the shortcomings of the general knowledge FMs by stacking pre-processing and post-processing applications on top of the FMs; these applications include tiling to produce uniform patches of the original image for faster detection, outlier rejection of redundant bounding boxes using statistical and machine learning methods. The pipeline was tested with UAV, aerial and satellite images taken over multiple areas. The accuracy for the semantic segmentation showed improvement from the original 64% to approximately 80%–99% by utilizing the pipeline and techniques proposed in this work. GitHub Repository: MohanadDiab/LangRS.
人工智能技术在遥感(RS)任务中的应用一直是专业和学术领域许多人关注的焦点。这种集成提供了更易于访问的接口和工具,使很少或没有经验的人能够直观地与多种格式的RS数据进行交互。然而,使用人工智能和人工智能代理来帮助自动化rs相关任务仍处于起步阶段,一些框架和接口建立在知名的视觉语言模型(VLM)之上,如GPT-4、分段任何模型(SAM)和接地DINO。这些工具确实承诺并绘制了关于使用上述模型的现有解决方案的潜力和局限性的指导方针。在这项工作中,对最先进的人工智能基础模型(FM)进行了回顾,并以多模式方式使用它们来摄取RS图像输入并使用自然语言执行零射击目标检测。然后使用自然语言输入来定义模型应该查找的类或标签,然后将两个输入都提供给管道。本文提出的流水线通过在流水线上叠加预处理和后处理应用,弥补了一般知识模型的不足;这些应用包括平铺以产生原始图像的均匀补丁,以便更快地检测,使用统计和机器学习方法拒绝冗余边界框的异常值。该管道用无人机进行了测试,在多个地区拍摄了空中和卫星图像。利用本文提出的管道和技术,将语义分割的准确率从原来的64%提高到80%-99%左右。GitHub Repository: mohanadiab / langs。
{"title":"Optimizing zero-shot text-based segmentation of remote sensing imagery using SAM and Grounding DINO","authors":"Mohanad Diab ,&nbsp;Polychronis Kolokoussis ,&nbsp;Maria Antonia Brovelli","doi":"10.1016/j.aiig.2025.100105","DOIUrl":"10.1016/j.aiig.2025.100105","url":null,"abstract":"<div><div>The use of AI technologies in remote sensing (RS) tasks has been the focus of many individuals in both the professional and academic domains. Having more accessible interfaces and tools that allow people of little or no experience to intuitively interact with RS data of multiple formats is a potential provided by this integration. However, the use of AI and AI agents to help automate RS-related tasks is still in its infancy stage, with some frameworks and interfaces built on top of well-known vision language models (VLM) such as GPT-4, segment anything model (SAM), and grounding DINO. These tools do promise and draw guidelines on the potentials and limitations of existing solutions concerning the use of said models. In this work, the state of the art AI foundation models (FM) are reviewed and used in a multi-modal manner to ingest RS imagery input and perform zero-shot object detection using natural language. The natural language input is then used to define the classes or labels the model should look for, then, both inputs are fed to the pipeline. The pipeline presented in this work makes up for the shortcomings of the general knowledge FMs by stacking pre-processing and post-processing applications on top of the FMs; these applications include tiling to produce uniform patches of the original image for faster detection, outlier rejection of redundant bounding boxes using statistical and machine learning methods. The pipeline was tested with UAV, aerial and satellite images taken over multiple areas. The accuracy for the semantic segmentation showed improvement from the original 64% to approximately 80%–99% by utilizing the pipeline and techniques proposed in this work. <strong>GitHub Repository:</strong> <span><span>MohanadDiab/LangRS</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"6 1","pages":"Article 100105"},"PeriodicalIF":0.0,"publicationDate":"2025-02-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143429918","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Loosening rocks detection at Draa Sfar deep underground mine in Morocco using infrared thermal imaging and image segmentation models 基于红外热成像和图像分割模型的摩洛哥Draa Sfar深部地下矿松动岩探测
Pub Date : 2025-01-27 DOI: 10.1016/j.aiig.2025.100106
Kaoutar Clero , Said Ed-Diny , Mohammed Achalhi , Mouhamed Cherkaoui , Imad El Harraki , Sanaa El Fkihi , Intissar Benzakour , Tarik Soror , Said Rziki , Hamd Ait Abdelali , Hicham Tagemouati , François Bourzeix
Rockfalls are among the frequent hazards in underground mines worldwide, requiring effective methods for detecting unstable rock blocks to ensure miners' and equipment's safety. This study proposes a novel approach for identifying potential rockfall zones using infrared thermal imaging and image segmentation techniques. Infrared images of rock blocks were captured at the Draa Sfar deep underground mine in Morocco using the FLUKE TI401 PRO thermal camera. Two segmentation methods were applied to locate the potential unstable areas: the classical thresholding and the K-means clustering model. The results show that while thresholding allows a binary distinction between stable and unstable areas, K-means clustering is more accurate, especially when using multiple clusters to show different risk levels. The close match between the clustering masks of unstable blocks and their corresponding visible light images further validated this. The findings confirm that thermal image segmentation can serve as an alternative method for predicting rockfalls and monitoring geotechnical issues in underground mines. Underground operators worldwide can apply this approach to monitor rock mass stability. However, further research is recommended to enhance these results, particularly through deep learning-based segmentation and object detection models.
岩崩是世界范围内地下矿山频发的灾害之一,为保证矿工和设备的安全,需要有效的检测不稳定岩块的方法。本研究提出了一种利用红外热成像和图像分割技术识别潜在岩崩带的新方法。利用FLUKE TI401 PRO热像仪在摩洛哥Draa Sfar深部地下矿山拍摄了岩石块的红外图像。采用经典阈值分割和k均值聚类两种分割方法定位潜在的不稳定区域。结果表明,虽然阈值允许对稳定和不稳定区域进行二元区分,但K-means聚类更准确,特别是当使用多个聚类来显示不同的风险水平时。不稳定块的聚类掩模与其对应的可见光图像的紧密匹配进一步验证了这一点。研究结果证实,热图像分割可以作为地下矿山岩崩预测和岩土工程问题监测的替代方法。世界各地的地下运营商都可以应用这种方法来监测岩体的稳定性。然而,建议进一步研究以增强这些结果,特别是通过基于深度学习的分割和目标检测模型。
{"title":"Loosening rocks detection at Draa Sfar deep underground mine in Morocco using infrared thermal imaging and image segmentation models","authors":"Kaoutar Clero ,&nbsp;Said Ed-Diny ,&nbsp;Mohammed Achalhi ,&nbsp;Mouhamed Cherkaoui ,&nbsp;Imad El Harraki ,&nbsp;Sanaa El Fkihi ,&nbsp;Intissar Benzakour ,&nbsp;Tarik Soror ,&nbsp;Said Rziki ,&nbsp;Hamd Ait Abdelali ,&nbsp;Hicham Tagemouati ,&nbsp;François Bourzeix","doi":"10.1016/j.aiig.2025.100106","DOIUrl":"10.1016/j.aiig.2025.100106","url":null,"abstract":"<div><div>Rockfalls are among the frequent hazards in underground mines worldwide, requiring effective methods for detecting unstable rock blocks to ensure miners' and equipment's safety. This study proposes a novel approach for identifying potential rockfall zones using infrared thermal imaging and image segmentation techniques. Infrared images of rock blocks were captured at the Draa Sfar deep underground mine in Morocco using the FLUKE TI401 PRO thermal camera. Two segmentation methods were applied to locate the potential unstable areas: the classical thresholding and the K-means clustering model. The results show that while thresholding allows a binary distinction between stable and unstable areas, K-means clustering is more accurate, especially when using multiple clusters to show different risk levels. The close match between the clustering masks of unstable blocks and their corresponding visible light images further validated this. The findings confirm that thermal image segmentation can serve as an alternative method for predicting rockfalls and monitoring geotechnical issues in underground mines. Underground operators worldwide can apply this approach to monitor rock mass stability. However, further research is recommended to enhance these results, particularly through deep learning-based segmentation and object detection models.</div></div>","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"6 1","pages":"Article 100106"},"PeriodicalIF":0.0,"publicationDate":"2025-01-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143143632","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Convolutional sparse coding network for sparse seismic time-frequency representation 用于稀疏地震时频表示的卷积稀疏编码网络
Pub Date : 2024-11-04 DOI: 10.1016/j.aiig.2024.100104
Qiansheng Wei , Zishuai Li , Haonan Feng , Yueying Jiang , Yang Yang , Zhiguo Wang
Seismic time-frequency (TF) transforms are essential tools in reservoir interpretation and signal processing, particularly for characterizing frequency variations in non-stationary seismic data. Recently, sparse TF transforms, which leverage sparse coding (SC), have gained significant attention in the geosciences due to their ability to achieve high TF resolution. However, the iterative approaches typically employed in sparse TF transforms are computationally intensive, making them impractical for real seismic data analysis. To address this issue, we propose an interpretable convolutional sparse coding (CSC) network to achieve high TF resolution. The proposed model is generated based on the traditional short-time Fourier transform (STFT) transform and a modified UNet, named ULISTANet. In this design, we replace the conventional convolutional layers of the UNet with learnable iterative shrinkage thresholding algorithm (LISTA) blocks, a specialized form of CSC. The LISTA block, which evolves from the traditional iterative shrinkage thresholding algorithm (ISTA), is optimized for extracting sparse features more effectively. Furthermore, we create a synthetic dataset featuring complex frequency-modulated signals to train ULISTANet. Finally, the proposed method's performance is subsequently validated using both synthetic and field data, demonstrating its potential for enhanced seismic data analysis.
地震时频(TF)变换是储层解释和信号处理的重要工具,特别是用于描述非稳态地震数据中的频率变化。最近,利用稀疏编码(SC)的稀疏时频变换因其实现高时频分辨率的能力而在地球科学领域备受关注。然而,稀疏 TF 变换通常采用的迭代方法需要大量计算,因此在实际地震数据分析中并不实用。为解决这一问题,我们提出了一种可解释卷积稀疏编码(CSC)网络,以实现高 TF 分辨率。我们提出的模型是基于传统的短时傅立叶变换(STFT)和改进的 UNet(名为 ULISTANet)生成的。在这一设计中,我们用可学习的迭代收缩阈值算法(LISTA)块(一种专门的 CSC 形式)取代了 UNet 的传统卷积层。LISTA 块由传统的迭代收缩阈值算法(ISTA)演化而来,经过优化,能更有效地提取稀疏特征。此外,我们还创建了一个以复杂频率调制信号为特征的合成数据集来训练 ULISTANet。最后,我们利用合成数据和野外数据对所提出方法的性能进行了验证,证明了该方法在增强地震数据分析方面的潜力。
{"title":"Convolutional sparse coding network for sparse seismic time-frequency representation","authors":"Qiansheng Wei ,&nbsp;Zishuai Li ,&nbsp;Haonan Feng ,&nbsp;Yueying Jiang ,&nbsp;Yang Yang ,&nbsp;Zhiguo Wang","doi":"10.1016/j.aiig.2024.100104","DOIUrl":"10.1016/j.aiig.2024.100104","url":null,"abstract":"<div><div>Seismic time-frequency (TF) transforms are essential tools in reservoir interpretation and signal processing, particularly for characterizing frequency variations in non-stationary seismic data. Recently, sparse TF transforms, which leverage sparse coding (SC), have gained significant attention in the geosciences due to their ability to achieve high TF resolution. However, the iterative approaches typically employed in sparse TF transforms are computationally intensive, making them impractical for real seismic data analysis. To address this issue, we propose an interpretable convolutional sparse coding (CSC) network to achieve high TF resolution. The proposed model is generated based on the traditional short-time Fourier transform (STFT) transform and a modified UNet, named ULISTANet. In this design, we replace the conventional convolutional layers of the UNet with learnable iterative shrinkage thresholding algorithm (LISTA) blocks, a specialized form of CSC. The LISTA block, which evolves from the traditional iterative shrinkage thresholding algorithm (ISTA), is optimized for extracting sparse features more effectively. Furthermore, we create a synthetic dataset featuring complex frequency-modulated signals to train ULISTANet. Finally, the proposed method's performance is subsequently validated using both synthetic and field data, demonstrating its potential for enhanced seismic data analysis.</div></div>","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"6 1","pages":"Article 100104"},"PeriodicalIF":0.0,"publicationDate":"2024-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142661128","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Research on the prediction method for fluvial-phase sandbody connectivity based on big data analysis--a case study of Bohai a oilfield 基于大数据分析的流相砂体连通性预测方法研究--以渤海某油田为例
Pub Date : 2024-10-16 DOI: 10.1016/j.aiig.2024.100095
Cai Li, Fei Ma, Yuxiu Wang, Delong Zhang
The connectivity of sandbodies is a key constraint to the exploration effectiveness of Bohai A Oilfield. Conventional connectivity studies often use methods such as seismic attribute fusion, while the development of contiguous composite sandbodies in this area makes it challenging to characterize connectivity changes with conventional seismic attributes. Aiming at the above problem in the Bohai A Oilfield, this study proposes a big data analysis method based on the Deep Forest algorithm to predict the sandbody connectivity. Firstly, by compiling the abundant exploration and development sandbodies data in the study area, typical sandbodies with reliable connectivity were selected. Then, sensitive seismic attribute were extracted to obtain training samples. Finally, based on the Deep Forest algorithm, mapping model between attribute combinations and sandbody connectivity was established through machine learning. This method achieves the first quantitative determination of the connectivity for continuous composite sandbodies in the Bohai Oilfield. Compared with conventional connectivity discrimination methods such as high-resolution processing and seismic attribute analysis, this method can combine the sandbody characteristics of the study area in the process of machine learning, and jointly judge connectivity by combining multiple seismic attributes. The study results show that this method has high accuracy and timeliness in predicting connectivity for continuous composite sandbodies. Applied to the Bohai A Oilfield, it successfully identified multiple sandbody connectivity relationships and provided strong support for the subsequent exploration potential assessment and well placement optimization. This method also provides a new idea and method for studying sandbody connectivity under similar complex geological conditions.
砂体连通性是渤海 A 油田勘探有效性的关键制约因素。传统的连通性研究通常采用地震属性融合等方法,而该地区连片复合砂体的发育使得用常规地震属性表征连通性变化具有挑战性。针对渤海A油田的上述问题,本研究提出了一种基于深林算法的大数据分析方法来预测砂体连通性。首先,通过整理研究区丰富的勘探开发沙体数据,筛选出具有可靠连通性的典型沙体。然后,提取敏感地震属性,获得训练样本。最后,基于深林算法,通过机器学习建立属性组合与沙体连通性之间的映射模型。该方法首次实现了对渤海油田连续复合砂体连通性的定量判定。与传统的高分辨率处理、地震属性分析等连通性判别方法相比,该方法在机器学习过程中能够结合研究区的沙体特征,综合多种地震属性共同判断连通性。研究结果表明,该方法在预测连续复合砂体连通性方面具有较高的准确性和时效性。应用于渤海A油田,成功识别了多个砂体的连通性关系,为后续的勘探潜力评估和井位优化提供了有力支持。该方法也为研究类似复杂地质条件下的砂体连通性提供了新的思路和方法。
{"title":"Research on the prediction method for fluvial-phase sandbody connectivity based on big data analysis--a case study of Bohai a oilfield","authors":"Cai Li,&nbsp;Fei Ma,&nbsp;Yuxiu Wang,&nbsp;Delong Zhang","doi":"10.1016/j.aiig.2024.100095","DOIUrl":"10.1016/j.aiig.2024.100095","url":null,"abstract":"<div><div>The connectivity of sandbodies is a key constraint to the exploration effectiveness of Bohai A Oilfield. Conventional connectivity studies often use methods such as seismic attribute fusion, while the development of contiguous composite sandbodies in this area makes it challenging to characterize connectivity changes with conventional seismic attributes. Aiming at the above problem in the Bohai A Oilfield, this study proposes a big data analysis method based on the Deep Forest algorithm to predict the sandbody connectivity. Firstly, by compiling the abundant exploration and development sandbodies data in the study area, typical sandbodies with reliable connectivity were selected. Then, sensitive seismic attribute were extracted to obtain training samples. Finally, based on the Deep Forest algorithm, mapping model between attribute combinations and sandbody connectivity was established through machine learning. This method achieves the first quantitative determination of the connectivity for continuous composite sandbodies in the Bohai Oilfield. Compared with conventional connectivity discrimination methods such as high-resolution processing and seismic attribute analysis, this method can combine the sandbody characteristics of the study area in the process of machine learning, and jointly judge connectivity by combining multiple seismic attributes. The study results show that this method has high accuracy and timeliness in predicting connectivity for continuous composite sandbodies. Applied to the Bohai A Oilfield, it successfully identified multiple sandbody connectivity relationships and provided strong support for the subsequent exploration potential assessment and well placement optimization. This method also provides a new idea and method for studying sandbody connectivity under similar complex geological conditions.</div></div>","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"5 ","pages":"Article 100095"},"PeriodicalIF":0.0,"publicationDate":"2024-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142532292","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Pore size classification and prediction based on distribution of reservoir fluid volumes utilizing well logs and deep learning algorithm in a complex lithology 在复杂岩性中利用测井记录和深度学习算法,基于储层流体体积分布进行孔隙尺寸分类和预测
Pub Date : 2024-10-12 DOI: 10.1016/j.aiig.2024.100094
Hassan Bagheri , Reza Mohebian , Ali Moradzadeh , Behnia Azizzadeh Mehmandost Olya
Pore size analysis plays a pivotal role in unraveling reservoir behavior and its intricate relationship with confined fluids. Traditional methods for predicting pore size distribution (PSD), relying on drilling cores or thin sections, face limitations associated with depth specificity. In this study, we introduce an innovative framework that leverages nuclear magnetic resonance (NMR) log data, encompassing clay-bound water (CBW), bound volume irreducible (BVI), and free fluid volume (FFV), to determine three PSDs (micropores, mesopores, and macropores). Moreover, we establish a robust pore size classification (PSC) system utilizing ternary plots, derived from the PSDs.
Within the three studied wells, NMR log data is exclusive to one well (well-A), while conventional well logs are accessible for all three wells (well-A, well-B, and well-C). This distinction enables PSD predictions for the remaining two wells (B and C). To prognosticate NMR outputs (CBW, BVI, FFV) for these wells, a two-step deep learning (DL) algorithm is implemented. Initially, three feature selection algorithms (f-classif, f-regression, and mutual-info-regression) identify the conventional well logs most correlated to NMR outputs in well-A. The three feature selection algorithms utilize statistical computations. These algorithms are utilized to systematically identify and optimize pertinent input features, thereby augmenting model interpretability and predictive efficacy within intricate data-driven endeavors. So, all three feature selection algorithms introduced the number of 4 logs as the most optimal number of inputs to the DL algorithm with different combinations of logs for each of the three desired outputs. Subsequently, the CUDA Deep Neural Network Long Short-Term Memory algorithm(CUDNNLSTM), belonging to the category of DL algorithms and harnessing the computational power of GPUs, is employed for the prediction of CBW, BVI, and FFV logs. This prediction leverages the optimal logs identified in the preceding step. Estimation of NMR outputs was done first in well-A (80% of data as training and 20% as testing). The correlation coefficient (CC) between the actual and estimated data for the three outputs CBW, BVI and FFV are 95%, 94%, and 97%, respectively, as well as root mean square error (RMSE) was obtained 0.0081, 0.098, and 0.0089, respectively. To assess the effectiveness of the proposed algorithm, we compared it with two traditional methods for log estimation: multiple regression and multi-resolution graph-based clustering methods. The results demonstrate the superior accuracy of our algorithm in comparison to these conventional approaches. This DL-driven approach facilitates PSD prediction grounded in fluid saturation for wells B and C.
Ternary plots are then employed for PSCs. Seven distinct PSCs within well-A employing actual NMR logs (CBW, BVI, FFV), in conjunction with an equivalent count within wells B and C utilizing three predicted
孔隙度分析在揭示储层行为及其与封闭流体的复杂关系方面起着举足轻重的作用。传统的孔径分布(PSD)预测方法依赖于钻井岩心或薄切片,面临着深度特异性的限制。在这项研究中,我们引入了一个创新框架,利用核磁共振(NMR)测井数据(包括粘土结合水(CBW)、不可还原结合体积(BVI)和自由流体体积(FFV))来确定三种孔径分布(微孔、中孔和大孔)。此外,我们还利用从 PSDs 得出的三元图建立了一个强大的孔径分类 (PSC) 系统。在所研究的三口井中,核磁共振测井数据仅适用于一口井(A 井),而常规测井数据则适用于所有三口井(A 井、B 井和 C 井)。这种区别使我们能够对其余两口井(B 井和 C 井)进行 PSD 预测。为了预测这些油井的 NMR 输出(CBW、BVI、FFV),采用了两步深度学习(DL)算法。首先,三种特征选择算法(f-classif、f-regression 和 mutual-info-regression)确定与 A 井 NMR 输出最相关的常规测井曲线。这三种特征选择算法利用统计计算。这些算法用于系统地识别和优化相关输入特征,从而在复杂的数据驱动工作中提高模型的可解释性和预测效力。因此,所有三种特征选择算法都将 4 个日志的数量作为 DL 算法的最佳输入数量,并为三种所需的输出分别引入不同的日志组合。随后,CUDA 深度神经网络长短期记忆算法(CUDNNLSTM)被用于预测 CBW、BVI 和 FFV 日志,该算法属于 DL 算法范畴,利用了 GPU 的计算能力。该预测利用了前一步中确定的最佳日志。首先在 A 井(80% 的数据作为训练数据,20% 作为测试数据)中对 NMR 输出进行估计。CBW、BVI 和 FFV 三项输出的实际数据与估计数据之间的相关系数(CC)分别为 95%、94% 和 97%,均方根误差(RMSE)分别为 0.0081、0.098 和 0.0089。为了评估所提出算法的有效性,我们将其与两种传统的对数估计方法进行了比较:多元回归和基于多分辨率图的聚类方法。结果表明,与这些传统方法相比,我们的算法具有更高的准确性。这种以 DL 为驱动的方法有助于根据 B 井和 C 井的流体饱和度预测 PSD。利用实际 NMR 测井(CBW、BVI、FFV)对 A 井中的七个不同 PSC 进行了和谐分类,同时利用三个预测测井对 B 井和 C 井中的等量 PSC 进行了和谐分类,从而确定了七个不同的孔径分类面 (PSCF)。由此产生的 PSCF 为生成精确、详细的储层三维模型提供了宝贵的见解。
{"title":"Pore size classification and prediction based on distribution of reservoir fluid volumes utilizing well logs and deep learning algorithm in a complex lithology","authors":"Hassan Bagheri ,&nbsp;Reza Mohebian ,&nbsp;Ali Moradzadeh ,&nbsp;Behnia Azizzadeh Mehmandost Olya","doi":"10.1016/j.aiig.2024.100094","DOIUrl":"10.1016/j.aiig.2024.100094","url":null,"abstract":"<div><div>Pore size analysis plays a pivotal role in unraveling reservoir behavior and its intricate relationship with confined fluids. Traditional methods for predicting pore size distribution (PSD), relying on drilling cores or thin sections, face limitations associated with depth specificity. In this study, we introduce an innovative framework that leverages nuclear magnetic resonance (NMR) log data, encompassing clay-bound water (CBW), bound volume irreducible (BVI), and free fluid volume (FFV), to determine three PSDs (micropores, mesopores, and macropores). Moreover, we establish a robust pore size classification (PSC) system utilizing ternary plots, derived from the PSDs.</div><div>Within the three studied wells, NMR log data is exclusive to one well (well-A), while conventional well logs are accessible for all three wells (well-A, well-B, and well-C). This distinction enables PSD predictions for the remaining two wells (B and C). To prognosticate NMR outputs (CBW, BVI, FFV) for these wells, a two-step deep learning (DL) algorithm is implemented. Initially, three feature selection algorithms (f-classif, f-regression, and mutual-info-regression) identify the conventional well logs most correlated to NMR outputs in well-A. The three feature selection algorithms utilize statistical computations. These algorithms are utilized to systematically identify and optimize pertinent input features, thereby augmenting model interpretability and predictive efficacy within intricate data-driven endeavors. So, all three feature selection algorithms introduced the number of 4 logs as the most optimal number of inputs to the DL algorithm with different combinations of logs for each of the three desired outputs. Subsequently, the CUDA Deep Neural Network Long Short-Term Memory algorithm(CUDNNLSTM), belonging to the category of DL algorithms and harnessing the computational power of GPUs, is employed for the prediction of CBW, BVI, and FFV logs. This prediction leverages the optimal logs identified in the preceding step. Estimation of NMR outputs was done first in well-A (80% of data as training and 20% as testing). The correlation coefficient (CC) between the actual and estimated data for the three outputs CBW, BVI and FFV are 95%, 94%, and 97%, respectively, as well as root mean square error (RMSE) was obtained 0.0081, 0.098, and 0.0089, respectively. To assess the effectiveness of the proposed algorithm, we compared it with two traditional methods for log estimation: multiple regression and multi-resolution graph-based clustering methods. The results demonstrate the superior accuracy of our algorithm in comparison to these conventional approaches. This DL-driven approach facilitates PSD prediction grounded in fluid saturation for wells B and C.</div><div>Ternary plots are then employed for PSCs. Seven distinct PSCs within well-A employing actual NMR logs (CBW, BVI, FFV), in conjunction with an equivalent count within wells B and C utilizing three predicted","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"5 ","pages":"Article 100094"},"PeriodicalIF":0.0,"publicationDate":"2024-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142437701","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Benchmarking data handling strategies for landslide susceptibility modeling using random forest workflows 利用随机森林工作流程为滑坡易发性建模的数据处理策略制定基准
Pub Date : 2024-10-05 DOI: 10.1016/j.aiig.2024.100093
Guruh Samodra , Ngadisih , Ferman Setia Nugroho
Machine learning (ML) algorithms are frequently used in landslide susceptibility modeling. Different data handling strategies may generate variations in landslide susceptibility modeling, even when using the same ML algorithm. This research aims to compare the combinations of inventory data handling, cross validation (CV), and hyperparameter tuning strategies to generate landslide susceptibility maps. The results are expected to provide a general strategy for landslide susceptibility modeling using ML techniques. The authors employed eight landslide inventory data handling scenarios to convert a landslide polygon into a landslide point, i.e., the landslide point is located on the toe (minimum height), on the scarp (maximum height), at the center of the landslide, randomly inside the polygon (1 point), randomly inside the polygon (3 points), randomly inside the polygon (5 points), randomly inside the polygon (10 points), and 15 m grid sampling. Random forest models using CV–nonspatial hyperparameter tuning, spatial CV–spatial hyperparameter tuning, and spatial CV–forward feature selection–no hyperparameter tuning were applied for each data handling strategy. The combination generated 24 random forest ML workflows, which are applied using a complete inventory of 743 landslides triggered by Tropical Cyclone Cempaka (2017) in Pacitan Regency, Indonesia, and 11 landslide controlling factors. The results show that grid sampling with spatial CV and spatial hyperparameter tuning is favorable because the strategy can minimize overfitting, generate a relatively high-performance predictive model, and reduce the appearance of susceptibility artifacts in the landslide area. Careful data inventory handling, CV, and hyperparameter tuning strategies should be considered in landslide susceptibility modeling to increase the applicability of landslide susceptibility maps in practical application.
机器学习(ML)算法经常用于滑坡易感性建模。即使使用相同的 ML 算法,不同的数据处理策略也可能导致滑坡易感性建模的差异。本研究旨在比较库存数据处理、交叉验证(CV)和超参数调整策略的组合,以生成滑坡易感性图。研究结果有望为使用 ML 技术进行滑坡易感性建模提供通用策略。作者采用了八种滑坡清单数据处理方案,将滑坡多边形转换为滑坡点,即滑坡点位于坡脚(最小高度)、坡面(最大高度)、滑坡中心、多边形内随机(1 点)、多边形内随机(3 点)、多边形内随机(5 点)、多边形内随机(10 点)和 15 米网格采样。每种数据处理策略都采用了 CV-非空间超参数调整、空间 CV-空间超参数调整和空间 CV-前向特征选择-无超参数调整的随机森林模型。组合生成了 24 个随机森林 ML 工作流,并将其应用于印尼帕契坦地区热带气旋 "肯帕卡"(2017 年)引发的 743 次滑坡的完整清单和 11 个滑坡控制因素。结果表明,网格采样加上空间 CV 和空间超参数调整是有利的,因为该策略可以最大限度地减少过拟合,生成性能相对较高的预测模型,并减少滑坡区域易感性假象的出现。在滑坡易感性建模中应考虑谨慎的数据清单处理、CV 和超参数调整策略,以提高滑坡易感性图在实际应用中的适用性。
{"title":"Benchmarking data handling strategies for landslide susceptibility modeling using random forest workflows","authors":"Guruh Samodra ,&nbsp;Ngadisih ,&nbsp;Ferman Setia Nugroho","doi":"10.1016/j.aiig.2024.100093","DOIUrl":"10.1016/j.aiig.2024.100093","url":null,"abstract":"<div><div>Machine learning (ML) algorithms are frequently used in landslide susceptibility modeling. Different data handling strategies may generate variations in landslide susceptibility modeling, even when using the same ML algorithm. This research aims to compare the combinations of inventory data handling, cross validation (CV), and hyperparameter tuning strategies to generate landslide susceptibility maps. The results are expected to provide a general strategy for landslide susceptibility modeling using ML techniques. The authors employed eight landslide inventory data handling scenarios to convert a landslide polygon into a landslide point, i.e., the landslide point is located on the toe (minimum height), on the scarp (maximum height), at the center of the landslide, randomly inside the polygon (1 point), randomly inside the polygon (3 points), randomly inside the polygon (5 points), randomly inside the polygon (10 points), and 15 m grid sampling. Random forest models using CV–nonspatial hyperparameter tuning, spatial CV–spatial hyperparameter tuning, and spatial CV–forward feature selection–no hyperparameter tuning were applied for each data handling strategy. The combination generated 24 random forest ML workflows, which are applied using a complete inventory of 743 landslides triggered by Tropical Cyclone Cempaka (2017) in Pacitan Regency, Indonesia, and 11 landslide controlling factors. The results show that grid sampling with spatial CV and spatial hyperparameter tuning is favorable because the strategy can minimize overfitting, generate a relatively high-performance predictive model, and reduce the appearance of susceptibility artifacts in the landslide area. Careful data inventory handling, CV, and hyperparameter tuning strategies should be considered in landslide susceptibility modeling to increase the applicability of landslide susceptibility maps in practical application.</div></div>","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"5 ","pages":"Article 100093"},"PeriodicalIF":0.0,"publicationDate":"2024-10-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142532855","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Artificial Intelligence in Geosciences
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1