Pub Date : 2025-03-01DOI: 10.1016/j.aiig.2025.100107
Jiaqi Yan , Li Ma , Tianqi Jiang , Jing Zheng , Dewei Li , Xingzhi Teng
This paper proposed a moment tensor regression prediction technology based on ResNet for microseismic events. Taking the great advantages of deep networks in classification and regression tasks, it can realize the great potential of fast and accurate inversion of microseismic moment tensors after the network trained. This ResNet-based moment tensor prediction technology, whose input is raw recordings, does not require the extraction of data features in advance. First, we tested the network using synthetic data and performed a quantitative assessment of the errors. The results demonstrate that the network exhibits high accuracy and efficiency during the prediction phase. Next, we tested the network using real microseismic data and compared the results with those from traditional inversion methods. The error in the results was relatively small compared to traditional methods. However, the network operates more efficiently without requiring manual intervention, making it highly valuable for near-real-time monitoring applications.
{"title":"Microseismic moment tensor inversion based on ResNet model","authors":"Jiaqi Yan , Li Ma , Tianqi Jiang , Jing Zheng , Dewei Li , Xingzhi Teng","doi":"10.1016/j.aiig.2025.100107","DOIUrl":"10.1016/j.aiig.2025.100107","url":null,"abstract":"<div><div>This paper proposed a moment tensor regression prediction technology based on ResNet for microseismic events. Taking the great advantages of deep networks in classification and regression tasks, it can realize the great potential of fast and accurate inversion of microseismic moment tensors after the network trained. This ResNet-based moment tensor prediction technology, whose input is raw recordings, does not require the extraction of data features in advance. First, we tested the network using synthetic data and performed a quantitative assessment of the errors. The results demonstrate that the network exhibits high accuracy and efficiency during the prediction phase. Next, we tested the network using real microseismic data and compared the results with those from traditional inversion methods. The error in the results was relatively small compared to traditional methods. However, the network operates more efficiently without requiring manual intervention, making it highly valuable for near-real-time monitoring applications.</div></div>","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"6 1","pages":"Article 100107"},"PeriodicalIF":0.0,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143551571","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-02-26DOI: 10.1016/j.aiig.2025.100110
A. Bolève, R. Eddies, M. Staring, Y. Benboudiaf, H. Pournaki, M. Nepveaux
Geotechnical parameters derived from an intrusive cone penetration test (CPT) are used to asses mechanical properties to inform the design phase of infrastructure projects. However, local, in situ 1D measurements can fail to capture 3D subsurface variations, which could mean less than optimal design decisions for foundation engineering. By coupling the localised measurements from CPTs with more global 3D measurements derived from geophysical methods, a higher fidelity 3D overview of the subsurface can be obtained. Machine Learning (ML) may offer an effective means to capture all types of geophysical information associated with CPT data at a site scale to build a 2D or 3D ground model. In this paper, we present an ML approach to build a 3D ground model of cone resistance and sleeve friction by combining several CPT measurements with Multichannel Analysis of Surface Waves (MASW) and Electrical Resistivity Tomography (ERT) data on a land site characterisation project in the United Arab Emirates (UAE). To avoid a potential overfitting problem inherent to the use of machine learning and a lack of data at certain locations, we explore the possibility of using a prior Geo-Statistical (GS) approach that attempts to constrain the overfitting process by “artificially” increasing the amount of input data. A sensitivity study is also performed on input features used to train the ML algorithm to better define the optimal combination of input features for the prediction. Our results showed that ERT data were not useful in capturing 3D variations of geotechnical properties compared to Vs due to the geographical location of the site (200 m east from the Oman Gulf) and the possible effect of saline water intrusion. Additionally, we demonstrate that the use of a prior GS phase could be a promising and interesting means to make the prediction of ground properties more robust, especially for this specific case study described in this paper. Looking ahead, better representation of the subsurface can lead to a number of benefits for stakeholders involved in developing assets. Better ground/geotechnical models mean better site calibration of design methods and fewer design assumptions for reliability-based design, creating an opportunity for value engineering in the form of lighter construction without compromising safety, shorter construction timelines, and reduced resource requirements.
{"title":"Innovative cone resistance and sleeve friction prediction from geophysics based on a coupled geo-statistical and machine learning process","authors":"A. Bolève, R. Eddies, M. Staring, Y. Benboudiaf, H. Pournaki, M. Nepveaux","doi":"10.1016/j.aiig.2025.100110","DOIUrl":"10.1016/j.aiig.2025.100110","url":null,"abstract":"<div><div>Geotechnical parameters derived from an intrusive cone penetration test (CPT) are used to asses mechanical properties to inform the design phase of infrastructure projects. However, local, in situ 1D measurements can fail to capture 3D subsurface variations, which could mean less than optimal design decisions for foundation engineering. By coupling the localised measurements from CPTs with more global 3D measurements derived from geophysical methods, a higher fidelity 3D overview of the subsurface can be obtained. Machine Learning (ML) may offer an effective means to capture all types of geophysical information associated with CPT data at a site scale to build a 2D or 3D ground model. In this paper, we present an ML approach to build a 3D ground model of cone resistance and sleeve friction by combining several CPT measurements with Multichannel Analysis of Surface Waves (MASW) and Electrical Resistivity Tomography (ERT) data on a land site characterisation project in the United Arab Emirates (UAE). To avoid a potential overfitting problem inherent to the use of machine learning and a lack of data at certain locations, we explore the possibility of using a prior Geo-Statistical (GS) approach that attempts to constrain the overfitting process by “artificially” increasing the amount of input data. A sensitivity study is also performed on input features used to train the ML algorithm to better define the optimal combination of input features for the prediction. Our results showed that ERT data were not useful in capturing 3D variations of geotechnical properties compared to V<sub>s</sub> due to the geographical location of the site (200 m east from the Oman Gulf) and the possible effect of saline water intrusion. Additionally, we demonstrate that the use of a prior GS phase could be a promising and interesting means to make the prediction of ground properties more robust, especially for this specific case study described in this paper. Looking ahead, better representation of the subsurface can lead to a number of benefits for stakeholders involved in developing assets. Better ground/geotechnical models mean better site calibration of design methods and fewer design assumptions for reliability-based design, creating an opportunity for value engineering in the form of lighter construction without compromising safety, shorter construction timelines, and reduced resource requirements.</div></div>","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"6 1","pages":"Article 100110"},"PeriodicalIF":0.0,"publicationDate":"2025-02-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143526930","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-02-25DOI: 10.1016/j.aiig.2025.100109
P. Zwartjes, J. Yoo
This work focuses on enhancing low frequency seismic data using a convolutional neural network trained on synthetic data. Traditional seismic data often lack both high and low frequencies, which are essential for detailed geological interpretation and various geophysical applications. Low frequency data is particularly valuable for reducing wavelet sidelobes and improving full waveform inversion (FWI). Conventional methods for bandwidth extension include seismic deconvolution and sparse inversion, which have limitations in recovering low frequencies. The study explores the potential of the U-net, which has been successful in other geophysical applications such as noise attenuation and seismic resolution enhancement. The novelty in our approach is that we do not rely on computationally expensive finite difference modelling to create training data. Instead, our synthetic training data is created from individual randomly perturbed events with variations in bandwidth, making it more adaptable to different data sets compared to previous deep learning methods. The method was tested on both synthetic and real seismic data, demonstrating effective low frequency reconstruction and sidelobe reduction. With a synthetic full waveform inversion to recover a velocity model and a seismic amplitude inversion to estimate acoustic impedance we demonstrate the validity and benefit of the proposed method. Overall, the study presents a robust approach to seismic bandwidth extension using deep learning, emphasizing the importance of diverse and well-designed but computationally inexpensive synthetic training data.
{"title":"Robust low frequency seismic bandwidth extension with a U-net and synthetic training data","authors":"P. Zwartjes, J. Yoo","doi":"10.1016/j.aiig.2025.100109","DOIUrl":"10.1016/j.aiig.2025.100109","url":null,"abstract":"<div><div>This work focuses on enhancing low frequency seismic data using a convolutional neural network trained on synthetic data. Traditional seismic data often lack both high and low frequencies, which are essential for detailed geological interpretation and various geophysical applications. Low frequency data is particularly valuable for reducing wavelet sidelobes and improving full waveform inversion (FWI). Conventional methods for bandwidth extension include seismic deconvolution and sparse inversion, which have limitations in recovering low frequencies. The study explores the potential of the U-net, which has been successful in other geophysical applications such as noise attenuation and seismic resolution enhancement. The novelty in our approach is that we do not rely on computationally expensive finite difference modelling to create training data. Instead, our synthetic training data is created from individual randomly perturbed events with variations in bandwidth, making it more adaptable to different data sets compared to previous deep learning methods. The method was tested on both synthetic and real seismic data, demonstrating effective low frequency reconstruction and sidelobe reduction. With a synthetic full waveform inversion to recover a velocity model and a seismic amplitude inversion to estimate acoustic impedance we demonstrate the validity and benefit of the proposed method. Overall, the study presents a robust approach to seismic bandwidth extension using deep learning, emphasizing the importance of diverse and well-designed but computationally inexpensive synthetic training data.</div></div>","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"6 1","pages":"Article 100109"},"PeriodicalIF":0.0,"publicationDate":"2025-02-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143512149","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-02-21DOI: 10.1016/j.aiig.2025.100108
Congcong Yuan , Jie Zhang
The availability of a tremendous amount of seismic data demands seismological researchers to analyze seismic phases efficiently. Recently, deep learning algorithms exhibit a powerful capability of detecting and picking on P- and S-wave phases. However, it remains a challenge to effeciently process enormous teleseismic phases, which are crucial to probe Earth's interior structures and their dynamics. In this study, we propose a scheme to detect and pick teleseismic phases, such as seismic phase that reflects off the core-mantle boundary (i.e., PcP) and that reflects off the inner-core boundary (i.e., PKiKP), from a seismic dataset in Japan. The scheme consists of three steps: 1) latent phase traces are truncated from the whole seismogram with theoretical arrival times; 2) latent phases are recognized and evaluated by convolutional neural network (CNN) models; 3) arrivals of good or fair phase are picked with another CNN models. The testing detection result on 7386 seismograms shows that the scheme recognizes 92.15% and 94.13% of PcP and PKiKP phases. The testing picking result has a mean absolute error of 0.0742 s and 0.0636 s for the PcP and PKiKP phases, respectively. These seismograms were processed in just 5 min for phase detection and picking, demonstrating the efficiency of the proposed scheme in automatic teleseismic phase analysis.
{"title":"Applying deep learning to teleseismic phase detection and picking: PcP and PKiKP cases","authors":"Congcong Yuan , Jie Zhang","doi":"10.1016/j.aiig.2025.100108","DOIUrl":"10.1016/j.aiig.2025.100108","url":null,"abstract":"<div><div>The availability of a tremendous amount of seismic data demands seismological researchers to analyze seismic phases efficiently. Recently, deep learning algorithms exhibit a powerful capability of detecting and picking on P- and S-wave phases. However, it remains a challenge to effeciently process enormous teleseismic phases, which are crucial to probe Earth's interior structures and their dynamics. In this study, we propose a scheme to detect and pick teleseismic phases, such as seismic phase that reflects off the core-mantle boundary (i.e., PcP) and that reflects off the inner-core boundary (i.e., PKiKP), from a seismic dataset in Japan. The scheme consists of three steps: 1) latent phase traces are truncated from the whole seismogram with theoretical arrival times; 2) latent phases are recognized and evaluated by convolutional neural network (CNN) models; 3) arrivals of good or fair phase are picked with another CNN models. The testing detection result on 7386 seismograms shows that the scheme recognizes 92.15% and 94.13% of PcP and PKiKP phases. The testing picking result has a mean absolute error of 0.0742 s and 0.0636 s for the PcP and PKiKP phases, respectively. These seismograms were processed in just 5 min for phase detection and picking, demonstrating the efficiency of the proposed scheme in automatic teleseismic phase analysis.</div></div>","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"6 1","pages":"Article 100108"},"PeriodicalIF":0.0,"publicationDate":"2025-02-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143488648","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-02-13DOI: 10.1016/j.aiig.2025.100105
Mohanad Diab , Polychronis Kolokoussis , Maria Antonia Brovelli
The use of AI technologies in remote sensing (RS) tasks has been the focus of many individuals in both the professional and academic domains. Having more accessible interfaces and tools that allow people of little or no experience to intuitively interact with RS data of multiple formats is a potential provided by this integration. However, the use of AI and AI agents to help automate RS-related tasks is still in its infancy stage, with some frameworks and interfaces built on top of well-known vision language models (VLM) such as GPT-4, segment anything model (SAM), and grounding DINO. These tools do promise and draw guidelines on the potentials and limitations of existing solutions concerning the use of said models. In this work, the state of the art AI foundation models (FM) are reviewed and used in a multi-modal manner to ingest RS imagery input and perform zero-shot object detection using natural language. The natural language input is then used to define the classes or labels the model should look for, then, both inputs are fed to the pipeline. The pipeline presented in this work makes up for the shortcomings of the general knowledge FMs by stacking pre-processing and post-processing applications on top of the FMs; these applications include tiling to produce uniform patches of the original image for faster detection, outlier rejection of redundant bounding boxes using statistical and machine learning methods. The pipeline was tested with UAV, aerial and satellite images taken over multiple areas. The accuracy for the semantic segmentation showed improvement from the original 64% to approximately 80%–99% by utilizing the pipeline and techniques proposed in this work. GitHub Repository:MohanadDiab/LangRS.
{"title":"Optimizing zero-shot text-based segmentation of remote sensing imagery using SAM and Grounding DINO","authors":"Mohanad Diab , Polychronis Kolokoussis , Maria Antonia Brovelli","doi":"10.1016/j.aiig.2025.100105","DOIUrl":"10.1016/j.aiig.2025.100105","url":null,"abstract":"<div><div>The use of AI technologies in remote sensing (RS) tasks has been the focus of many individuals in both the professional and academic domains. Having more accessible interfaces and tools that allow people of little or no experience to intuitively interact with RS data of multiple formats is a potential provided by this integration. However, the use of AI and AI agents to help automate RS-related tasks is still in its infancy stage, with some frameworks and interfaces built on top of well-known vision language models (VLM) such as GPT-4, segment anything model (SAM), and grounding DINO. These tools do promise and draw guidelines on the potentials and limitations of existing solutions concerning the use of said models. In this work, the state of the art AI foundation models (FM) are reviewed and used in a multi-modal manner to ingest RS imagery input and perform zero-shot object detection using natural language. The natural language input is then used to define the classes or labels the model should look for, then, both inputs are fed to the pipeline. The pipeline presented in this work makes up for the shortcomings of the general knowledge FMs by stacking pre-processing and post-processing applications on top of the FMs; these applications include tiling to produce uniform patches of the original image for faster detection, outlier rejection of redundant bounding boxes using statistical and machine learning methods. The pipeline was tested with UAV, aerial and satellite images taken over multiple areas. The accuracy for the semantic segmentation showed improvement from the original 64% to approximately 80%–99% by utilizing the pipeline and techniques proposed in this work. <strong>GitHub Repository:</strong> <span><span>MohanadDiab/LangRS</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"6 1","pages":"Article 100105"},"PeriodicalIF":0.0,"publicationDate":"2025-02-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143429918","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-27DOI: 10.1016/j.aiig.2025.100106
Kaoutar Clero , Said Ed-Diny , Mohammed Achalhi , Mouhamed Cherkaoui , Imad El Harraki , Sanaa El Fkihi , Intissar Benzakour , Tarik Soror , Said Rziki , Hamd Ait Abdelali , Hicham Tagemouati , François Bourzeix
Rockfalls are among the frequent hazards in underground mines worldwide, requiring effective methods for detecting unstable rock blocks to ensure miners' and equipment's safety. This study proposes a novel approach for identifying potential rockfall zones using infrared thermal imaging and image segmentation techniques. Infrared images of rock blocks were captured at the Draa Sfar deep underground mine in Morocco using the FLUKE TI401 PRO thermal camera. Two segmentation methods were applied to locate the potential unstable areas: the classical thresholding and the K-means clustering model. The results show that while thresholding allows a binary distinction between stable and unstable areas, K-means clustering is more accurate, especially when using multiple clusters to show different risk levels. The close match between the clustering masks of unstable blocks and their corresponding visible light images further validated this. The findings confirm that thermal image segmentation can serve as an alternative method for predicting rockfalls and monitoring geotechnical issues in underground mines. Underground operators worldwide can apply this approach to monitor rock mass stability. However, further research is recommended to enhance these results, particularly through deep learning-based segmentation and object detection models.
{"title":"Loosening rocks detection at Draa Sfar deep underground mine in Morocco using infrared thermal imaging and image segmentation models","authors":"Kaoutar Clero , Said Ed-Diny , Mohammed Achalhi , Mouhamed Cherkaoui , Imad El Harraki , Sanaa El Fkihi , Intissar Benzakour , Tarik Soror , Said Rziki , Hamd Ait Abdelali , Hicham Tagemouati , François Bourzeix","doi":"10.1016/j.aiig.2025.100106","DOIUrl":"10.1016/j.aiig.2025.100106","url":null,"abstract":"<div><div>Rockfalls are among the frequent hazards in underground mines worldwide, requiring effective methods for detecting unstable rock blocks to ensure miners' and equipment's safety. This study proposes a novel approach for identifying potential rockfall zones using infrared thermal imaging and image segmentation techniques. Infrared images of rock blocks were captured at the Draa Sfar deep underground mine in Morocco using the FLUKE TI401 PRO thermal camera. Two segmentation methods were applied to locate the potential unstable areas: the classical thresholding and the K-means clustering model. The results show that while thresholding allows a binary distinction between stable and unstable areas, K-means clustering is more accurate, especially when using multiple clusters to show different risk levels. The close match between the clustering masks of unstable blocks and their corresponding visible light images further validated this. The findings confirm that thermal image segmentation can serve as an alternative method for predicting rockfalls and monitoring geotechnical issues in underground mines. Underground operators worldwide can apply this approach to monitor rock mass stability. However, further research is recommended to enhance these results, particularly through deep learning-based segmentation and object detection models.</div></div>","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"6 1","pages":"Article 100106"},"PeriodicalIF":0.0,"publicationDate":"2025-01-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143143632","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-04DOI: 10.1016/j.aiig.2024.100104
Qiansheng Wei , Zishuai Li , Haonan Feng , Yueying Jiang , Yang Yang , Zhiguo Wang
Seismic time-frequency (TF) transforms are essential tools in reservoir interpretation and signal processing, particularly for characterizing frequency variations in non-stationary seismic data. Recently, sparse TF transforms, which leverage sparse coding (SC), have gained significant attention in the geosciences due to their ability to achieve high TF resolution. However, the iterative approaches typically employed in sparse TF transforms are computationally intensive, making them impractical for real seismic data analysis. To address this issue, we propose an interpretable convolutional sparse coding (CSC) network to achieve high TF resolution. The proposed model is generated based on the traditional short-time Fourier transform (STFT) transform and a modified UNet, named ULISTANet. In this design, we replace the conventional convolutional layers of the UNet with learnable iterative shrinkage thresholding algorithm (LISTA) blocks, a specialized form of CSC. The LISTA block, which evolves from the traditional iterative shrinkage thresholding algorithm (ISTA), is optimized for extracting sparse features more effectively. Furthermore, we create a synthetic dataset featuring complex frequency-modulated signals to train ULISTANet. Finally, the proposed method's performance is subsequently validated using both synthetic and field data, demonstrating its potential for enhanced seismic data analysis.
{"title":"Convolutional sparse coding network for sparse seismic time-frequency representation","authors":"Qiansheng Wei , Zishuai Li , Haonan Feng , Yueying Jiang , Yang Yang , Zhiguo Wang","doi":"10.1016/j.aiig.2024.100104","DOIUrl":"10.1016/j.aiig.2024.100104","url":null,"abstract":"<div><div>Seismic time-frequency (TF) transforms are essential tools in reservoir interpretation and signal processing, particularly for characterizing frequency variations in non-stationary seismic data. Recently, sparse TF transforms, which leverage sparse coding (SC), have gained significant attention in the geosciences due to their ability to achieve high TF resolution. However, the iterative approaches typically employed in sparse TF transforms are computationally intensive, making them impractical for real seismic data analysis. To address this issue, we propose an interpretable convolutional sparse coding (CSC) network to achieve high TF resolution. The proposed model is generated based on the traditional short-time Fourier transform (STFT) transform and a modified UNet, named ULISTANet. In this design, we replace the conventional convolutional layers of the UNet with learnable iterative shrinkage thresholding algorithm (LISTA) blocks, a specialized form of CSC. The LISTA block, which evolves from the traditional iterative shrinkage thresholding algorithm (ISTA), is optimized for extracting sparse features more effectively. Furthermore, we create a synthetic dataset featuring complex frequency-modulated signals to train ULISTANet. Finally, the proposed method's performance is subsequently validated using both synthetic and field data, demonstrating its potential for enhanced seismic data analysis.</div></div>","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"6 1","pages":"Article 100104"},"PeriodicalIF":0.0,"publicationDate":"2024-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142661128","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-16DOI: 10.1016/j.aiig.2024.100095
Cai Li, Fei Ma, Yuxiu Wang, Delong Zhang
The connectivity of sandbodies is a key constraint to the exploration effectiveness of Bohai A Oilfield. Conventional connectivity studies often use methods such as seismic attribute fusion, while the development of contiguous composite sandbodies in this area makes it challenging to characterize connectivity changes with conventional seismic attributes. Aiming at the above problem in the Bohai A Oilfield, this study proposes a big data analysis method based on the Deep Forest algorithm to predict the sandbody connectivity. Firstly, by compiling the abundant exploration and development sandbodies data in the study area, typical sandbodies with reliable connectivity were selected. Then, sensitive seismic attribute were extracted to obtain training samples. Finally, based on the Deep Forest algorithm, mapping model between attribute combinations and sandbody connectivity was established through machine learning. This method achieves the first quantitative determination of the connectivity for continuous composite sandbodies in the Bohai Oilfield. Compared with conventional connectivity discrimination methods such as high-resolution processing and seismic attribute analysis, this method can combine the sandbody characteristics of the study area in the process of machine learning, and jointly judge connectivity by combining multiple seismic attributes. The study results show that this method has high accuracy and timeliness in predicting connectivity for continuous composite sandbodies. Applied to the Bohai A Oilfield, it successfully identified multiple sandbody connectivity relationships and provided strong support for the subsequent exploration potential assessment and well placement optimization. This method also provides a new idea and method for studying sandbody connectivity under similar complex geological conditions.
砂体连通性是渤海 A 油田勘探有效性的关键制约因素。传统的连通性研究通常采用地震属性融合等方法,而该地区连片复合砂体的发育使得用常规地震属性表征连通性变化具有挑战性。针对渤海A油田的上述问题,本研究提出了一种基于深林算法的大数据分析方法来预测砂体连通性。首先,通过整理研究区丰富的勘探开发沙体数据,筛选出具有可靠连通性的典型沙体。然后,提取敏感地震属性,获得训练样本。最后,基于深林算法,通过机器学习建立属性组合与沙体连通性之间的映射模型。该方法首次实现了对渤海油田连续复合砂体连通性的定量判定。与传统的高分辨率处理、地震属性分析等连通性判别方法相比,该方法在机器学习过程中能够结合研究区的沙体特征,综合多种地震属性共同判断连通性。研究结果表明,该方法在预测连续复合砂体连通性方面具有较高的准确性和时效性。应用于渤海A油田,成功识别了多个砂体的连通性关系,为后续的勘探潜力评估和井位优化提供了有力支持。该方法也为研究类似复杂地质条件下的砂体连通性提供了新的思路和方法。
{"title":"Research on the prediction method for fluvial-phase sandbody connectivity based on big data analysis--a case study of Bohai a oilfield","authors":"Cai Li, Fei Ma, Yuxiu Wang, Delong Zhang","doi":"10.1016/j.aiig.2024.100095","DOIUrl":"10.1016/j.aiig.2024.100095","url":null,"abstract":"<div><div>The connectivity of sandbodies is a key constraint to the exploration effectiveness of Bohai A Oilfield. Conventional connectivity studies often use methods such as seismic attribute fusion, while the development of contiguous composite sandbodies in this area makes it challenging to characterize connectivity changes with conventional seismic attributes. Aiming at the above problem in the Bohai A Oilfield, this study proposes a big data analysis method based on the Deep Forest algorithm to predict the sandbody connectivity. Firstly, by compiling the abundant exploration and development sandbodies data in the study area, typical sandbodies with reliable connectivity were selected. Then, sensitive seismic attribute were extracted to obtain training samples. Finally, based on the Deep Forest algorithm, mapping model between attribute combinations and sandbody connectivity was established through machine learning. This method achieves the first quantitative determination of the connectivity for continuous composite sandbodies in the Bohai Oilfield. Compared with conventional connectivity discrimination methods such as high-resolution processing and seismic attribute analysis, this method can combine the sandbody characteristics of the study area in the process of machine learning, and jointly judge connectivity by combining multiple seismic attributes. The study results show that this method has high accuracy and timeliness in predicting connectivity for continuous composite sandbodies. Applied to the Bohai A Oilfield, it successfully identified multiple sandbody connectivity relationships and provided strong support for the subsequent exploration potential assessment and well placement optimization. This method also provides a new idea and method for studying sandbody connectivity under similar complex geological conditions.</div></div>","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"5 ","pages":"Article 100095"},"PeriodicalIF":0.0,"publicationDate":"2024-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142532292","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
<div><div>Pore size analysis plays a pivotal role in unraveling reservoir behavior and its intricate relationship with confined fluids. Traditional methods for predicting pore size distribution (PSD), relying on drilling cores or thin sections, face limitations associated with depth specificity. In this study, we introduce an innovative framework that leverages nuclear magnetic resonance (NMR) log data, encompassing clay-bound water (CBW), bound volume irreducible (BVI), and free fluid volume (FFV), to determine three PSDs (micropores, mesopores, and macropores). Moreover, we establish a robust pore size classification (PSC) system utilizing ternary plots, derived from the PSDs.</div><div>Within the three studied wells, NMR log data is exclusive to one well (well-A), while conventional well logs are accessible for all three wells (well-A, well-B, and well-C). This distinction enables PSD predictions for the remaining two wells (B and C). To prognosticate NMR outputs (CBW, BVI, FFV) for these wells, a two-step deep learning (DL) algorithm is implemented. Initially, three feature selection algorithms (f-classif, f-regression, and mutual-info-regression) identify the conventional well logs most correlated to NMR outputs in well-A. The three feature selection algorithms utilize statistical computations. These algorithms are utilized to systematically identify and optimize pertinent input features, thereby augmenting model interpretability and predictive efficacy within intricate data-driven endeavors. So, all three feature selection algorithms introduced the number of 4 logs as the most optimal number of inputs to the DL algorithm with different combinations of logs for each of the three desired outputs. Subsequently, the CUDA Deep Neural Network Long Short-Term Memory algorithm(CUDNNLSTM), belonging to the category of DL algorithms and harnessing the computational power of GPUs, is employed for the prediction of CBW, BVI, and FFV logs. This prediction leverages the optimal logs identified in the preceding step. Estimation of NMR outputs was done first in well-A (80% of data as training and 20% as testing). The correlation coefficient (CC) between the actual and estimated data for the three outputs CBW, BVI and FFV are 95%, 94%, and 97%, respectively, as well as root mean square error (RMSE) was obtained 0.0081, 0.098, and 0.0089, respectively. To assess the effectiveness of the proposed algorithm, we compared it with two traditional methods for log estimation: multiple regression and multi-resolution graph-based clustering methods. The results demonstrate the superior accuracy of our algorithm in comparison to these conventional approaches. This DL-driven approach facilitates PSD prediction grounded in fluid saturation for wells B and C.</div><div>Ternary plots are then employed for PSCs. Seven distinct PSCs within well-A employing actual NMR logs (CBW, BVI, FFV), in conjunction with an equivalent count within wells B and C utilizing three predicted
孔隙度分析在揭示储层行为及其与封闭流体的复杂关系方面起着举足轻重的作用。传统的孔径分布(PSD)预测方法依赖于钻井岩心或薄切片,面临着深度特异性的限制。在这项研究中,我们引入了一个创新框架,利用核磁共振(NMR)测井数据(包括粘土结合水(CBW)、不可还原结合体积(BVI)和自由流体体积(FFV))来确定三种孔径分布(微孔、中孔和大孔)。此外,我们还利用从 PSDs 得出的三元图建立了一个强大的孔径分类 (PSC) 系统。在所研究的三口井中,核磁共振测井数据仅适用于一口井(A 井),而常规测井数据则适用于所有三口井(A 井、B 井和 C 井)。这种区别使我们能够对其余两口井(B 井和 C 井)进行 PSD 预测。为了预测这些油井的 NMR 输出(CBW、BVI、FFV),采用了两步深度学习(DL)算法。首先,三种特征选择算法(f-classif、f-regression 和 mutual-info-regression)确定与 A 井 NMR 输出最相关的常规测井曲线。这三种特征选择算法利用统计计算。这些算法用于系统地识别和优化相关输入特征,从而在复杂的数据驱动工作中提高模型的可解释性和预测效力。因此,所有三种特征选择算法都将 4 个日志的数量作为 DL 算法的最佳输入数量,并为三种所需的输出分别引入不同的日志组合。随后,CUDA 深度神经网络长短期记忆算法(CUDNNLSTM)被用于预测 CBW、BVI 和 FFV 日志,该算法属于 DL 算法范畴,利用了 GPU 的计算能力。该预测利用了前一步中确定的最佳日志。首先在 A 井(80% 的数据作为训练数据,20% 作为测试数据)中对 NMR 输出进行估计。CBW、BVI 和 FFV 三项输出的实际数据与估计数据之间的相关系数(CC)分别为 95%、94% 和 97%,均方根误差(RMSE)分别为 0.0081、0.098 和 0.0089。为了评估所提出算法的有效性,我们将其与两种传统的对数估计方法进行了比较:多元回归和基于多分辨率图的聚类方法。结果表明,与这些传统方法相比,我们的算法具有更高的准确性。这种以 DL 为驱动的方法有助于根据 B 井和 C 井的流体饱和度预测 PSD。利用实际 NMR 测井(CBW、BVI、FFV)对 A 井中的七个不同 PSC 进行了和谐分类,同时利用三个预测测井对 B 井和 C 井中的等量 PSC 进行了和谐分类,从而确定了七个不同的孔径分类面 (PSCF)。由此产生的 PSCF 为生成精确、详细的储层三维模型提供了宝贵的见解。
{"title":"Pore size classification and prediction based on distribution of reservoir fluid volumes utilizing well logs and deep learning algorithm in a complex lithology","authors":"Hassan Bagheri , Reza Mohebian , Ali Moradzadeh , Behnia Azizzadeh Mehmandost Olya","doi":"10.1016/j.aiig.2024.100094","DOIUrl":"10.1016/j.aiig.2024.100094","url":null,"abstract":"<div><div>Pore size analysis plays a pivotal role in unraveling reservoir behavior and its intricate relationship with confined fluids. Traditional methods for predicting pore size distribution (PSD), relying on drilling cores or thin sections, face limitations associated with depth specificity. In this study, we introduce an innovative framework that leverages nuclear magnetic resonance (NMR) log data, encompassing clay-bound water (CBW), bound volume irreducible (BVI), and free fluid volume (FFV), to determine three PSDs (micropores, mesopores, and macropores). Moreover, we establish a robust pore size classification (PSC) system utilizing ternary plots, derived from the PSDs.</div><div>Within the three studied wells, NMR log data is exclusive to one well (well-A), while conventional well logs are accessible for all three wells (well-A, well-B, and well-C). This distinction enables PSD predictions for the remaining two wells (B and C). To prognosticate NMR outputs (CBW, BVI, FFV) for these wells, a two-step deep learning (DL) algorithm is implemented. Initially, three feature selection algorithms (f-classif, f-regression, and mutual-info-regression) identify the conventional well logs most correlated to NMR outputs in well-A. The three feature selection algorithms utilize statistical computations. These algorithms are utilized to systematically identify and optimize pertinent input features, thereby augmenting model interpretability and predictive efficacy within intricate data-driven endeavors. So, all three feature selection algorithms introduced the number of 4 logs as the most optimal number of inputs to the DL algorithm with different combinations of logs for each of the three desired outputs. Subsequently, the CUDA Deep Neural Network Long Short-Term Memory algorithm(CUDNNLSTM), belonging to the category of DL algorithms and harnessing the computational power of GPUs, is employed for the prediction of CBW, BVI, and FFV logs. This prediction leverages the optimal logs identified in the preceding step. Estimation of NMR outputs was done first in well-A (80% of data as training and 20% as testing). The correlation coefficient (CC) between the actual and estimated data for the three outputs CBW, BVI and FFV are 95%, 94%, and 97%, respectively, as well as root mean square error (RMSE) was obtained 0.0081, 0.098, and 0.0089, respectively. To assess the effectiveness of the proposed algorithm, we compared it with two traditional methods for log estimation: multiple regression and multi-resolution graph-based clustering methods. The results demonstrate the superior accuracy of our algorithm in comparison to these conventional approaches. This DL-driven approach facilitates PSD prediction grounded in fluid saturation for wells B and C.</div><div>Ternary plots are then employed for PSCs. Seven distinct PSCs within well-A employing actual NMR logs (CBW, BVI, FFV), in conjunction with an equivalent count within wells B and C utilizing three predicted","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"5 ","pages":"Article 100094"},"PeriodicalIF":0.0,"publicationDate":"2024-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142437701","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-05DOI: 10.1016/j.aiig.2024.100093
Guruh Samodra , Ngadisih , Ferman Setia Nugroho
Machine learning (ML) algorithms are frequently used in landslide susceptibility modeling. Different data handling strategies may generate variations in landslide susceptibility modeling, even when using the same ML algorithm. This research aims to compare the combinations of inventory data handling, cross validation (CV), and hyperparameter tuning strategies to generate landslide susceptibility maps. The results are expected to provide a general strategy for landslide susceptibility modeling using ML techniques. The authors employed eight landslide inventory data handling scenarios to convert a landslide polygon into a landslide point, i.e., the landslide point is located on the toe (minimum height), on the scarp (maximum height), at the center of the landslide, randomly inside the polygon (1 point), randomly inside the polygon (3 points), randomly inside the polygon (5 points), randomly inside the polygon (10 points), and 15 m grid sampling. Random forest models using CV–nonspatial hyperparameter tuning, spatial CV–spatial hyperparameter tuning, and spatial CV–forward feature selection–no hyperparameter tuning were applied for each data handling strategy. The combination generated 24 random forest ML workflows, which are applied using a complete inventory of 743 landslides triggered by Tropical Cyclone Cempaka (2017) in Pacitan Regency, Indonesia, and 11 landslide controlling factors. The results show that grid sampling with spatial CV and spatial hyperparameter tuning is favorable because the strategy can minimize overfitting, generate a relatively high-performance predictive model, and reduce the appearance of susceptibility artifacts in the landslide area. Careful data inventory handling, CV, and hyperparameter tuning strategies should be considered in landslide susceptibility modeling to increase the applicability of landslide susceptibility maps in practical application.
机器学习(ML)算法经常用于滑坡易感性建模。即使使用相同的 ML 算法,不同的数据处理策略也可能导致滑坡易感性建模的差异。本研究旨在比较库存数据处理、交叉验证(CV)和超参数调整策略的组合,以生成滑坡易感性图。研究结果有望为使用 ML 技术进行滑坡易感性建模提供通用策略。作者采用了八种滑坡清单数据处理方案,将滑坡多边形转换为滑坡点,即滑坡点位于坡脚(最小高度)、坡面(最大高度)、滑坡中心、多边形内随机(1 点)、多边形内随机(3 点)、多边形内随机(5 点)、多边形内随机(10 点)和 15 米网格采样。每种数据处理策略都采用了 CV-非空间超参数调整、空间 CV-空间超参数调整和空间 CV-前向特征选择-无超参数调整的随机森林模型。组合生成了 24 个随机森林 ML 工作流,并将其应用于印尼帕契坦地区热带气旋 "肯帕卡"(2017 年)引发的 743 次滑坡的完整清单和 11 个滑坡控制因素。结果表明,网格采样加上空间 CV 和空间超参数调整是有利的,因为该策略可以最大限度地减少过拟合,生成性能相对较高的预测模型,并减少滑坡区域易感性假象的出现。在滑坡易感性建模中应考虑谨慎的数据清单处理、CV 和超参数调整策略,以提高滑坡易感性图在实际应用中的适用性。
{"title":"Benchmarking data handling strategies for landslide susceptibility modeling using random forest workflows","authors":"Guruh Samodra , Ngadisih , Ferman Setia Nugroho","doi":"10.1016/j.aiig.2024.100093","DOIUrl":"10.1016/j.aiig.2024.100093","url":null,"abstract":"<div><div>Machine learning (ML) algorithms are frequently used in landslide susceptibility modeling. Different data handling strategies may generate variations in landslide susceptibility modeling, even when using the same ML algorithm. This research aims to compare the combinations of inventory data handling, cross validation (CV), and hyperparameter tuning strategies to generate landslide susceptibility maps. The results are expected to provide a general strategy for landslide susceptibility modeling using ML techniques. The authors employed eight landslide inventory data handling scenarios to convert a landslide polygon into a landslide point, i.e., the landslide point is located on the toe (minimum height), on the scarp (maximum height), at the center of the landslide, randomly inside the polygon (1 point), randomly inside the polygon (3 points), randomly inside the polygon (5 points), randomly inside the polygon (10 points), and 15 m grid sampling. Random forest models using CV–nonspatial hyperparameter tuning, spatial CV–spatial hyperparameter tuning, and spatial CV–forward feature selection–no hyperparameter tuning were applied for each data handling strategy. The combination generated 24 random forest ML workflows, which are applied using a complete inventory of 743 landslides triggered by Tropical Cyclone Cempaka (2017) in Pacitan Regency, Indonesia, and 11 landslide controlling factors. The results show that grid sampling with spatial CV and spatial hyperparameter tuning is favorable because the strategy can minimize overfitting, generate a relatively high-performance predictive model, and reduce the appearance of susceptibility artifacts in the landslide area. Careful data inventory handling, CV, and hyperparameter tuning strategies should be considered in landslide susceptibility modeling to increase the applicability of landslide susceptibility maps in practical application.</div></div>","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"5 ","pages":"Article 100093"},"PeriodicalIF":0.0,"publicationDate":"2024-10-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142532855","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}