首页 > 最新文献

Artificial Intelligence in Geosciences最新文献

英文 中文
Convolutional sparse coding network for sparse seismic time-frequency representation 用于稀疏地震时频表示的卷积稀疏编码网络
Pub Date : 2024-11-04 DOI: 10.1016/j.aiig.2024.100104
Qiansheng Wei , Zishuai Li , Haonan Feng , Yueying Jiang , Yang Yang , Zhiguo Wang
Seismic time-frequency (TF) transforms are essential tools in reservoir interpretation and signal processing, particularly for characterizing frequency variations in non-stationary seismic data. Recently, sparse TF transforms, which leverage sparse coding (SC), have gained significant attention in the geosciences due to their ability to achieve high TF resolution. However, the iterative approaches typically employed in sparse TF transforms are computationally intensive, making them impractical for real seismic data analysis. To address this issue, we propose an interpretable convolutional sparse coding (CSC) network to achieve high TF resolution. The proposed model is generated based on the traditional short-time Fourier transform (STFT) transform and a modified UNet, named ULISTANet. In this design, we replace the conventional convolutional layers of the UNet with learnable iterative shrinkage thresholding algorithm (LISTA) blocks, a specialized form of CSC. The LISTA block, which evolves from the traditional iterative shrinkage thresholding algorithm (ISTA), is optimized for extracting sparse features more effectively. Furthermore, we create a synthetic dataset featuring complex frequency-modulated signals to train ULISTANet. Finally, the proposed method's performance is subsequently validated using both synthetic and field data, demonstrating its potential for enhanced seismic data analysis.
地震时频(TF)变换是储层解释和信号处理的重要工具,特别是用于描述非稳态地震数据中的频率变化。最近,利用稀疏编码(SC)的稀疏时频变换因其实现高时频分辨率的能力而在地球科学领域备受关注。然而,稀疏 TF 变换通常采用的迭代方法需要大量计算,因此在实际地震数据分析中并不实用。为解决这一问题,我们提出了一种可解释卷积稀疏编码(CSC)网络,以实现高 TF 分辨率。我们提出的模型是基于传统的短时傅立叶变换(STFT)和改进的 UNet(名为 ULISTANet)生成的。在这一设计中,我们用可学习的迭代收缩阈值算法(LISTA)块(一种专门的 CSC 形式)取代了 UNet 的传统卷积层。LISTA 块由传统的迭代收缩阈值算法(ISTA)演化而来,经过优化,能更有效地提取稀疏特征。此外,我们还创建了一个以复杂频率调制信号为特征的合成数据集来训练 ULISTANet。最后,我们利用合成数据和野外数据对所提出方法的性能进行了验证,证明了该方法在增强地震数据分析方面的潜力。
{"title":"Convolutional sparse coding network for sparse seismic time-frequency representation","authors":"Qiansheng Wei ,&nbsp;Zishuai Li ,&nbsp;Haonan Feng ,&nbsp;Yueying Jiang ,&nbsp;Yang Yang ,&nbsp;Zhiguo Wang","doi":"10.1016/j.aiig.2024.100104","DOIUrl":"10.1016/j.aiig.2024.100104","url":null,"abstract":"<div><div>Seismic time-frequency (TF) transforms are essential tools in reservoir interpretation and signal processing, particularly for characterizing frequency variations in non-stationary seismic data. Recently, sparse TF transforms, which leverage sparse coding (SC), have gained significant attention in the geosciences due to their ability to achieve high TF resolution. However, the iterative approaches typically employed in sparse TF transforms are computationally intensive, making them impractical for real seismic data analysis. To address this issue, we propose an interpretable convolutional sparse coding (CSC) network to achieve high TF resolution. The proposed model is generated based on the traditional short-time Fourier transform (STFT) transform and a modified UNet, named ULISTANet. In this design, we replace the conventional convolutional layers of the UNet with learnable iterative shrinkage thresholding algorithm (LISTA) blocks, a specialized form of CSC. The LISTA block, which evolves from the traditional iterative shrinkage thresholding algorithm (ISTA), is optimized for extracting sparse features more effectively. Furthermore, we create a synthetic dataset featuring complex frequency-modulated signals to train ULISTANet. Finally, the proposed method's performance is subsequently validated using both synthetic and field data, demonstrating its potential for enhanced seismic data analysis.</div></div>","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"6 1","pages":"Article 100104"},"PeriodicalIF":0.0,"publicationDate":"2024-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142661128","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Research on the prediction method for fluvial-phase sandbody connectivity based on big data analysis--a case study of Bohai a oilfield 基于大数据分析的流相砂体连通性预测方法研究--以渤海某油田为例
Pub Date : 2024-10-16 DOI: 10.1016/j.aiig.2024.100095
Cai Li, Fei Ma, Yuxiu Wang, Delong Zhang
The connectivity of sandbodies is a key constraint to the exploration effectiveness of Bohai A Oilfield. Conventional connectivity studies often use methods such as seismic attribute fusion, while the development of contiguous composite sandbodies in this area makes it challenging to characterize connectivity changes with conventional seismic attributes. Aiming at the above problem in the Bohai A Oilfield, this study proposes a big data analysis method based on the Deep Forest algorithm to predict the sandbody connectivity. Firstly, by compiling the abundant exploration and development sandbodies data in the study area, typical sandbodies with reliable connectivity were selected. Then, sensitive seismic attribute were extracted to obtain training samples. Finally, based on the Deep Forest algorithm, mapping model between attribute combinations and sandbody connectivity was established through machine learning. This method achieves the first quantitative determination of the connectivity for continuous composite sandbodies in the Bohai Oilfield. Compared with conventional connectivity discrimination methods such as high-resolution processing and seismic attribute analysis, this method can combine the sandbody characteristics of the study area in the process of machine learning, and jointly judge connectivity by combining multiple seismic attributes. The study results show that this method has high accuracy and timeliness in predicting connectivity for continuous composite sandbodies. Applied to the Bohai A Oilfield, it successfully identified multiple sandbody connectivity relationships and provided strong support for the subsequent exploration potential assessment and well placement optimization. This method also provides a new idea and method for studying sandbody connectivity under similar complex geological conditions.
砂体连通性是渤海 A 油田勘探有效性的关键制约因素。传统的连通性研究通常采用地震属性融合等方法,而该地区连片复合砂体的发育使得用常规地震属性表征连通性变化具有挑战性。针对渤海A油田的上述问题,本研究提出了一种基于深林算法的大数据分析方法来预测砂体连通性。首先,通过整理研究区丰富的勘探开发沙体数据,筛选出具有可靠连通性的典型沙体。然后,提取敏感地震属性,获得训练样本。最后,基于深林算法,通过机器学习建立属性组合与沙体连通性之间的映射模型。该方法首次实现了对渤海油田连续复合砂体连通性的定量判定。与传统的高分辨率处理、地震属性分析等连通性判别方法相比,该方法在机器学习过程中能够结合研究区的沙体特征,综合多种地震属性共同判断连通性。研究结果表明,该方法在预测连续复合砂体连通性方面具有较高的准确性和时效性。应用于渤海A油田,成功识别了多个砂体的连通性关系,为后续的勘探潜力评估和井位优化提供了有力支持。该方法也为研究类似复杂地质条件下的砂体连通性提供了新的思路和方法。
{"title":"Research on the prediction method for fluvial-phase sandbody connectivity based on big data analysis--a case study of Bohai a oilfield","authors":"Cai Li,&nbsp;Fei Ma,&nbsp;Yuxiu Wang,&nbsp;Delong Zhang","doi":"10.1016/j.aiig.2024.100095","DOIUrl":"10.1016/j.aiig.2024.100095","url":null,"abstract":"<div><div>The connectivity of sandbodies is a key constraint to the exploration effectiveness of Bohai A Oilfield. Conventional connectivity studies often use methods such as seismic attribute fusion, while the development of contiguous composite sandbodies in this area makes it challenging to characterize connectivity changes with conventional seismic attributes. Aiming at the above problem in the Bohai A Oilfield, this study proposes a big data analysis method based on the Deep Forest algorithm to predict the sandbody connectivity. Firstly, by compiling the abundant exploration and development sandbodies data in the study area, typical sandbodies with reliable connectivity were selected. Then, sensitive seismic attribute were extracted to obtain training samples. Finally, based on the Deep Forest algorithm, mapping model between attribute combinations and sandbody connectivity was established through machine learning. This method achieves the first quantitative determination of the connectivity for continuous composite sandbodies in the Bohai Oilfield. Compared with conventional connectivity discrimination methods such as high-resolution processing and seismic attribute analysis, this method can combine the sandbody characteristics of the study area in the process of machine learning, and jointly judge connectivity by combining multiple seismic attributes. The study results show that this method has high accuracy and timeliness in predicting connectivity for continuous composite sandbodies. Applied to the Bohai A Oilfield, it successfully identified multiple sandbody connectivity relationships and provided strong support for the subsequent exploration potential assessment and well placement optimization. This method also provides a new idea and method for studying sandbody connectivity under similar complex geological conditions.</div></div>","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"5 ","pages":"Article 100095"},"PeriodicalIF":0.0,"publicationDate":"2024-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142532292","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Pore size classification and prediction based on distribution of reservoir fluid volumes utilizing well logs and deep learning algorithm in a complex lithology 在复杂岩性中利用测井记录和深度学习算法,基于储层流体体积分布进行孔隙尺寸分类和预测
Pub Date : 2024-10-12 DOI: 10.1016/j.aiig.2024.100094
Hassan Bagheri , Reza Mohebian , Ali Moradzadeh , Behnia Azizzadeh Mehmandost Olya
<div><div>Pore size analysis plays a pivotal role in unraveling reservoir behavior and its intricate relationship with confined fluids. Traditional methods for predicting pore size distribution (PSD), relying on drilling cores or thin sections, face limitations associated with depth specificity. In this study, we introduce an innovative framework that leverages nuclear magnetic resonance (NMR) log data, encompassing clay-bound water (CBW), bound volume irreducible (BVI), and free fluid volume (FFV), to determine three PSDs (micropores, mesopores, and macropores). Moreover, we establish a robust pore size classification (PSC) system utilizing ternary plots, derived from the PSDs.</div><div>Within the three studied wells, NMR log data is exclusive to one well (well-A), while conventional well logs are accessible for all three wells (well-A, well-B, and well-C). This distinction enables PSD predictions for the remaining two wells (B and C). To prognosticate NMR outputs (CBW, BVI, FFV) for these wells, a two-step deep learning (DL) algorithm is implemented. Initially, three feature selection algorithms (f-classif, f-regression, and mutual-info-regression) identify the conventional well logs most correlated to NMR outputs in well-A. The three feature selection algorithms utilize statistical computations. These algorithms are utilized to systematically identify and optimize pertinent input features, thereby augmenting model interpretability and predictive efficacy within intricate data-driven endeavors. So, all three feature selection algorithms introduced the number of 4 logs as the most optimal number of inputs to the DL algorithm with different combinations of logs for each of the three desired outputs. Subsequently, the CUDA Deep Neural Network Long Short-Term Memory algorithm(CUDNNLSTM), belonging to the category of DL algorithms and harnessing the computational power of GPUs, is employed for the prediction of CBW, BVI, and FFV logs. This prediction leverages the optimal logs identified in the preceding step. Estimation of NMR outputs was done first in well-A (80% of data as training and 20% as testing). The correlation coefficient (CC) between the actual and estimated data for the three outputs CBW, BVI and FFV are 95%, 94%, and 97%, respectively, as well as root mean square error (RMSE) was obtained 0.0081, 0.098, and 0.0089, respectively. To assess the effectiveness of the proposed algorithm, we compared it with two traditional methods for log estimation: multiple regression and multi-resolution graph-based clustering methods. The results demonstrate the superior accuracy of our algorithm in comparison to these conventional approaches. This DL-driven approach facilitates PSD prediction grounded in fluid saturation for wells B and C.</div><div>Ternary plots are then employed for PSCs. Seven distinct PSCs within well-A employing actual NMR logs (CBW, BVI, FFV), in conjunction with an equivalent count within wells B and C utilizing three predicted
孔隙度分析在揭示储层行为及其与封闭流体的复杂关系方面起着举足轻重的作用。传统的孔径分布(PSD)预测方法依赖于钻井岩心或薄切片,面临着深度特异性的限制。在这项研究中,我们引入了一个创新框架,利用核磁共振(NMR)测井数据(包括粘土结合水(CBW)、不可还原结合体积(BVI)和自由流体体积(FFV))来确定三种孔径分布(微孔、中孔和大孔)。此外,我们还利用从 PSDs 得出的三元图建立了一个强大的孔径分类 (PSC) 系统。在所研究的三口井中,核磁共振测井数据仅适用于一口井(A 井),而常规测井数据则适用于所有三口井(A 井、B 井和 C 井)。这种区别使我们能够对其余两口井(B 井和 C 井)进行 PSD 预测。为了预测这些油井的 NMR 输出(CBW、BVI、FFV),采用了两步深度学习(DL)算法。首先,三种特征选择算法(f-classif、f-regression 和 mutual-info-regression)确定与 A 井 NMR 输出最相关的常规测井曲线。这三种特征选择算法利用统计计算。这些算法用于系统地识别和优化相关输入特征,从而在复杂的数据驱动工作中提高模型的可解释性和预测效力。因此,所有三种特征选择算法都将 4 个日志的数量作为 DL 算法的最佳输入数量,并为三种所需的输出分别引入不同的日志组合。随后,CUDA 深度神经网络长短期记忆算法(CUDNNLSTM)被用于预测 CBW、BVI 和 FFV 日志,该算法属于 DL 算法范畴,利用了 GPU 的计算能力。该预测利用了前一步中确定的最佳日志。首先在 A 井(80% 的数据作为训练数据,20% 作为测试数据)中对 NMR 输出进行估计。CBW、BVI 和 FFV 三项输出的实际数据与估计数据之间的相关系数(CC)分别为 95%、94% 和 97%,均方根误差(RMSE)分别为 0.0081、0.098 和 0.0089。为了评估所提出算法的有效性,我们将其与两种传统的对数估计方法进行了比较:多元回归和基于多分辨率图的聚类方法。结果表明,与这些传统方法相比,我们的算法具有更高的准确性。这种以 DL 为驱动的方法有助于根据 B 井和 C 井的流体饱和度预测 PSD。利用实际 NMR 测井(CBW、BVI、FFV)对 A 井中的七个不同 PSC 进行了和谐分类,同时利用三个预测测井对 B 井和 C 井中的等量 PSC 进行了和谐分类,从而确定了七个不同的孔径分类面 (PSCF)。由此产生的 PSCF 为生成精确、详细的储层三维模型提供了宝贵的见解。
{"title":"Pore size classification and prediction based on distribution of reservoir fluid volumes utilizing well logs and deep learning algorithm in a complex lithology","authors":"Hassan Bagheri ,&nbsp;Reza Mohebian ,&nbsp;Ali Moradzadeh ,&nbsp;Behnia Azizzadeh Mehmandost Olya","doi":"10.1016/j.aiig.2024.100094","DOIUrl":"10.1016/j.aiig.2024.100094","url":null,"abstract":"&lt;div&gt;&lt;div&gt;Pore size analysis plays a pivotal role in unraveling reservoir behavior and its intricate relationship with confined fluids. Traditional methods for predicting pore size distribution (PSD), relying on drilling cores or thin sections, face limitations associated with depth specificity. In this study, we introduce an innovative framework that leverages nuclear magnetic resonance (NMR) log data, encompassing clay-bound water (CBW), bound volume irreducible (BVI), and free fluid volume (FFV), to determine three PSDs (micropores, mesopores, and macropores). Moreover, we establish a robust pore size classification (PSC) system utilizing ternary plots, derived from the PSDs.&lt;/div&gt;&lt;div&gt;Within the three studied wells, NMR log data is exclusive to one well (well-A), while conventional well logs are accessible for all three wells (well-A, well-B, and well-C). This distinction enables PSD predictions for the remaining two wells (B and C). To prognosticate NMR outputs (CBW, BVI, FFV) for these wells, a two-step deep learning (DL) algorithm is implemented. Initially, three feature selection algorithms (f-classif, f-regression, and mutual-info-regression) identify the conventional well logs most correlated to NMR outputs in well-A. The three feature selection algorithms utilize statistical computations. These algorithms are utilized to systematically identify and optimize pertinent input features, thereby augmenting model interpretability and predictive efficacy within intricate data-driven endeavors. So, all three feature selection algorithms introduced the number of 4 logs as the most optimal number of inputs to the DL algorithm with different combinations of logs for each of the three desired outputs. Subsequently, the CUDA Deep Neural Network Long Short-Term Memory algorithm(CUDNNLSTM), belonging to the category of DL algorithms and harnessing the computational power of GPUs, is employed for the prediction of CBW, BVI, and FFV logs. This prediction leverages the optimal logs identified in the preceding step. Estimation of NMR outputs was done first in well-A (80% of data as training and 20% as testing). The correlation coefficient (CC) between the actual and estimated data for the three outputs CBW, BVI and FFV are 95%, 94%, and 97%, respectively, as well as root mean square error (RMSE) was obtained 0.0081, 0.098, and 0.0089, respectively. To assess the effectiveness of the proposed algorithm, we compared it with two traditional methods for log estimation: multiple regression and multi-resolution graph-based clustering methods. The results demonstrate the superior accuracy of our algorithm in comparison to these conventional approaches. This DL-driven approach facilitates PSD prediction grounded in fluid saturation for wells B and C.&lt;/div&gt;&lt;div&gt;Ternary plots are then employed for PSCs. Seven distinct PSCs within well-A employing actual NMR logs (CBW, BVI, FFV), in conjunction with an equivalent count within wells B and C utilizing three predicted","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"5 ","pages":"Article 100094"},"PeriodicalIF":0.0,"publicationDate":"2024-10-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142437701","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Benchmarking data handling strategies for landslide susceptibility modeling using random forest workflows 利用随机森林工作流程为滑坡易发性建模的数据处理策略制定基准
Pub Date : 2024-10-05 DOI: 10.1016/j.aiig.2024.100093
Guruh Samodra , Ngadisih , Ferman Setia Nugroho
Machine learning (ML) algorithms are frequently used in landslide susceptibility modeling. Different data handling strategies may generate variations in landslide susceptibility modeling, even when using the same ML algorithm. This research aims to compare the combinations of inventory data handling, cross validation (CV), and hyperparameter tuning strategies to generate landslide susceptibility maps. The results are expected to provide a general strategy for landslide susceptibility modeling using ML techniques. The authors employed eight landslide inventory data handling scenarios to convert a landslide polygon into a landslide point, i.e., the landslide point is located on the toe (minimum height), on the scarp (maximum height), at the center of the landslide, randomly inside the polygon (1 point), randomly inside the polygon (3 points), randomly inside the polygon (5 points), randomly inside the polygon (10 points), and 15 m grid sampling. Random forest models using CV–nonspatial hyperparameter tuning, spatial CV–spatial hyperparameter tuning, and spatial CV–forward feature selection–no hyperparameter tuning were applied for each data handling strategy. The combination generated 24 random forest ML workflows, which are applied using a complete inventory of 743 landslides triggered by Tropical Cyclone Cempaka (2017) in Pacitan Regency, Indonesia, and 11 landslide controlling factors. The results show that grid sampling with spatial CV and spatial hyperparameter tuning is favorable because the strategy can minimize overfitting, generate a relatively high-performance predictive model, and reduce the appearance of susceptibility artifacts in the landslide area. Careful data inventory handling, CV, and hyperparameter tuning strategies should be considered in landslide susceptibility modeling to increase the applicability of landslide susceptibility maps in practical application.
机器学习(ML)算法经常用于滑坡易感性建模。即使使用相同的 ML 算法,不同的数据处理策略也可能导致滑坡易感性建模的差异。本研究旨在比较库存数据处理、交叉验证(CV)和超参数调整策略的组合,以生成滑坡易感性图。研究结果有望为使用 ML 技术进行滑坡易感性建模提供通用策略。作者采用了八种滑坡清单数据处理方案,将滑坡多边形转换为滑坡点,即滑坡点位于坡脚(最小高度)、坡面(最大高度)、滑坡中心、多边形内随机(1 点)、多边形内随机(3 点)、多边形内随机(5 点)、多边形内随机(10 点)和 15 米网格采样。每种数据处理策略都采用了 CV-非空间超参数调整、空间 CV-空间超参数调整和空间 CV-前向特征选择-无超参数调整的随机森林模型。组合生成了 24 个随机森林 ML 工作流,并将其应用于印尼帕契坦地区热带气旋 "肯帕卡"(2017 年)引发的 743 次滑坡的完整清单和 11 个滑坡控制因素。结果表明,网格采样加上空间 CV 和空间超参数调整是有利的,因为该策略可以最大限度地减少过拟合,生成性能相对较高的预测模型,并减少滑坡区域易感性假象的出现。在滑坡易感性建模中应考虑谨慎的数据清单处理、CV 和超参数调整策略,以提高滑坡易感性图在实际应用中的适用性。
{"title":"Benchmarking data handling strategies for landslide susceptibility modeling using random forest workflows","authors":"Guruh Samodra ,&nbsp;Ngadisih ,&nbsp;Ferman Setia Nugroho","doi":"10.1016/j.aiig.2024.100093","DOIUrl":"10.1016/j.aiig.2024.100093","url":null,"abstract":"<div><div>Machine learning (ML) algorithms are frequently used in landslide susceptibility modeling. Different data handling strategies may generate variations in landslide susceptibility modeling, even when using the same ML algorithm. This research aims to compare the combinations of inventory data handling, cross validation (CV), and hyperparameter tuning strategies to generate landslide susceptibility maps. The results are expected to provide a general strategy for landslide susceptibility modeling using ML techniques. The authors employed eight landslide inventory data handling scenarios to convert a landslide polygon into a landslide point, i.e., the landslide point is located on the toe (minimum height), on the scarp (maximum height), at the center of the landslide, randomly inside the polygon (1 point), randomly inside the polygon (3 points), randomly inside the polygon (5 points), randomly inside the polygon (10 points), and 15 m grid sampling. Random forest models using CV–nonspatial hyperparameter tuning, spatial CV–spatial hyperparameter tuning, and spatial CV–forward feature selection–no hyperparameter tuning were applied for each data handling strategy. The combination generated 24 random forest ML workflows, which are applied using a complete inventory of 743 landslides triggered by Tropical Cyclone Cempaka (2017) in Pacitan Regency, Indonesia, and 11 landslide controlling factors. The results show that grid sampling with spatial CV and spatial hyperparameter tuning is favorable because the strategy can minimize overfitting, generate a relatively high-performance predictive model, and reduce the appearance of susceptibility artifacts in the landslide area. Careful data inventory handling, CV, and hyperparameter tuning strategies should be considered in landslide susceptibility modeling to increase the applicability of landslide susceptibility maps in practical application.</div></div>","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"5 ","pages":"Article 100093"},"PeriodicalIF":0.0,"publicationDate":"2024-10-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142532855","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A 3D convolutional neural network model with multiple outputs for simultaneously estimating the reactive transport parameters of sandstone from its CT images 具有多输出的三维卷积神经网络模型,可同时从 CT 图像中估算砂岩的反应输运参数
Pub Date : 2024-09-22 DOI: 10.1016/j.aiig.2024.100092
Haiying Fu , Shuai Wang , Guicheng He , Zhonghua Zhu , Qing Yu , Dexin Ding
Porosity, tortuosity, specific surface area (SSA), and permeability are four key parameters of reactive transport modeling in sandstone, which are important for understanding solute transport and geochemical reaction processes in sandstone aquifers. These four parameters reflect the characteristics of pore structure of sandstone from different perspectives, and the traditional empirical formulas cannot make accurate predictions of them due to their complexity and heterogeneity. In this paper, eleven types of sandstone CT images were firstly segmented into numerous subsample images, the porosity, tortuosity, SSA, and permeability of the subsamples were calculated, and the dataset was established. The 3D convolutional neural network (CNN) models were subsequently established and trained to predict the key reactive transport parameters based on subsample CT images of sandstones. The results demonstrated that the 3D CNN model with multiple outputs exhibited excellent prediction ability for the four parameters compared to the traditional empirical formulas. In particular, for the prediction of tortuosity and permeability, the 3D CNN model with multiple outputs even showed slightly better prediction ability than its single-output variant model. Additionally, it demonstrated good generalization performance on sandstone CT images not included in the training dataset. The study showed that the 3D CNN model with multiple outputs has the advantages of simplifying operation and saving computational resources, which has the prospect of popularization and application.
孔隙度、曲折度、比表面积(SSA)和渗透率是砂岩反应运移模型的四个关键参数,对于理解砂岩含水层中的溶质运移和地球化学反应过程非常重要。这四个参数从不同角度反映了砂岩孔隙结构的特征,由于其复杂性和异质性,传统的经验公式无法对其进行准确预测。本文首先将 11 种砂岩 CT 图像分割成许多子样本图像,计算子样本的孔隙度、迂回度、SSA 和渗透率,并建立数据集。随后建立了三维卷积神经网络(CNN)模型,并对其进行了训练,以根据砂岩的子样本 CT 图像预测关键的反应输运参数。结果表明,与传统的经验公式相比,多输出的三维卷积神经网络模型对四个参数具有出色的预测能力。特别是在预测曲度和渗透率时,多输出三维 CNN 模型的预测能力甚至略优于其单输出变体模型。此外,该模型在未包含在训练数据集中的砂岩 CT 图像上也表现出了良好的泛化性能。研究表明,多输出的三维 CNN 模型具有简化操作和节省计算资源的优点,具有推广应用的前景。
{"title":"A 3D convolutional neural network model with multiple outputs for simultaneously estimating the reactive transport parameters of sandstone from its CT images","authors":"Haiying Fu ,&nbsp;Shuai Wang ,&nbsp;Guicheng He ,&nbsp;Zhonghua Zhu ,&nbsp;Qing Yu ,&nbsp;Dexin Ding","doi":"10.1016/j.aiig.2024.100092","DOIUrl":"10.1016/j.aiig.2024.100092","url":null,"abstract":"<div><div>Porosity, tortuosity, specific surface area (SSA), and permeability are four key parameters of reactive transport modeling in sandstone, which are important for understanding solute transport and geochemical reaction processes in sandstone aquifers. These four parameters reflect the characteristics of pore structure of sandstone from different perspectives, and the traditional empirical formulas cannot make accurate predictions of them due to their complexity and heterogeneity. In this paper, eleven types of sandstone CT images were firstly segmented into numerous subsample images, the porosity, tortuosity, SSA, and permeability of the subsamples were calculated, and the dataset was established. The 3D convolutional neural network (CNN) models were subsequently established and trained to predict the key reactive transport parameters based on subsample CT images of sandstones. The results demonstrated that the 3D CNN model with multiple outputs exhibited excellent prediction ability for the four parameters compared to the traditional empirical formulas. In particular, for the prediction of tortuosity and permeability, the 3D CNN model with multiple outputs even showed slightly better prediction ability than its single-output variant model. Additionally, it demonstrated good generalization performance on sandstone CT images not included in the training dataset. The study showed that the 3D CNN model with multiple outputs has the advantages of simplifying operation and saving computational resources, which has the prospect of popularization and application.</div></div>","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"5 ","pages":"Article 100092"},"PeriodicalIF":0.0,"publicationDate":"2024-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142318693","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Transfer learning for well logging formation evaluation using similarity weights 利用相似性权重进行测井地层评价的迁移学习
Pub Date : 2024-09-20 DOI: 10.1016/j.aiig.2024.100091
Binsen Xu , Zhou Feng , Jun Zhou , Rongbo Shao , Hongliang Wu , Peng Liu , Han Tian , Weizhong Li , Lizhi Xiao
Machine learning has been widely applied in well logging formation evaluation studies. However, several challenges negatively impacted the generalization capabilities of machine learning models in practical implementations, such as the mismatch of data domain between training and testing datasets, imbalances among sample categories, and inadequate representation of data model. These issues have led to substantial insufficient identification for reservoir and significant deviations in subsequent evaluations. To improve the transferability of machine learning models within limited sample sets, this study proposes a weight transfer learning framework based on the similarity of the labels. The similarity weighting method includes both hard weights and soft weights. By evaluating the similarity between test and training sets of logging data, the similarity results are used to estimate the weights of training samples, thereby optimizing the model learning process. We develop a double experts' network and a bidirectional gated neural network based on hierarchical attention and multi-head attention (BiGRU-MHSA) for well logs reconstruction and lithofacies classification tasks. Oil field data results for the shale strata in the Gulong area of the Songliao Basin of China indicate that the double experts’ network model performs well in curve reconstruction tasks. However, it may not be effective in lithofacies classification tasks, while BiGRU-MHSA performs well in that area. In the study of constructing large-scale well logging processing and formation interpretation models, it is maybe more beneficial by employing different expert models for combined evaluations. In addition, although the improvement is limited, hard or soft weighting methods is better than unweighted (i.e., average-weighted) in significantly different adjacent wells. The code and data are open and available for subsequent studies on other lithofacies layers.
机器学习已被广泛应用于测井地层评估研究。然而,在实际应用中,一些挑战对机器学习模型的泛化能力产生了负面影响,例如训练数据集和测试数据集之间的数据域不匹配、样本类别之间的不平衡以及数据模型的不充分表示。这些问题导致蓄水池的识别能力严重不足,并在后续评估中出现重大偏差。为了提高机器学习模型在有限样本集内的可转移性,本研究提出了一种基于标签相似性的权重转移学习框架。相似性加权方法包括硬加权和软加权。通过评估日志数据的测试集和训练集之间的相似性,利用相似性结果来估计训练样本的权重,从而优化模型学习过程。我们开发了基于分层注意力和多头注意力的双专家网络和双向门控神经网络(BiGRU-MHSA),用于井录重建和岩性分类任务。中国松辽盆地古隆地区页岩地层的油田数据结果表明,双专家网络模型在曲线重建任务中表现良好。但在岩性分类任务中可能效果不佳,而 BiGRU-MHSA 在该领域表现良好。在构建大规模测井处理和地层解释模型的研究中,采用不同的专家模型进行综合评价可能更有益处。此外,虽然改进有限,但在相邻油井差异明显的情况下,硬加权或软加权方法比不加权(即平均加权)方法更好。代码和数据是开放的,可用于其他岩性层的后续研究。
{"title":"Transfer learning for well logging formation evaluation using similarity weights","authors":"Binsen Xu ,&nbsp;Zhou Feng ,&nbsp;Jun Zhou ,&nbsp;Rongbo Shao ,&nbsp;Hongliang Wu ,&nbsp;Peng Liu ,&nbsp;Han Tian ,&nbsp;Weizhong Li ,&nbsp;Lizhi Xiao","doi":"10.1016/j.aiig.2024.100091","DOIUrl":"10.1016/j.aiig.2024.100091","url":null,"abstract":"<div><div>Machine learning has been widely applied in well logging formation evaluation studies. However, several challenges negatively impacted the generalization capabilities of machine learning models in practical implementations, such as the mismatch of data domain between training and testing datasets, imbalances among sample categories, and inadequate representation of data model. These issues have led to substantial insufficient identification for reservoir and significant deviations in subsequent evaluations. To improve the transferability of machine learning models within limited sample sets, this study proposes a weight transfer learning framework based on the similarity of the labels. The similarity weighting method includes both hard weights and soft weights. By evaluating the similarity between test and training sets of logging data, the similarity results are used to estimate the weights of training samples, thereby optimizing the model learning process. We develop a double experts' network and a bidirectional gated neural network based on hierarchical attention and multi-head attention (BiGRU-MHSA) for well logs reconstruction and lithofacies classification tasks. Oil field data results for the shale strata in the Gulong area of the Songliao Basin of China indicate that the double experts’ network model performs well in curve reconstruction tasks. However, it may not be effective in lithofacies classification tasks, while BiGRU-MHSA performs well in that area. In the study of constructing large-scale well logging processing and formation interpretation models, it is maybe more beneficial by employing different expert models for combined evaluations. In addition, although the improvement is limited, hard or soft weighting methods is better than unweighted (i.e., average-weighted) in significantly different adjacent wells. The code and data are open and available for subsequent studies on other lithofacies layers.</div></div>","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"5 ","pages":"Article 100091"},"PeriodicalIF":0.0,"publicationDate":"2024-09-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142323362","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhanced permeability prediction in porous media using particle swarm optimization with multi-source integration 利用多源集成的粒子群优化技术增强多孔介质的渗透性预测
Pub Date : 2024-09-16 DOI: 10.1016/j.aiig.2024.100090
Zhiping Chen , Jia Zhang , Daren Zhang , Xiaolin Chang , Wei Zhou

Accurately and efficiently predicting the permeability of porous media is essential for addressing a wide range of hydrogeological issues. However, the complexity of porous media often limits the effectiveness of individual prediction methods. This study introduces a novel Particle Swarm Optimization-based Permeability Integrated Prediction model (PSO-PIP), which incorporates a particle swarm optimization algorithm enhanced with dynamic clustering and adaptive parameter tuning (KGPSO). The model integrates multi-source data from the Lattice Boltzmann Method (LBM), Pore Network Modeling (PNM), and Finite Difference Method (FDM). By assigning optimal weight coefficients to the outputs of these methods, the model minimizes deviations from actual values and enhances permeability prediction performance. Initially, the computational performances of the LBM, PNM, and FDM are comparatively analyzed on datasets consisting of sphere packings and real rock samples. It is observed that these methods exhibit computational biases in certain permeability ranges. The PSO-PIP model is proposed to combine the strengths of each computational approach and mitigate their limitations. The PSO-PIP model consistently produces predictions that are highly congruent with actual permeability values across all prediction intervals, significantly enhancing prediction accuracy. The outcomes of this study provide a new tool and perspective for the comprehensive, rapid, and accurate prediction of permeability in porous media.

准确有效地预测多孔介质的渗透性对于解决各种水文地质问题至关重要。然而,多孔介质的复杂性往往限制了单个预测方法的有效性。本研究介绍了一种新颖的基于粒子群优化的渗透率综合预测模型(PSO-PIP),该模型采用了一种具有动态聚类和自适应参数调整(KGPSO)功能的粒子群优化算法。该模型整合了来自格子波尔兹曼法(LBM)、孔隙网络建模(PNM)和有限差分法(FDM)的多源数据。通过为这些方法的输出分配最佳权重系数,该模型最大限度地减少了与实际值的偏差,提高了渗透率预测性能。首先,在球状填料和实际岩石样本组成的数据集上对 LBM、PNM 和 FDM 的计算性能进行了比较分析。结果发现,这些方法在某些渗透率范围内存在计算偏差。我们提出了 PSO-PIP 模型,以结合每种计算方法的优势并减轻其局限性。在所有预测区间内,PSO-PIP 模型都能得出与实际渗透率值高度一致的预测结果,大大提高了预测精度。这项研究的成果为全面、快速、准确地预测多孔介质的渗透性提供了新的工具和视角。
{"title":"Enhanced permeability prediction in porous media using particle swarm optimization with multi-source integration","authors":"Zhiping Chen ,&nbsp;Jia Zhang ,&nbsp;Daren Zhang ,&nbsp;Xiaolin Chang ,&nbsp;Wei Zhou","doi":"10.1016/j.aiig.2024.100090","DOIUrl":"10.1016/j.aiig.2024.100090","url":null,"abstract":"<div><p>Accurately and efficiently predicting the permeability of porous media is essential for addressing a wide range of hydrogeological issues. However, the complexity of porous media often limits the effectiveness of individual prediction methods. This study introduces a novel Particle Swarm Optimization-based Permeability Integrated Prediction model (PSO-PIP), which incorporates a particle swarm optimization algorithm enhanced with dynamic clustering and adaptive parameter tuning (KGPSO). The model integrates multi-source data from the Lattice Boltzmann Method (LBM), Pore Network Modeling (PNM), and Finite Difference Method (FDM). By assigning optimal weight coefficients to the outputs of these methods, the model minimizes deviations from actual values and enhances permeability prediction performance. Initially, the computational performances of the LBM, PNM, and FDM are comparatively analyzed on datasets consisting of sphere packings and real rock samples. It is observed that these methods exhibit computational biases in certain permeability ranges. The PSO-PIP model is proposed to combine the strengths of each computational approach and mitigate their limitations. The PSO-PIP model consistently produces predictions that are highly congruent with actual permeability values across all prediction intervals, significantly enhancing prediction accuracy. The outcomes of this study provide a new tool and perspective for the comprehensive, rapid, and accurate prediction of permeability in porous media.</p></div>","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"5 ","pages":"Article 100090"},"PeriodicalIF":0.0,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2666544124000315/pdfft?md5=b1d09a2cb0aeba96843adc16601b4089&pid=1-s2.0-S2666544124000315-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142271080","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fully invertible hyperbolic neural networks for segmenting large-scale surface and sub-surface data 用于分割大规模地表和次地表数据的完全可逆双曲神经网络
Pub Date : 2024-08-24 DOI: 10.1016/j.aiig.2024.100087
Bas Peters , Eldad Haber , Keegan Lensink

The large spatial/temporal/frequency scale of geoscience and remote-sensing datasets causes memory issues when using convolutional neural networks for (sub-) surface data segmentation. Recently developed fully reversible or fully invertible networks can mostly avoid memory limitations by recomputing the states during the backward pass through the network. This results in a low and fixed memory requirement for storing network states, as opposed to the typical linear memory growth with network depth. This work focuses on a fully invertible network based on the telegraph equation. While reversibility saves the major amount of memory used in deep networks by the data, the convolutional kernels can take up most memory if fully invertible networks contain multiple invertible pooling/coarsening layers. We address the explosion of the number of convolutional kernels by combining fully invertible networks with layers that contain the convolutional kernels in a compressed form directly. A second challenge is that invertible networks output a tensor the same size as its input. This property prevents the straightforward application of invertible networks to applications that map between different input–output dimensions, need to map to outputs with more channels than present in the input data, or desire outputs that decrease/increase the resolution compared to the input data. However, we show that by employing invertible networks in a non-standard fashion, we can still use them for these tasks. Examples in hyperspectral land-use classification, airborne geophysical surveying, and seismic imaging illustrate that we can input large data volumes in one chunk and do not need to work on small patches, use dimensionality reduction, or employ methods that classify a patch to a single central pixel.

地球科学和遥感数据集的空间/时间/频率尺度较大,在使用卷积神经网络进行(次)表面数据分割时会产生内存问题。最近开发的完全可逆或完全可逆网络通过在网络后向传递过程中重新计算状态,在很大程度上避免了内存限制。这就使得存储网络状态所需的内存较低且固定,而不是典型的随网络深度线性增长的内存。这项工作的重点是基于电报方程的完全可逆网络。虽然可逆性节省了深度网络中数据所使用的大部分内存,但如果完全可逆网络包含多个可逆池/解析层,卷积核可能会占用大部分内存。我们通过将完全可逆网络与直接包含压缩形式卷积核的层结合起来,解决了卷积核数量激增的问题。第二个挑战是,可逆网络输出的张量与其输入大小相同。这一特性阻碍了可逆网络在以下应用中的直接应用:在不同的输入-输出维度之间进行映射,需要映射到比输入数据具有更多通道的输出,或者希望输出与输入数据相比降低/提高分辨率。不过,我们的研究表明,通过以非标准方式使用可逆网络,我们仍然可以将其用于这些任务。高光谱土地利用分类、机载地球物理勘测和地震成像中的例子说明,我们可以一次性输入大量数据,而无需处理小块数据、使用降维或采用将小块数据分类为单个中心像素的方法。
{"title":"Fully invertible hyperbolic neural networks for segmenting large-scale surface and sub-surface data","authors":"Bas Peters ,&nbsp;Eldad Haber ,&nbsp;Keegan Lensink","doi":"10.1016/j.aiig.2024.100087","DOIUrl":"10.1016/j.aiig.2024.100087","url":null,"abstract":"<div><p>The large spatial/temporal/frequency scale of geoscience and remote-sensing datasets causes memory issues when using convolutional neural networks for (sub-) surface data segmentation. Recently developed fully reversible or fully invertible networks can mostly avoid memory limitations by recomputing the states during the backward pass through the network. This results in a low and fixed memory requirement for storing network states, as opposed to the typical linear memory growth with network depth. This work focuses on a fully invertible network based on the telegraph equation. While reversibility saves the major amount of memory used in deep networks by the data, the convolutional kernels can take up most memory if fully invertible networks contain multiple invertible pooling/coarsening layers. We address the explosion of the number of convolutional kernels by combining fully invertible networks with layers that contain the convolutional kernels in a compressed form directly. A second challenge is that invertible networks output a tensor the same size as its input. This property prevents the straightforward application of invertible networks to applications that map between different input–output dimensions, need to map to outputs with more channels than present in the input data, or desire outputs that decrease/increase the resolution compared to the input data. However, we show that by employing invertible networks in a non-standard fashion, we can still use them for these tasks. Examples in hyperspectral land-use classification, airborne geophysical surveying, and seismic imaging illustrate that we can input large data volumes in one chunk and do not need to work on small patches, use dimensionality reduction, or employ methods that classify a patch to a single central pixel.</p></div>","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"5 ","pages":"Article 100087"},"PeriodicalIF":0.0,"publicationDate":"2024-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2666544124000285/pdfft?md5=aefb3645cc92ad5ad25d7d3f97a32057&pid=1-s2.0-S2666544124000285-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142087583","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
EQGraphNet: Advancing single-station earthquake magnitude estimation via deep graph networks with residual connections EQGraphNet:通过具有残差连接的深度图网络推进单站地震震级估算
Pub Date : 2024-08-22 DOI: 10.1016/j.aiig.2024.100089
Zhiguo Wang , Ziwei Chen , Huai Zhang

Magnitude estimation is a critical task in seismology, and conventional methods usually require dense seismic station arrays to provide data with sufficient spatiotemporal distribution. In this context, we propose the Earthquake Graph Network (EQGraphNet) to enhance the performance of single-station magnitude estimation. The backbone of the proposed model consists of eleven convolutional neural network layers and ten RCGL modules, where a RCGL combines a Residual Connection and a Graph convolutional Layer capable of mitigating the over-smoothing problem and simultaneously extracting temporal features of seismic signals. Our work uses the STanford EArthquake Dataset for model training and performance testing. Compared with three existing deep learning models, EQGraphNet demonstrates improved accuracy for both local magnitude and duration magnitude scales. To evaluate the robustness, we add natural background noise to the model input and find that EQGraphNet achieves the best results, particularly for signals with lower signal-to-noise ratios. Additionally, by replacing various network components and comparing their estimation performances, we illustrate the contribution of each part of EQGraphNet, validating the rationality of our approach. We also demonstrate the generalization capability of our model across different earthquakes occurring environments, achieving mean errors of ±0.1 units. Furthermore, by demonstrating the effectiveness of deeper architectures, this work encourages further exploration of deeper GNN models for both multi-station and single-station magnitude estimation.

震级估计是地震学中的一项关键任务,传统方法通常需要密集的地震台阵列来提供具有足够时空分布的数据。在这种情况下,我们提出了地震图网络(EQGraphNet)来提高单台站震级估计的性能。该模型的骨干由 11 个卷积神经网络层和 10 个 RCGL 模块组成,其中 RCGL 结合了残差连接和图卷积层,能够缓解过度平滑问题,同时提取地震信号的时间特征。我们的工作使用斯坦福大学地震数据集进行模型训练和性能测试。与现有的三个深度学习模型相比,EQGraphNet 在局部震级和持续时间震级尺度上的准确性都有所提高。为了评估鲁棒性,我们在模型输入中添加了自然背景噪声,结果发现 EQGraphNet 取得了最佳结果,尤其是对于信噪比较低的信号。此外,通过替换各种网络组件并比较其估计性能,我们说明了 EQGraphNet 各部分的贡献,验证了我们方法的合理性。我们还证明了我们的模型在不同地震发生环境下的泛化能力,其平均误差为 ±0.1 个单位。此外,通过证明更深层架构的有效性,这项工作鼓励进一步探索用于多站和单站震级估计的更深层 GNN 模型。
{"title":"EQGraphNet: Advancing single-station earthquake magnitude estimation via deep graph networks with residual connections","authors":"Zhiguo Wang ,&nbsp;Ziwei Chen ,&nbsp;Huai Zhang","doi":"10.1016/j.aiig.2024.100089","DOIUrl":"10.1016/j.aiig.2024.100089","url":null,"abstract":"<div><p>Magnitude estimation is a critical task in seismology, and conventional methods usually require dense seismic station arrays to provide data with sufficient spatiotemporal distribution. In this context, we propose the Earthquake Graph Network (EQGraphNet) to enhance the performance of single-station magnitude estimation. The backbone of the proposed model consists of eleven convolutional neural network layers and ten RCGL modules, where a RCGL combines a Residual Connection and a Graph convolutional Layer capable of mitigating the over-smoothing problem and simultaneously extracting temporal features of seismic signals. Our work uses the STanford EArthquake Dataset for model training and performance testing. Compared with three existing deep learning models, EQGraphNet demonstrates improved accuracy for both local magnitude and duration magnitude scales. To evaluate the robustness, we add natural background noise to the model input and find that EQGraphNet achieves the best results, particularly for signals with lower signal-to-noise ratios. Additionally, by replacing various network components and comparing their estimation performances, we illustrate the contribution of each part of EQGraphNet, validating the rationality of our approach. We also demonstrate the generalization capability of our model across different earthquakes occurring environments, achieving mean errors of <span><math><mo>±</mo></math></span>0.1 units. Furthermore, by demonstrating the effectiveness of deeper architectures, this work encourages further exploration of deeper GNN models for both multi-station and single-station magnitude estimation.</p></div>","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"5 ","pages":"Article 100089"},"PeriodicalIF":0.0,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2666544124000303/pdfft?md5=c749cb3c43c7b43360a083719cc07ae7&pid=1-s2.0-S2666544124000303-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142049374","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Wide & deep learning for predicting relative mineral compositions of sediment cores solely based on XRF scans, a case study from Pleistocene Paleolake Olduvai, Tanzania 完全基于 XRF 扫描预测沉积岩芯相对矿物成分的广泛和深度学习,坦桑尼亚更新世古湖奥杜威案例研究
Pub Date : 2024-08-22 DOI: 10.1016/j.aiig.2024.100088
Gayantha R.L. Kodikara , Lindsay J. McHenry , Ian G. Stanistreet , Harald Stollhofen , Jackson K. Njau , Nicholas Toth , Kathy Schick

This study develops a method to use deep learning models to predict the mineral assemblages and their relative abundances in paleolake cores using high-resolution XRF core scan elemental data and X-ray diffraction (XRD) mineralogical results from the same core taken at coarser resolution. It uses the XRF core scan data along with published mineralogical information from the Olduvai Gorge Coring Project (OGCP) 2014 sediment cores 1A, 2A, and 3A from Paleolake Olduvai, Tanzania. Both regression and classification models were developed using a Keras deep learning framework to assess the predictability of mineral assemblages with their relative abundances (in regression models) or at least the mineral assemblages (in classification models) using XRF core scan data. Models were created using the Sequential class and Functional API with different model architectures. The correlation matrix of element ratios calculated from XRF element intensity records from the cores and XRD-derived mineralogical information was used to select the most useful features to train the models. 1057 training data records were used for the models. Lithological classes were also used for some models using Wide & Deep neural networks since those combine the benefits of memorization and generalization for mineral prediction. The results were validated using 265 validation data records unseen by the model and discuss the accuracy of models using six test records. The optimized Deep Neural Network (DNN) classification model achieved over 86% binary accuracy while the regression models were also able to predict the relative mineral abundances of samples with high accuracies. Overall, the study shows the efficacy of a carefully crafted Deep Learning (DL) model for predicting mineral assemblages and abundances using high-resolution XRF core scan data.

本研究开发了一种方法,利用高分辨率 XRF 岩心扫描元素数据和同一岩心较粗分辨率的 X 射线衍射 (XRD) 矿物学结果,使用深度学习模型预测古湖岩心中的矿物组合及其相对丰度。该研究使用了 XRF 岩心扫描数据以及已公布的来自坦桑尼亚奥杜威峡谷岩芯采集项目(OGCP)2014 年奥杜威古湖 1A、2A 和 3A 号沉积岩芯的矿物学信息。我们使用 Keras 深度学习框架开发了回归和分类模型,以评估使用 XRF 岩心扫描数据的矿物组合及其相对丰度(在回归模型中)或至少矿物组合(在分类模型中)的可预测性。使用具有不同模型结构的序列类和功能应用程序接口创建了模型。根据岩心的 XRF 元素强度记录和 XRD 衍生矿物学信息计算出的元素比率相关矩阵用于选择最有用的特征来训练模型。模型使用了 1057 条训练数据记录。由于深度神经网络结合了矿物预测的记忆性和概括性优势,因此一些模型还使用了宽amp;深度神经网络的岩性类别。使用模型未见过的 265 条验证数据对结果进行了验证,并使用 6 条测试数据对模型的准确性进行了讨论。经过优化的深度神经网络(DNN)分类模型达到了 86% 以上的二元准确率,而回归模型也能以较高的准确率预测样本的相对矿物丰度。总之,这项研究表明,精心设计的深度学习(DL)模型能够有效地利用高分辨率 XRF 岩心扫描数据预测矿物组合和丰度。
{"title":"Wide & deep learning for predicting relative mineral compositions of sediment cores solely based on XRF scans, a case study from Pleistocene Paleolake Olduvai, Tanzania","authors":"Gayantha R.L. Kodikara ,&nbsp;Lindsay J. McHenry ,&nbsp;Ian G. Stanistreet ,&nbsp;Harald Stollhofen ,&nbsp;Jackson K. Njau ,&nbsp;Nicholas Toth ,&nbsp;Kathy Schick","doi":"10.1016/j.aiig.2024.100088","DOIUrl":"10.1016/j.aiig.2024.100088","url":null,"abstract":"<div><p>This study develops a method to use deep learning models to predict the mineral assemblages and their relative abundances in paleolake cores using high-resolution XRF core scan elemental data and X-ray diffraction (XRD) mineralogical results from the same core taken at coarser resolution. It uses the XRF core scan data along with published mineralogical information from the Olduvai Gorge Coring Project (OGCP) 2014 sediment cores 1A, 2A, and 3A from Paleolake Olduvai, Tanzania. Both regression and classification models were developed using a Keras deep learning framework to assess the predictability of mineral assemblages with their relative abundances (in regression models) or at least the mineral assemblages (in classification models) using XRF core scan data. Models were created using the Sequential class and Functional API with different model architectures. The correlation matrix of element ratios calculated from XRF element intensity records from the cores and XRD-derived mineralogical information was used to select the most useful features to train the models. 1057 training data records were used for the models. Lithological classes were also used for some models using Wide &amp; Deep neural networks since those combine the benefits of memorization and generalization for mineral prediction. The results were validated using 265 validation data records unseen by the model and discuss the accuracy of models using six test records. The optimized Deep Neural Network (DNN) classification model achieved over 86% binary accuracy while the regression models were also able to predict the relative mineral abundances of samples with high accuracies. Overall, the study shows the efficacy of a carefully crafted Deep Learning (DL) model for predicting mineral assemblages and abundances using high-resolution XRF core scan data.</p></div>","PeriodicalId":100124,"journal":{"name":"Artificial Intelligence in Geosciences","volume":"5 ","pages":"Article 100088"},"PeriodicalIF":0.0,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2666544124000297/pdfft?md5=8bd7402c96f4a311b4dbf3ffa0c2ef1b&pid=1-s2.0-S2666544124000297-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142075965","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Artificial Intelligence in Geosciences
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1