Pub Date : 2024-01-09DOI: 10.1007/s11004-023-10121-6
Abstract
We introduce a new methodology for inference of fluid composition from measurements of mineralogical or chemical compositions, expanding upon the use of reactive transport models to understand hydrothermal alteration processes. The reactive transport models are used to impute a latent variable explanatory mechanism in the formation of hydrothermal alteration zones and mineral deposits. An expectation maximisation algorithm is then employed to solve the joint problem of identifying alteration zones in the measured data and estimating the fluid composition, based on the fit between the mineral abundances in the measured and predicted alteration zones. Using the hydrothermal alteration of granite as a test case (greisenisation), a range of synthetic tests is presented to illustrate how the methodology enables objective inference of the mineralising fluid. For field data from the East Kemptville tin deposit in Nova Scotia, the technique generates inferences for the fluid composition which compare favourably with previous independent estimates, demonstrating the feasibility of the proposed calibration methodology.
{"title":"Towards a Model-Based Interpretation of Measurements of Mineralogical and Chemical Compositions","authors":"","doi":"10.1007/s11004-023-10121-6","DOIUrl":"https://doi.org/10.1007/s11004-023-10121-6","url":null,"abstract":"<h3>Abstract</h3> <p>We introduce a new methodology for inference of fluid composition from measurements of mineralogical or chemical compositions, expanding upon the use of reactive transport models to understand hydrothermal alteration processes. The reactive transport models are used to impute a latent variable explanatory mechanism in the formation of hydrothermal alteration zones and mineral deposits. An expectation maximisation algorithm is then employed to solve the joint problem of identifying alteration zones in the measured data and estimating the fluid composition, based on the fit between the mineral abundances in the measured and predicted alteration zones. Using the hydrothermal alteration of granite as a test case (greisenisation), a range of synthetic tests is presented to illustrate how the methodology enables objective inference of the mineralising fluid. For field data from the East Kemptville tin deposit in Nova Scotia, the technique generates inferences for the fluid composition which compare favourably with previous independent estimates, demonstrating the feasibility of the proposed calibration methodology.</p>","PeriodicalId":51117,"journal":{"name":"Mathematical Geosciences","volume":null,"pages":null},"PeriodicalIF":2.6,"publicationDate":"2024-01-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139411506","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-08DOI: 10.1007/s11004-023-10127-0
Yuming Ba, Dean S. Oliver
Because it is generally impossible to completely characterize the uncertainty in complex model variables after assimilation of data, it is common to approximate the uncertainty by sampling from approximations of the posterior distribution for model variables. When minimization methods are used for the sampling, the weights on each of the samples depend on the magnitude of the data mismatch at the critical points and on the Jacobian of the transformation from the prior density to the sample proposal density. For standard iterative ensemble smoothers, the Jacobian is identical for all samples, and the weights depend only on the data mismatch. In this paper, a hybrid data assimilation method is proposed which makes it possible for each ensemble member to have a distinct Jacobian and for the approximation to the posterior density to be multimodal. For the proposed hybrid iterative ensemble smoother, it is necessary that a part of the mapping from the prior Gaussian random variable to the data be analytic. Examples might include analytic transformation from a latent Gaussian random variable to permeability followed by a black-box transformation from permeability to state variables in porous media flow, or a Gaussian hierarchical model for variables followed by a similar black-box transformation from permeability to state variables. In this paper, the application of weighting to both hybrid and standard iterative ensemble smoothers is investigated using a two-dimensional, two-phase flow problem in porous media with various degrees of nonlinearity. As expected, the weights in a standard iterative ensemble smoother become degenerate for problems with large amounts of data. In the examples, however, the weights for the hybrid iterative ensemble smoother were useful for improving forecast reliability.
{"title":"Importance Weighting in Hybrid Iterative Ensemble Smoothers for Data Assimilation","authors":"Yuming Ba, Dean S. Oliver","doi":"10.1007/s11004-023-10127-0","DOIUrl":"https://doi.org/10.1007/s11004-023-10127-0","url":null,"abstract":"<p>Because it is generally impossible to completely characterize the uncertainty in complex model variables after assimilation of data, it is common to approximate the uncertainty by sampling from approximations of the posterior distribution for model variables. When minimization methods are used for the sampling, the weights on each of the samples depend on the magnitude of the data mismatch at the critical points and on the Jacobian of the transformation from the prior density to the sample proposal density. For standard iterative ensemble smoothers, the Jacobian is identical for all samples, and the weights depend only on the data mismatch. In this paper, a hybrid data assimilation method is proposed which makes it possible for each ensemble member to have a distinct Jacobian and for the approximation to the posterior density to be multimodal. For the proposed hybrid iterative ensemble smoother, it is necessary that a part of the mapping from the prior Gaussian random variable to the data be analytic. Examples might include analytic transformation from a latent Gaussian random variable to permeability followed by a black-box transformation from permeability to state variables in porous media flow, or a Gaussian hierarchical model for variables followed by a similar black-box transformation from permeability to state variables. In this paper, the application of weighting to both hybrid and standard iterative ensemble smoothers is investigated using a two-dimensional, two-phase flow problem in porous media with various degrees of nonlinearity. As expected, the weights in a standard iterative ensemble smoother become degenerate for problems with large amounts of data. In the examples, however, the weights for the hybrid iterative ensemble smoother were useful for improving forecast reliability.</p>","PeriodicalId":51117,"journal":{"name":"Mathematical Geosciences","volume":null,"pages":null},"PeriodicalIF":2.6,"publicationDate":"2024-01-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139397498","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-03DOI: 10.1007/s11004-023-10125-2
Herbert Rakotonirina, Ignacio Guridi, Paul Honeine, Olivier Atteia, Antonin Van Exem
Kriging is the most widely used spatial interpolation method in geostatistics. For many environmental applications, kriging may have to satisfy the stationarity and isotropy hypothesis, and new techniques using machine learning suffer from a lack of labeled data. In this paper, we propose the use of deep image prior, which is a U-net-like deep neural network designed for image reconstruction, to perform spatial interpolation and conditional map generation without any prior learning. This approach allows us to overcome the assumptions for kriging as well as the lack of labeled data when proposing uncertainty and probability above a certain threshold. The proposed method is based on a convolutional neural network that generates a map from random values by minimizing the difference between the output map and the observed values. With this new method of spatial interpolation, we generate n maps to obtain a map of uncertainty and a map of probability of exceeding the threshold. Experiments demonstrate the relevance of the proposed methods for spatial interpolation on both the well-known digital elevation model data and the more challenging case of pollution mapping. The results obtained with the three datasets demonstrate competitive performance compared with state-of-the-art methods.
克里金法是地质统计学中应用最广泛的空间插值方法。对于许多环境应用来说,克里金法可能必须满足静止性和各向同性假设,而使用机器学习的新技术又受到缺乏标记数据的困扰。在本文中,我们提出使用深度图像先验(一种专为图像重建设计的类 U-net 深度神经网络)来执行空间插值和条件地图生成,而无需任何先验学习。这种方法使我们能够克服克里金法的假设,以及在提出不确定性和概率超过一定阈值时缺乏标记数据的问题。所提出的方法以卷积神经网络为基础,通过最小化输出地图与观测值之间的差异,从随机值生成地图。利用这种新的空间插值方法,我们生成了 n 幅地图,从而获得了不确定性地图和超过阈值的概率地图。实验证明,无论是在众所周知的数字高程模型数据上,还是在更具挑战性的污染地图绘制上,所提出的空间插值方法都非常实用。与最先进的方法相比,使用这三种数据集获得的结果表明了具有竞争力的性能。
{"title":"Spatial Interpolation and Conditional Map Generation Using Deep Image Prior for Environmental Applications","authors":"Herbert Rakotonirina, Ignacio Guridi, Paul Honeine, Olivier Atteia, Antonin Van Exem","doi":"10.1007/s11004-023-10125-2","DOIUrl":"https://doi.org/10.1007/s11004-023-10125-2","url":null,"abstract":"<p>Kriging is the most widely used spatial interpolation method in geostatistics. For many environmental applications, kriging may have to satisfy the stationarity and isotropy hypothesis, and new techniques using machine learning suffer from a lack of labeled data. In this paper, we propose the use of deep image prior, which is a U-net-like deep neural network designed for image reconstruction, to perform spatial interpolation and conditional map generation without any prior learning. This approach allows us to overcome the assumptions for kriging as well as the lack of labeled data when proposing uncertainty and probability above a certain threshold. The proposed method is based on a convolutional neural network that generates a map from random values by minimizing the difference between the output map and the observed values. With this new method of spatial interpolation, we generate <i>n</i> maps to obtain a map of uncertainty and a map of probability of exceeding the threshold. Experiments demonstrate the relevance of the proposed methods for spatial interpolation on both the well-known digital elevation model data and the more challenging case of pollution mapping. The results obtained with the three datasets demonstrate competitive performance compared with state-of-the-art methods.\u0000</p>","PeriodicalId":51117,"journal":{"name":"Mathematical Geosciences","volume":null,"pages":null},"PeriodicalIF":2.6,"publicationDate":"2024-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139102929","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Tight sandstone reservoirs have low porosity, low permeability, and a complex pore structure. The seepage from tight sandstones is a key factor in evaluating the oil and gas accumulation in these reservoirs. Therefore, reservoir permeability prediction has become the focus of researchers. Using nuclear magnetic resonance (NMR), high-pressure mercury injection, scanning electron microscopy, and other experimental methods, scholars have established various permeability prediction models, which have obvious advantages and disadvantages. However, there is less research conducted on predicting the permeability of tight sandstone reservoirs according to their single-peak NMR T2 distribution. Based on NMR experiments and the bimodal Gaussian density formula, this study identified the criteria for determining the types of reservoir pore structures with single-peak NMR T2 distribution and established the parameters (η1 and η2) that can be used in the evaluation of reservoir pore structure. A novel model for predicting the permeability of tight sandstone reservoirs was established using η1 and η2. The results of the prediction model proposed in this study were found to be superior to the results of eight permeability prediction models established by other scholars in the studied case of the Huangliu Formation. However, permeability prediction models established using the NMR experimental results of different sources were found to be ineffective. Additionally, the new model is suitable for use with sandstone reservoirs with both single-peak and double-peak NMR T2 distributions in the studied case of the Yanchang Formation. Logging curves can be used to predict η1 and η2, and the permeability of a single well of a tight sandstone reservoir. The study findings would be useful for predicting tight sandstone reservoir permeability.
{"title":"A Permeability Prediction Model of Single-Peak NMR T2 Distribution in Tight Sandstones: A Case Study on the Huangliu Formation, Yinggehai Basin, China","authors":"Jing Zhao, Zhilong Huang, Jin Dong, Jingyuan Zhang, Rui Wang, Chonglin Ma, Guangjun Deng, Maguang Xu","doi":"10.1007/s11004-023-10118-1","DOIUrl":"https://doi.org/10.1007/s11004-023-10118-1","url":null,"abstract":"<p>Tight sandstone reservoirs have low porosity, low permeability, and a complex pore structure. The seepage from tight sandstones is a key factor in evaluating the oil and gas accumulation in these reservoirs. Therefore, reservoir permeability prediction has become the focus of researchers. Using nuclear magnetic resonance (NMR), high-pressure mercury injection, scanning electron microscopy, and other experimental methods, scholars have established various permeability prediction models, which have obvious advantages and disadvantages. However, there is less research conducted on predicting the permeability of tight sandstone reservoirs according to their single-peak NMR <i>T</i><sub>2</sub> distribution. Based on NMR experiments and the bimodal Gaussian density formula, this study identified the criteria for determining the types of reservoir pore structures with single-peak NMR <i>T</i><sub>2</sub> distribution and established the parameters (<i>η</i><sub>1</sub> and <i>η</i><sub>2</sub>) that can be used in the evaluation of reservoir pore structure. A novel model for predicting the permeability of tight sandstone reservoirs was established using <i>η</i><sub>1</sub> and <i>η</i><sub>2</sub>. The results of the prediction model proposed in this study were found to be superior to the results of eight permeability prediction models established by other scholars in the studied case of the Huangliu Formation. However, permeability prediction models established using the NMR experimental results of different sources were found to be ineffective. Additionally, the new model is suitable for use with sandstone reservoirs with both single-peak and double-peak NMR <i>T</i><sub>2</sub> distributions in the studied case of the Yanchang Formation. Logging curves can be used to predict <i>η</i><sub>1</sub> and <i>η</i><sub>2</sub>, and the permeability of a single well of a tight sandstone reservoir. The study findings would be useful for predicting tight sandstone reservoir permeability.</p>","PeriodicalId":51117,"journal":{"name":"Mathematical Geosciences","volume":null,"pages":null},"PeriodicalIF":2.6,"publicationDate":"2024-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139082663","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-02DOI: 10.1007/s11004-023-10120-7
João Lino Pereira, Emmanouil A. Varouchakis, George P. Karatzas, Leonardo Azevedo
Groundwater resources in Mediterranean coastal aquifers are under several threats including saltwater intrusion. This situation is exacerbated by the absence of sustainable management plans for groundwater resources. Management and monitoring of groundwater systems require an integrated approach and the joint interpretation of any available information. This work investigates how uncertainty can be integrated within the geo-modelling workflow when creating numerical three-dimensional aquifer models with electrical resistivity borehole logs, geostatistical simulation and Bayesian model averaging. Multiple geological scenarios of electrical resistivity are created with geostatistical simulation by removing one borehole at a time from the set of available boreholes. To account for the spatial uncertainty simultaneously reflected by the multiple geostatistical scenarios, Bayesian model averaging is used to combine the probability distribution functions of each scenario into a global one, thus providing more credible uncertainty intervals. The proposed methodology is applied to a water-stressed groundwater system located in Crete that is threatened by saltwater intrusion. The results obtained agree with the general knowledge of this complex environment and enable sustainable groundwater management policies to be devised considering optimistic and pessimistic scenarios.
{"title":"Uncertainty Quantification in Geostatistical Modelling of Saltwater Intrusion at a Coastal Aquifer System","authors":"João Lino Pereira, Emmanouil A. Varouchakis, George P. Karatzas, Leonardo Azevedo","doi":"10.1007/s11004-023-10120-7","DOIUrl":"https://doi.org/10.1007/s11004-023-10120-7","url":null,"abstract":"<p>Groundwater resources in Mediterranean coastal aquifers are under several threats including saltwater intrusion. This situation is exacerbated by the absence of sustainable management plans for groundwater resources. Management and monitoring of groundwater systems require an integrated approach and the joint interpretation of any available information. This work investigates how uncertainty can be integrated within the geo-modelling workflow when creating numerical three-dimensional aquifer models with electrical resistivity borehole logs, geostatistical simulation and Bayesian model averaging. Multiple geological scenarios of electrical resistivity are created with geostatistical simulation by removing one borehole at a time from the set of available boreholes. To account for the spatial uncertainty simultaneously reflected by the multiple geostatistical scenarios, Bayesian model averaging is used to combine the probability distribution functions of each scenario into a global one, thus providing more credible uncertainty intervals. The proposed methodology is applied to a water-stressed groundwater system located in Crete that is threatened by saltwater intrusion. The results obtained agree with the general knowledge of this complex environment and enable sustainable groundwater management policies to be devised considering optimistic and pessimistic scenarios.</p>","PeriodicalId":51117,"journal":{"name":"Mathematical Geosciences","volume":null,"pages":null},"PeriodicalIF":2.6,"publicationDate":"2024-01-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139078232","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-01Epub Date: 2023-06-02DOI: 10.1007/s11004-023-10064-y
Lea Friedli, Niklas Linde
A geophysical Bayesian inversion problem may target the posterior distribution of geological or hydrogeological parameters given geophysical data. To account for the scatter in the petrophysical relationship linking the target parameters to the geophysical properties, this study treats the intermediate geophysical properties as latent (unobservable) variables. To perform inversion in such a latent variable model, the intractable likelihood function of the (hydro)geological parameters given the geophysical data needs to be estimated. This can be achieved by approximation with a Gaussian probability density function based on local linearization of the geophysical forward operator, thereby, accounting for the noise in the petrophysical relationship by a corresponding addition to the data covariance matrix. The new approximate method is compared against the general correlated pseudo-marginal method, which estimates the likelihood by Monte Carlo averaging over samples of the latent variable. First, the performances of the two methods are tested on a synthetic test example, in which a multivariate Gaussian porosity field is inferred using crosshole ground-penetrating radar first-arrival travel times. For this example with rather small petrophysical uncertainty, the two methods provide near-identical estimates, while an inversion that ignores petrophysical uncertainty leads to biased estimates. The results of a sensitivity analysis are then used to suggest that the linearized Gaussian approach, while attractive due to its relative computational speed, suffers from a decreasing accuracy with increasing scatter in the petrophysical relationship. The computationally more expensive correlated pseudo-marginal method performs very well even for settings with high petrophysical uncertainty.
{"title":"Solving Geophysical Inversion Problems with Intractable Likelihoods: Linearized Gaussian Approximations Versus the Correlated Pseudo-marginal Method.","authors":"Lea Friedli, Niklas Linde","doi":"10.1007/s11004-023-10064-y","DOIUrl":"10.1007/s11004-023-10064-y","url":null,"abstract":"<p><p>A geophysical Bayesian inversion problem may target the posterior distribution of geological or hydrogeological parameters given geophysical data. To account for the scatter in the petrophysical relationship linking the target parameters to the geophysical properties, this study treats the intermediate geophysical properties as latent (unobservable) variables. To perform inversion in such a latent variable model, the intractable likelihood function of the (hydro)geological parameters given the geophysical data needs to be estimated. This can be achieved by approximation with a Gaussian probability density function based on local linearization of the geophysical forward operator, thereby, accounting for the noise in the petrophysical relationship by a corresponding addition to the data covariance matrix. The new approximate method is compared against the general correlated pseudo-marginal method, which estimates the likelihood by Monte Carlo averaging over samples of the latent variable. First, the performances of the two methods are tested on a synthetic test example, in which a multivariate Gaussian porosity field is inferred using crosshole ground-penetrating radar first-arrival travel times. For this example with rather small petrophysical uncertainty, the two methods provide near-identical estimates, while an inversion that ignores petrophysical uncertainty leads to biased estimates. The results of a sensitivity analysis are then used to suggest that the linearized Gaussian approach, while attractive due to its relative computational speed, suffers from a decreasing accuracy with increasing scatter in the petrophysical relationship. The computationally more expensive correlated pseudo-marginal method performs very well even for settings with high petrophysical uncertainty.</p>","PeriodicalId":51117,"journal":{"name":"Mathematical Geosciences","volume":null,"pages":null},"PeriodicalIF":2.6,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10817994/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79619150","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-19DOI: 10.1007/s11004-023-10122-5
Manfred Mudelsee
The linear calibration model is a powerful statistical tool that can be utilized to predict an unknown response variable, Y, through observations of a proxy or predictor variable, X. Since calibration involves estimation of regression model parameters on the basis of a limited amount of noisy data, an unbiased calibration slope estimation is of utmost importance. This can be achieved by means of state-of-the-art, data-driven statistical techniques. The present paper shows that weighted least-squares for both variables estimation (WLSXY) is able to deliver unbiased slope estimations under heteroscedasticity. In the case of homoscedasticity, besides WLSXY, ordinary least-squares (OLS) estimation with bias correction (OLSBC) also performs well. For achieving unbiasedness, it is further necessary to take the correct regression direction (i.e., of Y on X) into account. The present paper introduces a pairwise moving block bootstrap resampling approach for obtaining accurate estimation confidence intervals (CIs) under real-world climate conditions (i.e., non-Gaussian distributional shapes and autocorrelations in the noise components). A Monte Carlo simulation experiment confirms the feasibility and validity of this approach. The parameter estimates and bootstrap replications serve to predict the response with CIs. The methodological approach to unbiased calibration is illustrated for a paired time series dataset of sea-surface temperature and coral oxygen isotopic composition. Fortran software with implementation of OLSBC and WLSXY accompanies this paper.
线性校准模型是一种强大的统计工具,可用于通过观测替代变量或预测变量 X 来预测未知响应变量 Y。这可以通过最先进的数据驱动统计技术来实现。本文表明,双变量加权最小二乘法估计(WLSXY)能够在异方差情况下提供无偏的斜率估计。在同方差情况下,除 WLSXY 外,带偏差修正的普通最小二乘法(OLS)估计(OLSBC)也有很好的表现。为了实现无偏,还需要考虑正确的回归方向(即 Y 对 X 的回归方向)。本文介绍了一种成对移动块引导重采样方法,用于在实际气候条件下(即噪声成分的非高斯分布形状和自相关性)获得准确的估计置信区间(CI)。蒙特卡罗模拟实验证实了这种方法的可行性和有效性。参数估计和引导复制可用于预测具有 CIs 的响应。以海面温度和珊瑚氧同位素组成的成对时间序列数据集为例,说明了无偏校准的方法。本文附有实现 OLSBC 和 WLSXY 的 Fortran 软件。
{"title":"Unbiased Proxy Calibration","authors":"Manfred Mudelsee","doi":"10.1007/s11004-023-10122-5","DOIUrl":"https://doi.org/10.1007/s11004-023-10122-5","url":null,"abstract":"<p>The linear calibration model is a powerful statistical tool that can be utilized to predict an unknown response variable, <i>Y</i>, through observations of a proxy or predictor variable, <i>X</i>. Since calibration involves estimation of regression model parameters on the basis of a limited amount of noisy data, an unbiased calibration slope estimation is of utmost importance. This can be achieved by means of state-of-the-art, data-driven statistical techniques. The present paper shows that weighted least-squares for both variables estimation (WLSXY) is able to deliver unbiased slope estimations under heteroscedasticity. In the case of homoscedasticity, besides WLSXY, ordinary least-squares (OLS) estimation with bias correction (OLSBC) also performs well. For achieving unbiasedness, it is further necessary to take the correct regression direction (i.e., of <i>Y</i> on <i>X</i>) into account. The present paper introduces a pairwise moving block bootstrap resampling approach for obtaining accurate estimation confidence intervals (CIs) under real-world climate conditions (i.e., non-Gaussian distributional shapes and autocorrelations in the noise components). A Monte Carlo simulation experiment confirms the feasibility and validity of this approach. The parameter estimates and bootstrap replications serve to predict the response with CIs. The methodological approach to unbiased calibration is illustrated for a paired time series dataset of sea-surface temperature and coral oxygen isotopic composition. Fortran software with implementation of OLSBC and WLSXY accompanies this paper.</p>","PeriodicalId":51117,"journal":{"name":"Mathematical Geosciences","volume":null,"pages":null},"PeriodicalIF":2.6,"publicationDate":"2023-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138741196","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-12-19DOI: 10.1007/s11004-023-10124-3
Artur Posenato Garcia, Zoya Heidari
A thorough understanding of the interplay between polarization mechanisms is pivotal for the interpretation of electrical measurements, since sub-megahertz electrical measurements in sedimentary rocks are dominated by interfacial polarization mechanisms. Nonetheless, rock-physics models oversimplify pore-network geometry and the interaction of electric double layers relating to adjacent grains. Numerical algorithms present the best possible framework in which to characterize the electrical response of sedimentary rocks, avoiding the constraints intrinsic to rock-physics models. Recently, an algorithm was introduced that can simulate the interactions of electric fields with the ions in solution. The sub-kilohertz permittivity enhancement in sedimentary rocks is dominated by Stern-layer polarization, but a model for the polarization mechanism associated with the Stern layer has not been developed. Hence, the aim of this paper is to develop and test a numerical simulation framework to quantify the influence of Stern- and diffuse-layer polarization, temperature, ion concentration, and pore-network geometry on multi-frequency complex electrical measurements. The algorithm numerically solves the Poisson–Nernst–Planck equations in the time domain and a mineral-dependent electrochemical adsorption/desorption equilibrium model to determine surface charge distribution. Then, the numerical simulator is utilized to perform a sensitivity analysis to quantify the influence of electrolyte and interfacial properties on the permittivity of pore-scale samples at different frequencies.
{"title":"A New Numerical Simulation Framework to Model the Electric Interfacial Polarization Effects and Corresponding Impacts on Complex Dielectric Permittivity Measurements in Sedimentary Rocks","authors":"Artur Posenato Garcia, Zoya Heidari","doi":"10.1007/s11004-023-10124-3","DOIUrl":"https://doi.org/10.1007/s11004-023-10124-3","url":null,"abstract":"<p>A thorough understanding of the interplay between polarization mechanisms is pivotal for the interpretation of electrical measurements, since sub-megahertz electrical measurements in sedimentary rocks are dominated by interfacial polarization mechanisms. Nonetheless, rock-physics models oversimplify pore-network geometry and the interaction of electric double layers relating to adjacent grains. Numerical algorithms present the best possible framework in which to characterize the electrical response of sedimentary rocks, avoiding the constraints intrinsic to rock-physics models. Recently, an algorithm was introduced that can simulate the interactions of electric fields with the ions in solution. The sub-kilohertz permittivity enhancement in sedimentary rocks is dominated by Stern-layer polarization, but a model for the polarization mechanism associated with the Stern layer has not been developed. Hence, the aim of this paper is to develop and test a numerical simulation framework to quantify the influence of Stern- and diffuse-layer polarization, temperature, ion concentration, and pore-network geometry on multi-frequency complex electrical measurements. The algorithm numerically solves the Poisson–Nernst–Planck equations in the time domain and a mineral-dependent electrochemical adsorption/desorption equilibrium model to determine surface charge distribution. Then, the numerical simulator is utilized to perform a sensitivity analysis to quantify the influence of electrolyte and interfacial properties on the permittivity of pore-scale samples at different frequencies.</p>","PeriodicalId":51117,"journal":{"name":"Mathematical Geosciences","volume":null,"pages":null},"PeriodicalIF":2.6,"publicationDate":"2023-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138741194","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Machine learning (ML)-based landslide susceptibility mapping (LSM) has achieved substantial success in landslide risk management applications. However, the complexity of classically trained ML is often beyond nonexperts. With the rapid growth of practical applications, an “off-the-shelf” ML technique that can be easily used by nonexperts is highly relevant. In the present study, a new paradigm for an end-to-end ML modeling was adopted for LSM in the Three Gorges Reservoir area (TGRA) using automated machine learning (AutoML) as the backend model support for the paradigm. A well-defined database consisting of data from 290 landslides and 11 conditioning factors was collected for implementing AutoML and compared with classically trained ML approaches. The stacked ensemble model from AutoML achieved the best performance (0.954), surpassing the support vector machine with artificial bee colony optimization (ABC-SVM, 0.931), gray wolf optimization (GWO-SVM, 0.925), particle swarm optimization (PSO-SVM, 0.925), water cycle algorithm (WCA-SVM, 0.925), grid search (GS-SVM, 0.920), multilayer perceptron (MLP, 0.908), classification and regression tree (CART, 0.891), K-nearest neighbor (KNN, 0.898), and random forest (RF, 0.909) in terms of the area under the receiver operating characteristic curve (AUC). Notable improvements of up to 11% in AUC demonstrate that the AutoML approach succeeded in LSM and could be used to select the best model with minimal effort or intervention from the user. Moreover, a simple model that has been customarily ignored by practitioners and researchers has been identified with performance satisfying practical requirements. The experimental results indicate that AutoML provides an attractive alternative to manual ML practice, especially for practitioners with little expert knowledge in ML, by delivering a high-performance off-the-shelf solution for ML model development for LSM.
{"title":"Automated Machine Learning-Based Landslide Susceptibility Mapping for the Three Gorges Reservoir Area, China","authors":"Junwei Ma, Dongze Lei, Zhiyuan Ren, Chunhai Tan, Ding Xia, Haixiang Guo","doi":"10.1007/s11004-023-10116-3","DOIUrl":"https://doi.org/10.1007/s11004-023-10116-3","url":null,"abstract":"<p>Machine learning (ML)-based landslide susceptibility mapping (LSM) has achieved substantial success in landslide risk management applications. However, the complexity of classically trained ML is often beyond nonexperts. With the rapid growth of practical applications, an “off-the-shelf” ML technique that can be easily used by nonexperts is highly relevant. In the present study, a new paradigm for an end-to-end ML modeling was adopted for LSM in the Three Gorges Reservoir area (TGRA) using automated machine learning (AutoML) as the backend model support for the paradigm. A well-defined database consisting of data from 290 landslides and 11 conditioning factors was collected for implementing AutoML and compared with classically trained ML approaches. The stacked ensemble model from AutoML achieved the best performance (0.954), surpassing the support vector machine with artificial bee colony optimization (ABC-SVM, 0.931), gray wolf optimization (GWO-SVM, 0.925), particle swarm optimization (PSO-SVM, 0.925), water cycle algorithm (WCA-SVM, 0.925), grid search (GS-SVM, 0.920), multilayer perceptron (MLP, 0.908), classification and regression tree (CART, 0.891), K-nearest neighbor (KNN, 0.898), and random forest (RF, 0.909) in terms of the area under the receiver operating characteristic curve (AUC). Notable improvements of up to 11% in AUC demonstrate that the AutoML approach succeeded in LSM and could be used to select the best model with minimal effort or intervention from the user. Moreover, a simple model that has been customarily ignored by practitioners and researchers has been identified with performance satisfying practical requirements. The experimental results indicate that AutoML provides an attractive alternative to manual ML practice, especially for practitioners with little expert knowledge in ML, by delivering a high-performance off-the-shelf solution for ML model development for LSM.</p>","PeriodicalId":51117,"journal":{"name":"Mathematical Geosciences","volume":null,"pages":null},"PeriodicalIF":2.6,"publicationDate":"2023-11-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"138514896","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}