首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 477 毫秒
1.
A Bayesian linear inversion methodology based on Gaussian mixture models and its application to geophysical inverse problems are presented in this paper. The proposed inverse method is based on a Bayesian approach under the assumptions of a Gaussian mixture random field for the prior model and a Gaussian linear likelihood function. The model for the latent discrete variable is defined to be a stationary first-order Markov chain. In this approach, a recursive exact solution to an approximation of the posterior distribution of the inverse problem is proposed. A Markov chain Monte Carlo algorithm can be used to efficiently simulate realizations from the correct posterior model. Two inversion studies based on real well log data are presented, and the main results are the posterior distributions of the reservoir properties of interest, the corresponding predictions and prediction intervals, and a set of conditional realizations. The first application is a seismic inversion study for the prediction of lithological facies, P- and S-impedance, where an improvement of 30% in the root-mean-square error of the predictions compared to the traditional Gaussian inversion is obtained. The second application is a rock physics inversion study for the prediction of lithological facies, porosity, and clay volume, where predictions slightly improve compared to the Gaussian inversion approach.  相似文献   

2.
Empirical Maximum Likelihood Kriging: The General Case   总被引:4,自引:0,他引:4  
Although linear kriging is a distribution-free spatial interpolator, its efficiency is maximal only when the experimental data follow a Gaussian distribution. Transformation of the data to normality has thus always been appealing. The idea is to transform the experimental data to normal scores, krige values in the “Gaussian domain” and then back-transform the estimates and uncertainty measures to the “original domain.” An additional advantage of the Gaussian transform is that spatial variability is easier to model from the normal scores because the transformation reduces effects of extreme values. There are, however, difficulties with this methodology, particularly, choosing the transformation to be used and back-transforming the estimates in such a way as to ensure that the estimation is conditionally unbiased. The problem has been solved for cases in which the experimental data follow some particular type of distribution. In general, however, it is not possible to verify distributional assumptions on the basis of experimental histograms calculated from relatively few data and where the uncertainty is such that several distributional models could fit equally well. For the general case, we propose an empirical maximum likelihood method in which transformation to normality is via the empirical probability distribution function. Although the Gaussian domain simple kriging estimate is identical to the maximum likelihood estimate, we propose use of the latter, in the form of a likelihood profile, to solve the problem of conditional unbiasedness in the back-transformed estimates. Conditional unbiasedness is achieved by adopting a Bayesian procedure in which the likelihood profile is the posterior distribution of the unknown value to be estimated and the mean of the posterior distribution is the conditionally unbiased estimate. The likelihood profile also provides several ways of assessing the uncertainty of the estimation. Point estimates, interval estimates, and uncertainty measures can be calculated from the posterior distribution.  相似文献   

3.
Histograms of observations from spatial phenomena are often found to be more heavy-tailed than Gaussian distributions, which makes the Gaussian random field model unsuited. A T-distributed random field model with heavy-tailed marginal probability density functions is defined. The model is a generalization of the familiar Student-T distribution, and it may be given a Bayesian interpretation. The increased variability appears cross-realizations, contrary to in-realizations, since all realizations are Gaussian-like with varying variance between realizations. The T-distributed random field model is analytically tractable and the conditional model is developed, which provides algorithms for conditional simulation and prediction, so-called T-kriging. The model compares favourably with most previously defined random field models. The Gaussian random field model appears as a special, limiting case of the T-distributed random field model. The model is particularly useful whenever multiple, sparsely sampled realizations of the random field are available, and is clearly favourable to the Gaussian model in this case. The properties of the T-distributed random field model is demonstrated on well log observations from the Gullfaks field in the North Sea. The predictions correspond to traditional kriging predictions, while the associated prediction variances are more representative, as they are layer specific and include uncertainty caused by using variance estimates.  相似文献   

4.
Application of EM algorithms for seismic facices classification   总被引:1,自引:0,他引:1  
Identification of the geological facies and their distribution from seismic and other available geological information is important during the early stage of reservoir development (e.g. decision on initial well locations). Traditionally, this is done by manually inspecting the signatures of the seismic attribute maps, which is very time-consuming. This paper proposes an application of the Expectation-Maximization (EM) algorithm to automatically identify geological facies from seismic data. While the properties within a certain geological facies are relatively homogeneous, the properties between geological facies can be rather different. Assuming that noisy seismic data of a geological facies, which reflect rock properties, can be approximated with a Gaussian distribution, the seismic data of a reservoir composed of several geological facies are samples from a Gaussian mixture model. The mean of each Gaussian model represents the average value of the seismic data within each facies while the variance gives the variation of the seismic data within a facies. The proportions in the Gaussian mixture model represent the relative volumes of different facies in the reservoir. In this setting, the facies classification problem becomes a problem of estimating the parameters defining the Gaussian mixture model. The EM algorithm has long been used to estimate Gaussian mixture model parameters. As the standard EM algorithm does not consider spatial relationship among data, it can generate spatially scattered seismic facies which is physically unrealistic. We improve the standard EM algorithm by adding a spatial constraint to enhance spatial continuity of the estimated geological facies. By applying the EM algorithms to acoustic impedance and Poisson’s ratio data for two synthetic examples, we are able to identify the facies distribution.  相似文献   

5.
A methodology was presented for observation-based settlement prediction with consideration of the spatial correlation structure of soil. The spatial correlation is introduced among the settlement model parameters and the settlements at various points are spatially correlated through these geotechnical parameters, which naturally describe the phenomenon. The method is based on Bayesian estimation by considering both prior information, including spatial correlation and observed settlement, to search for the best estimates of the parameters at any arbitrary points on the ground. Within the Bayesian framework, the optimised selection of auto-correlation distance by Akaike's Bayesian Information Criterion (ABIC) is also proposed. The application of the proposed approach in consolidation settlement prediction using Asaoka's method is presented in this paper. Several case studies were carried out using simulated settlement data to investigate the performance the proposed approach. It is concluded that the accuracy of the settlement prediction can be improved by taking into account the spatial correlation structure and the proposed approach gives the rational prediction of the settlement at any location at any time with quantified uncertainty.  相似文献   

6.
A multivariate probability transformation between random variables, known as the Nataf transformation, is shown to be the appropriate transformation for multi-Gaussian kriging. It assumes a diagonal Jacobian matrix for the transformation of the random variables between the original space and the Gaussian space. This allows writing the probability transformation between the local conditional probability density function in the original space and the local conditional Gaussian probability density function in the Gaussian space as a ratio equal to the ratio of their respective marginal distributions. Under stationarity, the marginal distribution in the original space is modeled from the data histogram. The stationary marginal standard Gaussian distribution is obtained from the normal scores of the data and the local conditional Gaussian distribution is modeled from the kriging mean and kriging variance of the normal scores of the data. The equality of ratios of distributions has the same form as the Bayes’ rule and the assumption of stationarity of the data histogram can be re-interpreted as the gathering of the prior distribution. Multi-Gaussian kriging can be re-interpreted as an updating of the data histogram by a Gaussian likelihood. The Bayes’ rule allows for an even more general interpretation of spatial estimation in terms of equality for the ratio of the conditional distribution over the marginal distribution in the original data uncertainty space with the same ratio for a model of uncertainty with a distribution that can be modeled using the mean and variance from direct kriging of the original data values. It is based on the principle of conservation of probability ratio and no transformation is required. The local conditional distribution has a variance that is data dependent. When used in sequential simulation mode, it reproduces histogram and variogram of the data, thus providing a new approach for direct simulation in the original value space.  相似文献   

7.
Kriging Prediction Intervals Based on Semiparametric Bootstrap   总被引:1,自引:0,他引:1  
Kriging is a widely used method for prediction, which, given observations of a (spatial) process, yields the best linear unbiased predictor of the process at a new location. The construction of corresponding prediction intervals typically relies on Gaussian assumptions. Here we show that the distribution of kriging predictors for non-Gaussian processes may be far from Gaussian, even asymptotically. This emphasizes the need for other ways to construct prediction intervals. We propose a semiparametric bootstrap method with focus on the ordinary kriging predictor. No distributional assumptions about the data generating process are needed. A simulation study for Gaussian as well as lognormal processes shows that the semiparametric bootstrap method works well. For the lognormal process we see significant improvement in coverage probability compared to traditional methods relying on Gaussian assumptions.  相似文献   

8.
In this paper, the maximum likelihood method for inferring the parameters of spatial covariances is examined. The advantages of the maximum likelihood estimation are discussed and it is shown that this method, derived assuming a multivariate Gaussian distribution for the data, gives a sound criterion of fitting covariance models irrespective of the multivariate distribution of the data. However, this distribution is impossible to verify in practice when only one realization of the random function is available. Then, the maximum entropy method is the only sound criterion of assigning probabilities in absence of information. Because the multivariate Gaussian distribution has the maximum entropy property for a fixed vector of means and covariance matrix, the multinormal distribution is the most logical choice as a default distribution for the experimental data. Nevertheless, it should be clear that the assumption of a multivariate Gaussian distribution is maintained only for the inference of spatial covariance parameters and not necessarily for other operations such as spatial interpolation, simulation or estimation of spatial distributions. Various results from simulations are presented to support the claim that the simultaneous use of maximum likelihood method and the classical nonparametric method of moments can considerably improve results in the estimation of geostatistical parameters.  相似文献   

9.
The majority of geostatistical estimation and simulation algorithms rely on a covariance model as the sole characteristic of the spatial distribution of the attribute under study. The limitation to a single covariance implicitly calls for a multivariate Gaussian model for either the attribute itself or for its normal scores transform. The Gaussian model could be justified on the basis that it is both analytically simple and it is a maximum entropy model, i.e., a model that minimizes unwarranted structural properties. As a consequence, the Gaussian model also maximizes spatial disorder (beyond the imposed covariance) which can cause flow simulation results performed on multiple stochastic images to be very similar; thus, the space of response uncertainty could be too narrow entailing a misleading sense of safety. The ability of the sole covariance to adequately describe spatial distributions for flow studies, and the assumption that maximum spatial disorder amounts to either no additional information or a safe prior hypothesis are questioned. This paper attempts to clarify the link between entropy and spatial disorder and to provide, through a detailed case study, an appreciation for the impact of entropy of prior random function models on the resulting response distributions.  相似文献   

10.
The majority of geostatistical estimation and simulation algorithms rely on a covariance model as the sole characteristic of the spatial distribution of the attribute under study. The limitation to a single covariance implicitly calls for a multivariate Gaussian model for either the attribute itself or for its normal scores transform. The Gaussian model could be justified on the basis that it is both analytically simple and it is a maximum entropy model, i.e., a model that minimizes unwarranted structural properties. As a consequence, the Gaussian model also maximizes spatial disorder (beyond the imposed covariance) which can cause flow simulation results performed on multiple stochastic images to be very similar; thus, the space of response uncertainty could be too narrow entailing a misleading sense of safety. The ability of the sole covariance to adequately describe spatial distributions for flow studies, and the assumption that maximum spatial disorder amounts to either no additional information or a safe prior hypothesis are questioned. This paper attempts to clarify the link between entropy and spatial disorder and to provide, through a detailed case study, an appreciation for the impact of entropy of prior random function models on the resulting response distributions.  相似文献   

11.
The Bayesian framework is the standard approach for data assimilation in reservoir modeling. This framework involves characterizing the posterior distribution of geological parameters in terms of a given prior distribution and data from the reservoir dynamics, together with a forward model connecting the space of geological parameters to the data space. Since the posterior distribution quantifies the uncertainty in the geologic parameters of the reservoir, the characterization of the posterior is fundamental for the optimal management of reservoirs. Unfortunately, due to the large-scale highly nonlinear properties of standard reservoir models, characterizing the posterior is computationally prohibitive. Instead, more affordable ad hoc techniques, based on Gaussian approximations, are often used for characterizing the posterior distribution. Evaluating the performance of those Gaussian approximations is typically conducted by assessing their ability at reproducing the truth within the confidence interval provided by the ad hoc technique under consideration. This has the disadvantage of mixing up the approximation properties of the history matching algorithm employed with the information content of the particular observations used, making it hard to evaluate the effect of the ad hoc approximations alone. In this paper, we avoid this disadvantage by comparing the ad hoc techniques with a fully resolved state-of-the-art probing of the Bayesian posterior distribution. The ad hoc techniques whose performance we assess are based on (1) linearization around the maximum a posteriori estimate, (2) randomized maximum likelihood, and (3) ensemble Kalman filter-type methods. In order to fully resolve the posterior distribution, we implement a state-of-the art Markov chain Monte Carlo (MCMC) method that scales well with respect to the dimension of the parameter space, enabling us to study realistic forward models, in two space dimensions, at a high level of grid refinement. Our implementation of the MCMC method provides the gold standard against which the aforementioned Gaussian approximations are assessed. We present numerical synthetic experiments where we quantify the capability of each of the ad hoc Gaussian approximation in reproducing the mean and the variance of the posterior distribution (characterized via MCMC) associated to a data assimilation problem. Both single-phase and two-phase (oil–water) reservoir models are considered so that fundamental differences in the resulting forward operators are highlighted. The main objective of our controlled experiments was to exhibit the substantial discrepancies of the approximation properties of standard ad hoc Gaussian approximations. Numerical investigations of the type we present here will lead to the greater understanding of the cost-efficient, but ad hoc, Bayesian techniques used for data assimilation in petroleum reservoirs and hence ultimately to improved techniques with more accurate uncertainty quantification.  相似文献   

12.
In this paper, a Bayesian approach for updating a semi-empirical model for predicting excavation-induced maximum ground settlement using centrifuge test data is presented. The Bayesian approach involves three steps: (1) prior estimate of the maximum ground settlement and model bias factor, (2) establishment of the likelihood function and posterior distribution of the model bias factor using the settlement measurement in the centrifuge test, and (3) development of posterior distribution of the predicted maximum settlement. This Bayesian approach is demonstrated with a case study of a well-documented braced excavation, and the results show that the accuracy of the maximum settlement prediction can be improved and the model uncertainty can be reduced with Bayesian updating.  相似文献   

13.

Spatial data analytics provides new opportunities for automated detection of anomalous data for data quality control and subsurface segmentation to reduce uncertainty in spatial models. Solely data-driven anomaly detection methods do not fully integrate spatial concepts such as spatial continuity and data sparsity. Also, data-driven anomaly detection methods are challenged in integrating critical geoscience and engineering expertise knowledge. The proposed spatial anomaly detection method is based on the semivariogram spatial continuity model derived from sparsely sampled well data and geological interpretations. The method calculates the lag joint cumulative probability for each matched pair of spatial data, given their lag vector and the semivariogram under the assumption of bivariate Gaussian distribution. For each combination of paired spatial data, the associated head and tail Gaussian standardized values of a pair of spatial data are mapped to the joint probability density function informed from the lag vector and semivariogram. The paired data are classified as anomalous if the associated head and tail Gaussian standardized values fall within a low probability zone. The anomaly decision threshold can be decided based on a loss function quantifying the cost of overestimation or underestimation. The proposed spatial correlation anomaly detection method is able to integrate domain expertise knowledge through trend and correlogram models with sparse spatial data to identify anomalous samples, region, segmentation boundaries, or facies transition zones. This is a useful automation tool for identifying samples in big spatial data on which to focus professional attention.

  相似文献   

14.
Building of models in the Earth Sciences often requires the solution of an inverse problem: some unknown model parameters need to be calibrated with actual measurements. In most cases, the set of measurements cannot completely and uniquely determine the model parameters; hence multiple models can describe the same data set. Bayesian inverse theory provides a framework for solving this problem. Bayesian methods rely on the fact that the conditional probability of the model parameters given the data (the posterior) is proportional to the likelihood of observing the data and a prior belief expressed as a prior distribution of the model parameters. In case the prior distribution is not Gaussian and the relation between data and parameters (forward model) is strongly non-linear, one has to resort to iterative samplers, often Markov chain Monte Carlo methods, for generating samples that fit the data likelihood and reflect the prior model statistics. While theoretically sound, such methods can be slow to converge, and are often impractical when the forward model is CPU demanding. In this paper, we propose a new sampling method that allows to sample from a variety of priors and condition model parameters to a variety of data types. The method does not rely on the traditional Bayesian decomposition of posterior into likelihood and prior, instead it uses so-called pre-posterior distributions, i.e. the probability of the model parameters given some subset of the data. The use of pre-posterior allows to decompose the data into so-called, “easy data” (or linear data) and “difficult data” (or nonlinear data). The method relies on fast non-iterative sequential simulation to generate model realizations. The difficult data is matched by perturbing an initial realization using a perturbation mechanism termed “probability perturbation.” The probability perturbation method moves the initial guess closer to matching the difficult data, while maintaining the prior model statistics and the conditioning to the linear data. Several examples are used to illustrate the properties of this method.  相似文献   

15.
目前针对模型结构不确定性的研究方法主要为贝叶斯模型平均方法,而该方法受到模型权重计算困难等影响,应用受限。基于数据驱动的模型结构误差统计学习方法最近得到关注。研究采用高斯过程回归方法对地下水模型结构误差进行统计模拟,并将DREAMzs算法与高斯过程回归相结合,对地下水模型和统计模型的参数同时进行识别。基于此方法,分别以理想岩溶裂隙海水入侵过程和溶质运移柱体实验为例,进行地下水数值模拟及预测结果的不确定性分析。相对于不考虑模型结构误差条件的不确定性分析,结果表明,考虑结构误差之后,能够明显减少参数识别过程中的参数补偿影响,且能显著提高模型的预测性能。因此,基于高斯过程回归的模型结构不确定性分析可以一定程度控制地下水数值模拟的不确定性,提高模型预测可靠性。  相似文献   

16.
岩爆是地下工程开挖过程中硬质岩体存储的弹性应变能突然、迅速释放的动态过程。我国西南山区正在建设或拟建大量深埋长大隧道,勘察阶段岩爆的准确预测对有效设计和控制投资十分重要。从隧道工程勘察阶段线路比选与设计需求出发,针对隧道勘查期岩爆灾害预测指标获取难、预测精度低的问题,以该阶段岩爆预测指标的易获取性为前提,利用贝叶斯网络解决不确定性问题的有效性来反映岩爆烈度与各影响因素的相关关系。基于473组岩爆灾害案例,采用4个预测指标(地应力、地质构造、围岩级别和岩石强度)来构建岩爆烈度朴素贝叶斯概率分级预测模型,利用十折交叉验证方法确定模型预测精度达84.47%。将该模型应用于雅安—叶城高速公路跑马山1号隧道岩爆段落,预测结果显示:28次岩爆预测中有24次正确、4次错误,准确率高达85.71%;其中2组错误预测中,现场判别为轻微-中等岩爆,而本文模型预测为轻微岩爆。验证结果表明所建立的贝叶斯网络模型具有良好的预测性能,研究成果可为我国西南山区深埋长大硬岩隧道勘察设计期岩爆灾害预测提供技术支撑。  相似文献   

17.
针对现有深度学习水文模型未能充分刻画气象要素空间特征的问题, 本文基于主成分分析(PCA)方法提取气象要素空间特征, 利用长短时记忆神经网络(LSTM)学习长时序过程规律, 构建融合气象要素时空特性的深度学习水文模型PCA-LSTM。以黄河源区为研究区域, 利用LSTM模型和物理水文模型THREW作为比对模型, 基于高斯噪音法系统评估PCA-LSTM模型的适用性和鲁棒性。结果显示: PCA-LSTM模型径流模拟纳什效率系数为0.92, 高于比对模型LSTM和THREW, 表明模型具有较高的精度。研究结果可为流域高精度水文模拟提供参考。  相似文献   

18.
Variograms of Order ω: A Tool to Validate a Bivariate Distribution Model   总被引:1,自引:0,他引:1  
The multigaussian model is used in mining geostatistics to simulate the spatial distribution of grades or to estimate the recoverable reserves of an ore deposit. Checking the suitability of such model to the available data often constitutes a critical step of the geostatistical study. In general, the marginal distribution is not a problem because the data can be transformed to normal scores, so the check is usually restricted to the bivariate distributions. In this work, several tests for diagnosing the two-point normality of a set of Gaussian data are reviewed and commented. An additional criterion is proposed, based on the comparison between the usual variogram and the variograms of lower order: the latter are defined as half the mean absolute increments of the attribute raised to a power between 0 and 2. This criterion is then extended to other bivariate models, namely the bigamma, Hermitian and Laguerrian models. The concepts are illustrated on two real data-sets. Finally, some conditions to ensure the internal consistency of the variogram under a given model are given.  相似文献   

19.
Soil erosion is one of most widespread process of degradation. The erodibility of a soil is a measure of its susceptibility to erosion and depends on many soil properties. Soil erodibility factor varies greatly over space and is commonly estimated using the revised universal soil loss equation. Neglecting information about estimation uncertainty may lead to improper decision-making. One geostatistical approach to spatial analysis is sequential Gaussian simulation, which draws alternative, equally probable, joint realizations of a regionalised variable. Differences between the realizations provide a measure of spatial uncertainty and allow us to carry out an error analysis. The objective of this paper was to assess the model output error of soil erodibility resulting from the uncertainties in the input attributes (texture and organic matter). The study area covers about 30 km2 (Calabria, southern Italy). Topsoil samples were collected at 175 locations within the study area in 2006 and the main chemical and physical soil properties were determined. As soil textural size fractions are compositional data, the additive-logratio (alr) transformation was used to remove the non-negativity and constant-sum constraints on compositional variables. A Monte Carlo analysis was performed, which consisted of drawing a large number (500) of identically distributed input attributes from the multivariable joint probability distribution function. We incorporated spatial cross-correlation information through joint sequential Gaussian simulation, because model inputs were spatially correlated. The erodibility model was then estimated for each set of the 500 joint realisations of the input variables and the ensemble of the model outputs was used to infer the erodibility probability distribution function. This approach has also allowed for delineating the areas characterised by greater uncertainty and then to suggest efficient supplementary sampling strategies for further improving the precision of K value predictions.  相似文献   

20.
Model calibration and history matching are important techniques to adapt simulation tools to real-world systems. When prediction uncertainty needs to be quantified, one has to use the respective statistical counterparts, e.g., Bayesian updating of model parameters and data assimilation. For complex and large-scale systems, however, even single forward deterministic simulations may require parallel high-performance computing. This often makes accurate brute-force and nonlinear statistical approaches infeasible. We propose an advanced framework for parameter inference or history matching based on the arbitrary polynomial chaos expansion (aPC) and strict Bayesian principles. Our framework consists of two main steps. In step 1, the original model is projected onto a mathematically optimal response surface via the aPC technique. The resulting response surface can be viewed as a reduced (surrogate) model. It captures the model’s dependence on all parameters relevant for history matching at high-order accuracy. Step 2 consists of matching the reduced model from step 1 to observation data via bootstrap filtering. Bootstrap filtering is a fully nonlinear and Bayesian statistical approach to the inverse problem in history matching. It allows to quantify post-calibration parameter and prediction uncertainty and is more accurate than ensemble Kalman filtering or linearized methods. Through this combination, we obtain a statistical method for history matching that is accurate, yet has a computational speed that is more than sufficient to be developed towards real-time application. We motivate and demonstrate our method on the problem of CO2 storage in geological formations, using a low-parametric homogeneous 3D benchmark problem. In a synthetic case study, we update the parameters of a CO2/brine multiphase model on monitored pressure data during CO2 injection.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号