首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 406 毫秒
1.
—?A maximum-likelihood (ML) estimator of the correlation dimension d 2 of fractal sets of points not affected by the left-hand truncation of their inter-distances is defined. Such truncation might produce significant biases of the ML estimates of d 2 when the observed scale range of the phenomenon is very narrow, as often occurs in seismological studies. A second very simple algorithm based on the determination of the first two moments of the inter-distances distribution (SOM) is also proposed, itself not biased by the left-hand truncation effect. The asymptotic variance of the ML estimates is given. Statistical tests carried out on data samples with different sizes extracted from populations of inter-distances following a power law, suggested that the sample variance of the estimates obtained by the proposed methods are not significantly different, and are well estimated by the asymptotic variance also for samples containing a few hundred inter-distances. To examine the effects of different sources of systematic errors, the two estimators were also applied to sets of inter-distances between points belonging to statistical fractal distributions, baker's maps and experimental distributions of earthquake epicentres. For a full evaluation of the results achieved by the methods proposed here, these were compared with those obtained by the ML estimator for untruncated samples or by the least-squares algorithm.  相似文献   

2.
Asymptotic properties of maximum likelihood parameter and quantile estimators of the 2-parameter kappa distribution are studied. Eight methods for obtaining large sample confidence intervals for the shape parameter and for quantiles of this distribution are proposed and compared by using Monte Carlo simulation. The best method is highlighted on the basis of the coverage probability of the confidence intervals that it produces for sample sizes commonly found in practice. For such sample sizes, confidence intervals for quantiles and for the shape parameter are shown to be more accurate if the quantile estimators are assumed to be log normally distributed rather than normally distributed (same for the shape parameter estimator). Also, confidence intervals based on the observed Fisher information matrix perform slightly better than those based on the expected value of this matrix. A hydrological example is provided in which the obtained theoretical results are applied.  相似文献   

3.
A new bivariate pseudo Pareto distribution is proposed, and its distributional characteristics are investigated. The parameters of this distribution are estimated by the moment-, the maximum likelihood- and the Bayesian method. Point estimators of the parameters are presented for different sample sizes. Asymptotic confidence intervals are constructed and the parameter modeling the dependency between two variables is checked. The performance of the different estimation methods is investigated by using the bootstrap method. A Markov Chain Monte Carlo simulation is conducted to estimate the Bayesian posterior distribution for different sample sizes. For illustrative purposes, a real set of drought data is investigated.  相似文献   

4.
We consider kernel-based non-parametric estimation of second-order product densities of spatial point patterns. We present a new family of optimal and positive kernels that shows less variance and more flexibility than optimal kernels. This family generalises most of the classical and widely used kernel functions, such as Box or Epanechnikov kernels. We propose an alternative asymptotically unbiased estimator for the second-order product density function, and compare the performance of the estimator for several members of the family of optimal and positive kernels through MISE and relative efficiency. We present a simulation study to analyse the behaviour of such kernel functions, for three different spatial structures, for which we know the exact analytical form of the product density, and under small sample sizes. Some known datasets are revisited, and we also analyse the IMD dataset in the Rhineland Regional Council in Germany.  相似文献   

5.
In the analysis of regionalized data, irregular sampling patterns are often responsible for large deviations (fluctuations) between the theoretical and sample semi-variograms. This article proposes a new semi-variogram estimator that is unbiased irrespective of the actual multivariate distribution of the data (provided an assumption of stationarity) and has the minimal variance under a given multivariate distribution model. Such an estimator considerably reduces fluctuations in the sample semi-variogram when the data are strongly correlated and clustered in space, and proves to be robust to a misspecification of the multivariate distribution model. The traditional and proposed semi-variogram estimators are compared through an application to a pollution dataset.  相似文献   

6.
The key problem in nonparametric frequency analysis of flood and droughts is the estimation of the bandwidth parameter which defines the degree of smoothing. Most of the proposed bandwidth estimators have been based on the density function rather than the cumulative distribution function or the quantile that are the primary interest in frequency analysis. We propose a new bandwidth estimator derived from properties of quantile estimators. The estimator builds on work by Altman and Léger (1995). The estimator is compared to the well-known method of least squares cross-validation (LSCV) using synthetic data generated from various parametric distributions used in hydrologic frequency analysis. Simulations suggest that our estimator performs at least as well as, and in many cases better than, the method of LSCV. In particular, the use of the proposed plug-in estimator reduces bias in the estimation as compared to LSCV. When applied to data sets containing observations with identical values, typically the result of rounding or truncation, the LSCV and most other techniques generally underestimates the bandwidth. The proposed technique performs very well in such situations.  相似文献   

7.
Bayesian improver of a distribution   总被引:1,自引:0,他引:1  
 An estimate of a distribution obtained from a sample by any method of classical statistics may be erroneous when the sample is not representative of the population. A subjective distribution elicited from an expert may be miscalibrated when information is scanty and experience limited. The Bayesian Improver of a Distribution (BID) exploits a coherence principle and improves, in the ex ante sense, an initial estimate of a continuous distribution by using (i) the known distribution of a related variate and (ii) information about the dependence structure between the two variates. The theory of BID is developed into an applied (ABID) procedure. The ABID estimator is applicable to any continuous, monotone likelihood ratio dependent variates with arbitrary, strictly increasing marginal distributions, parametric or nonparametric; it is analytic in form and easy to implement via statistical or judgmental methods; it converges to the true distribution, provided the initial estimator does, as the sample size n→∞; it outperforms the initial estimator in the expected Kolmogorov–Smirnov distance for all n; and it offers the greatest gains when n is small – precisely when improved estimates are needed most.  相似文献   

8.
In this paper we study the mean and standard deviation of concentrations using random walk models. Two-particle models that take into account the space correlation of the turbulence are introduced and some properties of the distribution of the particle concentration are studied. In order to reduce the CPU time of the calculation a new estimator based on reverse time diffusion is applied. This estimator has been recently introduced by Milstein et al. (Bernoulli 10(2):281–312, 2004). Some numerical aspects of the implementation are discussed for relatively simple test problems. Finally, a realistic application to predict the spreading of the pollutant in the Dutch coastal zone is described.  相似文献   

9.
Various methods for estimating the self-similarity parameter (Hurst parameter, H) of a Hurst–Kolmogorov stochastic process (HKp) from a time series are available. Most of them rely on some asymptotic properties of processes with Hurst–Kolmogorov behaviour and only estimate the self-similarity parameter. Here we show that the estimation of the Hurst parameter affects the estimation of the standard deviation, a fact that was not given appropriate attention in the literature. We propose the least squares based on variance estimator, and we investigate numerically its performance, which we compare to the least squares based on standard deviation estimator, as well as the maximum likelihood estimator after appropriate streamlining of the latter. These three estimators rely on the structure of the HKp and estimate simultaneously its Hurst parameter and standard deviation. In addition, we test the performance of the three methods for a range of sample sizes and H values, through a simulation study and we compare it with other estimators of the literature.  相似文献   

10.
ABSTRACT

Several commonly-used nonparametric change-point detection methods are analysed in terms of power, ability and accuracy of the estimated change-point location. The analysis is performed with synthetic data for different sample sizes, two types of change and different magnitudes of change. The methods studied are the Pettitt method, a method based on the Cramér von Mises (CvM) two-sample test statistic and a variant of the CUSUM method. The methods differ considerably in behaviour. For all methods the spread of estimated change-point location increases significantly for points near one of the ends of the sample. Series of annual maximum runoff for four stations on the Yangtze River in China are used to examine the performance of the methods on real data. It was found that the CvM-based test gave the best results, but all three methods suffer from bias and low detection rates for change points near the ends of the series.  相似文献   

11.
The combination of Stokes formula and an Earth Gravity Model (EGM) for geoid determination has become a standard procedure. However, the way of modifying Stokes formula vary from author to author, and numerous methods of modification exist. Most methods are deterministic, with the primary goal of reducing the truncation bias committed by limiting the area of Stokes integration around the computation point, but there are also some stochastic methods with the explicit goal to reduce the global mean square error of the geoid height estimator stemming from the truncation bias as well as the random errors of the EGM and the gravity data. The latter estimators are thus, at least from a theoretical point of view, optimal in a global mean sense, but in a local sense they may be far from optimality.Here we take advantage of the error variance-covariance matrices of the EGM and the terrestrial gravity data to derive the modification parameters of Stokes kernel in a local least-squares sense. The solution is given for the unbiased type of modification of Stokes formula of Sjöberg (1991).  相似文献   

12.
A new approach for streamflow simulation using nonparametric methods was described in a recent publication (Sharma et al. 1997). Use of nonparametric methods has the advantage that they avoid the issue of selecting a probability distribution and can represent nonlinear features, such as asymmetry and bimodality that hitherto were difficult to represent, in the probability structure of hydrologic variables such as streamflow and precipitation. The nonparametric method used was kernel density estimation, which requires the selection of bandwidth (smoothing) parameters. This study documents some of the tests that were conduced to evaluate the performance of bandwidth estimation methods for kernel density estimation. Issues related to selection of optimal smoothing parameters for kernel density estimation with small samples (200 or fewer data points) are examined. Both reference to a Gaussian density and data based specifications are applied to estimate bandwidths for samples from bivariate normal mixture densities. The three data based methods studied are Maximum Likelihood Cross Validation (MLCV), Least Square Cross Validation (LSCV) and Biased Cross Validation (BCV2). Modifications for estimating optimal local bandwidths using MLCV and LSCV are also examined. We found that the use of local bandwidths does not necessarily improve the density estimate with small samples. Of the global bandwidth estimators compared, we found that MLCV and LSCV are better because they show lower variability and higher accuracy while Biased Cross Validation suffers from multiple optimal bandwidths for samples from strongly bimodal densities. These results, of particular interest in stochastic hydrology where small samples are common, may have importance in other applications of nonparametric density estimation methods with similar sample sizes and distribution shapes. Received: November 12, 1997  相似文献   

13.
A new approach for streamflow simulation using nonparametric methods was described in a recent publication (Sharma et al. 1997). Use of nonparametric methods has the advantage that they avoid the issue of selecting a probability distribution and can represent nonlinear features, such as asymmetry and bimodality that hitherto were difficult to represent, in the probability structure of hydrologic variables such as streamflow and precipitation. The nonparametric method used was kernel density estimation, which requires the selection of bandwidth (smoothing) parameters. This study documents some of the tests that were conduced to evaluate the performance of bandwidth estimation methods for kernel density estimation. Issues related to selection of optimal smoothing parameters for kernel density estimation with small samples (200 or fewer data points) are examined. Both reference to a Gaussian density and data based specifications are applied to estimate bandwidths for samples from bivariate normal mixture densities. The three data based methods studied are Maximum Likelihood Cross Validation (MLCV), Least Square Cross Validation (LSCV) and Biased Cross Validation (BCV2). Modifications for estimating optimal local bandwidths using MLCV and LSCV are also examined. We found that the use of local bandwidths does not necessarily improve the density estimate with small samples. Of the global bandwidth estimators compared, we found that MLCV and LSCV are better because they show lower variability and higher accuracy while Biased Cross Validation suffers from multiple optimal bandwidths for samples from strongly bimodal densities. These results, of particular interest in stochastic hydrology where small samples are common, may have importance in other applications of nonparametric density estimation methods with similar sample sizes and distribution shapes. Received: November 12, 1997  相似文献   

14.
Sequential kriging and cokriging: Two powerful geostatistical approaches   总被引:1,自引:0,他引:1  
A sequential linear estimator is developed in this study to progressively incorporate new or different spatial data sets into the estimation. It begins with a classical linear estimator (i.e., kriging or cokriging) to estimate means conditioned to a given observed data set. When an additional data set becomes available, the sequential estimator improves the previous estimate by using linearly weighted sums of differences between the new data set and previous estimates at sample locations. Like the classical linear estimator, the weights used in the sequential linear estimator are derived from a system of equations that contains covariances and cross-covariances between sample locations and the location where the estimate is to be made. However, the covariances and cross-covariances are conditioned upon the previous data sets. The sequential estimator is shown to produce the best, unbiased linear estimate, and to provide the same estimates and variances as classic simple kriging or cokriging with the simultaneous use of the entire data set. However, by using data sets sequentially, this new algorithm alleviates numerical difficulties associated with the classical kriging or cokriging techniques when a large amount of data are used. It also provides a new way to incorporate additional information into a previous estimation.  相似文献   

15.
The test for exponentiality of a dataset in terms of a specific aging property constitutes an interesting problem in reliability analysis. To this end, a wide variety of tests are proposed in the literature. In this paper, the excess-wealth function is recalled and new asymptotic properties are studied. By using the characterization of the exponential distribution based on the excess-wealth function, a new exponentiality test is proposed. Through simulation techniques, it is shown that this new test works well on small sample sizes. The exact null distribution and normality asymptotic is also obtained for the statistic proposed. This test and a new empirical graph based on the excess-wealth function are applied to extreme-value examples.  相似文献   

16.
The problem of approximating the probability distribution of maximum water discharges during rain-induced floods is considered. A truncation procedure applied to the analyzed samples is shown to yield acceptable results. The application of a procedure of joint data analysis demonstrated that sample truncation at the median is optimal as a trade-off between approximation quality and information loss.  相似文献   

17.
Compositional Bayesian indicator estimation   总被引:1,自引:1,他引:0  
Indicator kriging is widely used for mapping spatial binary variables and for estimating the global and local spatial distributions of variables in geosciences. For continuous random variables, indicator kriging gives an estimate of the cumulative distribution function, for a given threshold, which is then the estimate of a probability. Like any other kriging procedure, indicator kriging provides an estimation variance that, although not often used in applications, should be taken into account as it assesses the uncertainty of the estimate. An alternative approach to indicator estimation is proposed in this paper. In this alternative approach the complete probability density function of the indicator estimate is evaluated. The procedure is described in a Bayesian framework, using a multivariate Gaussian likelihood and an a priori distribution which are both combined according to Bayes theorem in order to obtain a posterior distribution for the indicator estimate. From this posterior distribution, point estimates, interval estimates and uncertainty measures can be obtained. Among the point estimates, the median of the posterior distribution is the maximum entropy estimate because there is a fifty-fifty chance of the unknown value of the estimate being larger or smaller than the median; that is, there is maximum uncertainty in the choice between two alternatives. Thus in some sense, the latter is an indicator estimator, alternative to the kriging estimator, that includes its own uncertainty. On the other hand, the mode of the posterior distribution estimator, assuming a uniform prior, is coincidental with the simple kriging estimator. Additionally, because the indicator estimate can be considered as a two-part composition which domain of definition is the simplex, the method is extended to compositional Bayesian indicator estimation. Bayesian indicator estimation and compositional Bayesian indicator estimation are illustrated with an environmental case study in which the probability of the content of a geochemical element in soil being over a particular threshold is of interest. The computer codes and its user guides are public domain and freely available.  相似文献   

18.
孙文科 《地震学报》1987,9(2):176-186
本文研究了海潮重力负荷效应的计算方法和中国大陆海潮(M2波)负荷效应的空间分布特征.给出了顾及测站高程的计算负荷效应的公式.从理论上证明了应该改化'n因子,这样不仅消除了由伪潮高所产生的误差|2fH*|,而且使引力项加快了一阶收敛速度.对已有的远近区结合法作了进一步的研究,确定了截断角0和逼近阶数 m 之间的取值关系,认为取0=12和 m=17是最合适的.计算了中国大陆共200个点的负荷效应,给出振幅分布图,并且就其分布特征进行了讨论.   相似文献   

19.
Spatial statistics of clustered data   总被引:7,自引:7,他引:0  
Modern spatial statistics techniques are widely used to make predictions for natural processes that are continuously distributed over some convex domain. Implementation of these techniques often relies on the adequate estimation of certain spatial correlation functions such as the covariance and the variogram from the data sets available. This work studies the practical estimation of such spatial correlation functions in the case of clustered data. The coefficient of variation of the dimensionless spatial density of the point pattern of sample locations is suggested as a useful metric for degree of clusteredness of the clustered data set. We show that the common variogram estimator becomes increasingly unreliable with increasing coefficient of variation of the dimensionless spatial density of the point pattern of sample locations. Moreover, we present a modified form of the variogram estimator that incorporates declustering weights, and propose a scheme for estimating the declustering weights based on zones of proximity. Finally, insight is gained in terms of a numerical application of the common and modified methods on piezometric head data collected over an irregular network.Acknowledgments. This work has been supported by grants from the National Institute of Environmental Health Sciences (P42 ES05948-02), the Army Research Office (DAAG55-98-1-0289), and the National Aeronautics and Space Administration (60-00RFQ041). Some of the calculations conducted in support of this work were done on the SGI Origin 2400 at the North Carolina Supercomputing Center, RTP, NC.  相似文献   

20.
The robustness of large quantile estimates of largest elements in a small sample by the methods of moments (MOM), L‐moments (LMM) and maximum likelihood (MLM) was evaluated and compared. Bias (B) and mean square error (MSE) were used to measure the estimation methods performance. Quantiles were estimated by eight two‐parameter probability distributions with the variation coefficient being the shape parameter. The effect of dropping largest elements of the series on large quantile values was assessed for various variation coefficient (CV)/sample size (n) ‘combinations’ with n = 30 as the basic value. To that end, both the Monte Carlo sampling experiments and an asymptotic approach consisting in distribution truncation were applied. In general, both sampling and asymptotic approaches point to MLM as the most robust method of the three considered, with respect to bias of large quantiles. Comparing the performance of two other methods, the MOM estimates were found to be more robust for small and moderate hydrological samples drawn from distributions with zero lower‐bound than were the LMM estimates. Extending the evaluation to outliers, it was shown that all the above findings remain valid. However, using the MSE variation as a measure of performance, the LMM was found to be the best for most distribution/variation coefficient combinations, whereas MOM was found to be the worst. Moreover, removal of the largest sample element need not result in a loss of estimation efficiency. The gain in accuracy is observed for the heavy‐tailed and log‐normal distributions, being particularly distinctive for LMM. In practice, while dealing with a single sample deprived of its largest element, one should choose the estimation method giving the lowest MSE of large quantiles. For n = 30 and several distribution/variation coefficient combinations, the MLM outperformed the two other methods in this respect and its supremacy grew with sample size, while MOM was usually the worst. Copyright © 2006 John Wiley & Sons, Ltd.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号