On selection of kernel parametes in relevance vector machines for hydrologic applications |
| |
Authors: | Shivam Tripathi Rao S. Govindaraju |
| |
Affiliation: | (1) School of Civil Engineering, Purdue University, West Lafayette, IN 47907, USA |
| |
Abstract: | Recent advances in statistical learning theory have yielded tools that are improving our capabilities for analyzing large and complex datasets. Among such tools, relevance vector machines (RVMs) are finding increasing applications in hydrology because of (1) their excellent generalization properties, and (2) the probabilistic interpretation associated with this technique that yields prediction uncertainty. RVMs combine the strengths of kernel-based methods and Bayesian theory to establish relationships between a set of input vectors and a desired output. However, a bias–variance analysis of RVM estimates revealed that a careful selection of kernel parameters is of paramount importance for achieving good performance from RVMs. In this study, several analytic methods are presented for selection of kernel parameters. These methods rely on structural properties of the data rather than expensive re-sampling approaches commonly used in RVM applications. An analytical expression for prediction risk in leave-one-out cross validation is derived. For brevity, the effectiveness of the proposed methods is assessed first by data generated from the benchmark sinc function, followed by an example involving estimation of hydraulic conductivity values over a field based on observations. It is shown that a straightforward maximization of likelihood function can lead to misleading results. The proposed methods are found to yield robust estimates of parameters for kernel functions. |
| |
Keywords: | Bayesian learning Relevance vector machines Interpolation Leave-one-out cross-validation VC dimension Bayes information criterion Power spectrum |
本文献已被 SpringerLink 等数据库收录! |
|