HOME

TheInfoList



OR:

In
probability theory Probability theory or probability calculus is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expre ...
and
statistics Statistics (from German language, German: ', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data. In applying statistics to a s ...
, a Gaussian process is a
stochastic process In probability theory and related fields, a stochastic () or random process is a mathematical object usually defined as a family of random variables in a probability space, where the index of the family often has the interpretation of time. Sto ...
(a collection of random variables indexed by time or space), such that every finite collection of those random variables has a
multivariate normal distribution In probability theory and statistics, the multivariate normal distribution, multivariate Gaussian distribution, or joint normal distribution is a generalization of the one-dimensional ( univariate) normal distribution to higher dimensions. One d ...
. The distribution of a Gaussian process is the
joint distribution A joint or articulation (or articular surface) is the connection made between bones, ossicles, or other hard structures in the body which link an animal's skeletal system into a functional whole.Saladin, Ken. Anatomy & Physiology. 7th ed. McGraw- ...
of all those (infinitely many) random variables, and as such, it is a distribution over functions with a continuous domain, e.g. time or space. The concept of Gaussian processes is named after
Carl Friedrich Gauss Johann Carl Friedrich Gauss (; ; ; 30 April 177723 February 1855) was a German mathematician, astronomer, geodesist, and physicist, who contributed to many fields in mathematics and science. He was director of the Göttingen Observatory and ...
because it is based on the notion of the Gaussian distribution (
normal distribution In probability theory and statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is f(x) = \frac ...
). Gaussian processes can be seen as an infinite-dimensional generalization of multivariate normal distributions. Gaussian processes are useful in
statistical model A statistical model is a mathematical model that embodies a set of statistical assumptions concerning the generation of Sample (statistics), sample data (and similar data from a larger Statistical population, population). A statistical model repre ...
ling, benefiting from properties inherited from the normal distribution. For example, if a
random process In probability theory and related fields, a stochastic () or random process is a mathematical object usually defined as a family of random variables in a probability space, where the index of the family often has the interpretation of time. Stoc ...
is modelled as a Gaussian process, the distributions of various derived quantities can be obtained explicitly. Such quantities include the average value of the process over a range of times and the error in estimating the average using sample values at a small set of times. While exact models often scale poorly as the amount of data increases, multiple approximation methods have been developed which often retain good accuracy while drastically reducing computation time.


Definition

A time continuous
stochastic process In probability theory and related fields, a stochastic () or random process is a mathematical object usually defined as a family of random variables in a probability space, where the index of the family often has the interpretation of time. Sto ...
\left\ is Gaussian
if and only if In logic and related fields such as mathematics and philosophy, "if and only if" (often shortened as "iff") is paraphrased by the biconditional, a logical connective between statements. The biconditional is true in two cases, where either bo ...
for every
finite set In mathematics, particularly set theory, a finite set is a set that has a finite number of elements. Informally, a finite set is a set which one could in principle count and finish counting. For example, is a finite set with five elements. Th ...
of indices t_1,\ldots,t_k in the index set T \mathbf_ = (X_, \ldots, X_) is a
multivariate Gaussian In probability theory and statistics, the multivariate normal distribution, multivariate Gaussian distribution, or joint normal distribution is a generalization of the one-dimensional (univariate) normal distribution to higher dimensions. One de ...
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a Mathematics, mathematical formalization of a quantity or object which depends on randomness, random events. The term 'random variable' in its mathema ...
. As the sum of independent and Gaussian distributed random variables is again Gaussian distributed, that is the same as saying every linear combination of (X_, \ldots, X_) has a univariate Gaussian (or normal) distribution. Using characteristic functions of random variables with i denoting the
imaginary unit The imaginary unit or unit imaginary number () is a mathematical constant that is a solution to the quadratic equation Although there is no real number with this property, can be used to extend the real numbers to what are called complex num ...
such that i^2 =-1, the Gaussian property can be formulated as follows: \left\ is Gaussian if and only if, for every finite set of indices t_1,\ldots,t_k, there are real-valued \sigma_, \mu_\ell with \sigma_ > 0 such that the following equality holds for all s_1,s_2,\ldots,s_k\in\mathbb, \left exp\left(i \sum_^k s_\ell \, \mathbf_\right)\right= \exp \left(-\tfrac \sum_ \sigma_ s_\ell s_j + i \sum_\ell \mu_\ell s_\ell\right), or \left ^ \right= ^. The numbers \sigma_ and \mu_\ell can be shown to be the
covariance In probability theory and statistics, covariance is a measure of the joint variability of two random variables. The sign of the covariance, therefore, shows the tendency in the linear relationship between the variables. If greater values of one ...
s and
means Means may refer to: * Means LLC, an anti-capitalist media worker cooperative * Means (band), a Christian hardcore band from Regina, Saskatchewan * Means, Kentucky, a town in the US * Means (surname) * Means Johnston Jr. (1916–1989), US Navy ...
of the variables in the process.


Variance

The variance of a Gaussian process is finite at any time t, formally \operatorname (t)= \left X(t)-\operatorname ^2\right< \infty \quad \text t \in T.


Stationarity

For general stochastic processes strict-sense stationarity implies wide-sense stationarity but not every wide-sense stationary stochastic process is strict-sense stationary. However, for a Gaussian stochastic process the two concepts are equivalent. A Gaussian stochastic process is strict-sense stationary if and only if it is wide-sense stationary.


Example

There is an explicit representation for stationary Gaussian processes. A simple example of this representation is X_t = \cos(at)\, \xi_1 + \sin(at)\, \xi_2 where \xi_1 and \xi_2 are independent random variables with the
standard normal distribution In probability theory and statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is f(x) = \frac e^ ...
.


Covariance functions

A key fact of Gaussian processes is that they can be completely defined by their second-order statistics. Thus, if a Gaussian process is assumed to have mean zero, defining the
covariance function In probability theory and statistics, the covariance function describes how much two random variables change together (their ''covariance'') with varying spatial or temporal separation. For a random field or stochastic process ''Z''(''x'') on a dom ...
completely defines the process' behaviour. Importantly the non-negative definiteness of this function enables its spectral decomposition using the Karhunen–Loève expansion. Basic aspects that can be defined through the covariance function are the process' stationarity,
isotropy In physics and geometry, isotropy () is uniformity in all orientations. Precise definitions depend on the subject area. Exceptions, or inequalities, are frequently indicated by the prefix ' or ', hence ''anisotropy''. ''Anisotropy'' is also u ...
,
smoothness In mathematical analysis, the smoothness of a function is a property measured by the number of continuous derivatives (''differentiability class)'' it has over its domain. A function of class C^k is a function of smoothness at least ; t ...
and periodicity. Stationarity refers to the process' behaviour regarding the separation of any two points x and x'. If the process is stationary, the covariance function depends only on x-x'. For example, the
Ornstein–Uhlenbeck process In mathematics, the Ornstein–Uhlenbeck process is a stochastic process with applications in financial mathematics and the physical sciences. Its original application in physics was as a model for the velocity of a massive Brownian particle ...
is stationary. If the process depends only on , x-x', , the Euclidean distance (not the direction) between x and x', then the process is considered isotropic. A process that is concurrently stationary and isotropic is considered to be
homogeneous Homogeneity and heterogeneity are concepts relating to the uniformity of a substance, process or image. A homogeneous feature is uniform in composition or character (i.e., color, shape, size, weight, height, distribution, texture, language, i ...
; in practice these properties reflect the differences (or rather the lack of them) in the behaviour of the process given the location of the observer. Ultimately Gaussian processes translate as taking priors on functions and the smoothness of these priors can be induced by the covariance function. If we expect that for "near-by" input points x and x' their corresponding output points y and y' to be "near-by" also, then the assumption of continuity is present. If we wish to allow for significant displacement then we might choose a rougher covariance function. Extreme examples of the behaviour is the Ornstein–Uhlenbeck covariance function and the squared exponential where the former is never differentiable and the latter infinitely differentiable. Periodicity refers to inducing periodic patterns within the behaviour of the process. Formally, this is achieved by mapping the input x to a two dimensional vector u(x) = \left( \cos(x), \sin(x) \right).


Usual covariance functions

There are a number of common covariance functions: *Constant : K_\operatorname(x,x') = C *Linear: K_\operatorname(x,x') = x^\mathsf x' *white Gaussian noise: K_\operatorname(x,x') = \sigma^2 \delta_ *Squared exponential: K_\operatorname(x,x') = \exp \left(-\tfrac \right) *Ornstein–Uhlenbeck: K_\operatorname(x,x') = \exp \left(-\tfrac \ell \right) *Matérn: K_\operatorname(x,x') = \tfrac \left(\tfrac \right)^\nu K_\nu \left(\tfrac \right) *Periodic: K_\operatorname(x,x') = \exp\left(-\tfrac \sin^2 (d/2) \right) *Rational quadratic: K_\operatorname(x,x') = \left(1+d^2\right)^, \quad \alpha \geq 0 Here d = , x- x', . The parameter \ell is the characteristic length-scale of the process (practically, "how close" two points x and x' have to be to influence each other significantly), ''\delta'' is the
Kronecker delta In mathematics, the Kronecker delta (named after Leopold Kronecker) is a function of two variables, usually just non-negative integers. The function is 1 if the variables are equal, and 0 otherwise: \delta_ = \begin 0 &\text i \neq j, \\ 1 &\ ...
and \sigma the
standard deviation In statistics, the standard deviation is a measure of the amount of variation of the values of a variable about its Expected value, mean. A low standard Deviation (statistics), deviation indicates that the values tend to be close to the mean ( ...
of the noise fluctuations. Moreover, K_\nu is the
modified Bessel function Bessel functions, named after Friedrich Bessel who was the first to systematically study them in 1824, are canonical solutions of Bessel's differential equation x^2 \frac + x \frac + \left(x^2 - \alpha^2 \right)y = 0 for an arbitrary complex ...
of order \nu and \Gamma(\nu) is the
gamma function In mathematics, the gamma function (represented by Γ, capital Greek alphabet, Greek letter gamma) is the most common extension of the factorial function to complex numbers. Derived by Daniel Bernoulli, the gamma function \Gamma(z) is defined ...
evaluated at \nu. Importantly, a complicated covariance function can be defined as a linear combination of other simpler covariance functions in order to incorporate different insights about the data-set at hand. The inferential results are dependent on the values of the hyperparameters \theta (e.g. \ell and \sigma) defining the model's behaviour. A popular choice for \theta is to provide ''
maximum a posteriori An estimation procedure that is often claimed to be part of Bayesian statistics is the maximum a posteriori (MAP) estimate of an unknown quantity, that equals the mode of the posterior density with respect to some reference measure, typically ...
'' (MAP) estimates of it with some chosen prior. If the prior is very near uniform, this is the same as maximizing the
marginal likelihood A marginal likelihood is a likelihood function that has been integrated over the parameter space. In Bayesian statistics, it represents the probability of generating the observed sample for all possible values of the parameters; it can be under ...
of the process; the marginalization being done over the observed process values y. This approach is also known as ''maximum likelihood II'', ''evidence maximization'', or '' empirical Bayes''.


Continuity

For a Gaussian process, continuity in probability is equivalent to mean-square continuity and continuity with probability one is equivalent to sample continuity. The latter implies, but is not implied by, continuity in probability. Continuity in probability holds if and only if the mean and autocovariance are continuous functions. In contrast, sample continuity was challenging even for stationary Gaussian processes (as probably noted first by
Andrey Kolmogorov Andrey Nikolaevich Kolmogorov ( rus, Андре́й Никола́евич Колмого́ров, p=ɐnˈdrʲej nʲɪkɐˈlajɪvʲɪtɕ kəlmɐˈɡorəf, a=Ru-Andrey Nikolaevich Kolmogorov.ogg, 25 April 1903 – 20 October 1987) was a Soviet ...
), and more challenging for more general processes. As usual, by a sample continuous process one means a process that admits a sample continuous
modification Modification may refer to: * Modifications of school work for students with special educational needs * Modifications (genetics), changes in appearance arising from changes in the environment * Posttranslational modifications, changes to prote ...
.


Stationary case

For a stationary Gaussian process X=(X_t)_, some conditions on its spectrum are sufficient for sample continuity, but fail to be necessary. A necessary and sufficient condition, sometimes called Dudley–Fernique theorem, involves the function \sigma defined by \sigma(h) = \sqrt (the right-hand side does not depend on t due to stationarity). Continuity of X in probability is equivalent to continuity of \sigma at 0. When convergence of \sigma(h) to 0 (as h\to 0) is too slow, sample continuity of X may fail. Convergence of the following integrals matters: I(\sigma) = \int_0^1 \frac \, dh = \int_0^\infty 2\sigma( e^) \, dx , these two integrals being equal according to
integration by substitution In calculus, integration by substitution, also known as ''u''-substitution, reverse chain rule or change of variables, is a method for evaluating integrals and antiderivatives. It is the counterpart to the chain rule for differentiation, and c ...
h = e^, x = \sqrt . The first integrand need not be bounded as h\to 0+, thus the integral may converge (I(\sigma)<\infty) or diverge (I(\sigma)=\infty). Taking for example \sigma( e^) = \tfrac for large x, that is, \sigma(h) = (\log(1/h))^ for small h, one obtains I(\sigma)<\infty when a>1, and I(\sigma)=\infty when 0 < a\le 1. In these two cases the function \sigma is increasing on [0,\infty), but generally it is not. Moreover, the condition does not follow from continuity of \sigma and the evident relations \sigma(h) \ge 0 (for all h) and \sigma(0) = 0. Some history. Sufficiency was announced by Xavier Fernique in 1964, but the first proof was published by Richard M. Dudley in 1967. Necessity was proved by Michael B. Marcus and Lawrence Shepp in 1970. There exist sample continuous processes X such that I(\sigma)=\infty; they violate condition (*). An example found by Marcus and Shepp is a random lacunary Fourier series X_t = \sum_^\infty c_n ( \xi_n \cos \lambda_n t + \eta_n \sin \lambda_n t ) , where \xi_1,\eta_1,\xi_2,\eta_2,\dots are independent random variables with
standard normal distribution In probability theory and statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is f(x) = \frac e^ ...
; frequencies 0<\lambda_1<\lambda_2<\dots are a fast growing sequence; and coefficients c_n>0 satisfy \sum_n c_n < \infty. The latter relation implies \sum_n c_n ( , \xi_n, + , \eta_n, ) = \sum_n c_n \xi_n, + , \eta_n, = \text \cdot \sum_n c_n < \infty, whence \sum_n c_n ( , \xi_n, + , \eta_n, ) < \infty almost surely, which ensures uniform convergence of the Fourier series almost surely, and sample continuity of X. Its autocovariation function _t X_= \sum_^\infty c_n^2 \cos \lambda_n h is nowhere monotone (see the picture), as well as the corresponding function \sigma, \sigma(h) = \sqrt = 2 \sqrt .


Brownian motion as the integral of Gaussian processes

A
Wiener process In mathematics, the Wiener process (or Brownian motion, due to its historical connection with Brownian motion, the physical process of the same name) is a real-valued continuous-time stochastic process discovered by Norbert Wiener. It is one o ...
(also known as Brownian motion) is the integral of a white noise generalized Gaussian process. It is not stationary, but it has stationary increments. The
Ornstein–Uhlenbeck process In mathematics, the Ornstein–Uhlenbeck process is a stochastic process with applications in financial mathematics and the physical sciences. Its original application in physics was as a model for the velocity of a massive Brownian particle ...
is a stationary Gaussian process. The Brownian bridge is (like the Ornstein–Uhlenbeck process) an example of a Gaussian process whose increments are not
independent Independent or Independents may refer to: Arts, entertainment, and media Artist groups * Independents (artist group), a group of modernist painters based in Pennsylvania, United States * Independentes (English: Independents), a Portuguese artist ...
. The fractional Brownian motion is a Gaussian process whose covariance function is a generalisation of that of the Wiener process.


RKHS structure and Gaussian process

Let f be a mean-zero Gaussian process \left\ with a non-negative definite covariance function K and let R be a symmetric and positive semidefinite function. Then, there exists a Gaussian process X which has the covariance R. Moreover, the
reproducing kernel Hilbert space In functional analysis, a reproducing kernel Hilbert space (RKHS) is a Hilbert space of functions in which point evaluation is a continuous linear functional. Specifically, a Hilbert space H of functions from a set X (to \mathbb or \mathbb) is ...
(RKHS) associated to R coincides with the
Cameron–Martin theorem In mathematics, the Cameron–Martin theorem or Cameron–Martin formula (named after Robert Horton Cameron and W. T. Martin) is a theorem of measure theory that describes how abstract Wiener measure changes under translation by certain eleme ...
associated space R(H) of X, and all the spaces R(H), H_X, and \mathcal(K) are isometric. From now on, let \mathcal(R) be a
reproducing kernel Hilbert space In functional analysis, a reproducing kernel Hilbert space (RKHS) is a Hilbert space of functions in which point evaluation is a continuous linear functional. Specifically, a Hilbert space H of functions from a set X (to \mathbb or \mathbb) is ...
with positive definite kernel R. Driscoll's zero-one law is a result characterizing the sample functions generated by a Gaussian process: \lim_ \operatorname _n R_n^< \infty, where K_n and R_n are the covariance matrices of all possible pairs of n points, implies \Pr \in \mathcal(R)= 1. Moreover, \lim_ \operatorname _n R_n^= \infty implies \Pr \in \mathcal(R)= 0. This has significant implications when K = R, as \lim_ \operatorname _n R_n^= \lim_\operatorname = \lim_ n = \infty. As such, almost all sample paths of a mean-zero Gaussian process with positive definite kernel K will lie outside of the Hilbert space \mathcal(K).


Linearly constrained Gaussian processes

For many applications of interest some pre-existing knowledge about the system at hand is already given. Consider e.g. the case where the output of the Gaussian process corresponds to a magnetic field; here, the real magnetic field is bound by Maxwell's equations and a way to incorporate this constraint into the Gaussian process formalism would be desirable as this would likely improve the accuracy of the algorithm. A method on how to incorporate linear constraints into Gaussian processes already exists: Consider the (vector valued) output function f(x) which is known to obey the linear constraint (i.e. \mathcal_X is a linear operator) \mathcal_X(f(x)) = 0. Then the constraint \mathcal_X can be fulfilled by choosing f(x) = \mathcal_X(g(x)), where g(x) \sim \mathcal(\mu_g, K_g) is modelled as a Gaussian process, and finding \mathcal_X such that \mathcal_X(\mathcal_X(g)) = 0 \qquad \forall g. Given \mathcal_X and using the fact that Gaussian processes are closed under linear transformations, the Gaussian process for f obeying constraint \mathcal_X becomes f(x) = \mathcal_X g \sim \mathcal ( \mathcal_X \mu_g, \mathcal_X K_g \mathcal_^\mathsf ). Hence, linear constraints can be encoded into the mean and covariance function of a Gaussian process.


Applications

A Gaussian process can be used as a
prior probability distribution A prior probability distribution of an uncertain quantity, simply called the prior, is its assumed probability distribution before some evidence is taken into account. For example, the prior could be the probability distribution representing the ...
over functions in
Bayesian inference Bayesian inference ( or ) is a method of statistical inference in which Bayes' theorem is used to calculate a probability of a hypothesis, given prior evidence, and update it as more information becomes available. Fundamentally, Bayesian infer ...
. Given any set of ''N'' points in the desired domain of your functions, take a
multivariate Gaussian In probability theory and statistics, the multivariate normal distribution, multivariate Gaussian distribution, or joint normal distribution is a generalization of the one-dimensional (univariate) normal distribution to higher dimensions. One de ...
whose covariance
matrix Matrix (: matrices or matrixes) or MATRIX may refer to: Science and mathematics * Matrix (mathematics), a rectangular array of numbers, symbols or expressions * Matrix (logic), part of a formula in prenex normal form * Matrix (biology), the m ...
parameter is the Gram matrix of your ''N'' points with some desired kernel, and sample from that Gaussian. For solution of the multi-output prediction problem, Gaussian process regression for vector-valued function was developed. In this method, a 'big' covariance is constructed, which describes the correlations between all the input and output variables taken in ''N'' points in the desired domain. This approach was elaborated in detail for the matrix-valued Gaussian processes and generalised to processes with 'heavier tails' like Student-t processes. Inference of continuous values with a Gaussian process prior is known as Gaussian process regression, or
kriging In statistics, originally in geostatistics, kriging or Kriging (), also known as Gaussian process regression, is a method of interpolation based on Gaussian process governed by prior covariances. Under suitable assumptions of the prior, kriging g ...
; extending Gaussian process regression to multiple target variables is known as ''cokriging''. Gaussian processes are thus useful as a powerful non-linear multivariate
interpolation In the mathematics, mathematical field of numerical analysis, interpolation is a type of estimation, a method of constructing (finding) new data points based on the range of a discrete set of known data points. In engineering and science, one ...
tool. Kriging is also used to extend Gaussian process in the case of mixed integer inputs. Gaussian processes are also commonly used to tackle numerical analysis problems such as numerical integration, solving differential equations, or optimisation in the field of
probabilistic numerics Probabilistic numerics is aactivefield of study at the intersection of applied mathematics, statistics, and machine learning centering on the concept of uncertainty in computation. In probabilistic numerics, tasks in numerical analysis such as find ...
. Gaussian processes can also be used in the context of mixture of experts models, for example. The underlying rationale of such a learning framework consists in the assumption that a given mapping cannot be well captured by a single Gaussian process model. Instead, the observation space is divided into subsets, each of which is characterized by a different mapping function; each of these is learned via a different Gaussian process component in the postulated mixture. In the natural sciences, Gaussian processes have found use as probabilistic models of astronomical time series and as predictors of molecular properties. They are also being increasingly used as surrogate models for force field optimization.


Gaussian process prediction, or Kriging

When concerned with a general Gaussian process regression problem (Kriging), it is assumed that for a Gaussian process f observed at coordinates x, the vector of values is just one sample from a multivariate Gaussian distribution of dimension equal to number of observed coordinates . Therefore, under the assumption of a zero-mean distribution, , where is the covariance matrix between all possible pairs for a given set of hyperparameters ''θ''. As such the log marginal likelihood is: \log p(f(x')\mid\theta,x) = -\frac \left(f(x)^\mathsf K(\theta,x,x')^ f(x') + \log \det(K(\theta,x,x')) + n \log 2\pi \right) and maximizing this marginal likelihood towards provides the complete specification of the Gaussian process . One can briefly note at this point that the first term corresponds to a penalty term for a model's failure to fit observed values and the second term to a penalty term that increases proportionally to a model's complexity. Having specified , making predictions about unobserved values at coordinates is then only a matter of drawing samples from the predictive distribution p(y^*\mid x^*,f(x),x) = N(y^*\mid A,B) where the posterior mean estimate is defined as A = K(\theta,x^*,x) K(\theta,x,x')^ f(x) and the posterior variance estimate ''B'' is defined as: B = K(\theta,x^*,x^*) - K(\theta,x^*,x) K(\theta,x,x')^ K(\theta,x^*,x)^\mathsf where is the covariance between the new coordinate of estimation ''x''* and all other observed coordinates ''x'' for a given hyperparameter vector , and are defined as before and is the variance at point as dictated by . It is important to note that practically the posterior mean estimate of (the "point estimate") is just a linear combination of the observations ; in a similar manner the variance of is actually independent of the observations . A known bottleneck in Gaussian process prediction is that the computational complexity of inference and likelihood evaluation is cubic in the number of points , ''x'', , and as such can become unfeasible for larger data sets. Works on sparse Gaussian processes, that usually are based on the idea of building a ''representative set'' for the given process ''f'', try to circumvent this issue. The
kriging In statistics, originally in geostatistics, kriging or Kriging (), also known as Gaussian process regression, is a method of interpolation based on Gaussian process governed by prior covariances. Under suitable assumptions of the prior, kriging g ...
method can be used in the latent level of a nonlinear mixed-effects model for a spatial functional prediction: this technique is called the latent kriging. Other classes of scalable Gaussian process for analyzing massive datasets have emerged from the Vecchia approximation and Nearest Neighbor Gaussian Processes (NNGP). Often, the covariance has the form K(\theta, x,x') = \frac \tilde(\theta,x,x'), where \sigma^2 is a scaling parameter. Examples are the Matérn class covariance functions. If this scaling parameter \sigma^2 is either known or unknown (i.e. must be marginalized), then the posterior probability, p(\theta \mid D), i.e. the probability for the hyperparameters \theta given a set of data pairs D of observations of x and f(x), admits an analytical expression.


Bayesian neural networks as Gaussian processes

Bayesian neural networks are a particular type of
Bayesian network A Bayesian network (also known as a Bayes network, Bayes net, belief network, or decision network) is a probabilistic graphical model that represents a set of variables and their conditional dependencies via a directed acyclic graph (DAG). Whi ...
that results from treating
deep learning Deep learning is a subset of machine learning that focuses on utilizing multilayered neural networks to perform tasks such as classification, regression, and representation learning. The field takes inspiration from biological neuroscience a ...
and
artificial neural network In machine learning, a neural network (also artificial neural network or neural net, abbreviated ANN or NN) is a computational model inspired by the structure and functions of biological neural networks. A neural network consists of connected ...
models probabilistically, and assigning a
prior distribution A prior probability distribution of an uncertain quantity, simply called the prior, is its assumed probability distribution before some evidence is taken into account. For example, the prior could be the probability distribution representing the ...
to their parameters. Computation in artificial neural networks is usually organized into sequential layers of
artificial neuron An artificial neuron is a mathematical function conceived as a model of a biological neuron in a neural network. The artificial neuron is the elementary unit of an ''artificial neural network''. The design of the artificial neuron was inspired ...
s. The number of neurons in a layer is called the layer width. As layer width grows large, many Bayesian neural networks reduce to a Gaussian process with a closed form compositional kernel. This Gaussian process is called the Neural Network Gaussian Process (NNGP) (not to be confused with the Nearest Neighbor Gaussian Process ). It allows predictions from Bayesian neural networks to be more efficiently evaluated, and provides an analytic tool to understand
deep learning Deep learning is a subset of machine learning that focuses on utilizing multilayered neural networks to perform tasks such as classification, regression, and representation learning. The field takes inspiration from biological neuroscience a ...
models.


Computational issues

In practical applications, Gaussian process models are often evaluated on a grid leading to multivariate normal distributions. Using these models for prediction or parameter estimation using maximum likelihood requires evaluating a multivariate Gaussian density, which involves calculating the determinant and the inverse of the covariance matrix. Both of these operations have cubic computational complexity which means that even for grids of modest sizes, both operations can have a prohibitive computational cost. This drawback led to the development of multiple approximation methods.


See also

* Bayes linear statistics * Bayesian interpretation of regularization *
Kriging In statistics, originally in geostatistics, kriging or Kriging (), also known as Gaussian process regression, is a method of interpolation based on Gaussian process governed by prior covariances. Under suitable assumptions of the prior, kriging g ...
* Gaussian free field * Gauss–Markov process * Gradient-enhanced kriging (GEK) * Student's t-process


References


External links


Literature


The Gaussian Processes Web Site, including the text of Rasmussen and Williams' Gaussian Processes for Machine Learning
*
A Review of Gaussian Random Fields and Correlation Functions

Efficient Reinforcement Learning using Gaussian Processes


Software


GPML: A comprehensive Matlab toolbox for GP regression and classification

STK: a Small (Matlab/Octave) Toolbox for Kriging and GP modeling

Kriging module in UQLab framework (Matlab)

CODES Toolbox: implementations of Kriging, variational kriging and multi-fidelity models (Matlab)

Matlab/Octave function for stationary Gaussian fields

Yelp MOE – A black box optimization engine using Gaussian process learning

ooDACE
– A flexible object-oriented Kriging Matlab toolbox.
GPstuff – Gaussian process toolbox for Matlab and Octave

GPy – A Gaussian processes framework in Python

GSTools - A geostatistical toolbox, including Gaussian process regression, written in Python

Interactive Gaussian process regression demo

Basic Gaussian process library written in C++11

scikit-learn
– A machine learning library for Python which includes Gaussian process regression and classification
SAMBO Optimization
library for Python supports sequential optimization driven by Gaussian process regressor from
scikit-learn scikit-learn (formerly scikits.learn and also known as sklearn) is a free and open-source machine learning library for the Python programming language. It features various classification, regression and clustering algorithms including support ...
.

- The Kriging toolKit (KriKit) is developed at the Institute of Bio- and Geosciences 1 (IBG-1) of Forschungszentrum Jülich (FZJ)


Video tutorials


Gaussian Process Basics by David MacKay

Learning with Gaussian Processes by Carl Edward Rasmussen

Bayesian inference and Gaussian processes by Carl Edward Rasmussen
{{DEFAULTSORT:Gaussian Process Stochastic processes Kernel methods for machine learning Nonparametric Bayesian statistics Normal distribution