Stochastic processes can be represented from (at least) two perspectives:
Random field is a measurable function on a product space, where one of the subdomains is a probability space, to the real line.
Symbolically, random field $\alpha: D\times \Omega \to \mathbb{R}$ is a $(\mathcal{S} \times \Sigma, \mathcal{B})$-measurable or $( (\mathcal{S} \times \Sigma)^∗ , \mathcal{L} )$-measurable function. Here $(D, \mathcal{S}, \mu)$ is a deterministic measure space, and $(\Omega, \Sigma, P)$ is a probability space. Their product space is $( D \times \Omega, \mathcal{S} \times \Sigma, \mu \times P)$, with completion $( D \times \Omega, (\mathcal{S} \times \Sigma)^{*}, \mu \times P)$. Either Borel measure or Lebesgue measure is applied to the real line.
Think of the deterministic field $D$ as an index set. When $D$ is finite, the RF can be seen as a random vector, i.e. a finite collection of random variables, or a random sample. When the number of samples approaches infinity, the RF can be seen as a random sequence. When $D$ is a space with power of the continuum, e.g. the real line or a Euclidean space, the RF is often seen as a general random process.
Given an undirected graph $G=(V,E)$, a set of random variables $X = {X_v}, v\in V$ form a Markov random field (MRF) with respect to $G$ if they satisfy the local Markov properties: $P(X_i = x_i \mid X_j = x_j, j = V \setminus \{i\}) = P(X_i = x_i \mid X_j = x_j, j = \partial_i)$.
A Gibbs random field (GRF) is a set of random variables that if all degrees of freedom outside an arbitrary subset are frozen, the canonical ensemble for the subset subject to these boundary conditions matches the probabilities in the Gibbs measure conditional on the frozen degrees of freedom: $P(X=x) = \frac{1}{Z(\beta)} \exp ( - \beta E(x))$.
The fundamental theorem of random fields (Hammersley–Clifford theorem): [@Hammersley1971] Given an undirected graph, any positive probability measure with the Markov properties is a Gibbs measure for some locally-defined energy function.
Gaussian process is a real random process, of which all finite-dimensional distributions are Gaussian. That is, for any finite collection of instances $t_1, \dots, t_n \in T$, the characteristic function of random vector $(X(t_1), \dots, X(t_n))$ has the form $\phi(u_1, \dots, u_n; t_1, \dots, t_n) = \exp{ i \sum_k A(t_k) u_k - \frac{1}{2} \sum_{k,j} B(t_k, t_j) u_k u_j }$, where $A(t)$ is the mathematical expectation and $B(t, s)$ is the covariance function. Gaussian processes can be seen as an infinite-dimensional generalization of multivariate Gaussian distributions. Gaussian processes are useful in statistical modelling for its analytical convenience.
Gaussian process regression, Wiener–Kolmogorov prediction, or kriging [@Krige1951; @Matheron1960], is a method of interpolation where the interpolated values are modeled by a Gaussian process governed by prior covariances. Simple kriging starts with a prior distribution over functions as a stationary Gaussian process with mean 0 and covariance function $C(d)$. Regarding the observations as a single realization of a random field, each with a Gaussian likelihood function, the posterior distribution is another Gaussian process which can be simply computed. Kriging can be seen as a spline in a reproducing kernel Hilbert space (RKHS) whose reproducing kernel is the covariance function $C(d)$.
Kriging originated in geostatistics, and is widely used in computer models [@Kennedy2001]. It is the best linear unbiased predictor (BLUP) under certain (strong) assumptions on the priors. Besides, there might be better nonlinear and/or biased methods. In case of no spatial dependence, the kriging interpolation is only as good as the arithmetic mean.
Empirical process is a random process from the underlying set $X$ of a random variable $(X, Σ, P)$ to the scaled deviation of the empirical measure $P_n$: $\mathcal{G}(x) = \sqrt{n} (F_n(x) - F(x))$. The (classical) empirical process can be generalized to the sigma-algebra $Σ$ or classes of measurable functions ${f: (X, Σ) \mapsto (\mathbb{R}, \mathcal{B(T_d)})}$. Empirical processes are used to prove propositions uniform in the underlying set, esp. Glivenko–Cantelli theorems (uniform laws of large numbers), laws of the iterated logarithm, central limit theorems, and probability inequalities. For certain classes (Vapnik–Chervonenkis, Donsker) of the underlying set, the empirical process converges weakly to a Gaussian process.
(Chap. 14.2, 15.1)