1. Introduction
With increasing availability of multivariate data and considerable improvements in computational feasibility, multivariate random fields have become a significant part of geostatistical modelling in recent years.
These random fields are usually assumed to be either second-order stationary or intrinsically stationary. An m-variate random field
is called second-order stationary if it has a constant mean and if its auto- and cross-covariances
exist and are functions of the lag $\boldsymbol{h}$ only. It is called intrinsically stationary if the increment process $\{\boldsymbol{Z}(\boldsymbol{x}+\boldsymbol{h})-\boldsymbol{Z}(\boldsymbol{x}), \boldsymbol{x} \in {{\mathbb{R}}}^d\}$ is second-order stationary for all $\boldsymbol{h} \in {{\mathbb{R}}}^d$ . In this case, the function ${\boldsymbol {\tilde\gamma}}\,:\, {{\mathbb{R}}}^d \rightarrow {{\mathbb{R}}}^{m \times m}$ ,
is well-defined and is called a cross-variogram [Reference Myers19]. If we additionally assume that $Z_i(\boldsymbol{x}+\boldsymbol{h})-Z_j(\boldsymbol{x})$ is square integrable and that ${Var}(Z_i(\boldsymbol{x} + \boldsymbol{h}) - Z_j(\boldsymbol{x}))$ does not depend on $\boldsymbol{x}$ for all $\boldsymbol{x}, \boldsymbol{h}\in {{\mathbb{R}}}^d$ , $i,j=1,\ldots,m$ , then we can also define the so-called pseudo cross-variogram ${\boldsymbol \gamma}\,:\, {{\mathbb{R}}}^d \rightarrow {{\mathbb{R}}}^{m \times m}$ [Reference Myers20] via
Obviously, the diagonal entries of pseudo cross-variograms and cross-variograms coincide and contain univariate variograms $\gamma_{ii}(\boldsymbol{h})= \frac12{Var}(Z_i(\boldsymbol{x}+\boldsymbol{h})-Z_i(\boldsymbol{x}))$ , $i=1,\ldots,m$ [Reference Matheron15, Reference Matheron16].
Both cross- and pseudo cross-variograms are commonly used in geostatistics to capture the degree of spatial dependence [Reference Chen and Genton5]. There is some controversy as to which one to use, since both have their benefits and drawbacks. The cross-variogram, on the one hand, is well-defined under weaker assumptions, but requires measurements of the quantities of interest at the same locations for estimation in practical applications [Reference Chen and Genton5]. Moreover, it only reproduces the symmetric part of a cross-covariance function of a stationary random field; see e.g. [Reference Wackernagel30]. The pseudo cross-variogram, on the other hand, can capture asymmetry, and provides optimal co-kriging predictors without imposing any symmetry assumption on the cross-dependence structure [Reference Cressie and Wikle6, Reference Ver Hoef and Cressie29], but is difficult to interpret in practice due to considering differences of generally different physical quantities; cf. [Reference Cressie and Wikle6] and their account of it.
From a theoretical perspective, pseudo cross-variograms are interesting objects, since they are not only found in multivariate geostatistics but also appear naturally in extreme value theory in the context of multivariate Brown–Resnick processes [Reference Genton, Padoan and Sang8, Reference Oesting, Schlather and Friederichs21]. However, pseudo cross-variograms, in contrast to cross-variograms, have not yet been sufficiently well understood. So far, elementary properties [Reference Du and Ma7] are known, such as their relation to cross-variograms and cross-covariance functions [Reference Myers20, Reference Papritz, Künsch and Webster22], their applicability to co-kriging [Reference Cressie and Wikle6, Reference Myers20, Reference Ver Hoef and Cressie29], and limiting behaviour [Reference Oesting, Schlather and Friederichs21, Reference Papritz, Künsch and Webster22], but a concise necessary and sufficient criterion for a matrix-valued function to be a pseudo cross-variogram is missing. This lack of an equivalent characterization makes it immensely difficult to show the validity of a function as a pseudo cross-variogram (cf. e.g. [Reference Ma13, p. 239]), unless it can be led back to an explicit construction of a random field as in [Reference Chen and Genton5] or [Reference Oesting, Schlather and Friederichs21].
Equivalent characterizations are well known for univariate variograms (see e.g. [Reference Gneiting, Sasvári and Schlather11], [Reference Matheron15]), and involve the notion of conditional negative definiteness. These characteristics are intimately connected to a result which can be mainly attributed to Schoenberg [Reference Berg, Christensen and Ressel2, Reference Schoenberg28], implying that a function $\gamma\,:\, {{\mathbb{R}}}^d \rightarrow {{\mathbb{R}}}$ is a univariate variogram if and only if $\exp({-}t\gamma)$ is a correlation function for all $t >0$ [Reference Gneiting, Sasvári and Schlather11]. Such a characterization of $\gamma$ in the multivariate case, however, seems to be untreated in the geostatistical literature. For cross-variograms, there is a result for the ‘if’ part [Reference Ma14, Theorem 10]. The ‘only if’ part is false in general; see e.g. [Reference Schlather27, Remark 2]. See also [Reference Du and Ma7].
The aim of this article is to fill these gaps. The key ingredient is to apply a stronger notion of conditional negative definiteness for matrix-valued functions than the predominant one in geostatistical literature. We discuss this notion in Section 2, and provide a first characterization of pseudo cross-variograms in these terms. This characterization leads to a Schoenberg-type result in terms of pseudo cross-variograms in Section 3, thus making a case for proponents of pseudo cross-variograms, at least from a theoretical standpoint. In Section 4 we apply this characterization and illustrate its power by extending versions of the very popular space–time covariance model of Gneiting [Reference Gneiting10] to the multivariate case.
Our presentation here is carried out in terms of pseudo cross-variograms in their original stationary form as introduced above. It is important to note that, with the exception of Corollary 2, all results presented here, which involve pseudo cross-variograms or conditionally negative definite matrix-valued functions as defined below, also hold for their respective non-stationary versions or kernel-based forms by straightforward adaptations. A non-stationary version of Corollary 2 is also available. The proofs follow the same lines.
2. Conditional negative definiteness for matrix-valued functions
Real-valued conditionally negative definite functions are essential to characterizing univariate variograms. A function $\gamma\,:\, {{\mathbb{R}}}^d \rightarrow {{\mathbb{R}}}$ is a univariate variogram if and only if $\gamma(0)=0$ and $\gamma$ is conditionally negative definite, that is, $\gamma$ is symmetric and for all $n \ge 2$ , $\boldsymbol{x}_{\textbf{1}},\ldots,\boldsymbol{x}_{\boldsymbol{n}} \in {{\mathbb{R}}}^d$ , $a_1,\ldots,a_n \in {{\mathbb{R}}}$ such that $\sum_{k=1}^n a_k=0$ , the inequality $\sum_{i=1}^n\sum_{j=1}^n a_i \gamma(\boldsymbol{x}_{\boldsymbol{i}}-\boldsymbol{x}_{\boldsymbol{j}})a_j \le 0$ holds [Reference Matheron15]. An extended notion of conditional negative definiteness for matrix-valued functions is part of a characterization of cross-variograms. A function ${\boldsymbol {\tilde\gamma}}\,:\, {{\mathbb{R}}}^d \rightarrow {{\mathbb{R}}}^{m \times m}$ is a cross-variogram if and only if ${\boldsymbol {\tilde\gamma}}(\textbf{0}) =\textbf{0}$ , ${\boldsymbol {\tilde\gamma}}(\boldsymbol{h})={\boldsymbol {\tilde\gamma}}({-}\boldsymbol{h})={\boldsymbol {\tilde\gamma}}(\boldsymbol{h})^\top$ and
for $n \ge 2$ , $\boldsymbol{x}_{\textbf{1}},\ldots,\boldsymbol{x}_{\boldsymbol{n}} \in {{\mathbb{R}}}^d$ , $\boldsymbol{a}_{\textbf{1}},\ldots,\boldsymbol{a}_{\boldsymbol{n}} \in {{\mathbb{R}}}^m$ such that $\sum_{k=1}^n \boldsymbol{a}_{\boldsymbol{k}} =\textbf{0}$ [Reference Ma14]. A function satisfying condition (1) is called an almost negative definite matrix-valued function in [Reference Xie31, p. 40].
A pseudo cross-variogram ${\boldsymbol \gamma}$ has similar, but only necessary properties; see [Reference Du and Ma7]. It holds that $\gamma_{ii}(\textbf{0})=0$ , and $\gamma_{ij}(\boldsymbol{h})=\gamma_{ji}({-}\boldsymbol{h})$ , $i,j=1,\ldots,m$ . Additionally, a pseudo cross-variogram is an almost negative definite matrix-valued function as well, but inequality (1), loosely speaking, cannot enforce non-negativity on the secondary diagonals. Therefore we consider the following stronger notion of conditional negative definiteness; see [Reference Gesztesy and Pang9].
Definition 1. A function ${\boldsymbol \gamma}\,:\, {{\mathbb{R}}}^d \rightarrow {{\mathbb{R}}}^{m \times m}$ is called conditionally negative definite if
for all $n \in \mathbb{N}$ , $\boldsymbol{x}_{\textbf{1}},\ldots,\boldsymbol{x}_{\boldsymbol{n}} \in {{\mathbb{R}}}^d$ , $\boldsymbol{a}_{\textbf{1}},\ldots,\boldsymbol{a}_{\boldsymbol{n}} \in {{\mathbb{R}}}^m$ such that
with $\textbf{1}_{\boldsymbol{m}}\,:\!=\, (1,\ldots, 1)^\top \in {{\mathbb{R}}}^m$ .
Obviously, the set of conditionally negative definite matrix-valued functions is a convex cone which is closed under integration and pointwise limits, if existing. In the univariate case, the concepts of conditionally and almost negative definite functions coincide, reproducing the traditional notion of real-valued conditionally negative definite functions. The main difference between them is the broader spectrum of vectors for which inequality (2b) has to hold, in that the sum of all components has to be zero instead of each component of the sum itself. This modification in particular includes sets of linearly independent vectors in the pool of admissible test vector families, resulting in more restrictive conditions on the secondary diagonals. Indeed, choosing $n=2$ , $\boldsymbol{x}_{\textbf{1}}=\boldsymbol{h} \in {{\mathbb{R}}}^d$ , $\boldsymbol{x}_{\textbf{2}}=\textbf{0}$ , and $\boldsymbol{a}_{\textbf{1}}=\boldsymbol{e}_{\boldsymbol{i}}$ , $\boldsymbol{a}_{\textbf{2}}=-\boldsymbol{e}_{\boldsymbol{j}}$ in Definition 1 with $\{\boldsymbol{e}_{\textbf{1}},\ldots,\boldsymbol{e}_{\boldsymbol{m}}\}$ denoting the canonical basis in ${{\mathbb{R}}}^m$ , we have $\gamma_{ij}(\boldsymbol{h}) \ge 0$ for a conditionally negative definite function ${\boldsymbol \gamma}\,:\, {{\mathbb{R}}}^d \rightarrow {{\mathbb{R}}}^{m \times m}$ with $\gamma_{ii}(\textbf{0}) = 0$ , $i,j=1,\ldots,m$ , fitting the non-negativity of a pseudo cross-variogram. In fact, the latter condition on the main diagonal and the conditional negative definiteness property are sufficient to characterize pseudo cross-variograms.
Theorem 1. Let ${\boldsymbol \gamma}\,:\, {{\mathbb{R}}}^d \rightarrow {{\mathbb{R}}}^{m \times m}$ . Then there exists a centred Gaussian random field $\boldsymbol{Z}$ on ${{\mathbb{R}}}^d$ with pseudo cross-variogram ${\boldsymbol \gamma}$ if and only if $\gamma_{ii}(\textbf{0})=0$ , $i=1,\ldots,m$ , and ${\boldsymbol \gamma}$ is conditionally negative definite.
Proof. The proof is analogous to the univariate one in [Reference Matheron15]. Let $\boldsymbol{Z}$ be an m-variate random field with pseudo cross-variogram ${\boldsymbol \gamma}$ . Obviously, $\gamma_{ii}(\textbf{0}) = 0$ and $\gamma_{ij}(\boldsymbol{h})=\gamma_{ji}({-}\boldsymbol{h})$ for all $\boldsymbol{h} \in {{\mathbb{R}}}^d$ , $i,j=1,\ldots, m$ . Define an m-variate random field $\tilde{\boldsymbol{Z}}$ via $\tilde Z_i(\boldsymbol{x})=Z_i(\boldsymbol{x})-Z_1(\textbf{0})$ , $\boldsymbol{x} \in {{\mathbb{R}}}^d$ , $i=1,\ldots,m$ . Then $\boldsymbol{Z}$ and $\tilde{\boldsymbol{Z}}$ have the same pseudo cross-variogram, and
(see also [Reference Papritz, Künsch and Webster22, equation (6)]), that is,
with $\boldsymbol{\gamma}_{\textbf{1}}(\boldsymbol{x})\,:\!=\, (\gamma_{11}(\boldsymbol{x}),\ldots,\gamma_{m1}(\boldsymbol{x}))^\top$ . For $\textbf{1}_{\boldsymbol{m}}^\top \sum_{k=1}^m \boldsymbol{a}_{\boldsymbol{k}}=0$ , we thus have
Now let ${\boldsymbol \gamma}$ be conditionally negative definite and $\gamma_{ii}(\textbf{0})=0$ , $i=1,\ldots,m$ . Let $\boldsymbol{a}_{\textbf{1}},\ldots, \boldsymbol{a}_{\boldsymbol{n}} \in {{\mathbb{R}}}^m, \boldsymbol{x}_{\textbf{1}}, \ldots,\boldsymbol{x}_{\boldsymbol{n}} \in {{\mathbb{R}}}^d$ be arbitrary, $\boldsymbol{x}_{\textbf{0}}=\textbf{0} \in {{\mathbb{R}}}^d$ and
Then
Since $\gamma_{11}(\textbf{0})=0$ , and $\boldsymbol{a}_{\textbf{0}}^\top {\boldsymbol \gamma}(\boldsymbol{x}_{\textbf{0}} -\boldsymbol{x}_{\boldsymbol{j}}) \boldsymbol{a}_{\boldsymbol{j}} =\boldsymbol{a}_{\boldsymbol{j}}^\top {\boldsymbol \gamma}(\boldsymbol{x}_{\boldsymbol{j}}) \boldsymbol{a}_{\textbf{0}}$ due to property (2a), we get that
that is,
is a matrix-valued positive definite kernel. Let
be a corresponding centred Gaussian random field. We have to show that ${Var}(Z_i(\boldsymbol{x}+\boldsymbol{h}) - Z_j(\boldsymbol{x}))$ is independent of $\boldsymbol{x}$ for all $\boldsymbol{x}, \boldsymbol{h} \in {{\mathbb{R}}}^d$ , $i,j=1,\ldots,m$ . We even show that $\boldsymbol{x} \mapsto Z_i(\boldsymbol{x}+\boldsymbol{h}) - Z_j(\boldsymbol{x})$ is weakly stationary for $i,j=1,\ldots,m$ :
Theorem 1 answers the questions raised in [Reference Du and Ma7, p. 422] and also settles a question in [Reference Ma13, p. 239] in a more general framework with regard to the intersection of the sets of pseudo and cross-variograms. It turns out that this intersection is trivial in the following sense.
Corollary 1. Let
Then we have
Proof. Let ${\boldsymbol \gamma} \in \mathcal{P} \cap \mathcal{C}$ . Without loss of generality, assume $m=2$ . Since ${\boldsymbol \gamma} \in \mathcal{P}\cap \mathcal{C}$ , for $n=2$ , $\boldsymbol{x}_{\textbf{1}}=\boldsymbol{h}$ , $\boldsymbol{x}_{\textbf{2}}=\textbf{0}$ , $\boldsymbol{a}_{\textbf{1}},\boldsymbol{a}_{\textbf{2}} \in {{\mathbb{R}}}^2$ with $\textbf{1}_{\textbf{2}}^\top \sum_{k=1}^2 \boldsymbol{a}_{\boldsymbol{k}}=0$ , using the symmetry of ${\boldsymbol \gamma}$ and ${\boldsymbol \gamma}(\textbf{0})=\textbf{0}$ we have
Choosing $\boldsymbol{a}_{\textbf{1}}=({-}1,0)^\top$ , $\boldsymbol{a}_{\textbf{2}}= (1-k,k)^\top$ , $k \ge 2$ , and applying the Cauchy–Schwarz inequality due to ${\boldsymbol \gamma} \in \mathcal{C}$ gives
By symmetry, we also have
Assume first that, without loss of generality, $\gamma_{11}(\boldsymbol{h})=0$ . Then $\gamma_{12}(\boldsymbol{h}) = 0$ and $\gamma_{22}(\boldsymbol{h}) =0$ due to inequalities (4) and (5). Suppose now that $\gamma_{11}(\boldsymbol{h}), \gamma_{22}(\boldsymbol{h}) \neq 0$ . Letting $k \rightarrow \infty$ in inequalities (4) and (5) yields $\gamma_{11}(\boldsymbol{h}) = \gamma_{22}(\boldsymbol{h})$ . Inserting this into inequality (5) gives
and consequently the result for $k \rightarrow \infty$ .
Remark 1. Let $\boldsymbol{Z}$ be a random field on ${{\mathbb{R}}}^d$ with cross-variogram ${\boldsymbol {\tilde\gamma}} \in \mathcal{P} \cap \mathcal{C}$ . Then the pseudo and cross-variogram of $\boldsymbol{Z}$ do not necessarily coincide; in fact, the pseudo cross-variogram might not even exist. For instance, let $\boldsymbol{Y}$ be a random field on ${{\mathbb{R}}}^d$ with cross-variogram ${\boldsymbol {\tilde\gamma}}$ , and take $Z_i(\boldsymbol{x})\,:\!=\, Y_i(\boldsymbol{x})+U_i$ , for i.i.d. random variables $U_i$ , $i=1,\ldots,m$ , without existing variance. However, if $\boldsymbol{Z}$ is a random field with existing pseudo cross-variogram of the form (3), then we have $2\gamma_{ij}(\textbf{0})={Var}(Z_i(\boldsymbol{x})-Z_j(\boldsymbol{x}))=0$ for all $\boldsymbol{x} \in {{\mathbb{R}}}^d$ , $i,j=1,\ldots,m$ . Consequently, the difference between $Z_i(\boldsymbol{x})$ and $Z_j(\boldsymbol{x})$ is almost surely constant for all $\boldsymbol{x} \in {{\mathbb{R}}}^d$ , $i,j=1,\ldots,m$ , implying that cross- and pseudo cross-variogram of $\boldsymbol{Z}$ coincide in that case.
Corollary 1 can also be proved by means of a result in [Reference Oesting, Schlather and Friederichs21]. In fact, Theorem 1 enables us to reproduce their result, which was originally derived in a stochastic manner, by a direct proof.
Corollary 2. Let ${\boldsymbol \gamma}\,:\, {{\mathbb{R}}}^d \rightarrow {{\mathbb{R}}}^{m \times m}$ be a pseudo cross-variogram. Then ${\boldsymbol \gamma}$ fulfils
Proof. Without loss of generality, assume $m=2$ . We present a proof for $i=1, j=2$ and $\gamma_{11}(\boldsymbol{h})$ , $\gamma_{12}(\boldsymbol{h}) > 0$ . Then, for $n=2$ , $\boldsymbol{x}_{\textbf{1}}=\boldsymbol{h}$ , $\boldsymbol{x}_{\textbf{2}}=\textbf{0}$ , $\boldsymbol{a}_{\textbf{1}},\boldsymbol{a}_{\textbf{2}} \in {{\mathbb{R}}}^2$ with $\textbf{1}_{\textbf{2}}^\top \sum_{k=1}^2 \boldsymbol{a}_{\boldsymbol{k}}=0$ , we have
Assuming $a_{12}=0$ , $a_{22}>0$ and $a_{11} + a_{22} = -a_{21} >0$ , inequality (6) simplifies to
for $x \,:\!=\, {{a_{11}}/{a_{22}}}$ . Maximization of the function
leads to
Inserting $x^\ast$ into (7) gives
3. A Schoenberg-type characterization
The stochastically motivated proof of Theorem 1 contains an important relation between matrix-valued positive definite kernels and conditionally negative definite functions we have not yet emphasized. Due to its significance, we formulate it in a separate lemma. As readily seen, the assumption on the main diagonal stemming from our consideration of pseudo cross-variograms can be dropped, resulting in the matrix-valued version of Lemma 3.2.1 in [Reference Berg, Christensen and Ressel2].
Lemma 1. Let ${\boldsymbol \gamma}\,:\, {{\mathbb{R}}}^d \rightarrow {{\mathbb{R}}}^{m \times m}$ be a matrix-valued function with $\gamma_{ij}(\boldsymbol{h})=\gamma_{ji}({-}\boldsymbol{h})$ , $i,j=1,\ldots,m$ . Define
with $\boldsymbol{\gamma_{\boldsymbol{k}}}(\boldsymbol{h})=(\gamma_{1k}(\boldsymbol{h}),\ldots,\gamma_{mk}(\boldsymbol{h}))^\top$ , $k \in \{1,\ldots,m\}$ . Then $\boldsymbol{C}_{\boldsymbol{k}}$ is a positive definite matrix-valued kernel for $k \in \{1,\ldots,m\}$ if and only if ${\boldsymbol \gamma}$ is conditionally negative definite. If $\gamma_{kk}(\textbf{0}) \ge 0$ for $k=1,\ldots,m$ , then
is a positive definite matrix-valued kernel for $k \in \{1,\ldots,m\}$ if and only if ${\boldsymbol \gamma}$ is conditionally negative definite.
The kernel construction in Lemma 1 leads to a matrix-valued version of Schoenberg’s theorem [Reference Berg, Christensen and Ressel2].
Theorem 2. A function ${\boldsymbol \gamma}\,:\, {{\mathbb{R}}}^d \rightarrow {{\mathbb{R}}}^{m \times m}$ is conditionally negative definite if and only if $\exp^\ast({-}t{\boldsymbol \gamma})$ , with $\exp^\ast({-}t{\boldsymbol \gamma}(\boldsymbol{h}))_{ij} \,:\!=\, \exp({-}t\gamma_{ij}(\boldsymbol{h}))$ , is positive definite for all $t >0$ .
Remark 2. Theorem 2, in the form presented here, has recently been formulated in [Reference Gesztesy and Pang9] in terms of conditionally positive definite matrix-valued functions with complex entries. Nonetheless, we present an alternative proof of the ‘only if’ part, which relies on the kernel construction in Lemma 1 and follows the lines of the proof of Theorem 3.2.2 in [Reference Berg, Christensen and Ressel2].
Proof of Theorem 2.Assume that ${\boldsymbol \gamma}$ is conditionally negative definite. Then
is a positive definite kernel due to Lemma 1. Since positive definite matrix-valued kernels are closed with regard to sums, Hadamard products, and pointwise limits (see e.g. [Reference Schlather27]), the kernel
is again positive definite for all $t >0$ . The same holds true for the kernel
since
with the product separable structure implying positive definiteness. Again using the stability of positive definite kernels under Hadamard products, the first part of the assertion follows.
Assume now that $\exp^\ast({-}t{\boldsymbol \gamma})$ is a positive definite function for all $t>0$ . Then
and thus
is a conditionally negative definite function. The assertion follows for $t\rightarrow 0$ .
Combining Theorems 1 and 2, and recalling that the classes of matrix-valued positive definite functions and covariance functions for multivariate random fields coincide, we immediately get the following characterization of pseudo cross-variograms.
Corollary 3. A function ${\boldsymbol \gamma}\,:\, {{\mathbb{R}}}^d \rightarrow {{\mathbb{R}}}^{m \times m}$ is a pseudo cross-variogram if and only if $\exp^\ast({-}t{\boldsymbol \gamma})$ is a matrix-valued correlation function for all $t>0$ .
Corollary 3 establishes a direct link between matrix-valued correlation functions and pseudo cross-variograms. Together with Corollary 1, it shows that the cross-variograms for which Theorem 10 in [Reference Ma14] holds are necessarily of the form (3), and it explains the findings in the first part of Remark 2 in [Reference Schlather27].
Remark 3. The correspondence of the proofs of the matrix-valued and real-valued versions of Schoenberg’s theorem (Theorem 2 here and Theorem 3.2.2 in [Reference Berg, Christensen and Ressel2]) is no coincidence. Since [Reference Berg, Christensen and Ressel2] does not impose any assumption on the underlying domain X of the negative definite function there, we could also choose $X= {{\mathbb{R}}}^d \times \{1,\ldots,m\}$ , which translates into Theorem 2. The same holds true for Lemma 1. With this ‘dimension expansion view’, it is no surprise that the pseudo cross-variogram turns out to be the natural multivariate analogue of the variogram from a theoretical standpoint. Due to our interest in pseudo cross-variograms, we chose a stochastically/pseudo cross-variogram driven derivation of Lemma 1 and Theorem 2.
As demonstrated in Theorem 3.2.3 in [Reference Berg, Christensen and Ressel2], Schoenberg’s theorem can be further generalized for conditionally negative definite functions with non-negative components in terms of componentwise Laplace transforms. Here we present the corresponding matrix-valued version explicitly for clarity, and combine it with our previous results concerning pseudo cross-variograms. With regard to Remark 3, and since we have already presented the matrix-valued proof of Theorem 2, we omit the proof here.
Theorem 3. Let $\mu$ be a probability measure on $[0,\infty)$ such that
Let $\mathcal{L}$ denote its Laplace transform, that is,
Then ${\boldsymbol \gamma}\,:\, {{\mathbb{R}}}^d \rightarrow [0,\infty)^{m \times m}$ is conditionally negative definite if and only if $(\mathcal{L}\mu(t\gamma_{ij}))_{i,j=1,\ldots,m}$ is positive definite for all $t >0$ . In particular, ${\boldsymbol \gamma}$ is a pseudo cross-variogram if and only if $(\mathcal{L}\mu(t\gamma_{ij}))_{i,j=1,\ldots,m}$ is an m-variate correlation function for all $t >0$ .
Corollary 4. Let ${\boldsymbol \gamma}\,:\, {{\mathbb{R}}}^d \rightarrow [0,\infty)^{m \times m}$ be a matrix-valued function. If ${\boldsymbol \gamma}$ is a pseudo cross-variogram, then for all $\lambda >0$ , the function $\boldsymbol{C}\,:\, {{\mathbb{R}}}^d \rightarrow {{\mathbb{R}}}^{m \times m}$ with
is a correlation function of an m-variate random field for all $t >0$ . Conversely, if a $\lambda >0$ exists such that $C_{ii}(\textbf{0})=1$ , $i=1,\ldots,m$ , and such that $\boldsymbol{C}$ is positive definite for all $t >0$ , then ${\boldsymbol \gamma}$ is a pseudo cross-variogram.
Proof. Choose
in Theorem 3.
Similarly to Theorem 3, we can translate the ‘univariate’ result that Bernstein functions operate on real-valued conditionally negative definite functions [Reference Berg, Christensen and Ressel2] to the matrix-valued case, which can thus be used to derive novel pseudo cross-variograms from known ones. Again, we omit the proof for the reasons given above.
Proposition 1. Let ${\boldsymbol \gamma}\,:\, {{\mathbb{R}}}^d \rightarrow [0,\infty)^{m\times m}$ be conditionally negative definite. Let $g\,:\, [0, \infty) \rightarrow [0,\infty)$ denote the continuous extension of a Bernstein function. Then $\boldsymbol{g} \circ {\boldsymbol \gamma}$ with $((\boldsymbol{g} \circ {\boldsymbol \gamma})(\boldsymbol{h}))_{ij}\,:\!=\, (g\circ \gamma_{ij})(\boldsymbol{h})$ , $i,j=1,\ldots,m$ , is conditionally negative definite. In particular, if $g(0)=0$ and ${\boldsymbol \gamma}$ is a pseudo cross-variogram, then $\boldsymbol{g} \circ \boldsymbol\gamma$ is again a pseudo cross-variogram.
4. Multivariate versions of Gneiting’s space–time model
Schoenberg’s result is often an integral part of proving the validity of univariate covariance models. Here we use its matrix-valued counterparts derived in the previous section to naturally extend covariance models of Gneiting type to the multivariate case.
Gneiting’s original space–time model is a univariate covariance function on ${{\mathbb{R}}}^d\times {{\mathbb{R}}}$ defined via
where $\psi\,:\, 0, \infty) \rightarrow (0,\infty)$ is the continuous extension of a Bernstein function, and $\varphi\,:\, [0,\infty) \rightarrow [0,\infty)$ is the continuous extension of a bounded completely monotone function [Reference Gneiting10]. For convenience, we simply speak of bounded completely monotone functions henceforth. Model (9) is very popular in practice due to its versatility and ability to model space–time interactions; see [Reference Porcu, Furrer and Nychka24] for a list of several applications. Its special structure has attracted and still attracts interest from a theoretical perspective as well, resulting in several extensions and refinements of the original model (9); see e.g. [Reference Menegatto17], [Reference Menegatto, Oliveira and Porcu18], [Reference Porcu, Bevilacqua and Genton23], [Reference Qadir and Sun25], [Reference Zastavnyi and Porcu32]. Only recently, specific simulation methods have been proposed [Reference Allard, Emery, Lacaux and Lantuéjoul1] for the so-called extended Gneiting class, a special case of [Reference Zastavnyi and Porcu32, Theorem 2.1],
with $\gamma$ denoting a continuous variogram. One of these methods is based on an explicit construction of a random field, where the continuity assumption on $\gamma$ is not needed [Reference Allard, Emery, Lacaux and Lantuéjoul1], and which can be directly transferred to the multivariate case via pseudo cross-variograms.
Theorem 4. Let R be a non-negative random variable with distribution $\mu$ , $\boldsymbol{\Omega} \sim N(\textbf{0},\mathbf{1}_{d \times d})$ with $\mathbf{1}_{d \times d} \in {{\mathbb{R}}}^{d \times d}$ denoting the identity matrix, $U \sim U(0,1)$ , $\Phi\sim U(0,2\pi)$ , and let $\boldsymbol{W}$ be a centred, m-variate Gaussian random field on ${{\mathbb{R}}}^l$ with pseudo cross-variogram ${\boldsymbol \gamma}$ , all independent. Then the m-variate random field $\boldsymbol{Z}$ on ${{\mathbb{R}}}^d \times {{\mathbb{R}}}^l$ defined via
has the extended Gneiting-type covariance function
where $\varphi$ denotes a bounded completely monotone function.
Proof. The proof follows the lines of the proof of Theorem 3 in [Reference Allard, Emery, Lacaux and Lantuéjoul1]. In the multivariate case, the cross-covariance function reads
$(\boldsymbol{x},\boldsymbol{t}), (\boldsymbol{y},\boldsymbol{s}) \in {{\mathbb{R}}}^d \times {{\mathbb{R}}}^l, i,j=1,\ldots,m$ . Due to the assumptions, $W_i(\boldsymbol{t})-W_j(\boldsymbol{s})$ is a Gaussian random variable with mean zero and variance $2\gamma_{ij}(\boldsymbol{t}-\boldsymbol{s})$ . Proceeding further as in [Reference Allard, Emery, Lacaux and Lantuéjoul1] gives the result.
Theorem 4 provides a multivariate extension of the extended Gneiting class, and lays the foundations for a simulation algorithm for an approximately Gaussian random field with the respective cross-covariance function; see [Reference Allard, Emery, Lacaux and Lantuéjoul1]. The existence of a Gaussian random field with a preset pseudo cross-variogram ${\boldsymbol \gamma}$ and the possibility of sampling from it are ensured by Theorem 1 and Lemma 1, respectively.
Due to our results in previous sections, Theorem 4 can easily be generalized further, replacing $d/2$ in the denominator in equation (10) with a general parameter $r \ge d/2$ .
Corollary 5. Let ${\boldsymbol \gamma}\,:\, {{\mathbb{R}}}^l \rightarrow {{\mathbb{R}}}^{m \times m}$ be a pseudo cross-variogram. Then the function $\boldsymbol{G}\,:\, {{\mathbb{R}}}^d \times {{\mathbb{R}}}^l \rightarrow {{\mathbb{R}}}^{m \times m}$ with
is positive definite for $r \ge {{d}/{2}}$ and a bounded completely monotone function $\varphi$ .
Proof. We proved the assertion for $r={{d}/{2}}$ in Theorem 4. Now let $\lambda > 0$ and $r = \lambda + {{d}/{2}}$ . Then the matrix-valued function (11) is the componentwise product of positive definite functions of the form (8) and (10), and consequently positive definite itself.
Even further refinements of Corollary 5 are possible. We can replace $\textbf{1}_{\boldsymbol{m}}\textbf{1}_{\boldsymbol{m}}^\top + {\boldsymbol \gamma}$ in (11) with general conditionally negative definite matrix-valued functions, but for a subclass of completely monotone functions, the so-called generalized Stieltjes functions of order $\lambda$ . This leads to a multivariate version of a result in [Reference Menegatto17]. A bounded generalized Stieltjes function $S\,:\, (0,\infty) \rightarrow [0,\infty)$ of order $\lambda >0$ has a representation
where $a\ge 0$ and the so-called Stieltjes measure $\mu$ is a positive measure on $(0,\infty)$ , such that $\int_{(0,\infty)}v^{-\lambda}\,{{d}} \mu(v) < \infty$ [Reference Menegatto17]. As for completely monotone functions, in the following we do not distinguish between a generalized Stieltjes function and its continuous extension. Several examples of generalized Stieltjes functions can be found in [Reference Berg, Koumandos and Pedersen3] and [Reference Menegatto17].
Theorem 5. Let $S_{ij}, i,j=1,\ldots,m$ , be generalized Stieltjes functions of order $\lambda >0$ . Let the associated Stieltjes measures have densities $\varphi_{ij}$ such that $(\varphi_{ij}(v))_{i,j=1,\ldots,m}$ is a symmetric positive semidefinite matrix for all $v > 0$ . Let
be conditionally negative definite functions. Then the function $\boldsymbol{G}\,:\, {{\mathbb{R}}}^d\times {{\mathbb{R}}}^l \rightarrow {{\mathbb{R}}}^{m \times m}$ with
is an m-variate covariance function for $r \ge \lambda$ .
Proof. We follow the proof in [Reference Menegatto17]. It holds that
The function $ x \mapsto {{1}/{x^\alpha}}$ is completely monotone for $\alpha \ge 0$ and thus the Laplace transform of a measure on $[0,\infty)$ [Reference Schilling, Song and Vondracek26, Theorem 1.4]. Therefore $(1/f_{ij}^r)_{i,j=1,\ldots,m}$ and $(1/f_{ij}^{r-\lambda})_{i,j=1,\ldots,m}$ are positive definite functions due to Theorem 2 as mixtures of positive definite functions. Furthermore, we have
The functions $\bigl({{e}}^{-sg_{ij}(\boldsymbol{h})}\bigr)_{i,j=1,\ldots,m}$ and $\bigl({{e}}^{-svf_{ij}(\boldsymbol{u})}\bigr)_{i,j=1,\ldots,m}$ are again positive definite due to Theorem 2 for all $s,v >0$ , and so is their componentwise product. Since positive definite functions are closed under integration,
is positive definite for all $v > 0$ . Therefore the function
is also positive definite for all $v > 0$ . Combining and applying the above arguments shows our claim.
Theorem 5 provides a very flexible model. In a space–time framework, it allows for different cross-covariance structures in both space and time, and it does not require assumptions such as continuity of the conditionally negative definite functions involved, or isotropy, which distinguishes it from the multivariate Gneiting-type models presented in [Reference Bourotte, Allard and Porcu4] and [Reference Guella12], respectively.
Funding information
C. D. was supported by the German Research Foundation (DFG) via RTG 1953: Statistical Modelling of Complex Systems and Processes – Advanced Nonparametric Approaches.
Competing interests
There were no competing interests to declare which arose during the preparation or publication process of this article.