Table of Contents
International Journal of Analysis
Volume 2016, Article ID 5967218, 5 pages
Research Article

On the Marginal Distribution of the Diagonal Blocks in a Blocked Wishart Random Matrix

1Nord University, 8700 Nesna, Norway
2Instituto de Alta Investigación, Universidad de Tarapacá, Antofagasta N. 1520, Arica, Chile
3Universidad Católica del Norte, Región de Antofagasta, Chile
4Departamento de Matemáticas, Universidad Católica del Norte, Casilla 1280, Antofagasta, Chile

Received 29 June 2016; Accepted 20 September 2016

Academic Editor: Shamsul Qamar

Copyright © 2016 Kjetil B. Halvorsen et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.


Let be a blocked Wishart random matrix with diagonal blocks of orders and . The goal of the paper is to find the exact marginal distribution of the two diagonal blocks of . We find an expression for this marginal density involving the matrix-variate generalized hypergeometric function. We became interested in this problem because of an application in spatial interpolation of random fields of positive definite matrices, where this result will be used for parameter estimation, using composite likelihood methods.

1. Introduction

The goal of this paper is to find an exact and useful form for the marginal distribution of the diagonal blocks of a blocked Wishart random matrix. This problem arises in an applied problem, to estimate the parameters of a Wishart random field, which will be reported elsewhere.

Let be a Wishart random matrix, where the diagonal blocks are of orders and , respectively. In our intended application , will be small integers (and , but we choose to treat the more general case). Write .

Denote the number of freedom parameters by and the scale parameter, which is a matrix blocked in the same way as , by . We are mostly interested in the special case where the absolute value of is less than one, but the general case is not more difficult.

All matrices are real. Notation: we use for the trace of the square matrix and . We write for the convex cone of real positive definite matrices, and we write for the orthogonal group, that is, the set of orthogonal matrices. The Stiefel manifold, that is, the set of column orthogonal matrices is written as . We indicate the transpose of a matrix by superscript .

In the convex cone of positive definite matrices, we use the cone order, defined by meaning that is positive definite, written as . Integrals over cones are written as meaning the integral is taken over the cone . The multivariate gamma function is denoted by for ; see Muirhead [1] for proofs and properties.

In Section 2 we give some background information, especially about the Jacobians which we need to evaluate the integrals. In Section 3 we state our results and give proofs. In Section 4 we give some comments on the result.

2. Background

The single most important reference for background material for this paper is Muirhead [1]. Some results therefrom will not be cited directly.

When doing change of variables in a multiple integral we need to know the Jacobian. Here we will list the ones we need; most can be found in Muirhead [1] or in Mathai [2]. We are following the notation of Muirhead [1]. First there is a very brief summary.

For any matrix , let denote the matrix of differentials . For an arbitrary matrix , the symbol denotes the exterior product of the elements of :If is a symmetric matrix, the symbol will denote the exterior product of the distinct elements of :with similar definitions for other kinds of structured matrices.

The following invariant form in the orthogonal group represents the Haar measure, . Here represents an orthogonal matrix. This form normalized to have total mass unity is represented by . We also need to integrate over a Stiefel manifold; then represents a similarly defined invariant form; see Muirhead [1].

Some needed Jacobians are not in Muirhead [1], so we cite those Jacobians here, from Díaz-García et al. [3, 4].

Lemma 1 (Jacobian of the symmetric square root of a positive definite matrix). Let and be in such that and let be a diagonal matrix with the eigenvalues of on the diagonal. Then,

This result can also be found in Mathai [2].

We need the generalized polar decomposition of a rectangular matrix. Let be rectangular matrix with . Then we always have where is positive semidefinite and positive definite if has full rank, and is column orthogonal matrix. In that last case, is unique; see Higham [5].

Lemma 2 (Generalized polar decomposition). Let be matrix with and of rank , with distinct singular values. Write , with and . Then has distinct eigenvalues. Also let be the diagonal matrix with the eigenvalues of on the diagonal. Then

Note that since those results are used for integration, the assumption of distinct singular values is unimportant, since the subset where the singular values are equal has measure zero.

3. Results

Let us state our main result.

Theorem 3 (The marginal distribution of the diagonal blocks of a blocked Wishart random matrix with blocks of unequal sizes). Let be a blocked Wishart random matrix, where the diagonal blocks are of sizes and , respectively. The Wishart distribution of has degrees of freedom and positive definite scale matrix blocked in the same way as . The marginal distributions of the two diagonal blocks and have density function given by where , , and . . is the generalized matrix-variate hypergeometric function, as defined in Muirhead [1].

Note that the definition of the matrix-variate hypergeometric function is by a series expansion, which is convergent in all cases we need; see Muirhead [1]. The rest of this section consists of a proof of this theorem.

Introduce the following notation: the Schur complements of is and . Then define .

In the following we will be using some standard results on blocked matrices without quoting them.

The Wishart density function of written as a function of the blocks iswhere and . In the following we will work with the density concentrating on the factors depending on . To prove the theorem we need to integrate out the variable . The other variables, which are constant under the integration, will be concentrated in one constant factor. So we repeat formula (6) written as a differential form with the constants left outwhere . Now, to find the marginal distribution of the diagonal blocks, we need to integrate over the off-diagonal block . Under this integration the value of the diagonal blocks and will remain fixed, and the region of integration will be a subset of consisting of the matrices such that the block matrix is positive definite. This seems like a complicated set, but we can give a simple description of it using the polar decomposition of a matrix. Note that this is one of the key observations for the proof, and this author has not seen any use of this observation earlier.

Now we need to assume that . For the opposite inequality a parallel development can be given, using the other factorization . From, for instance, Theorem  1.12 in Zhang [6] it follows that the region of integration is the setIntroduce where we use the usual symmetric square root. Then in terms of the new variable the region of integration becomesand with the generalized polar decomposition in the form with , so the region of integration can be written aswhich is a Cartesian product of a cone interval with a Stiefel manifold.

The Jacobian of the transformation from to is . The Jacobian of the polar decomposition is , where is a diagonal matrix with the eigenvalues of on the diagonal; see Lemma 2. A last transformation will be useful. Define . The Jacobian of this transformation is ; is as above. See Lemma 1.

Applying this transformation the integral of (7) can be written aswhere the constant

We are ready to perform the integration over the Stiefel manifold. For this purpose we need a generalization of Theorem  7.4.1 from Muirhead [1], which we cite here.

Let be real matrix with and orthogonal matrix, where is . ThenBut we have an integral over the Stiefel manifold, not the orthogonal group, so we need now to generalize the result (13) to an integral over the Stiefel manifold. What we need is the following. Let be the manifold of column orthogonal matrices with , and let be a function defined on the Stiefel manifold. We can extend this function to a function defined on in the following way. Let be orthogonal matrix, and write it in block form as such that . How can we characterize the set of which is complementing to form an orthogonal matrix? First, let be a fixed but arbitrary matrix complementing . Then clearly any other column orthogonal matrix with the same column space also works. The common column space is the orthogonal complement of the column space of . The set of such matrices can be described as . For this set we write . As a set we can identify this with . Specifically, we can identify with the very special column orthogonal matrix , where which clearly forms a proper submanifold of the Stiefel manifold . The function can now be extended to the orthogonal group by defining and for the integral we find thatReturning to our integral, the integral over the Stiefel manifold occurring in (11) can now be written aswhere consists of the first columns of where we did use (13). Here is the volume of the orthogonal group; see Muirhead [1]. The differential form denotes Haar measure normalized to total mass unity.

Now write ; then we can write (11) asand to evaluate this integral we need Theorem  7.2.10 from Muirhead [1]; we do not state it here.

Using this we find a result we need for the integral of a hypergeometric function, by using the series expansion definition of the hypergeometric function and integrating term by term.

Theorem 4. If is a symmetric matrix one has thatso both degrees of the hypergeometric function are raised by one.

The proof is a simple calculation that we leave out.

Now using (18) to calculate (17) we get, finally, the result but note that one pair of upper and lower arguments to the hypergeometric function are equal with those arguments canceled.

With a little algebra we complete the proof of our main theorem.

4. Some Final Comments

To help interpret our main result, we calculated the conditional distribution of the matrix given the matrix . We will not give the full details of the calculation here but only give the result. The density of given that has the density given bywhere we have given the conditional density only for the special case . For this case we have, with the notation from the main theorem, , , and . We have defined , which can be seen as a noncentrality parameter. The density above is equal to the noncentral Wishart distribution given in Theorem  10.3.2 in Muirhead [1]. We see that the conditional distribution is a kind of noncentral Wishart distribution, where the noncentrality parameter depends on the conditioning matrix . In this way, the effect of the conditioning is to change the distribution of , which in the marginal case is central Wishart, to a noncentral Wishart distribution, with noncentrality parameter depending on the conditioning matrix.

As said in Introduction, this result will be used for modelling of a spatial random field of tensors, where we will estimate the parameters using composite likelihood. This application will be reported elsewhere. For that application we will need to calculate values of matrix-variate hypergeometric functions numerically. A paper giving an efficient method for summing the defining series is Koev and Edelman [7], with associated Matlab implementation. Butler and Wood [8] give a Laplace approximation for the case we need, the function.

Competing Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.


Kjetil B. Halvorsen was partially supported by Proyecto Mecesup UCN0711 and Convenio de Cooperación Minera Escondida, Universidad Católica del Norte. Victor Ayala partially was supported by Proyecto Fondecyt no. 1100375 and no. 1150292. Eduardo Fierro was partially supported by Proyecto Mecesup UCN0711.


  1. R. J. Muirhead, Aspects of Multivariate Statistical Theory, John Wiley & Sons, 1982. View at MathSciNet
  2. A. M. Mathai, Jacobians of Matrix Transformations and Functions of Matrix Arguments, World Scientific, 1997. View at Publisher · View at Google Scholar · View at MathSciNet
  3. J. A. Díaz-García and G. González-Farías, “Singular random matrix decompositions: Jacobians,” Journal of Multivariate Analysis, vol. 93, no. 2, pp. 296–312, 2005. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus
  4. J. A. Díaz-García, R. Gutierrez Jaimez, and K. V. Mardia, “Wishart and pseudo-Wishart distributions and some applications to shape theory,” Journal of Multivariate Analysis, vol. 63, no. 1, pp. 73–87, 1997. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus
  5. N. J. Higham, Functions of Matrices, Theory and Computation, SIAM, 2008. View at Publisher · View at Google Scholar · View at MathSciNet
  6. F. Zhang, Ed., The Schur Complement and Its Applications, vol. 4, Springer, New York, NY, USA, 2005. View at Publisher · View at Google Scholar · View at MathSciNet
  7. P. Koev and A. Edelman, “The efficient evaluation of the hypergeometric function of a matrix argument,” Mathematics of Computation, vol. 75, no. 254, pp. 833–846, 2006. View at Publisher · View at Google Scholar
  8. R. W. Butler and A. T. Wood, “Laplace approximation for Bessel functions of matrix argument,” Journal of Computational and Applied Mathematics, vol. 155, no. 2, pp. 359–382, 2003. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet · View at Scopus