Schur complement


In linear algebra and the theory of matrices, the Schur complement of a block matrix is defined as follows.
Suppose A, B, C, D are respectively p × p, p × q, q × p, and q × q matrices, and D is invertible. Let
so that M is a × matrix.
Then the Schur complement of the block D of the matrix M is the p × p matrix defined by
and, if A is invertible, the Schur complement of the block A of the matrix M is the q × q matrix defined by
In the case that A or D is singular, substituting a generalized inverse for the inverses on M/A and M/D yields the generalized Schur complement.
The Schur complement is named after Issai Schur who used it to prove Schur's lemma, although it had been used previously. Emilie Virginia Haynsworth was the first to call it the Schur complement. The Schur complement is a key tool in the fields of numerical analysis, statistics and matrix analysis.

Background

The Schur complement arises as the result of performing a block Gaussian elimination by multiplying the matrix M from the right with a block lower triangular matrix
Here Ip denotes a p×p identity matrix. After multiplication with the matrix L the Schur complement appears in the upper p×p block. The product matrix is
This is analogous to an LDU decomposition. That is, we have shown that
and inverse of M thus may be expressed involving D−1 and the inverse of Schur's complement only as
Cf. matrix inversion lemma which illustrates relationships between the above and the equivalent derivation with the roles of A and D interchanged.

Properties

The Schur complement arises naturally in solving a system of linear equations such as
where x, a are p-dimensional column vectors, y, b are q-dimensional column vectors, and A, B, C, D are as above. Multiplying the bottom equation by and then subtracting from the top equation one obtains
Thus if one can invert D as well as the Schur complement of D, one can solve for x, and
then by using the equation one can solve for y. This reduces the problem of inverting a matrix to that of inverting a p × p matrix and a q × q matrix. In practice, one needs D to be well-conditioned in order for this algorithm to be numerically accurate.
In electrical engineering this is often referred to as node elimination or Kron reduction.

Applications to probability theory and statistics

Suppose the random column vectors X, Y live in Rn and Rm respectively, and the vector in Rn + m has a multivariate normal distribution whose covariance is the symmetric positive-definite matrix
where is the covariance matrix of X, is the covariance matrix of Y and is the covariance matrix between X and Y.
Then the conditional covariance of X given Y is the Schur complement of C in :
If we take the matrix above to be, not a covariance of a random vector, but a sample covariance, then it may have a Wishart distribution. In that case, the Schur complement of C in also has a Wishart distribution.

Schur complement condition for positive definiteness and positive semi-definiteness

Let X be a symmetric matrix given by
Let X/A be the Schur complement of A in X; i.e.,
and X/C be the Schur complement of C in X; i.e.,
Then
The first and third statements can be derived by considering the minimizer of the quantity
as a function of v.
Furthermore, since
and similarly for positive semi-definite matrices, the second statement is immediate from the first statement.
There is also a sufficient and necessary condition for the positive semi-definiteness of X in terms of a generalized Schur complement. Precisely,
where denotes the generalized inverse of.