The Bunch-Parlett Decomposition
The classical
decomposition of a symmetric matrix
is defined by
with a lower triangular matrix
with unit diagonal and a diagonal
matrix
. This can be computed by the following compact algorithm:
|
For a positive definite matrix (
is positive definite if and only if all
the
are strictly positive)
this is a perfectly backward stable algorithm.
It is associated with the Cholesky decomposition: the so called
Cholesky factror is simply
.
As long as none of the
becomes zero this can be performed also for
indefinite matrices, but it is numerically unstable.
One never should use it in this
situation. The alternative is the Bunch-Parlett decomposition, which is
stable also in the indefinite case and which discerns from the above
twofold: the
-part is now block diagonal with
and
blocks, and the introduction of a symmetric permutation is
indispensable: it reads
Here
denotes the permutation matrix. We describe here the version with
complete pivoting.
The total algorithm consists of up to
steps which themselves decompose into two parts.
The first part searches for the pivot: this is the element of largest absolute
value in the current right lower submatrix of dimension
. If this occurs on the diagonal, it becomes a
block
in
and is permuted to the position
in the current matrix.
Then a normal Gaussian elimination step is performed,
with the multipliers stored
in column
and
is increased by one.
If however the
element of largest absolute value appears in position
(with
and
) then the submatrix
becomes a
block in
and the rows
and
and
and
are swapped and equally the columns, of course.
Hence we have now to perform a block elimination with a
block.
This uses the formula
|
with
|
The two columns of the left factor go into
and the new remaining
submatrix to be processed is
,
also known as the Schurcomplement.
In this case
is increased by 2.
Finally, from Sylvesters theorem of inertia, the matrix
has as
many negative eigenvalues as
has
blocks and negative
blocks, since each
block has exactly one positive
and one negative eigenvalue by construction.
This decomposition is used here in a twofold manner: by shifting all
eigenvalues of
above some level
by addition of
a multiple of the identity matrix
we construct a positive
definite regularization of the Hessian,
and a strongly gradient related direction of descent
by
and a direction of negative curvature
from
where
is composed from the eigenvectors corresponding to negative eigenvalues
of
.
We have namely
We normalize
such that
and use
as
a direction of descent if this promises better progress than
.
File translated from
TEX
by
TTM Unregistered,
version 4.03.
On 16 Jun 2016, 18:51.