Sensitivity of spectral preconditoners
Description of the preconditioners
Convergence of CG v.s. eigenvalues (I)
Role played by large and small eigenvalues in the convergence of CG not necessarily symmetric in practice.
Symmetry broken for many reasons : role of clusters, role of the right-hand side. Numerous experimental evidence
An analysis of the convergence in terms of eigenmodes reveals classes of problems where smallest eigenmodes are difficult to capture For instance, regularized least-squares, satisfying the discrete Picard assumption (Fourier coefficients of the RHS decay faster to 0 than singular values). In this case CG even used as a regularization tool ! [P.C. Hansen, SIAM 1998]
“Removing” the effect of small eigenvalues is often targeted
Mitigate the slowdown effect of the smallest eigenvalues (I)
Eigen-information
(small, approximated)
Subspace deflation technique
Example,
,
. Minimize the A-norm of the error on span(
) and solve for error equation iteratively [Frank, Vuik, 2001]
Or minimize the residual
on
to get (
) and minimize the residual
on a subspace orthogonal to
[Parks, de Sturler, Mackey, Johnson and Maiti, 2004]
Note : comparable approaches used in a restarted framework to improve the convergence for small restarts [Bjorck, Grimme, van Dooren, 1994], [Morgan,1995,2000,2002], ...
Questions related to the use of approximate eigen-information
Question 1 : Study the role of the inaccuracy in the spectral elements on the ``quality'' of the spectral preconditionners
Question 2 : Is it possible to find criteria to screen the spectral information(keep only the relevant ones in the preconditioner)
Question 3 : The preconditioners involve quantities such as
. For large systems computing
may be very expensive. What happens if we replace
by
( rayl. } approx.) or
( eig. approx.)
Sensitivity analysis of spectral preconditioners
Sensitivity of the eigenvalues of the preconditioned matrix
Nonlinear perturbation of an eigenvalue problem
{Use first order perturbation theory: condition number approach
Consider the eigenelements as the eigenelements of a perturbed matrix
denoted by
and
. This pairs are assumed to be simple so that first order expansions of
in
is available (recent overview in [Stewart, 2000]).
Compute the first order expansion in
of the eigenvalues
of the preconditioned matrix
.
is an upper-bound of
the backward error
The preconditioners considered - Shift to one
The preconditioners considered - Shift to one (I)
.
In particular, this does not assume any scaling on the eigenvectors
For
,
is replaced by
For
,
is replaced by
For
,
is a semi-simple eigenvalue of
, with multiplicity
. (Perturbation theory for semi-simple eigenvalues needed {\color{violet} [Sun, 1990]}). Furthermore
For
the traditional expression
is recovered
The preconditioners considered - Shift to one (II)
Using the orthogonality property of the eigenvectors, one may use
For
,
is a semi-simple eigenvalues of
, with multiplicity
. (Perturbation theory for semi-simple eigenvalues again needed)
The preconditioners considered - Shift to one (III)
Coarse space preconditioners
and
defined as previously}
For
, the eigenvalues
of
are simple in general,
Theorem I : First order expansion
Fundamental :
The eigenvalues of the preconditioned matrices
admit the first order expansions
where
, and the
depend on the preconditioner. We can therefore define a condition number
for the eigenvalue
in the direction of
for
by [Rice, 66]
In addition,
[Horn, Johnson, 1995]
Matrices X1 and X2 for the various preconditioners
Define
The entries of
are large if some
are small
The matrices
and
may have large entries if some
are clustered
Prec |
|
| some cases of ill-conditioning |
| 0
| 0 0
| none cluster, small cluster, small |
|
| 0
| cluster, small cluster, small |
| 0 0 0 | 0 0
| none none small |
The terms “cluster” and “small” refer respectively to the presence of cluster or of small isolated eigenvalues.
Condition number and Backward Error
Backward error problem :
Let
be approximate eigenpairs of
.
The backward error associated is
Using the same rule of thumb as for linear systems,
For the Frobenius norm,
.This quantity is an upper-bound for the backward error in the spectral norm
This can be used to get an upper-bound on the forward-error in order to screen the eigen-information
Numerical illustrations
Difference between the computed and first order estimate
The eigenvalues of
are
Precond |
|
|
|
|
| 3.8e-07 1.5e-06 2.3e-05 | 7.1e-08 1.5e-07 3.4e-07 | 2.8e-08 1.2e-06 3.5e-07 | 3.8e-08 3.8e-08 3.9e-08 |
| 1.4e-06 2.3e-05 | 8.1e-08 6.7e-07 | 9.6e-07 5.1e-07 | 3.8e-08 3.9e-08 |
| 5.7e-07 2.0e-06 2.2e-05 | 2.5e-07 3.0e-07 2.2e-07 | 3.9e-05 3.9e-05 1.2e-05 | 3.9e-05 4.0e-05 1.3e-05 |
The first order approximation accurately approximates the computed eigenvalues
Departure of the first order and approximate eigenvalues from 1
Numerical examples where the 3 smallest eigenvalues are targeted
Distance to one
The exact radius
The computed radius
The estimated radius
Some clues to analyze the results
: validity of the first order approximation
: asymptotic sensitivity
Case I: a small isolated eigenvalue
We take
,
, the three smallest eigenvalues of
are
the
others are nicely located around
.
Precond |
|
|
|
| 3.7e-07 2.0e-03 1.8e-01 | 0.0e+00 2.0e-03 2.1e-01 | 0.0e+00 0.0e+00 0.0e+00 |
| 2.0e-03 1.8e-01 | 2.0e-03 2.1e-01 | 0.0e+00 0.0e+00 |
| 1.0e-01 1.0e-01 1.7e-01 | 1.0e-01 1.0e-01 2.1e-01 | 1.0e-01 1.0e-01 1.0e-01 |
Case II: a cluster of 3 eigenvalues around 1e-1
We take
,
, the three smallest eigenvalues of $A$ are such that
the 7 others are nicely located around
.
Precond |
|
|
|
| 4.2e-10 2.5e-02 2.5e-02 | 0.0e+00 2.7e-02 2.7e-02 | 0.0e+00 0.0e+00 0.0e+00 |
| 2.3e-02 2.3e-02 | 2.4e-02 2.4e-02 | 0.0e+00 0.0e+00 |
| 1.0e-01 1.0e-01 1.0e-01 | 1.0e-01 1.0e-01 1.0e-01 | 1.0e-01 1.0e-01 1.0e-01 |
A more complicated example
Not really covered by our theory
{Matrix
(Harwell-Boeing)
Symmetrically preconditioned by Incomplete Cholesky factorization (threshold
)
The right-hand side
Stopping criterion : unpreconditioned residual norm reduced by
(
)
We take
and
and report the number of PCG iterations to reach convergence when the number of eigenvectors
is varied
number of PCG iterations
| ||||||||||
|
|
|
|
|
|
|
|
|
|
|
All Prec | 161 | 147 | 132 | 118 | 106 | 95 | 88 | 81 | 81 | 79 |
| ||||||||||
|
|
|
|
|
|
|
|
|
|
|
| 168 168 168 | 154 155 156 | 142 152 154 | 129 142 141 | 122 144 143 | 114 139 140 | 107 133 134 | 104 131 194 | 101 126 183 | 98 121 183 |
| 168 168 | 155 156 | 151 152 | 139 140 | 138 139 | 133 133 | 125 125 | 124 184 | 121 173 | 117 171 |
| 168 168 168 | 154 155 156 | 142 152 152 | 129 139 140 | 121 138 139 | 114 133 133 | 107 125 125 | 104 124 184 | 101 121 174 | 98 117 171 |
The ten smallest eigenvalues of the IC preconditioned matrix
5.67e-04 |
2.91e-03 |
3.58e-03 |
4.71e-03 |
6.19e-03 |
8.35e-03 |
1.85e-02 |
2.05e-02 |
3.04e-02 |
3.05e-02 |
Concluding remarks
The sensitivity of a set of spectral preconditioners studied
Approximations in the preconditioners may results in poor performance, as observed numerically
Comparison of spectral preconditioners has to take into account their sensitivity to inexact eigen-infomation
Implemented in the CERFACS ocean data assimilation system
Study the unsymmetric case ...