Complete problem of eigen values
complete eigen value problem
The problem of calculating all the eigen values (as distinct from the partial problem of eigen values) of a square, usually real or complex, matrix. Often, one has not only to calculate the eigen values but also to construct a basis composed of eigen vectors, or root (principal) vectors, for the matrix.
The solution to the complete problem of eigen values for a matrix amounts theoretically to constructing the characteristic polynomial
for this matrix and calculating the, real or complex, roots of it (this makes it impossible to compute the eigen values by a finite computational process). For each eigen value
, one can determine the corresponding eigen vectors from a homogeneous system of linear equations:
(
is the identity matrix). In calculations over the field of complex numbers, a sufficient condition for the existence of a basis of eigen vectors is simplicity of the spectrum, while a necessary and sufficient condition is that the algebraic multiplicity of each eigen value
(i.e. its multiplicity as a root of the characteristic polynomial
) coincides with its geometric multiplicity (i.e. the defect of the matrix
). If it is necessary to calculate the root (principal) vectors of multiplicity (degree) exceeding one, one has to consider homogeneous systems of the form
![]() |
According to this scheme numerical methods of solving the complete problem of eigen values were constructed up to the end of the 1940-s. In the 1930-s, highly effective algorithms (as regards the number of arithmetic operations) were developed for calculating the characteristic polynomial of a matrix in terms of its coefficients. For example, in Danilevskii's method, the computation of the characteristic polynomial for a matrix of order involves about
multiplications [1], [2].
Methods in this group have received the name of direct, or exact, ones for the reason that if carried out with exact arithmetic they give the exact values of the coefficients in the characteristic polynomial. The behaviour in real calculations with rounding-off errors could not be tested for problems of any substantial order until digital computers became available. Such tests were carried out in the 1950-s, and as a result, direct methods were completely displaced from numerical practice. There is a catastrophic instability in calculating the eigen values by these methods, and there are two major reasons for this. First, the coefficients in the characteristic polynomial in the majority of exact methods are determined directly or indirectly as components of the solution of a system of linear equations, the matrix of which is built up by columns from the vectors , where
is the initial vector in the method. Such a matrix is usually very poorly conditioned, as is evident in particular from the fact that the lengths of the columns are usually very different, and this is the more so the larger
. Therefore, the coefficients in the characteristic polynomial in general are calculated with very large errors. Secondly, the calculation of the roots of a polynomial is frequently numerically unstable. In that respect, the following example is noteworthy [3]: If in the polynomial
![]() |
![]() |
one alters the coefficient of from
to
, the perturbed polynomial
acquires five pairs of complex-conjugate roots; for one of these pairs, the imaginary parts are
.
This sensitivity of the eigen values to perturbations in the coefficients of the characteristic polynomial is usually not accompanied by a comparable sensitivity with respect to the elements of the matrix itself. For example, if this polynomial is the characteristic polynomial of a symmetric matrix
, then perturbations of the order of
in any element in the matrix produce changes at most of the same order in the eigen values.
Modern numerical methods for solving the complete problem of eigen values yield their results without calculating the characteristic polynomial (see Iteration methods). The number of multiplications involved in the best methods is , where
is the order of the matrix and
is a constant independent of
and having the meaning of the average number of iterations required to calculate one eigen value. The values of
in the
algorithm are usually between 1.5 and 2.
Approximate eigen values (vectors) calculated for an ()-matrix
by a method
based on orthogonal transformations can be interpreted as the exact eigen values (vectors) of a perturbed matrix or matrices
(the
algorithm for matrices of general form, or Jacobi's method or methods based on dividing the spectrum for symmetric and Hermitian matrices are used). Here
is the equivalent perturbation matrix for the method
that satisfies bounds of the form
![]() | (1) |
where is the relative accuracy of the machine arithmetic,
is the Euclidean matrix norm and
is a function of the form
. The number
has been described above, while the exact values of the constant
and the exponent
depend on details of the computation such as the method of rounding-off, the use of scalar-product accumulation, etc. The usual value of
is 2.
The a priori bound (1) allows one to estimate the accuracy in calculating the eigen values (vectors) by the method . This accuracy depends on the conditioning of the individual eigen values (eigen subspaces).
Let be a simple eigen value of a matrix
, let
be the corresponding normalized eigen vector, and let
be the normalized eigen vector of the transposed matrix
for the same eigen value. When the matrix
is perturbed by a matrix
, the perturbation in the eigen value
is expressed by the following quantity, accurate up to the second order:
![]() | (2) |
and is estimated as
![]() | (3) |
( is the spectral norm). Then the sensitivity of
to perturbations in the matrix
is characterized by the number
, which is called the (individual) condition number of that eigen value. When both vectors
and
are real, the number
has a simple geometrical meaning: It is the cosine of the angle between
and
, which explains the other name of
, the distortion coefficient corresponding to
.
If the matrix is diagonalizable (i.e. has a basis of eigen vectors), then the conditioning of its eigen values
can be characterized completely. Let
be the matrix composed by the columns of eigen vectors at
and having the least condition number among all such matrices. The following theorem applies [4]: All the eigen values of the perturbed matrix
are enclosed in the region in the complex plane that is the union of the circles
![]() | (4) |
If this region splits up into connected components, then each of them contains as many eigen values of the perturbed matrix as the number of circles constituting that component (as the norm in (4), one can take the spectral norm, while can be taken as the spectral condition number).
The number is called the condition number of the matrix
in relation to the complete problem of eigen values. Expressed in the spectral norm, it is related to the distortion coefficients as follows:
![]() |
The perturbation in an eigen vector related to the simple eigen value
depends on the perturbation of the matrix
in a more complicated way. It is determined, generally speaking, not only by the distortion coefficient corresponding to
but also by those for the other eigen values. The sensitivity of
also increases if there are eigen values close to
. In the limiting case where
becomes a multiple eigen value, it becomes meaningless to consider the sensitivity of an individual eigen vector, and it is necessary to speak of the sensitivity of an eigen (or invariant) subspace.
Qualitatively, the estimates (3) and (4) mean that the perturbations in the eigen values of a diagonalizable matrix are proportional to the size of the perturbation
, while the proportionality factors are represented by the condition numbers, individual or global. If the Jordan form of
is non-diagonal and the eigen value
corresponds to an elementary divisor
, then the perturbation of
in the general case is proportional not to
, but to
.
The most important particular case of the complete problem of eigen values is to calculate all eigen values (eigen vectors) of a real symmetric or complex Hermitian matrix . The distortion coefficients for such a matrix are equal to one, and the approximate estimate (3) becomes an exact inequality:
![]() |
The matrix in (4) can be taken as orthogonal or unitary, and therefore the global condition number in the spectral norm is equal to one. No matter what the multiplicity of the points in the spectrum of
, there exists an ordering of the eigen values
of the matrix
such that for all
:
![]() |
Even sharper bounds are possible if not only the matrix itself is symmetric (Hermitian) but the same applies to the perturbation
[5].
In addition, there are some a posteriori estimates for the accuracy of the calculated eigen values and eigen vectors. They are most effective for symmetric and Hermitian matrices .
Let be an approximate eigen vector and let
be the exact eigen vector at a simple eigen value
of the matrix
. Both vectors are assumed to be normalized. The best estimate of
that can be obtained by means of the vector
is given by the value of the Rayleigh functional
![]() |
corresponding to , i.e. the number
(this applies for any matrix
, not only Hermitian ones). The vector
is called the discrepancy vector (or residual).
Let
![]() |
A bound for can readily be derived from the calculated eigen values
. The following apply:
![]() |
If is a multiple eigen value or if there is a group of eigen values near to
, it is necessary to estimate the over-all perturbation for the whole group and the perturbation for the corresponding invariant subspace [5].
Along with this ordinary complete eigen value problem, it is often also necessary to solve the so-called generalized eigen value problem:
![]() | (5) |
In the most important case, the matrices and
are symmetric (Hermitian) and one of them is positive definite. The theory and methods for numerically solving (5) are parallel to those for an ordinary symmetric (Hermitian) eigen value problem.
If neither of the matrices and
is definite or if at least one of them is not symmetric, one uses the
algorithm [6], which is a generalization of the
algorithm. Here new effects can occur that are not observed in ordinary complete eigen value problems: infinite eigen values and continuous spectra.
Even more complicated features occur in non-linear eigen value problems:
![]() |
Such problems are usually solved by reducing them to linear ones of a higher order [7].
References
[1] | A.M. Danilevskii, Mat. Sb. , 2 : 1 (1937) pp. 169–171 |
[2] | D.K. Faddeev, V.N. Faddeeva, "Computational methods of linear algebra" , Freeman (1963) (Translated from Russian) |
[3] | J.H. Wilkinson, "Rounding errors in algebraic processes" , Prentice-Hall (1963) |
[4] | J.H. Wilkinson, "The algebraic eigenvalue problem" , Oxford Univ. Press (1969) |
[5] | B.N. Parlett, "The symmetric eigenvalue problem" , Prentice-Hall (1980) |
[6] | C.B. Moler, G.W. Stewart, "An algorithm for generalized matrix eigenvalue problems" Siam J. Numer. Anal. , 10 (1973) pp. 241–256 |
[7] | V.N. Kublanovskaya, "Numerical methods of linear algebra" , Novosibirsk (1980) pp. 37–53 (In Russian) |
Comments
In recent years much research has been carried out on the eigen value problem for sparse matrices (cf. Sparse matrix). Because of storage problems, the challenges are even greater here. For the symmetric (Hermitian) case the Lanczos algorithm has proved very successful.
Cf. Iteration methods for a description of the QR algorithm and other iterative algorithms for calculating eigen values.
References
[a1] | G.H. Golub, C.F. van Loan, "Matrix computations" , North Oxford Acad. (1983) |
Complete problem of eigen values. Encyclopedia of Mathematics. URL: http://encyclopediaofmath.org/index.php?title=Complete_problem_of_eigen_values&oldid=16275