Difference between revisions of "Iteration algorithm"
(Importing text file) |
m (fix tex) |
||
(One intermediate revision by one other user not shown) | |||
Line 1: | Line 1: | ||
− | + | <!-- | |
+ | i0530101.png | ||
+ | $#A+1 = 38 n = 0 | ||
+ | $#C+1 = 38 : ~/encyclopedia/old_files/data/I053/I.0503010 Iteration algorithm | ||
+ | Automatically converted into TeX, above some diagnostics. | ||
+ | Please remove this comment and the {{TEX|auto}} line below, | ||
+ | if TeX found to be correct. | ||
+ | --> | ||
− | + | {{TEX|auto}} | |
+ | {{TEX|done}} | ||
− | The operation (1) is called an iteration, while the sequence | + | A recursive algorithm realizing a sequence of point-to-set mappings $ A _ {k} : V \rightarrow V $, |
+ | where $ V $ | ||
+ | is a topological space, that can be used to compute for an initial point $ u ^ {0} \in V $ | ||
+ | a sequence of points $ u ^ {k} \in V $ | ||
+ | by the formulas | ||
+ | |||
+ | $$ \tag{1 } | ||
+ | u ^ {k+1} = A _ {k} u ^ {k} ,\ \ | ||
+ | k = 0 , 1 ,\dots . | ||
+ | $$ | ||
+ | |||
+ | The operation (1) is called an iteration, while the sequence $ \{ u ^ {k} \} $ | ||
+ | is called an iterative sequence. | ||
Iteration methods (or methods of iterative approximation) are used both for finding a solution to an operator equation | Iteration methods (or methods of iterative approximation) are used both for finding a solution to an operator equation | ||
− | + | $$ \tag{2 } | |
+ | A u = f , | ||
+ | $$ | ||
− | a minimum of a functional, eigen values and eigen vectors of an equation | + | a minimum of a functional, eigen values and eigen vectors of an equation $ A u = \lambda u $, |
+ | etc., as well as for proving the existence of solutions to these problems. An iteration method (1) is called convergent for the initial approximation $ u ^ {0} $ | ||
+ | to a solution $ u $ | ||
+ | of a problem considered if $ u ^ {k} \rightarrow u $ | ||
+ | as $ k \rightarrow \infty $. | ||
− | Operators | + | Operators $ A _ {k} $ |
+ | for solving (2), given on a metric linear space $ V $, | ||
+ | are usually constructed by the formulas | ||
− | + | $$ \tag{3 } | |
+ | A _ {k} u ^ {k} = u ^ {k} H _ {k} ( A u ^ {k} - f ) , | ||
+ | $$ | ||
− | where | + | where $ \{ {H _ {k} } : {V \rightarrow V } \} $ |
+ | is some sequence of operators determined by the type of the iteration method. The [[Contracting-mapping principle|contracting-mapping principle]] and its generalizations, or variational minimization methods for a functional related to the problem, lie at the basis of constructing iteration methods of type (1), (3). Various methods for constructing $ A _ {k} $ | ||
+ | are used, e.g. variants of the [[Newton method|Newton method]] or method of descent (cf. [[Descent, method of|Descent, method of]]). One tries to choose the $ H _ {k} $ | ||
+ | so that a fast convergence $ u ^ {k} \rightarrow u $ | ||
+ | is ensured under the conditions that the numerical realization of the operations $ A _ {k} u ^ {k} $, | ||
+ | for a given amount of computer memory, is sufficiently simple, has as low complexity as possible and is numerically stable. Iteration methods for solving linear problems have been well-developed and were well-studied. The iteration methods are divided here into linear and non-linear ones. The [[Gauss method|Gauss method]], the [[Seidel method|Seidel method]], the successive overrelaxation method (cf. [[Relaxation method|Relaxation method]]), and iteration methods with Chebyshev parameters belong to the linear methods; variational methods belong to the non-linear methods (e.g. the methods of steepest descent and conjugate gradients, the [[Minimal discrepancy method|minimal discrepancy method]], etc., cf. [[Steepest descent, method of|Steepest descent, method of]]; [[Conjugate gradients, method of|Conjugate gradients, method of]]). One of the efficient iteration methods is the method using Chebyshev parameters, where $ A $ | ||
+ | is a self-adjoint operator with spectrum on $ [ m , M ] $, | ||
+ | $ M > m > 0 $. | ||
+ | This method provides an optimum (for given information on the boundaries of the spectrum) estimate of the convergence at a pre-assigned $ N $- | ||
+ | th step. The method can be written in the form | ||
− | + | $$ | |
+ | u ^ {k+1} = u - \alpha _ {k+1} ( A u ^ {k} - f ) ,\ \ | ||
+ | k = 0 \dots N - 1 , | ||
+ | $$ | ||
where | where | ||
− | + | $$ | |
+ | \alpha _ {k+1} = 2 | ||
+ | \left ( | ||
+ | M + m - ( M - m ) \ | ||
+ | \cos | ||
+ | \frac{2 j _ {k} - 1 }{2 N } | ||
− | + | \right ) ^ {-1} | |
+ | $$ | ||
− | + | and $ N $ | |
+ | is the expected number of iterations, and one uses in it a special permutation $ \kappa _ {N} = ( j _ {1} \dots j _ {N} ) $ | ||
+ | of order $ N $ | ||
+ | that mixes well for the stability of the roots of the Chebyshev polynomials. For $ N = 2 ^ {n} $ | ||
+ | the permutation $ \kappa _ {N} $ | ||
+ | can, e.g., be constructed as follows: $ \kappa _ {2} = ( 1 , 2 ) $, | ||
+ | and if $ \kappa _ {2 ^ {i-1} } = ( j _ {1} \dots j _ {2 ^ {i-1} } ) $ | ||
+ | has already been constructed, then $ \kappa _ {2 ^ {i} } = ( j _ {1} , 2 ^ {i} + 1 - j _ {1} , j _ {2} , 2 ^ {i} + 1- j _ {2} ,\dots ) $. | ||
+ | For $ N = 16 $ | ||
+ | it has the form (1, 16, 8, 9, 4, 13, 5, 12, 2, 15, 7, 10, 3, 14, 6, 11). | ||
+ | |||
+ | There exist iteration methods using $ r $ | ||
+ | previous approximations $ u ^ {k} \dots u ^ {k- r+ 1} $. | ||
+ | They are called $ r $-step methods and have an increased rate of convergence. | ||
Iteration methods are extensively used in solving multi-dimensional problems in mathematical physics, and for some classes of problems there exist special fast-converging iteration methods. Examples are: the method of variable directions, the methods explained in [[#References|[7]]] for elliptic boundary-initial value problems, and some methods for the problem of particle transfer or radiation (cf. [[#References|[1]]]–[[#References|[7]]]; [[Variable-directions method|Variable-directions method]]). | Iteration methods are extensively used in solving multi-dimensional problems in mathematical physics, and for some classes of problems there exist special fast-converging iteration methods. Examples are: the method of variable directions, the methods explained in [[#References|[7]]] for elliptic boundary-initial value problems, and some methods for the problem of particle transfer or radiation (cf. [[#References|[1]]]–[[#References|[7]]]; [[Variable-directions method|Variable-directions method]]). | ||
Line 31: | Line 92: | ||
====References==== | ====References==== | ||
<table><TR><TD valign="top">[1]</TD> <TD valign="top"> L.V. Kantorovich, G.P. Akilov, "Functionalanalysis in normierten Räumen" , Akademie Verlag (1964) (Translated from Russian)</TD></TR><TR><TD valign="top">[2]</TD> <TD valign="top"> L. Collatz, "Funktionalanalysis und numerische Mathematik" , Springer (1964)</TD></TR><TR><TD valign="top">[3]</TD> <TD valign="top"> G.I. Marchuk, V.I. Lebedev, "Numerical methods in the theory of neutron transport" , Harwood (1986) (Translated from Russian)</TD></TR><TR><TD valign="top">[4]</TD> <TD valign="top"> N.S. Bakhvalov, "Numerical methods: analysis, algebra, ordinary differential equations" , MIR (1977) (Translated from Russian)</TD></TR><TR><TD valign="top">[5]</TD> <TD valign="top"> M.A. Krasnosel'skii, G.M. Vainikko, P.P. Zabreiko, et al., "Approximate solution of operator equations" , Wolters-Noordhoff (1972) (Translated from Russian)</TD></TR><TR><TD valign="top">[6]</TD> <TD valign="top"> V.I. Lebedev, "Optimization in iteration methods" , ''Proc. 2nd Math. School Optim. Organ. Comp. 1971'' , Inst. Kibernet. Akad. Nauk SSSR (1972) pp. 109–135 (In Russian)</TD></TR><TR><TD valign="top">[7]</TD> <TD valign="top"> R.P. Fedorenko, "Iterative methods for elliptic difference equations" ''Russian Math. Surveys'' , '''2''' (1973) pp. 129–195 ''Uspekhi Mat. Nauk'' , '''28''' (1973) pp. 121–182</TD></TR></table> | <table><TR><TD valign="top">[1]</TD> <TD valign="top"> L.V. Kantorovich, G.P. Akilov, "Functionalanalysis in normierten Räumen" , Akademie Verlag (1964) (Translated from Russian)</TD></TR><TR><TD valign="top">[2]</TD> <TD valign="top"> L. Collatz, "Funktionalanalysis und numerische Mathematik" , Springer (1964)</TD></TR><TR><TD valign="top">[3]</TD> <TD valign="top"> G.I. Marchuk, V.I. Lebedev, "Numerical methods in the theory of neutron transport" , Harwood (1986) (Translated from Russian)</TD></TR><TR><TD valign="top">[4]</TD> <TD valign="top"> N.S. Bakhvalov, "Numerical methods: analysis, algebra, ordinary differential equations" , MIR (1977) (Translated from Russian)</TD></TR><TR><TD valign="top">[5]</TD> <TD valign="top"> M.A. Krasnosel'skii, G.M. Vainikko, P.P. Zabreiko, et al., "Approximate solution of operator equations" , Wolters-Noordhoff (1972) (Translated from Russian)</TD></TR><TR><TD valign="top">[6]</TD> <TD valign="top"> V.I. Lebedev, "Optimization in iteration methods" , ''Proc. 2nd Math. School Optim. Organ. Comp. 1971'' , Inst. Kibernet. Akad. Nauk SSSR (1972) pp. 109–135 (In Russian)</TD></TR><TR><TD valign="top">[7]</TD> <TD valign="top"> R.P. Fedorenko, "Iterative methods for elliptic difference equations" ''Russian Math. Surveys'' , '''2''' (1973) pp. 129–195 ''Uspekhi Mat. Nauk'' , '''28''' (1973) pp. 121–182</TD></TR></table> | ||
− | |||
− | |||
====Comments==== | ====Comments==== |
Latest revision as of 11:46, 17 June 2020
A recursive algorithm realizing a sequence of point-to-set mappings $ A _ {k} : V \rightarrow V $,
where $ V $
is a topological space, that can be used to compute for an initial point $ u ^ {0} \in V $
a sequence of points $ u ^ {k} \in V $
by the formulas
$$ \tag{1 } u ^ {k+1} = A _ {k} u ^ {k} ,\ \ k = 0 , 1 ,\dots . $$
The operation (1) is called an iteration, while the sequence $ \{ u ^ {k} \} $ is called an iterative sequence.
Iteration methods (or methods of iterative approximation) are used both for finding a solution to an operator equation
$$ \tag{2 } A u = f , $$
a minimum of a functional, eigen values and eigen vectors of an equation $ A u = \lambda u $, etc., as well as for proving the existence of solutions to these problems. An iteration method (1) is called convergent for the initial approximation $ u ^ {0} $ to a solution $ u $ of a problem considered if $ u ^ {k} \rightarrow u $ as $ k \rightarrow \infty $.
Operators $ A _ {k} $ for solving (2), given on a metric linear space $ V $, are usually constructed by the formulas
$$ \tag{3 } A _ {k} u ^ {k} = u ^ {k} H _ {k} ( A u ^ {k} - f ) , $$
where $ \{ {H _ {k} } : {V \rightarrow V } \} $ is some sequence of operators determined by the type of the iteration method. The contracting-mapping principle and its generalizations, or variational minimization methods for a functional related to the problem, lie at the basis of constructing iteration methods of type (1), (3). Various methods for constructing $ A _ {k} $ are used, e.g. variants of the Newton method or method of descent (cf. Descent, method of). One tries to choose the $ H _ {k} $ so that a fast convergence $ u ^ {k} \rightarrow u $ is ensured under the conditions that the numerical realization of the operations $ A _ {k} u ^ {k} $, for a given amount of computer memory, is sufficiently simple, has as low complexity as possible and is numerically stable. Iteration methods for solving linear problems have been well-developed and were well-studied. The iteration methods are divided here into linear and non-linear ones. The Gauss method, the Seidel method, the successive overrelaxation method (cf. Relaxation method), and iteration methods with Chebyshev parameters belong to the linear methods; variational methods belong to the non-linear methods (e.g. the methods of steepest descent and conjugate gradients, the minimal discrepancy method, etc., cf. Steepest descent, method of; Conjugate gradients, method of). One of the efficient iteration methods is the method using Chebyshev parameters, where $ A $ is a self-adjoint operator with spectrum on $ [ m , M ] $, $ M > m > 0 $. This method provides an optimum (for given information on the boundaries of the spectrum) estimate of the convergence at a pre-assigned $ N $- th step. The method can be written in the form
$$ u ^ {k+1} = u - \alpha _ {k+1} ( A u ^ {k} - f ) ,\ \ k = 0 \dots N - 1 , $$
where
$$ \alpha _ {k+1} = 2 \left ( M + m - ( M - m ) \ \cos \frac{2 j _ {k} - 1 }{2 N } \right ) ^ {-1} $$
and $ N $ is the expected number of iterations, and one uses in it a special permutation $ \kappa _ {N} = ( j _ {1} \dots j _ {N} ) $ of order $ N $ that mixes well for the stability of the roots of the Chebyshev polynomials. For $ N = 2 ^ {n} $ the permutation $ \kappa _ {N} $ can, e.g., be constructed as follows: $ \kappa _ {2} = ( 1 , 2 ) $, and if $ \kappa _ {2 ^ {i-1} } = ( j _ {1} \dots j _ {2 ^ {i-1} } ) $ has already been constructed, then $ \kappa _ {2 ^ {i} } = ( j _ {1} , 2 ^ {i} + 1 - j _ {1} , j _ {2} , 2 ^ {i} + 1- j _ {2} ,\dots ) $. For $ N = 16 $ it has the form (1, 16, 8, 9, 4, 13, 5, 12, 2, 15, 7, 10, 3, 14, 6, 11).
There exist iteration methods using $ r $ previous approximations $ u ^ {k} \dots u ^ {k- r+ 1} $. They are called $ r $-step methods and have an increased rate of convergence.
Iteration methods are extensively used in solving multi-dimensional problems in mathematical physics, and for some classes of problems there exist special fast-converging iteration methods. Examples are: the method of variable directions, the methods explained in [7] for elliptic boundary-initial value problems, and some methods for the problem of particle transfer or radiation (cf. [1]–[7]; Variable-directions method).
References
[1] | L.V. Kantorovich, G.P. Akilov, "Functionalanalysis in normierten Räumen" , Akademie Verlag (1964) (Translated from Russian) |
[2] | L. Collatz, "Funktionalanalysis und numerische Mathematik" , Springer (1964) |
[3] | G.I. Marchuk, V.I. Lebedev, "Numerical methods in the theory of neutron transport" , Harwood (1986) (Translated from Russian) |
[4] | N.S. Bakhvalov, "Numerical methods: analysis, algebra, ordinary differential equations" , MIR (1977) (Translated from Russian) |
[5] | M.A. Krasnosel'skii, G.M. Vainikko, P.P. Zabreiko, et al., "Approximate solution of operator equations" , Wolters-Noordhoff (1972) (Translated from Russian) |
[6] | V.I. Lebedev, "Optimization in iteration methods" , Proc. 2nd Math. School Optim. Organ. Comp. 1971 , Inst. Kibernet. Akad. Nauk SSSR (1972) pp. 109–135 (In Russian) |
[7] | R.P. Fedorenko, "Iterative methods for elliptic difference equations" Russian Math. Surveys , 2 (1973) pp. 129–195 Uspekhi Mat. Nauk , 28 (1973) pp. 121–182 |
Comments
See also the editorial comments to Chebyshev iteration method.
References
[a1] | D.M. Young, "Iterative solution of large linear systems" , Acad. Press (1971) |
[a2] | A. George, J.W.-H. Liu, "Computer solution of large sparse positive definite systems" , Prentice-Hall (1981) |
[a3] | J.E., jr. Dennis, R. Schnable, "Least change secant updates for quasi-Newton methods" SIAM Review , 21 (1979) pp. 443–459 |
[a4] | J.M. Ortega, W.C. Rheinboldt, "Iterative solution of non-linear equations in several variables" , Acad. Press (1970) |
[a5] | W.C. Rheinboldt, "Methods for solving systems of nonlinear equations" , SIAM (1970) |
[a6] | J.F. Traub, "Iterative methods for the solution of equations" , Prentice-Hall (1964) |
[a7] | R. Wait, "The numerical solution of algebraic equations" , Wiley (1979) |
[a8] | E. Wasserstrom, "Numerical solution by the continuation method" SIAM Review , 15 (1973) pp. 89–119 |
[a9] | R.S. Varga, "Matrix iterative analysis" , Prentice-Hall (1962) |
Iteration algorithm. Encyclopedia of Mathematics. URL: http://encyclopediaofmath.org/index.php?title=Iteration_algorithm&oldid=16800