Iterative method
   HOME

TheInfoList



OR:

In computational mathematics, an iterative method is a mathematical procedure that uses an initial value to generate a sequence of improving approximate solutions for a class of problems, in which the ''n''-th approximation is derived from the previous ones. A specific implementation of an iterative method, including the termination criteria, is an
algorithm In mathematics and computer science, an algorithm () is a finite sequence of rigorous instructions, typically used to solve a class of specific problems or to perform a computation. Algorithms are used as specifications for performing ...
of the iterative method. An iterative method is called convergent if the corresponding sequence converges for given initial approximations. A mathematically rigorous convergence analysis of an iterative method is usually performed; however,
heuristic A heuristic (; ), or heuristic technique, is any approach to problem solving or self-discovery that employs a practical method that is not guaranteed to be optimal, perfect, or rational, but is nevertheless sufficient for reaching an immediate ...
-based iterative methods are also common. In contrast, direct methods attempt to solve the problem by a finite sequence of operations. In the absence of
rounding error A roundoff error, also called rounding error, is the difference between the result produced by a given algorithm using exact arithmetic and the result produced by the same algorithm using finite-precision, rounded arithmetic. Rounding errors are d ...
s, direct methods would deliver an exact solution (for example, solving a linear system of equations A\mathbf=\mathbf by Gaussian elimination). Iterative methods are often the only choice for
nonlinear equation In mathematics and science, a nonlinear system is a system in which the change of the output is not proportional to the change of the input. Nonlinear problems are of interest to engineers, biologists, physicists, mathematicians, and many other ...
s. However, iterative methods are often useful even for linear problems involving many variables (sometimes on the order of millions), where direct methods would be prohibitively expensive (and in some cases impossible) even with the best available computing power.


Attractive fixed points

If an equation can be put into the form ''f''(''x'') = ''x'', and a solution x is an attractive fixed point of the function ''f'', then one may begin with a point ''x''1 in the
basin of attraction In the mathematical field of dynamical systems, an attractor is a set of states toward which a system tends to evolve, for a wide variety of starting conditions of the system. System values that get close enough to the attractor values remain ...
of x, and let ''x''''n''+1 = ''f''(''x''''n'') for ''n'' ≥ 1, and the sequence ''n'' ≥ 1 will converge to the solution x. Here ''x''''n'' is the ''n''th approximation or iteration of ''x'' and ''x''''n''+1 is the next or ''n'' + 1 iteration of ''x''. Alternately, superscripts in parentheses are often used in numerical methods, so as not to interfere with subscripts with other meanings. (For example, ''x''(''n''+1) = ''f''(''x''(''n'')).) If the function ''f'' is continuously differentiable, a sufficient condition for convergence is that the
spectral radius In mathematics, the spectral radius of a square matrix is the maximum of the absolute values of its eigenvalues. More generally, the spectral radius of a bounded linear operator is the supremum of the absolute values of the elements of its spectru ...
of the derivative is strictly bounded by one in a neighborhood of the fixed point. If this condition holds at the fixed point, then a sufficiently small neighborhood (basin of attraction) must exist.


Linear systems

In the case of a system of linear equations, the two main classes of iterative methods are the stationary iterative methods, and the more general
Krylov subspace In linear algebra, the order-''r'' Krylov subspace generated by an ''n''-by-''n'' matrix ''A'' and a vector ''b'' of dimension ''n'' is the linear subspace spanned by the images of ''b'' under the first ''r'' powers of ''A'' (starting from A^0=I), ...
methods.


Stationary iterative methods


Introduction

Stationary iterative methods solve a linear system with an operator approximating the original one; and based on a measurement of the error in the result ( the residual), form a "correction equation" for which this process is repeated. While these methods are simple to derive, implement, and analyze, convergence is only guaranteed for a limited class of matrices.


Definition

An ''iterative method'' is defined by : \mathbf^ := \Psi ( \mathbf^k ) \,, \quad k\geq0 and for a given linear system A\mathbf x= \mathbf b with exact solution \mathbf^* the ''error'' by : \mathbf^k := \mathbf^k - \mathbf^* \,, \quad k\geq0\,. An iterative method is called ''linear'' if there exists a matrix C \in \R^ such that : \mathbf^ = C \mathbf^k \quad \forall \, k\geq0 and this matrix is called the ''iteration matrix''. An iterative method with a given iteration matrix C is called ''convergent'' if the following holds : \lim_ C^k=0\,. An important theorem states that for a given iterative method and its iteration matrix C it is convergent if and only if its
spectral radius In mathematics, the spectral radius of a square matrix is the maximum of the absolute values of its eigenvalues. More generally, the spectral radius of a bounded linear operator is the supremum of the absolute values of the elements of its spectru ...
\rho(C) is smaller than unity, that is, : \rho(C) < 1 \,. The basic iterative methods work by
splitting Splitting may refer to: * Splitting (psychology) * Lumpers and splitters, in classification or taxonomy * Wood splitting * Tongue splitting * Splitting, railway operation Mathematics * Heegaard splitting * Splitting field * Splitting principle ...
the matrix A into : A = M - N and here the matrix M should be easily
invertible In mathematics, the concept of an inverse element generalises the concepts of opposite () and reciprocal () of numbers. Given an operation denoted here , and an identity element denoted , if , one says that is a left inverse of , and that is ...
. The iterative methods are now defined as : M \mathbf^ = N \mathbf^k + b \,, \quad k\geq0\,. From this follows that the iteration matrix is given by : C = I - M^A = M^N\,.


Examples

Basic examples of stationary iterative methods use a splitting of the matrix A such as : A = D+L+U\,,\quad D := \text( (a_)_i) where D is only the diagonal part of A , and L is the strict lower triangular part of A . Respectively, U is the strict upper triangular part of A . * Richardson method: M:=\frac I \quad (\omega \neq 0) *
Jacobi method In numerical linear algebra, the Jacobi method is an iterative algorithm for determining the solutions of a strictly diagonally dominant system of linear equations. Each diagonal element is solved for, and an approximate value is plugged in. The ...
: M:=D * Damped Jacobi method: M:=\fracD \quad (\omega \neq 0) *
Gauss–Seidel method In numerical linear algebra, the Gauss–Seidel method, also known as the Liebmann method or the method of successive displacement, is an iterative method used to solve a system of linear equations. It is named after the German mathematicians Carl ...
: M:=D+L * Successive over-relaxation method (SOR): M:=\fracD+L \quad (\omega \neq 0) *
Symmetric successive over-relaxation In applied mathematics, symmetric successive over-relaxation (SSOR), is a preconditioner. If the original matrix can be split into diagonal, lower and upper triangular as A=D+L+L^\mathsf then the SSOR preconditioner matrix is defined as M=(D+L) D^ ...
(SSOR): M := \frac (D+\omega L) D^ (D+\omega U) \quad (\omega \not \in \) Linear stationary iterative methods are also called relaxation methods.


Krylov subspace methods

Krylov subspace methods work by forming a
basis Basis may refer to: Finance and accounting * Adjusted basis, the net cost of an asset after adjusting for various tax-related items *Basis point, 0.01%, often used in the context of interest rates * Basis trading, a trading strategy consisting ...
of the sequence of successive matrix powers times the initial residual (the Krylov sequence). The approximations to the solution are then formed by minimizing the residual over the subspace formed. The prototypical method in this class is the conjugate gradient method (CG) which assumes that the system matrix A is
symmetric Symmetry (from grc, συμμετρία "agreement in dimensions, due proportion, arrangement") in everyday language refers to a sense of harmonious and beautiful proportion and balance. In mathematics, "symmetry" has a more precise definiti ...
positive-definite. For symmetric (and possibly indefinite) A one works with the
minimal residual method In mathematics, the generalized minimal residual method (GMRES) is an iterative method for the numerical solution of an indefinite nonsymmetric system of linear equations. The method approximates the solution by the vector in a Krylov subspace with ...
(MINRES). In the case of non-symmetric matrices, methods such as the
generalized minimal residual method In mathematics, the generalized minimal residual method (GMRES) is an iterative method for the numerical solution of an indefinite nonsymmetric system of linear equations. The method approximates the solution by the vector in a Krylov subspace wit ...
(GMRES) and the
biconjugate gradient method In mathematics, more specifically in numerical linear algebra, the biconjugate gradient method is an algorithm to solve systems of linear equations :A x= b.\, Unlike the conjugate gradient method, this algorithm does not require the matrix A to ...
(BiCG) have been derived.


Convergence of Krylov subspace methods

Since these methods form a basis, it is evident that the method converges in ''N'' iterations, where ''N'' is the system size. However, in the presence of rounding errors this statement does not hold; moreover, in practice ''N'' can be very large, and the iterative process reaches sufficient accuracy already far earlier. The analysis of these methods is hard, depending on a complicated function of the
spectrum A spectrum (plural ''spectra'' or ''spectrums'') is a condition that is not limited to a specific set of values but can vary, without gaps, across a continuum. The word was first used scientifically in optics to describe the rainbow of colors ...
of the operator.


Preconditioners

The approximating operator that appears in stationary iterative methods can also be incorporated in Krylov subspace methods such as
GMRES In mathematics, the generalized minimal residual method (GMRES) is an iterative method for the numerical solution of an indefinite nonsymmetric system of linear equations. The method approximates the solution by the vector in a Krylov subspace wit ...
(alternatively, preconditioned Krylov methods can be considered as accelerations of stationary iterative methods), where they become transformations of the original operator to a presumably better conditioned one. The construction of preconditioners is a large research area.


History

Jamshīd al-Kāshī Ghiyāth al-Dīn Jamshīd Masʿūd al-Kāshī (or al-Kāshānī) ( fa, غیاث الدین جمشید کاشانی ''Ghiyās-ud-dīn Jamshīd Kāshānī'') (c. 1380 Kashan, Iran – 22 June 1429 Samarkand, Transoxania) was a Persian astronomer ...
used iterative methods to calculate the sine of 1° and in ''The Treatise of Chord and Sine'' to high precision. An early iterative method for solving a linear system appeared in a letter of
Gauss Johann Carl Friedrich Gauss (; german: Gauß ; la, Carolus Fridericus Gauss; 30 April 177723 February 1855) was a German mathematician and physicist who made significant contributions to many fields in mathematics and science. Sometimes refer ...
to a student of his. He proposed solving a 4-by-4 system of equations by repeatedly solving the component in which the residual was the largest . The theory of stationary iterative methods was solidly established with the work of D.M. Young starting in the 1950s. The conjugate gradient method was also invented in the 1950s, with independent developments by Cornelius Lanczos, Magnus Hestenes and
Eduard Stiefel Eduard L. Stiefel (21 April 1909 – 25 November 1978) was a Swiss mathematician. Together with Cornelius Lanczos and Magnus Hestenes, he invented the conjugate gradient method, and gave what is now understood to be a partial construction of the ...
, but its nature and applicability were misunderstood at the time. Only in the 1970s was it realized that conjugacy based methods work very well for partial differential equations, especially the elliptic type.


See also

*
Closed-form expression In mathematics, a closed-form expression is a mathematical expression that uses a finite number of standard operations. It may contain constants, variables, certain well-known operations (e.g., + − × ÷), and functions (e.g., ''n''th ro ...
* Iterative refinement * Kaczmarz method *
Non-linear least squares Non-linear least squares is the form of least squares analysis used to fit a set of ''m'' observations with a model that is non-linear in ''n'' unknown parameters (''m'' ≥ ''n''). It is used in some forms of nonlinear regression. The ...
*
Numerical analysis Numerical analysis is the study of algorithms that use numerical approximation (as opposed to symbolic manipulations) for the problems of mathematical analysis (as distinguished from discrete mathematics). It is the study of numerical methods ...
*
Root-finding algorithm In mathematics and computing, a root-finding algorithm is an algorithm for finding zeros, also called "roots", of continuous functions. A zero of a function , from the real numbers to real numbers or from the complex numbers to the complex numbers ...


References


External links


Templates for the Solution of Linear Systems
{{Authority control Numerical analysis