In
numerical analysis, fixed-point iteration is a method of computing
fixed points of a function.
More specifically, given a function
defined on the
real numbers with real values and given a point
in the
domain of
, the fixed-point iteration is
:
which gives rise to the
sequence of
iterated function applications
which is hoped to
converge to a point
. If
is continuous, then one can prove that the obtained
is a fixed point of
, i.e.,
:
More generally, the function
can be defined on any
metric space with values in that same space.
Examples
* A first simple and useful example is the
Babylonian method for computing the
square root of ''a''>0, which consists in taking
, i.e. the mean value of ''x'' and ''a/x'', to approach the limit
(from whatever starting point
). This is a special case of
Newton's method
In numerical analysis, Newton's method, also known as the Newton–Raphson method, named after Isaac Newton and Joseph Raphson, is a root-finding algorithm which produces successively better approximations to the roots (or zeroes) of a real-valu ...
quoted below.
* The fixed-point iteration
converges to the unique fixed point of the function
for any starting point
This example does satisfy (at the latest after the first iteration step) the assumptions of the
Banach fixed-point theorem. Hence, the error after n steps satisfies
(where we can take
, if we start from
.) When the error is less than a multiple of
for some constant ''q'', we say that we have
linear convergence. The Banach fixed-point theorem allows one to obtain fixed-point iterations with linear convergence.
* The requirement that ''f'' is continuous is important, as the following example shows. The iteration
converges to 0 for all values of
. However, 0 is ''not'' a fixed point of the function
as this function is ''not'' continuous at
, and in fact has no fixed points.
Attracting fixed points
An ''attracting fixed point'' of a function ''f'' is a
fixed point ''x''
fix of ''f'' such that for any value of ''x'' in the domain that is close enough to ''x''
fix, the fixed-point iteration sequence
:
converges to ''x''
fix.
The natural
cosine function ("natural" means in
radians, not degrees or other units) has exactly one fixed point, and that fixed point is attracting. In this case, "close enough" is not a stringent criterion at all—to demonstrate this, start with ''any'' real number and repeatedly press the ''cos'' key on a calculator (checking first that the calculator is in "radians" mode). It eventually converges to the
Dottie number (about 0.739085133), which is a fixed point. That is where the graph of the cosine function intersects the line
.
Not all fixed points are attracting. For example, 0 is a fixed point of the function ''f''(''x'') = 2''x'', but iteration of this function for any value other than zero rapidly diverges. We say that the fixed point of
is repelling.
An attracting fixed point is said to be a ''stable fixed point'' if it is also
Lyapunov stable.
A fixed point is said to be a ''neutrally stable fixed point'' if it is
Lyapunov stable but not attracting. The center of a
linear homogeneous differential equation of the second order is an example of a neutrally stable fixed point.
Multiple attracting points can be collected in an ''attracting fixed set''.
Banach fixed-point theorem
The
Banach fixed-point theorem gives a sufficient condition for the existence of attracting fixed points. A
contraction mapping function
defined on a
complete metric space has precisely one fixed point, and the fixed-point iteration is attracted towards that fixed point for any initial guess
in the domain of the function. Common special cases are that (1)
is defined on the real line with real values and is
Lipschitz continuous with Lipschitz constant
, and (2) the function ''f'' is continuously differentiable in an open neighbourhood of a fixed point ''x''
fix, and
.
Although there are other
fixed-point theorems, this one in particular is very useful because not all fixed-points are attractive. When constructing a fixed-point iteration, it is very important to make sure it converges to the fixed point. We can usually use the Banach fixed-point theorem to show that the fixed point is attractive.
Attractors
Attracting fixed points are a special case of a wider mathematical concept of
attractors. Fixed-point iterations are a discrete
dynamical system on one variable.
Bifurcation theory studies dynamical systems and classifies various behaviors such as attracting fixed points,
periodic orbits, or
strange attractor
In the mathematical field of dynamical systems, an attractor is a set of states toward which a system tends to evolve, for a wide variety of starting conditions of the system. System values that get close enough to the attractor values remain ...
s. An example system is the
logistic map
The logistic map is a polynomial mapping (equivalently, recurrence relation) of degree 2, often referred to as an archetypal example of how complex, chaotic behaviour can arise from very simple non-linear dynamical equations. The map was popular ...
.
Iterative methods
In computational mathematics, an iterative method is a mathematical procedure that uses an initial value to generate a sequence of improving approximate solutions for a class of problems, in which the n-th approximation is derived from the previous ones. Convergent fixed-point iterations are mathematically rigorous formalizations of iterative methods.
Iterative method examples
Convergence acceleration
The speed of convergence of the iteration sequence can be increased by using a
convergence acceleration method such as
Anderson acceleration and
Aitken's delta-squared process. The application of Aitken's method to fixed-point iteration is known as
Steffensen's method
In numerical analysis, Steffensen's method is a root-finding technique named after Johan Frederik Steffensen which is similar to Newton's method. Steffensen's method also achieves quadratic convergence, but without using derivatives as Newton's me ...
, and it can be shown that Steffensen's method yields a rate of convergence that is at least quadratic.
Chaos game
The term ''chaos game'' refers to a method of generating the
fixed point of any
iterated function system
In mathematics, iterated function systems (IFSs) are a method of constructing fractals; the resulting fractals are often self-similar. IFS fractals are more related to set theory than fractal geometry. They were introduced in 1981.
IFS fractals, ...
(IFS). Starting with any point ''x''
0, successive iterations are formed as ''x''
''k''+1 = ''f''
''r''(''x''
''k''), where ''f''
''r'' is a member of the given IFS randomly selected for each iteration. Hence the chaos game is a randomized fixed-point iteration. The chaos game allows plotting the general shape of a
fractal
In mathematics, a fractal is a geometric shape containing detailed structure at arbitrarily small scales, usually having a fractal dimension strictly exceeding the topological dimension. Many fractals appear similar at various scales, as illu ...
such as the
Sierpinski triangle by repeating the iterative process a large number of times. More mathematically, the iterations converge to the fixed point of the IFS. Whenever ''x''
0 belongs to the attractor of the IFS, all iterations ''x''
''k'' stay inside the attractor and, with probability 1, form a
dense set in the latter.
See also
*
Fixed-point combinator
*
Cobweb plot
*
Markov chain
A Markov chain or Markov process is a stochastic model describing a sequence of possible events in which the probability of each event depends only on the state attained in the previous event. Informally, this may be thought of as, "What happe ...
*
Infinite compositions of analytic functions
*
Convergence and fixed point
References
Further reading
*
*
*
*
*
*
External links
Fixed-point algorithms onlineFixed-point iteration online calculator (Mathematical Assistant on Web)
{{DEFAULTSORT:Fixed-Point Iteration
Root-finding algorithms
Iterative methods