Fixed-point Computation
   HOME

TheInfoList



OR:

Fixed-point computation refers to the process of computing an exact or approximate fixed point of a given function. In its most common form, the given function f satisfies the condition to the
Brouwer fixed-point theorem Brouwer's fixed-point theorem is a fixed-point theorem in topology, named after Luitzen Egbertus Jan Brouwer, L. E. J. (Bertus) Brouwer. It states that for any continuous function f mapping a nonempty compactness, compact convex set to itself, the ...
: that is, f is continuous and maps the unit ''d''-cube to itself. The
Brouwer fixed-point theorem Brouwer's fixed-point theorem is a fixed-point theorem in topology, named after Luitzen Egbertus Jan Brouwer, L. E. J. (Bertus) Brouwer. It states that for any continuous function f mapping a nonempty compactness, compact convex set to itself, the ...
guarantees that f has a fixed point, but the proof is not constructive. Various algorithms have been devised for computing an approximate fixed point. Such algorithms are used in economics for computing a
market equilibrium In economics, economic equilibrium is a situation in which the economic forces of supply and demand are balanced, meaning that economic variables will no longer change. Market equilibrium in this case is a condition where a market price is esta ...
, in
game theory Game theory is the study of mathematical models of strategic interactions. It has applications in many fields of social science, and is used extensively in economics, logic, systems science and computer science. Initially, game theory addressed ...
for computing a
Nash equilibrium In game theory, the Nash equilibrium is the most commonly used solution concept for non-cooperative games. A Nash equilibrium is a situation where no player could gain by changing their own strategy (holding all other players' strategies fixed) ...
, and in
dynamic system In mathematics, a dynamical system is a system in which a function describes the time dependence of a point in an ambient space, such as in a parametric curve. Examples include the mathematical models that describe the swinging of a clock ...
analysis.


Definitions

The unit interval is denoted by E :=
, 1 The comma is a punctuation mark that appears in several variants in different languages. Some typefaces render it as a small line, slightly curved or straight, but inclined from the vertical; others give it the appearance of a miniature fille ...
/math>, and the unit ''d''-dimensional cube is denoted by E^d. A
continuous function In mathematics, a continuous function is a function such that a small variation of the argument induces a small variation of the value of the function. This implies there are no abrupt changes in value, known as '' discontinuities''. More preci ...
f is defined on E^d (from E^d to itself)''.'' Often, it is assumed that f is not only continuous but also
Lipschitz continuous In mathematical analysis, Lipschitz continuity, named after Germany, German mathematician Rudolf Lipschitz, is a strong form of uniform continuity for function (mathematics), functions. Intuitively, a Lipschitz continuous function is limited in h ...
, that is, for some constant L, , f(x)-f(y), \leq L\cdot , x-y, for all x,y in E^d. A fixed point of f is a point x in E^d such that f(x) = x. By the
Brouwer fixed-point theorem Brouwer's fixed-point theorem is a fixed-point theorem in topology, named after Luitzen Egbertus Jan Brouwer, L. E. J. (Bertus) Brouwer. It states that for any continuous function f mapping a nonempty compactness, compact convex set to itself, the ...
, any continuous function from E^d to itself has a fixed point. But for general functions, it is impossible to compute a fixed point precisely, since it can be an arbitrary real number. Fixed-point computation algorithms look for ''approximate'' fixed points. There are several criteria for an approximate fixed point. Several common criteria are: * The residual criterion: given an approximation parameter \varepsilon>0 , An -residual fixed-point of f is a point x in E^d' such that , f(x)-x, \leq \varepsilon, where here , \cdot, denotes the
maximum norm In mathematical analysis, the uniform norm (or ) assigns, to real- or complex-valued bounded functions defined on a set , the non-negative number :\, f\, _\infty = \, f\, _ = \sup\left\. This norm is also called the , the , the , or, when th ...
. That is, all d coordinates of the difference f(x)-x should be at most . * The absolute criterion: given an approximation parameter \delta>0, A δ-absolute fixed-point of f is a point x in E^d such that , x-x_0, \leq \delta, where x_0 is any fixed-point of f. * The relative criterion: given an approximation parameter \delta>0, A δ-relative fixed-point of f is a point ''x'' in E^d such that , x-x_0, /, x_0, \leq \delta, where x_0 is any fixed-point of f. For Lipschitz-continuous functions, the absolute criterion is stronger than the residual criterion: If f is Lipschitz-continuous with constant L, then , x-x_0, \leq \delta implies , f(x)-f(x_0), \leq L\cdot \delta. Since x_0 is a fixed-point of f, this implies , f(x)-x_0, \leq L\cdot \delta, so , f(x)-x, \leq (1+L)\cdot \delta. Therefore, a δ-absolute fixed-point is also an -residual fixed-point with \varepsilon = (1+L)\cdot \delta. The most basic step of a fixed-point computation algorithm is a value query: given any x in E^d, the algorithm is provided with an oracle \tilde to f that returns the value f(x). The accuracy of the approximate fixed-point depends upon the error in the oracle \tilde(x). The function f is accessible via evaluation queries: for any x, the algorithm can evaluate f(x). The run-time complexity of an algorithm is usually given by the number of required evaluations.


Contractive functions

A Lipschitz-continuous function with constant L is called contractive if L<1; it is called weakly-contractive if L\le 1. Every contractive function satisfying Brouwer's conditions has a ''unique'' fixed point. Moreover, fixed-point computation for contractive functions is easier than for general functions. The first algorithm for fixed-point computation was the
fixed-point iteration In numerical analysis, fixed-point iteration is a method of computing fixed points of a function. More specifically, given a function f defined on the real numbers with real values and given a point x_0 in the domain of f, the fixed-point itera ...
algorithm of Banach. Banach's fixed-point theorem implies that, when fixed-point iteration is applied to a contraction mapping, the error after t iterations is in O(L^t). Therefore, the number of evaluations required for a \delta-relative fixed-point is approximately \log_L(\delta) = \log(\delta)/\log(L) = \log(1/\delta)/\log(1/L) . Sikorski and Wozniakowski showed that Banach's algorithm is optimal when the dimension is large. Specifically, when d\geq \log(1/\delta)/\log(1/L) , the number of required evaluations of ''any'' algorithm for \delta-relative fixed-point is larger than 50% the number of evaluations required by the iteration algorithm. Note that when L approaches 1, the number of evaluations approaches infinity. No finite algorithm can compute a \delta-absolute fixed point for all functions with L=1. When L < 1 and ''d'' = 1, the optimal algorithm is the Fixed Point Envelope (FPE) algorithm of Sikorski and Wozniakowski. It finds a ''δ''-relative fixed point using O(\log(1/\delta) + \log \log(1/(1-L))) queries, and a ''δ''-absolute fixed point using O(\log(1/\delta)) queries. This is faster than the fixed-point iteration algorithm. When d>1 but not too large, and L\le 1, the optimal algorithm is the interior-ellipsoid algorithm (based on the
ellipsoid method In mathematical optimization, the ellipsoid method is an iterative method for convex optimization, minimizing convex functions over convex sets. The ellipsoid method generates a sequence of ellipsoids whose volume uniformly decreases at every ste ...
). It finds an -residual fixed-point using O(d\cdot \log(1/\varepsilon)) evaluations. When L<1, it finds a \delta-absolute fixed point using O(d\cdot log(1/\delta) + \log(1/(1-L)) evaluations. Shellman and Sikorski presented an algorithm called BEFix (Bisection Envelope Fixed-point) for computing an -residual fixed-point of a two-dimensional function with 'L\le 1, using only 2 \lceil\log_2(1/\varepsilon)\rceil+1 queries. They later presented an improvement called BEDFix (Bisection Envelope Deep-cut Fixed-point), with the same worst-case guarantee but better empirical performance. When L<1, BEDFix can also compute a \delta-absolute fixed-point using O(\log(1/\varepsilon)+\log(1/(1-L))) queries. Shellman and Sikorski presented an algorithm called PFix for computing an -residual fixed-point of a ''d''-dimensional function with ''L ≤'' 1, using O(\log^d(1/\varepsilon)) queries. When L < 1, PFix can be executed with \varepsilon = (1-L)\cdot \delta, and in that case, it computes a δ-absolute fixed-point, using O(\log^d(1/ 1-L)\delta) queries. It is more efficient than the iteration algorithm when L is close to 1. The algorithm is recursive: it handles a ''d''-dimensional function by recursive calls on (''d''-1)-dimensional functions.


Algorithms for differentiable functions

When the function f is differentiable, and the algorithm can evaluate its derivative (not only f itself), the Newton method can be used and it is much faster.


General functions

For functions with Lipschitz constant L > 1, computing a fixed-point is much harder.


One dimension

For a 1-dimensional function (''d'' = 1), a \delta-absolute fixed-point can be found using O(\log(1/\delta)) queries using the
bisection method In mathematics, the bisection method is a root-finding method that applies to any continuous function for which one knows two values with opposite signs. The method consists of repeatedly bisecting the interval defined by these values and t ...
: start with the interval E :=
, 1 The comma is a punctuation mark that appears in several variants in different languages. Some typefaces render it as a small line, slightly curved or straight, but inclined from the vertical; others give it the appearance of a miniature fille ...
/math>; at each iteration, let x be the center of the current interval, and compute f(x); if f(x) > x then recurse on the sub-interval to the right of x; otherwise, recurse on the interval to the left of x. Note that the current interval always contains a fixed point, so after O(\log(1/\delta)) queries, any point in the remaining interval is a \delta-absolute fixed-point of f Setting \delta := \varepsilon/(L+1), where L is the Lipschitz constant, gives an -residual fixed-point, using O(\log(L/\varepsilon) = \log(L) + \log(1/\varepsilon)) queries.


Two or more dimensions

For functions in two or more dimensions, the problem is much more challenging. Shellman and Sikorski proved that for any integers ''d'' ≥ 2 and L > 1, finding a δ-absolute fixed-point of ''d''-dimensional L-Lipschitz functions might require infinitely many evaluations. The proof idea is as follows. For any integer ''T'' > 1 and any sequence of ''T'' of evaluation queries (possibly adaptive), one can construct two functions that are Lipschitz-continuous with constant L, and yield the same answer to all these queries, but one of them has a unique fixed-point at (''x'', 0) and the other has a unique fixed-point at (''x'', 1). Any algorithm using ''T'' evaluations cannot differentiate between these functions, so cannot find a δ-absolute fixed-point. This is true for any finite integer ''T''. Several algorithms based on function evaluations have been developed for finding an -residual fixed-point * The first algorithm to approximate a fixed point of a general function was developed by
Herbert Scarf Herbert Eli "Herb" Scarf (July 25, 1930 – November 15, 2015) was an American mathematical economist and Sterling Professor of Economics at Yale University. Education and career Scarf was born in Philadelphia, the son of Jewish emigrants from ...
in 1967. Scarf's algorithm finds an -residual fixed-point by finding a fully labeled "primitive set", in a construction similar to
Sperner's lemma In mathematics, Sperner's lemma is a combinatorial result on colorings of triangulations, analogous to the Brouwer fixed point theorem, which is equivalent to it. It states that every Sperner coloring (described below) of a triangulation of an ...
. * A later algorithm by Harold Kuhn used simplices and simplicial partitions instead of primitive sets. * Developing the simplicial approach further, Orin Harrison Merrill presented the ''restart algorithm''. * B. Curtis Eaves presented the ''
homotopy In topology, two continuous functions from one topological space to another are called homotopic (from and ) if one can be "continuously deformed" into the other, such a deformation being called a homotopy ( ; ) between the two functions. ...
algorithm''. The algorithm works by starting with an affine function that approximates f, and deforming it towards f while following the fixed point''.'' A book by Michael Todd surveys various algorithms developed until 1976. *
David Gale David Gale (December 13, 1921 – March 7, 2008) was an American mathematician and economist. He was a professor emeritus at the University of California, Berkeley, affiliated with the departments of mathematics, economics, and industrial ...
showed that computing a fixed point of an ''n''-dimensional function (on the unit ''d''-dimensional cube) is equivalent to deciding who is the winner in a ''d''-dimensional game of Hex (a game with ''d'' players, each of whom needs to connect two opposite faces of a ''d''-cube). Given the desired accuracy ' ** Construct a Hex board of size ''kd'', where k > 1/\varepsilon. Each vertex ''z'' corresponds to a point ''z''/''k'' in the unit ''n''-cube. ** Compute the difference f(''z''/''k'') - ''z''/''k''; note that the difference is an ''n''-vector. ** Label the vertex ''z'' by a label in 1, ..., ''d'', denoting the largest coordinate in the difference vector. ** The resulting labeling corresponds to a possible play of the ''d''-dimensional Hex game among ''d'' players. This game must have a winner, and Gale presents an algorithm for constructing the winning path. ** In the winning path, there must be a point in which ''fi''(''z''/''k'') - ''z''/''k'' is positive, and an adjacent point in which ''fi''(''z''/''k'') - ''z''/''k'' is negative. This means that there is a fixed point of f between these two points. In the worst case, the number of function evaluations required by all these algorithms is exponential in the binary representation of the accuracy, that is, in \Omega(1/\varepsilon).


Query complexity

Hirsch, Papadimitriou and Vavasis proved that ''any'' algorithm based on function evaluations, that finds an -residual fixed-point of ''f,'' requires \Omega(L'/\varepsilon) function evaluations, where L' is the Lipschitz constant of the function f(x)-x (note that L-1 \leq L' \leq L+1). More precisely: * For a 2-dimensional function (''d''=2), they prove a tight bound \Theta(L'/\varepsilon). * For any d ≥ 3, finding an -residual fixed-point of a ''d''-dimensional function requires \Omega((L'/\varepsilon)^) queries and O((L'/\varepsilon)^) queries. The latter result leaves a gap in the exponent. Chen and Deng closed the gap. They proved that, for any ''d'' ≥ 2 and 1/\varepsilon > 4 d and L'/\varepsilon > 192 d^3, the number of queries required for computing an -residual fixed-point is in \Theta((L'/\varepsilon)^).


Discrete fixed-point computation

A discrete function is a function defined on a subset of ''\mathbb^d'' (the ''d''-dimensional integer grid). There are several discrete fixed-point theorems, stating conditions under which a discrete function has a fixed point. For example, the Iimura-Murota-Tamura theorem states that (in particular) if f is a function from a rectangle subset of ''\mathbb^d'' to itself, and f is ''hypercubic direction-preserving'', then f has a fixed point. Let f be a direction-preserving function from the integer cube \^d to itself. Chen and Deng prove that, for any ''d'' ≥ 2 and ''n'' > 48''d'', computing such a fixed point requires \Theta(n^) function evaluations. Chen and Deng define a different discrete-fixed-point problem, which they call 2D-BROUWER. It considers a discrete function f on \^2 such that, for every ''x'' on the grid, f(''x'') - ''x'' is either (0, 1) or (1, 0) or (-1, -1). The goal is to find a square in the grid, in which all three labels occur. The function f must map the square \^2to itself, so it must map the lines ''x'' = 0 and ''y'' = 0 to either (0, 1) or (1, 0); the line ''x'' = ''n'' to either (-1, -1) or (0, 1); and the line ''y'' = ''n'' to either (-1, -1) or (1,0). The problem can be reduced to 2D-SPERNER (computing a fully-labeled triangle in a triangulation satisfying the conditions to
Sperner's lemma In mathematics, Sperner's lemma is a combinatorial result on colorings of triangulations, analogous to the Brouwer fixed point theorem, which is equivalent to it. It states that every Sperner coloring (described below) of a triangulation of an ...
), and therefore it is
PPAD-complete In computer science, PPAD ("Polynomial Parity Arguments on Directed graphs") is a complexity class introduced by Christos Papadimitriou in 1994. PPAD is a subclass of TFNP based on functions that can be shown to be total by a parity argument. The ...
. This implies that computing an approximate fixed-point is PPAD-complete even for very simple functions.


Relation between fixed-point computation and root-finding algorithms

Given a function g from E^d to ''R'', a root of g is a point ''x'' in E^d such that g(''x'')=0. An -root of g is a point ''x'' in E^d such that g(x)\leq \varepsilon. Fixed-point computation is a special case of root-finding: given a function f on E^d, define g(x) := , f(x)-x, . ''X'' is a fixed-point of f if and only if ''x'' is a root of g, and ''x'' is an -residual fixed-point of f if and only if ''x'' is an -root of g. Therefore, any
root-finding algorithm In numerical analysis, a root-finding algorithm is an algorithm for finding zeros, also called "roots", of continuous functions. A zero of a function is a number such that . As, generally, the zeros of a function cannot be computed exactly nor ...
(an algorithm that computes an approximate root of a function) can be used to find an approximate fixed-point. The opposite is not true: finding an approximate root of a general function may be harder than finding an approximate fixed point. In particular, Sikorski proved that finding an -root requires \Omega(1/\varepsilon^d) function evaluations. This gives an exponential lower bound even for a one-dimensional function (in contrast, an -residual fixed-point of a one-dimensional function can be found using O(\log(1/\varepsilon)) queries using the
bisection method In mathematics, the bisection method is a root-finding method that applies to any continuous function for which one knows two values with opposite signs. The method consists of repeatedly bisecting the interval defined by these values and t ...
). Here is a proof sketch. Construct a function g that is slightly larger than everywhere in E^d except in some small cube around some point ''x''0, where ''x''0 is the unique root of g. If g is
Lipschitz continuous In mathematical analysis, Lipschitz continuity, named after Germany, German mathematician Rudolf Lipschitz, is a strong form of uniform continuity for function (mathematics), functions. Intuitively, a Lipschitz continuous function is limited in h ...
with constant L, then the cube around ''x''0 can have a side-length of \varepsilon/L. Any algorithm that finds an -root of g must check a set of cubes that covers the entire E^d; the number of such cubes is at least (L/\varepsilon)^d. However, there are classes of functions for which finding an approximate root is equivalent to finding an approximate fixed point. One example is the class of functions g such that g(x)+x maps E^d to itself (that is: g(x)+x is in E^d for all x in E^d). This is because, for every such function, the function f(x) := g(x)+x satisfies the conditions of Brouwer's fixed-point theorem. ''X'' is a fixed-point of f if and only if ''x'' is a root of g, and ''x'' is an -residual fixed-point of f if and only if ''x'' is an -root of g. Chen and Deng show that the discrete variants of these problems are computationally equivalent: both problems require \Theta(n^) function evaluations.


Communication complexity

Roughgarden and Weinstein studied the
communication complexity In theoretical computer science, communication complexity studies the amount of communication required to solve a problem when the input to the problem is distributed among two or more parties. The study of communication complexity was first intro ...
of computing an approximate fixed-point. In their model, there are two agents: one of them knows a function f and the other knows a function g. Both functions are Lipschitz continuous and satisfy Brouwer's conditions. The goal is to compute an approximate fixed point of the
composite function In mathematics, the composition operator \circ takes two functions, f and g, and returns a new function h(x) := (g \circ f) (x) = g(f(x)). Thus, the function is applied after applying to . (g \circ f) is pronounced "the composition of an ...
g\circ f. They show that the deterministic communication complexity is in \Omega(2^d).


References


Further reading

* {{cite journal , last1=Yannakakis , first1=Mihalis , title=Equilibria, fixed points, and complexity classes , journal=Computer Science Review , date=May 2009 , volume=3 , issue=2 , pages=71–85 , doi=10.1016/j.cosrev.2009.03.004 , url=https://drops.dagstuhl.de/opus/volltexte/2008/1311/, arxiv=0802.2831 Fixed-point theorems Numerical analysis