Adomian decomposition method
   HOME

TheInfoList



OR:

The Adomian decomposition method (ADM) is a semi-analytical method for solving ordinary and partial
nonlinear In mathematics and science, a nonlinear system is a system in which the change of the output is not proportional to the change of the input. Nonlinear problems are of interest to engineers, biologists, physicists, mathematicians, and many othe ...
differential equation In mathematics, a differential equation is an equation that relates one or more unknown functions and their derivatives. In applications, the functions generally represent physical quantities, the derivatives represent their rates of change, an ...
s. The method was developed from the 1970s to the 1990s by George Adomian, chair of the Center for Applied Mathematics at the
University of Georgia , mottoeng = "To teach, to serve, and to inquire into the nature of things.""To serve" was later added to the motto without changing the seal; the Latin motto directly translates as "To teach and to inquire into the nature of things." , establ ...
. It is further extensible to
stochastic system In probability theory and related fields, a stochastic () or random process is a mathematical object usually defined as a family of random variables. Stochastic processes are widely used as mathematical models of systems and phenomena that appea ...
s by using the
Ito integral Ito may refer to: Places * Ito Island, an island of Milne Bay Province, Papua New Guinea * Ito Airport, an airport in the Democratic Republic of the Congo * Ito District, Wakayama, a district located in Wakayama Prefecture, Japan * Itō, Shizuo ...
. The aim of this method is towards a unified theory for the solution of partial differential equations (PDE); an aim which has been superseded by the more general theory of the homotopy analysis method. The crucial aspect of the method is employment of the "Adomian polynomials" which allow for solution convergence of the nonlinear portion of the equation, without simply linearizing the system. These
polynomial In mathematics, a polynomial is an expression consisting of indeterminates (also called variables) and coefficients, that involves only the operations of addition, subtraction, multiplication, and positive-integer powers of variables. An example ...
s mathematically generalize to a
Maclaurin series Maclaurin or MacLaurin is a surname. Notable people with the surname include: * Colin Maclaurin (1698–1746), Scottish mathematician * Normand MacLaurin (1835–1914), Australian politician and university administrator * Henry Normand MacLaurin ...
about an arbitrary external parameter; which gives the solution method more flexibility than direct
Taylor series In mathematics, the Taylor series or Taylor expansion of a function is an infinite sum of terms that are expressed in terms of the function's derivatives at a single point. For most common functions, the function and the sum of its Taylor ser ...
expansion.


Ordinary differential equations

Adomian method is well suited to solve
Cauchy problem A Cauchy problem in mathematics asks for the solution of a partial differential equation that satisfies certain conditions that are given on a hypersurface in the domain. A Cauchy problem can be an initial value problem or a boundary value prob ...
s, an important class of problems which include initial conditions problems.


Application to a first order nonlinear system

An example of initial condition problem for an ordinary differential equation is the following: : y^\prime(t) + y^(t) = -1, : y(0) = 0. To solve the problem, the highest degree differential operator (written here as ''L'') is put on the left side, in the following way: : Ly = -1 -y^, with ''L'' = d/d''t'' and L^=\int_^() . Now the solution is assumed to be an infinite series of contributions: : y = y_ + y_ + y_ + y_ + \cdots. Replacing in the previous expression, we obtain: : (y_ + y_ + y_ + y_ + \cdots) = y(0) + L^ 1 - (y_ + y_ + y_ + y_ + \cdots)^ Now we identify ''y''0 with some explicit expression on the right, and ''y''''i'', ''i'' = 1, 2, 3, ..., with some expression on the right containing terms of lower order than ''i''. For instance: : \begin &y_ &=&\ y(0)+ L^(-1) &=& -t \\ &y_ &=& -L^(y_^) =-L^(t^) &=& -t^/3 \\ &y_ &=& -L^(2y_y_) &=& -2t^/15 \\ &y_ &=& -L^(y_^+2y_y_) &=& -17t^/315. \end In this way, any contribution can be explicitly calculated at any order. If we settle for the four first terms, the approximant is the following: : \begin y &= y_ + y_ + y_ + y_ + \cdots \\ & = -\left t + \frac t^ + \frac t^ + \frac t^ + \cdots \right \end


Application to Blasius equation

A second example, with more complex boundary conditions is the Blasius equation for a flow in a
boundary layer In physics and fluid mechanics, a boundary layer is the thin layer of fluid in the immediate vicinity of a bounding surface formed by the fluid flowing along the surface. The fluid's interaction with the wall induces a no-slip boundary cond ...
: : \frac + \frac u \frac = 0 With the following conditions at the boundaries: : \begin u(0) &= 0 \\ u^(0) &= 0 \\ u^(x) &\to 1, \qquad x \to \infty \end Linear and non-linear operators are now called L = \frac and N =\frac u \frac, respectively. Then, the expression becomes: : L u + N u = 0 and the solution may be expressed, in this case, in the following simple way: : u = \alpha + \beta x + \gamma x^/2 - L^ N u where: L^ \xi (x) = \int dx \int \mathrmx \int \mathrmx \;\; \xi(x) If: : \begin u &= u^ + u^ + u^ + \cdots + u^ \\ &=\alpha + \beta x + \gamma x^/2 - \frac L^ (u^+u^+u^+\cdots+u^) \frac(u^ + u^ + u^ + \cdots + u^) \end and: : \begin u^ &= \alpha + \beta x + \gamma x^/2\\ u^ &= -\frac L^(u^u^) &=& -L^ A_ \\ u^ &= -\frac L^(u^u^+u^u^) &=& -L^ A_ \\ u^ &= -\frac L^(u^u^+u^u^+u^u^) &=& -L^ A_ \\ &\cdots \end Adomian’s polynomials to linearize the non-linear term can be obtained systematically by using the following rule: : A_ = \frac \frac f(u(\lambda))\mid_, where: \frac u(\lambda)\mid_ = n! u_ Boundary conditions must be applied, in general, at the end of each approximation. In this case, the integration constants must be grouped into three final independent constants. However, in our example, the three constants appear grouped from the beginning in the form shown in the formal solution above. After applying the two first boundary conditions we obtain the so-called Blasius series: : u = \frac x^2 - \frac\left(\frac\right) + \frac\left(\frac\right) - \frac \left(\frac\right) + \cdots To obtain γ we have to apply boundary conditions at ∞, which may be done by writing the series as a
Padé approximant In mathematics, a Padé approximant is the "best" approximation of a function near a specific point by a rational function of given order. Under this technique, the approximant's power series agrees with the power series of the function it is ap ...
: : f(z) = \sum_^ c_ z^ = \frac where ''L'' = ''M''. The limit at \infty of this expression is ''a''''L''/''b''''M''. If we choose ''b''0 = 1, ''M'' linear equations for the ''b'' coefficients are obtained: : \left \begin c_ & c_ & \cdots & c_ \\ c_ & c_ & \cdots & c_ \\ \vdots & \vdots & & \vdots \\ c_ & c_ & \cdots & c_ \end \right \left \begin b_ \\ b_ \\ \vdots \\ b_ \end \right = - \left \begin c_ \\ c_ \\ \vdots \\ c_ \end \right Then, we obtain the ''a'' coefficients by means of the following sequence: : \begin a_ &= c_ \\ a_ &= c_ + b_ c_ \\ a_ &= c_ + b_c_+b_c_ \\ &\cdots \\ a_ &= c_ + \sum_^ b_ c_. \end In our example: : u'(x) = \gamma x - \frac \left(\frac\right) + \frac \left(\frac\right) - \frac \left(\frac\right) Which when γ = 0.0408 becomes: : u'(x) = \frac, with the limit: : \lim_ u'(x) = 1.004. Which is approximately equal to 1 (from boundary condition (3)) with an accuracy of 4/1000.


Partial differential equations


Application to a rectangular system with nonlinearity

One of the most frequent problems in physical sciences is to obtain the solution of a (linear or nonlinear) partial differential equation which satisfies a set of functional values on a rectangular boundary. An example is the following problem: : \frac + \frac - b \frac = \rho(x, y) \qquad (1) with the following boundary conditions defined on a rectangle: : u(x=0) = f_(y) \quad\text\quad u(x=x_) = f_(y) \qquad \text : u(y=-y_) = g_(x) \quad\text\quad u(y=y_) = g_(x) \qquad \text This kind of partial differential equation appears frequently coupled with others in
science Science is a systematic endeavor that Scientific method, builds and organizes knowledge in the form of Testability, testable explanations and predictions about the universe. Science may be as old as the human species, and some of the earli ...
and
engineering Engineering is the use of scientific principles to design and build machines, structures, and other items, including bridges, tunnels, roads, vehicles, and buildings. The discipline of engineering encompasses a broad range of more speciali ...
. For instance, in the
incompressible fluid In fluid mechanics or more generally continuum mechanics, incompressible flow ( isochoric flow) refers to a flow in which the material density is constant within a fluid parcel—an infinitesimal volume that moves with the flow velocity. An eq ...
flow problem, the
Navier–Stokes equations In physics, the Navier–Stokes equations ( ) are partial differential equations which describe the motion of viscous fluid substances, named after French engineer and physicist Claude-Louis Navier and Anglo-Irish physicist and mathematician Geo ...
must be solved in parallel with a
Poisson equation Poisson's equation is an elliptic partial differential equation of broad utility in theoretical physics. For example, the solution to Poisson's equation is the potential field caused by a given electric charge or mass density distribution; with t ...
for the pressure.


Decomposition of the system

Let us use the following notation for the problem (1): : L_ u + L_ u + N u = \rho(x, y) \qquad (2) where ''L''x, ''L''y are double derivate operators and ''N'' is a non-linear operator. The formal solution of (2) is: : u = a(y) + b(y) x + L_^ \rho(x, y) - L_^ L_ u - L_^ N u \qquad (3) Expanding now u as a set of contributions to the solution we have: : u = u_ + u_ + u_ + u_ + \cdots By substitution in (3) and making a one-to-one correspondence between the contributions on the left side and the terms on the right side we obtain the following iterative scheme: : \begin u_ &= a_(y) + b_(y) x + L_^ \rho(x, y) \\ u_ &= a_(y) + b_(y) x - L_^ L_ u_ + b \int dx A_ \\ &\cdots \\ u_ &= a_(y) + b_(y) x - L_^ L_ u_ + b \int dx A_ \quad 0 < n < \infty \end where the couple is the solution of the following system of equations: : \begin \varphi^(x=0) &= f_(y) \\ \varphi^(x=x_) &= f_(y), \end here \varphi^ \equiv \sum_^ u_ is the ''n''th-order approximant to the solution and ''N u'' has been consistently expanded in Adomian polynomials: : \begin N u &= -b \partial_ u^ = -b \partial_ (u_ + u_ + u_ + u_ + \cdots)(u_ + u_ + u_ + u_ + \cdots) \\ &= -b \partial_ (u_ u_ + 2 u_ u_ + u_ u_ + 2 u_ u_ + \cdots) \\ &= -b \partial_ \sum_^ A(n-1), \end where A_ = \sum_^ C(\nu, n) f^(u_) and ''f''(''u'') = ''u''2 in the example (1). Here ''C''(ν, ''n'') are products (or sum of products) of ν components of ''u'' whose subscripts sum up to ''n'', divided by the factorial of the number of repeated subscripts. It is only a thumb-rule to order systematically the decomposition to be sure that all the combinations appearing are utilized sooner or later. The \sum_^ A_ is equal to the sum of a generalized Taylor series about ''u''0. For the example (1) the Adomian polynomials are: : \begin A_ &= u_^ \\ A_ &= 2 u_ u_ \\ A_ &= u_^ + 2 u_ u_ \\ A_ &= 2 u_ u_ + 2 u_ u_ \\ & \cdots \end Other possible choices are also possible for the expression of ''A''''n''.


Series solutions

Cherruault established that the series terms obtained by Adomian's method approach zero as 1/(''mn'')! if ''m'' is the order of the highest linear differential operator and that \lim_ \varphi^ = u. With this method the solution can be found by systematically integrating along any of the two directions: in the ''x''-direction we would use expression (3); in the alternative ''y''-direction we would use the following expression: : u = c(x) + d(x) y + L_^ \rho(x, y) - L_^ L_ u - L_^ N u where: ''c''(''x''), ''d''(''x'') is obtained from the boundary conditions at ''y'' = - ''y''''l'' and ''y'' = ''y''''l'': : \begin u(y=-y_) &= g_(x) \\ u(y=y_) &= g_(x) \end If we call the two respective solutions ''x-partial solution'' and ''y-partial solution'', one of the most interesting consequences of the method is that the ''x-partial solution'' uses only the two boundary conditions (1-a) and the ''y-partial solution'' uses only the conditions (1-b). Thus, one of the two sets of boundary functions or is redundant, and this implies that a partial differential equation with boundary conditions on a rectangle cannot have arbitrary boundary conditions on the borders, since the conditions at ''x'' = ''x''1, ''x'' = ''x''2 must be consistent with those imposed at ''y'' = ''y''1 and ''y'' = ''y''2. An example to clarify this point is the solution of the Poisson problem with the following boundary conditions: : \begin u(x=0) &= f_(y) = 0 \\ u(x=x_) &= f_(y) = 0 \end By using Adomian's method and a symbolic processor (such as Mathematica or
Maple ''Acer'' () is a genus of trees and shrubs commonly known as maples. The genus is placed in the family Sapindaceae.Stevens, P. F. (2001 onwards). Angiosperm Phylogeny Website. Version 9, June 2008 nd more or less continuously updated since http ...
) it is easy to obtain the third order approximant to the solution. This approximant has an error lower than 5×10−16 in any point, as it can be proved by substitution in the initial problem and by displaying the absolute value of the residual obtained as a function of (''x'', ''y''). The solution at ''y'' = -0.25 and ''y'' = 0.25 is given by specific functions that in this case are: : g_(x) = 0.0520833\, x -0.347222\, x^ + 9.25186 \times 10^ x^ + 0.833333\, x^ -0.555556\, x^ and ''g''2(''x'') = ''g''1(''x'') respectively. If a (double) integration is now performed in the ''y''-direction using these two boundary functions the same solution will be obtained, which satisfy ''u''(''x''=0, ''y'') = 0 and ''u''(''x''=0.5, ''y'') = 0 and cannot satisfy any other condition on these borders. Some people are surprised by these results; it seems strange that not all initial-boundary conditions must be explicitly used to solve a differential system. However, it is a well established fact that any
elliptic equation An elliptic equation can mean: * The equation of an ellipse * An elliptic curve, describing the relationships between invariants of an ellipse * A differential equation with an elliptic operator * An elliptic partial differential equation Second ...
has one and only one solution for any functional conditions in the four sides of a rectangle provided there is no discontinuity on the edges. The cause of the misconception is that scientists and engineers normally think in a boundary condition in terms of weak convergence in a Hilbert space (the distance to the boundary function is small enough to practical purposes). In contrast, Cauchy problems impose a point-to-point convergence to a given boundary function and to all its derivatives (and this is a quite strong condition!). For the first ones, a function satisfies a boundary condition when the area (or another functional distance) between it and the true function imposed in the boundary is so small as desired; for the second ones, however, the function must tend to the true function imposed in any and every point of the interval. The commented Poisson problem does not have a solution for any functional boundary conditions ''f''1, ''f''2, ''g''1, ''g''2; however, given ''f''1, ''f''2 it is always possible to find boundary functions ''g''1*, ''g''2* so close to ''g''1, ''g''2 as desired (in the weak convergence meaning) for which the problem has solution. This property makes it possible to solve Poisson's and many other problems with arbitrary boundary conditions but never for analytic functions exactly specified on the boundaries. The reader can convince himself (herself) of the high sensitivity of PDE solutions to small changes in the boundary conditions by solving this problem integrating along the ''x''-direction, with boundary functions slightly different even though visually not distinguishable. For instance, the solution with the boundary conditions: : f_(y) = 0.00413682 - 0.0813801\, y^ + 0.260416\, y^ - 0.277778\, y^ at ''x'' = 0 and ''x'' = 0.5, and the solution with the boundary conditions: : \begin f_(y) = 0.00413683 &- 0.00040048\, y - 0.0813802\, y^ + 0.0101279\, y^ + 0.260417\, y^ \\ &- 0.0694455\, y^ - 0.277778\, y^ + 0.15873\, y^ + \cdots \end at ''x'' = 0 and ''x'' = 0.5, produce lateral functions with different sign convexity even though both functions are visually not distinguishable. Solutions of elliptic problems and other partial differential equations are highly sensitive to small changes in the boundary function imposed when only two sides are used. And this sensitivity is not easily compatible with models that are supposed to represent real systems, which are described by means of measurements containing experimental errors and are normally expressed as initial-boundary value problems in a Hilbert space.


Improvements to the decomposition method

At least three methods have been reported
/ref>
/ref> OI: 10.1108/03684920310463939

/ref> to obtain the boundary functions ''g''1*, ''g''2* that are compatible with any lateral set of conditions imposed. This makes it possible to find the analytical solution of any PDE boundary problem on a closed rectangle with the required accuracy, so allowing to solve a wide range of problems that the standard Adomian's method was not able to address. The first one perturbs the two boundary functions imposed at ''x'' = 0 and ''x'' = ''x''1 (condition 1-a) with a ''N''th-order polynomial in ''y'': ''p''1, ''p''2 in such a way that: ''f''1' = ''f''1 + ''p''1, ''f''2' = ''f''2 + ''p''2, where the norm of the two perturbation functions are smaller than the accuracy needed at the boundaries. These ''p''1, ''p''2 depend on a set of polynomial coefficients ''c''''i'', ''i'' = 1, ..., ''N''. Then, the Adomian method is applied and functions are obtained at the four boundaries which depend on the set of ''c''''i'', ''i'' = 1, ..., ''N''. Finally, a boundary function ''F''(''c''1, ''c''2, ..., ''c''''N'') is defined as the sum of these four functions, and the distance between ''F''(''c''1, ''c''2, ..., ''c''''N'') and the real boundary functions ((1-a) and (1-b)) is minimized. The problem has been reduced, in this way, to the global minimization of the function ''F''(''c''1, ''c''2, ..., ''c''''N'') which has a global minimum for some combination of the parameters ''c''''i'', ''i'' = 1, ..., ''N''. This minimum may be found by means of a genetic algorithm or by using some other optimization method, as the one proposed by Cherruault (1999). A second method to obtain analytic approximants of initial-boundary problems is to combine Adomian decomposition with spectral methods. Finally, the third method proposed by García-Olivares is based on imposing analytic solutions at the four boundaries, but modifying the original differential operator in such a way that it is different from the original one only in a narrow region close to the boundaries, and it forces the solution to satisfy exactly analytic conditions at the four boundaries.


Integral Equations

The Adomian decomposition method may also be applied to linear and nonlinear
integral equation In mathematics, integral equations are equations in which an unknown function appears under an integral sign. In mathematical notation, integral equations may thus be expressed as being of the form: f(x_1,x_2,x_3,...,x_n ; u(x_1,x_2,x_3,...,x_n) ...
s to obtain solutions. This corresponds to the fact that many differential equation can be converted into integral equations.


Adomian Decomposition Method

The Adomian decomposition method for nonhomogenous
Fredholm integral equation In mathematics, the Fredholm integral equation is an integral equation whose solution gives rise to Fredholm theory, the study of Fredholm kernels and Fredholm operators. The integral equation was studied by Ivar Fredholm. A useful method to so ...
of the second kind goes as follows: Given an integral equation of the form: : u(x) = f(x) + \lambda \int_^b K(x,t) u(t) dt We assume we may express the solution in series form: : u(x) = \sum_^\infty u_n(x) Plugging the series form into the integral equation then yields: : \sum_^\infty u_n(x) = f(x) + \lambda \int_^b K(x,t) (\sum_^\infty u_n(t)) dt Assuming that the sum converges absolutely to u(x) we may integerchange the sum and integral as follows: : \sum_^\infty u_n(x)= f(x) + \lambda \int_^b \sum_^\infty K(x,t) u_n(t)dt : \sum_^\infty u_n(x) = f(x) + \lambda \sum_^\infty \int_^b K(x,t) u_n(t)dt Expanding the sum on both sides yields: : u_0(x)+u_1(x)+u_2(x)+... = f(x) + \lambda \int_^b K(x,t)u_0(t)dt+ \lambda \int_^b K(x,t)u_1(t)dt + \lambda \int_^b K(x,t)u_2(t)dt+ ... Hence we may associate each u_i(x) in the following recurrent manner: : u_0(x) = f(x) : u_i(x)=\lambda \int_a^b K(x,t)u_dt, \, \, \, \, \, \, \, \, i \geq 1 which gives us the solution u(x) in the solution form above.


Example

Given the Fredholm integral equation: : u(x) = \cos(x) +2x + \int_^\pi xt \cdot u(t) dt Since f(x) = \cos(x)+2x , we can set: : u_0(x) = \cos(x)+2x : u_1(x)=\int_0^\pi xt\cdot u_0 dt =\int_0^\pi xt\cdot (cosx+2x) dt = (-2+\frac)x : u_2(x)=\int_0^\pi xt\cdot u_1 dt =\int_0^\pi xt\cdot (-2+\frac)t \,dt = (\frac+ \frac)x :... Hence the solution u(x) may be written as: : u(x)= \cos(x)+2x+ (-2+\frac)x + (\frac+ \frac{9})x + ... Since this is a telescoping series, we can see that every terms after cos(x) cancels and may be regarded as "noise", Thus, u(x) becomes: : u(x) = \cos(x)


Gallery


See also

*
Order of approximation In science, engineering, and other quantitative disciplines, order of approximation refers to formal or informal expressions for how accurate an approximation is. Usage in science and engineering In formal expressions, the ordinal number used b ...


References

Differential equations