Parker–Sochacki Method
   HOME

TheInfoList



OR:

In
mathematics Mathematics is an area of knowledge that includes the topics of numbers, formulas and related structures, shapes and the spaces in which they are contained, and quantities and their changes. These topics are represented in modern mathematics ...
, the Parker–Sochacki method is an
algorithm In mathematics and computer science, an algorithm () is a finite sequence of rigorous instructions, typically used to solve a class of specific Computational problem, problems or to perform a computation. Algorithms are used as specificat ...
for solving systems of ordinary
differential equation In mathematics, a differential equation is an equation that relates one or more unknown functions and their derivatives. In applications, the functions generally represent physical quantities, the derivatives represent their rates of change, an ...
s (ODEs), developed by G. Edgar Parker and
James Sochacki James is a common English language surname and given name: *James (name), the typically masculine first name James * James (surname), various people with the last name James James or James City may also refer to: People * King James (disambiguat ...
, of the
James Madison University James Madison University (JMU, Madison, or James Madison) is a public research university in Harrisonburg, Virginia. Founded in 1908 as the State Normal and Industrial School for Women at Harrisonburg, the institution was renamed Madison Coll ...
Mathematics Department. The method produces
Maclaurin series Maclaurin or MacLaurin is a surname. Notable people with the surname include: * Colin Maclaurin (1698–1746), Scottish mathematician * Normand MacLaurin (1835–1914), Australian politician and university administrator * Henry Normand MacLaurin ( ...
solutions to systems of differential equations, with the coefficients in either algebraic or numerical form.


Summary

The Parker–Sochacki method rests on two simple observations: * If a set of ODEs has a particular form, then the Picard method can be used to find their solution in the form of a
power series In mathematics, a power series (in one variable) is an infinite series of the form \sum_^\infty a_n \left(x - c\right)^n = a_0 + a_1 (x - c) + a_2 (x - c)^2 + \dots where ''an'' represents the coefficient of the ''n''th term and ''c'' is a const ...
. * If the ODEs do not have the required form, it is nearly always possible to find an expanded set of equations that do have the required form, such that a subset of the solution is a solution of the original ODEs. Several coefficients of the power series are calculated in turn, a time step is chosen, the series is evaluated at that time, and the process repeats. The end result is a high order piecewise solution to the original ODE problem. The order of the solution desired is an adjustable variable in the program that can change between steps. The order of the solution is only limited by the floating point representation on the machine running the program. And in some cases can be either extended by using arbitrary precision floating point numbers, or for special cases by finding solution with only integer or rational coefficients.


Advantages

The method requires only addition, subtraction, and multiplication, making it very convenient for high-speed computation. (The only divisions are inverses of small integers, which can be precomputed.) Use of a high order—calculating many coefficients of the power series—is convenient. (Typically a higher order permits a longer time step without loss of accuracy, which improves efficiency.) The order and step size can be easily changed from one step to the next. It is possible to calculate a guaranteed error bound on the solution. Arbitrary precision floating point libraries allow this method to compute arbitrarily accurate solutions. With the Parker–Sochacki method, information between integration steps is developed at high order. As the Parker–Sochacki method integrates, the program can be designed to save the power series coefficients that provide a smooth solution between points in time. The coefficients can be saved and used so that polynomial evaluation provides the high order solution between steps. With most other classical integration methods, one would have to resort to interpolation to get information between integration steps, leading to an increase of error. There is an A-priori error bound for a single step with the Parker–Sochacki method. This allows a Parker–Sochacki program to calculate the step size that guarantees that the error is below any non-zero given tolerance. Using this calculated step size with an error tolerance of less than half of the machine epsilon yields a symplectic integration.


Disadvantages

Most methods for numerically solving ODEs require only the evaluation of derivatives for chosen values of the variables, so systems like MATLAB include implementations of several methods all sharing the same calling sequence. Users can try different methods by simply changing the name of the function called. The Parker–Sochacki method requires more work to put the equations into the proper form, and cannot use the same calling sequence.


References


External links

* . A thorough explanation of the paradigm and application of the Parker–Sochacki method * . A demonstration of the theory and usage of the Parker–Sochacki method, including a solution for the classical Newtonian ''N''-body problem with mutual gravitational attraction. * . A collection of papers and some Matlab code. {{DEFAULTSORT:Parker-Sochacki method Mathematical analysis