Finite Difference Method
   HOME

TheInfoList



OR:

In
numerical analysis Numerical analysis is the study of algorithms that use numerical approximation (as opposed to symbolic computation, symbolic manipulations) for the problems of mathematical analysis (as distinguished from discrete mathematics). It is the study of ...
, finite-difference methods (FDM) are a class of numerical techniques for solving
differential equations In mathematics, a differential equation is an equation that relates one or more unknown functions and their derivatives. In applications, the functions generally represent physical quantities, the derivatives represent their rates of change, an ...
by approximating
derivatives The derivative of a function is the rate of change of the function's output relative to its input value. Derivative may also refer to: In mathematics and economics * Brzozowski derivative in the theory of formal languages * Formal derivative, an ...
with
finite differences A finite difference is a mathematical expression of the form . If a finite difference is divided by , one gets a difference quotient. The approximation of derivatives by finite differences plays a central role in finite difference methods for the ...
. Both the spatial domain and time interval (if applicable) are discretized, or broken into a finite number of steps, and the value of the solution at these discrete points is approximated by solving algebraic equations containing finite differences and values from nearby points. Finite difference methods convert
ordinary differential equations In mathematics, an ordinary differential equation (ODE) is a differential equation whose unknown(s) consists of one (or more) function(s) of one variable and involves the derivatives of those functions. The term ''ordinary'' is used in contrast w ...
(ODE) or
partial differential equations In mathematics, a partial differential equation (PDE) is an equation which imposes relations between the various partial derivatives of a multivariable function. The function is often thought of as an "unknown" to be solved for, similarly to ...
(PDE), which may be
nonlinear In mathematics and science, a nonlinear system is a system in which the change of the output is not proportional to the change of the input. Nonlinear problems are of interest to engineers, biologists, physicists, mathematicians, and many othe ...
, into a system of linear equations that can be solved by matrix algebra techniques. Modern computers can perform these
linear algebra Linear algebra is the branch of mathematics concerning linear equations such as: :a_1x_1+\cdots +a_nx_n=b, linear maps such as: :(x_1, \ldots, x_n) \mapsto a_1x_1+\cdots +a_nx_n, and their representations in vector spaces and through matrices. ...
computations efficiently which, along with their relative ease of implementation, has led to the widespread use of FDM in modern numerical analysis. Today, FDM are one of the most common approaches to the numerical solution of PDE, along with
finite element methods The finite element method (FEM) is a popular method for numerically solving differential equations arising in engineering and mathematical modeling. Typical problem areas of interest include the traditional fields of structural analysis, heat ...
.


Derivation from Taylor's polynomial

First, assuming the function whose derivatives are to be approximated is properly behaved, by
Taylor's theorem In calculus, Taylor's theorem gives an approximation of a ''k''-times differentiable function around a given point by a polynomial of degree ''k'', called the ''k''th-order Taylor polynomial. For a smooth function, the Taylor polynomial is the t ...
, we can create a
Taylor series In mathematics, the Taylor series or Taylor expansion of a function is an infinite sum of terms that are expressed in terms of the function's derivatives at a single point. For most common functions, the function and the sum of its Taylor serie ...
expansion :f(x_0 + h) = f(x_0) + \frach + \frach^2 + \cdots + \frach^n + R_n(x), where ''n''! denotes the
factorial In mathematics, the factorial of a non-negative denoted is the product of all positive integers less than or equal The factorial also equals the product of n with the next smaller factorial: \begin n! &= n \times (n-1) \times (n-2) \t ...
of ''n'', and ''R''''n''(''x'') is a remainder term, denoting the difference between the Taylor polynomial of degree ''n'' and the original function. We will derive an approximation for the first derivative of the function "f" by first truncating the Taylor polynomial: :f(x_0 + h) = f(x_0) + f'(x_0)h + R_1(x), Setting, x0=a we have, :f(a+h) = f(a) + f'(a)h + R_1(x), Dividing across by ''h'' gives: : = + f'(a)+ Solving for f'(a): :f'(a) = - Assuming that R_1(x) is sufficiently small, the approximation of the first derivative of "f" is: :f'(a)\approx . This is, not coincidentally, similar to the definition of derivative, which is given as: :f'(a)=\lim_\frac. except for the limit towards zero (the method is named after this).


Accuracy and order

The error in a method's solution is defined as the difference between the approximation and the exact analytical solution. The two sources of error in finite difference methods are
round-off error A roundoff error, also called rounding error, is the difference between the result produced by a given algorithm using exact arithmetic and the result produced by the same algorithm using finite-precision, rounded arithmetic. Rounding errors are d ...
, the loss of precision due to computer rounding of decimal quantities, and
truncation error In numerical analysis and scientific computing, truncation error is an error caused by approximating a mathematical process. Examples Infinite series A summation series for e^x is given by an infinite series such as e^x=1+ x+ \frac + \frac ...
or
discretization error In numerical analysis, computational physics, and simulation, discretization error is the error resulting from the fact that a function of a continuous variable is represented in the computer by a finite number of evaluations, for example, on a ...
, the difference between the exact solution of the original differential equation and the exact quantity assuming perfect arithmetic (that is, assuming no round-off). To use a finite difference method to approximate the solution to a problem, one must first discretize the problem's domain. This is usually done by dividing the domain into a uniform grid (see image to the right). This means that finite-difference methods produce sets of discrete numerical approximations to the derivative, often in a "time-stepping" manner. An expression of general interest is the
local truncation error Truncation errors in numerical integration are of two kinds: * ''local truncation errors'' – the error caused by one iteration, and * ''global truncation errors'' – the cumulative error caused by many iterations. Definitions Suppose we have ...
of a method. Typically expressed using
Big-O notation Big ''O'' notation is a mathematical notation that describes the limiting behavior of a function when the argument tends towards a particular value or infinity. Big O is a member of a family of notations invented by Paul Bachmann, Edmund Land ...
, local truncation error refers to the error from a single application of a method. That is, it is the quantity f'(x_i) - f'_i if f'(x_i) refers to the exact value and f'_i to the numerical approximation. The remainder term of a Taylor polynomial is convenient for analyzing the local truncation error. Using the Lagrange form of the remainder from the Taylor polynomial for f(x_0 + h), which is R_n(x_0 + h) = \frac (h)^ , where x_0 < \xi < x_0 + h, the dominant term of the local truncation error can be discovered. For example, again using the forward-difference formula for the first derivative, knowing that f(x_i)=f(x_0+i h), : f(x_0 + i h) = f(x_0) + f'(x_0)i h + \frac (i h)^, and with some algebraic manipulation, this leads to : \frac = f'(x_0) + \frac i h, and further noting that the quantity on the left is the approximation from the finite difference method and that the quantity on the right is the exact quantity of interest plus a remainder, clearly that remainder is the local truncation error. A final expression of this example and its order is: : \frac = f'(x_0) + O(h). This means that, in this case, the local truncation error is proportional to the step sizes. The quality and duration of simulated FDM solution depends on the discretization equation selection and the step sizes (time and space steps). The data quality and simulation duration increase significantly with smaller step size. Therefore, a reasonable balance between data quality and simulation duration is necessary for practical usage. Large time steps are useful for increasing simulation speed in practice. However, time steps which are too large may create instabilities and affect the data quality. The
von Neumann Von Neumann may refer to: * John von Neumann (1903–1957), a Hungarian American mathematician * Von Neumann family * Von Neumann (surname), a German surname * Von Neumann (crater), a lunar impact crater See also * Von Neumann algebra * Von Ne ...
and Courant-Friedrichs-Lewy criteria are often evaluated to determine the numerical model stability.


Example: ordinary differential equation

For example, consider the ordinary differential equation : u'(x) = 3u(x) + 2. \, The
Euler method In mathematics and computational science, the Euler method (also called forward Euler method) is a first-order numerical procedure for solving ordinary differential equations (ODEs) with a given initial value. It is the most basic explicit m ...
for solving this equation uses the finite difference quotient :\frac \approx u'(x) to approximate the differential equation by first substituting it for u'(x) then applying a little algebra (multiplying both sides by h, and then adding u(x) to both sides) to get : u(x+h) = u(x) + h(3u(x)+2). \, The last equation is a finite-difference equation, and solving this equation gives an approximate solution to the differential equation.


Example: The heat equation

Consider the normalized
heat equation In mathematics and physics, the heat equation is a certain partial differential equation. Solutions of the heat equation are sometimes known as caloric functions. The theory of the heat equation was first developed by Joseph Fourier in 1822 for t ...
in one dimension, with homogeneous
Dirichlet boundary condition In the mathematical study of differential equations, the Dirichlet (or first-type) boundary condition is a type of boundary condition, named after Peter Gustav Lejeune Dirichlet (1805–1859). When imposed on an ordinary or a partial differential ...
s : U_t=U_ \, : U(0,t)=U(1,t)=0 \, (boundary condition) : U(x,0) =U_0(x) \, (initial condition) One way to numerically solve this equation is to approximate all the derivatives by finite differences. We partition the domain in space using a mesh x_0, ..., x_J and in time using a mesh t_0, ...., t_N . We assume a uniform partition both in space and in time, so the difference between two consecutive space points will be ''h'' and between two consecutive time points will be ''k''. The points : u(x_j,t_n) = u_^n will represent the numerical approximation of u(x_j, t_n).


Explicit method

Using a
forward difference A finite difference is a mathematical expression of the form . If a finite difference is divided by , one gets a difference quotient. The approximation of derivatives by finite differences plays a central role in finite difference methods for the ...
at time t_n and a second-order
central difference A finite difference is a mathematical expression of the form . If a finite difference is divided by , one gets a difference quotient. The approximation of derivatives by finite differences plays a central role in finite difference methods for the ...
for the space derivative at position x_j ( FTCS) we get the recurrence equation: : \frac = \frac. \, This is an
explicit method Explicit and implicit methods are approaches used in numerical analysis for obtaining numerical approximations to the solutions of time-dependent ordinary and partial differential equations, as is required in computer simulations of physical pro ...
for solving the one-dimensional
heat equation In mathematics and physics, the heat equation is a certain partial differential equation. Solutions of the heat equation are sometimes known as caloric functions. The theory of the heat equation was first developed by Joseph Fourier in 1822 for t ...
. We can obtain u_j^ from the other values this way: : u_^ = (1-2r)u_^ + ru_^ + ru_^ where r=\alpha k/h^2. So, with this recurrence relation, and knowing the values at time ''n'', one can obtain the corresponding values at time ''n''+1. u_0^n and u_J^n must be replaced by the boundary conditions, in this example they are both 0. This explicit method is known to be
numerically stable In the mathematical subfield of numerical analysis, numerical stability is a generally desirable property of numerical algorithms. The precise definition of stability depends on the context. One is numerical linear algebra and the other is algorit ...
and convergent whenever r\le 1/2 . The numerical errors are proportional to the time step and the square of the space step: : \Delta u = O(k)+O(h^2) \,


Implicit method

If we use the
backward difference A finite difference is a mathematical expression of the form . If a finite difference is divided by , one gets a difference quotient. The approximation of derivatives by finite differences plays a central role in finite difference methods for the ...
at time t_ and a second-order central difference for the space derivative at position x_j (The Backward Time, Centered Space Method "BTCS") we get the recurrence equation: : \frac =\frac. \, This is an
implicit method Explicit and implicit methods are approaches used in numerical analysis for obtaining numerical approximations to the solutions of time-dependent ordinary and partial differential equations, as is required in computer simulations of physical p ...
for solving the one-dimensional
heat equation In mathematics and physics, the heat equation is a certain partial differential equation. Solutions of the heat equation are sometimes known as caloric functions. The theory of the heat equation was first developed by Joseph Fourier in 1822 for t ...
. We can obtain u_j^ from solving a system of linear equations: : (1+2r)u_j^ - ru_^ - ru_^= u_^ The scheme is always
numerically stable In the mathematical subfield of numerical analysis, numerical stability is a generally desirable property of numerical algorithms. The precise definition of stability depends on the context. One is numerical linear algebra and the other is algorit ...
and convergent but usually more numerically intensive than the explicit method as it requires solving a system of numerical equations on each time step. The errors are linear over the time step and quadratic over the space step: : \Delta u = O(k)+O(h^2). \,


Crank–Nicolson method

Finally if we use the central difference at time t_ and a second-order central difference for the space derivative at position x_j ("CTCS") we get the recurrence equation: : \frac = \frac \left(\frac+\frac\right).\, This formula is known as the Crank–Nicolson method. We can obtain u_j^ from solving a system of linear equations: : (2+2r)u_j^ - ru_^ - ru_^= (2-2r)u_j^n + ru_^n + ru_^n The scheme is always
numerically stable In the mathematical subfield of numerical analysis, numerical stability is a generally desirable property of numerical algorithms. The precise definition of stability depends on the context. One is numerical linear algebra and the other is algorit ...
and convergent but usually more numerically intensive as it requires solving a system of numerical equations on each time step. The errors are quadratic over both the time step and the space step: : \Delta u = O(k^2)+O(h^2). \,


Comparison

To summarize, usually the Crank–Nicolson scheme is the most accurate scheme for small time steps. For larger time steps, the implicit scheme works better since it is less computationally demanding. The explicit scheme is the least accurate and can be unstable, but is also the easiest to implement and the least numerically intensive. Here is an example. The figures below present the solutions given by the above methods to approximate the heat equation : U_t = \alpha U_, \quad \alpha = \frac, with the boundary condition : U(0, t) = U(1, t) = 0. The exact solution is :U(x, t) = \frace^\sin(\pi x).


Example: The Laplace operator

The (continuous) Laplace operator in n -dimensions is given by \Delta u(x) = \sum_^n \partial_i^2 u(x) . The discrete Laplace operator \Delta_h u depends on the dimension n . In 1D the Laplace operator is approximated as : \Delta u(x) = u''(x) \approx \frac =: \Delta_h u(x) \,. This approximation is usually expressed via the following
stencil Stencilling produces an image or pattern on a surface, by applying pigment to a surface through an intermediate object, with designed holes in the intermediate object, to create a pattern or image on a surface, by allowing the pigment to reach ...
: \Delta_h = \frac \begin 1 & -2 & 1 \end and which represents a symmetric, tridiagonal matrix. For an equidistant grid one gets a
Toeplitz matrix In linear algebra, a Toeplitz matrix or diagonal-constant matrix, named after Otto Toeplitz, is a matrix in which each descending diagonal from left to right is constant. For instance, the following matrix is a Toeplitz matrix: :\qquad\begin a & b ...
. The 2D case shows all the characteristics of the more general nD case. Each second partial derivative needs to be approximated similar to the 1D case : \begin \Delta u(x,y) &= u_(x,y)+u_(x,y) \\ &\approx \frac + \frac \\ &= \frac \\ &=: \Delta_h u(x, y) \,, \end which is usually given by the following
stencil Stencilling produces an image or pattern on a surface, by applying pigment to a surface through an intermediate object, with designed holes in the intermediate object, to create a pattern or image on a surface, by allowing the pigment to reach ...
: \Delta_h = \frac \begin & 1 \\ 1 & -4 & 1 \\ & 1 \end \,.


Consistency

Consistency of the above-mentioned approximation can be shown for highly regular functions, such as u \in C^4(\Omega) . The statement is : \Delta u - \Delta_h u = \mathcal(h^2) \,. To prove this, one needs to substitute
Taylor Series In mathematics, the Taylor series or Taylor expansion of a function is an infinite sum of terms that are expressed in terms of the function's derivatives at a single point. For most common functions, the function and the sum of its Taylor serie ...
expansions up to order 3 into the discrete Laplace operator.


Properties


Subharmonic

Similar to continuous subharmonic functions one can define ''subharmonic functions'' for finite-difference approximations u_h : -\Delta_h u_h \leq 0 \,.


Mean value

One can define a general
stencil Stencilling produces an image or pattern on a surface, by applying pigment to a surface through an intermediate object, with designed holes in the intermediate object, to create a pattern or image on a surface, by allowing the pigment to reach ...
of ''positive type'' via : \begin & \alpha_N \\ \alpha_W & -\alpha_C & \alpha_E \\ & \alpha_S \end \,, \quad \alpha_i >0\,, \quad \alpha_C = \sum_ \alpha_i \,. If u_h is (discrete) subharmonic then the following'' mean value property'' holds : u_h(x_C) \leq \frac \,, where the approximation is evaluated on points of the grid, and the stencil is assumed to be of positive type. A similar mean value property also holds for the continuous case.


Maximum principle

For a (discrete) subharmonic function u_h the following holds : \max_ u_h \leq \max_ u_h \,, where \Omega_h, \partial\Omega_h are discretizations of the continuous domain \Omega , respectively the boundary \partial \Omega . A similar
maximum principle In the mathematical fields of partial differential equations and geometric analysis, the maximum principle is any of a collection of results and techniques of fundamental importance in the study of elliptic and parabolic differential equations. ...
also holds for the continuous case.


The SBP-SAT method

The SBP-SAT (''
summation by parts In mathematics, summation by parts transforms the summation of products of sequences into other summations, often simplifying the computation or (especially) estimation of certain types of sums. It is also called Abel's lemma or Abel transformati ...
- simultaneous approximation term'') method is a stable and accurate technique for discretizing and imposing boundary conditions of a well-posed
partial differential equation In mathematics, a partial differential equation (PDE) is an equation which imposes relations between the various partial derivatives of a Multivariable calculus, multivariable function. The function is often thought of as an "unknown" to be sol ...
using high order finite differences. The method is based on finite differences where the differentiation operators exhibit summation-by-parts properties. Typically, these operators consist of differentiation matrices with central difference stencils in the interior with carefully chosen one-sided boundary stencils designed to mimic integration-by-parts in the discrete setting. Using the SAT technique, the boundary conditions of the PDE are imposed weakly, where the boundary values are "pulled" towards the desired conditions rather than exactly fulfilled. If the tuning parameters (inherent to the SAT technique) are chosen properly, the resulting system of ODE's will exhibit similar energy behavior as the continuous PDE, i.e. the system has no non-physical energy growth. This guarantees stability if an integration scheme with a stability region that includes parts of the imaginary axis, such as the fourth order Runge-Kutta method, is used. This makes the SAT technique an attractive method of imposing boundary conditions for higher order finite difference methods, in contrast to for example the injection method, which typically will not be stable if high order differentiation operators are used.


See also

*
Finite element method The finite element method (FEM) is a popular method for numerically solving differential equations arising in engineering and mathematical modeling. Typical problem areas of interest include the traditional fields of structural analysis, heat ...
*
Finite difference A finite difference is a mathematical expression of the form . If a finite difference is divided by , one gets a difference quotient. The approximation of derivatives by finite differences plays a central role in finite difference methods for t ...
* Finite difference time domain * Infinite difference method * Stencil (numerical analysis) *
Finite difference coefficients In mathematics, to approximate a derivative to an arbitrary order of accuracy, it is possible to use the finite difference. A finite difference can be central, forward or backward. Central finite difference This table contains the coefficients o ...
*
Five-point stencil In numerical analysis, given a square grid in one or two dimensions, the five-point stencil of a point in the grid is a stencil made up of the point itself together with its four "neighbors". It is used to write finite difference approximations to ...
* Lax–Richtmyer theorem *
Finite difference methods for option pricing Finite difference methods for option pricing are numerical methods used in mathematical finance for the valuation of options. Finite difference methods were first applied to option pricing by Eduardo Schwartz in 1977. In general, finite differ ...
*
Upwind differencing scheme for convection The upwind differencing scheme is a method used in numerical methods in computational fluid dynamics for convection–diffusion problems. This scheme is specific for Peclet number greater than 2 or less than −2 Description By taking in ...
*
Central differencing scheme In applied mathematics, the central differencing scheme is a finite difference method that optimizes the approximation for the differential operator in the central node of the considered patch and provides numerical solutions to differential equat ...
*
Discrete Poisson equation In mathematics, the discrete Poisson equation is the finite difference analog of the Poisson equation. In it, the discrete Laplace operator takes the place of the Laplace operator. The discrete Poisson equation is frequently used in numerical an ...
*
Discrete Laplace operator In mathematics, the discrete Laplace operator is an analog of the continuous Laplace operator, defined so that it has meaning on a graph or a discrete grid. For the case of a finite-dimensional graph (having a finite number of edges and vertice ...


References


Further reading

* K.W. Morton and D.F. Mayers, ''Numerical Solution of Partial Differential Equations, An Introduction''. Cambridge University Press, 2005. * Autar Kaw and E. Eric Kalu, ''Numerical Methods with Applications'', (2008

Contains a brief, engineering-oriented introduction to FDM (for ODEs) i
Chapter 08.07
* * * . * Randall J. LeVeque,
Finite Difference Methods for Ordinary and Partial Differential Equations
', SIAM, 2007. * Sergey Lemeshevsky, Piotr Matus, Dmitriy Poliakov(Eds): "Exact Finite-Difference Schemes", De Gruyter (2016). DOI: https://doi.org/10.1515/9783110491326 . {{DEFAULTSORT:Finite Difference Method Finite differences Numerical differential equations