In
mathematics
Mathematics is an area of knowledge that includes the topics of numbers, formulas and related structures, shapes and the spaces in which they are contained, and quantities and their changes. These topics are represented in modern mathematics ...
, separation of variables (also known as the Fourier method) is any of several methods for solving
ordinary and
partial differential equation
In mathematics, a partial differential equation (PDE) is an equation which imposes relations between the various partial derivatives of a Multivariable calculus, multivariable function.
The function is often thought of as an "unknown" to be sol ...
s, in which algebra allows one to rewrite an equation so that each of two variables occurs on a different side of the equation.
Ordinary differential equations (ODE)
Suppose a differential equation can be written in the form
:
which we can write more simply by letting
:
:
As long as ''h''(''y'') ≠ 0, we can rearrange terms to obtain:
:
so that the two variables ''x'' and ''y'' have been separated. ''dx'' (and ''dy'') can be viewed, at a simple level, as just a convenient notation, which provides a handy mnemonic aid for assisting with manipulations. A formal definition of ''dx'' as a
differential (infinitesimal)
In mathematics, differential refers to several related notions derived from the early days of calculus, put on a rigorous footing, such as infinitesimal differences and the derivatives of functions.
The term is used in various branches of mathema ...
is somewhat advanced.
Alternative notation
Those who dislike
Leibniz's notation
In calculus, Leibniz's notation, named in honor of the 17th-century German philosopher and mathematician Gottfried Wilhelm Leibniz, uses the symbols and to represent infinitely small (or infinitesimal) increments of and , respectively, just a ...
may prefer to write this as
:
but that fails to make it quite as obvious why this is called "separation of variables". Integrating both sides of the equation with respect to
, we have
or equivalently,
:
because of the
substitution rule for integrals.
If one can evaluate the two integrals, one can find a solution to the differential equation. Observe that this process effectively allows us to treat the
derivative
In mathematics, the derivative of a function of a real variable measures the sensitivity to change of the function value (output value) with respect to a change in its argument (input value). Derivatives are a fundamental tool of calculus. F ...
as a fraction which can be separated. This allows us to solve separable differential equations more conveniently, as demonstrated in the example below.
(Note that we do not need to use two
constants of integration, in equation () as in
:
because a single constant
is equivalent.)
Example
Population growth is often modeled by the differential equation
:
where
is the population with respect to time
,
is the rate of growth, and
is the
carrying capacity
The carrying capacity of an environment is the maximum population size of a biological species that can be sustained by that specific environment, given the food, habitat, water, and other resources available. The carrying capacity is defined as t ...
of the environment.
Separation of variables may be used to solve this differential equation.
:
To evaluate the integral on the left side, we simplify the fraction
:
and then, we decompose the fraction into partial fractions
:
Thus we have
:
Let
.
:
Therefore, the solution to the logistic equation is
:
To find
, let
and
. Then we have
:
Noting that
, and solving for ''A'' we get
:
Generalization of separable ODEs to the nth order
Much like one can speak of a separable first-order ODE, one can speak of a separable second-order, third-order or ''n''th-order ODE. Consider the separable first-order ODE:
:
The derivative can alternatively be written the following way to underscore that it is an operator working on the unknown function, ''y'':
:
Thus, when one separates variables for first-order equations, one in fact moves the ''dx'' denominator of the operator to the side with the ''x'' variable, and the ''d''(''y'') is left on the side with the ''y'' variable. The second-derivative operator, by analogy, breaks down as follows:
:
The third-, fourth- and ''n''th-derivative operators break down in the same way. Thus, much like a first-order separable ODE is reducible to the form
:
a separable second-order ODE is reducible to the form
:
and an nth-order separable ODE is reducible to
:
Example
Consider the simple nonlinear second-order differential equation:
This equation is an equation only of ''y
'''' and ''y'
'', meaning it is reducible to the general form described above and is, therefore, separable. Since it is a second-order separable equation, collect all ''x'' variables on one side and all ''y'
'' variables on the other to get:
Now, integrate the right side with respect to ''x'' and the left with respect to ''y:
This gives
which simplifies to:
This is now a simple integral problem that gives the final answer:
Partial differential equations
The method of separation of variables is also used to solve a wide range of linear partial differential equations with boundary and initial conditions, such as the
heat equation
In mathematics and physics, the heat equation is a certain partial differential equation. Solutions of the heat equation are sometimes known as caloric functions. The theory of the heat equation was first developed by Joseph Fourier in 1822 for t ...
,
wave equation
The (two-way) wave equation is a second-order linear partial differential equation for the description of waves or standing wave fields — as they occur in classical physics — such as mechanical waves (e.g. water waves, sound waves and s ...
,
Laplace equation
In mathematics and physics, Laplace's equation is a second-order partial differential equation named after Pierre-Simon Laplace, who first studied its properties. This is often written as
\nabla^2\! f = 0 or \Delta f = 0,
where \Delta = \nab ...
,
Helmholtz equation
In mathematics, the eigenvalue problem for the Laplace operator is known as the Helmholtz equation. It corresponds to the linear partial differential equation
\nabla^2 f = -k^2 f,
where is the Laplace operator (or "Laplacian"), is the eigenv ...
and
biharmonic equation
In mathematics, the biharmonic equation is a fourth-order partial differential equation which arises in areas of continuum mechanics, including linear elasticity theory and the solution of Stokes flows. Specifically, it is used in the modeling ...
.
The analytical method of separation of variables for solving partial differential equations has also been generalized into a computational method of decomposition in invariant structures that can be used to solve systems of partial differential equations.
Example: homogeneous case
Consider the one-dimensional
heat equation
In mathematics and physics, the heat equation is a certain partial differential equation. Solutions of the heat equation are sometimes known as caloric functions. The theory of the heat equation was first developed by Joseph Fourier in 1822 for t ...
. The equation is
The variable ''u'' denotes temperature. The boundary condition is homogeneous, that is
Let us attempt to find a solution which is not identically zero satisfying the boundary conditions but with the following property: ''u'' is a product in which the dependence of ''u'' on ''x'', ''t'' is separated, that is:
Substituting ''u'' back into equation and using the
product rule
In calculus, the product rule (or Leibniz rule or Leibniz product rule) is a formula used to find the derivatives of products of two or more functions. For two functions, it may be stated in Lagrange's notation as (u \cdot v)' = u ' \cdot v + ...
,
Since the right hand side depends only on ''x'' and the left hand side only on ''t'', both sides are equal to some constant value −''λ''. Thus:
and
−''λ'' here is the
eigenvalue
In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denoted b ...
for both differential operators, and ''T''(''t'') and ''X''(''x'') are corresponding
eigenfunction
In mathematics, an eigenfunction of a linear operator ''D'' defined on some function space is any non-zero function f in that space that, when acted upon by ''D'', is only multiplied by some scaling factor called an eigenvalue. As an equation, th ...
s.
We will now show that solutions for ''X''(''x'') for values of ''λ'' ≤ 0 cannot occur:
Suppose that ''λ'' < 0. Then there exist real numbers ''B'', ''C'' such that
:
From we get
and therefore ''B'' = 0 = ''C'' which implies ''u'' is identically 0.
Suppose that ''λ'' = 0. Then there exist real numbers ''B'', ''C'' such that
:
From we conclude in the same manner as in 1 that ''u'' is identically 0.
Therefore, it must be the case that ''λ'' > 0. Then there exist real numbers ''A'', ''B'', ''C'' such that
:
and
:
From we get ''C'' = 0 and that for some positive integer ''n'',
:
This solves the heat equation in the special case that the dependence of ''u'' has the special form of .
In general, the sum of solutions to which satisfy the boundary conditions also satisfies and . Hence a complete solution can be given as
:
where ''D''
''n'' are coefficients determined by initial condition.
Given the initial condition
:
we can get
:
This is the
sine series expansion of ''f''(''x''). Multiplying both sides with
and integrating over results in
:
This method requires that the eigenfunctions ''X'', here
, are
orthogonal
In mathematics, orthogonality is the generalization of the geometric notion of ''perpendicularity''.
By extension, orthogonality is also used to refer to the separation of specific features of a system. The term also has specialized meanings in ...
and
complete
Complete may refer to:
Logic
* Completeness (logic)
* Completeness of a theory, the property of a theory that every formula in the theory's language or its negation is provable
Mathematics
* The completeness of the real numbers, which implies t ...
. In general this is guaranteed by
Sturm–Liouville theory In mathematics and its applications, classical Sturm–Liouville theory is the theory of ''real'' second-order ''linear'' ordinary differential equations of the form:
for given coefficient functions , , and , an unknown function ''y = y''(''x'') ...
.
Example: nonhomogeneous case
Suppose the equation is nonhomogeneous,
with the boundary condition the same as .
Expand ''h''(''x,t''), ''u''(''x'',''t'') and ''f''(''x'') into
where ''h''
''n''(''t'') and ''b''
''n'' can be calculated by integration, while ''u''
''n''(''t'') is to be determined.
Substitute and back to and considering the orthogonality of sine functions we get
:
which are a sequence of
linear differential equations
In mathematics, a linear differential equation is a differential equation that is defined by a linear polynomial in the unknown function and its derivatives, that is an equation of the form
:a_0(x)y + a_1(x)y' + a_2(x)y'' \cdots + a_n(x)y^ = ...
that can be readily solved with, for instance,
Laplace transform
In mathematics, the Laplace transform, named after its discoverer Pierre-Simon Laplace (), is an integral transform
In mathematics, an integral transform maps a function from its original function space into another function space via integra ...
, or
Integrating factor
In mathematics, an integrating factor is a function that is chosen to facilitate the solving of a given equation involving differentials. It is commonly used to solve ordinary differential equations, but is also used within multivariable calcul ...
. Finally, we can get
:
If the boundary condition is nonhomogeneous, then the expansion of and is no longer valid. One has to find a function ''v'' that satisfies the boundary condition only, and subtract it from ''u''. The function ''u-v'' then satisfies homogeneous boundary condition, and can be solved with the above method.
Example: mixed derivatives
For some equations involving mixed derivatives, the equation does not separate as easily as the heat equation did in the first example above, but nonetheless separation of variables may still be applied. Consider the two-dimensional
biharmonic equation
In mathematics, the biharmonic equation is a fourth-order partial differential equation which arises in areas of continuum mechanics, including linear elasticity theory and the solution of Stokes flows. Specifically, it is used in the modeling ...
:
Proceeding in the usual manner, we look for solutions of the form
:
and we obtain the equation
:
Writing this equation in the form
:
Taking the derivative of this expression with respect to
gives
which means
and likewise, taking derivative with respect to
leads to
and thus
, hence either ''F''(''x'') or ''G''(''y'') must be a constant, say −λ. This further implies that either
or
are constant. Returning to the equation for ''X'' and ''Y'', we have two cases
:
and
:
which can each be solved by considering the separate cases for
and noting that
.
Curvilinear coordinates
In
orthogonal curvilinear coordinates, separation of variables can still be used, but in some details different from that in Cartesian coordinates. For instance, regularity or periodic condition may determine the eigenvalues in place of boundary conditions. See
spherical harmonics
In mathematics and physical science, spherical harmonics are special functions defined on the surface of a sphere. They are often employed in solving partial differential equations in many scientific fields.
Since the spherical harmonics form a ...
for example.
Applicability
Partial differential equations
For many PDEs, such as the wave equation, Helmholtz equation and Schrodinger equation, the applicability of separation of variables is a result of the
spectral theorem
In mathematics, particularly linear algebra and functional analysis, a spectral theorem is a result about when a linear operator or matrix (mathematics), matrix can be Diagonalizable matrix, diagonalized (that is, represented as a diagonal matrix i ...
. In some cases, separation of variables may not be possible. Separation of variables may be possible in some coordinate systems but not others,
[''John Renze, Eric W. Weisstein'', Separation of variables] and which coordinate systems allow for separation depends on the symmetry properties of the equation.
[Willard Miller(1984) ''Symmetry and Separation of Variables'', Cambridge University Press] Below is an outline of an argument demonstrating the applicability of the method to certain linear equations, although the precise method may differ in individual cases (for instance in the biharmonic equation above).
Consider an initial boundary value problem for a function
on
in two variables:
:
where
is a differential operator with respect to
and
is a differential operator with respect to
with boundary data:
:
for
:
for
where
is a known function.
We look for solutions of the form
. Dividing the PDE through by
gives
:
The right hand side depends only on
and the left hand side only on
so both must be equal to a constant
, which gives two ordinary differential equations
:
which we can recognize as eigenvalue problems for the operators for
and
. If
is a compact, self-adjoint operator on the space