Introduction
A function of three variables is " harmonic" or "a solution of the Laplace equation" if it satisfies the condition Such functions were widely studied in the 19th century due to their relevance forDefinition
A partial differential equation is an equation that involves an unknown function of variables and (some of) its partial derivatives. That is, for the unknown function of variables belonging to the open subset of , the -order partial differential equation is defined as where and is the partial derivative operator.Notation
When writing PDEs, it is common to denote partial derivatives using subscripts. For example: In the general situation that is a function of variables, then denotes the first partial derivative relative to the -th input, denotes the second partial derivative relative to the -th and -th inputs, and so on. The Greek letter denotes the Laplace operator; if is a function of variables, then In the physics literature, the Laplace operator is often denoted by ; in the mathematics literature, may also denote the Hessian matrix of .Classification
Linear and nonlinear equations
A PDE is called linear if it is linear in the unknown and its derivatives. For example, for a function of and , a second order linear PDE is of the form where and are functions of the independent variables and only. (Often the mixed-partial derivatives and will be equated, but this is not required for the discussion of linearity.) If the are constants (independent of and ) then the PDE is called linear with constant coefficients. If is zero everywhere then the linear PDE is homogeneous, otherwise it is inhomogeneous. (This is separate from asymptotic homogenization, which studies the effects of high-frequency oscillations in the coefficients upon solutions to PDEs.) Nearest to linear PDEs are semi-linear PDEs, where only the highest order derivatives appear as linear terms, with coefficients that are functions of the independent variables. The lower order derivatives and the unknown function may appear arbitrarily. For example, a general second order semi-linear PDE in two variables is In a quasilinear PDE the highest order derivatives likewise appear only as linear terms, but with coefficients possibly functions of the unknown and lower-order derivatives: Many of the fundamental PDEs in physics are quasilinear, such as the Einstein equations of general relativity and the Navier–Stokes equations describing fluid motion. A PDE without any linearity properties is called fully nonlinear, and possesses nonlinearities on one or more of the highest-order derivatives. An example is the Monge–Ampère equation, which arises inSecond order equations
The elliptic/parabolic/hyperbolic classification provides a guide to appropriate initial- and boundary conditions and to the smoothness of the solutions. Assuming , the general linear second-order PDE in two independent variables has the form where the coefficients , , ... may depend upon and . If over a region of the -plane, the PDE is second-order in that region. This form is analogous to the equation for a conic section: More precisely, replacing by , and likewise for other variables (formally this is done by a Fourier transform), converts a constant-coefficient PDE into a polynomial of the same degree, with the terms of the highest degree (a homogeneous polynomial, here a quadratic form) being most significant for the classification. Just as one classifies conic sections and quadratic forms into parabolic, hyperbolic, and elliptic based on the discriminant , the same can be done for a second-order PDE at a given point. However, the discriminant in a PDE is given by due to the convention of the term being rather than ; formally, the discriminant (of the associated quadratic form) is , with the factor of 4 dropped for simplicity. # ('' elliptic partial differential equation''): Solutions of elliptic PDEs are as smooth as the coefficients allow, within the interior of the region where the equation and solutions are defined. For example, solutions of Laplace's equation are analytic within the domain where they are defined, but solutions may assume boundary values that are not smooth. The motion of a fluid at subsonic speeds can be approximated with elliptic PDEs, and the Euler–Tricomi equation is elliptic where . By change of variables, the equation can always be expressed in the form: where x and y correspond to changed variables. This justifies Laplace equation as an example of this type. # ('' parabolic partial differential equation''): Equations that are parabolic at every point can be transformed into a form analogous to the heat equation by a change of independent variables. Solutions smooth out as the transformed time variable increases. The Euler–Tricomi equation has parabolic type on the line where . By change of variables, the equation can always be expressed in the form: where x correspond to changed variables. This justifies heat equation, which are of form , as an example of this type. # ('' hyperbolic partial differential equation''): hyperbolic equations retain any discontinuities of functions or derivatives in the initial data. An example is the wave equation. The motion of a fluid at supersonic speeds can be approximated with hyperbolic PDEs, and the Euler–Tricomi equation is hyperbolic where . By change of variables, the equation can always be expressed in the form: where x and y correspond to changed variables. This justifies wave equation as an example of this type. If there are independent variables , a general linear partial differential equation of second order has the form The classification depends upon the signature of the eigenvalues of the coefficient matrix . # Elliptic: the eigenvalues are all positive or all negative. # Parabolic: the eigenvalues are all positive or all negative, except one that is zero. # Hyperbolic: there is only one negative eigenvalue and all the rest are positive, or there is only one positive eigenvalue and all the rest are negative. # Ultrahyperbolic: there is more than one positive eigenvalue and more than one negative eigenvalue, and there are no zero eigenvalues. The theory of elliptic, parabolic, and hyperbolic equations have been studied for centuries, largely centered around or based upon the standard examples of the Laplace equation, the heat equation, and the wave equation. However, the classification only depends on linearity of the second-order terms and is therefore applicable to semi- and quasilinear PDEs as well. The basic types also extend to hybrids such as the Euler–Tricomi equation; varying from elliptic to hyperbolic for different regions of the domain, as well as higher-order PDEs, but such knowledge is more specialized.Systems of first-order equations and characteristic surfaces
The classification of partial differential equations can be extended to systems of first-order equations, where the unknown is now a vector with components, and the coefficient matrices are by matrices for . The partial differential equation takes the form where the coefficient matrices and the vector may depend upon and . If a hypersurface is given in the implicit form where has a non-zero gradient, then is a characteristic surface for the operator at a given point if the characteristic form vanishes: The geometric interpretation of this condition is as follows: if data for are prescribed on the surface , then it may be possible to determine the normal derivative of on from the differential equation. If the data on and the differential equation determine the normal derivative of on , then is non-characteristic. If the data on and the differential equation ''do not'' determine the normal derivative of on , then the surface is characteristic, and the differential equation restricts the data on : the differential equation is ''internal'' to . # A first-order system is ''elliptic'' if no surface is characteristic for : the values of on and the differential equation always determine the normal derivative of on . # A first-order system is ''hyperbolic'' at a point if there is a spacelike surface with normal at that point. This means that, given any non-trivial vector orthogonal to , and a scalar multiplier , the equation has real roots . The system is strictly hyperbolic if these roots are always distinct. The geometrical interpretation of this condition is as follows: the characteristic form defines a cone (the normal cone) with homogeneous coordinates ζ. In the hyperbolic case, this cone has sheets, and the axis runs inside these sheets: it does not intersect any of them. But when displaced from the origin by η, this axis intersects every sheet. In the elliptic case, the normal cone has no real sheets.Analytical solutions
Separation of variables
Linear PDEs can be reduced to systems of ordinary differential equations by the important technique of separation of variables. This technique rests on a feature of solutions to differential equations: if one can find any solution that solves the equation and satisfies the boundary conditions, then it is ''the'' solution (this also applies to ODEs). We assume as an ansatz that the dependence of a solution on the parameters space and time can be written as a product of terms that each depend on a single parameter, and then see if this can be made to solve the problem. In the method of separation of variables, one reduces a PDE to a PDE in fewer variables, which is an ordinary differential equation if in one variable – these are in turn easier to solve. This is possible for simple PDEs, which are called separable partial differential equations, and the domain is generally a rectangle (a product of intervals). Separable PDEs correspond to diagonal matrices – thinking of "the value for fixed " as a coordinate, each coordinate can be understood separately. This generalizes to the method of characteristics, and is also used in integral transforms.Method of characteristics
The characteristic surface in dimensional space is called a characteristic curve. In special cases, one can find characteristic curves on which the first-order PDE reduces to an ODE – changing coordinates in the domain to straighten these curves allows separation of variables, and is called the method of characteristics. More generally, applying the method to first-order PDEs in higher dimensions, one may find characteristic surfaces.Integral transform
An integral transform may transform the PDE to a simpler one, in particular, a separable PDE. This corresponds to diagonalizing an operator. An important example of this is Fourier analysis, which diagonalizes the heat equation using the eigenbasis of sinusoidal waves. If the domain is finite or periodic, an infinite sum of solutions such as a Fourier series is appropriate, but an integral of solutions such as a Fourier integral is generally required for infinite domains. The solution for a point source for the heat equation given above is an example of the use of a Fourier integral.Change of variables
Often a PDE can be reduced to a simpler form with a known solution by a suitable change of variables. For example, the Black–Scholes equation is reducible to the heat equation by the change of variablesFundamental solution
Inhomogeneous equations can often be solved (for constant coefficient PDEs, always be solved) by finding the fundamental solution (the solution for a point source ), then taking the convolution with the boundary conditions to get the solution. This is analogous inSuperposition principle
The superposition principle applies to any linear system, including linear systems of PDEs. A common visualization of this concept is the interaction of two waves in phase being combined to result in a greater amplitude, for example . The same principle can be observed in PDEs where the solutions may be real or complex and additive. If and are solutions of linear PDE in some function space , then with any constants and are also a solution of that PDE in the same function space.Methods for non-linear equations
There are no generally applicable analytical methods to solve nonlinear PDEs. Still, existence and uniqueness results (such as the Cauchy–Kowalevski theorem) are often possible, as are proofs of important qualitative and quantitative properties of solutions (getting these results is a major part of analysis). Nevertheless, some techniques can be used for several types of equations. The -principle is the most powerful method to solve underdetermined equations. The Riquier–Janet theory is an effective method for obtaining information about many analytic overdetermined systems. The method of characteristics can be used in some very special cases to solve nonlinear partial differential equations. In some cases, a PDE can be solved via perturbation analysis in which the solution is considered to be a correction to an equation with a known solution. Alternatives areLie group method
From 1870 Sophus Lie's work put the theory of differential equations on a more satisfactory foundation. He showed that the integration theories of the older mathematicians can, by the introduction of what are now called Lie groups, be referred, to a common source; and that ordinary differential equations which admit the same infinitesimal transformations present comparable difficulties of integration. He also emphasized the subject of transformations of contact. A general approach to solving PDEs uses the symmetry property of differential equations, the continuous infinitesimal transformations of solutions to solutions ( Lie theory). Continuous group theory, Lie algebras andSemi-analytical methods
The Adomian decomposition method, the Lyapunov artificial small parameter method, and his homotopy perturbation method are all special cases of the more general homotopy analysis method. These are series expansion methods, and except for the Lyapunov method, are independent of small physical parameters as compared to the well known perturbation theory, thus giving these methods greater flexibility and solution generality.Numerical solutions
The three most widely used numerical methods to solve PDEs are the finite element method (FEM), finite volume methods (FVM) and finite difference methods (FDM), as well other kind of methods called meshfree methods, which were made to solve problems where the aforementioned methods are limited. The FEM has a prominent position among these methods and especially its exceptionally efficient higher-order version hp-FEM. Other hybrid versions of FEM and Meshfree methods include the generalized finite element method (GFEM), extended finite element method (XFEM), spectral finite element method (SFEM), meshfree finite element method, discontinuous Galerkin finite element method (DGFEM), element-free Galerkin method (EFGM), interpolating element-free Galerkin method (IEFGM), etc.Finite element method
The finite element method (FEM) (its practical application often known as finite element analysis (FEA)) is a numerical technique for approximating solutions of partial differential equations (PDE) as well as of integral equations using a finite set of functions. The solution approach is based either on eliminating the differential equation completely (steady state problems), or rendering the PDE into an approximating system of ordinary differential equations, which are then numerically integrated using standard techniques such as Euler's method, Runge–Kutta, etc.Finite difference method
Finite-difference methods are numerical methods for approximating the solutions to differential equations usingFinite volume method
Similar to the finite difference method or finite element method, values are calculated at discrete places on a meshed geometry. "Finite volume" refers to the small volume surrounding each node point on a mesh. In the finite volume method, surface integrals in a partial differential equation that contain a divergence term are converted to volume integrals, using the divergence theorem. These terms are then evaluated as fluxes at the surfaces of each finite volume. Because the flux entering a given volume is identical to that leaving the adjacent volume, these methods conserve mass by design.Neural networks
Weak solutions
Weak solutions are functions that satisfy the PDE, yet in other meanings than regular sense. The meaning for this term may differ with context, and one of the most commonly used definitions is based on the notion of distributions. An example for the definition of a weak solution is as follows: Consider the boundary-value problem given by: where denotes a second-order partial differential operator in divergence form. We say a is a weak solution if for every , which can be derived by a formal integral by parts. An example for a weak solution is as follows: is a weak solution satisfying in distributional sense, as formally,Theoretical Studies
As a branch of pure mathematics, the theoretical studies of PDEs focus on the criteria for a solution to exist, the properties of a solution, and finding its formula is often secondary.Well-posedness
Well-posedness refers to a common schematic package of information about a PDE. To say that a PDE is well-posed, one must have: * an existence and uniqueness theorem, asserting that by the prescription of some freely chosen functions, one can single out one specific solution of the PDE * by continuously changing the free choices, one continuously changes the corresponding solution This is, by the necessity of being applicable to several different PDE, somewhat vague. The requirement of "continuity", in particular, is ambiguous, since there are usually many inequivalent means by which it can be rigorously defined. It is, however, somewhat unusual to study a PDE without specifying a way in which it is well-posed.Regularity
Regularity refers to the integrability and differentiability of weak solutions, which can often be represented by Sobolev spaces. This problem arise due to the difficulty in searching for classical solutions. Researchers often tend to find weak solutions at first and then find out whether it is smooth enough to be qualified as a classical solution. Results fromSee also
Some common PDEs * Acoustic wave equation * Burgers' equation * Continuity equation * Heat equation * Helmholtz equation * Klein–Gordon equation * Jacobi equation * Lagrange equation * Lorenz equation * Laplace's equation * Maxwell's equations * Navier-Stokes equation * Poisson's equation * Reaction–diffusion system * Schrödinger equation * Wave equation Types of boundary conditions * Dirichlet boundary condition * Neumann boundary condition * Robin boundary condition * Cauchy problem Various topics * Jet bundle * Laplace transform applied to differential equations * List of dynamical systems and differential equations topics * Matrix differential equation * Numerical partial differential equations * Partial differential algebraic equation * Recurrence relation * Stochastic processes and boundary value problemsNotes
References
* . * * * . * . * . * . * . * . * . * . * . * * . * * * . * . * . * . * . * *Further reading
* * Nirenberg, Louis (1994). "Partial differential equations in the first half of the century." Development of mathematics 1900–1950 (Luxembourg, 1992), 479–515, Birkhäuser, Basel. *External links
*