HOME

TheInfoList



OR:

Pontryagin's maximum principle is used in
optimal control Optimal control theory is a branch of control theory that deals with finding a control for a dynamical system over a period of time such that an objective function is optimized. It has numerous applications in science, engineering and operations ...
theory to find the best possible control for taking a
dynamical system In mathematics, a dynamical system is a system in which a Function (mathematics), function describes the time dependence of a Point (geometry), point in an ambient space, such as in a parametric curve. Examples include the mathematical models ...
from one state to another, especially in the presence of constraints for the state or input controls. It states that it is necessary for any optimal control along with the optimal state trajectory to solve the so-called Hamiltonian system, which is a two-point
boundary value problem In the study of differential equations, a boundary-value problem is a differential equation subjected to constraints called boundary conditions. A solution to a boundary value problem is a solution to the differential equation which also satis ...
, plus a maximum condition of the control Hamiltonian. These necessary conditions become sufficient under certain convexity conditions on the objective and constraint functions. The maximum principle was formulated in 1956 by the Russian mathematician Lev Pontryagin and his students, and its initial application was to the maximization of the terminal speed of a rocket. The result was derived using ideas from the classical
calculus of variations The calculus of variations (or variational calculus) is a field of mathematical analysis that uses variations, which are small changes in Function (mathematics), functions and functional (mathematics), functionals, to find maxima and minima of f ...
. After a slight perturbation of the optimal control, one considers the first-order term of a
Taylor Taylor, Taylors or Taylor's may refer to: People * Taylor (surname) ** List of people with surname Taylor * Taylor (given name), including Tayla and Taylah * Taylor sept, a branch of Scottish clan Cameron * Justice Taylor (disambiguation) ...
expansion with respect to the perturbation; sending the perturbation to zero leads to a variational inequality from which the maximum principle follows. Widely regarded as a milestone in optimal control theory, the significance of the maximum principle lies in the fact that maximizing the Hamiltonian is much easier than the original infinite-dimensional control problem; rather than maximizing over a
function space In mathematics, a function space is a set of functions between two fixed sets. Often, the domain and/or codomain will have additional structure which is inherited by the function space. For example, the set of functions from any set into a ve ...
, the problem is converted to a
pointwise In mathematics, the qualifier pointwise is used to indicate that a certain property is defined by considering each value f(x) of some Function (mathematics), function f. An important class of pointwise concepts are the ''pointwise operations'', that ...
optimization. A similar logic leads to Bellman's principle of optimality, a related approach to optimal control problems which states that the optimal trajectory remains optimal at intermediate points in time. The resulting
Hamilton–Jacobi–Bellman equation The Hamilton-Jacobi-Bellman (HJB) equation is a nonlinear partial differential equation that provides necessary and sufficient conditions for optimality of a control with respect to a loss function. Its solution is the value function of the opti ...
provides a necessary and sufficient condition for an optimum, and admits a straightforward extension to stochastic optimal control problems, whereas the maximum principle does not. However, in contrast to the Hamilton–Jacobi–Bellman equation, which needs to hold over the entire state space to be valid, Pontryagin's Maximum Principle is potentially more computationally efficient in that the conditions which it specifies only need to hold over a particular trajectory.


Notation

For set \mathcal and functions :\Psi : \reals^n \to \reals, :H : \reals^n \times \mathcal \times \reals^n \times \reals \to \reals, :L : \reals^n \times \mathcal \to \reals, :f : \reals^n \times \mathcal \to \reals^n, we use the following notation: : \Psi_T(x(T))= \left.\frac\_ \, , : \Psi_x(x(T))=\begin \left.\frac\_ & \cdots & \left.\frac \_ \end , : H_x(x^*,u^*,\lambda^*,t)=\begin \left.\frac\_ & \cdots & \left.\frac\_ \end , : L_x(x^*,u^*)=\begin \left.\frac\_ & \cdots & \left.\frac\_ \end , : f_x(x^*,u^*)=\begin \left.\frac\_ & \cdots & \left.\frac\_ \\ \vdots & \ddots & \vdots \\ \left.\frac\_ & \ldots & \left.\frac\_ \end .


Formal statement of necessary conditions for minimization problems

Here the necessary conditions are shown for minimization of a functional. Consider an n-dimensional
dynamical system In mathematics, a dynamical system is a system in which a Function (mathematics), function describes the time dependence of a Point (geometry), point in an ambient space, such as in a parametric curve. Examples include the mathematical models ...
, with state variable x \in \R^n, and control variable u \in \mathcal, where \mathcal is the set of admissible controls. The evolution of the system is determined by the state and the control, according to the differential equation \dot=f(x,u). Let the system's initial state be x_0 and let the system's evolution be controlled over the time-period with values t \in , T/math>. The latter is determined by the following differential equation: : \dot=f(x,u), \quad x(0)=x_0, \quad u(t) \in \mathcal, \quad t \in ,T The control trajectory u: , T\to \mathcal is to be chosen according to an objective. The objective is a functional J defined by : J=\Psi(x(T))+\int^T_0 L\big(x(t),u(t)\big) \,dt , where L(x, u) can be interpreted as the ''rate'' of cost for exerting control u in state x, and \Psi(x) can be interpreted as the cost for ending up at state x. The specific choice of L, \Psi depends on the application. The constraints on the system dynamics can be adjoined to the Lagrangian L by introducing time-varying
Lagrange multiplier In mathematical optimization, the method of Lagrange multipliers is a strategy for finding the local maxima and minima of a function (mathematics), function subject to constraint (mathematics), equation constraints (i.e., subject to the conditio ...
vector \lambda, whose elements are called the ''costates'' of the system. This motivates the construction of the
Hamiltonian Hamiltonian may refer to: * Hamiltonian mechanics, a function that represents the total energy of a system * Hamiltonian (quantum mechanics), an operator corresponding to the total energy of that system ** Dyall Hamiltonian, a modified Hamiltonian ...
H defined for all t \in ,T/math> by: : H\big(x(t),u(t),\lambda(t),t\big)=\lambda^(t)\cdot f\big(x(t),u(t)\big) + L\big(x(t),u(t)\big) where \lambda^ is the transpose of \lambda. Pontryagin's minimum principle states that the optimal state trajectory x^*, optimal control u^*, and corresponding Lagrange multiplier vector \lambda^* must minimize the Hamiltonian H so that for all time t \in ,T/math> and for all permissible control inputs u \in \mathcal. Here, the trajectory of the Lagrangian multiplier vector \lambda is the solution to the costate equation and its terminal conditions: If x(T) is fixed, then these three conditions in (1)-(3) are the necessary conditions for an optimal control. If the final state x(T) is not fixed (i.e., its differential variation is not zero), there is an additional condition These four conditions in (1)-(4) are the necessary conditions for an optimal control.


See also

*
Lagrange multipliers on Banach spaces In the field of calculus of variations in mathematics, the method of Lagrange multipliers on Banach spaces can be used to solve certain infinite-dimensional constrained optimization problems. The method is a generalization of the classical method ...
, Lagrangian method in calculus of variations


Notes


References


Further reading

* * * *


External links

* {{DEFAULTSORT:Pontryagin's Minimum Principle Optimal control Principles fr:Commande optimale#Principe du maximum ru:Оптимальное управление#Принцип максимума Понтрягина