Singular Control
   HOME

TheInfoList



OR:

In
optimal control Optimal control theory is a branch of mathematical optimization that deals with finding a control for a dynamical system over a period of time such that an objective function is optimized. It has numerous applications in science, engineering and ...
, problems of singular control are problems that are difficult to solve because a straightforward application of Pontryagin's minimum principle fails to yield a complete solution. Only a few such problems have been solved, such as
Merton's portfolio problem Merton's portfolio problem is a well known problem in continuous-time finance and in particular intertemporal portfolio choice. An investor must choose how much to consume and must allocate their wealth between stocks and a risk-free asset so as ...
in
financial economics Financial economics, also known as finance, is the branch of economics characterized by a "concentration on monetary activities", in which "money of one type or another is likely to appear on ''both sides'' of a trade".William F. Sharpe"Financial ...
or
trajectory optimization Trajectory optimization is the process of designing a trajectory that minimizes (or maximizes) some measure of performance while satisfying a set of constraints. Generally speaking, trajectory optimization is a technique for computing an open-loop ...
in aeronautics. A more technical explanation follows. The most common difficulty in applying Pontryagin's principle arises when the
Hamiltonian Hamiltonian may refer to: * Hamiltonian mechanics, a function that represents the total energy of a system * Hamiltonian (quantum mechanics), an operator corresponding to the total energy of that system ** Dyall Hamiltonian, a modified Hamiltonian ...
depends linearly on the control u, i.e., is of the form: H(u)=\phi(x,\lambda,t)u+\cdots and the control is restricted to being between an upper and a lower bound: a\le u(t)\le b. To minimize H(u), we need to make u as big or as small as possible, depending on the sign of \phi(x,\lambda,t), specifically: : u(t) = \begin b, & \phi(x,\lambda,t)<0 \\ ?, & \phi(x,\lambda,t)=0 \\ a, & \phi(x,\lambda,t)>0.\end If \phi is positive at some times, negative at others and is only zero instantaneously, then the solution is straightforward and is a
bang-bang control Bang Bang or Bang Bang Bang or similar may refer to: Food * Bang bang chicken, a Chinese dish *Bang bang shrimp, a Chinese dish People * Abdul Razzaq (cricketer) (born 1979), nicknamed Bang Bang Razzaq * Bang Bang (Dubliner) (1906–1981), ...
that switches from b to a at times when \phi switches from negative to positive. The case when \phi remains at zero for a finite length of time t_1\le t\le t_2 is called the singular control case. Between t_1 and t_2 the maximization of the Hamiltonian with respect to u gives us no useful information and the solution in that time interval is going to have to be found from other considerations. (One approach would be to repeatedly differentiate \partial H/\partial u with respect to time until the control u again explicitly appears, though this is not guaranteed to happen eventually. One can then set that expression to zero and solve for u. This amounts to saying that between t_1 and t_2 the control u is determined by the requirement that the singularity condition continues to hold. The resulting so-called singular arc, if it is optimal, will satisfy the Kelley condition: :(-1)^k \frac \left ^ H_u \right\ge 0 ,\, k=0,1,\cdots Others refer to this condition as the generalized
Legendre–Clebsch condition __NOTOC__ In the calculus of variations the Legendre–Clebsch condition is a second-order condition which a solution of the Euler–Lagrange equation must satisfy in order to be a minimum. For the problem of minimizing : \int_^ L(t,x,x')\, dt . ...
. The term bang-singular control refers to a control that has a bang-bang portion as well as a singular portion.


References


External links

* {{DEFAULTSORT:Singular Control Control theory Optimal control