HOME

TheInfoList



OR:

The theory of
optimal control Optimal control theory is a branch of control theory that deals with finding a control for a dynamical system over a period of time such that an objective function is optimized. It has numerous applications in science, engineering and operations ...
is concerned with operating a
dynamic system In mathematics, a dynamical system is a system in which a function describes the time dependence of a point in an ambient space, such as in a parametric curve. Examples include the mathematical models that describe the swinging of a clock ...
at minimum cost. The case where the system dynamics are described by a set of
linear differential equation In mathematics, a linear differential equation is a differential equation that is linear equation, linear in the unknown function and its derivatives, so it can be written in the form a_0(x)y + a_1(x)y' + a_2(x)y'' \cdots + a_n(x)y^ = b(x) wher ...
s and the cost is described by a quadratic
function Function or functionality may refer to: Computing * Function key, a type of key on computer keyboards * Function model, a structured representation of processes in a system * Function object or functor or functionoid, a concept of object-orie ...
is called the LQ problem. One of the main results in the theory is that the solution is provided by the linear–quadratic regulator (LQR), a feedback controller whose equations are given below. LQR controllers possess inherent robustness with guaranteed gain and
phase margin In electronic amplifiers, the phase margin (PM) is the difference between the phase (waves), phase lag (< 0) and -180°, for an amplifier's output signal (relative to its input) at zero dB gain - i.e. unity gain, or that the output signal has the ...
, and they also are part of the solution to the LQG (linear–quadratic–Gaussian) problem. Like the LQR problem itself, the LQG problem is one of the most fundamental problems in
control theory Control theory is a field of control engineering and applied mathematics that deals with the control system, control of dynamical systems in engineered processes and machines. The objective is to develop a model or algorithm governing the applic ...
.


General description

The settings of a (regulating) controller governing either a machine or process (like an airplane or chemical reactor) are found by using a mathematical algorithm that minimizes a cost function with weighting factors supplied by the operator. The cost function is often defined as a sum of the deviations of key measurements, like altitude or process temperature, from their desired values. The algorithm thus finds those controller settings that minimize undesired deviations. The magnitude of the control action itself may also be included in the cost function. The LQR algorithm reduces the amount of work done by the control systems engineer to optimize the controller. However, the engineer still needs to specify the cost function parameters, and compare the results with the specified design goals. Often this means that controller construction will be an iterative process in which the engineer judges the "optimal" controllers produced through simulation and then adjusts the parameters to produce a controller more consistent with design goals. The LQR algorithm is essentially an automated way of finding an appropriate state-feedback controller. As such, it is not uncommon for control engineers to prefer alternative methods, like
full state feedback Full state feedback (FSF), or pole placement, is a method employed in feedback control system theory to place the closed-loop poles of a Plant (control theory), plant in predetermined locations in the s-plane.* Placing poles is desirable because t ...
, also known as pole placement, in which there is a clearer relationship between controller parameters and controller behavior. Difficulty in finding the right weighting factors limits the application of the LQR based controller synthesis.


Versions


Finite-horizon, continuous-time

For a continuous-time linear system, defined on described by: \dot = A\mathbf + B\mathbf where \mathbf \in \mathbb^ (that is, \mathbf is an n-dimensional real-valued vector) is the state of the system and \mathbf \in \mathbb^ is the control input. Given a quadratic cost function for the system, defined as: J = \mathbf^\mathsf\!(t_1) F(t_1) \mathbf(t_1) + \int_^ \left( \mathbf^\mathsf Q \mathbf + \mathbf^\mathsf R \mathbf + 2 \mathbf^\mathsf N \mathbf \right) dt where F is the terminal cost matrix, Q is the state cost matrix, R is the control cost matrix, and N is the cross-term (control and state) cost matrix, the feedback control law that minimizes the value of the cost is: \mathbf = -K \mathbf where K is given by: K = R^ \left(B^\mathsf P(t) + N^\mathsf\right) and P is found by solving the continuous time Riccati differential equation: A^\mathsf P(t) + P(t) A - \left (t) B + N\rightR^ \left ^\mathsf P(t) + N^\mathsf\right+ Q = - \dot(t) with the boundary condition: P(t_1) = F(t_1). The first order conditions for J_\min are: # State equation \dot = A\mathbf + B\mathbf # Co-state equation -\dot = Q\mathbf + N\mathbf + A^\mathsf \boldsymbol\lambda # Stationary equation \mathbf = R \mathbf + N^\mathsf \mathbf + B^\mathsf \boldsymbol\lambda # Boundary conditions \mathbf(t_0) = \mathbf_0 and \boldsymbol\lambda(t_1) = F(t_1) \mathbf(t_1)


Infinite-horizon, continuous-time

For a continuous-time linear system described by: \dot = A\mathbf + B\mathbf with a cost function defined as: J = \int_0^\infty \left( \mathbf^\mathsf Q \mathbf + \mathbf^\mathsf R \mathbf + 2 \mathbf^\mathsf N \mathbf \right) dt the feedback control law that minimizes the value of the cost is: \mathbf = -K \mathbf where K is given by: K = R^ \left(B^\mathsf P + N^\mathsf\right) and P is found by solving the continuous time
algebraic Riccati equation An algebraic Riccati equation is a type of nonlinear equation that arises in the context of infinite-horizon optimal control problems in continuous time or discrete time. A typical algebraic Riccati equation is similar to one of the following: t ...
: A^\mathsf P + P A - \left(P B + N\right) R^ \left(B^\mathsf P + N^\mathsf\right) + Q = 0 This can be also written as: \mathcal A^\mathsf P + P \mathcal A - P B R^ B^\mathsf P + \mathcal Q = 0 with \mathcal A = A - B R^ N^\mathsf, \qquad \mathcal Q = Q - N R^ N^\mathsf


Finite-horizon, discrete-time

For a discrete-time linear system described by: \mathbf_ = A \mathbf_k + B \mathbf_k with a performance index defined as: J = \mathbf_^\mathsf Q_ \mathbf_ + \sum_^ \left( \mathbf_k^\mathsf Q \mathbf_k + \mathbf_k^\mathsf R \mathbf_k + 2 \mathbf_k^\mathsf N \mathbf_k \right), where H_p is the time horizon. the optimal control sequence minimizing the performance index is given by: \mathbf_k = -F_k \mathbf_k where F_k = ^ \left(B^\mathsf P_ A + N^\mathsf\right) and P_k is found iteratively backwards in time by the dynamic Riccati equation: P_ = A^\mathsf P_k A - \left(A^\mathsf P_k B + N\right) \left( R + B^\mathsf P_k B \right)^ \left(B^\mathsf P_k A + N^\mathsf\right) + Q from terminal condition P_ = Q_. Note that \mathbf_ is not defined, since x is driven to its final state \mathbf_ by A \mathbf_ + B \mathbf_.


Infinite-horizon, discrete-time

For a discrete-time linear system described by: \mathbf_ = A \mathbf_k + B \mathbf_k with a performance index defined as: J = \sum_^ \left( \mathbf_k^\mathsf Q \mathbf_k + \mathbf_k^\mathsf R \mathbf_k + 2 \mathbf_k^\mathsf N \mathbf_k \right) the optimal control sequence minimizing the performance index is given by: \mathbf_k = -F \mathbf_k where: F = ^ \left(B^\mathsf P A + N^\mathsf\right) and P is the unique positive definite solution to the discrete time
algebraic Riccati equation An algebraic Riccati equation is a type of nonlinear equation that arises in the context of infinite-horizon optimal control problems in continuous time or discrete time. A typical algebraic Riccati equation is similar to one of the following: t ...
(DARE): P = A^\mathsf P A - \left(A^\mathsf P B + N\right) \left( R + B^\mathsf P B \right)^ \left(B^\mathsf P A + N^\mathsf\right) + Q . This can be also written as: P = \mathcal A^\mathsf P \mathcal A - \mathcal A^\mathsf P B \left( R + B^\mathsf P B \right)^ B^\mathsf P \mathcal A + \mathcal Q with: \mathcal A = A - B R^ N^\mathsf, \qquad \mathcal Q = Q - N R^ N^\mathsf . Note that one way to solve the algebraic Riccati equation is by iterating the dynamic Riccati equation of the finite-horizon case until it converges.


Constraints

In practice, not all values of \mathbf_k may be allowed. One common constraint is the linear one: C \mathbf + D\mathbf \leq \mathbf. The finite horizon version of this is a
convex optimization Convex optimization is a subfield of mathematical optimization that studies the problem of minimizing convex functions over convex sets (or, equivalently, maximizing concave functions over convex sets). Many classes of convex optimization problems ...
problem, and so the problem is often solved repeatedly with a receding horizon. This is a form of
model predictive control Model predictive control (MPC) is an advanced method of process control that is used to control a process while satisfying a set of constraints. It has been in use in the process industries in chemical plants and oil refineries since the 1980s. I ...
.


Related controllers


Quadratic-quadratic regulator

If the state equation is quadratic then the problem is known as the quadratic-quadratic regulator (QQR). The Al'Brekht algorithm can be applied to reduce this problem to one that can be solved efficiently using tensor based linear solvers.


Polynomial-quadratic regulator

If the state equation is
polynomial In mathematics, a polynomial is a Expression (mathematics), mathematical expression consisting of indeterminate (variable), indeterminates (also called variable (mathematics), variables) and coefficients, that involves only the operations of addit ...
then the problem is known as the polynomial-quadratic regulator (PQR). Again, the Al'Brekht algorithm can be applied to reduce this problem to a large linear one which can be solved with a generalization of the Bartels-Stewart algorithm; this is feasible provided that the degree of the polynomial is not too high.


Model predictive control

Model predictive control Model predictive control (MPC) is an advanced method of process control that is used to control a process while satisfying a set of constraints. It has been in use in the process industries in chemical plants and oil refineries since the 1980s. I ...
(MPC) and linear-quadratic regulators are two types of optimal control methods that have distinct approaches for setting the optimization costs. In particular, when the LQR is run repeatedly with a receding horizon, it becomes a form of MPC. In general, however, MPC is not limited to linear system and can naturally incorporate constraints.


References

* *


External links


MATLAB function for Linear Quadratic Regulator design


{{DEFAULTSORT:Linear-quadratic regulator Optimal control