Linear–quadratic–Gaussian Control
   HOME

TheInfoList



OR:

In
control theory Control theory is a field of mathematics that deals with the control of dynamical systems in engineered processes and machines. The objective is to develop a model or algorithm governing the application of system inputs to drive the system to a ...
, the linear–quadratic–Gaussian (LQG) control problem is one of the most fundamental optimal control problems, and it can also be operated repeatedly for
model predictive control Model predictive control (MPC) is an advanced method of process control that is used to control a process while satisfying a set of constraints. It has been in use in the process industries in chemical plants and oil refineries since the 1980s. In r ...
. It concerns
linear system In systems theory, a linear system is a mathematical model of a system based on the use of a linear operator. Linear systems typically exhibit features and properties that are much simpler than the nonlinear case. As a mathematical abstracti ...
s driven by
additive white Gaussian noise Additive white Gaussian noise (AWGN) is a basic noise model used in information theory to mimic the effect of many random processes that occur in nature. The modifiers denote specific characteristics: * ''Additive'' because it is added to any nois ...
. The problem is to determine an output feedback law that is optimal in the sense of minimizing the expected value of a quadratic
cost In Production (economics), production, research, retail, and accounting, a cost is the value of money that has been used up to produce something or deliver a service, and hence is not available for use anymore. In business, the cost may be one o ...
criterion. Output measurements are assumed to be corrupted by Gaussian noise and the initial state, likewise, is assumed to be a Gaussian random vector. Under these assumptions an optimal control scheme within the class of linear control laws can be derived by a completion-of-squares argument. This control law which is known as the LQG controller, is unique and it is simply a combination of a
Kalman filter For statistics and control theory, Kalman filtering, also known as linear quadratic estimation (LQE), is an algorithm that uses a series of measurements observed over time, including statistical noise and other inaccuracies, and produces estima ...
(a linear–quadratic state estimator (LQE)) together with a
linear–quadratic regulator The theory of optimal control is concerned with operating a dynamic system at minimum cost. The case where the system dynamics are described by a set of linear differential equations and the cost is described by a quadratic function is called t ...
(LQR). The separation principle states that the state estimator and the state feedback can be designed independently. LQG control applies to both linear time-invariant systems as well as linear time-varying systems, and constitutes a linear dynamic feedback control law that is easily computed and implemented: the LQG controller itself is a dynamic system like the system it controls. Both systems have the same state dimension. A deeper statement of the separation principle is that the LQG controller is still optimal in a wider class of possibly nonlinear controllers. That is, utilizing a nonlinear control scheme will not improve the expected value of the cost functional. This version of the separation principle is a special case of the separation principle of stochastic control which states that even when the process and output noise sources are possibly non-Gaussian martingales, as long as the system dynamics are linear, the optimal control separates into an optimal state estimator (which may no longer be a Kalman filter) and an LQR regulator.. In the classical LQG setting, implementation of the LQG controller may be problematic when the dimension of the system state is large. The reduced-order LQG problem (fixed-order LQG problem) overcomes this by fixing ''a priori'' the number of states of the LQG controller. This problem is more difficult to solve because it is no longer separable. Also, the solution is no longer unique. Despite these facts numerical algorithms are available Associated software download from Matlab Central
Associated software download from Matlab Central
to solve the associated optimal projection equations which constitute necessary and sufficient conditions for a locally optimal reduced-order LQG controller. LQG optimality does not automatically ensure good robustness properties. The robust stability of the closed loop system must be checked separately after the LQG controller has been designed. To promote robustness some of the system parameters may be assumed stochastic instead of deterministic. The associated more difficult control problem leads to a similar optimal controller of which only the controller parameters are different. It is possible to compute the expected value of the cost function for the optimal gains, as well as any other set of stable gains. The LQG controller is also used to control perturbed non-linear systems.


Mathematical description of the problem and solution


Continuous time

Consider the
continuous-time In mathematical dynamics, discrete time and continuous time are two alternative frameworks within which variables that evolve over time are modeled. Discrete time Discrete time views values of variables as occurring at distinct, separate "po ...
linear dynamic system : \dot(t) = A(t) \mathbf(t) + B(t) \mathbf(t) + \mathbf(t), : \mathbf(t) = C(t) \mathbf(t) + \mathbf(t), where represents the vector of state variables of the system, the vector of control inputs and the vector of measured outputs available for feedback. Both additive white Gaussian system noise \mathbf(t) and additive white Gaussian measurement noise \mathbf(t) affect the system. Given this system the objective is to find the control input history (t) which at every time t may depend linearly only on the past measurements (t'), 0 \leq t' < t such that the following cost function is minimized: : J = \mathbb\left T)F(T)+ \int_^ (t)Q(t)(t) + (t)R(t)(t)\,dt \right : F \ge 0,\quad Q(t) \ge 0,\quad R(t) > 0, where \mathbb denotes the
expected value In probability theory, the expected value (also called expectation, expectancy, mathematical expectation, mean, average, or first moment) is a generalization of the weighted average. Informally, the expected value is the arithmetic mean of a ...
. The final time (horizon) T may be either finite or infinite. If the horizon tends to infinity the first term ^\mathrm T(T)F(T) of the cost function becomes negligible and irrelevant to the problem. Also to keep the costs finite the cost function has to be taken to be J/T. The LQG controller that solves the LQG control problem is specified by the following equations: : \dot(t) = A(t)\hat(t) + B(t)(t)+L(t) \left( (t)-C(t)\hat(t) \right), \quad \hat(0) = \mathbb \left (0) \right : (t)= -K(t) \hat(t). The matrix L(t) is called the Kalman gain of the associated
Kalman filter For statistics and control theory, Kalman filtering, also known as linear quadratic estimation (LQE), is an algorithm that uses a series of measurements observed over time, including statistical noise and other inaccuracies, and produces estima ...
represented by the first equation. At each time t this filter generates estimates \hat(t) of the state (t) using the past measurements and inputs. The Kalman gain L(t) is computed from the matrices A(t), C(t), the two intensity matrices \mathbfV(t), W(t) associated to the white Gaussian noises \mathbf(t) and \mathbf(t) and finally \mathbb\left 0)^\mathrm T(0) \right/math>. These five matrices determine the Kalman gain through the following associated matrix Riccati differential equation: : \dot(t) = A(t)P(t)+P(t)A^\mathrm T(t)-P(t)C^\mathrm T(t)W^(t) C(t)P(t)+V(t), : P(0)= \mathbb \left 0)^\mathrm T(0) \right Given the solution P(t), 0 \leq t \leq T the Kalman gain equals : L(t) = P(t)C^\mathrm T(t)W^(t). The matrix K(t) is called the feedback gain matrix. This matrix is determined by the matrices A(t), B(t), Q(t), R(t) and F through the following associated matrix Riccati differential equation: : -\dot(t) = A^\mathrm T(t)S(t)+S(t)A(t)-S(t)B(t)R^(t)B^\mathrm T(t)S(t)+Q(t), : S(T) = F. Given the solution S(t), 0 \leq t \leq T the feedback gain equals : K(t) = R^(t)B^\mathrm T(t)S(t). Observe the similarity of the two matrix Riccati differential equations, the first one running forward in time, the second one running backward in time. This similarity is called duality. The first matrix Riccati differential equation solves the linear–quadratic estimation problem (LQE). The second matrix Riccati differential equation solves the
linear–quadratic regulator The theory of optimal control is concerned with operating a dynamic system at minimum cost. The case where the system dynamics are described by a set of linear differential equations and the cost is described by a quadratic function is called t ...
problem (LQR). These problems are dual and together they solve the linear–quadratic–Gaussian control problem (LQG). So the LQG problem separates into the LQE and LQR problem that can be solved independently. Therefore, the LQG problem is called separable. When A(t), B(t), C(t), Q(t), R(t) and the noise intensity matrices \mathbfV(t), \mathbfW(t) do not depend on t and when T tends to infinity the LQG controller becomes a time-invariant dynamic system. In that case the second matrix Riccati differential equation may be replaced by the associated algebraic Riccati equation.


Discrete time

Since the
discrete-time In mathematical dynamics, discrete time and continuous time are two alternative frameworks within which variables that evolve over time are modeled. Discrete time Discrete time views values of variables as occurring at distinct, separate "po ...
LQG control problem is similar to the one in continuous-time, the description below focuses on the mathematical equations. The discrete-time linear system equations are : _ = A_i\mathbf_i + B_i \mathbf_i + \mathbf_i, : \mathbf_ = C_ \mathbf_i + \mathbf_i. Here \mathbfi represents the discrete time index and \mathbf_, \mathbf_ represent discrete-time Gaussian white noise processes with covariance matrices \mathbfV_, W_, respectively, and are independent of each other. The quadratic cost function to be minimized is : J = \mathbb\left \mathrm T_F_+ \sum_^( \mathbf_i^\mathrm T Q_i \mathbf_i + \mathbf_i^\mathrm T R_i \mathbf_i )\right : F \ge 0, Q_i \ge 0, R_i > 0. \, The discrete-time LQG controller is :\hat_=A_i\hat_i+B_i_i+L_ \left(_-C_ \left\ \right), \qquad \hat_0=\mathbb 0/math>, : \mathbf_i=-K_i\hat_i. \, and \hat_i corresponds to the predictive estimate \hat_i = \mathbb \mathbf^i, \mathbf^. The Kalman gain equals : L_i = P_iC ^\mathrm T _i(C_iP_iC ^\mathrm T _i + W_i)^, where P_i is determined by the following matrix Riccati difference equation that runs forward in time: : P_ = A_i \left( P_i - P_i C ^\mathrm T _i \left( C_i P_i C ^\mathrm T _i+W_i \right)^ C_i P_i \right) A ^\mathrm T _i+V_i, \qquad P_0=\mathbb left( _0 - \hat_0\right)\left(_0- \hat_0\right)^\mathrm T The feedback gain matrix equals : K_i = (B^\mathrm T_iS_B_i + R_i)^B^\mathrm T_iS_A_i where S_i is determined by the following matrix Riccati difference equation that runs backward in time: : S_i = A^\mathrm T_i \left( S_ - S_B_i \left( B^\mathrm T_iS_B_i+R_i \right)^ B^\mathrm T_i S_ \right) A_i+Q_i, \quad S_N=F. If all the matrices in the problem formulation are time-invariant and if the horizon N tends to infinity the discrete-time LQG controller becomes time-invariant. In that case the matrix Riccati difference equations may be replaced by their associated discrete-time algebraic Riccati equations. These determine the time-invariant linear–quadratic estimator and the time-invariant
linear–quadratic regulator The theory of optimal control is concerned with operating a dynamic system at minimum cost. The case where the system dynamics are described by a set of linear differential equations and the cost is described by a quadratic function is called t ...
in discrete-time. To keep the costs finite instead of J one has to consider J/N in this case.


See also

* Stochastic control *
Witsenhausen's counterexample Witsenhausen's counterexample, shown in the figure below, is a deceptively simple toy problem in decentralized stochastic control. It was formulated by Hans Witsenhausen in 1968. It is a counterexample to a natural conjecture that one can general ...


References


Further reading

* {{DEFAULTSORT:Linear-quadratic-Gaussian control Control loop theory Optimal control Stochastic control