Continuous time Consider the
state-space representation of a
continuous-time linear dynamic system \begin{align}\dot{\mathbf{x}}(t) &= A(t) \mathbf{x}(t) + B(t) \mathbf{u}(t) + \mathbf{v}(t),\\ \mathbf{y}(t) &= C(t) \mathbf{x}(t) + \mathbf{w}(t),\end{align} where \mathbf{x} represents the vector of state variables of the system, \mathbf{u} the vector of control inputs and \mathbf{y} the vector of measured outputs available for feedback. Both additive white Gaussian system noise \mathbf{v}(t) and additive white Gaussian measurement noise \mathbf{w}(t) affect the system. Given this system the objective is to find the control input history {\mathbf{u}}(t) which at every time t may depend linearly only on the past measurements {\mathbf{y}}(t'), where 0 \leq t' , such that the following cost function is minimized: J = \mathbb{E}\left[{\mathbf{x}^\mathrm T}(T)F{\mathbf{x}}(T)+ \int_{0}^{T} {\mathbf{x}^\mathrm T}(t)Q(t){\mathbf{x}}(t) + {\mathbf{u}^\mathrm T}(t)R(t){\mathbf{u}}(t)\,dt \right], F \ge 0,\quad Q(t) \ge 0,\quad R(t) > 0, where \mathbb{E} denotes the
expected value. The final time (horizon) {\mathbf{}}T may be either finite or infinite. If the horizon tends to infinity the first term {\mathbf{x}}^\mathrm T(T)F{\mathbf{x}}(T) of the cost function becomes negligible and irrelevant to the problem. Also to keep the costs finite the cost function has to be taken to be J/T. The LQG controller that solves the LQG control problem is specified by the following equations: \dot{\hat{\mathbf{x}}}(t) = A(t)\hat{\mathbf{x}}(t) + B(t){\mathbf{u}}(t)+L(t) \left( {\mathbf{y}}(t)-C(t)\hat{\mathbf{x}}(t) \right), \quad \hat{\mathbf{x}}(0) = \mathbb{E} \left[ {\mathbf{x}}(0) \right], \mathbf{u}(t)= -K(t) \hat{\mathbf{x}}(t). The matrix L(t) is called the
Kalman gain of the associated
Kalman filter represented by the first equation. At each time t this filter generates estimates \hat{\mathbf{x}}(t) of the state {\mathbf{x}}(t) using the past measurements and inputs. The Kalman gain L(t) is computed from the matrices A(t), C(t), the two intensity matrices V(t), W(t) associated to the white Gaussian noises \mathbf{v}(t) and \mathbf{w}(t) and finally \mathbb{E}\left[{\mathbf{x}}(0){\mathbf{x}}^\mathrm T(0) \right]. These five matrices determine the Kalman gain through the following associated matrix Riccati
differential equation: \dot{P}(t) = A(t)P(t)+P(t)A^\mathrm T(t)-P(t)C^\mathrm T(t){\mathbf{}}W^{-1}(t) C(t)P(t)+V(t), P(0)= \mathbb{E} \left[{\mathbf{x}}(0){\mathbf{x}}^\mathrm T(0) \right]. Given the solution P(t), 0 \leq t \leq T the Kalman gain equals L(t) = P(t)C^\mathrm T(t)W^{-1}(t). The matrix {\mathbf{}}K(t) is called the
feedback gain matrix. This matrix is determined by the matrices {\mathbf{}}A(t), B(t), Q(t), R(t) and {\mathbf{}}F through the following associated matrix Riccati differential equation: \begin{align} -\dot{S}(t) &= A^\mathrm T(t)S(t)+S(t)A(t)-S(t)B(t)R^{-1}(t)B^\mathrm T(t)S(t)+Q(t),\\ S(T) &= F. \end{align} Given the solution S(t), 0 \leq t \leq T the feedback gain equals K(t) = R^{-1}(t)B^\mathrm T(t)S(t). Observe the similarity of the two matrix Riccati differential equations, the first one running forward in time, the second one running backward in time. This similarity is called
duality. The first matrix Riccati differential equation solves the linear–quadratic estimation problem (LQE). The second matrix Riccati differential equation solves the
linear–quadratic regulator problem (LQR). These problems are dual and together they solve the linear–quadratic–Gaussian control problem (LQG). So the LQG problem separates into the LQE and LQR problem that can be solved independently. Therefore, the LQG problem is called
separable. When A(t), B(t), C(t), Q(t), R(t) and the noise intensity matrices V(t), \mathbf{}W(t) do not depend on t and when {\mathbf{}}T tends to infinity the LQG controller becomes a time-invariant dynamic system. In that case the second matrix Riccati differential equation may be replaced by the associated
algebraic Riccati equation.
Discrete time Since the
discrete-time LQG control problem is similar to the one in continuous-time, the description below focuses on the mathematical equations. The discrete-time linear system equations are \begin{align}\mathbf{x}_{i+1} &= A_i\mathbf{x}_i + B_i \mathbf{u}_i + \mathbf{v}_i,\\ \mathbf{y}_{i} &= C_{i} \mathbf{x}_i + \mathbf{w}_i. \end{align} Here \mathbf{}i represents the discrete time index and \mathbf{v}_{i}, \mathbf{w}_{i} represent discrete-time Gaussian white noise processes with covariance matrices V_{i}, W_{i}, respectively, and are independent of each other. The quadratic cost function to be minimized is J = \mathbb{E}\left[{\mathbf{x}}^\mathrm T_{N}F{\mathbf{x}}_{N}+ \sum_{i=0}^{N-1}( \mathbf{x}_i^\mathrm T Q_i \mathbf{x}_i + \mathbf{u}_i^\mathrm T R_i \mathbf{u}_i )\right], F \ge 0, Q_i \ge 0, R_i > 0. \, The discrete-time LQG controller is \hat{\mathbf{x}}_{i+1}=A_i\hat{\mathbf{x}}_i+B_i{\mathbf{u}}_i+L_{i+1} \left({\mathbf{y}}_{i+1}-C_{i+1} \left\{ A_i \hat{\mathbf{x}}_i + B_i \mathbf{u}_i \right\} \right), \qquad \hat{\mathbf{x}}_0=\mathbb{E}[{\mathbf{x}}_0] \mathbf{u}_i=-K_i\hat{\mathbf{x}}_i. \, and \hat{\mathbf{x}}_i corresponds to the predictive estimate \hat{\mathbf{x}}_i = \mathbb{E}[\mathbf{x}_i|\mathbf{y}^i, \mathbf{u}^{i-1}] . The Kalman gain equals L_i = P_iC ^\mathrm T _i(C_iP_iC ^\mathrm T _i + W_i)^{-1}, where {\mathbf{}}P_i is determined by the following matrix Riccati difference equation that runs forward in time: P_{i+1} = A_i \left( P_i - P_i C ^\mathrm T _i \left( C_i P_i C ^\mathrm T _i+W_i \right)^{-1} C_i P_i \right) A ^\mathrm T _i+V_i, \qquad P_0=\mathbb{E} [\left( {\mathbf{x}}_0 - \hat{\mathbf{x}}_0\right)\left({\mathbf{x}}_0- \hat{\mathbf{x}}_0\right)^\mathrm T]. The feedback gain matrix equals K_i = (B^\mathrm T_iS_{i+1}B_i + R_i)^{-1}B^\mathrm T_iS_{i+1}A_i where S_i is determined by the following matrix Riccati difference equation that runs backward in time: S_i = A^\mathrm T_i \left( S_{i+1} - S_{i+1}B_i \left( B^\mathrm T_iS_{i+1}B_i+R_i \right)^{-1} B^\mathrm T_i S_{i+1} \right) A_i+Q_i, \quad S_N=F. If all the matrices in the problem formulation are time-invariant and if the horizon N tends to infinity the discrete-time LQG controller becomes time-invariant. In that case the matrix Riccati difference equations may be replaced by their associated discrete-time
algebraic Riccati equations. These determine the time-invariant linear–quadratic estimator and the time-invariant
linear–quadratic regulator in discrete-time. To keep the costs finite instead of J one has to consider J/N in this case. ==See also==