next up previous contents index
Next: 6.1.4 Global existence of Up: 6.1 Finite-horizon LQR problem Previous: 6.1.2 Riccati differential equation   Contents   Index


6.1.3 Value function and optimality

We now proceed to show that the control law (6.12) identified by the maximum principle is globally optimal. We already asked the reader to examine this issue via a direct analysis of the second variation in part b) of Exercise 3.8. Since then, however, we learned a general method for establishing optimality--namely, the sufficient condition for optimality from Section 5.1.4--and it is instructive to see it in action here.

Specialized to our present LQR problem, the HJB equation (5.10) becomes

$\displaystyle -{V}_{t}(t,x)=\inf_{u\in\mathbb{R}^m}\left\{x^TQ(t)x+u^TR(t)u+\left\langle {V}_{x}(t,x),A(t)x+B(t)u\right\rangle \right\}$ (6.15)

and the boundary condition (5.3) reads

$\displaystyle V(t_1,x)= x^TMx.$ (6.16)

Since $ R(t)>0$ , it is easy to see that the infimum of the quadratic function of $ u$ in (6.16) is a minimum and to calculate (similarly to how we arrived at the formula (6.3) earlier) that the minimizing control is

$\displaystyle u=-\frac12 R^{-1}(t)B^T(t){V}_{x}(t,x).$ (6.17)

We substitute this control into (6.16) and, after some term cancellations, bring the HJB equation to the following form:

$\displaystyle -{V}_{t}(t,x)=x^TQ(t)x-\frac 14\left({V}_{x}(t,x)\right)^T\!B(t)R^{-1}(t)B^T(t){V}_{x}(t,x)+\left({V}_{x}(t,x)\right)^T\!A(t)x.$ (6.18)

In order to apply the sufficient condition for optimality proved in Section 5.1.4, we need to find a solution $ V(\cdot,\cdot)$ of (6.19). Then, for the feedback law (6.12) to be optimal, it should match the feedback law given by (6.18) for this $ V$ . (The precise meaning of the last statement is provided by the formula (5.22) on page [*].) This will in turn be true if we have

$\displaystyle \frac12{V}_{x}(t,x)=P(t)x$ (6.19)

for all $ t$ and $ x$ . The equation (6.20) suggests that we should look for a function $ V$ that is defined in terms of $ P$ . Another observation that supports this idea is that in view of (6.11), the boundary condition (6.17) for the HJB equation can be written as

$\displaystyle V(t_1,x)= x^TP(t_1)x.$ (6.20)

Armed with these facts, let us apply a certain amount of ``wishful thinking" in solving the partial differential equation (6.19). Namely, let us try to guess a function $ V$ that satisfies the simple conditions (6.20) and (6.21), and then see if it satisfies the complicated equation (6.19). If it does, then it must be the value function and by the previous reasoning our control (6.12) must be optimal, hence everything will be proved. (In this last step, we are using the fact that (6.12) is a feedback law which does not depend on the initial condition; see the remarks immediately following the proof of the sufficient condition for optimality on page [*].)

For the moment, let us proceed under the assumption (which will be validated momentarily) that $ P(t)$ is symmetric for all $ t$ . A guess for $ V$ is actually fairly obvious, and might have already occurred to the reader:

$\displaystyle V(t,x)=x^TP(t)x.$ (6.21)

This function clearly satisfies (6.21), and since its gradient with respect to $ x$ is $ {V}_{x}(t,x)=2P(t)x
$ we see that (6.20) is also fulfilled. Now, let us check whether the function (6.22) satisfies (6.19). Since $ V\in\mathcal C^1$ , the viscosity solution concept is not needed here. Noting that $ {V}_{t}(t,x)=x^T\dot P(t)x
$ and plugging the two expressions for the partial derivatives of $ V$ into (6.19), we obtain

$\displaystyle -x^T\dot P(t)x=
x^TQ(t)x-x^TP(t)B(t)R^{-1}(t)B^T(t)P(t)x+
{2x^TP(t)A(t)x}
$

or, equivalently,

$\displaystyle -x^T\dot P(t)x=x^T\big(Q(t)- P(t)B(t)R^{-1}(t)B^T(t)P(t)+P(t)A(t)+A^T(t)P(t)\big)x.$ (6.22)

Since $ P(t)$ satisfies the RDE (6.14), it immediately follows that (6.23) is a true identity. We conclude that, indeed, the function (6.22) is the value function (optimal cost-to-go) and the linear feedback law (6.12) is the optimal control. (We already know from Section 6.1.1 that an optimal control must be unique, and we also know that the sufficient condition of Section 5.1.4 guarantees global optimality.)

It is useful to reflect on how we found the optimal control. First, we singled out a candidate optimal control by using the maximum principle. Second, we identified a candidate value function and verified that this function and the candidate control satisfy the sufficient condition for optimality. Thus we followed the typical path outlined in Section 5.1.4. The next exercise takes a closer look at properties of $ P(t)$ and closes a gap that we left in the above argument.


\begin{Exercise}
Let $t\le t_1$\ be an arbitrary time at which the solution $P(t...
...ou prove this
by strengthening one of the standing
assumptions?
\end{Exercise}

It was both insightful and convenient to employ the sufficient condition for optimality in terms of the HJB equation to find the expression for the optimal cost and confirm optimality of the control (6.12). However, having the solution $ P(t)$ of the RDE (6.14) with the boundary condition (6.11) in hand, it is also possible to solve the LQR problem by direct algebraic manipulations without relying on any prior theory.


\begin{Exercise}Confirm the facts that~\eqref{e-oplife} is the unique optimal co...
... is the value function using nothing more than square completion.
\end{Exercise}


next up previous contents index
Next: 6.1.4 Global existence of Up: 6.1 Finite-horizon LQR problem Previous: 6.1.2 Riccati differential equation   Contents   Index
Daniel 2010-12-20