Make sure you have the most recent version of the notes. I have redesigned Chapter 13. Previously, the notes covered only the Gaussian noise case. However, I think it easier to understand how certainty equivalence if I do it without the Gaussian assumptions, as I have now in Theorem 13.1.
When Gaussian assumption are made then
ˆxt=E[xt∣Wt] is particularly simple. It is a linear function of
ˆx0 and
Wt and can be computed recursively using the Kalman filter.
The name "Kalman filter" refers to the estimation equation (13.12) and takes its name from
Rudolf Kalman (1930 –), who developed it in the years 1958-64. He also coined the terms
controllable and
observable, and gave the criteria that we have seen in previous lectures. The fact that a system is controllable iff the matrix
[B AB ⋯ An−1B] is of full rank is sometimes called
Kalman's criteria. In the
IEEE biography of Kalman it is stated
The Kalman filter, and its later extensions to nonlinear problems, represents perhaps the most widely applied by-product of modern control theory. It has been used in space vehicle navigation and control (e.g. the Apollo vehicle), radar tracking algorithms for ABM applications, process control, and socioeconomic systems.
The theory in this lecture is admittedly quite tricky - partly because the notation. As a test of memory, can you say what roles in the theory are taken by each of these?
xt, ut, A, B, ϵt, yt, C, ηt, ˆxt, Δt, ξt, ζt, R, S, Q, Kt, Πt, N, L, M, Ht, Vt.
You will understand the ideas better once you have worked through the details of a scalar example (in which
n=m=p=1). You do this in
Example Sheet 3 Question 2. When you do this question, start by supposing that
ˆxt=ˆxt−1+ut−1+ht(yt−ˆxt−1), and then find the value of
ht that minimizes the variance of
ˆxt. You can start by subtracting
xt=xt−1+ut−1+3ϵt and using
yt=xt−1+2ηt. You get,
ˆxt+1−xt+1=Δt+1=Δt−3ϵt−htΔt+ht2ηt.
Then square, take the expected value, and minimize the variance of
ˆxt with respect to
ht, to find a formula for
Vt+1 in terms of
Vt.
You will not be asked to reproduce the proof of Theorem 13.1 or 13.3 in examinations. You should simply know that
ˆxt is computed from
ˆxt−1 and
yt in the linear manner specified by (13.12), and that the covariance matrix
Vt satisfies a Riccati equation. You are not expected to memorize Riccati equations.
Notice that the Riccati equation for
Vt, i.e.
Vt=gVt−1 runs in the opposite time direction to the one we had for
Πt in lecture 10, where
Πt−1=fΠt. We are given
V0 and
Πh.