next up previous contents
Next: Tasks Up: Well-known graphical models Previous: Hidden Markov models   Contents


State-space models

Linear state-space models (see textbook by Chen, 1999) share the same structure as hidden Markov models but now the states $ \mathbf{s}(t)$ and observations $ \mathbf{x}(t)$ are continuous valued vectors. The conditional probabilities are represented as a linear mapping with additive Gaussian noise:

$\displaystyle \mathbf{s}(t)$ $\displaystyle = \mathbf{B}\mathbf{s}(t-1) + \mathbf{n}_s(t)$ (3.20)
$\displaystyle \mathbf{x}(t)$ $\displaystyle = \mathbf{A}\mathbf{s}(t) + \mathbf{n}_x(t)$ (3.21)

Note that Equation (3.22) is exactly the same as (3.14), that is, state-space models can be seen as a dynamic extension of factor analysis. The dynamics in Equation (3.21) correspond to linear dynamical systems discretised in the time domain.

Again, the belief propagation algorithm has another name, and it was originally derived from a different starting point probably by a Danish statistician T.N. Thiele in 1880 and later popularised by Kalman (1960) (see also the textbook by Anderson and Moore, 1979). In the context of state-space models, belief propagation is known as the Kalman smoother. It is popular in many applications fields, including econometrics (Engle and Watson, 1987), radar tracking (Chui and Chen, 1991), control systems (Maybeck, 1979), signal processing, navigation, and robotics.

In control systems, dynamics can be affected by control inputs $ \mathbf{u}(t)$. Equation (3.21) for dynamics is replaced by

$\displaystyle \mathbf{s}(t)$ $\displaystyle = \mathbf{B}\mathbf{s}(t-1) + \mathbf{C}\mathbf{u}(t) + \mathbf{n}_s(t).$ (3.22)

Note that the control inputs $ \mathbf{u}(t)$ are coming from outside the generative model. One possibility is to use feedback control (see textbook by Doyle et al., 1992),

$\displaystyle \mathbf{u}(t)$ $\displaystyle = -\mathbf{K}\mathbf{x}(t-1) + \mathbf{r}(t),$ (3.23)

but note that $ \mathbf{K}$ and $ \mathbf{r}(t)$ should be chosen so as to accomplish the control goal, not by inference or learning.

Section 4.3 studies an extension to nonlinear state-space models where the linear mappings $ \mathbf{A}$ and $ \mathbf{B}$ are replaced by multi-layer perceptron networks. Section 4.3.2 and Publication IV study control in nonlinear state-space models.


next up previous contents
Next: Tasks Up: Well-known graphical models Previous: Hidden Markov models   Contents
Tapani Raiko 2006-11-21