next up previous contents
Next: Learning algorithm for the Up: Learning algorithm for the Previous: Learning procedure   Contents


Continuing learning with new data

Continuing the learning process with the old model but new data requires initial estimates for the new hidden states. If the new data is a direct continuation of the old, the predictions of the old states provide a reasonable initial estimate for the new ones and the algorithm can continue the adaptation from there.

If the new data forms an entirely separate sequence, the problem is more difficult. Knowing the model, we can still do much better than starting at random or using the same initialisation as in the very beginning.

One way to find the estimates is to use an auxiliary MLP network to model the inverse of the observation mapping $ \mathbf{f}$ [59]. This MLP can be trained using standard supervised back-propagation with the estimated means of $ \mathbf{s}(t)$ and $ \mathbf{x}(t)$ as training set. Their roles are of course inverted so that $ \mathbf{x}(t)$ are the inputs and $ \mathbf{s}(t)$ the outputs. The auxiliary MLP cannot give perfect estimates for the states $ \mathbf{s}(t)$, but they can usually be adapted very quickly by using the standard learning algorithm to update only the hidden states.


next up previous contents
Next: Learning algorithm for the Up: Learning algorithm for the Previous: Learning procedure   Contents
Antti Honkela 2001-05-30