next up previous contents
Next: Initialisation Up: Linear and Nonlinear Factor Previous: Ensemble Learning

The Missing Values

For efficiency reasons in Matlab code, the data elements, that are missing, are set to any value (e.g. zero) and the fact that they are missing is otherwise taken account of.

The approximation $Q(\mathbf{s},\boldsymbol{\theta})$ is assumed to be a Gaussian density with a diagonal covariance matrix. This simplifies the cost function CKL given by (7) into expectations of many simple terms. The terms concerning the reconstruction error of the missing values are not included in the cost function.

The partial derivatives of CKL with respect to parameters $\boldsymbol{\theta}$ are used for adaptation of $\boldsymbol{\theta}$. The effect of i(t) on CKL is transferred through the noise vectors  e(t) which are also the reconstruction errors as seen in formula (3). When the variance of noise for the missing elements is set to infinity, the algorithm effectively ignores the reconstruction errors of the missing values.

 \begin{displaymath}\xi_{k}(t)^{-1}=\frac{i_{k}(t)}{\xi_{k}}
\end{displaymath} (8)

The factors are thus estimated based only on the available observations (Figure 2).

The reconstruction of data is obtained by the mapping f from the estimated factors. There is an analogy with the SOM. Factor vector scorresponds to the winning map unit in the SOM and f(s)corresponds to the model vector of the winner.


  
Figure 2: Left figure: MLP net as a generative model, Right figure: Error propagates only from observed values to factors.
\begin{figure}
\begin{tabular}{cccccc}
\epsfig{file=gene1b.eps,width=0.48\textwidth} &
\epsfig{file=gene2b.eps,width=0.48\textwidth} \end{tabular} \end{figure}


next up previous contents
Next: Initialisation Up: Linear and Nonlinear Factor Previous: Ensemble Learning
Tapani Raiko
2001-09-26