We shall now summarise the algorithm. Figure 3.6
schematically shows the structure of the network. The network has a
layer of neurons, which compute projections p of the input
in the direction of their weight vectors
.
The weight vectors are normalised to unity.
Figure 3.6:
The network has a layer of linear neurons and a mechanism which
assings winning strengths for each neuron.
The efficiency of the algorithm is based on selecting for each input a
set of winners. The selection is based on winning ratios which
are defined in equation 3.6.
The ratios are combined according to equation 3.7, which
has to be solved numerically since the preliminary winning strengths
appear in both sides of the equation. On the other hand, the
computation needs to be done for the set of winners only. The
preliminary winning strengths
have been derived with the
following competition mechanism in mind: the stronger is the
correlation
between the weight vectors of neurons i and j
the stronger is the competition between them, and if
there is no mutual competition between neurons i and j.
Equations 3.10, 3.11 and 3.12 are used to
compute the final winning strengths from the preliminary
winning strengths
. The final outputs
are obtained by
multiplying the projections with the winning strengths.
These equations have been chosen so that if we use the following reconstruction mapping for the input
then the reconstruction error is close to minimum.
The learning rule for the network is a combination of minimising the
reconstruction error and using neighbourhoods. It is advisable to use
the batch version of the learning rule (equation 3.18),
because the computation of correlations between the weight vectors
( ) is quite a heavy operation.