Next: Why LVQ?
Up: Minimization of classification errors
Previous: Minimization of classification errors
In LVQ and stochastic gradient descent algorithms
the number of expected classification errors is
minimized using misclassifications or near-misses
observed in stochastic training samples.
For the small model adjustment steps corresponding to
each such sample,
some continuous measure for the degree of misclassification is useful.
For two-class problems the measure is based
on the difference between the Bayesian class discriminants gi
defined using the distances to the closest references
from the competing classes.
For example, the discriminant for class i can be simply
| |
(10) |
where the index c of the BMU (1) is separately
determined among the set of references of each class.
It is also possible to use an appropriately weighted sum of
contributions of several closest references or
reference sequences as in
[Chang and Juang, 1992,McDermott and Katagiri, 1994].
For decisions between multiple classes (M)
there exist several possibilities.
In
[Juang and Katagiri, 1992]
the misclassification of is measured by
| |
(11) |
which is a continuous extension of the one used in
[Amari, 1967],
where only the discriminant differences
between the correct class and the confusing classes (gi > gk)
are averaged.
In (12) is a positive number to control
the relative effect between larger and smaller discriminants.
In
[Juang and Katagiri, 1992]
definition (12) is used to define a
framework of discriminative training methods called GPD
(Generalized Probabilistic Descent) to minimize the
expected cost of the misclassifications.
In the extreme case of (12),
where ,
only the largest discriminant (class Ci) affects
the misclassification measure
| |
(12) |
This approach leads to the LVQ2 algorithm
[Komori and Katagiri, 1992].
In
[Katagiri et al., 1991]
it is shown that the practically efficient simple LVQ2
actually does approximate well the more complicated gradient
search GPD implementations.
Next: Why LVQ?
Up: Minimization of classification errors
Previous: Minimization of classification errors
Mikko Kurimo
11/7/1997