next up previous contents
Next: HMM/ANN hybrids. Up: Description of the model Previous: Probability of the sequences.

Output density models.

The output probability density models aim at providing the conditional probabilities for the output features of the system associated with each state. The different ways to generate the output probabilities have been used to classify the HMMs.

A traditional way is to estimate a parametric density function (often a Gaussian) for each state [Baker, 1975,Jelinek, 1976]. Because a simple function is clearly insufficient to represent the density of, for example, a phoneme, the model has been extended to a linear combination of PDFs for (16) [Liporace, 1982,Juang, 1985]  
 \begin{displaymath}
b_i(\mbox{\boldmath$O$}_t) = \sum_{m=1}^M c_{im} b_{im}(\mbox{\boldmath$O$}_t) \:,\end{displaymath} (19)
where the mixture weights satisfy the conditions $c_{im} \geq 0$ and $\sum_{m=1}^M c_{im} = 1$for all states i. The mixture components can be, for example, Gaussian PDFs $b_{im}(\mbox{\boldmath$O$}_t) \sim N (\mu_{im},\Sigma_{im}) \:$,where $\mu_{im} \in {\cal R}^D$ and $\Sigma_{im} \in {\cal R}^{D \times D}$ are the mean vector and the covariance matrix, respectively. HMMs which have parametric output density functions are generally called continuous density HMMs (CDHMMs) and those having mixture density functions mixture density HMMs (MDHMMs).

Another traditional way is to use vector quantization (VQ) [Makhoul et al., 1985] to transform the features into a set of output symbols and then estimate their discrete probability distribution in each state [Rabiner et al., 1983].

By combining the gains of the continuous and discrete models, good results have been obtained by a class of methods called semi-continuous HMMs (SCHMMs) [Huang and Jack, 1989,Paul, 1989,Bellegarda and Nahamoo, 1990] and multi-codebook vector quantization HMMs [Segura et al., 1994,Peinado et al., 1994]. In SCHMMs a single set of Gaussians is shared between all states of all HMMs so that it is not necessary to train individual sets of Gaussians for each state. By forming a large set of Gaussians, which can be regarded actually as a codebook of Gaussian densities, the number of the mixtures used for each state can be much larger without increasing the total number of parameters to be estimated. Also the covariance matrices are often shared between the Gaussians as a trade-off to be able to increase the size of the codebook. In practice, SCHMMs are easier to train than CDHMMs, because the training of the codebook can be separated from the segmentation and even be trained mostly with unlabeled data (e.g., see Publication 1).

The SCHMMs offer an extension to the discrete observation HMMs, since it enhances the VQ performance by smoothing the quantization effects with weighted average of the probabilities given by several nearest units instead of only the BMU. The weights are inversely proportional to the distance of the units. This weighting leads actually to density model family known as the kernel density estimators or Parzen estimators [Parzen, 1962] with suitably chosen kernel functions. In the broad sense, as applied in this work, all the SCHMMs and CDHMMs, where the density model consists of a finite set of weighted multivariate unimodal kernel functions, belong to the mixture density HMMs (MDHMMs).

An illustration of a MDHMM and its parameters is shown in Figure 1. The temporal structure for this phoneme model is a typical simple uni-directional chain without skips.


  
Figure 1: In phoneme-wise tied MDHMMs the same mixtures are used for the states representing the same phoneme. The model is defined by the set of transition probabilities aij, mixture weights cik and mixture densities bk(x). The output probability of state Ci at time t is computed as weighted average of the values provided by the mixture Gaussians for the current observation vector xt. The observed feature vector is here a scalar only for illustrative purposes.
\begin{figure}
 \centerline{
 \renewcommand {
\setlength 
 }[2]{ ...

Publication 3 as several other recent studies, e.g. [Young and Woodland, 1994,Digalakis et al., 1996], suggest to improve the performance of SCHMMs by tying some parameters for several similar states. The problem in practice is that the finite training data cannot provide reliable weights for the whole Gaussian codebook for every HMM state. For large vocabulary recognition tasks the tying of parameters is often crucial for developing a required amount of context dependent phoneme models. A common choice for tying mixtures is to have a separate Gaussian codebook for each HMM, which can then be called as a phoneme-wise tied MDHMMs. In Publication 3 there is a brief comparison of different HMM classes.


next up previous contents
Next: HMM/ANN hybrids. Up: Description of the model Previous: Probability of the sequences.
Mikko Kurimo
11/7/1997