Hidden Markov models (HMMs) (Rabiner and Juang, 1986) are dynamic Bayesian
networks with a certain structure. Observed data are assumed to be
generated as a function of an unobserved state which is evolving
stochastically in time. The observations are conditioned on
state
. The discrete state
at time
is a latent
variable and the conditional probability for a state transition
does not depend on
. Here we will use a
non-standard formulation where the state transition is conditioned on
the observation as well, that is,
. It is
fairly easy to see that the two approaches are equivalent3.3 in representational capacity. Appendix B of
Publication VII shows that the two approaches are
equivalent in a setting generalised to first-order logic.
Figure 3.3 shows an example of a HMM. There are two bent coins
in a bag, one gives more heads and the other more tails. A person
draws a coin randomly from the bag and tosses it until it lays tails
up. At that point it is put back to the bag and one of the coins is
drawn and tossed again. At time 0, the system is always in the start state (start). Afterwards, it is always either in state coin1 or
coin2 possibly going through an auxiliary state bag. The state
being hidden (or latent) means that one does not know which coin is
used, only the sequence of heads and tails is observed.
![]() ![]() ![]() |
The instance of belief propagation algorithm for HMMs is called the forward-backward algorithm. In the example, this would give the probabilities for which coin was drawn from the bag at each stage. It is also possible to learn the parameters from observations, say, estimate how badly the coins are bent. That can be done using the Baum-Welch algorithm, which is the instance of EM algorithm for HMMs.
Hidden Markov models are very popular for analysing sequential data. Application areas include computational biology (Koski, 2001), speech recognition (Rabiner and Juang, 1986), natural language processing (Charniak, 1993), user modelling (Lane, 1999), and robotics (Meila and Jordan, 1996).
There are several extensions of HMMs such as factorial HMMs by Ghahramani and Jordan (1997), relational Markov models by Anderson et al. (2002), and extension based on tree automata by Frasconi et al. (2002). Section 6.2 and Publication VII introduce an extension of hidden Markov models to first-order logic, which generalises all of the above.