Hidden Markov models [13] (HMMs) are among the most widely and successfully used tools for the analysis of sequential data. Areas of application include computational biology, user modeling, speech recognition, (stochastic) natural language processing, and robotics. The analyzed sequences consist of symbols that could be named as In a logical sense, the alphabets are propositional. In case the propositional elements are replaced by logical atoms [11], we end up with sequences like . Logical sequences are a natural representation e.g. for the secondary structure of proteins [10] and Unix shell command logs.
If we would like to analyze logical sequences with a propositional hidden Markov model, the sequence has to be propositionalized by e.g. leaving the arguments out and getting or flattening the structure getting . In the former case, we lose a lot of information. In the latter case, the number of parameters becomes very high and we lose the shared information between and . Logical hidden Markov models [10] (LOHMMs), overcome these weaknesses. In this paper, we go into details on how to learn a LOHMM from data.
This paper is organized as follows. In the next Section, we will describe logical hidden Markov models. In Section 3, we discuss learning them in general from Bayesian point of view. Section 4 describes the adaptation of the Baum-Welch re-estimation in more detail. Section 5 shows some synthetic experiments. Subsequently, we list related work.