This section describes how a LOHMM can be learned from data using the Bayesian approach.
Let us first separate a LOHMM into its parameters
(the
probabilities) and the structure
(the rest). Let us also assume
a prior which generates a structure
with probability
and parameters for it with probability density
. Learning a
single LOHMM
from data
corresponds to finding a good
representative of the posterior probability mass:
Instead of finding just one LOHMM, one could use an ensemble of them represented with a set of sampled points [5] or a distribution with a simple form [7]. This is out of scope of this paper.
Finding a good representative structure
involves a combinatorial
search in the structure space, where for each structure candidate, the
parameters
need to be estimated. One could first try structures
that resemble good candidates by using inductive logic programming
[12] techniques. Also, the information from
other structures can be used to guide the parameter estimation. This
is further research.
The representative parameters
for a given structure
can be
found using estimation. There are two commonly used estimators: the
maximum a posteriori (map) estimate
and the Bayes
estimate
. They are defined as
One can also use the Bayes estimator componentwise (cB). Each
component is estimated by