next up previous
Next: INTRODUCTION

BAYESIAN LEARNING OF LOGICAL HIDDEN MARKOV MODELS

T. Raiko1,2 - K. Kersting1 - J. Karhunen2 - L. De Raedt1
minipage[b]7cm center 1
Institute for Computer Science
Machine Learning Lab
Albert-Ludwigs University of Freiburg
Georges-Koehler-Allee, Building 079
79112 Freiburg, Germany minipage[b]7cm center 2Helsinki University of Technology
Laboratory of Computer and
Information Science,
P.O. Box 5400,
02015 HUT, Finland

Abstract:

Logical hidden Markov models (LOHMMs) are a generalisation of hidden Markov models to analyze sequences of logical atoms. Transitions are factorized into two steps, selecting an atom and instantiating the variables. Unification is used to share information among states, and between states and observations. In this paper, we show how LOHMMs can be learned using Bayesian methods. Some estimators are compared and parameter estimation is tested with synthetic data.





Tapani Raiko 2003-07-09