Hidden markov model classification based on empirical frequencies of observed symbols
Date
2014Publisher
IFAC SecretariatEdition
IFAC Proceedings Volumes (IFACPapersOnline)Source
IFAC Proceedings Volumes (IFACPapersOnline)IFAC Proceedings Volumes (IFACPapersOnline)
Volume
9Pages
712Google Scholar check
Keyword(s):
Metadata
Show full item recordAbstract
Given a sequence of observations, classification among two known hidden Markov models (HMMs) can be accomplished with a classifier that minimizes the probability of error (i.e., the probability of misclassification) by enforcing the maximum a posteriori probability (MAP) rule. For this MAP classifier, the a priori probability of error (before any observations are made) can be obtained, as a function of the length of the sequence of observations, by summing up the probability of error over all possible observation sequences of the given length, which is a computationally expensive task. In this paper, we obtain an upper bound on the probability of error of the MAP classifier. Our results are based on a suboptimal decision rule that ignores the order with which observations occur and relies solely on the empirical frequencies with which different symbols appear. We describe necessary and sufficient conditions under which this bound on the probability of error decreases exponentially with the length of the observation sequence. Apart from the usefulness of the suboptimal rule in bounding the probability of misclassification, its numerous advantages (such as low computational complexity, reduced storage requirements, and potential applicability to distributed or decentralized decision schemes) could prove a useful alternative to the MAP rule for HMM classification in many applications. © IFAC.
Collections
Cite as
Related items
Showing items related by title, author, creator and subject.

Conference Object
Bounds on the probability of misclassification among hidden Markov models
Keroglou, C.; Hadjicostis, Christoforos N. (2011)Given a sequence of observations, classification among two known hidden Markov models (HMMs) can be accomplished with a classifier that minimizes the probability of error (i.e., the probability of misclassification) by ...

Conference Object
Bound on the probability of HMM misclassification
Keroglou, C.; Hadjicostis, Christoforos N. (2011)In this paper we consider the problem of classifying a sequence of observations among two known hidden Markov models (HMMs). We use a classifier that minimizes the probability of error (i.e., the probability of misclassification), ...

Article
Approximation of Markov Processes by Lower Dimensional Processes via Total Variation Metrics
Tzortzis, I.; Charalambous, Charalambos D.; Charalambous, T.; Hadjicostis, Christoforos N.; Johansson, M. (2017)The aim of this paper is to approximate a FiniteState Markov (FSM) process by another process defined on a lower dimensional state space, called the approximating process, with respect to a total variation distance fidelity ...