Hidden Markov Models
Hidden Markov Models
• However this condition is little relaxed for i.i.d., by considering a Markov model.
• The product rule can be used to express the joint distribution for a sequence of observations
• If we assume that each of the conditional distributions on the RHS is independent of all
previous observations except the most recent, we obtain the first-order Markov chain
• first order Markov model
• A first-order Markov chain of observations {xn}
in which the distribution p(xn|xn−1) of a
particular observation xn is conditioned on the
value of the previous observation xn−1.
• The hidden Markov model can be viewed as a specific instance of the state
space model in which the latent variables are discrete.
• The HMM is widely used in
• speech recognition
• natural language modelling
• handwriting recognition
• analysis of biological sequences such as proteins and DNA
• ....
• the latent variables are the discrete multinomial variables zn describing which component of the
mixture is responsible for generating the corresponding observation xn.
• We allow the probability distribution of zn to depend on the state of the previous latent variable zn−1
through a conditional distribution p(zn|zn−1).
• Suppose we observe gradually rising temperatures over a week: temps = [-2, 0, 2, 4, 6, 8, 10]
• We can then compute the most probable sequence of hidden states, which results into [0 0 0 0 0 1 1] telling us that
the transition from "cold" to "hot" most likely happened between the 5th and 6th days.