Siddiqi et al 2009 Reduced-Rank Hidden Markov Models

From Cohen Courses
Revision as of 12:16, 10 October 2011 by Jmflanig (talk | contribs)
Jump to navigationJump to search

Citation

Online version

[1]

Summary

This paper introduces Reduced-Rank Hidden Markov Models (RR-HMMs). RR-HMMs are similar to standard HMMs, except the rank of the transition matrix is less than the number of hidden states. Thus the dynamics evolve in a subspace of the hidden state probability space.

Method

Sample singles, doubles, and triples from the from the observed output of the RR-HMM. Then let:

Siddiqi et al 2009 Definition of P.png

The learning algorithm uses a singular value decomposition (SVD) of the correlation matrix between past and future observations. The algorithm is borrowed from Hsu et al 2009, with no change for the reduced-rank case. Learning is O(

Siddiqi et al 2009 Algorithm.png

is the initial state distribution, is the final state distribution, and is the transition matrix when x is observed. Note that denotes the Moore-Penrose pseudo-inverse of the matrix .


Inference can be performed using the model parameters:

Siddiqi et al 2009 Inference.png

Experimental Result

Related Papers

In progress by User:Jmflanig