1
1

Dear Group,

If we read Rabiner's paper on HMM, suggested as standard tutorial of it available in http://www.cs.ubc.ca/~murphyk/Bayes/rabiner.pdf suggests that HMM training is done by Forward-Backward Probability, which if I read it correctly acts as EM algorithm. But I am not getting how to do EM with Forward Backward. If any one kindly suggest me any net based tutorial preferably with some examples, etc. it may be of big help.

Best Regards, Subhabrata.

asked Apr 02 '11 at 15:22

Subhabrata%20Banerjee's gravatar image

Subhabrata Banerjee
40222224


5 Answers:

Consider training an HMM where you have a training set with both state and observation labels. You train the parameters to maximize likelihood, which ends up with a simple procedure for setting parameters from counts, like with Naive Bayes.

Now consider a scenario where your training set does not have observation labels. You can use Forward-Backward to create "artificial training data" and use likelihood maximization as in the scenario above. IE, set your parameters to some initial numbers and get observation labels using forward backward. Those labels will be "soft", and you can treat them as partial counts, ie if forward backward gives 0.5 of state 1 at some position, that will contribute 1/2 to the count of state1 observations. Now you have your training data with labels, and set parameters as in the fully observed case, by maximizing likelihood. If you run forward-backward after these steps, you get different set of "training data", so you repeat the whole procedure until convergence.

A good introduction to Forward-Backward training is Eisner's "An Interactive Spreadsheet for Teaching the Forward-Backward Algorithm"

answered Apr 02 '11 at 20:26

Yaroslav%20Bulatov's gravatar image

Yaroslav Bulatov
2333214365

I can recommend you this lecture notes by a student of Andrew NG

http://www.stanford.edu/class/cs229/section/cs229-hmm.pdf

Here he explains quite well the different algorithms you may use in HMM to get the transition probabilities or the effect probabilities, or even both.

Hope it is of any use

answered Apr 03 '11 at 11:28

Leon%20Palafox's gravatar image

Leon Palafox ♦
40857194128

Forward-Backward algorithm is not EM. It is a dynamic programming approach to calculate the probabilities p(x_1,...x_n, z_n) and p(z_n| x_n+1,...) . These probabilities are needed in the M-step of the EM algorithm for HMM. A very readable explanation is in chapter 13 of the "Pattern Recognition and Machine Learning" of Chris Bishop. You can read chapter 9 to understand EM also.

answered Apr 02 '11 at 16:29

WeakLearner's gravatar image

WeakLearner
51126

Thanks Sir, for kindly answering it. Chris Bishop I may not be able to access, but if you can kindly suggest me an internet based reading material. Best Regards, Subhabrata.

(Apr 02 '11 at 16:48) Subhabrata Banerjee

just to add to this, the method presented in Rabiner's paper to train the parameters is the Baum-Welch algorithm, and as WeakLearner rightly said, the BW algorithm makes use of the Forward-Backward algorithm. It is the BW algorithm that is a special case of the general EM algorithm.

(Apr 05 '11 at 18:43) Avneesh Saluja

Dear Sir, Thank you for your kind note. Everytime you send some note they are just so nice. Some other colleagues in the room also took lot of pain and their valuable times to make me understand the issue. I found another back note of mine of HMM of Biological Sequence Analysis by Jiawei Han and Micheline Kamber also a very nice and lucid one. Thank you all for your kind time and valuable effort. Wishing A Happy Day Ahead, Best Regards, Subhabrata Banerjee.

answered Apr 03 '11 at 16:09

Subhabrata%20Banerjee's gravatar image

Subhabrata Banerjee
40222224

Dear Sir, Thanks for your kind reply. I got it now. Best Regards, Subhabrata.

answered Apr 03 '11 at 03:18

Subhabrata%20Banerjee's gravatar image

Subhabrata Banerjee
40222224

Your answer
toggle preview

powered by OSQA

User submitted content is under Creative Commons: Attribution - Share Alike; Other things copyright (C) 2010, MetaOptimize LLC.