|
|
Hi,
I hope that helps. Regards. Thank you for your explanation. It helped a lot. But I am still not sure about this: If I was to use a DBN as a generative model, do I still have to add those labels? Another question is about the “classifier” that you just mentioned. Does it refer to merely one layer of neurons, or does it also include labels?
(May 14 '13 at 12:51)
Chen You
A good generative model doesn't necessarily imply a good discriminative model. Getting a good generative model only implies you were able to model the underlying distribution of the data. In order for the model to be better at classification, one can use the labelled data to fine tune the model. In DBN you can do that with back-prop.
(May 14 '13 at 13:28)
Rakesh Chalasani
1
If you just want to use the network to generate samples with the distribution of the training patterns, then you don't need labels. I'm not sure, but i think that, once trained, if you propagate forward from the input layer to the last hidden layer, and then backpropagate from the last hidden layer back to the input layer, you should get in the input layer the samples you want. Just like a BM but with a multilayer structure. But for that purpose probably a DBM is a better choice. When I say "classifier layer" I'm not being very accurate, sorry. I mean a layer where you can "write" your labels and then train using backprop. What we call an output layer in a classical multilayer perceptron trained for classification. It's just that I'm a little bit lazy and my English is not very good ;) Also, as Rakesh has pointed, the generative pre-training doesn't give you a good classifier. Notice that the objective of such training is not to predict any target, but more or less to make the network "learn the distribution" of the training patterns. However, according to my little experience with this, it is usually a good idea to pre-train a network that will then be trained for a more specific purpose. Maybe you can find something of use here http://arantxa.ii.uam.es/~gaa/. There is a little bit more detailed explanation of these ideas in my master's thesis and also I uploaded a very simple (and not totally finished and tested, I'm afraid) octave code that reproduces some experiments with DBNs and stacked autoencoders. Of course, the papers of Hinton, Bengio, Ng and others are much better, but maybe a master's thesis is easier to follow if you are a beginner with these models.
(May 14 '13 at 14:28)
David Diaz Vico
|
|