|
I have trained and RBM using the matlab Deep learning Toolbox by rasmusbergpalm. I have modified the rbmtrain.m code to do CDk Here is the original code.
Changed Code.
Architectures: [784 100 100 100 10]
|
|
I suppose that you have plotted the reconstruction error because likelihood estimation is quite expensive procedure. If this is the case then: One way to look at CD training of RBMs is as a process that lowers the free energy of the data v in the positive phase and raises the free energy of the kth state of a Gibbs chain started at the data v. If the negative samples vk have the same distribution as the data v, the loss function will be zero on average, meaning that we have successfully learned the data distribution p(v). However, since the negative phase Gibbs chain starts at the data, this objective function will also be small if the chain is mixing slowly, vk will be very close to v. So reconstruction error can be small but at the same time likelihood is small too. Reconstruction error is not what you actually optimize. If you run you Markov chain long enough then sample is from distribution close to true distribution. So CD-k better than CD-1 for estimating true gradient of likelihood function w.r.t. parameters of RBM. Thanks for the answer. The errors plotted are the classification errors(on MNIST) after fine-tuning.
(Jun 26 '13 at 18:17)
noname
|
|
I have same problem. For all k values error is same. Is that true? |

What error are you plotting?