Revision history[back]
click to hide/show revision 1
Revision n. 1

Jul 07 '13 at 07:51

Midas's gravatar image

Midas
42151017

Hi, it seems you definitely have something wrong with your math. Visible units are continuous and alt text

To begin with energy function that mentioned in Hinton's practical guide is

alt text

So if you do some math(for detail look at this) you will have:

alt text

Now visible units is continuous random variable:

alt text

Contrastive divergence learning algorithm will be the same. You just sample visible units from normal distribution.

Also another parameterization of energy function can be done (look at thisfor more details)

alt text

Now mean of visible unit does not scale by standard deviation.

alt text

As long as learning of standard deviation is not quite stable, common practice is to fix sigma to 1 and normalize your input to zero mean and standart deviation 1. But sometimes it is a good idea to learn standard deviation by using different parameterization of the variance parameters alt text Since we learn log-variances alt text, alt text is naturally constrained to stay positive.

Bernoulli-Bernoulli RBM can model arbitrary binary distribution but a Gaussian-Bernoulli RBM is extremely limited in the class of distributions it can represent. I extremely recommend you to read 4.1 Conceptual Understanding of Gaussian-Binary RBMs from link text thesis.

So I actually do not understand why do you need both hidden and visible layers continuous?

Articles that may be useful:

  1. Learning Multiple Layers of Features from Tiny Images
  2. Learning Natural Image Statistics with Gaussian-Binary Restricted Boltzmann Machines
  3. Improved Learning of Gaussian-Bernoulli Restricted Boltzmann Machines
  4. An Analysis of Gaussian-Binary Restricted Boltzmann Machines for Natural Images
click to hide/show revision 2
Revision n. 2

Jul 07 '13 at 07:51

Midas's gravatar image

Midas
42151017

Hi, it seems you definitely have something wrong with your math. Visible units are continuous and alt text

To begin with energy function that mentioned in Hinton's practical guide is

alt text

So if you do some math(for detail look at this) you will have:

alt text

Now visible units is continuous random variable:

alt text

Contrastive divergence learning algorithm will be the same. You just sample visible units from normal distribution.

Also another parameterization of energy function can be done (look at thisfor more details)

alt text

Now mean of visible unit does not scale by standard deviation.

alt text

As long as learning of standard deviation is not quite stable, common practice is to fix sigma to 1 and normalize your input to zero mean and standart deviation 1. But sometimes it is a good idea to learn standard deviation by using different parameterization of the variance parameters alt text Since we learn log-variances alt text, alt text is naturally constrained to stay positive.

Bernoulli-Bernoulli RBM can model arbitrary binary distribution but a Gaussian-Bernoulli RBM is extremely limited in the class of distributions it can represent. I extremely recommend you to read 4.1 Conceptual Understanding of Gaussian-Binary RBMs from link textthis thesis.

So I actually do not understand why do you need both hidden and visible layers continuous?

Articles that may be useful:

  1. Learning Multiple Layers of Features from Tiny Images
  2. Learning Natural Image Statistics with Gaussian-Binary Restricted Boltzmann Machines
  3. Improved Learning of Gaussian-Bernoulli Restricted Boltzmann Machines
  4. An Analysis of Gaussian-Binary Restricted Boltzmann Machines for Natural Images
click to hide/show revision 3
Revision n. 3

Jul 07 '13 at 07:57

Midas's gravatar image

Midas
42151017

Hi, it seems you definitely have something wrong with your math. Visible units are continuous and alt textalt text

To begin with energy function that mentioned in Hinton's practical guide is

alt text

So if you do some math(for detail look at this) you will have:

alt text

Now visible units is continuous random variable:

alt text

Contrastive divergence learning algorithm will be the same. You just sample visible units from normal distribution.

Also another parameterization of energy function can be done (look at thisfor more details)

alt text

Now mean of visible unit does not scale by standard deviation.

alt text

As long as learning of standard deviation is not quite stable, common practice is to fix sigma to 1 and normalize your input to zero mean and standart deviation 1. But sometimes it is a good idea to learn standard deviation by using different parameterization of the variance parameters alt text Since we learn log-variances alt text, alt text is naturally constrained to stay positive.

Bernoulli-Bernoulli RBM can model arbitrary binary distribution but a Gaussian-Bernoulli RBM is extremely limited in the class of distributions it can represent. I extremely recommend you to read 4.1 Conceptual Understanding of Gaussian-Binary RBMs from this thesis.

So I actually do not understand why do you need both hidden and visible layers continuous?

Articles that may be useful:

  1. Learning Multiple Layers of Features from Tiny Images
  2. Learning Natural Image Statistics with Gaussian-Binary Restricted Boltzmann Machines
  3. Improved Learning of Gaussian-Bernoulli Restricted Boltzmann Machines
  4. An Analysis of Gaussian-Binary Restricted Boltzmann Machines for Natural Images
click to hide/show revision 4
Revision n. 4

Jul 07 '13 at 07:57

Midas's gravatar image

Midas
42151017

Hi, it seems you definitely have something wrong with your math. Visible units are continuous and alt textalt text

To begin with energy function that mentioned in Hinton's practical guide is

alt text

So if you do some math(for detail look at this) you will have:

alt text

Now visible units is continuous random variable:

alt text

Contrastive divergence learning algorithm will be the same. You just sample visible units from normal distribution.

Also another parameterization of energy function can be done (look at thisfor more details)

alt text

Now mean of visible unit does not scale by standard deviation.

alt text

As long as learning of standard deviation is not quite stable, common practice is to fix sigma to 1 and normalize your input to zero mean and standart deviation 1. But sometimes it is a good idea to learn standard deviation by using different parameterization of the variance parameters alt text Since we learn log-variances alt text, alt text is naturally constrained to stay positive.

Bernoulli-Bernoulli RBM can model arbitrary binary distribution but a Gaussian-Bernoulli RBM is extremely limited in the class of distributions it can represent. I extremely recommend you to read 4.1 Conceptual Understanding of Gaussian-Binary RBMs from this thesis.

So I actually do not understand why do you need both hidden and visible layers continuous?

Articles that may be useful:

  1. Learning Multiple Layers of Features from Tiny Images
  2. Learning Natural Image Statistics with Gaussian-Binary Restricted Boltzmann Machines
  3. Improved Learning of Gaussian-Bernoulli Restricted Boltzmann Machines
  4. An Analysis of Gaussian-Binary Restricted Boltzmann Machines for Natural Images
click to hide/show revision 5
Revision n. 5

Jul 07 '13 at 08:03

Midas's gravatar image

Midas
42151017

Hi, it seems you definitely have something wrong with your math. Visible units are continuous and alt textalt text

To begin with energy function that mentioned in Hinton's practical guide is

alt text

So if you do some math(for detail look at this) you will have:

alt text

Now visible units is continuous random variable:

alt text

Contrastive divergence learning algorithm will be the same. You just sample visible units from normal distribution.

Also another parameterization of energy function can be done (look at thisfor more details)

alt text

Now mean of visible unit does not scale by standard deviation.

alt text

As long as learning of standard deviation is not quite stable, common practice is to fix sigma to 1 and normalize your input to zero mean and standart deviation 1. But sometimes it is a good idea to learn standard deviation by using different parameterization of the variance parameters alt text Since we learn log-variances alt text, alt text is naturally constrained to stay positive.

Bernoulli-Bernoulli RBM can model arbitrary binary distribution but a Gaussian-Bernoulli RBM is extremely limited in the class of distributions it can represent. I extremely recommend you to read 4.1 Conceptual Understanding of Gaussian-Binary RBMs from this thesis.

So I actually do not understand why do you need both hidden and visible layers continuous?

Articles that may be useful:

  1. Learning Multiple Layers of Features from Tiny Images
  2. Learning Natural Image Statistics with Gaussian-Binary Restricted Boltzmann Machines
  3. Improved Learning of Gaussian-Bernoulli Restricted Boltzmann Machines
  4. An Analysis of Gaussian-Binary Restricted Boltzmann Machines for Natural Images
click to hide/show revision 6
Revision n. 6

Jul 07 '13 at 08:03

Midas's gravatar image

Midas
42151017

Hi, it seems you definitely have something wrong with your math. Visible units are continuous and alt text

To begin with energy function that mentioned in Hinton's practical guide is

alt text

So if you do some math(for detail look at this) you will have:

alt text

Now visible units is continuous random variable:

alt text

Contrastive divergence learning algorithm will be the same. You just sample visible units from normal distribution.

Also another parameterization of energy function can be done (look at this for more details)

alt text

Now mean of visible unit does not scale by standard deviation.

alt text

As long as learning of standard deviation is not quite stable, common practice is to fix sigma to 1 and normalize your input to zero mean and standart deviation 1. But sometimes it is a good idea to learn standard deviation by using different parameterization of the variance parameters alt text Since we learn log-variances alt text, alt text is naturally constrained to stay positive.

Bernoulli-Bernoulli RBM can model arbitrary binary distribution but a Gaussian-Bernoulli RBM is extremely limited in the class of distributions it can represent. I extremely recommend you to read 4.1 Conceptual Understanding of Gaussian-Binary RBMs from this thesis.

So I actually do not understand why do you need both hidden and visible layers continuous?

Articles that may be useful:

  1. Learning Multiple Layers of Features from Tiny Images
  2. Learning Natural Image Statistics with Gaussian-Binary Restricted Boltzmann Machines
  3. Improved Learning of Gaussian-Bernoulli Restricted Boltzmann Machines
  4. An Analysis of Gaussian-Binary Restricted Boltzmann Machines for Natural Images
click to hide/show revision 7
Revision n. 7

Jul 07 '13 at 08:05

Midas's gravatar image

Midas
42151017

Hi, it seems you definitely have something wrong with your math. Visible units are continuous and alt text

To begin with energy function that mentioned in Hinton's practical guide is

alt text

So if you do some math(for detail look at this) you will have:

alt text

Now visible units is continuous random variable:

alt text

Contrastive divergence learning algorithm will be the same. You just sample visible units from normal distribution.

Also another parameterization of energy function can be done (look at this for more details)

alt text

Now mean of visible unit does not scale by standard deviation.

alt text

As long as learning of standard deviation is not quite stable, common practice is to fix sigma to 1 and normalize your input to zero mean and standart deviation 1. But sometimes it is a good idea to learn standard deviation by using different parameterization of the variance parameters alt text Since we learn log-variances alt text, alt text is naturally constrained to stay positive.

Bernoulli-Bernoulli RBM can model arbitrary binary distribution but a Gaussian-Bernoulli RBM is extremely limited in the class of distributions it can represent. I extremely recommend you to read 4.1 Conceptual Understanding of Gaussian-Binary RBMs from this thesis.

So I actually do not understand why do you need both hidden and visible layers continuous?

Articles that may be useful:

  1. Learning Multiple Layers of Features from Tiny Images
  2. Learning Natural Image Statistics with Gaussian-Binary Restricted Boltzmann Machines
  3. Improved Learning of Gaussian-Bernoulli Restricted Boltzmann Machines
  4. An Analysis of Gaussian-Binary Restricted Boltzmann Machines for Natural Images
click to hide/show revision 8
Revision n. 8

Jul 07 '13 at 08:08

Midas's gravatar image

Midas
42151017

Hi, it seems you definitely have something wrong with your math. Visible units are continuous and alt text

To begin with energy function that mentioned in Hinton's practical guide is

alt text

So if you do some math(for detail look at this) you will have:

alt text

Now visible units is continuous random variable:

alt text

Contrastive divergence learning algorithm will be the same. You just sample visible units from normal distribution.

Also another parameterization of energy function can be done (look at this for more details)

alt text

Now mean of visible unit does not scale by standard deviation.

alt text

As long as learning of standard deviation is not quite stable, common practice is to fix sigma to 1 and normalize your input to zero mean and standart deviation 1. But sometimes it is a good idea to learn standard deviation by using different parameterization of the variance parameters alt text Since we learn log-variances alt text, alt text is naturally constrained to stay positive.

Bernoulli-Bernoulli RBM can model arbitrary binary distribution but a Gaussian-Bernoulli RBM is extremely limited in the class of distributions it can represent. I extremely recommend you to read 4.1 Conceptual Understanding of Gaussian-Binary RBMs from this thesis.

So I actually do not understand why do you need both hidden and visible layers continuous?

Articles that may be useful:

  1. Learning Multiple Layers of Features from Tiny Images
  2. Learning Natural Image Statistics with Gaussian-Binary Restricted Boltzmann Machines
  3. Improved Learning of Gaussian-Bernoulli Restricted Boltzmann Machines
  4. An Analysis of Gaussian-Binary Restricted Boltzmann Machines for Natural Images
click to hide/show revision 9
Revision n. 9

Jul 07 '13 at 10:03

Midas's gravatar image

Midas
42151017

Hi, it seems you definitely have something wrong with your math. Visible units are continuous and alt text

To begin with energy function that mentioned in Hinton's practical guide is

alt text

So if you do some math(for detail look at this) you will have:

alt text

Now visible units is continuous random variable:

alt text

Contrastive divergence learning algorithm will be the same. You just sample visible units from normal distribution.

Also another parameterization of energy function can be done (look at this for more details)

alt text

Now mean of visible unit does not scale by standard deviation.

alt text

As long as learning of standard deviation is not quite stable, common practice is to fix sigma to 1 and normalize your input to zero mean and standart deviation 1. But sometimes it is a good idea to learn standard deviation by using different parameterization of the variance parameters alt text Since we learn log-variances alt text, alt text is naturally constrained to stay positive.

Bernoulli-Bernoulli RBM can model arbitrary binary distribution but a Gaussian-Bernoulli RBM is extremely limited limited in the class of distributions it can represent. I extremely recommend you to read 4.1 Conceptual Understanding of Gaussian-Binary RBMs from this thesis.

So I actually do not understand why do you need both hidden and visible layers continuous?

Articles that may be useful:

  1. Learning Multiple Layers of Features from Tiny Images
  2. Learning Natural Image Statistics with Gaussian-Binary Restricted Boltzmann Machines
  3. Improved Learning of Gaussian-Bernoulli Restricted Boltzmann Machines
  4. An Analysis of Gaussian-Binary Restricted Boltzmann Machines for Natural Images
click to hide/show revision 10
Revision n. 10

Jul 08 '13 at 05:39

Midas's gravatar image

Midas
42151017

Hi, it seems you definitely have something wrong with your math. Visible units are continuous and alt text

To begin with with, energy function that mentioned in Hinton's practical guide is

alt text

So if you do some math(for detail look at this) you will have:

alt text

Now visible units is continuous random variable:

alt text

Contrastive divergence learning algorithm will be the same. You just sample visible units from normal distribution.

Also another parameterization of energy function can be done (look at this for more details)

alt text

Now mean of visible unit does not scale by standard deviation.

alt text

As long as learning of standard deviation is not quite stable, common practice is to fix sigma to 1 and normalize your input to zero mean and standart deviation 1. But sometimes it is a good idea to learn standard deviation by using different parameterization of the variance parameters alt text Since we learn log-variances alt text, alt text is naturally constrained to stay positive.

Bernoulli-Bernoulli RBM can model arbitrary binary distribution but a Gaussian-Bernoulli RBM is extremely limited in the class of distributions it can represent. I extremely recommend you to read 4.1 Conceptual Understanding of Gaussian-Binary RBMs from this thesis.

So I actually do not understand why do you need both hidden and visible layers continuous?

Articles that may be useful:

  1. Learning Multiple Layers of Features from Tiny Images
  2. Learning Natural Image Statistics with Gaussian-Binary Restricted Boltzmann Machines
  3. Improved Learning of Gaussian-Bernoulli Restricted Boltzmann Machines
  4. An Analysis of Gaussian-Binary Restricted Boltzmann Machines for Natural Images

powered by OSQA

User submitted content is under Creative Commons: Attribution - Share Alike; Other things copyright (C) 2010, MetaOptimize LLC.