Contrastive divergence learning:
A quick way to learn an RBM
Start with a training vector on the
Update all the hidden units in
Update all the visible units in
parallel to get a “reconstruction”.
Update all the hidden units again.
This is not following the gradient of the log likelihood. But it works well.
It is approximately following the gradient of another objective function.