etc.
•
Then freeze the first layer of weights
in both directions and learn the
remaining weights (still tied
together).
–
This is equivalent to learning
another RBM, using the
aggregated posterior distribution
of h0 as the data.
h
2
v
2
h
1
v
1
v
1
h
0
h
0
v
0