Restricted Boltzmann Machines (RBMS)
Restricted Boltzmann Machines (RBMS)
• The second layer is known as the “hidden” layer. Hidden units are feature
detectors, learning features from the input data
• Each layer has a bias unit with its state always set to on.
• Each node performs computation based on the input to the node and outputs a
result based on a stochastic decision whether or not to transmit data through an
activation.
• The activation computation is based on weights on the connections and the input
values. The initial weights are randomly generated.
Restricted Boltzmann Machine (RBM)
• Connections and weights
• All connections are visible-hidden; none are visible-visible or hidden-
hidden.
• Those circles, or nodes, act like human neurons. They are decision
units.
• Reconstruction Phase
• It is like the forward pass but in the opposite direction.
• h1: hidden layer vector, v0: input vector, W: weight matrix, a: hidden layer bias vector, S:
sigmoid activation function, b: visible layer bias
• The di erence (v0-v1) can be considered as the reconstruction error that we need to reduce
in subsequent steps of the training process. The weights are adjusted in each iteration to
minimize this error.
ff
Training RBM
Training RBM
• Suppose we have two normal
distributions,
• KL-divergence (Kullback-Leibler
divergence) measures the non-
overlapping areas under the two
graphs.
ff
Case Study on RBM
• MNIST (Mixed National Institute of
Standards and Technology)
• Cross entropy
• Cross means comparison
between two distributions