Professional Documents
Culture Documents
restricted boltzmann machine
restricted boltzmann machine
restricted boltzmann machine
Wiki
A Beginner’s Guide to Important Topics in AI, Machine Learning, and Deep
Learning.
Search articles..
A Beginner's Guide to
Bayes' Theorem & Naive RBMs are shallow, two-layer neural nets that constitute
Bayes Classifiers (/bayes- the building blocks of deep-belief networks. The first
theorem-naive-bayes) layer of the RBM is called the visible, or input, layer, and
Convolutional Neural the second is the hidden layer. (Editor’s note: While
RBMs are occasionally used, most practitioners in the
Network (CNN)
machine-learning community have deprecated them
(/convolutional-network)
in favor of generative adversarial networks or
Data for Deep Learning variational autoencoders (generative-adversarial-
(/data-for-deep- network-gan). RBMs are the Model T’s of neural
networks – interesting for historical reasons, but
learning)
surpassed by more up-to-date models.)
Datasets and Machine
Learning (/datasets-ml)
Decision Tree (/decision-
tree)
Deep Autoencoders
(/deep-autoencoder)
Deep-Belief Networks
(/deep-belief-network)
Deep Reinforcement
Learning (/deep-
reinforcement-learning)
Deep Learning Resources
(/deeplearning-
research-papers)
Define Artificial Each circle in the graph above represents a neuron-like
Intelligence (AI) (/define- unit called a node, and nodes are simply where
Gaussian Processes & (MNIST images have 784 pixels, so neural nets
processing them must have 784 input nodes on the
Machine Learning
visible layer.)
(/gaussian-processes)
Generative AI & Now let’s follow that single pixel value, x, through the
Python Tooling for AI input nodes x 3 hidden nodes). The weights between
two layers will always form a matrix where the rows are
(/python-ai)
equal to the input nodes, and the columns are equal to
the output nodes.
Questions When Applying Each hidden node receives the four inputs multiplied by
p (S) p (S)
__
6 1 _
=
0.16 36 6
0.14 __
5
36
0.12 __
4 1 _
=
0.10 36 9
__
3 __ 1
0.08 =
36 12
0.06 __
2 __ 1
=
36 18
0.04
__
1
0.02 36
2 3 4 5 6 7 8 9 10 11 12
S
That is, 7s are the most likely because there are more
ways to get to 7 (3+4, 1+6, 2+5) than there are ways to
arrive at any other sum between 2 and 12. Any formula
attempting to predict the outcome of dice rolls needs
to take seven’s greater frequency into account.
One last point: You’ll notice that RBMs have two biases.
This is one aspect that distinguishes them from other
autoencoders. The hidden bias helps the RBM produce
the activations on the forward pass (since biases
impose a floor so that at least some nodes fire no
matter how sparse the data), while the visible layer’s
biases help the RBM learn the reconstructions on the
backward pass.
Multiple Layers
Once this RBM learns the structure of the input data as
it relates to the activations of the first hidden layer, then
the data is passed one layer down the net. Your first
hidden layer takes on the role of visible layer. The
activations now effectively become your input, and
they are multiplied by weights at the nodes of the
second hidden layer, to produce another set of
activations.
Parameters & k
The variable k is the number of times you run
contrastive divergence
(glossary#contrastivedivergence). Contrastive
divergence is the method used to calculate the
gradient (the slope representing the relationship
between a network’s weights and its error), without
which no learning can occur.
Continuous RBMs
A continuous restricted Boltzmann machine is a form of
RBM that accepts continuous input (i.e. numbers cut
finer than integers) via a different type of contrastive
divergence sampling. This allows the CRBM to handle
things like image pixels or word-count vectors that are
normalized to decimals between zero and one.
Other Resources
Geoff Hinton on Boltzmann Machines
(http://www.scholarpedia.org/article/Boltzmann_machine)
Deeplearning.net’s Restricted Boltzmann Machine Tutorial
(http://deeplearning.net/tutorial/rbm.html)
A Practical Guide to Training Restricted Boltzmann Machines
(https://www.cs.toronto.edu/~hinton/absps/guideTR.pdf);
Geoff Hinton
Share Post
Chris V. Nicholson
Chris V. Nicholson is a venture
partner at Page One Ventures
(https://www.pageone.vc/). He
previously led Pathmind and
Skymind. In a prior life, Chris spent a
decade reporting on tech and
finance for The New York Times,
Businessweek and Bloomberg,
among others.