0% found this document useful (0 votes)
20 views34 pages

Unsupervised Learning

The document discusses unsupervised learning networks. It describes three competitive networks - Maxnet, Mexican hat, and Hamming net. Unsupervised learning involves clustering similar input patterns without supervision. The neural network must discover patterns and relationships in the input data on its own. Competitive learning involves output units competing so that only one unit responds to each input pattern. Kohonen learning is commonly used, where weights are updated based on the input and closest matching unit. Maxnet and Hamming nets are examples of fixed-weight unsupervised learning networks.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
20 views34 pages

Unsupervised Learning

The document discusses unsupervised learning networks. It describes three competitive networks - Maxnet, Mexican hat, and Hamming net. Unsupervised learning involves clustering similar input patterns without supervision. The neural network must discover patterns and relationships in the input data on its own. Competitive learning involves output units competing so that only one unit responds to each input pattern. Kohonen learning is commonly used, where weights are updated based on the input and closest matching unit. Maxnet and Hamming nets are examples of fixed-weight unsupervised learning networks.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 34

06-02-2019

CHAPTER 5
UNSUPERVISED
LEARNING NETWORKS

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 1
by: (Chiranji Lal)

UNSUPERVISED LEARNING

• These competitive nets arc those where the weights remain fixed,
even during training process. The idea of competition is used
among neurons for enhancement of contrast in their activation
functions.
• Three such nets are- Maxnet, Mexican hat and Hamming net.
Original Slides: (Sivanandam & Deepa). Edited
2/6/2019 2
by: (Chiranji Lal)

1
06-02-2019

• Unsupervised learning is not having supervision of


a teacher.
• The learning process is independent.
• During the training of ANN under unsupervised
learning, the input vector of similar type are
combined to form clusters.
• When the new pattern is applied, then the NN gives
an output response indicating the class to which
input pattern belongs.
• No feedback from the environment.
• In this type of learning the network itself must
discover the patterns, feature from the input data
and the relation for the input data over the output.
Original Slides: (Sivanandam & Deepa). Edited
2/6/2019 3
by: (Chiranji Lal)

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 4
by: (Chiranji Lal)

2
06-02-2019

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 5
by: (Chiranji Lal)

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 6
by: (Chiranji Lal)

3
06-02-2019

COMPETITIVE LEARNING
 Output units compete, so that eventually only one neuron (the one
with the most input) is active in response to each output pattern.

 The total weight from the input layer to each output neuron is
limited. If some connections are strengthened, others must be
weakened.

 A consequence is that the winner is the output neuron


whose weights best match the activation pattern.

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 7
by: (Chiranji Lal)

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 8
by: (Chiranji Lal)

4
06-02-2019

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 9
by: (Chiranji Lal)

• The learning algorithm used ·in most of these nets is known as


Kohonen learning.
• In this learning, the units update their weights by forming a new
weight vector, which is a linear combination of the old weight vector
and the new input vector.
• Also, the learning continues for the unit whose weight vector is
closest to the input vector. The weight updation formula used in
Kohonen learning for output cluster unit j is given as:

• where x is the input vector; wj the weight vector for unit j; alpha the
learning rare whose value decreases monotonically as training
continues.
• There exist two methods to determine the winner of the network
during competition: (a) Euclidean distance and (b) dot product.
Original Slides: (Sivanandam & Deepa). Edited
2/6/2019 10
by: (Chiranji Lal)

5
06-02-2019

The networks based on


unsupervised learning.
Fixed Weight Networks

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 11
by: (Chiranji Lal)

Maxnet
• This is also a fixed weight network, which serves as a subnet for
selecting the node having the highest input. All the nodes are fully
interconnected and there exists symmetrical weights in all these
weighted interconnections.

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 12
by: (Chiranji Lal)

6
06-02-2019

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 13
by: (Chiranji Lal)

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 14
by: (Chiranji Lal)

7
06-02-2019

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 15
by: (Chiranji Lal)

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 16
by: (Chiranji Lal)

8
06-02-2019

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 17
by: (Chiranji Lal)

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 18
by: (Chiranji Lal)

9
06-02-2019

•Hamming Network
• In most of the neural networks using unsupervised learning, it is
essential to compute the distance and perform comparisons.
This kind of network is Hamming network, where for every
given input vectors, it would be clustered into different groups.
• Following are some important features of Hamming Networks −
• Lippmann started working on Hamming networks in 1987.
• It is a single layer network.
• The inputs can be either binary {0, 1} or bipolar {-1, 1}.
• The weights of the net are calculated by the exemplar vectors.
• (the weight vector for an output unit in a clustering net is exemplar
vector or code book vector for the pattern of inputs, which the net
has placed on that output unit)
• It is a fixed weight network which means the weights would
remain the same even during training.

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 19
by: (Chiranji Lal)

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 20
by: (Chiranji Lal)

10
06-02-2019

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 21
by: (Chiranji Lal)

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 22
by: (Chiranji Lal)

11
06-02-2019

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 23
by: (Chiranji Lal)

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 24
by: (Chiranji Lal)

12
06-02-2019

UNSUPERVISED LEARNING
 No help from the outside.

 No training data, no information available on the desired output.

 Learning by doing.

 Used to pick out structure in the input:


• Clustering,
• Reduction of dimensionality  compression.

 Example: Kohonen’s Learning Law.


 nodes compete for the right to respond to a
subset of the input data

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 25
by: (Chiranji Lal)

FEW UNSUPERVISED LEARNING NETWORKS


There exists several networks under this category, such as

 Kohonen Self-organizing Feature Maps,

 Learning Vector Quantization,

 Adaptive Resonance Theory.

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 26
by: (Chiranji Lal)

13
06-02-2019

Introduction
 The network by itself should discover any relationship of interest,
such as features, patterns, contours, correlations or categories,
classification in the input data, and thereby translate the
discovered relationship into output.
 Such network are also called as Self-organizing network.

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 27
by: (Chiranji Lal)

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 28
by: (Chiranji Lal)

14
06-02-2019

SELF-ORGANIZATION
 Network Organization is fundamental to the brain

• Functional structure.
• Layered structure.
• Both parallel processing and serial processing require organization
of the brain.

• What is Topology Preservation


• Preservation of the neighborhood relation of the observations in the output
space. It means that the observations which are neighbors in the input space
should be projected in neighbor nodes.

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 29
by: (Chiranji Lal)

SELF-ORGANIZING NETWORKS
 Discover significant patterns or features in the input data.

 Discovery is done without a teacher.

 Synaptic weights are changed according to local rules.

 The changes affect a neuron’s immediate environment until a final


configuration develops.

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 30
by: (Chiranji Lal)

15
06-02-2019

KOHONEN SELF-ORGANIZING FEATURE MAP


(KSOFM)

 The Kohonen model provides a topological mapping.


 (Preservation of the neighborhood relation of the
observations in the output space. It means that the
observations which are neighbors in the input space should be
projected in neighbor nodes.)
 It places a fixed number of input patterns from the input
layer into output or Kohonen layer.

 Training in the Kohonen network begins with the winner’s


neighborhood of a fairly large size. Then, as training
proceeds, the neighborhood size gradually decreases.

 Kohonen SOMs result from the synergy of three basic processes


• Competition,
• Cooperation,
Original Slides: (Sivanandam & Deepa). Edited
• Adaptation.
2/6/2019
by: (Chiranji Lal)
31

Architecture

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 32
by: (Chiranji Lal)

16
06-02-2019

ARCHITECTURE OF KSOFM

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 33
by: (Chiranji Lal)

COMPETITION OF KSOFM
 Each neuron in an SOM is
assigned a weight vector with the
same dimensionality N as the
input space.

 Any given input pattern is


compared to the weight
vector of each neuron and the
closest neuron is declared the
winner.

 The Euclidean norm is commonly


used to measure distance.

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 34
by: (Chiranji Lal)

17
06-02-2019

CO-OPERATION OF KSOFM
 The activation of the winning neuron is spread to neurons in its
immediate neighborhood.
• This allows topologically close neurons to become
sensitive to similar patterns.

 The winner’s neighborhood is determined on the lattice topology.


• Distance in the lattice is a function of the number of lateral
connections to the winner.

 The size of the neighborhood is initially large, but shrinks over


time.
• An initially large neighborhood promotes a topology-preserving
mapping.
• Smaller neighborhoods allow neurons to specialize in the latter
stages of training.
Original Slides: (Sivanandam & Deepa). Edited
2/6/2019 35
by: (Chiranji Lal)

ADAPTATION OF KSOFM
During training, the winner neuron
and its topological neighbors are
adapted to make their weight
vectors more similar to the input
pattern that caused the activation.

Neurons that are closer to the winner


will adapt more heavily than neurons
that are further away.

The magnitude of the adaptation is


controlled with a learning rate,
which decays over time to ensure
convergence of the SOM.

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 36
by: (Chiranji Lal)

18
06-02-2019

KSOFM ALGORITHM

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 37
by: (Chiranji Lal)

EXAMPLE OF KSOFM

Find the winning neuron using the Euclidean distance:

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 38
by: (Chiranji Lal)

19
06-02-2019

Neuron 3 is the winner and its weight vector W3 is updated


according to the competitive learning rule:

The updated weight vector W3 at iteration (p+1) is


determined as:

The weight vector W3 of the winning neuron 3 becomes closer


to the input vector X with each iteration.

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 39
by: (Chiranji Lal)

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 40
by: (Chiranji Lal)

20
06-02-2019

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 41
by: (Chiranji Lal)

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 42
by: (Chiranji Lal)

21
06-02-2019

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 43
by: (Chiranji Lal)

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 44
by: (Chiranji Lal)

22
06-02-2019

LEARNING VECTOR
QUANTIZATION
(LVQ)
Supervised or Unsupervised?

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 45
by: (Chiranji Lal)

Supervised or Unsupervised?
• The Learning Vector Quantization algorithm is a
supervised neural network that uses a competitive
(winner-take-all) learning strategy.
• It is related to other supervised neural networks
such as the Perceptron and the Back-propagation
algorithm.
• It is related to other competitive learning neural
networks such as the Self-Organizing Map algorithm
that is a similar algorithm for unsupervised learning
with the addition of connections between the
neurons.
Original Slides: (Sivanandam & Deepa). Edited
2/6/2019 46
by: (Chiranji Lal)

23
06-02-2019

LEARNING VECTOR QUANTIZATION (LVQ)


 LVQ is a process of classifying the patterns, wherein each output
unit represents a particular class.

 The output unit weight vector is called the reference vector or


code book vector for the class which the unit represents.

 The LVQ is a classifier paradigm that adjusts the boundaries


between categories to minimize existing misclassifications.

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 47
by: (Chiranji Lal)

LEARNING VECTOR QUANTIZATION (LVQ)


 This is a (special case of competitive net which uses)
supervised version of vector quantization. Classes are predefined
and we have a set of labeled data.

 The goal is to determine a set of prototypes that best represent


each class.

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 48
by: (Chiranji Lal)

24
06-02-2019

BASIC SCHEME OF LVQ


Step 1: Initialize prototype vectors for different classes.
Step 2: Present a single input.
Step 3: Identify the closest prototype, i.e., the so-called winner.
(which has weight vector very close to the input vector)
Step 4: Move the winner
• closer toward the data (same class),
• away from the data (different class).

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 49
by: (Chiranji Lal)

LVQ: Flowchart

• In case of
training, a set of
training input
vectors with a
known
classification is
provided with
some initial
distribution of
reference vector.
• Assign initial
weights and
classifications
randomely.
• K-mean
clustering
method.
Original Slides: (Sivanandam & Deepa). Edited
2/6/2019 50
by: (Chiranji Lal)

25
06-02-2019

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 51
by: (Chiranji Lal)

You also read

• VARIANTS OF LVQ
• LVQ 1
• LVQ 2
• LVQ 2.1
• LVQ 3

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 52
by: (Chiranji Lal)

26
06-02-2019

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 53
by: (Chiranji Lal)

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 54
by: (Chiranji Lal)

27
06-02-2019

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 55
by: (Chiranji Lal)

ADAPTIVE
RESONANCE THEORY
(ART) NETWORK
• Adaptive Resonance Theory (ART) networks is
always open to new learning (adaptive) without
losing the old patterns (resonance).
• Basically, ART network is a vector classifier which
accepts an input vector and classifies it into one of
the categories depending upon which of the stored
pattern it resembles the most.

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 56
by: (Chiranji Lal)

28
06-02-2019

ART Operating Principal


• Recognition phase − The input vector is compared with the
classification presented at every node in the output layer.
The output of the neuron becomes “1” if it best matches
with the classification applied, otherwise it becomes “0”.
• Comparison phase − In this phase, a comparison of the
input vector to the comparison layer vector is done. The
condition for reset is that the degree of similarity would be
less than vigilance parameter.
• Search phase − In this phase, the network will search for
reset as well as the match done in the above phases. Hence,
if there would be no reset and the match is quite good, then
the classification is over. Otherwise, the process would be
repeated and the other stored pattern must be sent to find
the correct match.
Original Slides: (Sivanandam & Deepa). Edited
2/6/2019 57
by: (Chiranji Lal)

ADAPTIVE RESONANCE THEORY (ART)


NETWORK

 Adaptive Resonance Theory (ART) is a family of algorithms for


unsupervised learning developed by Carpenter and Grossberg.

 ART is similar to many iterative clustering algorithms where each


pattern is processed by

• finding the "nearest" cluster (a.k.a. prototype or template) to


that exemplar (desired).
• updating that cluster to be "closer" to the exemplar.

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 58
by: (Chiranji Lal)

29
06-02-2019

ARCHITECTURES OF ART NETWORK


 ART1, designed for binary features.

 ART2, designed for continuous (analog) features.

 ARTMAP, a supervised version of ART.

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 59
by: (Chiranji Lal)

Fundamental Architecture
• Three groups of neurons are used to build an ART
network.
• Input processing unit (F1 layer)
• Input portion
• Interface portion
• Clustering units (F2)
• Control mechanism (control degree of similarity of patterns place on
the same cluster)

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 60
by: (Chiranji Lal)

30
06-02-2019

FUNDAMENTAL ALGORITHM OF ART NETWORK

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 61
by: (Chiranji Lal)

ART1
• In ART1 network, it is not necessary to present an input
pattern in a particular order; it can be presented in any
order.
• ART 1 network can be practically implemented by analog
circuits and governing the different equations, i.e., the
bottom-up and top-down weights are controlled by
differential equations.
• ART 1 network runs throughout autonomously.
• It doesn't require any external control signals and can run
stably with infinite patterns of input data.
• ART1 net is trained using fast learning method, in which the
weights reach equilibrium during each learning trial.
• During the resonance phase, the activations of F1 units do
not change; hence the equilibrium weights can be
determined exactly.
Original Slides: (Sivanandam & Deepa). Edited
2/6/2019 62
by: (Chiranji Lal)

31
06-02-2019

BASIC ARCHITECTURE OF ART1

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 63
by: (Chiranji Lal)

ART1 UNITS
ART1 Network is made up of two units

 Computational units
• Input unit (F1 unit – input and interface).
• Cluster unit (F2 unit – output).
• Reset control unit (controls degree of similarity).

 Supplemental units
• One reset control unit.
• Two gain control units.

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 64
by: (Chiranji Lal)

32
06-02-2019

ART2 NETWORK

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 65
by: (Chiranji Lal)

BASIC ARCHITECTURE OF ART2

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 66
by: (Chiranji Lal)

33
06-02-2019

ART2 ALGORITHM

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 67
by: (Chiranji Lal)

SUMMARY
This chapter discussed on the various unsupervised learning networks
like:

 Kohonen Self-organizing Feature Maps

 Learning Vector Quantization

 Adaptive Resonance Theory

Original Slides: (Sivanandam & Deepa). Edited


2/6/2019 68
by: (Chiranji Lal)

34

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy