Wa0020.
Wa0020.
• The idea of soft computing was initiated in 1981 when Lotfi A. Zadeh published his first
paper on soft data analysis “What is Soft Computing”, Soft Computing. Springer-Verlag
Germany/USA 1997.]
• Zadeh, defined Soft Computing into one multidisciplinary system as the fusion of the
fields of Fuzzy Logic, Neuro-Computing, Evolutionary and Genetic Computing, and
Probabilistic Computing.
• Soft Computing is the fusion of methodologies designed to model and enable solutions to
real world problems, which are not modeled or too difficult to model mathematically.
• The aim of Soft Computing is to exploit the tolerance for imprecision, uncertainty,
approximate reasoning, and partial truth in order to achieve close resemblance with
human like decision making.
• The Soft Computing – development history
SC = EC + NN + FL
EC = GP + ES + EP + GA
2
Definitions of Soft Computing (SC)
Lotfi A. Zadeh, 1992 : “Soft Computing is an emerging approach to computing which parallel
the remarkable ability of the human mind to reason and learn in a environment of uncertainty
and imprecision”.
Hybridization of these three creates a successful synergic effect; that is, hybridization creates a
situation where different entities cooperate advantageously for a final outcome.
Hence, a clear definite agreement on what comprises Soft Computing has not yet been reached.
More new sciences are still merging into Soft Computing.
• The main goal of Soft Computing is to develop intelligent machines to provide solutions
to real world problems, which are not modeled, or too difficult to model mathematically.
• Its aim is to exploit the tolerance for Approximation, Uncertainty, Imprecision, and
Partial Truth in order to achieve close resemblance with human like decision making.
Approximation : here the model features are similar to the real ones, but not the same.
3
Uncertainty : here we are not sure that the features of the model are the same as that of the
entity (belief).
Imprecision : here the model features (quantities) are not the same as that of the real ones,
but close to them.
exploit these tolerance to achieve tractability, robustness and low solution cost. In
effect, the role model for soft computing is the human mind.
The four fields that constitute Soft Computing (SC) are : Fuzzy Computing (FC),
Evolutionary Computing (EC), Neural computing (NC), and Probabilistic
Computing (PC), with the latter subsuming belief networks, chaos theory and parts
of learning theory.
4
Fuzzy Computing
In the real world there exists much fuzzy knowledge, that is, knowledge which
is vague, imprecise, uncertain, ambiguous, inexact, or probabilistic in nature.
Human can use such information because the human thinking and reasoning
frequently involve fuzzy information, possibly originating from inherently
inexact human concepts and matching of similar rather then identical
experiences.
The computing systems, based upon classical set theory and two-valued logic,
can not answer to some questions, as human does, because they do not have
completely true answers.
We want, the computing systems should not only give human like answers but
also describe their reality levels. These levels need to be calculated using
imprecision and the uncertainty of facts and rules that were applied.
Fuzzy Sets
5
then we can state explicitly whether each element x of space X "is or
is not" an element of A.
Α : Χ → [0, 1]
1 if x∈ X
A (x) = Eq.(2)
0 otherwise
− Thus, in classical set theory A (x) has only the values 0 ('false') and 1
('true''). Such sets are called crisp sets.
6
• Crisp and Non-crisp Set
The characteristic function A(x) of Eq. (2) for the crisp set is
generalized for the Non-crisp sets.
membership function.
− The proposition of Fuzzy Sets are motivated by the need to capture and
represent real world data with uncertainty due to imprecise
measurement.
7
• Example 1 : Heap Paradox
This example represents a situation where vagueness and uncertainty are inevitable.
- If we remove one grain from a heap of grains, we will still have a heap.
- However, if we keep removing one-by-one grain from a heap of grains, there will be a
time when we do not have a heap anymore.
- The question is, at what time does the heap turn into a countable collection of grains
that do not form a heap? There is no one correct answer to this question.
■
Non-Crisp Representation to represent the notion of a tall person.
A student of height 1.79m would belong to both tall and not tall sets with a particular degree of
membership.As the height increases the membership grade within the tall set would increase whilst
the membership grade within the not-tall set would decrease.
8
• Capturing Uncertainty
Therefore, Crisp Sets ⊆ Fuzzy Sets In other words, Crisp Sets are
Special cases of Fuzzy Sets.
9
Example 2: Set of SMALL ( as non-crisp set) Example 1: Set of prime
numbers ( a crisp set)
10
• Definition of Fuzzy Set
The value A(x) is the membership grade of the element x in a fuzzy set A.
to 12.
Assume:
SMALL(1) = 1, SMALL(2) = 1, SMALL(3) = 0.9, SMALL(4) = 0.6,
11
SMALL(5) = 0.4, SMALL(6) = 0.3, SMALL(7) = 0.2, SMALL(8) = 0.1,
SMALL(u) = 0 for u >= 9.
Set SMALL = {{1, 1 }, {2, 1 }, {3, 0.9}, {4, 0.6}, {5, 0.4}, {6, 0.3}, {7, 0.2},
Note that a fuzzy set can be defined precisely by associating with each x ,
its grade of membership in SMALL.
Originally the universal space for fuzzy sets in fuzzy logic was defined only
on the integers. Now, the universal space for fuzzy sets and fuzzy
relations is defined with three numbers. The first two numbers specify the
start and end of the universal space, and the third argument specifies the
increment between elements. This gives the user more flexibility in
choosing the universal space.
The fuzzy set SMALL of small numbers, defined in the universal space
12
X = { xi } = {1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12} is presented as
SMALL = FuzzySet {{1, 1 }, {2, 1 }, {3, 0.9}, {4, 0.6}, {5, 0.4}, {6, 0.3},
{7, 0.2}, {8, 0.1}, {9, 0 }, {10, 0 }, {11, 0}, {12, 0}}
13
• Graphic Interpretation of Fuzzy Sets PRIME Numbers
PRIME = FuzzySet {{1, 0}, {2, 1}, {3, 1}, {4, 0}, {5, 1}, {6, 0}, {7, 1}, {8, 0}, {9, 0}, {10, 0}, {11, 1},
{12, 0}}
{8, 0}, {9, 0}, {10, 0}, {11, 1}, {12, 0}} , UniversalSpace → {1, 12, 1}]
14
• Graphic Interpretation of Fuzzy Sets UNIVERSALSPACE
In any application of sets or fuzzy sets theory, all sets are subsets of
UNIVERSALSPACE = FuzzySet {{1, 1}, {2, 1}, {3, 1}, {4, 1}, {5, 1}, {6, 1},
{7, 1}, {8, 1}, {9, 1}, {10, 1}, {11, 1}, {12, 1}}
15
Finite and Infinite Universal Space
Examples:
16
• Graphic Interpretation of Fuzzy Sets EMPTY
EMPTY = FuzzySet {{1, 0}, {2, 0}, {3, 0}, {4, 0}, {5, 0}, {6, 0}, {7, 0}, {8,
{8, 0}, {9, 0}, {10, 0}, {11, 0}, {12, 0}} , UniversalSpace → {1, 12, 1}]
17
Fuzzy Operations
A fuzzy set operations are the operations on fuzzy sets. The fuzzy set
operations are generalization of crisp set operations. Zadeh [1965]
formulated the fuzzy set theory in the terms of standard operations:
Complement, Union, Intersection, and Difference.
Inclusion :
FuzzyInclude [VERYSMALL, SMALL]
Equality :
FuzzyEQUALITY [SMALL, STILLSMALL]
Complement :
FuzzyNOTSMALL = FuzzyCompliment [Small]
Union :
FuzzyUNION = [SMALL MEDIUM]
∪
Intersection :
FUZZYINTERSECTON = [SMALL MEDIUM]
∩
18
• Inclusion
The fuzzy set A is included in the fuzzy set B if and only if for every x in
Example :
SMALL = FuzzySet {{1, 1 }, {2, 1 }, {3, 0.9}, {4, 0.6}, {5, 0.4}, {6, 0.3},
{7, 0.2}, {8, 0.1}, {9, 0 }, {10, 0 }, {11, 0}, {12, 0}}
19
SetVerySmall = FuzzySet [{{1,1},{2,0.8}, {3,0.7}, {4,0.4}, {5,0.2},{6,0.1},
{7,0}, {8, 0}, {9, 0}, {10, 0}, {11, 0}, {12, 0}} , UniversalSpace → {1, 12, 1}]
Include
[VERYSMALL,
SMALL]
20
• Comparability
if one of the fuzzy sets is a subset of the other set, they are comparable.
Example 1:
Example 2 :
D is not a subset of C.
21
• Equality
Let A and B
be fuzzy sets defined in the same space X.
Then A and B if
and only if
are equal, which is denoted X = Y
Example.
SMALL = FuzzySet {{1, 1 }, {2, 1 }, {3, 0.9}, {4, 0.6}, {5, 0.4}, {6, 0.3},
{7, 0.2}, {8, 0.1}, {9, 0 }, {10, 0 }, {11, 0}, {12, 0}}
STILLSMALL = FuzzySet {{1, 1 }, {2, 1 }, {3, 0.9}, {4, 0.6}, {5, 0.4},
{6, 0.3}, {7, 0.2}, {8, 0.1}, {9, 0 }, {10, 0 }, {11, 0}, {12, 0}}
22
• Complement
Then the fuzzy set B is a complement of the fuzzy set A, if and only if,
Example 1.
SMALL = FuzzySet {{1, 1 }, {2, 1 }, {3, 0.9}, {4, 0.6}, {5, 0.4}, {6, 0.3},
{7, 0.2}, {8, 0.1}, {9, 0 }, {10, 0 }, {11, 0}, {12, 0}}
NOTSMALL = FuzzySet {{1, 0 }, {2, 0 }, {3, 0.1}, {4, 0.4}, {5, 0.6}, {6, 0.7},
{7, 0.8}, {8, 0.9}, {9, 1 }, {10, 1 }, {11, 1}, {12, 1}}
23
SC – Fuzzy Computing
Example 2.
The empty set Φ and the universal set X, as fuzzy sets, are
Φ'= X' =
X , Φ
Empty = FuzzySet {{1, 0 }, {2, 0 }, {3, 0}, {4, 0}, {5, 0}, {6, 0},
{7, 0}, {8, 0}, {9, 0 }, {10, 0 }, {11, 0}, {12, 0}}
Universal = FuzzySet {{1, 1 }, {2, 1 }, {3, 1}, {4, 1}, {5, 1}, {6, 1},
{7, 1}, {8, 1}, {9, 1 }, {10, 1 }, {11, 1}, {12, 1}}
24
• Union
The union is defined as the smallest fuzzy set that contains both A and
B. The union of A and B is denoted by A ∪ B.
A(x) = 0.6 and B(x) = 0.4 ∴ (A ∪ B)(x) = max [0.6, 0.4] = 0.6
SMALL = FuzzySet {{1, 1 }, {2, 1 }, {3, 0.9}, {4, 0.6}, {5, 0.4}, {6, 0.3},
{7, 0.2}, {8, 0.1}, {9, 0 }, {10, 0 }, {11, 0}, {12, 0}}
MEDIUM = FuzzySet {{1, 0 }, {2, 0 }, {3, 0}, {4, 0.2}, {5, 0.5}, {6, 0.8},
{7, 1}, {8, 1}, {9, 0.7 }, {10, 0.4 }, {11, 0.1}, {12, 0}}
FUZZYUNION = [SMALL
∪ MEDIUM]
The notion of the union is closely related to that of the connective "or".
25
If "David is Young" or "David is Bald," then David is associated with the
union of A and B. Implies David is a member of A ∪ B.
■
Identity:
A∪
Φ =A
input
= Equality [SMALL ∪ EMPTY , SMALL]
output = True
A
∪ X=X
input
= Equality [SMALL ∪ UnivrsalSpace , UnivrsalSpace]
output = True
■ Idempotence :
A∪A=A
output = True
■
Commutativity :
A ∪ B =B ∪ A
26
input = Equality [SMALL ∪ MEDIUM, MEDIUM ∪ SMALL]
output = True
■
Associativity:
A
∪ (B C) = B) C
∪ (A∪ ∪
output = True
MEDIUM = FuzzySet {{1, 0 }, {2, 0 }, {3, 0}, {4, 0.2}, {5, 0.5}, {6, 0.8},
{7, 1}, {8, 1}, {9, 0 }, {10, 0 }, {11, 0.1}, {12, 0}}
BIG
= FuzzySet [{{1,0}, {2,0}, {3,0}, {4,0}, {5,0}, {6,0.1}, {7,0.2},
{8,0.4}, {9,0.6}, {10,0.8}, {11,1}, {12,1}}]
27
27
SC – Fuzzy Computing
• Intersection
The intersection is defined as the greatest fuzzy set included both A and
B. The intersection of A and B is denoted by A ∩ B.
A(x) = 0.6 and B(x) = 0.4 ∴ (A ∩ B)(x) = min [0.6, 0.4] = 0.4
SMALL = FuzzySet {{1, 1 }, {2, 1 }, {3, 0.9}, {4, 0.6}, {5, 0.4}, {6, 0.3},
{7, 0.2}, {8, 0.1}, {9, 0 }, {10, 0 }, {11, 0}, {12, 0}}
MEDIUM = FuzzySet {{1, 0 }, {2, 0 }, {3, 0}, {4, 0.2}, {5, 0.5}, {6, 0.8},
{7, 1}, {8, 1}, {9, 0.7 }, {10, 0.4 }, {11, 0.1}, {12, 0}}
{5,0.4}, {6,0.3}, {7,0.2}, {8, 0.1}, {9, 0},{10, 0}, {11, 0}, {12, 0}} , UniversalSpace →
{1, 12, 1}]
28
Neural Computing
29
Biological Model:
The human brain consists of a large number (more than a billion) of neural
cells that process information. Each cell works like a simple processor. The
massive interaction between all cells and their parallel processing, makes
the brain's abilities possible. The
structure of neuron is shown below.
Myelin Sheath consists of fat-containing cells that insulate the axon from electrical
activity. This insulation acts to increase the rate of transmission of signals. A gap
exists between each myelin sheath cell along the axon. Since fat inhibits the
propagation of electricity, the signals jump from one gap to the next.
30
Nodes of Ranvier are the gaps (about 1 m) between myelin sheath cells long
axons. Since fat serves as a good insulator, the myelin sheaths speed the rate of
transmission of an electrical impulse along the axon.
Synapse is the point of connection between two neurons or a neuron and a muscle or
a gland. Electrochemical communication between neurons takes place at these
junctions.
Terminal Buttons of a neuron are the small knobs at the end of an axon that
release chemicals called neurotransmitters.
31
• Information flow in a Neural Cell
The input /output and the propagation of information are shown below.
32
Artificial Neuron
The signals build up in the cell. Finally the cell fires (discharges)
through the output. The cell can start building up signals again.
33
• Functions :
34
• McCulloch-Pitts (M-P) Neuron Equation
34
35
• Basic Elements of an Artificial Neuron
Weighting Factors
Threshold
y= Σn Xi Wi - Φ k
i=1
36
Activation Function
37
• Example :
y (threshold) = 1
38
• Single and Multi - Layer Perceptrons
The neurons are shown as circles in the diagram. It has several inputs and
a single output. The neurons have gone under various names.
multi-layer perceptron
Output
Output Output
Input InputInput
39
• Perceptron
40
Genetic Algorithms
Genetic Algorithms (GAs) were invented by John Holland in early 1970's to mimic
some of the processes observed in natural evolution.
Later in 1992 John Koza used GAs to evolve programs to perform certain tasks.
He called his method "Genetic Programming" (GP).
GAs simulate the survival of the fittest, among individuals over consecutive
generation for solving a problem. Each generation consists of a population of
character strings that are analogous to the chromosome in our DNA
(Deoxyribonucleic acid). DNA contains the genetic instructions used in the
development and functioning of all known living organisms.
search into the region of better performance within the search space.
■ In nature, competition among individuals for scanty resources results
in the fittest individuals dominating over the weaker ones.
41
• Why Genetic Algorithms
42
■ Each gene represents a specific trait (feature) of the organism and has
several different settings, e.g. setting for a hair color gene may be
black or brown.
■ When two organisms mate they share their genes. The resultant
offspring may end up having half the genes from one parent and half
from the other parent. This process is called crossover
(recombination).
Mutation means, that the elements of DNA are a bit changed. This
change is mainly caused by errors in copying genes from parents.
43
Artificial Evolution and Search Optimization
44
• Taxonomy of Evolution & Search Optimization Classes
■ Enumerative Methods
These are the traditional search and control strategies. They search for
a solution in a problem space within the domain of artificial
intelligence. There are many control structures for search. The depth-
first search and breadth-first search are the two most
45
categories : uninformed and informed methods.
46
■ Calculus based techniques
47
■ Guided Random Search techniques
48
Evolutionary Algorithms (EAs)
new solutions that are biased towards regions of the space for
Each feasible solution is marked by its value or fitness for the problem.
- Search for a solution point, means finding which one point (or more)
among many feasible solution points in the search space is the solution.
This requires looking for some extremes, minimum or maximum.
- Search space can be whole known, but usually we know only a few
points and we are generating other points as the process of finding
solution continues.
49
- Search can be very complicated. One does not know where to look
- What we find is some suitable solution, not necessarily the best solution.
The solution found is often considered as a good solution, because it is not
often possible to prove what is the real optimum solution.
Associative Memory
50
Description of Associative Memory
■
A content-addressable memory is a type of memory that allows, the
recall of data based on the degree of similarity between the input
pattern and the patterns stored in memory.
■
It refers to a memory organization in which the memory is accessed by
its content and not or opposed to an explicit address in the traditional
computer memory system.
■
This type of memory allows the recall of information based on partial
knowledge of its contents.
51
■ It is a system that “associates” two patterns (X, Y) such that
- auto-associative and
- hetero-associative.
52
Adaptive Resonance Theory (ART)
Note : The terms nearest and closer are defined in many ways in clustering
algorithm. In ART, these two terms are defined in slightly different way by
introducing the concept of "resonance".
53
• Definitions of ART and other types of Learning
The unsupervised learning means that the network learns the significant
patterns on the basis of the inputs only. There is no feedback. There is no
external teacher that instructs the network or tells to which category a
certain input belongs. Learning in biological systems always starts as
unsupervised learning; Example : For the newly born, hardly any pre-
existing categories exist.
only limited feedback, like "on this input you performed well" or
"on this input you have made an error". In supervised mode of learning
54
• Description of Adaptive Resonance Theory
− a reset module.
best match is, the single neuron whose set of weights (weight
vector) matches most closely the input vector.
■ Vigilance parameter
55
■ Reset module
After the input vector is classified, the Reset module compares the
strength of the recognition match with the vigilance parameter.
- The Recognition neurons are disabled one by one by the reset function
until the vigilance parameter is satisfied by a recognition match.
There are two basic methods, the slow and fast learning.
- Fast learning method : here the algebraic equations are used to calculate
degree of weight adjustments to be made, and binary values are used.
56
Note : While fast learning is effective and efficient for a variety of tasks,
the slow learning method is more biologically plausible and can be used
with continuous-time networks (i.e. when the input vector can vary
continuously).
■ ART 1: The simplest variety of ART networks, accept only binary inputs.
■ Fuzzy ART : It Implements fuzzy logic into ART’s pattern recognition, thus
enhances generalizing ability. One very useful feature of fuzzy ART is
complement coding, a means of incorporating the absence of features into
pattern classifications, which goes a long way towards preventing
inefficient and unnecessary category proliferation.
57
Applications of Soft Computing
The relevance of soft computing for pattern recognition and image processing
is already established during the last few years. The subject has recently
gained importance because of its potential applications in problems like :
- Medical Imaging,
- Forensic Applications,
- Signature Verification,
- Multimedia,
- Target Recognition,
58
Fundamentals of Neural Networks
59
Introduction
60
Why Neural Network
■ The conventional computers are good for - fast arithmetic and does
what programmer programs, ask them to do.
61
Research History
The history is relevant because for nearly two decades the future of Neural
network remained uncertain.
McCulloch and Pitts (1943) are generally recognized as the designers of the
first neural network. They combined many simple processing units together
that could lead to an overall increase in computational power. They suggested
many ideas like : a neuron has a threshold level and once that level is
reached the neuron fires. It is still the fundamental way in which ANNs
operate. The McCulloch and Pitts's network had a fixed set of weights.
In the 1950 and 60's, many researchers (Block, Minsky, Papert, and
Rosenblatt worked on perceptron. The neural network model could be proved
to converge to the correct weights, that will solve the problem. The weight
adjustment (learning algorithm) used in the perceptron was found more
powerful than the learning rules used by Hebb. The perceptron caused great
excitement. It was thought to produce programs that could think.
Minsky & Papert (1969) showed that perceptron could not learn those
functions which are not linearly separable.
The neural networks research declined throughout the 1970 and until mid
80's because the perceptron could not learn certain important functions.
62
Neural network regained importance in 1985-86. The researchers, Parker
and LeCun discovered a learning algorithm for multi-layer networks called
back propagation that could solve problems that were not linearly
separable.
63
Biological Neuron Model
The human brain consists of a large number, more than a billion of neural
cells that process information. Each cell works like a simple processor. The
massive interaction between all cells and their parallel processing only
makes the brain's abilities possible.
Myelin Sheath consists of fat-containing cells that insulate the axon from electrical
activity. This insulation acts to increase the rate of transmission of signals. A gap
exists between each myelin sheath cell along the axon. Since fat inhibits the
propagation of electricity, the signals jump from one gap to the next.
64
Nodes of Ranvier are the gaps (about 1 m) between myelin sheath cells long
axons are Since fat serves as a good insulator, the myelin sheaths speed the rate of
transmission of an electrical impulse along the axon.
Synapse is the point of connection between two neurons or a neuron and a muscle or
a gland. Electrochemical communication between neurons takes place at these
junctions.
Terminal Buttons of a neuron are the small knobs at the end of an axon that
release chemicals called neurotransmitters.
The input /output and the propagation of information are shown below.
65
■ Axons act as transmission lines to send activation to other neurons.
66
SC - Neural Network – Introduction
■
A set of input connections brings in activations from other neurons.
■
A processing unit sums the inputs, and then applies a non-linear
activation function (i.e. squashing / transfer / threshold function).
■
An output line transmits the result to other neurons.
In other words ,
67
Single Layer Feed-forward Network
calculated in each neuron node, and if the value is above some threshold
(typically 0) the neuron fires and takes the activated value (typically 1);
otherwise it takes the deactivated value (typically -1).
input xi output yj
weights wij
w11
x1 y1
w21
w12
w22
x2 y2
w2m
w1m
wn1
wn2
xn ym
wnm
Single layer
Neurons
68
Multi Layer Feed-forward Network
Fig.
Multilayer feed-forward network in (ℓ – m – n) configuration.
■ The input layer neurons are linked to the hidden layer neurons; the
weights on these links are referred to as input-hidden layer weights.
■ The hidden layer neurons and the corresponding weights are referred to
as output-hidden layer weights.
69
■ A multi-layer feed-forward network with ℓ input neurons, m1 neurons in
the first hidden layers, m2 neurons in the second hidden layers, and n
output neurons in the output layers is written as (ℓ - m1 - m2 – n ).
Recurrent Networks
Example :
70
Learning Methods in Neural Networks
The learning methods in neural networks are classified into three basic types :
• Supervised Learning,
• Reinforced Learning
• Hebbian,
• Gradient descent,
• Competitive and
• Stochastic learning.
71
• Classification of Learning Algorithms
Neural Network
Learning algorithms
72
b Supervised Learning
c Unsupervised Learning
No teacher is present.
d Reinforced learning
Note : The Supervised and Unsupervised learning methods are most popular
forms of learning compared to Reinforced learning.
73
• Hebbian Learning
In this rule, the input-output pattern pairs (Xi , Yi) are associated by the
weight matrix W, known as correlation matrix computed as
W= Σn Xi Yi
T
i=1
There are many variations of this rule proposed by the other researchers
(Kosko, Anderson, Lippman) .
74
• Gradient descent Learning
• If ∆ Wij is the weight update of the link connecting the i th and the j th
Wij = ( E/ Wij )
η ∂ ∂
where
η is the learning rate parameters and E/ Wij ) is error
(∂ ∂
gradient
with reference to the weight Wij .
75
• Competitive Learning
• Stochastic Learning
76
• Taxonomy Of Neural Network Systems
In the previous sections, the Neural Network Architectures and the Learning
methods have been discussed. Here the popular neural network
systems are listed. The grouping of these systems in terms of architectures and
the learning methods are presented in the next slide.
– AM (Associative Memory)
– Boltzmann machines
– BSB ( Brain-State-in-a-Box)
– Cauchy machines
– Hopfield Network
– Neoconition
77
– Perceptron
30
78
• Classification of Neural Network
Learning Methods
descent
Percepton,
RBF
Hopfield, machines
31
79
■ Single-Layer NN Systems
input xi output yj
weights wij
w11
x1 y1
w21
w12
w22
x2 y2
w2m
w1m
wn1
wn2
xn ym
wnm
Single layer
Perceptron
where net j =
n
y j = f (net j) = 1 if net j ≥ 0 Σ xi wij
0 if net j 0 i=1
80
■ Learning Algorithm : Training Perceptron
K+1 K
i.e. W = W
ij ij
− If the output is 1 but should have been 0 then the weights are
K+1 K
W − α.
i.e. ij = W i j xi
− If the output is 0 but should have been 1 then the weights are
increased on the active input link
K+1 K
i.e. W ij
+ . xi
=W ij
α
Where
Wi j
81