0% found this document useful (0 votes)
99 views26 pages

UNDERSTANG PERCEPTRON and Perceptron LEARNING

Artificial intelligence

Uploaded by

dassomasree4006
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
99 views26 pages

UNDERSTANG PERCEPTRON and Perceptron LEARNING

Artificial intelligence

Uploaded by

dassomasree4006
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 26

UNDERSTANG PERCEPTRON

LEARNING

Machine learning (ML) is a field of study in


artificial intelligence concerned with the
development and study of statistical algorithms
that can learn from data and generalize to unseen
data and thus perform tasks without explicit
instructions. Recently, artificial neural networks
have been able to surpass many previous
approaches in performance. ML finds application
in many fields, including natural language
processing, computer vision, speech recognition,
email filtering, agriculture, and medicine. When
applied to business problems, it is known under
the name predictive analytics. Although not all
machine learning is statistically based,
computational statistics is an important source of
the field's methods.

The term machine learning was coined in 1959 by


Arthur Samuel, an IBM employee and pioneer in
the field of computer gaming and artificial
intelligence. The synonym self-teaching computers
was also used in this time period. The earliest
machine learning model was introduced in the
1950s when Arthur Samuel invented the a
programme which calculated the winning chance
in checkers for each side.

Machine learning is important because it gives


enterprises a view of trends in customer behaviour
and operational business patterns as well as
supports the developments of new products.
Machine learning enables the automation of tasks
which are complex and time consuming; ML
algorithms can process & analyze vast amounts of
data quickly; ML powers personalization engines,
enabling companies to offer tailored content ,
products and services to individual users. This is
widely seen in platforms like Netflix, Amazon and
Spotify and many others; ML powers intelligent
systems like chatbots and virtual assistants,
providing instant and accurate customer support
while reducing the need for human intervention,
etc,.
INTRODUCTION TO PERCEPTRON
LEARNING

A single-layer feed forward neural network was


introduced in the late 1950s by Frank Rosenblatt. It
was the starting phase of Deep Learning and
Artificial neural networks. During that time for
prediction, Statistical machine learning, or
Traditional code Programming is used. Perceptron
is one of the first and most straightforward
models of artificial neural networks. Despite being
a straightforward model, the perceptron has been
proven to be successful in solving specific
categorization issues.

The perceptron was invented in 1943 by Warren


McCulloch and Walter Pitts. The first hardware
implementation was Mark I Perceptron machine
built in 1957 at the Cornell Aeronautical Laboratory
by Frank Rosenblatt,funded by the Information
Systems Branch of the United States Office of
Naval Research and the Rome Air Development
Center. It was first publicly demonstrated on 23
June 1960.The machine was "part of a previously
secret four-year NPIC [the US' National
Photographic Interpretation Center] effort from
1963 through 1966 to develop this algorithm into a
useful tool for photo-interpreters".

Rosenblatt described the details of the perceptron


in a 1958 paper. His organization of a perceptron
is constructed of three kinds of cells ("units"): AI,
AII, R, which stand for "projection", "association"
and "response".

The Mark I Perceptron has 3 layers.

 An array of 400 photocells arranged in a 20x20


grid, named "sensory units" (S-units), or "input
retina". Each S-unit can connect to up to 40 A-
units.
 A hidden layer of 512 perceptrons, named
"association units" (A-units).
 An output layer of 8 perceptrons, named
"response units" (R-units).

Rosenblatt called this three-layered perceptron


network the alpha-perceptron, to distinguish it
from other perceptron models he experimented
with.

Rosenblatt described his experiments with many


variants of the Perceptron machine in a book
Principles of Neurodynamics (1962). The book is a
[16]
published version of the 1961 report.
Among the variants are:

 "cross-coupling" (connections between units


within the same layer) with possibly closed
loops,
 "back-coupling" (connections from units in a
later layer to units in a previous layer),
 four-layer perceptrons where the last two
layers have adjustible weights (and thus a
proper multilayer perceptron),
 incorporating time-delays to perceptron units,
to allow for processing sequential data,
 analyzing audio (instead of images).

The machine was shipped from Cornell to


Smithsonian in 1967, under a government transfer
administered by the Office of Naval Research.

 MARVIN MINSKY & SEYMOUR PAPERT VIEWS


ON PERCEPTRON

Although the perceptron initially seemed promising,


it was quickly proved that perceptrons could not be
trained to recognise many classes of patterns. This
caused the field of neural network research to
stagnate for many years, before it was recognised
that a feedforward neural network with two or more
layers (also called a multilayer perceptron) had
greater processing power than perceptrons with
one layer (also called a single-layer perceptron).

In 1969, a famous book entitled Perceptrons by


Marvin Minsky and Seymour Papert showed that it
was impossible for these classes of network to
learn an XOR function. It is often believed
(incorrectly) that they also conjectured that a
similar result would hold for a multi-layer
perceptron network. However, this is not true, as
both Minsky and Papert already knew that multi-
layer perceptrons were capable of producing an
XOR function. (See the page on Perceptrons (book)
for more information.) Nevertheless, the often-
miscited Minsky/Papert text caused a significant
decline in interest and funding of neural network
research. It took ten more years until neural
network research experienced a resurgence in the
1980s.This text was reprinted in 1987 as
"Perceptrons - Expanded Edition" where some
errors in the original text are shown and corrected.

 EVOLUTION OVER TIME

Rosenblatt continued working on perceptrons


despite diminishing funding. The last attempt was
Tobermory, built between 1961 and 1967, built for
speech recognition.It occupied an entire room.It
had 4 layers with 12,000 weights implemented by
toroidal magnetic cores. By the time of its
completion, simulation on digital computers had
become faster than purpose-built perceptron
machines. He died in a boating accident in 1971.

The kernel perceptron algorithm was already


introduced in 1964 by Aizerman et al.Margin
bounds guarantees were given for the Perceptron
algorithm in the general non-separable case first by
Freund and Schapire (1998), and more recently by
Mohri and Rostamizadeh (2013) who extend
previous results and give new and more favorable
L1 bounds.

The perceptron is a simplified model of a biological


neuron. While the complexity of biological neuron
models is often required to fully understand neural
behavior, research suggests a perceptron-like linear
model can produce some behavior seen in real
neurons.

THE PERCEPTRON MODEL


Perceptron model is also treated as one of the
best and simplest types of Artificial Neural
networks. However, it is a supervised learning
algorithm of binary classifiers. Hence, we can
consider it as a single-layer neural network with
four main parameters, i.e., input values, weights
and Bias, net sum, and an activation function.

In Machine Learning, binary classifiers are defined


as the function that helps in deciding whether
input data can be represented as vectors of
numbers and belongs to some specific class.

Basic Components of Perceptron

Mr. Frank Rosenblatt invented the perceptron model as


a binary classifier which contains three main
components. These are as follows:
 Input Nodes or Input Layer:

This is the primary component of Perceptron which


accepts the initial data into the system for further
processing. Each input node contains a real numerical
value.

 Wight and Bias:

Weight parameter represents the strength of the


connection between units. This is another most
important parameter of Perceptron components. Weight
is directly proportional to the strength of the associated
input neuron in deciding the output. Further, Bias can be
considered as the line of intercept in a linear equation.

 Activation Function:

These are the final and important components that help


to determine whether the neuron will fire or not.
Activation Function can be considered primarily as a
step function.

Types of Activation functions:

How does Perceptron work?


Step-1

In the first step first, multiply all input values with


corresponding weight values and then add them to
determine the weighted sum. Mathematically, we can
calculate the weighted sum as follows:

∑wi*xi = x1*w1 + x2*w2 +…wn*xn

Add a special term called bias 'b' to this weighted sum


to improve the model's performance.

∑wi*xi + b

Step-2

In the second step, an activation function is applied with


the above-mentioned weighted sum, which gives us
output either in binary form or a continuous value as
follows:

Y = f(∑wi*xi + b)

PERCEPTRON LEARNING ALGORITHM


APPLICATIONS OF PERCEPTRON
LEARNING

The early applications of perceptron learning are –

 One of the 1st practical applications of it was in the


area of character recognition and was used to
classify and recognise letters ,digits , symbols ,etc,.
 It was also applied to simple image recognition
tasks.
 In the early stages, the perceptron was explored for
applications in signal processing such as detecting
patterns in radar signals etc,.

The modern applications of perceptron learning are


 It is used in binary classification task like spam


detection and sentiment analysis.
 It influenced the development of SUPPORT
VECTOR MACHINES (SVMs).
 It is the building block of the modern neural
networks.
 It influenced the NATURAL LANGUAGE
PROCESSING , ROBOTICS and CONTROL
SYSTEMS and QUANTUM COMPUTING .

LIMITATION OF PERCEPTRON
LEARNING
Perceptual learning is the process by which
experience or practice results in a long-lasting
improvement in the ability to perceive sensory
information. While perceptual learning can lead
to significant improvements in sensory
processing and performance, it does have
several limitations:

### 1. **Specificity**

- **Task Specificity:** Perceptual learning tends


to be highly specific to the task or stimuli used
during training. Improvements in one task or
with one set of stimuli may not transfer to
other tasks or stimuli.

- **Context Specificity:** Learning can also be


specific to the context, including the
environment, the position of the stimulus, or
even the timing of the stimuli.

### 2. **Slower Learning Rate**

- Perceptual learning often requires extensive


practice over long periods, which can be time-
consuming. The rate of improvement might be
slow, especially in complex tasks or for older
individuals.

### 3. **Age-Related Constraints**

- The ability to engage in perceptual learning


declines with age. Younger individuals tend to
benefit more from perceptual learning than
older adults, who may show slower progress or
reduced capacity for change.

### 4. **Plateaus in Learning**

- After an initial period of improvement,


individuals may hit a plateau where further
practice does not lead to significant gains. This
can be due to reaching the limits of perceptual
sensitivity or the constraints of the training
method.

### 5. **Limited Transferability**


- Even when some transfer of learning occurs, it
is often limited to closely related tasks.
Generalization to broader or more distant tasks
is rare, which can reduce the practical utility of
perceptual learning in real-world settings.

### 6. **Resource Intensive**

- Effective perceptual learning often requires


controlled training environments, specialized
equipment, or software, and sometimes the
guidance of an expert, making it less
accessible for everyone.

### 7. **Interference**

- Prior learning can interfere with new learning.


For example, if an individual has learned to
perceive a certain stimulus in a particular way,
it may be difficult for them to adapt to
perceiving the same stimulus in a different way.

### 8. **Cognitive Load**

- Perceptual learning can sometimes require


significant cognitive resources, especially in
the early stages. This can lead to mental
fatigue, which may hinder learning or make it
less effective.
### 9. **Neural Adaptation**

- The neural changes associated with


perceptual learning can be highly specific and
not always beneficial. In some cases, these
changes may lead to maladaptive effects, such
as over-sensitivity to certain stimuli.

These limitations highlight the challenges in


applying perceptual learning to broader or more
complex tasks and the need for carefully
designed training protocols to maximize its
benefits.

SIGNIFICANCE AND IMPACTS

### Significance of Perceptron Learning

1. **Foundation of Neural Networks:**

- The perceptron is a fundamental building


block for more complex neural network
architectures. It introduced the concept of
neurons as binary classifiers, laying the
groundwork for multi-layer networks.

2. **Mathematical Formalization:**

- The perceptron formalized the idea of a


linear classifier. It uses a set of weights and a
bias to map input features to an output
decision, introducing key concepts such as
weight updates and activation functions.

3. **Learning Algorithm:**

- The perceptron learning algorithm is an


iterative process that adjusts the weights
based on the error of the prediction. This
concept of weight adjustment is central to
training neural networks.

4. **Convergence Proof:**

- For linearly separable data, the perceptron


learning algorithm guarantees convergence.
This was one of the first proofs in machine
learning showing that an algorithm could learn
from data and improve performance over time.

### Impact on Modern Machine Learning


1. **Inspiration for Deep Learning:**

- The perceptron inspired the development of


multi-layer perceptrons (MLPs) or feedforward
neural networks, which are the basis for deep
learning. The concepts of layers, neurons, and
backpropagation (an extension of the
perceptron learning rule) are critical in modern
deep learning.

2. **Development of Algorithms:**

- Techniques from the perceptron algorithm


have influenced the development of various
other algorithms, including support vector
machines (SVMs) and logistic regression. The
idea of adjusting parameters to minimize error
is a common thread in many machine learning
algorithms.

3. **Understanding Limitations:**

- The limitations of the perceptron,


particularly its inability to solve non-linearly
separable problems (as highlighted by Minsky
and Papert in their critique), spurred research
into more sophisticated models. This led to the
development of algorithms that could handle
non-linear relationships, such as neural
networks with hidden layers and kernel
methods in SVMs.

4. **Historical Context:**

- The initial excitement and subsequent


criticism of the perceptron marked important
phases in the history of artificial intelligence.
The eventual resurgence of interest in neural
networks in the 1980s and 1990s, culminating
in the deep learning revolution of the 2010s,
owes much to the foundational ideas of the
perceptron.

5. **Educational Value:**

- The perceptron is often used as an


introductory example in machine learning and
artificial intelligence courses. It helps students
understand basic concepts such as linear
classification, weight adjustment, and iterative
learning processes.

CRITICIZM AND LIMITATION

The perceptron, despite being a foundational


model in the history of artificial intelligence and
machine learning, has several criticisms and
limitations. These limitations largely stem from
its simplicity and the fact that it is a linear
classifier. Here are the key criticisms and
limitations of the perceptron:

### 1. **Inability to Solve Non-linearly


Separable Problems**

- **XOR Problem:** The most well-known


limitation of the perceptron is its inability to
solve problems where the data is not linearly
separable, such as the XOR problem. Since the
perceptron uses a linear decision boundary, it
cannot correctly classify datasets where the
classes cannot be separated by a straight line.

### 2. **Single-layer Limitation**

- **No Hidden Layers:** The original


perceptron model consists of a single layer of
neurons (the input layer directly connected to
the output layer), which limits its ability to
model complex relationships within data.
Without hidden layers, the perceptron cannot
learn hierarchical representations or capture
non-linear patterns.

### 3. **Binary Classification**


- **Limited to Two Classes:** The basic
perceptron is designed for binary classification
problems. Although it can be extended to
handle multi-class classification using
techniques like one-vs-all or one-vs-one, the
basic perceptron itself is not naturally equipped
for this task.

### 4. **Convergence on Linearly Separable


Data**

- **Convergence Guarantee:** While the


perceptron algorithm converges if the data is
linearly separable, there is no guarantee of
convergence for non-linearly separable data. In
such cases, the algorithm might keep adjusting
weights without ever finding a solution, leading
to an infinite loop or requiring additional rules
to terminate.

### 5. **Sensitivity to Data Representation**

- **Feature Scaling:** The performance of the


perceptron is sensitive to how the data is
represented, including feature scaling and
normalization. Poorly scaled data can lead to
poor convergence behavior and suboptimal
performance.
### 6. **Hard Threshold Activation Function**

- **Non-differentiable:** The perceptron uses


a hard threshold (step function) as its
activation function, which is not differentiable.
This limits the model’s ability to be trained
using gradient-based optimization methods,
which are crucial for training more complex
neural networks.

### 7. **Lack of Probabilistic Output**

- **No Probability Estimates:** The


perceptron produces a binary output without
providing a measure of confidence or
probability for its predictions. Modern machine
learning models often provide probabilistic
outputs, which are useful for decision-making
and uncertainty estimation.

### 8. **No Handling of Noisy Data**

- **No Robustness to Noise:** The perceptron


is not robust to noisy or overlapping data. If the
data is not perfectly separable or contains
outliers, the perceptron might struggle to find
an optimal decision boundary or may produce
suboptimal results.

### 9. **Limitations in Real-World


Applications**

- **Limited Practical Use:** Because of its


limitations, the perceptron has relatively few
practical applications in its original form. Most
real-world problems are not linearly separable
and require more sophisticated models, such
as multi-layer perceptrons (MLPs) or support
vector machines (SVMs).

### Historical Criticism: Minsky and Papert's


Critique

- **Perceptrons Book (1969):** In their book


*Perceptrons*, Marvin Minsky and Seymour
Papert highlighted many of the limitations of
the perceptron, particularly its inability to solve
the XOR problem. This critique led to a
temporary decline in research interest in neural
networks, a period sometimes referred to as
the "AI winter," until the development of multi-
layer networks and backpropagation in the
1980s.

### Conclusion

While the perceptron was groundbreaking for


its time and laid the groundwork for neural
network research, its limitations make it
inadequate for solving many real-world
problems. The main criticisms revolve around
its inability to handle non-linear relationships,
lack of hidden layers, and sensitivity to data
representation. These shortcomings led to the
development of more advanced neural network
models, which have addressed these issues
and enabled the modern deep learning
revolution.

CONCLUTION

The perceptron is an early model of a neural


network, introduced by Frank Rosenblatt in
1958. Here are the key points:

1. **Linear Classifier:** The perceptron is


designed to classify data that is linearly
separable, using a linear decision boundary.

2. **Single-Layer Model:** It consists of an


input layer connected directly to an output
neuron, without any hidden layers.
3. **Learning Algorithm:** The perceptron
adjusts its weights based on the error of its
predictions, using a simple rule that guarantees
convergence if the data is linearly separable.

4. **Limitations:** It cannot solve non-linearly


separable problems (like XOR), is limited to
binary classification, and uses a non-
differentiable activation function.

5. **Impact:** Despite its limitations, the


perceptron laid the foundation for modern
neural networks and inspired the development
of more complex models like multi-layer
perceptrons and deep learning architectures.

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy