0% found this document useful (0 votes)
10 views10 pages

Pattern Recognition 21BR551 MODULE 05 NOTES

The document contains notes for Module 05 on Pattern Recognition, focusing on Artificial Neural Networks (ANNs) and their various components, algorithms, and applications. It covers topics such as nets without hidden layers, backpropagation, Hopfield nets, Support Vector Machines, and concepts like risk minimization and VC dimension. The notes are prepared by Mr. Thanmay J S, Assistant Professor at Mysore University School of Engineering.

Uploaded by

Thanmay JS
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
10 views10 pages

Pattern Recognition 21BR551 MODULE 05 NOTES

The document contains notes for Module 05 on Pattern Recognition, focusing on Artificial Neural Networks (ANNs) and their various components, algorithms, and applications. It covers topics such as nets without hidden layers, backpropagation, Hopfield nets, Support Vector Machines, and concepts like risk minimization and VC dimension. The notes are prepared by Mr. Thanmay J S, Assistant Professor at Mysore University School of Engineering.

Uploaded by

Thanmay JS
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 10

Mysore University School of Engineering

8J99+QC7, Manasa Gangothiri, Mysuru, Karnataka 570006

Pattern Recognition
(21BR551)

MODULE 05 NOTES

Prepared by: Mr Thanmay J S, Assistant Professor, Bio-Medical & Robotics Engineering, UoM, SoE, Mysore 57006
Mysore University School of Engineering
8J99+QC7, Manasa Gangothiri, Mysuru, Karnataka 570006

Module 05: Course Content

5.0 Introduction to Artificial neural networks


5.1 Nets without hidden layers, examples
5.2 Sequential MSE algorithm
5.3 Steepest descent method, examples
5.4 Nets with hidden layers, examples
5.5 The back propagation algorithm
5.6 Hopfield nets, examples
5.7 Storage and retrieval algorithms
5.8 Support vector machines,
5.9 Support vector machine algorithms.
5.10 Risk minimization principles and the Concept of uniform Convergence
5.11 VC dimension

Prepared by: Mr Thanmay J S, Assistant Professor, Bio-Medical & Robotics Engineering, UoM, SoE, Mysore 57006
Mysore University School of Engineering
8J99+QC7, Manasa Gangothiri, Mysuru, Karnataka 570006
5.0 Introduction to Artificial Neural Networks (ANNs)
Artificial Neural Networks (ANNs) are computational models inspired by the human brain. They are
composed of nodes (also called neurons) that are interconnected by links, each with an associated weight.
ANNs are used to model complex patterns, data relationships, and to solve tasks such as classification,
regression, pattern recognition, and optimization.
Key Components of ANNs:
1. Neurons (Nodes): Basic units that process information. Each neuron receives inputs, processes them,
and produces an output.
2. Layers: ANNs consist of multiple layers:
o Input layer: Takes input data.
o Hidden layers: Process inputs from the input layer and pass the results to the output layer.
o Output layer: Produces the final result (prediction or classification).
3. Weights: Each connection between neurons has a weight that determines the influence of the input on
the output.
4. Bias: An additional parameter added to the weighted sum before applying the activation function,
allowing the network to fit data better.
5. Activation Function: Determines whether a neuron fires or not. Common functions include:

6. Training: The weights of the network are adjusted during training to minimize the error, typically
using backpropagation and optimization algorithms like gradient descent.

5.1 Nets without hidden layers


Nets without hidden layers, often referred to as linear models or single-layer networks, are neural networks
that consist only of an input layer and an output layer. These networks do not have intermediate (hidden) layers
between the input and output layers.

Prepared by: Mr Thanmay J S, Assistant Professor, Bio-Medical & Robotics Engineering, UoM, SoE, Mysore 57006
Mysore University School of Engineering
8J99+QC7, Manasa Gangothiri, Mysuru, Karnataka 570006
Key Features of Nets Without Hidden Layers:
1. Linear Transformation: A neural network without hidden layers applies a linear transformation to
the input data. It can be seen as a linear model that tries to map input data directly to output using
weights and biases.
2. Simple Function: These networks represent linear functions like 𝑦 = 𝑊 ⋅ 𝑥 + 𝑏, where:
o 𝑥 is the input vector,
o 𝑊 is the weight matrix,
o 𝑏 is the bias term,
o 𝑦 is the output vector.
3. No Non-Linearities: Without hidden layers, the network lacks the ability to model complex non-linear
relationships. Essentially, the output is a weighted sum of the input features.

Prepared by: Mr Thanmay J S, Assistant Professor, Bio-Medical & Robotics Engineering, UoM, SoE, Mysore 57006
Mysore University School of Engineering
8J99+QC7, Manasa Gangothiri, Mysuru, Karnataka 570006
5.2 Sequential MSE Algorithm
The Sequential MSE (Mean Squared Error) Algorithm is used for adjusting a model’s parameters (like
weights in a neural network) to minimize the error in predictions. The process works step by step, taking one
training example at a time and adjusting the model based on the error. The goal is to reduce the error by
iterating over the data multiple times.
Steps:
1. Pick a training example.
2. Calculate the prediction error (difference between predicted and actual values).
3. Adjust the model parameters using the error.
4. Repeat for each training example until the model improves.

5.3 Steepest Descent Method


The Steepest Descent method is used to minimize a function (like a loss function in machine learning). It
works by calculating the gradient (slope) at the current point and then moving in the opposite direction of that
gradient to minimize the function.
The Steepest Descent Method is a first-order optimization algorithm used to find the minimum of a function.
The method works by iteratively adjusting the parameters in the direction opposite to the gradient (the steepest
slope), aiming to minimize the function.

Prepared by: Mr Thanmay J S, Assistant Professor, Bio-Medical & Robotics Engineering, UoM, SoE, Mysore 57006
Mysore University School of Engineering
8J99+QC7, Manasa Gangothiri, Mysuru, Karnataka 570006

5.4 Nets with Hidden Layers


A neural network with hidden layers consists of an input layer, one or more hidden layers, and an output
layer. The hidden layers help the network learn complex patterns by transforming inputs before passing them
to the output layer.: In a neural network for digit recognition:
• Input Layer: Raw pixel data from an image.
• Hidden Layer: Neurons detect edges, corners, or more complex patterns.
• Output Layer: Classifies the image as a digit (e.g., 0, 1, 2).
Hidden layers allow the network to learn and make decisions based on the patterns in the data.

Prepared by: Mr Thanmay J S, Assistant Professor, Bio-Medical & Robotics Engineering, UoM, SoE, Mysore 57006
Mysore University School of Engineering
8J99+QC7, Manasa Gangothiri, Mysuru, Karnataka 570006
5.5 The Backpropagation Algorithm
Backpropagation is a key algorithm for training neural networks. It works by calculating the error at the
output and then propagating this error backward through the network to update the weights and minimize the
error.
Steps:
1. Forward pass: Input data is passed through the network, and predictions are made.
2. Calculate error: The error (difference between predicted and actual output) is calculated at the output
layer.
3. Backward pass: The error is propagated backward from the output layer to the input layer, using the
chain rule of calculus to compute gradients.
4. Update weights: Adjust the weights using the gradient descent method to minimize the error.
5. Repeat: This process is repeated for multiple epochs (iterations) over the training data until the
network learns.

5.6 Hopfield Nets


The Hopfield Neural Networks, invented by Dr John J. Hopfield consists of one layer of ‘n’ fully connected
recurrent neurons. It is generally used in performing auto-association and optimization tasks. It is calculated
using a converging interactive process and it generates a different response than our normal neural nets.
Structure & Architecture of Hopfield Network
• Each neuron has an inverting and a non-inverting output.
• Being fully connected, the output of each neuron is an input to all other neurons but not the self.
The below figure shows a sample representation of a Discrete Hopfield Neural Network architecture having
the following elements.

Prepared by: Mr Thanmay J S, Assistant Professor, Bio-Medical & Robotics Engineering, UoM, SoE, Mysore 57006
Mysore University School of Engineering
8J99+QC7, Manasa Gangothiri, Mysuru, Karnataka 570006
5.7 Storage and Retrieval (with algorithm steps)
Storage and Retrieval algorithms are used to store data and then retrieve it later, typically using a model like
Hopfield networks or associative memory.
Steps:
1. Storage:
o Present data patterns to the network.
o The network updates its internal states to store these
patterns.
2. Retrieval:
o Provide a noisy or partial input.
o The network computes the best match to the stored
patterns.
o Return the closest matching pattern.

5.8 Support Vector Machines (SVM)


A Support Vector Machine (SVM) is a supervised learning
algorithm used for classification. It works by finding the optimal
hyperplane that best separates data points into different classes. The
goal is to maximize the margin between the two classes.
Explanation: SVM is commonly used for tasks like text
classification, image classification, and more. It works by finding the best line (in 2D) or hyperplane (in higher
dimensions) that separates the classes while maximizing the distance from the closest points on either side.

5.9 Support Vector Machine Algorithm Steps


The SVM algorithm works by finding the optimal hyperplane that maximizes the margin between two classes
of data. Here's a simple step-by-step process:
Steps:
1. Input data: Collect labeled training data points (e.g., positive
and negative samples).
2. Find hyperplane: Calculate the hyperplane that best separates
the data points into different classes.
3. Maximize margin: Ensure the margin (distance between the hyperplane and the nearest data points)
is as wide as possible.
4. Solve optimization problem: Use a mathematical optimization technique (e.g., quadratic
programming) to solve for the optimal hyperplane.
5. Classify new data: After training, classify new data by determining which side of the hyperplane the
data point falls on.
These steps help SVM find the best possible model for classifying data into different categories.

Prepared by: Mr Thanmay J S, Assistant Professor, Bio-Medical & Robotics Engineering, UoM, SoE, Mysore 57006
Mysore University School of Engineering
8J99+QC7, Manasa Gangothiri, Mysuru, Karnataka 570006

5.10 Risk Minimization and Uniform Convergence


Risk Minimization involves choosing a model that minimizes the error on unseen data (generalization). The
goal is to make the model not only perform well on the training data but also on new data.
Uniform Convergence ensures that as the sample size grows, the error on the training data closely matches
the error on unseen data. This helps prevent overfitting, where a model performs well on training data but
poorly on new data.
Difference:
• Risk Minimization is about finding a model that works well on both training and unseen data.
• Uniform Convergence guarantees that, with more data, the model will improve and its error will
stabilize.
Risk Minimization Uniform Convergence

5.10 VC Dimension
The VC (Vapnik-Chervonenkis) Dimension measures the capacity of a model to learn different patterns. It
tells you the largest number of points a model can shatter (i.e., perfectly separate into different classes). A high
VC dimension means the model is complex and can handle more diverse patterns, but it also risks overfitting.
Example: VC dimension (h) for a set of function (F) is defined as the largest number of points that can be
shattered by F. VC dimension talks about the complexity of model. Here shattered means classified /
partitioned.

2 Class Classification

3 Class Classification

Prepared by: Mr Thanmay J S, Assistant Professor, Bio-Medical & Robotics Engineering, UoM, SoE, Mysore 57006
Mysore University School of Engineering
8J99+QC7, Manasa Gangothiri, Mysuru, Karnataka 570006
Modal Questions
3 to 5 Marks Questions
1) Explain Artificial neural networks and its main components
2) Explain Sequential MSE algorithm
3) Explain with a neat sketch Nets with hidden layers
4) Explain the back propagation algorithm
5) Describe Hopfield nets with a neat sketch
6) Explain the loop in Storage and retrieval algorithms
7) What are the significant features of Support vector machines,
8) Explain Risk minimization principles
9) Explain the Concept of uniform Convergence
10) Describe VC dimension for 2 class and 3 class classification

8 to 10 Marks Questions
1) Describe Nets without hidden layers with any two examples
Solved problems on Steepest descent method
2) Question: Solve the function with minimum 3 iteration using Steepest descent method

----------------------------------------------------------------------------------------------------------------------------------

Prepared by: Mr Thanmay J S, Assistant Professor, Bio-Medical & Robotics Engineering, UoM, SoE, Mysore 57006

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy