5 SVM
5 SVM
into classes so that we can easily put the new data point in the
called a hyperplane.
SVM chooses the extreme points/vectors
that help in creating the hyperplane. These
extreme cases are called as support
vectors, and hence algorithm is termed as
Support Vector Machine.
We will first train our model with lots of images of cats and dogs so that it can learn about
different features of cats and dogs, and then we test it with this strange creature.
So as support vector creates a decision boundary between these two data (cat and dog) and
choose extreme cases (support vectors), it will see the extreme case of cat and dog. On the
basis of the support vectors, it will classify it as a cat.
Types of SVM
Linear SVM: Linear SVM is used for linearly separable data, which means if a dataset can be
classified into two classes by using a single straight line, then such data is termed as linearly
separable data, and classifier is used called as Linear SVM classifier.
Non-linear SVM: Non-Linear SVM is used for non-linearly separated data, which means if a
dataset cannot be classified by using a straight line, then such data is termed as non-linear
data and classifier used is called as Non-linear SVM classifier.
Hyperplane and Support Vectors in the SVM algorithm:
Hyperplane: The hyperplane is the decision boundary used to separate data points of
different classes in a feature space. For linear classification, this is a linear equation
represented as wx+b=0
Support Vector Machine
Terminology
Hyperplane: The hyperplane is the decision boundary used to separate data points of
different classes in a feature space. For linear classification, this is a linear equation
represented as wx+b=0
Support Vectors: Support vectors are the closest data points to the hyperplane. These points are
critical in determining the hyperplane and the margin in Support Vector Machine (SVM)
Support Vector Machine
Terminology
Hyperplane: The hyperplane is the decision boundary used to separate data points of
different classes in a feature space. For linear classification, this is a linear equation
represented as wx+b=0
Support Vectors: Support vectors are the closest data points to the hyperplane. These points are
critical in determining the hyperplane and the margin in Support Vector Machine (SVM)
Margin:refers to the distance between the support vector and the hyperplane. The primary goal
of the SVM algorithm is to maximize this margin, as a wider margin typically results in better
classification performance.
Support Vector Machine
Terminology
Kernel: The kernel is a mathematical function used in SVM to map input data into a higher-dimensional feature
space. This allows the SVM to find a hyperplane in cases where data points are not linearly separable in the
original space. Common kernel functions include linear, polynomial, radial basis function (RBF), and sigmoid.
Hard Margin: A hard margin refers to the maximum-margin hyperplane that perfectly separates
the data points of different classes without any misclassifications.
Soft Margin: When data contains outliers or is not perfectly separable, SVM uses the soft
margin technique. This method introduces a slack variable for each data point to allow some
misclassifications
uation for the linear hyperplane can be written as:
uation for the linear hyperplane can be written as:
distance between a data point x_i and the decision boundary can be calculated as:
distance between a data point x_i and the decision boundary can be calculated as: