0% found this document useful (0 votes)
16 views4 pages

Support Vector Machine

A support vector machine (SVM) is a supervised learning algorithm primarily used for binary classification tasks in machine learning. It identifies a hyperplane that best separates two classes of data, aiming for the largest margin between the hyperplane and the nearest data points. In cases where data is not linearly separable, SVM employs techniques like kernelling to map data into higher dimensions for effective classification.

Uploaded by

THIRUNEELAKANDAN
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
16 views4 pages

Support Vector Machine

A support vector machine (SVM) is a supervised learning algorithm primarily used for binary classification tasks in machine learning. It identifies a hyperplane that best separates two classes of data, aiming for the largest margin between the hyperplane and the nearest data points. In cases where data is not linearly separable, SVM employs techniques like kernelling to map data into higher dimensions for effective classification.

Uploaded by

THIRUNEELAKANDAN
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 4

SUPPORT VECTOR MACHINE

• A support vector machine (SVM) is a type of supervised learning algorithm used in


machine learning to solve classification and regression tasks; SVMs are mainly used
for solving binary classification problems, which require classifying the elements of a
data set into two groups.
• SVMs can be used for a variety of tasks, such as text classification, image
classification, spam detection, handwriting identification, gene expression analysis,
face detection, and anomaly detection
What is a hyperplane?
• As a simple example, for a classification task with only two features, you can think of
a hyperplane as a line that linearly separates and classifies a set of data.
• Intuitively, the further from the hyperplane our data points lie, the more confident we
are that they have been correctly classified. We therefore want our data points to be as
far away from the hyperplane as possible, while still being on the correct side of it.
• So when new testing data are added, whatever side of the hyperplane it lands will
decide the class that we assign to it.
How do we find the right hyperplane?
• How do we best segregate the two classes within the data?
• The distance between the hyperplane and the nearest data point from either set is
known as the margin. The goal is to choose a hyperplane with the greatest possible
margin between the hyperplane and any point within the training set, giving a greater
chance of new data being classified correctly. There will never be any data point
inside the margin.

But what happens when there is no clear hyperplane?


• Data are rarely ever as clean as our simple example above. A dataset will often look
more like the jumbled balls below which represent a linearly non separable dataset.

• In order to classify a dataset like the one above it’s necessary to move away from a 2d
view of the data to a 3d view. Explaining this is easiest with another simplified
example. Imagine that our two sets of colored balls above are sitting on a sheet and
this sheet is lifted suddenly, launching the balls into the air. While the balls are up in
the air, you use the sheet to separate them. This ‘lifting’ of the balls represents the
mapping of data into a higher dimension. This is known as kernelling.
How does it work? How can we identify the right hyper-plane?
• You need to remember a thumb rule to identify the right hyper-plane:
“Select the hyper-plane which segregates the two classes better”.
Identify the right hyperplane (Scenario-1):
• Here, we have three hyperplanes (A, B and C). Now, identify the right hyperplane to
classify star and circle.


• Hyperplane “B” has excellently performed this job.
Identify the right hyperplane (Scenario-2):
• Here, we have three hyperplanes (A, B and C) and all are segregating the classes well.
Now, how can we identify the right hyperplane?
Here, maximizing the distances between nearest data point (either class) and hyperplane
will help us to decide the right hyperplane.
Scenario-2
This distance is called as Margin. Let’s look at the below snapshot:
We can see that the margin for hyperplane C is high as compared to both A and B. Hence, we
name the right hyperplane as C. Another lightning reason for selecting the hyperplane with
higher margin is robustness. If we select a hyperplane having low margin then there is high
chance of missclassification.
Identify the right hyperplane (Scenario-3)

• Some of you may have selected the hyper-plane B as it has higher margin compared
to A. But, here is the catch, SVM selects the hyperplane which classifies the classes
accurately prior to maximizing margin. Here, hyperplane B has a classification error
and A has classified all correctly. Therefore, the right hyperplane is A.

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy