0% found this document useful (0 votes)
6 views2 pages

BAR Machine Learning Notes Part 2

BAR_Machine_learning_notes_Part_2

Uploaded by

surya
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
6 views2 pages

BAR Machine Learning Notes Part 2

BAR_Machine_learning_notes_Part_2

Uploaded by

surya
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 2

The goal of clustering is to reduce the intra-cluster variability and increase the inter-cluster

variability.

K-Means Clustering is an unsupervised learning algorithm that is used to solve clustering problems
in machine learning or data science.

It allows us to cluster the data into different groups and is a convenient way to discover the
categories of groups in the unlabeled dataset on its own without the need for any training.

It is a centroid-based algorithm, where each cluster is associated with a centroid. The main aim of
this algorithm is to minimize the sum of distances between the data point and their corresponding
clusters.

The working of the K-Means algorithm is explained in the below steps:

Step-1: Select the number K to decide the number of clusters.

Step-2: Select random K points or centroids. (It can be other from the input dataset).

Step-3: Assign each data point to their closest centroid, which will form the predefined K clusters.

Step-4: Calculate the variance and place a new centroid of each cluster.

Step-5: Repeat the third steps, which means reassign each datapoint to the new closest centroid of
each cluster.

Step-6: If any reassignment occurs, then go to step-4 else go to FINISH.

Step-7: The model is ready.

Bias variance trade-off

This is one of the most obvious trade-offs in machine learning. This trade -off is in relation to

the error of the model. The error of the model is a sum of the bias error and variance error.

We have two prominent scenarios derived from the famous bull’s eye framework

Underfitting- high bias error and low variance error

Overfitting- low bias error and high variance error

Every model will have some element of bias error as well as variance error. Hence, we have
the problem of generalization which is inevitable. The challenge is to have the optimal mix of bias

error and variance error that provides the least total error for the model. Thus minimize the problem

of generalization.

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy