Nit ML Sugg
Nit ML Sugg
In PCA (Principal Component Analysis), eigenvectors are used to find new axes, and
eigenvalues tell us how much information (variance) is captured.
Why important: They help reduce data dimensions and focus on the most useful
features.
Keywords: direction, variance, PCA, dimensionality reduction, important features.
Logistic regression uses a sigmoid function to give results between 0 and 1 (as
probabilities).
When to use:
✅ No training time.
SVM tries to draw a line (or plane) that maximizes the gap (margin) between
classes.
o Builds many independent models and averages them (like Random Forest).
o Reduces variance.
Boosting:
o Builds models one by one, each correcting previous errors (like AdaBoost).
o Reduces bias.
Outliers:
o Can be removed or replaced (with mean/median).
Train the model with past data and test on new emails.
Keywords: supervised learning, labels, features, classification.
Solutions:
Two types: