Pro Tip: Hyperparameters can significantly impact your model’s performance. Experiment with different combinations to find what works best for your data.
1️⃣ Linear & Logistic Regression:
* L1/L2 Penalty: Regularization techniques to prevent overfitting.
* Solver: Optimization algorithm for finding model coefficients. * Class Weight (Logistic): Adjusts weights inversely proportional to class frequencies.
2️⃣ Naive Bayes:
* Alpha: Smoothing parameter to handle zero frequencies.
* Fit Prior: Option to learn class priors or not.
3️⃣ Decision Trees & Random Forests:
* Criterion: Metric to measure the quality of splits (e.g., Gini, Entropy).
* Max Depth: Maximum depth of the tree to control overfitting. * Min Sample Split: Minimum number of samples required to split a node.
4️⃣ Gradient Boosted Trees:
* Learning Rate: Shrinks contribution of each tree to prevent overfitting.
* N Estimators: Number of boosting stages to perform.
5️⃣ Principal Component Analysis:
* N Component: Number of principal components to keep.
* SVD Solver: Algorithm for computing the singular value decomposition.
6️⃣ K-Nearest Neighbor:
* N Neighbors: Number of neighbors to use for k-nearest neighbors. * Weights: Function to weight points (e.g., uniform, distance).
DATA MINING and MACHINE LEARNING. PREDICTIVE TECHNIQUES: ENSEMBLE METHODS, BOOSTING, BAGGING, RANDOM FOREST, DECISION TREES and REGRESSION TREES.: Examples with MATLAB