Decision Trees
Decision Trees
Decision Trees
Input data is called training data and has a known label or result such as spam/not-
spam or a stock price at a time.A model is prepared through a training process
where it is required to make predictions and is corrected when those predictions are
wrong. The training process continues until the model achieves a desired level of
accuracy on the training data.
Example algorithms include Logistic Regression and the Back Propagation Neural
Network.
Unsupervised Learning
Input data is not labelled and does not have a known result. A model is prepared by
deducing structures present in the input data. This may be to extract general rules.
It may through a mathematical process to systematically reduce redundancy, or it
may be to organize data by similarity.
2) CLUSTERING ALGORITHM
Clustering, like regression describes the class of problem and the class of
methods. Clustering methods are typically organized by the modelling
approaches such as centroid-based and hierarchal. All methods are
concerned with using the inherent structures in the data to best organize the
data into groups of maximum commonality.
k-Means
k-Medians
Expectation Maximisation (EM)
Hierarchical Clustering
4) FEATURE LEARNING
Feature learning is motivated by the fact that machine learning tasks such
as classification often require input that is mathematically and
computationally convenient to process. However, real-world data such as
images, video, and sensor measurement is usually complex, redundant, and
highly variable. Thus, it is necessary to discover useful features or
representations from raw data. Traditional hand-crafted features often
require expensive human labor and often rely on expert knowledge. Also,
they normally do not generalize well. This motivates the design of efficient
feature learning techniques, to automate and generalize this.