Forest
Forest
Forest
1 Introduction
The traditional machine learning algorithms usually give low classifier accuracy, and
easy got over-fitting .To improve the accuracy, many people research on the algorithm
of combining classifiers. Many scholar start the research on improve the classification
accuracy by means of combining classifiers. In 1996, Leo Breiman advanced Bagging
algorithm which is one of the early stage algorithm [1]. Amit and Geman define a large
number of geometric features and search over a random selection on these for the best
split at each note[2]. In 1998, Dietterich put forward the random split selection
theory[3]. At each node the split is randomly selected from the N best splits. Ho[4] has
done much study on “the random subspace” method which grows each tree by a random
selection of a subset of features. Breiman [5]generate new training sets by randomizing
the outputs in the original training set. Among these, the idea, in Amit and Geman’s
paper, influenced Breiman’s thinking about random forests.
Random forests are a combination machine learning algorithm. Which are
combined with a series of tree classifiers, each tree cast a unit vote for the most
popular class, then combining these results get the final sort result. RF posses high
classification accuracy, tolerate outliers and noise well and never got overfitting. RF
has been one of the most popular research methods in data mining area and
information to the biological field. In China there are little study on RF, so it is
necessary to systemic summarize the down to date theory and application about RF.
B. Liu, M. Ma, and J. Chang (Eds.): ICICA 2012, LNCS 7473, pp. 246–252, 2012.
© Springer-Verlag Berlin Heidelberg 2012
New Machine Learning Algorithm: Random Forest 247
identically distributed random vectors and each tree casts a unit vote for the most
popular class at input x .
This definition show RF is a combination of many tree-structure classifiers. In
Breiman’s RF model, every tree is planted on the basis of a training sample set and a
random variable, the random variable corresponding to the kth tree is denoted as Θk ,
between any two of these random variables are independent and identically
distributed, resulting in a classifier h (x,Θ k ) where x is the input vector. After k
times running ,we obtain classifiers sequence {h1 (x), h2 (x), hk (x)} ,and use these to
constitute more than one classification model system ,the final result of this system is
drown by ordinary majority vote, the decision function is
k
H ( x) = arg max I (hi ( x) = Y ) (1)
Y
i =1
The larger the margin value, the higher accuracy of the classification prediction, and
the more confidence in classification.