Professional Documents
Culture Documents
Random Forest
Random Forest
Learning
Prepared By:
Dr. Sara Sweidan
Random
forest
Trees
Random forest trees
• Random Forest is a popular machine learning algorithm that
belongs to the supervised learning technique. It can be used for
both Classification and Regression problems in ML. It is based on
the concept of ensemble learning, which is a process of
combining multiple classifiers to solve a complex problem and to
improve the performance of the model.
• It is a homogeneous weak
learners‘ model that learns
from each other
independently in parallel
and combines them for
determining the model
average.
Boosting
• is also a homogeneous
weak learners‘ model but
works differently from
Bagging. In this model,
learners learn sequentially
and adaptively to improve
model predictions of a
learning algorithm.
Bagging vs. Boosting
Bagging Boosting
Each model is built independently New models are influenced by the performance of
previously built models.
Aim to decrease variance, not bias. Aim to decrease bias, not variance
Different training data subsets are randomly drawn Every new subset contains the elements that were
with replacement from the entire training dataset. misclassified by previous models.
Bagging tries to solve the over-fitting problem. Boosting tries to reduce bias.
If the classifier is unstable (high variance), then If the classifier is stable and simple (high bias) then
apply bagging. apply boosting.
Underfitting &
overfitting
• Underfitting & overfitting
Underfitting & overfitting
• Variance
• Measures precision or specificity of the estimator
• Low variance implies the estimator does not change much as
the training set varies
underfit region
overfit region
underfit region
overfit region
•Models with too few parameters are inaccurate because
of a large bias (not enough flexibility).
Low Bias
Low Variance / / High Variance
High Bias