Random Forest - Learning Algorithm

Learning Algorithm

Each tree is constructed using the following algorithm:

  1. Let the number of training cases be N, and the number of variables in the classifier be M.
  2. We are told the number m of input variables to be used to determine the decision at a node of the tree; m should be much less than M.
  3. Choose a training set for this tree by choosing n times with replacement from all N available training cases (i.e. take a bootstrap sample). Use the rest of the cases to estimate the error of the tree, by predicting their classes.
  4. For each node of the tree, randomly choose m variables on which to base the decision at that node. Calculate the best split based on these m variables in the training set.
  5. Each tree is fully grown and not pruned (as may be done in constructing a normal tree classifier).

For prediction a new sample is pushed down the tree. It is assigned the label of the training sample in the terminal node it ends up in. This procedure is iterated over all trees in the ensemble, and the mode vote of all trees is reported as random forest prediction.

Read more about this topic:  Random Forest

Famous quotes containing the word learning:

    I think it is better to show love by meeting needs than to keep telling my son that I love him. Right now he is learning to tie his shoes. He is old enough, so even though it’s hard for him, sometimes I insist. But once in a while when I see he’s tired I still do it for him, and I have noticed that while I am tying his shoe, he says, “I love you, Mommy.” When he says, “I love you,” I know that he knows that he is loved.
    Anonymous Parent (20th century)