Lecture 9 PDF
Lecture 9 PDF
● Random Forests
Pruning
Random Forests
● Purpose
○ Improve prediction accuracy
● Principle
○ Encouraging diversity among the tree
● Definition
○ Collection of unpruned CARTs
○ Rule to combine individual tree decisions
● Solution: randomness
○ Bagging
○ Random decision trees (rCART)
Random Forests: Example
Bootstrapping
● L: original learning set composed of n samples
● Generate K learning sets Lk…
○ ... composed of m samples, m ≤ n,...
○ ... obtained by uniform sampling with replacement
from L
○ In consequences, Lk may contain repeated samples
0.632 Bootstrap
● m=n
● A particular training data has a probability of (1-1/n) of
not being picked
● Thus its probability of ending up in the test data (not
selected) is: